pith. machine review for the scientific record. sign in

arxiv: 2605.05787 · v1 · submitted 2026-05-07 · 💻 cs.DB

Recognition: unknown

Low-Latency Out-of-Core ANN Search in High-Dimensional Space

Bin Wang, Junhua Zhang, Xiaochun Yang, Ziwen Song

Authors on Pith no claims yet

Pith reviewed 2026-05-08 03:37 UTC · model grok-4.3

classification 💻 cs.DB
keywords approximate nearest neighborANN searchout-of-coredisk-memory hybridlow-latency searchgraph-based indexinglower bound filteringasynchronous I/O
0
0 comments X

The pith

SkipDisk achieves 63 percent of HNSW latency at 20 percent memory footprint for out-of-core high-dimensional ANN search.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper aims to resolve the tension between high memory use in graph-based in-memory ANN methods and high disk latency in out-of-core methods. It introduces SkipDisk, which keeps most data on disk but uses per-point pivots to create tight lower bounds on distances, an estimation filter to skip unneeded reads, three-level pruning to retain only key data in memory, and asynchronous I/O to overlap disk access with computation. A sympathetic reader cares because this hybrid design could let ANN scale to datasets far larger than available RAM while avoiding the slowdowns typical of pure disk approaches. Experiments claim the method reaches 85 percent of HNSW latency with roughly 10 percent memory or 63 percent latency with about 20 percent memory, all while preserving search quality.

Core claim

SkipDisk equips each point with its own dedicated pivot to strengthen the triangle inequality lower bound, then applies an estimation-based filter to eliminate most irrelevant disk accesses. A three-level pruning strategy keeps only informative data resident in memory, and neighbor nodes are stored in memory to enable asynchronous I/O that decouples the in-memory search from disk operations. These steps together produce search latency equal to 85 percent of HNSW at approximately 10 percent memory footprint or 63 percent of HNSW latency at around 20 percent memory footprint.

What carries the argument

Dedicated per-point pivot that tightens the triangle-inequality lower bound, paired with estimation-based filtering and three-level data pruning to control disk accesses and memory residency.

If this is right

  • Disk accesses drop sharply because the tightened lower bound filters out most irrelevant points before any read occurs.
  • Memory footprint falls to 10-20 percent of a pure in-memory graph while nearest-neighbor accuracy remains comparable.
  • Asynchronous I/O hides remaining disk latency by letting in-memory search continue while neighbor nodes are fetched.
  • The hybrid storage model scales ANN to datasets larger than available RAM without reverting to slow disk-only performance.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • If the per-point pivot choice works across many datasets, the same bound-tightening idea could improve filtering in other out-of-core indexes such as range or k-NN variants.
  • Keeping neighbor nodes in memory while pruning the rest of the graph to three levels may suggest selective caching policies for other I/O-bound database operators.
  • The latency gains might increase or diminish at higher dimensions; running controlled tests that vary dimensionality would test the method's robustness.

Load-bearing premise

The per-point pivot produces a lower bound tight enough that the filter removes most irrelevant disk accesses without dropping true neighbors and the three-level pruning retains enough information to keep search quality intact.

What would settle it

On a standard high-dimensional benchmark, measure that either recall falls more than a few percent below HNSW at the claimed memory levels or the observed disk access count fails to explain the reported latency reduction.

Figures

Figures reproduced from arXiv: 2605.05787 by Bin Wang, Junhua Zhang, Xiaochun Yang, Ziwen Song.

Figure 1
Figure 1. Figure 1: Illustration of memory footprint and search Latency view at source ↗
Figure 3
Figure 3. Figure 3: Latency of Trim on different parameters. view at source ↗
Figure 4
Figure 4. Figure 4: , is that very little data is actually filtered out. Consequently, 90.0 92.5 95.0 97.5 100.0 Recall(%) 5 10 Latency (ms) ARGILLA, K=20 90.0 92.5 95.0 97.5 100.0 Recall(%) 2.5 5.0 7.5 Latency (ms) BIGCODE, K=20 NO-Filtering Lower Bound 0.3 0.4 0.5 0.6 0.7 0.8 view at source ↗
Figure 5
Figure 5. Figure 5: Gorgeous analysis on different 𝜎. result requirements. Figure 5a illustrates the performance under dif￾ferent 𝜎 values. As 𝜎 decreases, latency gradually increases because more points need to be checked to compensate for the loss in preci￾sion, leading to a decline in overall performance. Figure 5b further demonstrates this trade-off by showing that a larger 𝐿 (checking more points during the search) is re… view at source ↗
Figure 6
Figure 6. Figure 6: Workflow of SkipDisk. We follow these steps: view at source ↗
Figure 7
Figure 7. Figure 7: Dimensional-Level pruning for a vector view at source ↗
Figure 8
Figure 8. Figure 8: Lower bound. The distance of point 𝑏 to 𝑞 exceeds the threshold 𝜏. The lower bound 𝐿𝐵1 (𝑞, 𝑏) calculated using the pivot point 𝑙 designed for points 𝑎 and 𝑏 is below 𝜏, so point 𝑏 cannot be filtered out. By designing a closer pivot point𝑙𝑏 specifically for point 𝑏, the new lower bound 𝐿𝐵2 (𝑞, 𝑏) exceeds 𝜏, thus filtering out point 𝑏. The triangle inequality is a widely used principle in index de￾sign [72].… view at source ↗
Figure 9
Figure 9. Figure 9: Bit-Level pruning for a float. Instead of finding a single pivot point to represent a group of points, we design a dedicated pivot point for each individ￾ual point. Why does this approach help improve our performance? First, this approach aims to provide a tighter lower bound for filter￾ing. From the triangle inequality, to obtain a tighter lower bound, we want |𝛿 (𝑝,𝑙) − 𝛿 (𝑞,𝑙)| to be as close as possibl… view at source ↗
Figure 10
Figure 10. Figure 10: Overall Performance. Our method achieves lower latency compared to disk-based approaches. Compared to HNSW, view at source ↗
Figure 11
Figure 11. Figure 11: Illustration of search latency and memory foot view at source ↗
Figure 14
Figure 14. Figure 14: Tightness of lower bound. 0.10 0.20 0.30 0.40 0.50 0.60 Sub-dimensions 0 100 200 Number of Candidates 35.9% 49.3% 85.9% 81.4% 77.0% 70.9% 64.1% 50.7% BIGCODE Recall@99%,K=20 Filtered I/O view at source ↗
Figure 15
Figure 15. Figure 15: I/O reduction on estimation based filtering. view at source ↗
Figure 13
Figure 13. Figure 13: Disk I/O reduction and memory footprint on dif view at source ↗
Figure 18
Figure 18. Figure 18: Latency of SkipDisk-Trim. 6.10 Extend with Trim Given that the Trim method can filter candidate points, we extend our approach to incorporate Trim’s p-relax lower bound method for filtering and evaluate its performance. We compare this exten￾sion with our other variants on the bigcode and argilla datasets, setting the gamma parameter to 0.6 and 0.35, respectively. Fig￾ure 18 illustrates the performance co… view at source ↗
Figure 16
Figure 16. Figure 16: illustrates the latency of our method under different top-𝐾 values. As the top-𝐾 value increases, the search becomes more challenging because more relevant items need to be returned to meet recall requirements, leading to more disk accesses and increased search latency. At top-𝐾=100, our method still achieves lower latency than HNSW. However, as 𝐾 increases to 500, our latency exceeds that of HNSW. Despit… view at source ↗
Figure 17
Figure 17. Figure 17: Latency of PipeANN with caching data points in view at source ↗
Figure 20
Figure 20. Figure 20: Time distribution of PipeANN and SkipDisk-PB view at source ↗
read the original abstract

In-memory graph-based approximate nearest neighbor (ANN) search has superior search performance but incurs significant memory footprint. Disk-based methods reduce memory usage but suffer from high disk access latency. A common challenge is how to achieve low-latency search while significantly reducing memory footprint. In this paper, we propose SkipDisk, a disk-memory hybrid ANN search that significantly reduces memory footprint while achieving search latency comparable to or lower than in-memory method HNSW. By analyzing existing disk-based methods, we observed that disk access remains the primary bottleneck, and existing lower bound based filtering methods are two loose to effectively reduce disk access. Therefore, we design SkipDisk to achieve tight lower bound with low memory footprint to reduce the search latency. First, we design a dedicated pivot for each point to improve the lower bound of the triangle inequality for effective filtering. We further design an estimation-based approach based on this lower bound. Second, to reduce the memory footprint, we employ a three-level data pruning strategy to preserve informative data in memory. Third, to further reduce search latency, we design an asynchronous I/O strategy based on the decoupling of in-memory search and disk access by storing neighbor nodes in memory. Experiments show that our method achieves a latency of 85 of HNSW's latency with approximately 10 memory footprint, and a latency to 63 of HNSW's with a slightly higher memory footprint of around 20.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

3 major / 2 minor

Summary. The manuscript introduces SkipDisk, a disk-memory hybrid ANN search method for high-dimensional data. It uses a dedicated per-point pivot to tighten triangle-inequality lower bounds, an estimation-based filter to prune disk accesses, a three-level pruning strategy to keep memory low, and asynchronous I/O that keeps neighbor nodes in RAM. The central claim is that the resulting system delivers 85% of HNSW latency at ~10% memory footprint and 63% of HNSW latency at ~20% memory footprint.

Significance. If the latency and memory claims are substantiated with reproducible experiments, SkipDisk would address a practical bottleneck in large-scale ANN: how to keep search latency close to in-memory graph indexes while dropping the memory requirement by an order of magnitude. The per-point pivot plus estimation filter is a concrete, if incremental, idea for tightening bounds without storing full distance matrices.

major comments (3)
  1. [Abstract, §4] Abstract and §4 (Experiments): the headline performance numbers (85% and 63% of HNSW latency at 10% and 20% memory) are stated without any description of the datasets, their dimensionality, the recall@K target, the HNSW baseline parameters, or the hardware (SSD vs. HDD, page size, etc.). Without these, the central latency/memory trade-off cannot be evaluated or reproduced.
  2. [§3.1] §3.1 (Dedicated pivot): the claim that a per-point pivot yields a sufficiently tight lower bound for the estimation-based filter to remove most irrelevant disk accesses rests on an untested assumption. No measurements of bound tightness (e.g., average ratio of lower bound to true distance), filter false-negative rate, or actual disk-I/O reduction factor are reported, leaving the method vulnerable to the well-known distance-concentration problem in high dimensions.
  3. [§3.2] §3.2 (Three-level pruning): the strategy is described at a high level, but no analysis shows how much information is retained versus discarded at each level, nor any guarantee that true neighbors survive the pruning with high probability. This directly affects whether the 10-20% memory figures can be achieved without quality loss.
minor comments (2)
  1. [Abstract] Abstract contains several typographical omissions: '85 of HNSW's latency' should read '85% of HNSW's latency'; 'a latency to 63 of HNSW's' should read '63% of HNSW's latency'; '10 memory footprint' and '20 memory footprint' are missing the percent sign.
  2. [§3] Notation for the per-point pivot and the estimation-based filter is introduced without a clear summary table or pseudocode, making the algorithmic description harder to follow than necessary.

Simulated Author's Rebuttal

3 responses · 0 unresolved

We thank the referee for the constructive feedback, which identifies key areas where additional details and analysis will strengthen the manuscript. We address each major comment below and commit to revisions that improve clarity and substantiation of the claims without altering the core contributions.

read point-by-point responses
  1. Referee: [Abstract, §4] Abstract and §4 (Experiments): the headline performance numbers (85% and 63% of HNSW latency at 10% and 20% memory) are stated without any description of the datasets, their dimensionality, the recall@K target, the HNSW baseline parameters, or the hardware (SSD vs. HDD, page size, etc.). Without these, the central latency/memory trade-off cannot be evaluated or reproduced.

    Authors: We agree that these details are essential for evaluation and reproducibility. In the revised manuscript, we will expand the abstract to include a brief mention of the datasets (sizes and dimensionalities), target recall@K, and hardware context. Section 4 will be augmented with a dedicated experimental setup subsection specifying dataset characteristics, exact recall targets (e.g., recall@10), HNSW parameters (M and efConstruction values), and hardware details including SSD type, page size, and I/O configuration. This will directly enable reproduction of the reported latency/memory trade-offs. revision: yes

  2. Referee: [§3.1] §3.1 (Dedicated pivot): the claim that a per-point pivot yields a sufficiently tight lower bound for the estimation-based filter to remove most irrelevant disk accesses rests on an untested assumption. No measurements of bound tightness (e.g., average ratio of lower bound to true distance), filter false-negative rate, or actual disk-I/O reduction factor are reported, leaving the method vulnerable to the well-known distance-concentration problem in high dimensions.

    Authors: The original submission did not include explicit quantitative measurements of bound tightness or filter performance, which leaves the assumption insufficiently supported. While end-to-end results in §4 show effective disk-access reduction and maintained recall, we will add a new analysis subsection (or appendix) reporting the average lower-bound-to-true-distance ratio, estimation-filter false-negative rates, and measured disk-I/O reduction factors on the evaluated high-dimensional datasets. This will provide direct evidence that the dedicated pivot produces sufficiently tight bounds despite distance concentration. revision: yes

  3. Referee: [§3.2] §3.2 (Three-level pruning): the strategy is described at a high level, but no analysis shows how much information is retained versus discarded at each level, nor any guarantee that true neighbors survive the pruning with high probability. This directly affects whether the 10-20% memory figures can be achieved without quality loss.

    Authors: We acknowledge that the high-level description of the three-level pruning lacks supporting analysis on retention rates and neighbor survival. In the revision, we will add empirical measurements quantifying the fraction of data retained/discarded at each pruning level and the empirical survival probability of true nearest neighbors. Additional experiments will demonstrate that recall is preserved at the reported 10-20% memory footprints, and we will discuss any probabilistic guarantees derivable from the estimation filter. These additions will substantiate that quality is maintained. revision: yes

Circularity Check

0 steps flagged

No significant circularity; performance claims are experimental comparisons

full rationale

The paper introduces SkipDisk with three design elements (per-point pivots for tighter triangle-inequality bounds, estimation-based filtering, three-level pruning, and asynchronous I/O) and validates them via direct latency and memory measurements against HNSW. No equations, fitted parameters renamed as predictions, or self-citation chains appear in the provided text; the central claims reduce to empirical results rather than any self-referential derivation or ansatz smuggled through prior work. The reader's assessment of score 2.0 aligns with the absence of load-bearing self-referential steps.

Axiom & Free-Parameter Ledger

1 free parameters · 1 axioms · 1 invented entities

The approach rests on the triangle inequality for distance lower bounds and on the empirical effectiveness of the proposed pruning and pivot choices; no external benchmarks or formal proofs are referenced in the abstract.

free parameters (1)
  • pruning thresholds
    Three-level data pruning strategy requires thresholds to decide which data to retain in memory; values are not specified in the abstract.
axioms (1)
  • standard math Triangle inequality holds for the distance metric used
    Invoked to justify the lower-bound filtering step.
invented entities (1)
  • Dedicated pivot per point no independent evidence
    purpose: To tighten the lower bound of the triangle inequality for filtering
    New per-point pivot design introduced to improve filtering effectiveness.

pith-pipeline@v0.9.0 · 5554 in / 1379 out tokens · 48316 ms · 2026-05-08T03:37:09.409578+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

73 extracted references · 28 canonical work pages

  1. [1]

    2026.Anserini: OpenAI-ada2 Embeddings for MS MARCO Passage.Retrieved April 17, 2026 from https://github.com/castorini/anserini/blob/master/docs/ experiments-msmarco-passage-openai-ada2.md?utm_source=chatgpt.com

  2. [2]

    2026.argilla-warehouse/personahub-fineweb-edu-4-embeddings.Retrieved April 17, 2026 from https://huggingface.co/datasets/argilla-warehouse/personahub- fineweb-edu-4-embeddings

  3. [3]

    Retrieved April 17, 2026 from https://github.com/axboe/ liburing

    2026.axboe/liburing. Retrieved April 17, 2026 from https://github.com/axboe/ liburing

  4. [4]

    2026.bigcode/stack-exchange-embeddings-20230914.Retrieved April 17, 2026 from https://huggingface.co/datasets/bigcode/stack-exchange-embeddings- 20230914

  5. [5]

    2026.CohereLabs/msmarco-v2.1-embed-english-v3.Retrieved April 17, 2026 from https://huggingface.co/datasets/CohereLabs/msmarco-v2.1-embed-english-v3

  6. [6]

    Retrieved April 17, 2026 from https://huggingface.co/

    2026.Hugging Face. Retrieved April 17, 2026 from https://huggingface.co/

  7. [7]

    Retrieved April 17, 2026 from https://en.wikipedia.org/wiki/ IEEE_754

    2026.IEEE 754. Retrieved April 17, 2026 from https://en.wikipedia.org/wiki/ IEEE_754

  8. [8]

    Alexandr Andoni and Piotr Indyk. 2008. Near-optimal hashing algorithms for approximate nearest neighbor in high dimensions.Commun. ACM51, 1 (2008), 117–122

  9. [9]

    Alexandr Andoni and Ilya Razenshteyn. 2015. Optimal data-dependent hashing for approximate near neighbors. InProceedings of the forty-seventh annual ACM symposium on Theory of computing. 793–801

  10. [10]

    Fabien André, Anne-Marie Kermarrec, and Nicolas Le Scouarnec. 2015. Cache locality is not enough: High-Performance Nearest Neighbor Search with Product Quantization Fast Scan.Proc. VLDB Endow.9, 4 (2015), 288–299

  11. [11]

    Ilias Azizi, Karima Echihabi, and Themis Palpanas. 2025. Graph-Based Vector Search: An Experimental Evaluation of the State-of-the-Art.Proc. ACM Manag. Data3, 1, Article 43 (Feb. 2025), 31 pages. https://doi.org/10.1145/3709693 Ziwen Song, Bin Wang, Xiaochun Yang, and Junhua Zhang

  12. [12]

    Jon Louis Bentley. 1975. Multidimensional Binary Search Trees Used for Asso- ciative Searching.Commun. ACM18, 9 (1975), 509–517. https://doi.org/10.1145/ 361002.361007

  13. [13]

    Kakade, and John Langford

    Alina Beygelzimer, Sham M. Kakade, and John Langford. 2006. Cover trees for nearest neighbor. InMachine Learning, Proceedings of the Twenty-Third In- ternational Conference (ICML 2006), Pittsburgh, Pennsylvania, USA, June 25-29, 2006 (ACM International Conference Proceeding Series), William W. Cohen and Andrew W. Moore (Eds.), Vol. 148. ACM, 97–104

  14. [14]

    Cheng Chen, Chenzhe Jin, Yunan Zhang, Sasha Podolsky, Chun Wu, Szu- Po Wang, Eric Hanson, Zhou Sun, Robert Walzer, and Jianguo Wang. 2024. SingleStore-V: An Integrated Vector Database System in SingleStore.Proc. VLDB Endow.17, 12 (2024), 3772–3785. https://doi.org/10.14778/3685800.3685805

  15. [15]

    Chen, Wei-Cheng Chang, Jyun-Yu Jiang, Hsiang-Fu Yu, Inderjit S

    Patrick H. Chen, Wei-Cheng Chang, Jyun-Yu Jiang, Hsiang-Fu Yu, Inderjit S. Dhillon, and Cho-Jui Hsieh. 2023. FINGER: Fast Inference for Graph-based Approximate Nearest Neighbor Search. InProceedings of the ACM Web Conference 2023, WWW 2023, Austin, TX, USA, 30 April 2023 - 4 May 2023, Ying Ding, Jie Tang, Juan F. Sequeda, Lora Aroyo, Carlos Castillo, and ...

  16. [16]

    Qi Chen, Bing Zhao, Haidong Wang, Mingqin Li, Chuanjie Liu, Zengzhong Li, Mao Yang, and Jingdong Wang. 2021. SPANN: Highly-efficient Billion-scale Approximate Nearest Neighborhood Search. InAdvances in Neural Information Processing Systems 34: Annual Conference on Neural Information Processing Sys- tems 2021, NeurIPS 2021, December 6-14, 2021, virtual, Ma...

  17. [17]

    Weijian Chen, Haotian Liu, Yangshen Deng, Long Xiang, Liang Huang, Gezi Li, and Bo Tang. 2026. AlayaLaser: Efficient Index Layout and Search Strategy for Large-scale High-dimensional Vector Similarity Search.CoRRabs/2602.23342 (2026). https://doi.org/10.48550/ARXIV.2602.23342 arXiv:2602.23342

  18. [18]

    Paolo Ciaccia, Marco Patella, and Pavel Zezula. 1997. M-tree: An Efficient Ac- cess Method for Similarity Search in Metric Spaces. InVLDB’97, Proceedings of 23rd International Conference on Very Large Data Bases, August 25-29, 1997, Athens, Greece, Matthias Jarke, Michael J. Carey, Klaus R. Dittrich, Frederick H. Lochovsky, Pericles Loucopoulos, and Manfr...

  19. [19]

    Mirrokni

    Mayur Datar, Nicole Immorlica, Piotr Indyk, and Vahab S. Mirrokni. 2004. Locality-sensitive hashing scheme based on p-stable distributions. InProceedings of the 20th ACM Symposium on Computational Geometry, Brooklyn, New York, USA, June 8-11, 2004, Jack Snoeyink and Jean-Daniel Boissonnat (Eds.). ACM, 253–262

  20. [20]

    Liwei Deng, Penghao Chen, Ximu Zeng, Tianfu Wang, Yan Zhao, and Kai Zheng. 2024. Efficient Data-aware Distance Comparison Operations for High- Dimensional Approximate Nearest Neighbor Search.Proc. VLDB Endow.18, 3 (2024), 812–821. https://doi.org/10.14778/3712221.3712244

  21. [21]

    Wei Dong, Moses Charikar, and Kai Li. 2011. Efficient k-nearest neighbor graph construction for generic similarity measures. InProceedings of the 20th Interna- tional Conference on World Wide Web, WWW 2011, Hyderabad, India, March 28 - April 1, 2011, Sadagopan Srinivasan, Krithi Ramamritham, Arun Kumar, M. P. Ravindra, Elisa Bertino, and Ravi Kumar (Eds.)...

  22. [22]

    Razenshteyn, and Tal Wagner

    Yihe Dong, Piotr Indyk, Ilya P. Razenshteyn, and Tal Wagner. 2020. Learning Space Partitions for Nearest Neighbor Search. In8th International Conference on Learning Representations, ICLR 2020, Addis Ababa, Ethiopia, April 26-30, 2020. OpenReview.net. https://openreview.net/forum?id=rkenmREFDr

  23. [23]

    Wenqi Fan, Yujuan Ding, Liangbo Ning, Shijie Wang, Hengyun Li, Dawei Yin, Tat-Seng Chua, and Qing Li. 2024. A Survey on RAG Meeting LLMs: To- wards Retrieval-Augmented Large Language Models. InProceedings of the 30th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, KDD 2024, Barcelona, Spain, August 25-29, 2024, Ricardo Baeza-Yates and France...

  24. [24]

    Cong Fu, Changxu Wang, and Deng Cai. 2021. High dimensional similarity search with satellite system graph: Efficiency, scalability, and unindexed query compatibility.IEEE Transactions on Pattern Analysis and Machine Intelligence44, 8 (2021), 4139–4150

  25. [25]

    Cong Fu, Chao Xiang, Changxu Wang, and Deng Cai. 2019. Fast Approximate Nearest Neighbor Search With The Navigating Spreading-out Graph.Proc. VLDB Endow.12, 5 (2019), 461–474

  26. [26]

    Junhao Gan, Jianlin Feng, Qiong Fang, and Wilfred Ng. 2012. Locality-sensitive hashing scheme based on dynamic collision counting. InProceedings of the ACM SIGMOD International Conference on Management of Data, SIGMOD 2012, Scottsdale, AZ, USA, May 20-24, 2012, K. Selçuk Candan, Yi Chen, Richard T. Snodgrass, Luis Gravano, and Ariel Fuxman (Eds.). ACM, 541–552

  27. [27]

    Jianyang Gao and Cheng Long. 2023. High-Dimensional Approximate Nearest Neighbor Search: with Reliable and Efficient Distance Comparison Operations. Proc. ACM Manag. Data1, 2 (2023), 137:1–137:27

  28. [28]

    Jianyang Gao and Cheng Long. 2024. RaBitQ: Quantizing High-Dimensional Vectors with a Theoretical Error Bound for Approximate Nearest Neighbor Search.Proc. ACM Manag. Data2, 3, Article 167 (May 2024), 27 pages. https: //doi.org/10.1145/3654970

  29. [29]

    Tiezheng Ge, Kaiming He, Qifa Ke, and Jian Sun. 2013. Optimized Product Quantization for Approximate Nearest Neighbor Search. In2013 IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, June 23-28, 2013. IEEE Computer Society, 2946–2953

  30. [30]

    Hao Guo and Youyou Lu. 2025. Achieving Low-Latency Graph-Based Vector Search via Aligning Best-First Search Algorithm with SSD. In19th USENIX Symposium on Operating Systems Design and Implementation, OSDI 2025, Boston, MA, USA, July 7-9, 2025, Lidong Zhou and Yuanyuan Zhou (Eds.). USENIX Association, 171–186. https://www.usenix.org/conference/osdi25/prese...

  31. [31]

    Rentong Guo, Xiaofan Luan, Long Xiang, Xiao Yan, Xiaomeng Yi, Jigao Luo, Qianya Cheng, Weizhi Xu, Jiarui Luo, Frank Liu, Zhenshan Cao, Yanliang Qiao, Ting Wang, Bo Tang, and Charles Xie. 2022. Manu: A Cloud Native Vector Database Management System.Proc. VLDB Endow.15, 12 (2022), 3548–3561

  32. [32]

    Kiana Hajebi, Yasin Abbasi-Yadkori, Hossein Shahbazi, and Hong Zhang. 2011. Fast Approximate Nearest-Neighbor Search with k-Nearest Neighbor Graph. InIJCAI 2011, Proceedings of the 22nd International Joint Conference on Artifi- cial Intelligence, Barcelona, Catalonia, Spain, July 16-22, 2011, Toby Walsh (Ed.). IJCAI/AAAI, 1312–1317

  33. [33]

    Ben Harwood and Tom Drummond. 2016. FANNG: Fast Approximate Nearest Neighbour Graphs. In2016 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2016, Las Vegas, NV, USA, June 27-30, 2016. IEEE Computer Society, 5713–5722. https://doi.org/10.1109/CVPR.2016.616

  34. [34]

    Chengying Huan, Lizheng Chen, Zhengyi Yang, Shaonan Ma, Rong Gu, Renjie Yao, Zhibin Wang, Mingxing Zhang, Fang Xi, Jie Tao, Gang Zhang, Guihai Chen, and Chen Tian. 2025. OrchANN: A Unified I/O Orchestration Framework for Skewed Out-of-Core Vector Search.CoRRabs/2512.22838 (2025). https: //doi.org/10.48550/ARXIV.2512.22838 arXiv:2512.22838

  35. [35]

    Qiang Huang, Jianlin Feng, Yikai Zhang, Qiong Fang, and Wilfred Ng. 2015. Query-Aware Locality-Sensitive Hashing for Approximate Nearest Neighbor Search.Proc. VLDB Endow.9, 1 (2015), 1–12

  36. [36]

    Piotr Indyk and Rajeev Motwani. 1998. Approximate nearest neighbors: towards removing the curse of dimensionality. InProceedings of the thirtieth annual ACM symposium on Theory of computing. 604–613

  37. [37]

    Suhas Jayaram Subramanya, Fnu Devvrit, Harsha Vardhan Simhadri, Ravishankar Krishnawamy, and Rohan Kadekodi. 2019. Diskann: Fast accurate billion-point nearest neighbor search on a single node.Advances in Neural Information Processing Systems32 (2019)

  38. [38]

    Hervé Jégou, Romain Tavenard, Matthijs Douze, and Laurent Amsaleg. 2011. Searching in one billion vectors: Re-rank with source coding. InProceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2011, May 22-27, 2011, Prague Congress Center, Prague, Czech Republic. IEEE, 861–864

  39. [39]

    Haodi Jiang, Hao Guo, Minhui Xie, Jiwu Shu, and Youyou Lu. 2025. High- Throughput, Cost-Effective Billion-Scale Vector Search with a Single GPU.Proc. ACM Manag. Data3, 6 (2025), 1–27. https://doi.org/10.1145/3769799

  40. [40]

    Yannis Kalantidis and Yannis Avrithis. 2014. Locally Optimized Product Quan- tization for Approximate Nearest Neighbor Search. In2014 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2014, Columbus, OH, USA, June 23-28, 2014. IEEE Computer Society, 2329–2336

  41. [41]

    Dingyi Kang, Dongming Jiang, Hanshen Yang, Hang Liu, and Bingzhe Li. 2025. Scalable Disk-Based Approximate Nearest Neighbor Search with Page-Aligned Graph.CoRRabs/2509.25487 (2025). https://doi.org/10.48550/ARXIV.2509.25487 arXiv:2509.25487

  42. [42]

    Leonardo Kuffo, Elena Krippner, and Peter Boncz. 2025. PDX: A Data Layout for Vector Similarity Search.Proc. ACM Manag. Data3, 3, Article 196 (June 2025), 26 pages. https://doi.org/10.1145/3725333

  43. [43]

    Binhong Li, Xiao Yan, and Shangqi Lu. 2025. Fast-Convergent Proximity Graphs for Approximate Nearest Neighbor Search.CoRRabs/2510.05975 (2025). https: //doi.org/10.48550/ARXIV.2510.05975 arXiv:2510.05975

  44. [44]

    Hui Li, Shiyuan Deng, Xiao Yan, Xiangyu Zhi, and James Cheng. 2025. SAQ: Pushing the Limits of Vector Quantization through Code Adjustment and Di- mension Segmentation.Proc. ACM Manag. Data3, 6 (2025), 1–25. https: //doi.org/10.1145/3769824

  45. [45]

    Sen Li, Fuyu Lv, Taiwei Jin, Guli Lin, Keping Yang, Xiaoyi Zeng, Xiao-Ming Wu, and Qianli Ma. 2021. Embedding-based Product Retrieval in Taobao Search. In KDD ’21: The 27th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, Virtual Event, Singapore, August 14-18, 2021, Feida Zhu, Beng Chin Ooi, and Chunyan Miao (Eds.). ACM, 3181–3189. https://d...

  46. [46]

    Kejing Lu, Mineichi Kudo, Chuan Xiao, and Yoshiharu Ishikawa. 2021. HVS: Hi- erarchical Graph Structure Based on Voronoi Diagrams for Solving Approximate Nearest Neighbor Search.Proc. VLDB Endow.15, 2 (2021), 246–258

  47. [47]

    Kejing Lu, Chuan Xiao, and Yoshiharu Ishikawa. 2024. Probabilistic Routing for Graph-Based Approximate Nearest Neighbor Search. InForty-first International Conference on Machine Learning, ICML 2024, Vienna, Austria, July 21-27, 2024 (Proceedings of Machine Learning Research), Ruslan Salakhutdinov, Zico Kolter, Katherine A. Heller, Adrian Weller, Nuria Oli...

  48. [48]

    Yu A Malkov and Dmitry A Yashunin. 2018. Efficient and robust approximate nearest neighbor search using hierarchical navigable small world graphs.IEEE transactions on pattern analysis and machine intelligence42, 4 (2018), 824–836

  49. [49]

    James Jie Pan, Jianguo Wang, and Guoliang Li. 2024. Survey of vector database management systems.VLDB J.33, 5 (2024), 1591–1615. https://doi.org/10.1007/ S00778-024-00864-X

  50. [50]

    Yun Peng, Byron Choi, Tsz Nam Chan, Jianye Yang, and Jianliang Xu. 2023. Efficient Approximate Nearest Neighbor Search in Multi-dimensional Databases. Proc. ACM Manag. Data1, 1 (2023), 54:1–54:27

  51. [51]

    Chanop Silpa-Anan and Richard I. Hartley. 2008. Optimised KD-trees for fast im- age descriptor matching. In2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR 2008), 24-26 June 2008, Anchorage, Alaska, USA. IEEE Computer Society

  52. [52]

    Yitong Song, Pengcheng Zhang, Chao Gao, Bin Yao, Kai Wang, Zongyuan Wu, and Lin Qu. 2025. TRIM: Accelerating High-Dimensional Vector Similarity Search with Enhanced Triangle-Inequality-Based Pruning.Proc. ACM Manag. Data3, 6 (2025), 1–26. https://doi.org/10.1145/3769838

  53. [53]

    Ziwen Song, Bin Wang, and Xiaochun Yang. 2025. Accelerating High- Dimensional ANN Search via Skipping Redundant Distance Computations.Proc. ACM Manag. Data3, 6 (2025), 1–29. https://doi.org/10.1145/3769753

  54. [54]

    Yifang Sun, Wei Wang, Jianbin Qin, Ying Zhang, and Xuemin Lin. 2014. SRS: Solving c-Approximate Nearest Neighbor Queries in High Dimensional Euclidean Space with a Tiny Index.Proc. VLDB Endow.8, 1 (2014), 1–12. https://doi.org/10. 14778/2735461.2735462

  55. [55]

    Yufei Tao, Ke Yi, Cheng Sheng, and Panos Kalnis. 2009. Quality and efficiency in high dimensional nearest neighbor search. InProceedings of the ACM SIGMOD International Conference on Management of Data, SIGMOD 2009, Providence, Rhode Island, USA, June 29 - July 2, 2009, Ugur Çetintemel, Stanley B. Zdonik, Donald Kossmann, and Nesime Tatbul (Eds.). ACM, 563–576

  56. [56]

    Bing Tian, Haikun Liu, Yuhang Tang, Shihai Xiao, Zhuohui Duan, Xiaofei Liao, Hai Jin, Xuecang Zhang, Junhua Zhu, and Yu Zhang. 2025. Towards High-throughput and Low-latency Billion-scale Vector Search via CPU/GPU Collaborative Filtering and Re-ranking. In23rd USENIX Conference on File and Storage Technologies, FAST 2025, Santa Clara, CA, February 25-27, 2...

  57. [57]

    Jingdong Wang and Shipeng Li. 2012. Query-driven iterated neighborhood graph search for large scale indexing. InProceedings of the 20th ACM Multime- dia Conference, MM ’12, Nara, Japan, October 29 - November 02, 2012, Noboru Babaguchi, Kiyoharu Aizawa, John R. Smith, Shin’ichi Satoh, Thomas Plagemann, Xian-Sheng Hua, and Rong Yan (Eds.). ACM, 179–188

  58. [58]

    Jun Wang, Wei Liu, Sanjiv Kumar, and Shih-Fu Chang. 2016. Learning to Hash for Indexing Big Data - A Survey.Proc. IEEE104, 1 (2016), 34–57

  59. [59]

    Jianguo Wang, Xiaomeng Yi, Rentong Guo, Hai Jin, Peng Xu, Shengjun Li, Xi- angyu Wang, Xiangzhou Guo, Chengming Li, Xiaohai Xu, Kun Yu, Yuxing Yuan, Yinghao Zou, Jiquan Long, Yudong Cai, Zhenxiang Li, Zhifeng Zhang, Yihua Mo, Jun Gu, Ruiyi Jiang, Yi Wei, and Charles Xie. 2021. Milvus: A Purpose-Built Vector Data Management System. InSIGMOD ’21: Internatio...

  60. [60]

    Mengzhao Wang, Weizhi Xu, Xiaomeng Yi, Songlin Wu, Zhangyang Peng, Xi- angyu Ke, Yunjun Gao, Xiaoliang Xu, Rentong Guo, and Charles Xie. 2024. Starling: An I/O-Efficient Disk-Resident Graph Index Framework for High- Dimensional Vector Similarity Search on Data Segment.Proc. ACM Manag. Data2, 1 (2024), V2mod014:1–V2mod014:27

  61. [61]

    Mengzhao Wang, Xiaoliang Xu, Qiang Yue, and Yuxiang Wang. 2021. A Com- prehensive Survey and Experimental Comparison of Graph-Based Approximate Nearest Neighbor Search.Proc. VLDB Endow.14, 11 (2021), 1964–1978

  62. [62]

    Likang Wu, Zhi Zheng, Zhaopeng Qiu, Hao Wang, Hongchao Gu, Tingjia Shen, Chuan Qin, Chen Zhu, Hengshu Zhu, Qi Liu, Hui Xiong, and Enhong Chen

  63. [63]

    https://doi.org/10.1007/S11280-024-01291-2

    A survey on large language models for recommendation.World Wide Web (WWW)27, 5 (2024), 60. https://doi.org/10.1007/S11280-024-01291-2

  64. [64]

    Jiadong Xie, Jeffrey Xu Yu, and Yingfan Liu. 2025. Graph Based K-Nearest Neighbor Search Revisited.ACM Trans. Database Syst.50, 4 (2025), 14:1–14:30. https://doi.org/10.1145/3736716

  65. [65]

    Yuming Xu, Hengyu Liang, Jin Li, Shuotao Xu, Qi Chen, Qianxi Zhang, Cheng Li, Ziyue Yang, Fan Yang, Yuqing Yang, Peng Cheng, and Mao Yang. 2023. SPFresh: Incremental In-Place Update for Billion-Scale Vector Search. InPro- ceedings of the 29th Symposium on Operating Systems Principles, SOSP 2023, Koblenz, Germany, October 23-26, 2023, Jason Flinn, Margo I....

  66. [66]

    Mingyu Yang, Liuchang Jing, Wentao Li, and Wei Wang. 2024. Quantization Meets Projection: A Happy Marriage for Approximate k-Nearest Neighbor Search. arXiv preprint arXiv:2411.06158(2024)

  67. [67]

    Mingyu Yang, Wentao Li, Jiabao Jin, Xiaoyao Zhong, Xiangyu Wang, Zhitao Shen, Wei Jia, and Wei Wang. 2025. Effective and General Distance Computation for Approximate Nearest Neighbor Search. In41st IEEE International Conference on Data Engineering, ICDE 2025, Hong Kong, May 19-23, 2025. IEEE, 1098–1110. https://doi.org/10.1109/ICDE65448.2025.00087

  68. [68]

    Peiqi Yin, Xiao Yan, Qihui Zhou, Hui Li, Xiaolu Li, Lin Zhang, Meiling Wang, Xin Yao, and James Cheng. 2025. Gorgeous: Revisiting the Data Layout for Disk-Resident High-Dimensional Vector Search.arXiv preprint arXiv:2508.15290 (2025)

  69. [69]

    Peiqi Yin, Xiao Yan, Qihui Zhou, Hui Li, Xiaolu Li, Lin Zhang, Meiling Wang, Xin Yao, and James Cheng. 2025. Gorgeous: Revisiting the Data Layout for Disk-Resident High-Dimensional Vector Search.CoRRabs/2508.15290 (2025). https://doi.org/10.48550/ARXIV.2508.15290 arXiv:2508.15290

  70. [70]

    Weichen Zhao, Yuncheng Lu, Yao Tian, Hao Zhang, Jiehui Li, Minghao Zhao, Yakun Li, and Weining Qian. 2026. Optimizing SSD-Resident Graph Indexing for High-Throughput Vector Search.CoRRabs/2602.22805 (2026). https://doi.org/ 10.48550/ARXIV.2602.22805 arXiv:2602.22805

  71. [71]

    Bolong Zheng, Xi Zhao, Lianggui Weng, Quoc Viet Hung Nguyen, Hang Liu, and Christian S. Jensen. 2022. PM-LSH: a fast and accurate in-memory framework for high-dimensional approximate NN and closest pair search.VLDB J.31, 6 (2022), 1339–1363. https://doi.org/10.1007/S00778-021-00680-7

  72. [72]

    Yuxin Zheng, Qi Guo, Anthony K. H. Tung, and Sai Wu. 2016. LazyLSH: Approx- imate Nearest Neighbor Search for Multiple Distance Functions with a Single Index. InProceedings of the 2016 International Conference on Management of Data, SIGMOD Conference 2016, San Francisco, CA, USA, June 26 - July 01, 2016, Fatma Özcan, Georgia Koutrika, and Sam Madden (Eds....

  73. [73]

    Yifan Zhu, Lu Chen, Yunjun Gao, and Christian S. Jensen. 2022. Pivot selection algorithms in metric spaces: a survey and experimental study.VLDB J.31, 1 (2022), 23–47. https://doi.org/10.1007/S00778-021-00691-4