{"work":{"id":"54c15172-6304-4008-a3b6-c4cc0803c054","openalex_id":null,"doi":null,"arxiv_id":"1802.03426","raw_key":null,"title":"UMAP: Uniform Manifold Approximation and Projection for Dimension Reduction","authors":null,"authors_text":"Leland McInnes, John Healy, James Melville","year":2018,"venue":"stat.ML","abstract":"UMAP (Uniform Manifold Approximation and Projection) is a novel manifold learning technique for dimension reduction. UMAP is constructed from a theoretical framework based in Riemannian geometry and algebraic topology. The result is a practical scalable algorithm that applies to real world data. The UMAP algorithm is competitive with t-SNE for visualization quality, and arguably preserves more of the global structure with superior run time performance. Furthermore, UMAP has no computational restrictions on embedding dimension, making it viable as a general purpose dimension reduction technique for machine learning.","external_url":"https://arxiv.org/abs/1802.03426","cited_by_count":null,"metadata_source":"pith","metadata_fetched_at":"2026-05-14T20:37:57.983533+00:00","pith_arxiv_id":"1802.03426","created_at":"2026-05-09T04:15:08.587774+00:00","updated_at":"2026-05-14T20:37:57.983533+00:00","title_quality_ok":true,"display_title":"UMAP: Uniform Manifold Approximation and Projection for Dimension Reduction","render_title":"UMAP: Uniform Manifold Approximation and Projection for Dimension Reduction"},"hub":{"state":{"work_id":"54c15172-6304-4008-a3b6-c4cc0803c054","tier":"super_hub","tier_reason":"100+ Pith inbound or 10,000+ external citations","pith_inbound_count":111,"external_cited_by_count":null,"distinct_field_count":28,"first_pith_cited_at":"2022-08-23T23:37:14+00:00","last_pith_cited_at":"2026-05-13T11:00:12+00:00","author_build_status":"needed","summary_status":"needed","contexts_status":"needed","graph_status":"needed","ask_index_status":"needed","reader_status":"not_needed","recognition_status":"not_needed","updated_at":"2026-05-14T20:46:11.209973+00:00","tier_text":"super_hub"},"tier":"super_hub","role_counts":[],"polarity_counts":[],"runs":{"ask_index":{"job_type":"ask_index","status":"succeeded","result":{"title":"UMAP: Uniform Manifold Approximation and Projection for Dimension Reduction","claims":[{"claim_text":"UMAP (Uniform Manifold Approximation and Projection) is a novel manifold learning technique for dimension reduction. UMAP is constructed from a theoretical framework based in Riemannian geometry and algebraic topology. The result is a practical scalable algorithm that applies to real world data. The UMAP algorithm is competitive with t-SNE for visualization quality, and arguably preserves more of the global structure with superior run time performance. Furthermore, UMAP has no computational restrictions on embedding dimension, making it viable as a general purpose dimension reduction technique","claim_type":"abstract","evidence_strength":"source_metadata"}],"why_cited":"Pith tracks UMAP: Uniform Manifold Approximation and Projection for Dimension Reduction because it crossed a citation-hub threshold.","role_counts":[]},"error":null,"updated_at":"2026-05-14T02:04:08.467036+00:00"},"author_expand":{"job_type":"author_expand","status":"succeeded","result":{"authors_linked":[{"id":"d7e2e150-6ff3-472f-9631-9eb39e6a05cd","orcid":null,"display_name":"Leland McInnes"},{"id":"813f31af-25ce-485c-904a-9ac65e359a81","orcid":null,"display_name":"John Healy"},{"id":"ade686ff-2101-4fae-a9ab-ae7437b159ed","orcid":null,"display_name":"James Melville"}]},"error":null,"updated_at":"2026-05-14T02:04:03.292764+00:00"},"context_extract":{"job_type":"context_extract","status":"succeeded","result":{"enqueued_papers":25},"error":null,"updated_at":"2026-05-14T02:04:16.549275+00:00"},"graph_features":{"job_type":"graph_features","status":"succeeded","result":{"co_cited":[{"title":"Adam: A Method for Stochastic Optimization","work_id":"1910796d-9b52-4683-bf5c-de9632c1028b","shared_citers":7},{"title":"BERTopic: Neural topic modeling with a class-based TF-IDF procedure","work_id":"fe7000f5-9930-49ed-ba20-6975a338b713","shared_citers":6},{"title":"On the Opportunities and Risks of Foundation Models","work_id":"a18039e9-928d-47c9-a836-32656a71bf71","shared_citers":6},{"title":"Scaling Laws for Neural Language Models","work_id":"b7dd8749-9c45-4977-ab9b-64478dce1ae8","shared_citers":6},{"title":"The Llama 3 Herd of Models","work_id":"1549a635-88af-4ac1-acfe-51ae7bb53345","shared_citers":6},{"title":"An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale","work_id":"e96730e3-129b-4db6-b981-15ab7932e297","shared_citers":5},{"title":"Auto-Encoding Variational Bayes","work_id":"97d95295-30e1-42b4-bbf6-85f0fa4edb44","shared_citers":5},{"title":"BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding","work_id":"ed240a10-5b19-406c-baa5-30803f465785","shared_citers":5},{"title":"Fashion-MNIST: a Novel Image Dataset for Benchmarking Machine Learning Algorithms","work_id":"6714d44f-1b5e-4141-9450-ea09a7e724b0","shared_citers":5},{"title":"Decoupled Weight Decay Regularization","work_id":"07ef7360-d385-4033-83f7-8384a6325204","shared_citers":4},{"title":"GPT-4 Technical Report","work_id":"b928e041-6991-4c08-8c81-0359e4097c7b","shared_citers":4},{"title":"Jiarui Zhang, Ollie Liu, Tianyu Yu, Jinyi Hu, and Willie Neiswanger","work_id":"eb18b0c2-9ed0-4254-b208-425469f09e64","shared_citers":4},{"title":"Llama 2: Open Foundation and Fine-Tuned Chat Models","work_id":"68a5177f-d644-44c1-bd4f-4e5278c22f5d","shared_citers":4},{"title":"Mistral 7B","work_id":"eb5e1305-ad11-4875-ad8d-ad8b8f697599","shared_citers":4},{"title":"Proximal Policy Optimization Algorithms","work_id":"240c67fe-d14d-4520-91c1-38a4e272ca19","shared_citers":4},{"title":"R., Millman, K","work_id":"b05b154d-0381-4d1b-911f-1b35eb7a6768","shared_citers":4},{"title":"Attention Is All You Need","work_id":"baafb5a2-5272-43bc-932f-09fa9ffe5316","shared_citers":3},{"title":"Can a suit of armor conduct electricity? a new dataset for open book question answering","work_id":"efba341e-9173-426f-928c-b9aa410afbee","shared_citers":3},{"title":"DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models","work_id":"c5006563-f3ec-438a-9e35-b7b484f34828","shared_citers":3},{"title":"DeepSeek-R1: Incentivizing Reasoning Capability in LLMs via Reinforcement Learning","work_id":"e6b75ad5-2877-4168-97c8-710407094d20","shared_citers":3},{"title":"DINOv3","work_id":"c8b07deb-8fe7-4e18-9620-f3569d3529ce","shared_citers":3},{"title":"doi: 10.18653/v1/D19-1006","work_id":"d83bae09-5488-413f-8412-c6a3a78ae329","shared_citers":3},{"title":"Efficient Estimation of Word Representations in Vector Space","work_id":"59edaa01-a696-45b3-9a08-5eae777a799e","shared_citers":3},{"title":"Flow Straight and Fast: Learning to Generate and Transfer Data with Rectified Flow","work_id":"a1989e1b-d66d-4533-be3a-fb9c5fd62290","shared_citers":3}],"time_series":[{"n":2,"year":2022},{"n":2,"year":2024},{"n":105,"year":2026}],"dependency_candidates":[]},"error":null,"updated_at":"2026-05-14T02:04:03.377045+00:00"},"identity_refresh":{"job_type":"identity_refresh","status":"succeeded","result":{"items":[{"title":"Qwen3 Technical Report","outcome":"unchanged","work_id":"25a4e30c-1232-48e7-9925-02fa12ba7c9e","resolver":"local_arxiv","confidence":0.98,"old_work_id":"25a4e30c-1232-48e7-9925-02fa12ba7c9e"}],"counts":{"fixed":0,"merged":0,"unchanged":1,"quarantined":0,"needs_external_resolution":0},"errors":[],"attempted":1},"error":null,"updated_at":"2026-05-14T02:04:08.361717+00:00"},"role_polarity":{"job_type":"role_polarity","status":"succeeded","result":{"title":"UMAP: Uniform Manifold Approximation and Projection for Dimension Reduction","claims":[{"claim_text":"UMAP (Uniform Manifold Approximation and Projection) is a novel manifold learning technique for dimension reduction. UMAP is constructed from a theoretical framework based in Riemannian geometry and algebraic topology. The result is a practical scalable algorithm that applies to real world data. The UMAP algorithm is competitive with t-SNE for visualization quality, and arguably preserves more of the global structure with superior run time performance. Furthermore, UMAP has no computational restrictions on embedding dimension, making it viable as a general purpose dimension reduction technique","claim_type":"abstract","evidence_strength":"source_metadata"}],"why_cited":"Pith tracks UMAP: Uniform Manifold Approximation and Projection for Dimension Reduction because it crossed a citation-hub threshold.","role_counts":[]},"error":null,"updated_at":"2026-05-14T02:04:08.369189+00:00"},"summary_claims":{"job_type":"summary_claims","status":"succeeded","result":{"title":"UMAP: Uniform Manifold Approximation and Projection for Dimension Reduction","claims":[{"claim_text":"UMAP (Uniform Manifold Approximation and Projection) is a novel manifold learning technique for dimension reduction. UMAP is constructed from a theoretical framework based in Riemannian geometry and algebraic topology. The result is a practical scalable algorithm that applies to real world data. The UMAP algorithm is competitive with t-SNE for visualization quality, and arguably preserves more of the global structure with superior run time performance. Furthermore, UMAP has no computational restrictions on embedding dimension, making it viable as a general purpose dimension reduction technique","claim_type":"abstract","evidence_strength":"source_metadata"}],"why_cited":"Pith tracks UMAP: Uniform Manifold Approximation and Projection for Dimension Reduction because it crossed a citation-hub threshold.","role_counts":[]},"error":null,"updated_at":"2026-05-14T02:04:08.378224+00:00"}},"summary":{"title":"UMAP: Uniform Manifold Approximation and Projection for Dimension Reduction","claims":[{"claim_text":"UMAP (Uniform Manifold Approximation and Projection) is a novel manifold learning technique for dimension reduction. UMAP is constructed from a theoretical framework based in Riemannian geometry and algebraic topology. The result is a practical scalable algorithm that applies to real world data. The UMAP algorithm is competitive with t-SNE for visualization quality, and arguably preserves more of the global structure with superior run time performance. Furthermore, UMAP has no computational restrictions on embedding dimension, making it viable as a general purpose dimension reduction technique","claim_type":"abstract","evidence_strength":"source_metadata"}],"why_cited":"Pith tracks UMAP: Uniform Manifold Approximation and Projection for Dimension Reduction because it crossed a citation-hub threshold.","role_counts":[]},"graph":{"co_cited":[{"title":"Adam: A Method for Stochastic Optimization","work_id":"1910796d-9b52-4683-bf5c-de9632c1028b","shared_citers":7},{"title":"BERTopic: Neural topic modeling with a class-based TF-IDF procedure","work_id":"fe7000f5-9930-49ed-ba20-6975a338b713","shared_citers":6},{"title":"On the Opportunities and Risks of Foundation Models","work_id":"a18039e9-928d-47c9-a836-32656a71bf71","shared_citers":6},{"title":"Scaling Laws for Neural Language Models","work_id":"b7dd8749-9c45-4977-ab9b-64478dce1ae8","shared_citers":6},{"title":"The Llama 3 Herd of Models","work_id":"1549a635-88af-4ac1-acfe-51ae7bb53345","shared_citers":6},{"title":"An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale","work_id":"e96730e3-129b-4db6-b981-15ab7932e297","shared_citers":5},{"title":"Auto-Encoding Variational Bayes","work_id":"97d95295-30e1-42b4-bbf6-85f0fa4edb44","shared_citers":5},{"title":"BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding","work_id":"ed240a10-5b19-406c-baa5-30803f465785","shared_citers":5},{"title":"Fashion-MNIST: a Novel Image Dataset for Benchmarking Machine Learning Algorithms","work_id":"6714d44f-1b5e-4141-9450-ea09a7e724b0","shared_citers":5},{"title":"Decoupled Weight Decay Regularization","work_id":"07ef7360-d385-4033-83f7-8384a6325204","shared_citers":4},{"title":"GPT-4 Technical Report","work_id":"b928e041-6991-4c08-8c81-0359e4097c7b","shared_citers":4},{"title":"Jiarui Zhang, Ollie Liu, Tianyu Yu, Jinyi Hu, and Willie Neiswanger","work_id":"eb18b0c2-9ed0-4254-b208-425469f09e64","shared_citers":4},{"title":"Llama 2: Open Foundation and Fine-Tuned Chat Models","work_id":"68a5177f-d644-44c1-bd4f-4e5278c22f5d","shared_citers":4},{"title":"Mistral 7B","work_id":"eb5e1305-ad11-4875-ad8d-ad8b8f697599","shared_citers":4},{"title":"Proximal Policy Optimization Algorithms","work_id":"240c67fe-d14d-4520-91c1-38a4e272ca19","shared_citers":4},{"title":"R., Millman, K","work_id":"b05b154d-0381-4d1b-911f-1b35eb7a6768","shared_citers":4},{"title":"Attention Is All You Need","work_id":"baafb5a2-5272-43bc-932f-09fa9ffe5316","shared_citers":3},{"title":"Can a suit of armor conduct electricity? a new dataset for open book question answering","work_id":"efba341e-9173-426f-928c-b9aa410afbee","shared_citers":3},{"title":"DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models","work_id":"c5006563-f3ec-438a-9e35-b7b484f34828","shared_citers":3},{"title":"DeepSeek-R1: Incentivizing Reasoning Capability in LLMs via Reinforcement Learning","work_id":"e6b75ad5-2877-4168-97c8-710407094d20","shared_citers":3},{"title":"DINOv3","work_id":"c8b07deb-8fe7-4e18-9620-f3569d3529ce","shared_citers":3},{"title":"doi: 10.18653/v1/D19-1006","work_id":"d83bae09-5488-413f-8412-c6a3a78ae329","shared_citers":3},{"title":"Efficient Estimation of Word Representations in Vector Space","work_id":"59edaa01-a696-45b3-9a08-5eae777a799e","shared_citers":3},{"title":"Flow Straight and Fast: Learning to Generate and Transfer Data with Rectified Flow","work_id":"a1989e1b-d66d-4533-be3a-fb9c5fd62290","shared_citers":3}],"time_series":[{"n":2,"year":2022},{"n":2,"year":2024},{"n":105,"year":2026}],"dependency_candidates":[]},"authors":[{"id":"ade686ff-2101-4fae-a9ab-ae7437b159ed","orcid":null,"display_name":"James Melville","source":"manual","import_confidence":0.72},{"id":"813f31af-25ce-485c-904a-9ac65e359a81","orcid":null,"display_name":"John Healy","source":"manual","import_confidence":0.72},{"id":"d7e2e150-6ff3-472f-9631-9eb39e6a05cd","orcid":null,"display_name":"Leland McInnes","source":"manual","import_confidence":0.72}]}}