Recognition: unknown
GlintMarkers: Spatial Perception on XR Eyewear using Corneal Reflections
Pith reviewed 2026-05-10 14:26 UTC · model grok-4.3
The pith
GlintMarkers uses reflections from passive markers on the cornea to let XR eyewear cameras estimate object positions and orientations.
A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.
Core claim
The cornea functions as a compact mirror that encodes both the user's gaze direction and visual details of the surrounding scene. Passive retroreflective markers placed on objects concentrate reflected near-infrared light into distinct glint patterns on this mirror surface. These patterns are captured by the inward-facing camera and fed into a custom Perspective-n-Point estimation framework adapted for corneal geometry, yielding reliable measurements of object orientation, distance, and unique identification.
What carries the argument
The passive retroreflective marker design that produces concentrated bright glints on the cornea, paired with a custom Perspective-n-Point (PnP) estimation framework adapted to the geometry of corneal reflections.
If this is right
- XR eyewear can determine the three-dimensional position and orientation of tagged objects using only its existing inward-facing cameras.
- Multiple objects can be uniquely identified and tracked simultaneously through their distinct glint signatures.
- Spatial perception becomes possible without outward-facing cameras or active electronic tags on the environment.
- Gaze direction and environmental layout are recovered from the same corneal image stream.
Where Pith is reading between the lines
- The approach could be tested with untagged everyday objects by analyzing natural specular highlights instead of engineered markers.
- Integration with existing eye-tracking pipelines in commercial XR headsets would require only software changes rather than new hardware.
- The same glint data might support additional inferences such as surface material or rough shape once more sophisticated decoding is added.
- Privacy implications arise because the system records reflections of the user's immediate surroundings on the eye itself.
Load-bearing premise
The cornea must act as a sufficiently faithful mirror whose small, low-contrast reflections can be made bright enough by passive markers to support accurate spatial calculations with the limited resolution of an inward-facing camera.
What would settle it
A controlled experiment that measures the root-mean-square error of distance and orientation estimates against ground-truth motion-capture data while varying lighting conditions and marker distances would show whether the glint patterns contain enough geometric information for reliable PnP results.
Figures
read the original abstract
We present GlintMarkers, the first system to perform gaze-driven spatial perception using the inward-facing cameras on XR eyewear. Our key observation is that the cornea acts as a mirror that encodes both gaze direction and visual information about the environment in a small, low-contrast reflection. To extract spatial and semantic information from this reflection despite the camera's limited pixel budget, we present a passive retroreflective marker design that concentrates reflected near-infrared light onto the cornea, producing bright glint patterns. We develop a custom Perspective-n-Point (PnP) estimation framework adapted to corneal imaging and perform orientation and distance estimation of tagged objects, as well as unique object identification.
Editorial analysis
A structured set of objections, weighed in public.
Referee Report
Summary. The paper presents GlintMarkers, the first system for gaze-driven spatial perception on XR eyewear that repurposes inward-facing cameras to capture corneal reflections of passive retroreflective markers. It introduces a marker design to concentrate NIR light into bright glint patterns and a custom PnP framework to recover 6-DoF pose (orientation and distance) of tagged objects along with unique identification, despite the limited pixel budget of the corneal reflection.
Significance. If the core technical assumptions hold, the work would offer a hardware-minimal approach to environmental sensing in XR by exploiting existing eye-tracking cameras and corneal optics, potentially enabling new gaze-based spatial interactions. The novelty of the retroreflective marker design and corneal-adapted PnP is notable, but the lack of any reported error metrics or validation data substantially reduces the assessed significance at present.
major comments (3)
- [Abstract and §4 (PnP Framework)] The abstract and introduction claim reliable orientation and distance estimation via the custom PnP framework, yet no quantitative results (e.g., mean reprojection error, angular accuracy, or distance error) are provided anywhere in the manuscript to demonstrate that the solver converges on the low-resolution, low-contrast glints.
- [§3 (Marker Design)] §3 (Marker Design): The feasibility claim that retroreflective markers produce glint patterns bright enough and geometrically stable enough for PnP rests on an untested assumption about signal-to-noise ratio; the text acknowledges the limited pixel budget but supplies no pixel-occupancy measurements, SNR analysis, or ambient-light robustness tests.
- [§4 (PnP Framework) and §5 (Evaluation)] The central 6-DoF estimation claim depends on the cornea acting as a sufficiently spherical mirror whose radius and asphericity variations across users do not push detected correspondences outside the PnP basin of convergence; no user-variability study or calibration-drift analysis is reported.
minor comments (2)
- [§4] Notation for the custom PnP objective function and the retroreflective marker geometry should be defined more explicitly with equations rather than prose descriptions.
- [Figures 3-5] Figure captions for the glint-pattern examples would benefit from scale bars or pixel-count annotations to illustrate the limited resolution.
Simulated Author's Rebuttal
We thank the referee for the constructive feedback on our manuscript. The comments correctly identify gaps in quantitative validation and analysis that we will address through targeted revisions and additional experiments. Below we respond point by point to the major comments.
read point-by-point responses
-
Referee: [Abstract and §4 (PnP Framework)] The abstract and introduction claim reliable orientation and distance estimation via the custom PnP framework, yet no quantitative results (e.g., mean reprojection error, angular accuracy, or distance error) are provided anywhere in the manuscript to demonstrate that the solver converges on the low-resolution, low-contrast glints.
Authors: We agree that the manuscript would be strengthened by explicit quantitative error metrics. While §5 presents functional demonstrations of 6-DoF estimation and identification, it does not report numerical values such as mean reprojection error or angular/distance accuracy. In the revision we will add these metrics, computed from our existing experimental datasets, to §4 and §5 to directly substantiate convergence and reliability on the low-resolution glints. revision: yes
-
Referee: [§3 (Marker Design)] §3 (Marker Design): The feasibility claim that retroreflective markers produce glint patterns bright enough and geometrically stable enough for PnP rests on an untested assumption about signal-to-noise ratio; the text acknowledges the limited pixel budget but supplies no pixel-occupancy measurements, SNR analysis, or ambient-light robustness tests.
Authors: The referee is correct that §3 relies on optical principles without supporting measurements. We will revise §3 to include pixel-occupancy statistics for the glint patterns, quantitative SNR values under controlled and ambient lighting, and robustness test results. These additions will be drawn from supplementary optical characterization experiments we will perform and report. revision: yes
-
Referee: [§4 (PnP Framework) and §5 (Evaluation)] The central 6-DoF estimation claim depends on the cornea acting as a sufficiently spherical mirror whose radius and asphericity variations across users do not push detected correspondences outside the PnP basin of convergence; no user-variability study or calibration-drift analysis is reported.
Authors: We acknowledge the absence of a dedicated user-variability study. Our current evaluation uses a fixed corneal model and limited participant data. In the revision we will expand §5 with a multi-user study that measures the effects of corneal radius and asphericity variation on correspondence accuracy and PnP convergence, together with an analysis of calibration drift over time. This will be accompanied by updated discussion of model assumptions. revision: yes
Circularity Check
No significant circularity; derivation relies on external CV primitives and empirical validation
full rationale
The paper's core pipeline—retroreflective marker design to produce corneal glints, followed by an adapted PnP solver for 6-DoF pose—applies standard computer-vision techniques (Perspective-n-Point) to a new imaging modality. No equation or claim reduces by construction to a fitted parameter defined by the result itself, nor does any load-bearing step rest on a self-citation chain whose prior work is unverified. The abstract and described framework treat the cornea's reflective properties and marker concentration as physical observations to be validated experimentally, not as tautological inputs. The derivation is therefore self-contained against external benchmarks such as reprojection error on real corneal images and marker visibility measurements.
Axiom & Free-Parameter Ledger
Reference graph
Works this paper leans on
- [1]
-
[2]
Hadi Alzayer, Kevin Zhang, Brandon Feng, Christopher A Metzler, and Jia-Bin Huang. 2024. Seeing the world through your eyes. InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 4864–4873
2024
-
[3]
Apple. 2026. Apple Vision Pro. https://www.apple.com/apple-vision-pro/
2026
-
[4]
Ananta Narayanan Balaji, Clayton Kimber, David Li, Shengzhi Wu, Ruofei Du, and David Kim. 2023. RetroSphere: Self-Contained Passive 3D Controller Tracking for Augmented Reality.Proc. ACM Interact. Mob. Wearable Ubiquitous Technol.6, 4, Article 157 (Jan. 2023), 36 pages. doi:10.1145/3569479
-
[5]
Toby Collins and Adrien Bartoli. 2014. Infinitesimal plane-based pose estimation. International journal of computer vision109, 3 (2014), 252–286
2014
-
[6]
Mustafa Doga Dogan, Raul Garcia-Martin, Patrick William Haertel, Jamison John O’Keefe, Ahmad Taka, Akarsh Aurora, Raul Sanchez-Reillo, and Stefanie Mueller
-
[7]
BrightMarker: 3D Printed Fluorescent Markers for Object Tracking. In Proceedings of the 36th Annual ACM Symposium on User Interface Software and Technology(San Francisco, CA, USA)(UIST ’23). Association for Computing Machinery, New York, NY, USA, Article 55, 13 pages. doi:10.1145/3586183.3606758
-
[8]
Mustafa Doga Dogan, Ahmad Taka, Michael Lu, Yunyi Zhu, Akshat Kumar, Aakar Gupta, and Stefanie Mueller. 2022. InfraredTags: Embedding Invisible AR Markers and Barcodes Using Low-Cost, Infrared-Based 3D Printing and Imaging Tools. In Proceedings of the 2022 CHI Conference on Human Factors in Computing Systems (New Orleans, LA, USA)(CHI ’22). Association fo...
-
[9]
Anjith George and Aurobinda Routray. 2016. Fast and accurate algorithm for eye localisation for gaze tracking in low-resolution images.IET Computer Vision10, 7 (2016), 660–669
2016
-
[10]
Michail Kalaitzakis, Brennan Cain, Sabrina Carroll, Anand Ambrosi, Camden Whitehead, and Nikolaos Vitzilaios. 2021. Fiducial markers for pose estimation: Overview, applications and experimental comparison of the artag, apriltag, aruco and stag markers.Journal of Intelligent & Robotic Systems101, 4 (2021), 71
2021
-
[11]
Tharindu Kaluarachchi, Shamane Siriwardhana, Elliott Wen, and Suranga Nanayakkara. 2023. A Corneal Surface Reflections-Based Intelligent System for Lifelogging Applications.International Journal of Human–Computer Interac- tion39, 9 (2023), 1963–1980. arXiv:https://doi.org/10.1080/10447318.2022.2163240 doi:10.1080/10447318.2022.2163240
-
[12]
H. Kato and M. Billinghurst. 1999. Marker tracking and HMD calibration for a video-based augmented reality conferencing system. InProceedings 2nd IEEE and ACM International Workshop on Augmented Reality (IW AR’99). 85–94. doi:10. 1109/IWAR.1999.803809
-
[13]
The story of life is quicker than the blink of an eye
Christian Lander, Antonio Krüger, and Markus Löchtefeld. 2016. "The story of life is quicker than the blink of an eye": using corneal imaging for life log- ging. InProceedings of the 2016 ACM International Joint Conference on Per- vasive and Ubiquitous Computing: Adjunct(Heidelberg, Germany)(UbiComp ’16). Association for Computing Machinery, New York, NY,...
-
[14]
LEDGUHON. n.d.. 10Pcs 850nm High Power SMD LED Chip 3W Led Emitting Diode Infrared Emitting Quadruple Chip Use for Infrared Transmitter (120deg, IR 850nm–856nm, 700mA). https://www.amazon.com/dp/B0CJ5C8TCF. Amazon product page. Accessed: 2026-04-01
2026
-
[15]
Meta. 2026. Meta Orion. https://www.meta.com/emerging-tech/orion
2026
-
[16]
MidOpt. n.d.. BP850 Near-IR Bandpass Filter for Machine Vision. https://midopt. com/filters/bp850/. Accessed: 2026-04-01
2026
-
[17]
Atsushi Nakazawa, Christian Nitschke, and Toyoaki Nishida. 2016. Registration of eye reflection and scene images using an aspherical eye model.J. Opt. Soc. Am. A33, 11 (Nov 2016), 2264–2276. doi:10.1364/JOSAA.33.002264
-
[18]
Ko Nishino and Shree K Nayar. 2004. Eyes for relighting.ACM Transactions on Graphics (TOG)23, 3 (2004), 704–711
2004
-
[19]
Christian Nitschke and Atsushi Nakazawa. 2012. Super-Resolution from Corneal Images. InBMVC. 1–12
2012
-
[20]
Christian Nitschke, Atsushi Nakazawa, and Haruo Takemura. 2013. Corneal imaging revisited: An overview of corneal reflection analysis and applications. IPSJ Transactions on Computer Vision and Applications5 (2013), 1–18
2013
-
[21]
Yuki Ohshima, Kyosuke Maeda, Yusuke Edamoto, and Atsushi Nakazawa. 2021. Visual place recognition from eye reflection.IEEE Access9 (2021), 57364–57371
2021
-
[22]
Edwin Olson. 2011. AprilTag: A robust and flexible visual fiducial system. In2011 IEEE international conference on robotics and automation. IEEE, 3400–3407
2011
-
[23]
OpenCV. 2026. Blob Detection using OpenCV. https://opencv.org/blob-detection- using-opencv/
2026
-
[24]
Erich Schubert, Jörg Sander, Martin Ester, Hans Peter Kriegel, and Xiaowei Xu
-
[25]
ACM transactions on database systems (tods)42, 3 (2017), 1–21
DBSCAN revisited, revisited: why and how you should (still) use DBSCAN. ACM transactions on database systems (tods)42, 3 (2017), 1–21
2017
-
[26]
Nusrat Sharmin and Remus Brad. 2012. Optimal filter estimation for Lucas- Kanade optical flow.Sensors12, 9 (2012), 12694–12709
2012
-
[27]
Huiqiong Wang, Stephen Lin, Xiuqing Ye, and Weikang Gu. 2008. Separating corneal reflections for illumination estimation.Neurocomputing71, 10-12 (2008), 1788–1797
2008
-
[28]
John Wang and Edwin Olson. 2016. AprilTag 2: Efficient and robust fiducial detection. In2016 IEEE/RSJ international conference on intelligent robots and systems (IROS). IEEE, 4193–4198
2016
-
[29]
Sunu Wibirama, Igi Ardiyanto, Thoriq Satriya, Teguh Bharata Adji, Noor Akhmad Setiawan, and Muhammad Taufiq Setiawan. 2019. An improved pupil localization technique for real-time video-oculography under extreme eyelid occlusion.In- ternational Journal of Innovative Computing, Information and Control15, 4 (2019), 1547–1563
2019
-
[30]
Flana- gin, Peter zu Eulenburg, and Seyed-Ahmad Ahmadi
Yuk-Hoi Yiu, Moustafa Aboulatta, Theresa Raiser, Leoni Ophey, Virginia L. Flana- gin, Peter zu Eulenburg, and Seyed-Ahmad Ahmadi. 2019. DeepVOG: Open- source pupil segmentation and gaze estimation in neuroscience using deep learn- ing.Journal of Neuroscience Methods324 (2019), 108307. doi:10.1016/j.jneumeth. 2019.05.016
discussion (0)
Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.