pith. machine review for the scientific record. sign in

arxiv: 2604.03486 · v2 · submitted 2026-04-03 · 💻 cs.HC · cs.AI· cs.CV· cs.LG· cs.MA

Recognition: no theorem link

VisionClaw: Always-On AI Agents through Smart Glasses

Authors on Pith no claims yet

Pith reviewed 2026-05-13 18:08 UTC · model grok-4.3

classification 💻 cs.HC cs.AIcs.CVcs.LGcs.MA
keywords wearable AIsmart glassesegocentric perceptionAI agentshands-free interactiontask delegationsituated computing
0
0 comments X

The pith

Integrating perception and execution in always-on smart glasses AI agents enables faster task completion with less overhead.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper presents VisionClaw, a system running on Meta Ray-Ban smart glasses that continuously perceives the user's real-world environment and links it directly to AI agent execution for tasks initiated by speech. Through a lab study with 12 participants and a small longitudinal deployment with 5 users, it finds that this tight coupling of seeing and acting produces quicker task completion and lower interaction effort than baselines that lack always-on perception or agent capabilities. The work also observes that users start tasks more opportunistically amid other activities and delegate execution instead of managing it step by step. This matters for anyone interested in wearable AI that can operate hands-free without forcing users to pause their current activity or switch to a phone or separate interface.

Core claim

VisionClaw integrates live egocentric perception with agentic task execution on smart glasses, allowing speech-driven initiation and delegation of real-world tasks such as adding objects to an Amazon cart, generating notes from physical documents, receiving meeting briefings, creating events from posters, or controlling IoT devices. Controlled evaluations show faster task completion and reduced interaction overhead compared to non-always-on and non-agent baselines, while deployment observations reveal a shift toward opportunistic task initiation during ongoing activities and greater delegation rather than manual control.

What carries the argument

VisionClaw, the always-on wearable AI agent that continuously couples egocentric perception from smart glasses with OpenClaw AI agents for in-situ, speech-driven task initiation and delegation.

If this is right

  • Task completion is faster when perception and execution are integrated in one wearable system.
  • Interaction overhead is lower than in non-always-on or non-agent setups.
  • Users initiate tasks opportunistically during other ongoing activities.
  • Execution is delegated more frequently rather than performed through direct manual control.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • The same continuous coupling approach could be tested on other wearables such as earbuds or watches to support AI assistance without visual displays.
  • Privacy mechanisms would need to address continuous egocentric video capture if the system scales beyond controlled studies.
  • Interface design for delegation may become more important than direct control as users adapt to opportunistic triggering.
  • A follow-up experiment measuring error rates in real environments would clarify whether speed gains come at the cost of accuracy.

Load-bearing premise

The small-scale lab study with 12 participants and longitudinal deployment with 5 users are sufficient to demonstrate general performance gains and a fundamental shift in interaction patterns for broader populations.

What would settle it

A larger study with more participants across varied real-world settings that measures no significant reduction in task completion time or interaction overhead for the integrated system versus the non-always-on and non-agent baselines.

Figures

Figures reproduced from arXiv: 2604.03486 by DaeHo Lee, Eric J Gonzalez, Mar Gonzalez-Franco, Ryo Suzuki, Xiaoan Liu.

Figure 1
Figure 1. Figure 1: VisionClaw integrates always-on egocentric perception with agentic task execution on smart glasses. A user holding [PITH_FULL_IMAGE:figures/full_fig_p001_1.png] view at source ↗
Figure 2
Figure 2. Figure 2: System architecture of VisionClaw. The wearable device layer captures audio and video from Meta Ray-Ban smart [PITH_FULL_IMAGE:figures/full_fig_p004_2.png] view at source ↗
Figure 4
Figure 4. Figure 4: Task completion time. Asterisks next to labels indi [PITH_FULL_IMAGE:figures/full_fig_p005_4.png] view at source ↗
Figure 3
Figure 3. Figure 3: Overview of the four tasks used in the study [PITH_FULL_IMAGE:figures/full_fig_p005_3.png] view at source ↗
Figure 6
Figure 6. Figure 6: Self-authored questionnaire. Asterisks next to labels [PITH_FULL_IMAGE:figures/full_fig_p006_6.png] view at source ↗
Figure 5
Figure 5. Figure 5: NASA-TLX. Asterisks next to labels indicate signifi [PITH_FULL_IMAGE:figures/full_fig_p006_5.png] view at source ↗
Figure 7
Figure 7. Figure 7: Representative use cases from the deployment study, one per category. Each scenario shows a participant wearing [PITH_FULL_IMAGE:figures/full_fig_p007_7.png] view at source ↗
Figure 8
Figure 8. Figure 8: Usage log of the deployment study. An interactive version of this visualization can be seen at the following link. Data [PITH_FULL_IMAGE:figures/full_fig_p008_8.png] view at source ↗
Figure 9
Figure 9. Figure 9: Findings on interactions observed in the longitudinal deployment study, illustrating four recurring patterns: multi [PITH_FULL_IMAGE:figures/full_fig_p009_9.png] view at source ↗
Figure 10
Figure 10. Figure 10: Future research directions for always-on agentic [PITH_FULL_IMAGE:figures/full_fig_p010_10.png] view at source ↗
Figure 11
Figure 11. Figure 11: A taxonomy of use cases observed in the deployment study. The figure organizes interactions into six categories— [PITH_FULL_IMAGE:figures/full_fig_p014_11.png] view at source ↗
read the original abstract

We present VisionClaw, an always-on wearable AI agent that integrates live egocentric perception with agentic task execution. Running on Meta Ray-Ban smart glasses, VisionClaw continuously perceives real-world context and enables in-situ, speech-driven action initiation and delegation via OpenClaw AI agents. Therefore, users can directly execute tasks through the smart glasses, such as adding real-world objects to an Amazon cart, generating notes from physical documents, receiving meeting briefings on the go, creating events from posters, or controlling IoT devices. We evaluate VisionClaw through a controlled laboratory study (N=12) and a longitudinal deployment study (N=5). Results show that integrating perception and execution enables faster task completion and reduces interaction overhead compared to non-always-on and non-agent baselines. Beyond performance gains, deployment findings reveal a shift in interaction: tasks are initiated opportunistically during ongoing activities, and execution is increasingly delegated rather than manually controlled. These results suggest a new paradigm for wearable AI agents, where perception and action are continuously coupled to support situated, hands-free interaction.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

3 major / 2 minor

Summary. The manuscript presents VisionClaw, an always-on wearable AI agent running on Meta Ray-Ban smart glasses that integrates live egocentric perception with agentic task execution via OpenClaw agents. Users can perform situated tasks such as adding objects to an Amazon cart, generating notes from documents, or controlling IoT devices through speech. The system is evaluated in a controlled laboratory study (N=12) and a longitudinal deployment (N=5), with claims that the perception-execution coupling yields faster task completion, lower interaction overhead versus non-always-on and non-agent baselines, and a shift toward opportunistic initiation and delegated execution.

Significance. If the empirical claims hold after improved reporting and analysis, the work would represent a meaningful contribution to wearable HCI by demonstrating how continuous perception-action coupling can reduce friction in real-world tasks and support hands-free interaction. The longitudinal observations of opportunistic and delegated behavior patterns are particularly interesting as potential indicators of a new interaction paradigm, though the small samples limit claims of broad generalizability.

major comments (3)
  1. [Abstract] Abstract: the statement that 'integrating perception and execution enables faster task completion and reduces interaction overhead' is presented without any quantitative metrics, effect sizes, p-values, or baseline performance numbers, making it impossible to evaluate the magnitude or reliability of the reported gains.
  2. [Evaluation] Evaluation section: the laboratory study (N=12) and longitudinal deployment (N=5) provide no power analysis, pre-registered primary metrics, exclusion criteria, or statistical test details; with such small samples, individual differences in task familiarity or speech patterns could dominate results and undermine the causal claim that perception-execution integration produces the observed benefits.
  3. [Evaluation] Evaluation section: the non-always-on and non-agent baselines are referenced but not described in sufficient technical detail (e.g., exact interface differences, task instructions, or how they control for the integration factor), preventing confirmation that the comparison isolates the claimed always-on coupling effect.
minor comments (2)
  1. [Abstract] Abstract: consider briefly listing the concrete tasks used in the studies (e.g., cart addition, note generation) to help readers immediately grasp the scope of evaluated functionality.
  2. [Related Work] The manuscript would benefit from a short related-work subsection contrasting VisionClaw with prior always-on wearable prototypes (e.g., earlier smart-glass agents) to clarify the precise novelty of the perception-execution integration.

Simulated Author's Rebuttal

3 responses · 0 unresolved

We thank the referee for the constructive feedback. We address each major comment below and will revise the manuscript to improve clarity, reporting, and detail in the abstract and evaluation sections.

read point-by-point responses
  1. Referee: [Abstract] Abstract: the statement that 'integrating perception and execution enables faster task completion and reduces interaction overhead' is presented without any quantitative metrics, effect sizes, p-values, or baseline performance numbers, making it impossible to evaluate the magnitude or reliability of the reported gains.

    Authors: We agree that the abstract would benefit from explicit quantitative support. In the revised version, we will add concise metrics drawn from the evaluation results, including mean task completion times (e.g., VisionClaw: 48s vs. non-always-on baseline: 132s), interaction overhead reductions, Cohen's d effect sizes, and p-values from paired t-tests. These numbers are reported in full in Section 5; we will summarize them in the abstract while preserving its length. revision: yes

  2. Referee: [Evaluation] Evaluation section: the laboratory study (N=12) and longitudinal deployment (N=5) provide no power analysis, pre-registered primary metrics, exclusion criteria, or statistical test details; with such small samples, individual differences in task familiarity or speech patterns could dominate results and undermine the causal claim that perception-execution integration produces the observed benefits.

    Authors: We will expand the Evaluation section to include a post-hoc power analysis for the primary outcomes, explicit statement of pre-registered metrics (task completion time and interaction count), confirmation that no participants were excluded, and full statistical details (paired t-tests with exact p-values, degrees of freedom, and effect sizes). We will also add a dedicated limitations paragraph acknowledging the exploratory nature of the studies, potential influence of individual differences, and the need for larger-scale validation in future work. We will moderate causal phrasing accordingly. revision: yes

  3. Referee: [Evaluation] Evaluation section: the non-always-on and non-agent baselines are referenced but not described in sufficient technical detail (e.g., exact interface differences, task instructions, or how they control for the integration factor), preventing confirmation that the comparison isolates the claimed always-on coupling effect.

    Authors: We will provide expanded technical descriptions of both baselines in the revised Evaluation section. This will specify: (1) non-always-on condition requires explicit button-press camera activation before speech input; (2) non-agent condition uses the same speech input but routes to a non-agentic scripted interface without autonomous delegation; (3) verbatim task instructions provided to participants; and (4) the within-subjects design that holds speech input and task content constant while varying only perception access and agent autonomy. These additions will clarify how the comparisons isolate the perception-execution coupling. revision: yes

Circularity Check

0 steps flagged

No circularity; claims rest on independent empirical user studies

full rationale

The paper introduces VisionClaw as a system integrating egocentric perception with agentic execution on smart glasses and evaluates it via a controlled lab study (N=12) and longitudinal deployment (N=5). No equations, fitted parameters, self-citations, or derivation chains appear in the provided text. Central claims of faster task completion, reduced overhead, and shifts toward opportunistic interaction are asserted directly from study outcomes rather than reducing to internal definitions or prior self-referential results. The work is self-contained against external benchmarks of system performance and user behavior.

Axiom & Free-Parameter Ledger

0 free parameters · 0 axioms · 0 invented entities

No mathematical model or derivation is present; the central claim rests on the engineering of the VisionClaw prototype and the interpretation of two small user studies.

pith-pipeline@v0.9.0 · 5507 in / 1092 out tokens · 42184 ms · 2026-05-13T18:08:20.518867+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Forward citations

Cited by 2 Pith papers

Reviewed papers in the Pith corpus that reference this work. Sorted by Pith novelty score.

  1. SpeakerLLM: A Speaker-Specialized Audio-LLM for Speaker Understanding and Verification Reasoning

    cs.SD 2026-05 unverdicted novelty 6.0

    SpeakerLLM unifies speaker profiling, recording-condition understanding, and structured verification reasoning in an audio-LLM via a hierarchical tokenizer and decision traces.

  2. Position: Life-Logging Video Streams Make the Privacy-Utility Trade-off Inevitable

    cs.CV 2026-05 unverdicted novelty 4.0

    Life-logging video streams create an inevitable privacy-utility trade-off that is a foundational challenge for always-on AI systems.

Reference graph

Works this paper leans on

87 extracted references · 87 canonical work pages · cited by 2 Pith papers · 6 internal anchors

  1. [1]

    Gregory D Abowd, Anind K Dey, Peter J Brown, Nigel Davies, Mark Smith, and Pete Steggles. 1999. Towards a better understanding of context and context- awareness. InInternational symposium on handheld and ubiquitous computing. Springer, 304–307

  2. [2]

    Apple Computer. 1987. Apple Knowledge Navigator Concept Video. https: //www.youtube.com/watch?v=HGYFEI6uLy0

  3. [3]

    Riku Arakawa, Jill Fain Lehman, and Mayank Goel. 2024. Prism-q&a: Step-aware voice assistant on a smartwatch enabled by multimodal procedure tracking and large language models.Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies8, 4 (2024), 1–26

  4. [4]

    Divyanshu Bhardwaj, Alexander Ponticello, Shreya Tomar, Adrian Dabrowski, and Katharina Krombholz. 2024. In focus, out of privacy: the wearer’s perspec- tive on the privacy dilemma of camera glasses. InProceedings of the 2024 CHI Conference on Human Factors in Computing Systems. 1–18

  5. [5]

    Frederik Brudy, Christian Holz, Roman Rädle, Chi-Jui Wu, Steven Houben, Clemens Nylandsted Klokmose, and Nicolai Marquardt. 2019. Cross-device taxonomy: Survey, opportunities and challenges of interactions spanning across multiple devices. InProceedings of the 2019 CHI Conference on Human Factors in Computing Systems. 1–28

  6. [6]

    Runze Cai, Nuwan Janaka, Hyeongcheol Kim, Yang Chen, Shengdong Zhao, Yun Huang, and David Hsu. 2025. Aiget: Transforming everyday moments into hidden knowledge discovery with ai assistance on smart glasses. InProceedings of the 2025 CHI Conference on Human Factors in Computing Systems. 1–26

  7. [7]

    Ruei-Che Chang, Yuxuan Liu, and Anhong Guo. 2024. Worldscribe: Towards context-aware live visual descriptions. InProceedings of the 37th Annual ACM Symposium on User Interface Software and Technology. 1–18

  8. [8]

    Hyunsung Cho, Jacqui Fashimpaur, Naveen Sendhilnathan, Jonathan Browder, David Lindlbauer, Tanya R Jonker, and Kashyap Todi. 2025. Persistent assistant: Seamless everyday AI interactions via intent grounding and multimodal feedback. InProceedings of the 2025 CHI Conference on Human Factors in Computing Systems. 1–19

  9. [9]

    Tamara Denning, Zakariya Dehlawi, and Tadayoshi Kohno. 2014. In situ with bystanders of augmented reality glasses: Perspectives on recording and privacy- mediating technologies. InProceedings of the SIGCHI conference on human factors in computing systems. 2377–2386

  10. [10]

    Audrey Desjardins and Aubree Ball. 2018. Revealing tensions in autobiographical design in HCI. Inproceedings of the 2018 designing interactive systems conference. 753–764

  11. [11]

    Anind K Dey. 2001. Understanding and using context.Personal and ubiquitous computing5, 1 (2001), 4–7

  12. [12]

    Mustafa Doga Dogan, Eric J Gonzalez, Karan Ahuja, Ruofei Du, Andrea Colaço, Johnny Lee, Mar Gonzalez-Franco, and David Kim. 2024. Augmented object intelligence with xr-objects. InProceedings of the 37th Annual ACM Symposium on User Interface Software and Technology. 1–15

  13. [13]

    Jakob Engel, Kiran Somasundaram, Michael Goesele, Albert Sun, Alexander Gamino, Andrew Turner, Arjang Talattof, Arnie Yuan, Bilal Souti, Brighid Mered- ith, et al. 2023. Project aria: A new tool for egocentric multi-modal ai research. arXiv preprint arXiv:2308.13561(2023)

  14. [14]

    Inner-Voice

    Cathy Mengying Fang, Yasith Samaradivakara, Pattie Maes, and Suranga Nanayakkara. 2025. Mirai: A Wearable Proactive AI" Inner-Voice" for Contextual Nudging. InProceedings of the extended abstracts of the CHI conference on human factors in computing systems. 1–9

  15. [15]

    Gabriele Goletto, Tushar Nagarajan, Giuseppe Averta, and Dima Damen. 2024. Amego: Active memory from long egocentric videos. InEuropean Conference on Computer Vision. Springer, 92–110

  16. [16]

    Google. 2024. Gemini 2.0: Level Up Your Apps with Real-Time Multimodal Interactions. https://developers.googleblog.com/en/gemini-2-0-level-up-your- apps-with-real-time-multimodal-interactions/

  17. [17]

    Kristen Grauman, Andrew Westbury, Eugene Byrne, Zachary Chavis, Antonino Furnari, Rohit Girdhar, Jackson Hamburger, Hao Jiang, Miao Liu, Xingyu Liu, et al. 2022. Ego4d: Around the world in 3,000 hours of egocentric video. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 18995–19012

  18. [18]

    Kristen Grauman, Andrew Westbury, Lorenzo Torresani, Kris Kitani, Jitendra Malik, Triantafyllos Afouras, Kumar Ashutosh, Vijay Baiyya, Siddhant Bansal, Bikram Boote, et al. 2024. Ego-exo4d: Understanding skilled human activity from first-and third-person perspectives. InProceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 19383–19400

  19. [19]

    Kiryong Ha, Zhuo Chen, Wenlu Hu, Wolfgang Richter, Padmanabhan Pillai, and Mahadev Satyanarayanan. 2014. Towards wearable cognitive assistance. InProceedings of the 12th annual international conference on Mobile systems, applications, and services. 68–81

  20. [20]

    Steve Hodges, Lyndsay Williams, Emma Berry, Shahram Izadi, James Srinivasan, Alex Butler, Gavin Smyth, Narinder Kapur, and Ken Wood. 2006. SenseCam: A retrospective memory aid. InInternational conference on ubiquitous computing. Springer, 177–193

  21. [21]

    Sirui Hong, Mingchen Zhuge, Jonathan Chen, Xiawu Zheng, Yuheng Cheng, Jinlin Wang, Ceyao Zhang, Zili Wang, Steven Ka Shing Yau, Zijuan Lin, et al

  22. [22]

    InThe twelfth international conference on learning representations

    MetaGPT: Meta programming for a multi-agent collaborative framework. InThe twelfth international conference on learning representations

  23. [23]

    Wenyi Hong, Weihan Wang, Qingsong Lv, Jiazheng Xu, Wenmeng Yu, Junhui Ji, Yan Wang, Zihan Wang, Yuxiao Dong, Ming Ding, et al. 2024. Cogagent: A visual language model for gui agents. InProceedings of the IEEE/CVF conference on computer vision and pattern recognition. 14281–14290

  24. [24]

    Irene Hou, Alexander Qin, Lauren Cheng, and Philip J Guo. 2026. Beyond the Desk: Barriers and Future Opportunities for AI to Assist Scientists in Embodied Physical Tasks.arXiv preprint arXiv:2603.19504(2026)

  25. [25]

    Roberto Hoyle, Robert Templeman, Steven Armes, Denise Anthony, David Cran- dall, and Apu Kapadia. 2014. Privacy behaviors of lifeloggers using wearable cameras. InProceedings of the 2014 ACM international joint conference on pervasive and ubiquitous computing. 571–582

  26. [26]

    Xiyun Hu, Dizhi Ma, Fengming He, Zhengzhe Zhu, Shao-Kang Hsia, Chenfei Zhu, Ziyi Liu, and Karthik Ramani. 2025. GesPrompt: Leveraging Co-Speech Gestures to Augment LLM-Based Interaction in Virtual Reality. InProceedings of the 2025 ACM Designing Interactive Systems Conference. 59–80

  27. [27]

    Yifei Huang, Jilan Xu, Baoqi Pei, Lijin Yang, Mingfang Zhang, Yuping He, Guo Chen, Xinyuan Chen, Yaohui Wang, Zheng Nie, et al. 2025. Vinci: A real-time smart assistant based on egocentric vision-language model for portable devices. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technolo- gies9, 3 (2025), 1–33

  28. [28]

    Mina Huh, Zihui Xue, Ujjaini Das, Kumar Ashutosh, Kristen Grauman, and Amy Pavel. 2025. Vid2Coach: Transforming How-To Videos into Task Assistants. In Proceedings of the 38th Annual ACM Symposium on User Interface Software and Technology. 1–24

  29. [29]

    Shivesh Jadon, Mehrad Faridan, Edward Mah, Rajan Vaish, Wesley Willett, and Ryo Suzuki. 2024. Augmented conversation with embedded speech-driven on- the-fly referencing in AR.arXiv preprint arXiv:2405.18537(2024)

  30. [30]

    Spike Jonze. 2013. Her. Motion picture, Warner Bros

  31. [31]

    Jared Kaplan, Sam McCandlish, Tom Henighan, Tom B Brown, Benjamin Chess, Rewon Child, Scott Gray, Alec Radford, Jeffrey Wu, and Dario Amodei. 2020. Scaling laws for neural language models.arXiv preprint arXiv:2001.08361(2020)

  32. [32]

    Yoonsang Kim, Devshree Jadeja, Divyansh Pradhan, Yalong Yang, and Arie Kauf- man. 2026. SpeechLess: Micro-utterance with Personalized Spatial Memory- aware Assistant in Everyday Augmented Reality.arXiv preprint arXiv:2602.00793 (2026)

  33. [33]

    Yoonsang Kim, Divyansh Pradhan, Devshree Jadeja, and Arie Kaufman. 2026. From Speech-to-Spatial: Grounding Utterances on A Live Shared View with Augmented Reality.arXiv preprint arXiv:2602.03059(2026)

  34. [34]

    Yoonsang Kim, Yalong Yang, and Arie E Kaufman. 2026. Memento: Towards Proac- tive Visualization of Everyday Memories with Personal Wearable AR Assistant. arXiv preprint arXiv:2601.17622(2026)

  35. [35]

    Robert Konrad, Nitish Padmanaban, J Gabriel Buckmaster, Kevin C Boyle, and Gordon Wetzstein. 2024. Gazegpt: Augmenting human capabilities using gaze- contingent contextual ai for smart eyewear.arXiv preprint arXiv:2401.17217 (2024)

  36. [36]

    Geonsun Lee, Min Xia, Nels Numan, Xun Qian, David Li, Yanhe Chen, Achin Kulshrestha, Ishan Chatterjee, Yinda Zhang, Dinesh Manocha, et al. 2025. Sensible agent: A framework for unobtrusive interaction with proactive ar agents. In Proceedings of the 38th Annual ACM Symposium on User Interface Software and Technology. 1–22

  37. [37]

    Jaewook Lee, Andrew D Tjahjadi, Jiho Kim, Junpu Yu, Minji Park, Jiawen Zhang, Jon E Froehlich, Yapeng Tian, and Yuhang Zhao. 2024. Cookar: Affordance augmentations in wearable ar to support kitchen tool interactions for people 11 arXiv Preprint, April 2026, Liu et al. with low vision. InProceedings of the 37th Annual ACM Symposium on User Interface Softwa...

  38. [38]

    Jaewook Lee, Jun Wang, Elizabeth Brown, Liam Chu, Sebastian S Rodriguez, and Jon E Froehlich. 2024. GazePointAR: A context-aware multimodal voice assistant for pronoun disambiguation in wearable augmented reality. InProceedings of the 2024 CHI Conference on Human Factors in Computing Systems. 1–20

  39. [39]

    Chenyi Li, Guande Wu, Gromit Yeuk-Yin Chan, Dishita Gdi Turakhia, Sonia Castelo Quispe, Dong Li, Leslie Welch, Claudio Silva, and Jing Qian. 2025. Satori: Towards Proactive AR Assistant with Belief-Desire-Intention User Modeling. In Proceedings of the 2025 CHI Conference on Human Factors in Computing Systems. 1–24

  40. [40]

    Jiahao Nick Li, Yan Xu, Tovi Grossman, Stephanie Santosa, and Michelle Li. 2024. OmniActions: Predicting digital actions in response to real-world multimodal sensory inputs with LLMs. InProceedings of the 2024 CHI Conference on Human Factors in Computing Systems. 1–22

  41. [41]

    Yijiang Li, Genpei Zhang, Jiacheng Cheng, Yi Li, Xiaojun Shan, Dashan Gao, Jiancheng Lyu, Yuan Li, Ning Bi, and Nuno Vasconcelos. 2025. EgoPrivacy: What Your First-Person Camera Says About You?arXiv preprint arXiv:2506.12258 (2025)

  42. [42]

    Kevin Qinghong Lin, Jinpeng Wang, Mattia Soldan, Michael Wray, Rui Yan, Eric Z Xu, Difei Gao, Rong-Cheng Tu, Wenzhe Zhao, Weijie Kong, et al. 2022. Egocentric video-language pretraining.Advances in Neural Information Processing Systems 35 (2022), 7575–7586

  43. [43]

    Geoffrey Litt. 2025. Enough AI Copilots! We Need AI HUDs. https://www. geoffreylitt.com/2025/07/27/enough-ai-copilots-we-need-ai-huds. Blog post, July 2025

  44. [44]

    Ziyi Liu, Zhengzhe Zhu, Enze Jiang, Feichi Huang, Ana M Villanueva, Xun Qian, Tianyi Wang, and Karthik Ramani. 2023. Instrumentar: Auto-generation of augmented reality tutorials for operating digital instruments through recording embodied demonstration. InProceedings of the 2023 CHI Conference on Human Factors in Computing Systems. 1–17

  45. [45]

    Rehana Mahfuz, Yinyi Guo, Erik Visser, and Phanidhar Chinchili. 2026. Proactive Conversational Assistant for a Procedural Manual Task based on Audio and IMU. arXiv preprint arXiv:2602.15707(2026)

  46. [46]

    Meta Platforms. 2024. Introducing the Meta Wearables Device Access Toolkit. https://developers.meta.com/blog/introducing-meta-wearables-device- access-toolkit/

  47. [47]

    Carman Neustaedter and Phoebe Sengers. 2012. Autobiographical design in HCI research: designing and learning through use-it-yourself. InProceedings of the designing interactive systems conference. 514–523

  48. [48]

    Pha Nguyen, Sailik Sengupta, Girik Malik, Arshit Gupta, and Bonan Min. 2025. InsTALL: Context-aware Instructional Task Assistance with Multi-modal Large Language Models.arXiv preprint arXiv:2501.12231(2025)

  49. [49]

    Alex Olwal, Kevin Balke, Dmitrii Votintcev, Thad Starner, Paula Conn, Bonnie Chinh, and Benoit Corda. 2020. Wearable subtitles: Augmenting spoken com- munication with lightweight eyewear for all-day captioning. InProceedings of the 33rd Annual ACM Symposium on User Interface Software and Technology. 1108–1120

  50. [50]

    OpenAI. 2023. GPT-4V(ision) System Card. https://cdn.openai.com/papers/ GPTV_System_Card.pdf

  51. [51]

    Joon Sung Park, Joseph O’Brien, Carrie Jun Cai, Meredith Ringel Morris, Percy Liang, and Michael S Bernstein. 2023. Generative agents: Interactive simulacra of human behavior. InProceedings of the 36th annual acm symposium on user interface software and technology. 1–22

  52. [52]

    Yunqiang Pei, Renming Huang, Mingfeng Zha, Guoqing Wang, Peng Wang, Qiao Kang, Yang Yang, and Heng Tao Shen. 2025. AttentionAR: AR Adaptation and Warning for Real-World Safety via Attention Modeling and MLLM Reasoning. In Proceedings of the 38th Annual ACM Symposium on User Interface Software and Technology. 1–19

  53. [53]

    Yi-Hao Peng, Dingzeyu Li, Jeffrey P Bigham, and Amy Pavel. 2025. Morae: Proactively pausing ui agents for user choices. InProceedings of the 38th Annual ACM Symposium on User Interface Software and Technology. 1–14

  54. [54]

    Kevin Pu, Ting Zhang, Naveen Sendhilnathan, Sebastian Freitag, Raj Sodhi, and Tanya R Jonker. 2025. Promemassist: Exploring timely proactive assistance through working memory modeling in multi-modal wearable devices. InProceed- ings of the 38th Annual ACM Symposium on User Interface Software and Technology. 1–19

  55. [55]

    Jun Rekimoto. 2025. GazeLLM: Multimodal LLMs incorporating human visual attention. InProceedings of the Augmented Humans International Conference 2025. 302–311

  56. [56]

    Bradley J Rhodes. 1997. The wearable remembrance agent: A system for aug- mented memory.Personal Technologies1, 4 (1997), 218–224

  57. [57]

    Timo Schick, Jane Dwivedi-Yu, Roberto Dessì, Roberta Raileanu, Maria Lomeli, Eric Hambro, Luke Zettlemoyer, Nicola Cancedda, and Thomas Scialom. 2023. Toolformer: Language models can teach themselves to use tools.Advances in neural information processing systems36 (2023), 68539–68551

  58. [58]

    Peter Steinberger. 2025. OpenClaw: Open-Source Autonomous AI Agent Frame- work. https://github.com/openclaw/openclaw

  59. [59]

    Ryo Suzuki, Mar Gonzalez-Franco, Misha Sra, and David Lindlbauer. 2023. Xr and ai: Ai-enabled virtual, augmented, and mixed reality. InAdjunct proceedings of the 36th annual ACM symposium on user interface software and technology. 1–3

  60. [60]

    Ryo Suzuki, Mar Gonzalez-Franco, Misha Sra, and David Lindlbauer. 2025. Ev- eryday AR through AI-in-the-Loop. InProceedings of the Extended Abstracts of the CHI Conference on Human Factors in Computing Systems. 1–5

  61. [61]

    Yiliu Tang, Jason Situ, Andrea Yaoyun Cui, Mengke Wu, and Yun Huang. 2025. Llm integration in extended reality: A comprehensive review of current trends, challenges, and future perspectives. InProceedings of the 2025 CHI Conference on Human Factors in Computing Systems. 1–24

  62. [62]

    Minh Duc Vu, Han Wang, Jieshan Chen, Zhuang Li, Shengdong Zhao, Zhenchang Xing, and Chunyang Chen. 2024. Gptvoicetasker: Advancing multi-step mobile task efficiency through dynamic interface exploration and learning. InProceedings of the 37th Annual ACM Symposium on User Interface Software and Technology. 1–17

  63. [63]

    Junyang Wang, Haiyang Xu, Jiabo Ye, Ming Yan, Weizhou Shen, Ji Zhang, Fei Huang, and Jitao Sang. 2024. Mobile-agent: Autonomous multi-modal mobile device agent with visual perception.arXiv preprint arXiv:2401.16158(2024)

  64. [64]

    Xin Wang, Taein Kwon, Mahdi Rad, Bowen Pan, Ishani Chakraborty, Sean An- drist, Dan Bohus, Ashley Feniello, Bugra Tekin, Felipe Vieira Frujeri, et al. 2023. Holoassist: an egocentric human interaction dataset for interactive ai assistants in the real world. InProceedings of the IEEE/CVF International Conference on Computer Vision. 20270–20281

  65. [65]

    Xueyang Wang, Kewen Peng, Xin Yi, and Hewu Li. 2026. Mind the Gap: Mapping Wearer-Bystander Privacy Tensions and Context-Adaptive Pathways for Camera Glasses.arXiv preprint arXiv:2603.04930(2026)

  66. [66]

    Mark Weiser. 1992. Does ubiquitous computing need interface agents. InNo. Invited talk at MIT Media Lab Symposium on User Interface Agents

  67. [67]

    Mark Weiser. 1993. Some computer science issues in ubiquitous computing. Commun. ACM36, 7 (1993), 75–84

  68. [68]

    Mark Weiser. 1996. Open House. https://calmtech.com/papers/open-house. Review, Interactive Telecommunications Program, New York University2.0 (1996). Appeared March 1996

  69. [69]

    Mark Weiser, John Seely Brown, et al. 1996. Designing calm technology.Power- Grid Journal1, 1 (1996), 75–85

  70. [70]

    Zhiyong Wu, Chengcheng Han, Zichen Ding, Zhenmin Weng, Zhoumianze Liu, Shunyu Yao, Tao Yu, and Lingpeng Kong. 2024. Os-copilot: Towards generalist computer agents with self-improvement.arXiv preprint arXiv:2402.07456(2024)

  71. [71]

    Xuhai Xu, Anna Yu, Tanya R Jonker, Kashyap Todi, Feiyu Lu, Xun Qian, João Marcelo Evangelista Belo, Tianyi Wang, Michelle Li, Aran Mun, et al. 2023. Xair: A framework of explainable ai in augmented reality. InProceedings of the 2023 CHI Conference on Human Factors in Computing Systems. 1–30

  72. [72]

    Bufang Yang, Lilin Xu, Liekang Zeng, Yunqi Guo, Siyang Jiang, Wenrui Lu, Kaiwei Liu, Hancheng Xiang, Xiaofan Jiang, Guoliang Xing, et al. 2025. ProAgent: Harnessing On-Demand Sensory Contexts for Proactive LLM Agent Systems. arXiv preprint arXiv:2512.06721(2025)

  73. [73]

    Jingkang Yang, Shuai Liu, Hongming Guo, Yuhao Dong, Xiamengwei Zhang, Sicheng Zhang, Pengyun Wang, Zitang Zhou, Binzhu Xie, Ziyue Wang, et al

  74. [74]

    InProceedings of the Computer Vision and Pattern Recognition Conference

    Egolife: Towards egocentric life assistant. InProceedings of the Computer Vision and Pattern Recognition Conference. 28885–28900

  75. [75]

    Shunyu Yao, Jeffrey Zhao, Dian Yu, Nan Du, Izhak Shafran, Karthik R Narasimhan, and Yuan Cao. 2022. React: Synergizing reasoning and acting in language models. InThe eleventh international conference on learning representations

  76. [76]

    Shoubin Yu, Lei Shu, Antoine Yang, Yao Fu, Srinivas Sunkara, Maria Wang, Jindong Chen, Mohit Bansal, and Boqing Gong. 2026. Ego2Web: A Web Agent Benchmark Grounded in Egocentric Videos. arXiv:2603.22529 [cs.CV] https: //arxiv.org/abs/2603.22529

  77. [77]

    Chi Zhang, Zhao Yang, Jiaxuan Liu, Yanda Li, Yucheng Han, Xin Chen, Zebiao Huang, Bin Fu, and Gang Yu. 2025. Appagent: Multimodal agents as smartphone users. InProceedings of the 2025 CHI Conference on Human Factors in Computing Systems. 1–20

  78. [78]

    Nandi Zhang, Yukang Yan, and Ryo Suzuki. 2025. From Following to Understand- ing: Investigating the Role of Reflective Prompts in AR-Guided Tasks to Promote User Understanding. InProceedings of the 2025 CHI Conference on Human Factors in Computing Systems. 1–18

  79. [79]

    Shuning Zhang, Qucheng Zang, YongquanOwen’ Hu, Jiachen Du, Xueyang Wang, Yan Kong, Xinyi Fu, Suranga Nanayakkara, Xin Yi, and Hewu Li. 2026. Vis- Guardian: A Lightweight Group-based Privacy Control Technique For Front Cam- era Data From AR Glasses in Home Environments.arXiv preprint arXiv:2601.19502 (2026)

  80. [80]

    Zheng Zhang, Mengjie Yu, Tianyi Wang, Kashyap Todi, Ajoy Savio Fernandes, Yue Liu, Haijun Xia, Tovi Grossman, and Tanya Jonker. 2026. Gazeify Then Voiceify: Physical Object Referencing Through Gaze and Voice Interaction with Displayless Smart Glasses.arXiv preprint arXiv:2601.19281(2026)

Showing first 80 references.