Embardiment: An Embodied AI Agent For Productivity In XR · The Large Language Model Bible Contribute to LLM-Bible

Embardiment: An Embodied AI Agent For Productivity In XR

Bovo Riccardo, Abreu Steven, Ahuja Karan, Gonzalez Eric J, Cheng Li-te, Gonzalez-franco Mar. Arxiv 2024

[Paper]    
Agentic Attention Mechanism Model Architecture Prompting RAG Tools

XR devices running chat-bots powered by Large Language Models (LLMs) have tremendous potential as always-on agents that can enable much better productivity scenarios. However, screen based chat-bots do not take advantage of the the full-suite of natural inputs available in XR, including inward facing sensor data, instead they over-rely on explicit voice or text prompts, sometimes paired with multi-modal data dropped as part of the query. We propose a solution that leverages an attention framework that derives context implicitly from user actions, eye-gaze, and contextual memory within the XR environment. This minimizes the need for engineered explicit prompts, fostering grounded and intuitive interactions that glean user insights for the chat-bot. Our user studies demonstrate the imminent feasibility and transformative potential of our approach to streamline user interaction in XR with chat-bots, while offering insights for the design of future XR-embodied LLM agents.

Similar Work