Large Language Models as A Paradigm Shift in Next-Generation Virtual Reality Interaction: A Comprehensive Investigation

Authors

  • Min Hu

DOI:

https://doi.org/10.54097/hw3wve60

Keywords:

Large Language Models; Virtual Reality; Human-Computer Interaction.

Abstract

The convergence of Large Language Models (LLMs) and Virtual Reality (VR) represents an emerging frontier in Human-computer interaction (HCI). Contemporary LLMs, such as GPT-4 and Gemini 1.5, demonstrate advanced multimodal reasoning capabilities, while VR has evolved from purely visual immersion to semantically rich, generative environments enabled by neural scene encoding and real-time holographic generation. However, current VR systems predominantly rely on preconfigured interactions and static interfaces, failing to harness the adaptive, generative, and dynamic potential of LLMs which is a significant research gap. Prior studies have focused on isolated aspects, such as graphical fidelity or voice-based navigation, without exploring the integration of LLMs as the central intelligence for real-time dialogue, agent behavior, and procedural content generation. Moreover, literature lacks a systematic analysis of the architectural frameworks, opportunities, and challenges inherent in deep LLM-VR integration. This review addresses these gaps through three primary objectives: (1) synthesizing recent advancements in multimodal LLMs and generative VR environments to assess their technical compatibility; (2) analyzing key application areas of LLMs in VR, including agent modeling, user experience adaptation, and procedural content generation, with detailed evaluation of methodologies, innovations, and empirical outcomes; and (3) proposing a structured framework to guide the development of next-generation intelligent VR systems. The findings affirm that LLM integration constitutes not merely an enhancement but a fundamental paradigm shift—enabling dynamic, context-aware, and highly immersive virtual experiences. This review provides a roadmap for researchers and practitioners aiming to leverage LLM-VR synergy across domains such as education, healthcare, and entertainment.

References

[1]OpenAI. GPT-4 technical report. 2023. Available from: https://cdn.openai.com/papers/gpt-4.pdf

[2]Google DeepMind. Gemini 1.5: unlocking multimodal understanding across millions of tokens. 2024. Available from: https://blog.google/technology/google-deepmind/gemini-1-5/

[3]Muller T, Evans A, Schied C, Keller A. Instant neural graphics primitives with a multiresolution hash encoding. ACM Trans Graph. 2022 Jul;41(4):102. doi: 10.1145/3528223.3530127

[4]Garon M, et al. Real-time high-fidelity facial performance capture. In: ACM SIGGRAPH 2023 Conference Proceedings. Los Angeles (CA), USA; 2023. Available from: https://dl.acm.org/doi/10.1145/3588432.3591544

[5]Lee D, Hoffmann K, Wolf L. The vision of tactile internet: bridging VR and AI. Proc IEEE. 2023 May;111(5):467-94. doi: 10.1109/JPROC.2023.3263587

[6]Meta AI. Llama 3: open and efficient foundation language models. 2024. Available from: https://ai.meta.com/blog/meta-llama-3/

[7]Zhang Y, Wang L, Liu M. Voice-driven navigation in VR using end-to-end ASR and LLMs. In: 2023 IEEE Conference on Virtual Reality and 3D User Interfaces (VR). Shanghai, China; 2023. p. 812-21. doi: 10.1109/VR55154.2023.00103

[8]Bhattacharya N, Overbeck R, Debevec P. Photorealistic rendering for immersive VR. In: ACM SIGGRAPH Asia 2022 Conference Papers. Daegu, Republic of Korea; 2022. p. 1-14. doi: 10.1145/3550469.3555394

[9]Kim S, Benoit A, Mueller FF. LLM-driven adaptive feedback in virtual reality language learning environments. Comput Assist Lang Learn. 2023;36(8):1234-56. doi: 10.1080/09588221.2023.2182525

[10]Anthropic. Claude 3 model card. 2024. Available from: https://www.anthropic.com/news/claude-3-model-card

[11]Zhang Y, et al. VR-LAA: a VR-LLM agent alignment framework for adaptive medical training. IEEE Trans Vis Comput Graph. 2024 May;30(5):2100-10. doi: 10.1109/TVCG.2024.3372050

[12]Kim J, et al. A multimodal LLM-based virtual therapist for anxiety management in VR. J Med Syst. 2024;48(1):28. doi: 10.1007/s10916-024-02052-4

[13]Lee S, et al. VR-LUXA: a personalized STEM education platform using LLM-driven adaptation in virtual reality. Comput Educ. 2024 Jan; 208:104952. doi: 10.1016/j.compedu.2023.104952

[14]Wang T, et al. NUXA: a narrative UX adapter for open-world VR games using large context window LLMs. In: Proceedings of the CHI Conference on Human Factors in Computing Systems. Honolulu (HI), USA; 2024. p. 1-16. doi: 10.1145/3613904.3642407

[15]Chen X, et al. VR-LCG: a safety-aware LLM-based content generation system for virtual chemistry labs. J Chem Educ. 2024 Mar;101(3):987-95. doi: 10.1021/acs.jchemed.3c00987

[16]Liu H, et al. VR-CG: an LLM-powered framework for generative urban planning in virtual reality. Landsc Urban Plan. 2024 Apr; 244:104998. doi: 10.1016/j.landurbplan.2023.104998

[17]Park J, et al. VR-MTG: an LLM-VR integration for generating and personalizing industrial maintenance training tasks. IEEE Trans Learn Technol. 2024; 17:567-80. doi: 10.1109/TLT.2024.3366011

[18]Miller R, et al. The impact of interaction latency on user trust and performance in LLM-driven virtual reality systems. In: 2024 IEEE Conference on Virtual Reality and 3D User Interfaces Abstracts and Workshops (VRW). Orlando (FL), USA; 2024. p. 943-4. doi: 10.1109/VRW62533.2024.00282

Downloads

Published

15-03-2026

Issue

Section

Articles

How to Cite

Hu, M. (2026). Large Language Models as A Paradigm Shift in Next-Generation Virtual Reality Interaction: A Comprehensive Investigation. Mathematical Modeling and Algorithm Application, 9(1), 246-251. https://doi.org/10.54097/hw3wve60