My caption

When Digital Twins Meet Large Language Models: Realistic, Interactive, and Editable Simulation for Autonomous Driving

Publication
IEEE Robotics and Automation Magazine (RA-M), 2026, DOI: 10.1109/MRA.2026.3683252

Abstract: Simulation frameworks have been key enablers for the development and validation of autonomous driving systems. However, existing methods struggle to comprehensively address the autonomy-oriented requirements of balancing 1) dynamical fidelity, 2) photorealistic rendering, 3) context-relevant scenario orchestration, and 4) real-time performance. To address these limitations, we present a unified framework for creating and curating high-fidelity digital twins to accelerate advancements in autonomous driving research. Our framework leverages a mix of physics-based and data-driven techniques for developing and simulating digital twins of autonomous vehicles and their operating environments. It is capable of reconstructing real-world scenes and assets with geometric and photorealistic accuracy (∼97% structural similarity) and infusing them with physical properties to enable real-time (>60-Hz) dynamical simulation of the ensuing driving scenarios. Additionally, it incorporates a large language model (LLM) interface to flexibly edit the driving scenarios online via natural language prompts, with ∼85% generalizability and ∼95% repeatability. Finally, an optional vision language model (VLM) provides ∼80% visual enhancement by blending the hybrid scene composition.

Related