Main
A practical reading workflow:
- Start with surveys to build a taxonomy (tasks, sensors, environments, metrics).
- Pick one benchmark/dataset and trace which papers report results on it.
- For each method you care about, find an implementation and record reproducibility notes.
README excerpt (verbatim)
🤖 Awesome Embodied Robotics and Agent 
This is a curated list of "Embodied robotics or agent with Vision-Language Models (VLMs) and Large Language Models (LLMs)" research which is maintained by haonan.
Watch this repository for the latest updates and feel free to raise pull requests if you find some interesting papers!
News🔥
[2026/05/11] 🎉 Add NavSpace: How Intelligent Agents Follow Spatial Intelligence Instructions (ICRA 2026), the first benchmark for evaluating spatial intelligence in embodied navigation, with open-sourced dataset, evaluation code, and baseline SNav. [arXiv] [Github]
[2025/10/30] 🎉 Our survey paper "A Survey on Efficient Vision-Language-Action Models" [arXiv] has been released!
[2025/04/23] Add π-0.5, a lightweight and modular framework designed to integrate perception, control, and learning directly within physical systems.
[2025/03/18] Add some popular vision-language action (VLA) models. 🦾
[2024/06/28] Created a new board about agent self-evolutionary research. 🤖
[2024/06/07] Add Mobile-Agent-v2, a mobile device operation assistant with effective navigation via multi-agent collaboration. 🚀
[2024/05/13] Add "Learning Interactive Real-World Simulators"——outstanding paper award in ICLR 2024 🥇.
[2024/04/24] Add "A Survey on Self-Evolution of Large Language Models", a systematic survey on self-evolution in LLMs! 💥
[2024/04/16] Add some CVPR 2024 papers.
[2024/04/15] Add MetaGPT, accepted for oral presentation (top 1.2%) at ICLR 2024, ranking #1 in the LLM-based Agent category. 🚀
[2024/03/13] Add CRADLE, an interesting paper exploring LLM-based agent in Red Dead Redemption II!🎮