Neighbor communities
0 / 0 papers shown
Top Contributors
| Name | # Papers | # Citations |
|---|---|---|
Social Events
| Date | Location | Event |
|---|---|---|
| Name | # Papers | # Citations |
|---|---|---|
| Date | Location | Event |
|---|---|---|
Investigates how language models can be integrated into robotic systems to enhance interaction, command interpretation, and autonomous decision-making.
Instruction-based Image Editing with Planning, Reasoning, and Generation Liya Ji Chenyang Qi Qifeng Chen | |||
Toward Personalized LLM-Powered Agents: Foundations, Evaluation, and Future Directions Yue Xu Qian Chen Zizhan Ma Dongrui Liu Wenxuan Wang Xiting Wang Li Xiong Wenjie Wang | |||
Cognitive Models and AI Algorithms Provide Templates for Designing Language Agents Ryan Liu Dilip Arumugam Cedegao E. Zhang Sean Escola Xaq Pitkow Thomas L. Griffiths | |||
MindDriver: Introducing Progressive Multimodal Reasoning for Autonomous Driving Lingjun Zhang Yujian Yuan Changjie Wu Xinyuan Chang Xin Cai Shuang Zeng Linzhe Shi Sijin Wang Hang Zhang Mu Xu | |||
Architecting AgentOS: From Token-Level Context to Emergent System-Level Intelligence ChengYou Li XiaoDong Liu XiangBao Meng XinYu Zhao | |||
How Foundational Skills Influence VLM-based Embodied Agents:A Native Perspective Bo Peng Pi Bu Keyu Pan Xinrun Xu Yinxiu Zhao Miao Chen Yang Du Lin Li Jun Song Tong Xu | |||
LongVideo-R1: Smart Navigation for Low-cost Long Video Understanding Jihao Qiu Lingxi Xie Xinyue Huo Qi Tian Qixiang Ye | |||
Inner Speech as Behavior Guides: Steerable Imitation of Diverse Behaviors for Human-AI coordination Rakshit Trivedi Kartik Sharma David C Parkes | |||
Grounding LLMs in Scientific Discovery via Embodied Actions Bo Zhang Jinfeng Zhou Yuxuan Chen Jianing Yin Minlie Huang Hongning Wang | |||
Tool-R0: Self-Evolving LLM Agents for Tool-Learning from Zero Data Emre Can Acikgoz Cheng Qian Jonas Hübotter Heng Ji Dilek Hakkani-Tür Gokhan Tur | |||
HALO: A Unified Vision-Language-Action Model for Embodied Multimodal Chain-of-Thought Reasoning Quanxin Shou Fangqi Zhu Shawn Chen Puxin Yan Zhengyang Yan ...Zicong Hong Ruikai Shi Hao Huang Jie Zhang Song Guo | |||
NovaPlan: Zero-Shot Long-Horizon Manipulation via Closed-Loop Video Language Planning Jiahui Fu Junyu Nan Lingfeng Sun Hongyu Li Jianing Qian Jennifer L. Barry Kris Kitani George Konidaris | |||
Vinedresser3D: Agentic Text-guided 3D Editing Yankuan Chi Xiang Li Zixuan Huang James M. Rehg | |||
The Price Is Not Right: Neuro-Symbolic Methods Outperform VLAs on Structured Long-Horizon Manipulation Tasks with Significantly Lower Energy Consumption Timothy Duggan Pierrick Lorang Hong Lu Matthias Scheutz | |||
MagicAgent: Towards Generalized Agent Planning Xuhui Ren Shaokang Dong Chen Yang Qing Gao Yunbin Zhao ...Xuanjing Huang Yu-Gang Jiang Tao Gui Qi Zhang Yunke Zhang | |||
VIGiA: Instructional Video Guidance via Dialogue Reasoning and Retrieval Diogo Glória-Silva David Semedo João Maglhães | |||
Global Commander and Local Operative: A Dual-Agent Framework for Scene Navigation Kaiming Jin Yuefan Wu Shengqiong Wu Bobo Li Shuicheng Yan Tat-Seng Chua | |||
CapNav: Benchmarking Vision Language Models on Capability-conditioned Indoor Navigation Xia Su Ruiqi Chen Benlin Liu Jingwei Ma Zonglin Di Ranjay Krishna Jon Froehlich | |||
SimVLA: A Simple VLA Baseline for Robotic Manipulation Yuankai Luo Woping Chen Tong Liang Baiqiao Wang Zhenguo Li | |||
Modeling Distinct Human Interaction in Web Agents Faria Huq Zora Zhiruo Wang Zhanqiu Guo Venu Arvind Arangarajan Tianyue Ou Frank Xu Shuyan Zhou Graham Neubig Jeffrey P. Bigham | |||
MALLVI: A Multi-Agent Framework for Integrated Generalized Robotics Manipulation Iman Ahmadi Mehrshad Taji Arad Mahdinezhad Kashani AmirHossein Jadidi Saina Kashani Babak Khalaj | |||
One Agent to Guide Them All: Empowering MLLMs for Vision-and-Language Navigation via Explicit World Representation Zerui Li Hongpei Zheng Fangguo Zhao Aidan Chan Jian Zhou Sihao Lin Shijie Li Qi Wu | |||
World Action Models are Zero-shot Policies Seonghyeon Ye Yunhao Ge Kaiyuan Zheng Shenyuan Gao Sihyun Yu ...Scott Reed Jan Kautz Yuke Zhu Linxi "Jim" Fan Joel Jang | |||
DM0: An Embodied-Native Vision-Language-Action Model towards Physical AI En Yu Haoran Lv Jianjian Sun Kangheng Lin Ruitao Zhang ...Wenbin Tang Xiangyu Zhang Zheng Ge Erjin Zhou Tiancai Wang | |||
Mobile-Agent-v3.5: Multi-platform Fundamental GUI Agents Haiyang Xu Xi Zhang Haowei Liu Junyang Wang Zhaozai Zhu ...Ze Xu Shuai Bai Junyang Lin Jingren Zhou Ming Yan | |||
Plan-MCTS: Plan Exploration for Action Exploitation in Web Navigation Weiming Zhang Jihong Wang Jiamu Zhou Qingyao Li Xinbei Ma ...Weiwen Liu Zhuosheng Zhang Jun Wang Yong Yu Weinan Zhang | |||
Ontological grounding for sound and natural robot explanations via large language models Alberto Olivares-Alarcos Muhammad Ahsan Satrio Sanjaya Hsien-I Lin Guillem Alenyà | |||
AgentRob: From Virtual Forum Agents to Hijacked Physical Robots Wenrui Liu Yaxuan Wang Xun Zhang Yanshu Wang Jiashen Wei ...Xinyang Chen Hengzhe Sun Jiyu Shen Jingjing He Tong Yang | |||
UniManip: General-Purpose Zero-Shot Robotic Manipulation with Agentic Operational Graph Haichao Liu Yuanjiang Xue Yuheng Zhou Haoyuan Deng Yinan Liang Lihua Xie Ziwei Wang | |||
Scaling Single Human Demonstrations for Imitation Learning using Generative Foundational Models Nick Heppert Minh Quang Nguyen Abhinav Valada | |||
How Do We Research Human-Robot Interaction in the Age of Large Language Models? A Systematic Review Yufeng Wang Yuan Xu Anastasia Nikolova Yuxuan Wang Jianyu Wang Chongyang Wang Xin Tong | |||
Xiaomi-Robotics-0: An Open-Sourced Vision-Language-Action Model with Real-Time Execution Rui Cai Jun Guo Xinze He Piaopiao Jin Jie Li ...Diyun Xiang Yu Yang Hangjun Ye Yuan Zhang Quanyun Zhou | |||
RynnBrain: Open Embodied Foundation Models Ronghao Dang Jiayan Guo Bohan Hou Sicong Leng Kehan Li ...Wenqiao Zhang Chengju Liu Jianfei Yang Shijian Lu Deli Zhao | |||
Agentic AI for Robot Control: Flexible but still Fragile Oscar Lima Marc Vinci Martin Günther Marian Renz Alexander Sung ...Zongyao Yi Felix Igelbrink Benjamin Kisliuk Martin Atzmueller Joachim Hertzberg | |||
In-Context Autonomous Network Incident Response: An End-to-End Large Language Model Agent Approach Yiran Gao Kim Hammar Tao Li | |||
Steerable Vision-Language-Action Policies for Embodied Reasoning and Hierarchical Control William Chen Jagdeep Singh Bhatia Catherine Glossop Nikhil Mathihalli Ria Doshi Andy Tang Danny Driess Karl Pertsch Sergey Levine | |||
Budget-Constrained Agentic Large Language Models: Intention-Based Planning for Costly Tool Use Hanbing Liu Chunhao Tian Nan An Ziyuan Wang Pinyan Lu Changyuan Yu Qi Qi | |||
ABot-N0: Technical Report on the VLA Foundation Model for Versatile Embodied Navigation Zedong Chu Shichao Xie Xiaolong Wu Yanfen Shen Minghua Luo ...Xiangpo Yang Menglin Yang Hongguang Xing Weiguo Li Mu Xu | |||
LAMP: Implicit Language Map for Robot NavigationIEEE Robotics and Automation Letters (IEEE RA-L), 2025 Sibaek Lee Hyeonwoo Yu Giseop Kim Sunwook Choi | |||
GigaBrain-0.5M*: a VLA That Learns From World Model-Based Reinforcement Learning GigaBrain Team Boyuan Wang Bohan Li Chaojun Ni Guan Huang ...Yilong Li Yukun Zhou Yun Ye Zhichao Liu Zheng Zhu | |||
Any House Any Task: Scalable Long-Horizon Planning for Abstract Human Tasks Zhihong Liu Yang Li Rengming Huang Cewu Lu Panpan Cai | |||
HoloBrain-0 Technical Report Xuewu Lin Tianwei Lin Yun Du Hongyu Xie Yiwei Jin ...Ziang Li Chaodong Huang Hongzhe Bi Lichao Huang Zhizhong Su | |||
3DGSNav: Enhancing Vision-Language Model Reasoning for Object Navigation via Active 3D Gaussian Splatting Wancai Zheng Hao Chen Xianlong Lu Linlin Ou Xinyi Yu | |||
MolmoSpaces: A Large-Scale Open Ecosystem for Robot Navigation and Manipulation Yejin Kim Wilbert Pumacay Omar Rayyan Max Argus Winson Han ...Georgia Chalvatzaki Yuchen Cui Ali Farhadi Dieter Fox Ranjay Krishna | |||
LAP: Language-Action Pre-Training Enables Zero-shot Cross-Embodiment Transfer Lihan Zha Asher J. Hancock Mingtong Zhang Tenny Yin Yixuan Huang Dhruv Shah Allen Z. Ren Anirudha Majumdar | |||
Scaling World Model for Hierarchical Manipulation Policies Qian Long Yueze Wang Jiaxi Song Junbo Zhang Peiyan Li ...Xinlong Wang Zhongyuan Wang Xuguang Lan Huaping Liu Xinghang Li | |||
Say, Dream, and Act: Learning Video World Models for Instruction-Driven Robot Manipulation Songen Gu Yunuo Cai Tianyu Wang Simo Wu Yanwei Fu | |||
Active Zero: Self-Evolving Vision-Language Models through Active Environment Exploration Jinghan He Junfeng Fang Feng Xiong Zijun Yao Fei Shen Haiyun Guo Jinqiao Wang Tat-Seng Chua | |||
| Name (-) |
|---|
| Name (-) |
|---|
| Name (-) |
|---|
| Date | Location | Event | |
|---|---|---|---|
| No social events available | |||