RT-2: 视觉-语言-动作模型在机器人控制中的迁移RT-2: Vision-Language-Action Models Transfer Web Knowledge to Robotic Control2023-07-28|CoRL 2023VLMVLA控制Google DeepMind提出RT-2,将大规模VLM直接微调为机器人动作策略模型(VLA),展示了互联网知识向具身智能迁移的可行性。arXiv