论文标题
基于模型的模仿学习城市驾驶
Model-Based Imitation Learning for Urban Driving
论文作者
论文摘要
环境的准确模型和作用的动态代理为改善运动计划提供了巨大的潜力。我们展示英里:一种基于模型的模仿学习方法,可以共同学习世界模型和自动驾驶政策。我们的方法将3D几何形状作为归纳偏见,并直接从专家演示的高分辨率视频中学习了高度紧凑的潜在空间。我们的模型经过了脱机城市驾驶数据的培训,而没有与环境进行任何在线互动。当部署在一个全新的城镇和新的天气状况时,Mile在Carla模拟器上的驾驶得分和新的天气状况时,Mile的驾驶得分提高了31%。我们的模型可以预测各种和合理的状态和行动,这些状态和行动可以解释为鸟眼观察语义分割。此外,我们证明它可以从完全预测的计划中执行复杂的驾驶操作。我们的方法是在城市驾驶环境中建模静态场景,动态场景和自我行为的第一个仅使用相机的方法。代码和型号的权重可从https://github.com/wayveai/mile获得。
An accurate model of the environment and the dynamic agents acting in it offers great potential for improving motion planning. We present MILE: a Model-based Imitation LEarning approach to jointly learn a model of the world and a policy for autonomous driving. Our method leverages 3D geometry as an inductive bias and learns a highly compact latent space directly from high-resolution videos of expert demonstrations. Our model is trained on an offline corpus of urban driving data, without any online interaction with the environment. MILE improves upon prior state-of-the-art by 31% in driving score on the CARLA simulator when deployed in a completely new town and new weather conditions. Our model can predict diverse and plausible states and actions, that can be interpretably decoded to bird's-eye view semantic segmentation. Further, we demonstrate that it can execute complex driving manoeuvres from plans entirely predicted in imagination. Our approach is the first camera-only method that models static scene, dynamic scene, and ego-behaviour in an urban driving environment. The code and model weights are available at https://github.com/wayveai/mile.