Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                

DriveVLM: The Convergence of Autonomous Driving and Large Vision-Language Models

Xiaoyu Tian1\phantom{}{}^{1}\phantom{}{}^{*}start_FLOATSUPERSCRIPT 1 end_FLOATSUPERSCRIPT start_FLOATSUPERSCRIPT ∗ end_FLOATSUPERSCRIPT   Junru Gu1\phantom{}{}^{1}\phantom{}{}^{*}start_FLOATSUPERSCRIPT 1 end_FLOATSUPERSCRIPT start_FLOATSUPERSCRIPT ∗ end_FLOATSUPERSCRIPT   Bailin Li2\phantom{}{}^{2}\phantom{}{}^{*}start_FLOATSUPERSCRIPT 2 end_FLOATSUPERSCRIPT start_FLOATSUPERSCRIPT ∗ end_FLOATSUPERSCRIPT   Yicheng Liu1\phantom{}{}^{1}\phantom{}{}^{*}start_FLOATSUPERSCRIPT 1 end_FLOATSUPERSCRIPT start_FLOATSUPERSCRIPT ∗ end_FLOATSUPERSCRIPT   Yang Wang22\phantom{}{}^{2}start_FLOATSUPERSCRIPT 2 end_FLOATSUPERSCRIPT   Zhiyong Zhao22\phantom{}{}^{2}start_FLOATSUPERSCRIPT 2 end_FLOATSUPERSCRIPT    Kun Zhan22\phantom{}{}^{2}start_FLOATSUPERSCRIPT 2 end_FLOATSUPERSCRIPT   Peng Jia22\phantom{}{}^{2}start_FLOATSUPERSCRIPT 2 end_FLOATSUPERSCRIPT   Xianpeng Lang22\phantom{}{}^{2}start_FLOATSUPERSCRIPT 2 end_FLOATSUPERSCRIPT   Hang Zhao1\phantom{}{}^{1}\phantom{}{}^{\text{\textdagger}}start_FLOATSUPERSCRIPT 1 end_FLOATSUPERSCRIPT start_FLOATSUPERSCRIPT † end_FLOATSUPERSCRIPT
11\phantom{}{}^{1}start_FLOATSUPERSCRIPT 1 end_FLOATSUPERSCRIPTIIIS, Tsinghua University   22\phantom{}{}^{2}start_FLOATSUPERSCRIPT 2 end_FLOATSUPERSCRIPTLi Auto
Abstract

A primary hurdle of autonomous driving in urban environments is understanding complex and long-tail scenarios, such as challenging road conditions and delicate human behaviors. We introduce DriveVLM, an autonomous driving system leveraging Vision-Language Models (VLMs) for enhanced scene understanding and planning capabilities. DriveVLM integrates a unique combination of reasoning modules for scene description, scene analysis, and hierarchical planning. Furthermore, recognizing the limitations of VLMs in spatial reasoning and heavy computational requirements, we propose DriveVLM-Dual, a hybrid system that synergizes the strengths of DriveVLM with the traditional autonomous driving pipeline. Experiments on both the nuScenes dataset and our SUP-AD dataset demonstrate the efficacy of DriveVLM and DriveVLM-Dual in handling complex and unpredictable driving conditions. Finally, we deploy the DriveVLM-Dual on a production vehicle, verifying it is effective in real-world autonomous driving environments.

{NoHyper}footnotetext: Equal contribution. Listing order is random.{NoHyper}footnotetext: {}^{\text{\textdagger}}start_FLOATSUPERSCRIPT † end_FLOATSUPERSCRIPTCorresponding to: hangzhao@mail.tsinghua.edu.cnfootnotetext: Project page: https://tsinghua-mars-lab.github.io/DriveVLM/

Keywords: Autonomous Driving, Vision Language Model, Dual System

1 Introduction

Autonomous driving, with its great promise to revolutionize transportation, has been an active research area over the past two decades. A primary hurdle to a fully autonomous driving system is scene understanding [1], which involves navigating complex, unpredictable scenarios such as adverse weather, intricate road layouts, and unforeseen human behaviors.

Existing autonomous driving systems, typically comprising 3D perception, motion prediction, and planning, struggle with these scene understanding challenges. Specifically, 3D perception [2, 3, 4, 5] is limited to detecting and tracking familiar objects, omitting rare objects and their unique attributes; motion prediction [6, 7, 8, 9, 10] and planning [11, 12, 13] focus on trajectory-level actions, often neglecting the decision-level interactions between objects and vehicles.

We introduce DriveVLM, a novel autonomous driving system that aims at these scene understanding challenges, capitalizing on the recent Vision-Language Models (VLMs) [14, 15, 16, 17] which have demonstrated exceptional prowess in visual comprehension and reasoning. Specifically, DriveVLM contains a Chain-of-Though (CoT) process with three key modules: scene description, scene analysis, and hierarchical planning. The scene description module linguistically depicts the driving environment and identifies critical objects in the scene; the scene analysis module delves into the characteristics of the critical objects and their influence on the ego vehicle; the hierarchical planning module formulates plans step-by-step, from meta-actions and decision descriptions to waypoints. These modules respectively correspond to the components of the traditional perception-prediction-planning pipeline, but the difference is that these modules tackle object perception, intention-level prediction and task-level planning, which were extremely challenging to cope with in the past.

While VLMs excel in visual understanding, they have limitations in spatial grounding and reasoning, and their computational intensity poses challenges for onboard inference speed. Therefore we further propose DriveVLM-Dual, a hybrid system that combines the strengths of both DriveVLM and traditional systems. DriveVLM-Dual optionally integrates DriveVLM with traditional 3D perception and planning modules, such as 3D object detectors, occupancy networks, and motion planners, enabling the system to achieve 3D grounding and high-frequency planning abilities. This dual system design, akin to the human brain’s slow and fast thinking processes, adapts efficiently to varying complexity in driving scenarios.

Meanwhile, we formally define the scene understanding and planning (SUP) task, and propose new evaluation metrics to assess the scene analysis and meta-action planning capabilities of DriveVLM and DriveVLM-Dual. We carry out a comprehensive data mining and annotation pipeline to construct an in-house SUP-AD dataset for the SUP task. Extensive experiments on both the nuScenes dataset and our own dataset demonstrate the superior performance of DriveVLM, particularly in few-shot scenarios. Furthermore, DriveVLM-Dual exceeds state-of-the-art end-to-end motion planning methods. We have also deployed the model on a production vehicle, confirming that DriveVLM-Dual is effective in real-world autonomous driving environments. Additionally, we have included a demo in the supplementary materials.

In summary, the contribution of this paper is three-fold:

  1. 1.

    We introduce DriveVLM, a novel autonomous driving system that leverages VLMs for effective scene understanding and planning. We further introduce DriveVLM-Dual, a hybrid system that incorporates DriveVLM and a traditional autonomous pipeline, which achieves improved spatial reasoning and real-time planning capabilities.

  2. 2.

    We present a comprehensive data mining and annotation pipeline to construct a scene understanding and planning dataset (SUP-AD), together with metrics for evaluation.

  3. 3.

    We have successfully deployed DriveVLM-Dual system in a production vehicle and test various effective strategies for accelerating VLM deployment in real driving scenarios.

2 Related Works

Vision-Language Models (VLMs). Recently, there has been a surge in research on large Vision-Language Models (VLMs), exemplified by works such as MiniGPT-4 [16], LLaVA [17], Qwen-VL [18], and others [19, 14, 20, 21]. VLMs can be used in various scenarios, especially robotics [22, 23, 24, 25, 26], where VLMs output corresponding actions that can be high-level instructions [22] or low-level robot actions [24]. DriveVLM focuses on utilizing VLMs to assist in autonomous driving, thereby establishing a novel framework. A Concurrent work [15] shares a similar motivation.

Learning-based Planning. The integration of learning frameworks into motion planning has been an active area of research since Pomerleau [11] pioneering contributions. One promising line of work is Reinforcement learning and imitation learning [27, 28, 29]. These methods can learn an end-to-end planning policy that directly maps raw sensory inputs to control actions [29]. Several works [30, 31, 32, 33] improve interpretability by explicitly building dense cost maps derived from learning-based modules. A recent trend involves training multiple blocks in an end-to-end fashion [32, 33, 34, 35]. These methods enhance overall performance, but rely on backpropagation from future trajectory predictions loss in a less interpretable decision-making process [36].

Driving Caption Datasets. Recent works [37, 15, 38] argue that language captions are an important medium to connect human knowledge with the driving objective, helping to inform decisions and actions. Refer-KITTI [39] annotates objects in the KITTI dataset [40] with language prompts that can reference a collection of objects. Talk2Car [41], NuPrompt [42] and nuScenes-QA [43] introduce free-form captions and QA annotation to the nuScenes dataset [44]. BDD-X [45] and BDD-OIA [46] offer datasets with language explanations for the ego vehicle’s actions or traffic scenarios [47] [48]. These datasets offer scenes for natural language use, but lack sufficient data on critical safety scenarios in self-driving systems.

Refer to caption
Figure 1: DriveVLM and DriveVLM-Dual model pipelines. DriveVLM takes images as input and, through a Chain-of-Thought (CoT) mechanism, outputs scene description, scene analysis, and hierarchical planning results. DriveVLM-Dual further incorporates traditional 3D perception and trajectory planning modules to achieve spatial reasoning capability and real-time trajectory planning.

3 DriveVLM

3.1 Overview

The overall pipeline of DriveVLM is illustrated in Figure 1. A sequence of images is processed by a Vision Language Model (VLM) to perform a special chain-of-thought (CoT) [49] reasoning to derive the driving planning results. The architecture of DriveVLM involves a vision transformer encoder [50] and a Large Language Model (LLM). The vision encoder produces image tokens; then an attention-based extractor aligns these tokens with the LLM. The reasoning process can be divided into three modules: scene description (Section 3.2), scene analysis (Section 3.3), and hierarchical planning (Section 3.4).

For real-world deployment, we propose a hybrid system, DriveVLM-Dual, in Section 3.5, which combines DriveVLM and the traditional autonomous driving pipeline, leveraging the strengths of both approaches.

3.2 Scene Description

The scene description module identifies driving environment description and critical objects.

Environment Description. Driving environments, such as weather and road conditions, have a non-negligible impact on driving difficulty. Therefore, the model is first prompted to output a linguistic description E𝐸Eitalic_E of the driving environment, including several conditions: E={Eweather,Etime,Eroad,Elane},𝐸subscript𝐸weathersubscript𝐸timesubscript𝐸roadsubscript𝐸laneE=\{E_{\text{weather}},E_{\text{time}},E_{\text{road}},E_{\text{lane}}\},italic_E = { italic_E start_POSTSUBSCRIPT weather end_POSTSUBSCRIPT , italic_E start_POSTSUBSCRIPT time end_POSTSUBSCRIPT , italic_E start_POSTSUBSCRIPT road end_POSTSUBSCRIPT , italic_E start_POSTSUBSCRIPT lane end_POSTSUBSCRIPT } , each representing a crucial aspect of the driving environment. The weather component, Eweathersubscript𝐸weatherE_{\text{weather}}italic_E start_POSTSUBSCRIPT weather end_POSTSUBSCRIPT, spans conditions from sunny to snowy, affecting visibility and traction. The time component, Etimesubscript𝐸timeE_{\text{time}}italic_E start_POSTSUBSCRIPT time end_POSTSUBSCRIPT, distinguishes between daytime and nighttime, impacting driving strategies due to visibility changes. Road types, Eroadsubscript𝐸roadE_{\text{road}}italic_E start_POSTSUBSCRIPT road end_POSTSUBSCRIPT, such as urban or highway, introduce different challenges, while lane conditions, Elanesubscript𝐸laneE_{\text{lane}}italic_E start_POSTSUBSCRIPT lane end_POSTSUBSCRIPT, focus on current lane positioning and possible maneuvers, crucial for safe driving decisions.

Critical Object Identification. In addition to environmental conditions, various objects in driving scenarios significantly influence driving behaviors. Unlike traditional autonomous driving perception modules, which detect all objects within a specific range, we solely focus on identifying critical objects that are most likely to influence the current scenario, inspired by human cognitive processes during driving. Each critical object, denoted as Ocsubscript𝑂𝑐O_{c}italic_O start_POSTSUBSCRIPT italic_c end_POSTSUBSCRIPT, contains two attributes: the object category c𝑐citalic_c and its approximate bounding box coordinates b(x1,y1,x2,y2)𝑏𝑥1𝑦1𝑥2𝑦2b(x1,y1,x2,y2)italic_b ( italic_x 1 , italic_y 1 , italic_x 2 , italic_y 2 ) on the image. The category and coordinates are mapped to their corresponding language token_id𝑡𝑜𝑘𝑒𝑛_𝑖𝑑token\_iditalic_t italic_o italic_k italic_e italic_n _ italic_i italic_d in the language modality, enabling seamless integration into the following modules. Moreover, taking advantage of the pre-trained vision encoder, DriveVLM can identify long-tail critical objects that may elude typical 3D object detectors, such as road debris or unusual animals.

3.3 Scene Analysis

In the traditional autonomous driving pipeline, the prediction module typically concentrates on forecasting the future trajectories of objects. The emergence of advanced vision-language models has provided us with the ability to perform a more comprehensive analysis of the current scene. The scene-level analysis summarizes all the critical objects together with the environmental description. This summary gives a comprehensive understanding of the scene, and is fed into the following planning module.

Critical Object Analysis. DriveVLM characterizes critical objects in three aspects: static attributes Cssubscript𝐶𝑠C_{s}italic_C start_POSTSUBSCRIPT italic_s end_POSTSUBSCRIPT, motion states Cmsubscript𝐶𝑚C_{m}italic_C start_POSTSUBSCRIPT italic_m end_POSTSUBSCRIPT, and particular behaviors Cbsubscript𝐶𝑏C_{b}italic_C start_POSTSUBSCRIPT italic_b end_POSTSUBSCRIPT. Static attributes Cssubscript𝐶𝑠C_{s}italic_C start_POSTSUBSCRIPT italic_s end_POSTSUBSCRIPT describe inherent properties of objects, such as a roadside billboard’s visual cues or a truck’s oversized cargo, which are critical in preempting and navigating potential hazards. Motion states Cmsubscript𝐶𝑚C_{m}italic_C start_POSTSUBSCRIPT italic_m end_POSTSUBSCRIPT describe an object’s dynamics over a period, including position, direction, and action—characteristics that are vital in predicting the object’s future trajectory and potential interactions with the ego vehicle. Particular behaviors Cbsubscript𝐶𝑏C_{b}italic_C start_POSTSUBSCRIPT italic_b end_POSTSUBSCRIPT refer to special actions or gestures of an object that could directly influence the ego vehicle’s next driving decisions. We do not require the model to analyze all three characteristics for all objects. In practice, only one or two characteristics apply to a critical object. Upon analyzing these characteristics, DriveVLM then predicts the potential influence I𝐼Iitalic_I of each critical object on the ego vehicle.

3.4 Hierarchical Planning

The scene-level summary is then combined with the route, ego pose and velocity to form a prompt for planning. Finally, DriveVLM progressively generates driving plans, in three stages: meta-actions, decision description, and trajectory waypoints.

Meta-actions A𝐴\boldsymbol{A}bold_italic_A. A meta-action, denoted as aisubscript𝑎𝑖a_{i}italic_a start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT, represents a short-term decision of the driving strategy. These actions fall into 17 categories, including but not limited to acceleration, deceleration, turning left, changing lanes, minor positional adjustments, and waiting. To plan the ego vehicle’s future maneuver over a certain period, we generate a sequence of meta-actions.

Decision Description D𝐷\boldsymbol{D}bold_italic_D. Decision description D𝐷Ditalic_D articulates the more fine-grained driving strategy the ego vehicle should adopt. It contains three elements: Action 𝒜𝒜\mathcal{A}caligraphic_A, Subject 𝒮𝒮\mathcal{S}caligraphic_S, and Duration 𝒟𝒟\mathcal{D}caligraphic_D. Action pertains to meta actions such as ‘turn’, ‘wait’, or ‘accelerate’. Subject refers to the interacting object, such as a pedestrian, a traffic signal, or a specific lane. Duration indicates the temporal aspect of the action, specifying how long it should be carried out or when it should start.

Trajectory Waypoints W𝑊\boldsymbol{W}bold_italic_W. Upon establishing the decision description D𝐷{D}italic_D, our next phase involves the generation of corresponding trajectory waypoints. These waypoints, denoted by W={w1,w2,,wn}𝑊subscript𝑤1subscript𝑤2subscript𝑤𝑛W=\{w_{1},w_{2},...,w_{n}\}italic_W = { italic_w start_POSTSUBSCRIPT 1 end_POSTSUBSCRIPT , italic_w start_POSTSUBSCRIPT 2 end_POSTSUBSCRIPT , … , italic_w start_POSTSUBSCRIPT italic_n end_POSTSUBSCRIPT }, wi=(xi,yi)subscript𝑤𝑖subscript𝑥𝑖subscript𝑦𝑖w_{i}=(x_{i},y_{i})italic_w start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT = ( italic_x start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT , italic_y start_POSTSUBSCRIPT italic_i end_POSTSUBSCRIPT ), depict the vehicle’s path over a certain future period with predetermined intervals ΔtΔ𝑡\Delta troman_Δ italic_t. We map these numerical waypoints into language tokens for auto-regressive generation.

3.5 DriveVLM-Dual

To mitigate the challenges of high latency and imprecise spatial and motion understanding in VLMs, we propose DriveVLM-Dual, a collaboration between DriveVLM and the traditional autonomous driving system. This novel approach involves two key strategies: incorporating 3D perception for critical object analysis, and high-frequency trajectory refinement.

Integrating 3D Perception. We represent objects detected by a 3D detector as O3D={c3Di,b3Di}subscript𝑂3Dsuperscriptsubscript𝑐3D𝑖superscriptsubscript𝑏3D𝑖O_{\text{3D}}=\{c_{\text{3D}}^{i},b_{\text{3D}}^{i}\}italic_O start_POSTSUBSCRIPT 3D end_POSTSUBSCRIPT = { italic_c start_POSTSUBSCRIPT 3D end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_i end_POSTSUPERSCRIPT , italic_b start_POSTSUBSCRIPT 3D end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_i end_POSTSUPERSCRIPT }, where b3Disuperscriptsubscript𝑏3D𝑖b_{\text{3D}}^{i}italic_b start_POSTSUBSCRIPT 3D end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_i end_POSTSUPERSCRIPT denotes the i𝑖iitalic_i-th bounding box and c3Disuperscriptsubscript𝑐3D𝑖c_{\text{3D}}^{i}italic_c start_POSTSUBSCRIPT 3D end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_i end_POSTSUPERSCRIPT denotes its category. These 3D bounding boxes are then back-projected onto 2D images to derive corresponding 2D bounding boxes b2Disuperscriptsubscript𝑏2D𝑖b_{\text{2D}}^{i}italic_b start_POSTSUBSCRIPT 2D end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_i end_POSTSUPERSCRIPT. We conduct IoU matching between these 2D bounding boxes b2Disuperscriptsubscript𝑏2D𝑖b_{\text{2D}}^{i}italic_b start_POSTSUBSCRIPT 2D end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_i end_POSTSUPERSCRIPT and bcjsuperscriptsubscript𝑏𝑐𝑗b_{c}^{j}italic_b start_POSTSUBSCRIPT italic_c end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_j end_POSTSUPERSCRIPT. bcjsuperscriptsubscript𝑏𝑐𝑗b_{c}^{j}italic_b start_POSTSUBSCRIPT italic_c end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_j end_POSTSUPERSCRIPT are the bounding boxes of previously identified critical objects Ocritical={ccj,bcj}subscript𝑂criticalsuperscriptsubscript𝑐𝑐𝑗superscriptsubscript𝑏𝑐𝑗O_{\text{critical}}=\{c_{c}^{j},b_{c}^{j}\}italic_O start_POSTSUBSCRIPT critical end_POSTSUBSCRIPT = { italic_c start_POSTSUBSCRIPT italic_c end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_j end_POSTSUPERSCRIPT , italic_b start_POSTSUBSCRIPT italic_c end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_j end_POSTSUPERSCRIPT }. We classify critical objects that meet a certain approximate IoU threshold and belong to the same category as matched critical objects Ocmatchedsuperscriptsubscript𝑂𝑐matchedO_{c}^{\text{matched}}italic_O start_POSTSUBSCRIPT italic_c end_POSTSUBSCRIPT start_POSTSUPERSCRIPT matched end_POSTSUPERSCRIPT, defined as

Ocmatched={ccj,bcj},superscriptsubscript𝑂𝑐matchedsuperscriptsubscript𝑐𝑐𝑗superscriptsubscript𝑏𝑐𝑗\displaystyle O_{c}^{\text{matched}}=\{c_{c}^{j},b_{c}^{j}\},\quaditalic_O start_POSTSUBSCRIPT italic_c end_POSTSUBSCRIPT start_POSTSUPERSCRIPT matched end_POSTSUPERSCRIPT = { italic_c start_POSTSUBSCRIPT italic_c end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_j end_POSTSUPERSCRIPT , italic_b start_POSTSUBSCRIPT italic_c end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_j end_POSTSUPERSCRIPT } , if ccj=c2Di and aIoU(bcj,b2Di)>τ, where aIoU(bcj,b2Di)=Sbcjb2DiSb2Di,formulae-sequenceif superscriptsubscript𝑐𝑐𝑗superscriptsubscript𝑐2D𝑖 and aIoUsuperscriptsubscript𝑏𝑐𝑗superscriptsubscript𝑏2D𝑖𝜏 where aIoUsuperscriptsubscript𝑏𝑐𝑗superscriptsubscript𝑏2D𝑖subscript𝑆superscriptsubscript𝑏𝑐𝑗superscriptsubscript𝑏2D𝑖subscript𝑆superscriptsubscript𝑏2D𝑖\displaystyle\text{if }c_{c}^{j}=c_{\text{2D}}^{i}\text{ and }\text{aIoU}(b_{c% }^{j},b_{\text{2D}}^{i})>\tau,\text{ where aIoU}(b_{c}^{j},b_{\text{2D}}^{i})=% \frac{S_{b_{c}^{j}\cap b_{\text{2D}}^{i}}}{S_{b_{\text{2D}}^{i}}},if italic_c start_POSTSUBSCRIPT italic_c end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_j end_POSTSUPERSCRIPT = italic_c start_POSTSUBSCRIPT 2D end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_i end_POSTSUPERSCRIPT and roman_aIoU ( italic_b start_POSTSUBSCRIPT italic_c end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_j end_POSTSUPERSCRIPT , italic_b start_POSTSUBSCRIPT 2D end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_i end_POSTSUPERSCRIPT ) > italic_τ , where aIoU ( italic_b start_POSTSUBSCRIPT italic_c end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_j end_POSTSUPERSCRIPT , italic_b start_POSTSUBSCRIPT 2D end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_i end_POSTSUPERSCRIPT ) = divide start_ARG italic_S start_POSTSUBSCRIPT italic_b start_POSTSUBSCRIPT italic_c end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_j end_POSTSUPERSCRIPT ∩ italic_b start_POSTSUBSCRIPT 2D end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_i end_POSTSUPERSCRIPT end_POSTSUBSCRIPT end_ARG start_ARG italic_S start_POSTSUBSCRIPT italic_b start_POSTSUBSCRIPT 2D end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_i end_POSTSUPERSCRIPT end_POSTSUBSCRIPT end_ARG ,

Those critical objects without a corresponding match in the 3D data are noted as Ocunmatchedsuperscriptsubscript𝑂𝑐unmatchedO_{c}^{\text{unmatched}}italic_O start_POSTSUBSCRIPT italic_c end_POSTSUBSCRIPT start_POSTSUPERSCRIPT unmatched end_POSTSUPERSCRIPT.

In the scene analysis module, for Ocmatchedsuperscriptsubscript𝑂𝑐matchedO_{c}^{\text{matched}}italic_O start_POSTSUBSCRIPT italic_c end_POSTSUBSCRIPT start_POSTSUPERSCRIPT matched end_POSTSUPERSCRIPT, the center coordinates, orientations, and historical trajectories of the corresponding 3D objects are used as language prompts for the model, assisting in object analysis. Conversely, for Ocunmatchedsuperscriptsubscript𝑂𝑐unmatchedO_{c}^{\text{unmatched}}italic_O start_POSTSUBSCRIPT italic_c end_POSTSUBSCRIPT start_POSTSUPERSCRIPT unmatched end_POSTSUPERSCRIPT, analysis relies solely on the language tokens derived from the image. This design enables DriveVLM-Dual to understand the locations and motions of critical objects more accurately, enhancing the overall performance.

High-frequency Trajectory Refinement. To achieve real-time, high-frequency inference capabilities, we integrate it with a conventional planner to form a slow-fast dual system, combining the advanced capabilities of DriveVLM with the efficiency of traditional planning methods. After obtaining a trajectory from DriveVLM at low frequency, denoted as Wslowsubscript𝑊slowW_{\text{slow}}italic_W start_POSTSUBSCRIPT slow end_POSTSUBSCRIPT, we take it as a reference trajectory for a classical planner for high-frequency trajectory refinement. In the case of an optimization-based planner, Wslowsubscript𝑊slowW_{\text{slow}}italic_W start_POSTSUBSCRIPT slow end_POSTSUBSCRIPT serves as the initial solution for the optimization solver. For a neural network-based planner, Wslowsubscript𝑊slowW_{\text{slow}}italic_W start_POSTSUBSCRIPT slow end_POSTSUBSCRIPT is used as an input query, combined with additional input features f𝑓fitalic_f, and then decoded into a new planning trajectory denoted as Wfastsubscript𝑊fastW_{\text{fast}}italic_W start_POSTSUBSCRIPT fast end_POSTSUBSCRIPT. The formulation of this process can be described as:

Wfast=Planner([Wslow,f]).subscript𝑊fastPlannersubscript𝑊slow𝑓\displaystyle W_{\text{fast}}=\text{Planner}([W_{\text{slow}},f]).italic_W start_POSTSUBSCRIPT fast end_POSTSUBSCRIPT = Planner ( [ italic_W start_POSTSUBSCRIPT slow end_POSTSUBSCRIPT , italic_f ] ) . (1)

This refinement step ensures that the trajectory produced by DriveVLM-Dual (1) achieves higher trajectory quality, and (2) meets real-time requirements. In practice, the two branches operate asynchronously in a slow-fast manner, where the planner module in the traditional autonomous driving branch can selectively receive trajectory from the VLM branch as additional input.

4 Task and Dataset

To fully exploit the potential of DriveVLM and DriveVLM-Dual in handling complex and long-tail driving scenarios, we formally define a task called Scene Understanding for Planning (Section 4.1), together with a set of evaluation metrics (Section 4.2). Furthermore, we propose a data mining and annotation protocol to curate a scene understanding and planning dataset (Section 4.3).

Refer to caption
Figure 2: An annotated sample of the SUP-AD dataset.

4.1 Task Definition

The Scene Understanding for Planning task is defined as follows. The input comprises multi-view videos 𝒱𝒱\mathcal{V}caligraphic_V from surrounding cameras and optionally 3D perception results 𝒫𝒫\mathcal{P}caligraphic_P from a perception module. The output includes the following components:

  1. 1.

    Scene Description E𝐸Eitalic_E: Composed of weather condition Eweathersubscript𝐸weatherE_{\text{weather}}italic_E start_POSTSUBSCRIPT weather end_POSTSUBSCRIPT, time Etimesubscript𝐸timeE_{\text{time}}italic_E start_POSTSUBSCRIPT time end_POSTSUBSCRIPT, road condition Eroadsubscript𝐸roadE_{\text{road}}italic_E start_POSTSUBSCRIPT road end_POSTSUBSCRIPT, and lane conditions Elanesubscript𝐸laneE_{\text{lane}}italic_E start_POSTSUBSCRIPT lane end_POSTSUBSCRIPT.

  2. 2.

    Scene Analysis S𝑆Sitalic_S: Including object-level analysis and scene-level summary S𝑆Sitalic_S.

  3. 3.

    Meta Actions A𝐴Aitalic_A: A sequence of actions representing task-level maneuvers.

  4. 4.

    Decision Description D𝐷Ditalic_D: A detailed account of the driving decisions.

  5. 5.

    Trajectory Waypoints W𝑊Witalic_W: The waypoints outlining the planned trajectory of the ego vehicle.

4.2 Evaluation Metrics

To comprehensively evaluate a model’s performance, we care about its interpretation of the driving scene and the decisions made. Therefore, our evaluation has two aspects: scene description/analysis evaluation and meta-action evaluation.

Scene Description/Analysis Evaluation. Given the subjective nature of human evaluation in scene description, we adopt a structured approach using a pre-trained LLM. This method entails comparing the generated scene description with a human-annotated ground truth description. The ground truth description encompasses structured data such as environmental conditions, navigation, lane information, and critical events with specific objects, verbs, and their influences. The LLM assesses and scores the generated descriptions based on their consistency with the ground truth.

Meta-action Evaluation. Meta-actions are a predefined set of decision-making options. A driving decision is formulated as a sequence of meta-actions. Our evaluation method employs a dynamic programming algorithm to compare the model-generated sequences with a manually annotated ground truth sequence. The evaluation should also weigh the relative importance of various meta-actions, designating some as ‘conservative actions’ with a lower impact on the sequence’s overall context. To increase robustness, we first use the LLM to generate semantically equivalent alternatives to the ground truth sequence to enhance robustness. The sequence with the highest similarity to these alternatives calculates the final driving decision score. More details of the proposed metric are available in the Appendix B.

4.3 Dataset Construction

We propose a comprehensive data mining and annotation pipeline, shown in Figure 3, to construct a Scene Understanding for Planning (SUP-AD) Dataset for the proposed task. Specifically, we perform long-tail object mining and challenging scenario mining from a large database to collect samples, then we select a keyframe from each sample and further perform scene annotation. Dataset statistics are available in the Appendix A.

Long-tail Object Mining. According to real-world road object distribution, we first define a list of long-tail object categories, such as weird-shaped vehicles, road debris, and animals crossing the road. Next, we mine these long-tail scenarios using a CLIP-based search engine, capable of mining driving data using language queries from a large collection of logs. Following that, we perform a manual inspection to filter out scenes inconsistent with the specified categories.

Challenging Scenario Mining. In addition to long-tail objects, we are also interested in challenging driving scenarios, where the driving strategy of the ego vehicle needs to be adapted according to the changing driving conditions. These scenarios are mined according to the variance of the recorded driving maneuvers.

Keyframe Selection. Each scene is a video clip, it is essential to identify the ‘keyframe’ to annotate. In most challenging scenarios, a keyframe is the moment before a significant change in speed or direction is required. We select this keyframe 0.5s to 1s earlier than the actual maneuver, based on comprehensive testing, to guarantee an optimal reaction time for decision-making. For scenes that do not involve changes in driving behavior, we select a frame that is relevant to the current driving scenario as the keyframe.

Scene Annotation. We employ a group of annotators to perform the scene annotation, including scene description, scene analysis, and planning, except for waypoints, which can be auto-labeled from the vehicle’s IMU recordings. To facilitate scene annotation, we make a video annotation tool with the following features: (1) the annotators can slide the progress bar back and forth to replay any part of a video; (2) while annotating a keyframe, the annotator can draw bounding boxes on the image together with language descriptions; (3) annotators can select from a list of action and decision candidates while annotating driving plans. Each annotation is meticulously verified by 3 annotators for accuracy and consistency, ensuring a reliable dataset for model training. Figure 2 illustrates a sample scenario with detailed annotations.

Refer to caption
Figure 3: The proposed data mining and annotation pipeline for constructing a scene understanding and planning dataset. Scenario examples randomly sampled from the dataset (below) demonstrate the diversity and complexity of the dataset.

5 Experiments

5.1 Settings

We test DriveVLM and DriveVLM-Dual on our proposed SUP-AD dataset and nuScenes dataset [44].

SUP-AD Dataset. The SUP-AD dataset is a dataset built by our proposed data mining and annotation pipeline. It is divided into train, validation, and test splits with a ratio of 7.5:1:1.5:7.51:1.57.5:1:1.57.5 : 1 : 1.5. We train models on the training split and use our proposed scene description and meta-action metrics to evaluate model performance on the test split. We also employ co-tuning with additional datasets to ensure the generalization of the LLM is not compromised.

nuScenes Dataset. The nuScenes dataset is a large-scale driving dataset of urban scenarios with 1000 scenes, where each scene lasts about 20 seconds. Following previous works [34, 51], we adopt Displacement Error (DE) and Collision Rate (CR) as metrics to evaluate models’ performance.

Table 1: Results on the test set of our proposed SUP-AD dataset. : Using the official API of GPT-4V. For Lynx and CogVLM, we utilize the training split for fine-tuning purposes. In contrast, for GPT-4V, we employ in-context learning.
{tblr}

colspec = l—cc, row1-Z = rowsep=0.6pt, hline1,2,6 = 0.8pt, Method & Scene Description Meta-actions
Fine-tuning w/ Lynx [14] 0.460.460.460.46 0.150.150.150.15
Fine-tuning w/ CogVLM [21] 0.490.490.490.49 0.220.220.220.22
GPT-4V [52] 0.380.380.380.38 0.190.190.190.19
DriveVLM w/ Qwen 0.71 0.37

Table 2: Planning results on the nuScenes validation dataset. DriveVLM-Dual achieves the best performance. denotes cooperating with VAD [51].

{tblr} colspec = l—cccc—cccc, rows = font=, row1-Z = rowsep=-2pt, hline1,12 = 0.6pt, hline3,10 = 0.5pt, cell11 = r=2l, cell12 = c=4, cell16 = c=4, Method & L2 (m) \downarrow Collision (%) \downarrow
1s 2s 3s Avg. 1s 2s 3s Avg.
NMP [30] - - 2.31 - - - 1.92 -
SA-NMP [30] - - 2.05 - - - 1.59 -
FF [32] 0.55 1.20 2.54 1.43 0.06 0.17 1.07 0.43
EO [53] 0.67 1.36 2.78 1.60 0.04 0.09 0.88 0.33
ST-P3 [54] 1.33 2.11 2.90 2.11 0.23 0.62 1.27 0.71
UniAD [34] 0.48 0.96 1.65 1.03 0.05 0.17 0.71 0.31
VAD-Base [51] 0.17 0.34 0.60 0.37 0.07 0.10 0.24 0.14
DriveVLM 0.18 0.34 0.68 0.40 0.10 0.22 0.45 0.27
DriveVLM-Dual 0.15 0.29 0.48 0.31 0.05 0.08 0.17 0.10

Co-tuning

To preserve the LLM’s generalization capabilities during the fine-tuning process, we employed co-tuning with several additional datasets. These include the Talk2Car [41], BDDX [45], Drama [48], SUTD [55], and LLAVA [17] datasets. For each dataset, we conducted random sampling in a 1:1 ratio corresponding to the data volume of the SUP-AD and nuScenes datasets. Following this co-tuning approach, we found that the scores on the SUP-AD dataset, under the evaluation metrics of Scene Description and Meta Action, remained virtually unchanged, simultaneously ensuring the preservation of the LLM’s original capabilities and its generalization capacity.

Base Model.

We use Qwen-VL [18] as our default large vision-language model, which exhibits remarkable performance in tasks like question answering, visual localization, and text recognition. It contains a total of 9.6 billion parameters, including a visual encoder (1.9 billion), a vision-language adapter (0.08 billion), and a large language model (Qwen, 7.7 billion). Images are resized to a resolution of 448×448448448448\times 448448 × 448 before being encoded by the vision encoder. During training, we randomly select a sequence of images at the current time T𝑇Titalic_Ts, T1𝑇1T-1italic_T - 1s, T2𝑇2T-2italic_T - 2s, and T3𝑇3T-3italic_T - 3s as input. The selected images ensure the inclusion of the current time frame and follow an ascending chronological order.

5.2 Main Results

SUP-AD. We present the performance of our proposed DriveVLM with several large vision-language models and compare them with GPT-4V, as shown in Table 1. DriveVLM, utilizing Qwen-VL as its backbone, achieves the best performance due to its strong capabilities in question answering and flexible interaction compared to the other open-source VLMs. Although GPT-4V exhibits robust capabilities in vision and language processing, its inability to undergo fine-tuning, restricting it solely to in-context learning, often results in the generation of extraneous information during scene description tasks. Under our evaluation metric, the additional information is frequently classified as hallucination, consequently leading to lower scores.

nuScenes. As shown in Table 2, DriveVLM-Dual achieves state-of-the-art performance on the nuScenes planning task when cooperating with VAD. It demonstrates that our method, although tailored for understanding complex scenes, also excels in ordinary scenarios.

5.3 Ablation Study

Model Design. To better understand the significance of our designed modules in DriveVLM, we conduct ablations on different combinations of modules, as shown in Table 3. The inclusion of critical object analysis enables our model to identify and prioritize important elements in the driving environment, enhancing the decision-making accuracy for safer navigation. Integrating 3D perception data, our model gains a refined understanding of the surroundings and achieves precise predictions.

Traditional AD Pipeline. To demonstrate the generalization of our dual system design, we test DriveVLM-Dual with different traditional autonomous driving pipelines on the validation set of nuScenes. As illustrated in Table 4, our proposed DriveVLM-Dual adapts well to different traditional AD pipelines. While a standalone MLP method shows a notable performance gap compared to VAD, both variants of DriveVLM-Dual achieve nearly identical performance, underscoring the efficacy and robustness of our dual system design.

Table 3: Ablations of design choices on the validation set of nuScenes. “Base” refers to only indicating the hierarchical planning results without our proposed CoT inference. “CO” represents the addition of critical object analysis. “3D” denotes the inclusion of 3D perception results as an auxiliary language prompt.

{tblr} colspec = l—ccc—cccc—cccc, rows = font=, row1-Z = rowsep=-2pt, hline1,3,6 = 0.6pt, cell11 = r=2c, cell12 = r=2c, cell13 = r=2c, cell14 = r=2c, cell15 = c=4, cell19 = c=4, ID & Base CO 3D L2 (m) \downarrow Collision (%) \downarrow
1s 2s 3s Avg. 1s 2s 3s Avg.
1 ✓ 0.19 0.41 0.89 0.49 0.16 0.28 0.63 0.36
2 ✓ ✓ 0.20 0.38 0.75 0.44 0.15 0.29 0.61 0.35
3 ✓ ✓ ✓ 0.18 0.34 0.68 0.40 0.10 0.22 0.45 0.27

Table 4: Ablations of traditional autonomous driving pipeline in DriveVLM-Dual. MLP stands for methods similar to AD-MLP [56].

{tblr} colspec = l—cccc—cccc, rows = font=, row1-Z = rowsep=-2pt, hline1,3,9 = 0.6pt, hline5,7 = 0.5pt, cell11 = r=2l, cell12 = c=4, cell16 = c=4, Method & L2 (m) \downarrow Collision (%) \downarrow
1s 2s 3s Avg. 1s 2s 3s Avg.
UniAD [34] 0.48 0.96 1.65 1.03 0.05 0.17 0.71 0.31
DriveVLM-Dual (UniAD) 0.17 0.37 0.63 0.39 0.08 0.18 0.35 0.20
MLP 0.25 0.46 0.62 0.44 0.14 0.18 0.28 0.20
DriveVLM-Dual (MLP) 0.14 0.35 0.30 0.31 0.09 0.13 0.18 0.13
VAD [51] 0.17 0.34 0.60 0.37 0.07 0.10 0.24 0.14
DriveVLM-Dual(VAD) 0.15 0.29 0.48 0.31 0.05 0.08 0.17 0.10

5.4 Qualitative Results

Qualitative results of DriveVLM are shown in Figure 4. In Figure 4(a), DriveVLM accurately predicts the current scene conditions and incorporates well-considered planning decisions regarding the cyclist approaching us. In Figure 4(b), DriveVLM effectively comprehends the gesture of the traffic police ahead, signaling the ego vehicle to proceed, and also considers the person riding a tricycle on the right side, thereby making sensible driving decisions. These qualitative results demonstrate our model’s exceptional ability to understand complex scenarios and make suitable driving plans. More visualization of our model’s output is shown in the Appendix C.

Refer to caption
(a)
Refer to caption
(b)
Figure 4: Qualitative results of DriveVLM. The orange curves represent the model’s planned future trajectories for the next 3 seconds.

6 Onboard Deployment and Testing

We deploy DriveVLM-Dual on an autonomous vehicle equipped with two OrinX processors, with a high-frequency end-to-end driving system on OrinX-1 and our DriveVLM on OrinX-2. These two systems operate and collaborate asynchronously. Furthermore, we optimize the DriveVLM and achieve an average inference speed of 410 ms on OrinX. The video demonstration of this dual-system deployment can be found at the following link: https://www.youtube.com/watch?v=MMCO0TLMT74.

In this section, we present several comparative experiments, which will provide insights for practical deployment on edge devices. Notice that all the VLMs are pre-trained and fine-tuned on common datasets as well as our proprietary datasets specifically designed for autonomous driving.

Base LLM Avg. MMMU [57] SEEDBench [58] RefCOCO [59] SUP-AD Drivelm-QA [60] Drivelm-Grounding [60] RealworldQA [61]
5% 20% 15% 15% 7.5% 7.5% 15%
MobileLLaMA1.4B [62] 0.457 0.331 0.590 0.421 0.520 0.686 0.735 0.501
Qwen-1.8B [63] 0.477 0.340 0.622 0.492 0.523 0.680 0.725 0.518
Gemma-2B [64] 0.439 0.345 0.571 0.330 0.510 0.680 0.721 0.507
MiniCPM-2.4B [65] 0.482 0.379 0.640 0.444 0.539 0.676 0.717 0.553
MobileLLaMA2.7B [62] 0.496 0.348 0.635 0.557 0.546 0.683 0.725 0.536
Phi3-3.8B [66] 0.538 0.435 0.688 0.608 0.604 0.697 0.743 0.592
Qwen-4B [63] 0.511 0.366 0.671 0.603 0.515 0.681 0.735 0.562
Qwen-4B 0.529 0.373 0.684 0.624 0.596 0.699 0.738 0.553
Table 5: Performance of different LLMs on various datasets using the LLAVA-1.5 [67] architecture with ViT-L-336 [50] as the image encoder. Note that indicates using SigLIP-L-384 [68] as the image encoder. Compared to general capabilities, we are more concerned with the performance of VLMs in abilities that are more relevant to autonomous driving. Therefore, we set different score weights for various benchmarks.
LLM Promt Length(toks) Prefill latency (s) Prefill (tok/s) Decode (tok/s) Output (toks) Decode latency (s) Model Size (GB) Layer Num Head Size Vocab Size
Gemma-2B [64] 1063 0.95 1121 40.9 59 1.44 4.7 18 256 256000
Phi3-4k [66] 1045 1.30 797.3 49.0 59 1.20 7.2 32 96 32064
MobileLLaMa-2.7B [62] 1047 0.92 1134 61.7 59 0.96 5.0 32 80 32000
MobileLLaMa-1.4B [62] 1047 0.23 4634 117.4 59 0.50 2.5 24 128 32000
Qwen4B [63] 1078 0.57 1882 44.5 59 1.33 7.5 40 128 151936
Qwen1.8B [63] 1078 0.23 4709 79.6 59 0.74 3.7 24 128 151936
Table 6: Inference performance of different LLMs after quantization and deployment on an OrinX chip. The Qwen series achieved the best performance.
Backbone Token Length Method Weighted Total MMMU [57] SEEDBench [58] RefCOCO [59] SUP-AD Drivelm-QA [60] Drivelm-Grounding [60] RealworldQA [61] PointVQA
5% 20% 15% 15% 7.5% 7.5% 15% 15%
ViT-L-336 [50] - - 0.421 0.368 0.657 0.502 0.553 0.688 0.742 0.541 0.583
ViT-L-336 - Fixed 1x2 + CA-256 0.413 0.388 0.62 0.327 0.543 0.699 0.733 0.536 0.563
ViT-L-336 - Fixed 2x2 + CA-256 0.405 0.376 0.618 0.306 0.518 0.692 0.731 0.523 0.594
ViT-L-448 - Fixed 1x2 + CA-256 0.383 0.391 0.542 0.184 0.528 0.689 0.718 0.470 0.562
ViT-L-336 - Fixed 1x2 + S2 0.368 0.383 0.511 0.198 0.469 0.689 0.718 0.469 0.314
ViT-L-336 - Fixed 1x2 + S2 0.368 0.381 0.537 0.207 0.544 0.698 0.729 0.460 0.554
ViT-L-336 - DM-4 + CA-256 0.409 0.377 0.610 0.277 0.544 0.698 0.726 0.525 0.554
ViT-L-336 - DM-6 + CA-256 0.381 0.381 0.500 0.162 0.523 0.700 0.729 0.466 0.584
ViT-L-336 - DM-4 + PT 0.426 0.376 0.653 0.557 0.585 0.700 0.743 0.540 0.598
ViT-L-336 - DM-4 + SM + LT 0.413 0.403 0.617 0.293 0.550 0.699 0.737 0.527 0.532
ViT-L-336 - DM-4 + SM + AAP + LT 0.406 0.384 0.610 0.273 0.518 0.68 0.726 0.520 0.533
ViT-L-336 - DM-4 + SM + CD + LT 0.409 0.388 0.630 0.480 0.515 0.700 0.734 0.524 0.577
SigLIP-L-384 [68] 576 - 0.432 0.389 0.631 0.615 0.624 0.707 0.749 0.556 0.560
SigLIP-L-768 576 - 0.438 0.377 0.642 0.580 0.638 0.712 0.764 0.561 0.556
SigLIP-L-1152 576 - 0.436 0.377 0.637 0.595 0.628 0.715 0.763 0.571 0.564
SigLIP-L-384-768 576 PE 0.434 0.367 0.632 0.581 0.631 0.716 0.762 0.556 0.554
SigLIP-L-512-960 480 PE 0.442 0.369 0.640 0.557 0.650 0.719 0.762 0.579 0.568
Table 7: Performance of different methods for scaling ViT’s original input resolution to higher resolution. “CA” stands for cross-attention, “DM” is the abbreviation for Dynamic Max, “PT” indicates the use of patch end token, “SM” stands for Spatial Merge, “LT” indicates the use of line end token, “AAP” is the abbreviation for Adaptive Average Pooling, and “CD” stands for Convolutional Downsampling. “PE” represents Position Embedding Interpolation. Among these methods, applying PE interpolation to SigLIP-L-384 and modifying it to take 768-resolution images as input achieves a good trade-off between inference speed and performance.
Base LLM

Due to the limited memory and bandwidth of the vehicle’s hardware, we cannot use overly large LLMs to maintain real-time inference. Therefore, we chose models with fewer than 4 billion parameters. As shown in Table 5 and 6, our experiments revealed that on the Orin architecture, the “wide and shallow” Qwen series (wider and fewer layers) models outperform “narrow and deep” models (narrower and more layers) in inference speed.

Visual Encoder

High-resolution images are essential for fine-grained visual understanding in autonomous driving. As shown in Table 7, compared to the basic ViT model used as a visual encoder, we explored several options, including different GridPatch strategies and PE (Position Embedding) interpolation. Ultimately, for real-time inference, we selected the simpler SigLIP-L-384 model with PE interpolation, achieving high-resolution input through original 384-resolution PE interpolation and fine-tuning parameters with additional convolution layers.

Projector Origin Compressed Output(ms) Prefill(ms) Avg. MMMU [57] SEEDV2 [58] BDD [bddx] Drivelm-QA [60] Drivelm-grounding [60] RealworldQA [61] RefCOCO [59] SUP-AD
MLP (Baseline) 576 576 666.60 707.93 56.27 37.90 64.00 53.90 67.60 71.70 55.30 44.40 53.90
LDPNetV2 [62] 576 576 661.70 707.42 50.53 37.67 56.71 53.36 68.90 73.20 48.30 21.29 54.70
Perceiver Resampler [69] 576 576 637.20 666.20 49.71 39.93 58.84 52.21 68.37 71.90 49.61 19.01 48.70
Pixel shuffle 576 256 610.70 655.92 52.02 40.42 60.20 55.19 68.72 71.52 51.44 28.23 48.40
LDPNetV2 576 256 605.32 652.79 54.73 38.93 62.19 54.82 68.57 72.78 50.59 34.32 59.30
Pixel shuffle 576 144 604.98 645.61 49.40 38.98 61.63 56.06 69.18 73.18 51.05 30.67 58.80
LDPNetV2 576 144 597.15 646.77 55.56 38.93 62.63 56.23 68.93 72.48 50.59 33.14 59.30
LDPNetV2 576 64 616.20 645.18 56.24 39.40 61.80 54.60 68.80 72.60 51.00 41.88 61.20
Table 8: Performance of different methods for visual token compression. Using LDPNetV2 to compress the original tokens to 75% of the original token count, achieves the best trade-off between performance and speed.
Visual Token Compression

To address the increased computational load from high-resolution images, we implemented LDPNetv2 [62] to reduce the number of image tokens by 75% without compromising performance, as shown in Table 8. Additionally, we enhanced performance by replacing the average pooling layer with a convolution layer in LDPNetv2.

Video Input

In autonomous driving scenarios, more temporal context is needed for accurately assessing object motion changes. We employ a short-term memory-bank [70] strategy, temporarily storing visual features from historical frames in a feature queue. Only the features from the current moment are extracted and fused with multiple historical frames before being projected into the LLM. Besides basic spatiotemporal pooling, we added SE [71] blocks to perform a weighted fusion of multiple temporal frames.

Base + q4f16_1 Eagle [72] + q4f16_1 Medusa [73] + q4f16_ft Eagle + q4f16_ft Eagle + q4f16_ft
+ Shrink Vocab Size(1024)
Quant Type q4f16_1 q4f16_1 q4f16_ft q4f16_ft q4f16_ft
Input Size (384, 960) (384, 960) (384, 960) (384, 960) (384, 960)
Prefill Tokens 604 604 613 604 613
Output Tokens 37 39 41 41 41
Prefill Speed (tok/s) 1818 1793 2160.64 2230 2175.35
Decode Speed (tok/s) 109 172 232.34 295.04 518.3
Prefill Latency (s) 0.332 0.340 0.284 0.276 0.274
Decode Latency (s) 0.328 0.216 0.176 0.130 0.071
Acceleration ratio 1 1.57 2.17 2.7 4.33
Table 9: Performance of different speculative sampling methods. “Shrink Vocab Size” means we reduce the vocabulary to the 1024 most frequently used words. “q4f16_1” is a 4-bit quantization method using a 16-bit floating-point representation for efficient model compression, while “q4f16_ft” includes subsequent fine-tuning to enhance performance post-quantization.
Speculative Sampling

Speculative Sampling is used to accelerate inference by preemptively generating likely outputs. This approach reduces the latency of generating predictions, achieving a significant speedup without substantial loss in accuracy. As shown in Table 9, we test two speculative sampling methods: Medusa [73] and Eagle [72] with our inference framework designed specifically for the OrinX chip. Eagle achieved a 2.7 ×\times× speedup in decode latency compared to Medusa’s 2.17 ×\times×, making real-time vehicle deployment feasible.

7 Conclusion

In summary, we introduce DriveVLM and DriveVLM-Dual. DriveVLM leverages VLMs, significantly progressing in interpreting complex driving environments. DriveVLM-Dual further enhances these capabilities by synergizing existing 3D perception and planning approaches, effectively addressing the spatial reasoning and computational challenges inherent in VLMs. Moreover, we define a scene understanding for planning task for autonomous driving, together with evaluation metrics and dataset construction protocol. DriveVLM and DriveVLM-Dual have surpassed the state-of-the-art methods on the public and our benchmarks, especially in handling intricate and dynamic scenarios. Finally, we have verified the effectiveness of DriveVLM-Dual through onboard deployment and testing on a production vehicle.

Acknowledgments

We thank Chenxu Hu at Tsinghua University for the help on paper writing. We thank Xu Bian, Hongkun Chen, Sui Cong, Chengze Guan, Mingyu Guo, Yue Jiang, Qi Jiang, Pengfei Ji, Wei Xiao, Dafeng Wei, Zijian Wang, Zhao Yang, Chenglong Zhao, Simeng Zhao, and Jian Zhou at Li Auto for their efforts in the experiments related to onboard deployment.

References

  • Barabas et al. [2017] I. Barabas, A. Todoruţ, N. Cordoş, and A. Molea. Current challenges in autonomous driving. In IOP conference series: materials science and engineering, volume 252, page 012096. IOP Publishing, 2017.
  • Qi et al. [2017] C. R. Qi, H. Su, K. Mo, and L. J. Guibas. Pointnet: Deep learning on point sets for 3d classification and segmentation. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 652–660, 2017.
  • Lang et al. [2019] A. H. Lang, S. Vora, H. Caesar, L. Zhou, J. Yang, and O. Beijbom. Pointpillars: Fast encoders for object detection from point clouds. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 12697–12705, 2019.
  • Wang et al. [2022] Y. Wang, V. C. Guizilini, T. Zhang, Y. Wang, H. Zhao, and J. Solomon. Detr3d: 3d object detection from multi-view images via 3d-to-2d queries. In Conference on Robot Learning, pages 180–191. PMLR, 2022.
  • Li et al. [2022] Z. Li, W. Wang, H. Li, E. Xie, C. Sima, T. Lu, Y. Qiao, and J. Dai. Bevformer: Learning bird’s-eye-view representation from multi-camera images via spatiotemporal transformers. In European conference on computer vision, pages 1–18. Springer, 2022.
  • Gao et al. [2020] J. Gao, C. Sun, H. Zhao, Y. Shen, D. Anguelov, C. Li, and C. Schmid. Vectornet: Encoding hd maps and agent dynamics from vectorized representation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June 2020.
  • Zhao et al. [2021] H. Zhao, J. Gao, T. Lan, C. Sun, B. Sapp, B. Varadarajan, Y. Shen, Y. Shen, Y. Chai, C. Schmid, C. Li, and D. Anguelov. Tnt: Target-driven trajectory prediction. In J. Kober, F. Ramos, and C. Tomlin, editors, Proceedings of the 2020 Conference on Robot Learning, volume 155 of Proceedings of Machine Learning Research, pages 895–904. PMLR, 16–18 Nov 2021.
  • Liu et al. [2021] Y. Liu, J. Zhang, L. Fang, Q. Jiang, and B. Zhou. Multimodal motion prediction with stacked transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 7577–7586, 2021.
  • Gu et al. [2021] J. Gu, C. Sun, and H. Zhao. Densetnt: End-to-end trajectory prediction from dense goal sets. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 15303–15312, 2021.
  • Nayakanti et al. [2023] N. Nayakanti, R. Al-Rfou, A. Zhou, K. Goel, K. S. Refaat, and B. Sapp. Wayformer: Motion forecasting via simple & efficient attention networks. In 2023 IEEE International Conference on Robotics and Automation (ICRA), pages 2980–2987. IEEE, 2023.
  • Pomerleau [1988] D. A. Pomerleau. Alvinn: An autonomous land vehicle in a neural network. Advances in neural information processing systems, 1, 1988.
  • Bansal et al. [2018] M. Bansal, A. Krizhevsky, and A. Ogale. Chauffeurnet: Learning to drive by imitating the best and synthesizing the worst. arXiv preprint arXiv:1812.03079, 2018.
  • Li et al. [2023] Z. Li, F. Nie, Q. Sun, F. Da, and H. Zhao. Uncertainty-aware decision transformer for stochastic driving environments. arXiv preprint arXiv:2309.16397, 2023.
  • Zeng et al. [2023] Y. Zeng, H. Zhang, J. Zheng, J. Xia, G. Wei, Y. Wei, Y. Zhang, and T. Kong. What matters in training a gpt4-style language model with multimodal inputs? arXiv preprint arXiv:2307.02469, 2023.
  • Xu et al. [2023] Z. Xu, Y. Zhang, E. Xie, Z. Zhao, Y. Guo, K. K. Wong, Z. Li, and H. Zhao. Drivegpt4: Interpretable end-to-end autonomous driving via large language model. arXiv preprint arXiv:2310.01412, 2023.
  • Zhu et al. [2023] D. Zhu, J. Chen, X. Shen, X. Li, and M. Elhoseiny. Minigpt-4: Enhancing vision-language understanding with advanced large language models. arXiv preprint arXiv:2304.10592, 2023.
  • Liu et al. [2023] H. Liu, C. Li, Q. Wu, and Y. J. Lee. Visual instruction tuning, 2023.
  • Bai et al. [2023] J. Bai, S. Bai, S. Yang, S. Wang, S. Tan, P. Wang, J. Lin, C. Zhou, and J. Zhou. Qwen-vl: A versatile vision-language model for understanding, localization, text reading, and beyond. arXiv preprint arXiv:2308.12966, 2023.
  • Dai et al. [2023] W. Dai, J. Li, D. Li, A. M. H. Tiong, J. Zhao, W. Wang, B. Li, P. Fung, and S. Hoi. Instructblip: Towards general-purpose vision-language models with instruction tuning, 2023.
  • Zhang et al. [2023] P. Zhang, X. Dong, B. Wang, Y. Cao, C. Xu, L. Ouyang, Z. Zhao, S. Ding, S. Zhang, H. Duan, W. Zhang, H. Yan, X. Zhang, W. Li, J. Li, K. Chen, C. He, X. Zhang, Y. Qiao, D. Lin, and J. Wang. Internlm-xcomposer: A vision-language large model for advanced text-image comprehension and composition, 2023.
  • Wang et al. [2023] W. Wang, Q. Lv, W. Yu, W. Hong, J. Qi, Y. Wang, J. Ji, Z. Yang, L. Zhao, X. Song, et al. Cogvlm: Visual expert for pretrained language models. arXiv preprint arXiv:2311.03079, 2023.
  • Driess et al. [2023] D. Driess, F. Xia, M. S. M. Sajjadi, C. Lynch, A. Chowdhery, B. Ichter, A. Wahid, J. Tompson, Q. Vuong, T. Yu, W. Huang, Y. Chebotar, P. Sermanet, D. Duckworth, S. Levine, V. Vanhoucke, K. Hausman, M. Toussaint, K. Greff, A. Zeng, I. Mordatch, and P. Florence. Palm-e: An embodied multimodal language model, 2023.
  • Brohan et al. [2022] A. Brohan, N. Brown, J. Carbajal, Y. Chebotar, J. Dabis, C. Finn, K. Gopalakrishnan, K. Hausman, A. Herzog, J. Hsu, et al. Rt-1: Robotics transformer for real-world control at scale. arXiv preprint arXiv:2212.06817, 2022.
  • Brohan et al. [2023] A. Brohan, N. Brown, J. Carbajal, Y. Chebotar, X. Chen, K. Choromanski, T. Ding, D. Driess, A. Dubey, C. Finn, P. Florence, C. Fu, M. G. Arenas, K. Gopalakrishnan, K. Han, K. Hausman, A. Herzog, J. Hsu, B. Ichter, A. Irpan, N. Joshi, R. Julian, D. Kalashnikov, Y. Kuang, I. Leal, L. Lee, T.-W. E. Lee, S. Levine, Y. Lu, H. Michalewski, I. Mordatch, K. Pertsch, K. Rao, K. Reymann, M. Ryoo, G. Salazar, P. Sanketi, P. Sermanet, J. Singh, A. Singh, R. Soricut, H. Tran, V. Vanhoucke, Q. Vuong, A. Wahid, S. Welker, P. Wohlhart, J. Wu, F. Xia, T. Xiao, P. Xu, S. Xu, T. Yu, and B. Zitkovich. Rt-2: Vision-language-action models transfer web knowledge to robotic control. In arXiv preprint arXiv:2307.15818, 2023.
  • Huang et al. [2023] W. Huang, C. Wang, R. Zhang, Y. Li, J. Wu, and L. Fei-Fei. Voxposer: Composable 3d value maps for robotic manipulation with language models. arXiv preprint arXiv:2307.05973, 2023.
  • Padalkar et al. [2023] A. Padalkar, A. Pooley, A. Jain, A. Bewley, A. Herzog, A. Irpan, A. Khazatsky, A. Rai, A. Singh, A. Brohan, et al. Open x-embodiment: Robotic learning datasets and rt-x models. arXiv preprint arXiv:2310.08864, 2023.
  • Chekroun et al. [2023] R. Chekroun, M. Toromanoff, S. Hornauer, and F. Moutarde. Gri: General reinforced imitation and its application to vision-based autonomous driving. Robotics, 12(5):127, 2023.
  • Chen et al. [2021] D. Chen, V. Koltun, and P. Krähenbühl. Learning to drive from a world on rails. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 15590–15599, 2021.
  • Toromanoff et al. [2020] M. Toromanoff, E. Wirbel, and F. Moutarde. End-to-end model-free reinforcement learning for urban driving using implicit affordances. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 7153–7162, 2020.
  • Zeng et al. [2019] W. Zeng, W. Luo, S. Suo, A. Sadat, B. Yang, S. Casas, and R. Urtasun. End-to-end interpretable neural motion planner. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 8660–8669, 2019.
  • Wei et al. [2021] B. Wei, M. Ren, W. Zeng, M. Liang, B. Yang, and R. Urtasun. Perceive, attend, and drive: Learning spatial attention for safe self-driving. In 2021 IEEE International Conference on Robotics and Automation (ICRA), pages 4875–4881. IEEE, 2021.
  • Hu et al. [2021] P. Hu, A. Huang, J. Dolan, D. Held, and D. Ramanan. Safe local motion planning with self-supervised freespace forecasting. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 12732–12741, 2021.
  • Casas et al. [2021] S. Casas, A. Sadat, and R. Urtasun. Mp3: A unified model to map, perceive, predict and plan. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 14403–14412, 2021.
  • Hu et al. [2023] Y. Hu, J. Yang, L. Chen, K. Li, C. Sima, X. Zhu, S. Chai, S. Du, T. Lin, W. Wang, et al. Planning-oriented autonomous driving. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 17853–17862, 2023.
  • Renz et al. [2022] K. Renz, K. Chitta, O.-B. Mercea, A. Koepke, Z. Akata, and A. Geiger. Plant: Explainable planning transformers via object-level representations. arXiv preprint arXiv:2210.14222, 2022.
  • Chen et al. [2023] L. Chen, P. Wu, K. Chitta, B. Jaeger, A. Geiger, and H. Li. End-to-end autonomous driving: Challenges and frontiers. arXiv preprint arXiv:2306.16927, 2023.
  • Hu et al. [2023] A. Hu, L. Russell, H. Yeo, Z. Murez, G. Fedoseev, A. Kendall, J. Shotton, and G. Corrado. Gaia-1: A generative world model for autonomous driving. arXiv preprint arXiv:2309.17080, 2023.
  • Yang et al. [2023] Z. Yang, X. Jia, H. Li, and J. Yan. A survey of large language models for autonomous driving. arXiv preprint arXiv:2311.01043, 2023.
  • Wu et al. [2023] D. Wu, W. Han, T. Wang, X. Dong, X. Zhang, and J. Shen. Referring multi-object tracking. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 14633–14642, 2023.
  • Geiger et al. [2013] A. Geiger, P. Lenz, C. Stiller, and R. Urtasun. Vision meets robotics: The kitti dataset. The International Journal of Robotics Research, 32(11):1231–1237, 2013.
  • Deruyttere et al. [2019] T. Deruyttere, S. Vandenhende, D. Grujicic, L. Van Gool, and M.-F. Moens. Talk2car: Taking control of your self-driving car. arXiv preprint arXiv:1909.10838, 2019.
  • Wu et al. [2023] D. Wu, W. Han, T. Wang, Y. Liu, X. Zhang, and J. Shen. Language prompt for autonomous driving. arXiv preprint arXiv:2309.04379, 2023.
  • Qian et al. [2023] T. Qian, J. Chen, L. Zhuo, Y. Jiao, and Y.-G. Jiang. Nuscenes-qa: A multi-modal visual question answering benchmark for autonomous driving scenario. arXiv preprint arXiv:2305.14836, 2023.
  • Caesar et al. [2020] H. Caesar, V. Bankiti, A. H. Lang, S. Vora, V. E. Liong, Q. Xu, A. Krishnan, Y. Pan, G. Baldan, and O. Beijbom. nuscenes: A multimodal dataset for autonomous driving. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 11621–11631, 2020.
  • Kim et al. [2018] J. Kim, A. Rohrbach, T. Darrell, J. Canny, and Z. Akata. Textual explanations for self-driving vehicles. In Proceedings of the European conference on computer vision (ECCV), pages 563–578, 2018.
  • Xu et al. [2020] Y. Xu, X. Yang, L. Gong, H.-C. Lin, T.-Y. Wu, Y. Li, and N. Vasconcelos. Explainable object-induced action decision for autonomous vehicles. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 9523–9532, 2020.
  • Sachdeva et al. [2023] E. Sachdeva, N. Agarwal, S. Chundi, S. Roelofs, J. Li, B. Dariush, C. Choi, and M. Kochenderfer. Rank2tell: A multimodal driving dataset for joint importance ranking and reasoning. arXiv preprint arXiv:2309.06597, 2023.
  • Malla et al. [2023] S. Malla, C. Choi, I. Dwivedi, J. H. Choi, and J. Li. Drama: Joint risk localization and captioning in driving. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, pages 1043–1052, 2023.
  • Wei et al. [2022] J. Wei, X. Wang, D. Schuurmans, M. Bosma, F. Xia, E. Chi, Q. V. Le, D. Zhou, et al. Chain-of-thought prompting elicits reasoning in large language models. Advances in Neural Information Processing Systems, 35:24824–24837, 2022.
  • Dosovitskiy et al. [2020] A. Dosovitskiy, L. Beyer, A. Kolesnikov, D. Weissenborn, X. Zhai, T. Unterthiner, M. Dehghani, M. Minderer, G. Heigold, S. Gelly, et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv preprint arXiv:2010.11929, 2020.
  • Jiang et al. [2023] B. Jiang, S. Chen, Q. Xu, B. Liao, J. Chen, H. Zhou, Q. Zhang, W. Liu, C. Huang, and X. Wang. Vad: Vectorized scene representation for efficient autonomous driving. arXiv preprint arXiv:2303.12077, 2023.
  • OpenAI [2023] OpenAI. Gpt-4 technical report, 2023.
  • Khurana et al. [2022] T. Khurana, P. Hu, A. Dave, J. Ziglar, D. Held, and D. Ramanan. Differentiable raycasting for self-supervised occupancy forecasting. In European Conference on Computer Vision, pages 353–369. Springer, 2022.
  • Hu et al. [2022] S. Hu, L. Chen, P. Wu, H. Li, J. Yan, and D. Tao. St-p3: End-to-end vision-based autonomous driving via spatial-temporal feature learning. In European Conference on Computer Vision, pages 533–549. Springer, 2022.
  • Xu et al. [2021] L. Xu, H. Huang, and J. Liu. Sutd-trafficqa: A question answering benchmark and an efficient network for video reasoning over traffic events. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 9878–9888, 2021.
  • Zhai et al. [2023] J.-T. Zhai, Z. Feng, J. Du, Y. Mao, J.-J. Liu, Z. Tan, Y. Zhang, X. Ye, and J. Wang. Rethinking the open-loop evaluation of end-to-end autonomous driving in nuscenes. arXiv preprint arXiv:2305.10430, 2023.
  • Yue et al. [2024] X. Yue, Y. Ni, K. Zhang, T. Zheng, R. Liu, G. Zhang, S. Stevens, D. Jiang, W. Ren, Y. Sun, et al. Mmmu: A massive multi-discipline multimodal understanding and reasoning benchmark for expert agi. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 9556–9567, 2024.
  • Li et al. [2023] B. Li, R. Wang, G. Wang, Y. Ge, Y. Ge, and Y. Shan. Seed-bench: Benchmarking multimodal llms with generative comprehension. arXiv preprint arXiv:2307.16125, 2023.
  • Yu et al. [2016] L. Yu, P. Poirson, S. Yang, A. C. Berg, and T. L. Berg. Modeling context in referring expressions. In Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, October 11-14, 2016, Proceedings, Part II 14, pages 69–85. Springer, 2016.
  • Sima et al. [2023] C. Sima, K. Renz, K. Chitta, L. Chen, H. Zhang, C. Xie, P. Luo, A. Geiger, and H. Li. Drivelm: Driving with graph visual question answering. arXiv preprint arXiv:2312.14150, 2023.
  • X.ai [2024] X.ai. Grok-1.5 vision preview. https://x.ai/blog/grok-1.5v, 2024.
  • Chu et al. [2023] X. Chu, L. Qiao, X. Lin, S. Xu, Y. Yang, Y. Hu, F. Wei, X. Zhang, B. Zhang, X. Wei, et al. Mobilevlm: A fast, reproducible and strong vision language assistant for mobile devices. arXiv preprint arXiv:2312.16886, 2023.
  • Team [2024] Q. Team. Introducing qwen1.5, February 2024. URL https://qwenlm.github.io/blog/qwen1.5/.
  • Team et al. [2024] G. Team, T. Mesnard, C. Hardin, R. Dadashi, S. Bhupatiraju, S. Pathak, L. Sifre, M. Rivière, M. S. Kale, J. Love, et al. Gemma: Open models based on gemini research and technology. arXiv preprint arXiv:2403.08295, 2024.
  • Hu et al. [2024] S. Hu, Y. Tu, X. Han, C. He, G. Cui, X. Long, Z. Zheng, Y. Fang, Y. Huang, W. Zhao, et al. Minicpm: Unveiling the potential of small language models with scalable training strategies. arXiv preprint arXiv:2404.06395, 2024.
  • Abdin et al. [2024] M. Abdin, S. A. Jacobs, A. A. Awan, J. Aneja, A. Awadallah, H. Awadalla, N. Bach, A. Bahree, A. Bakhtiari, H. Behl, et al. Phi-3 technical report: A highly capable language model locally on your phone. arXiv preprint arXiv:2404.14219, 2024.
  • Liu et al. [2024] H. Liu, C. Li, Y. Li, and Y. J. Lee. Improved baselines with visual instruction tuning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 26296–26306, 2024.
  • Zhai et al. [2023] X. Zhai, B. Mustafa, A. Kolesnikov, and L. Beyer. Sigmoid loss for language image pre-training. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 11975–11986, 2023.
  • Alayrac et al. [2022] J.-B. Alayrac, J. Donahue, P. Luc, A. Miech, I. Barr, Y. Hasson, K. Lenc, A. Mensch, K. Millican, M. Reynolds, et al. Flamingo: a visual language model for few-shot learning. Advances in neural information processing systems, 35:23716–23736, 2022.
  • He et al. [2024] B. He, H. Li, Y. K. Jang, M. Jia, X. Cao, A. Shah, A. Shrivastava, and S.-N. Lim. Ma-lmm: Memory-augmented large multimodal model for long-term video understanding. arXiv preprint arXiv:2404.05726, 2024.
  • Hu et al. [2018] J. Hu, L. Shen, and G. Sun. Squeeze-and-excitation networks. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 7132–7141, 2018.
  • Li et al. [2024] Y. Li, F. Wei, C. Zhang, and H. Zhang. Eagle: Speculative sampling requires rethinking feature uncertainty. arXiv preprint arXiv:2401.15077, 2024.
  • Cai et al. [2023] T. Cai, Y. Li, Z. Geng, H. Peng, and T. Dao. Medusa: Simple framework for accelerating llm generation with multiple decoding heads, 2023.

Appendix

Appendix A SUP-AD Dataset

A.1 Meta-actions

Meta-action statistics.

We use the meta-action sequence to formally represent the driving strategy. Meta actions are classified into 17 categories. We show the distribution of each meta-action being the first/second/third place in the meta-action sequence, as shown in Figure 5. It indicates that the meta-actions are quite diverse in the SUP-AD dataset. We also show the distribution of the length of meta-actions per scene in Figure 6. Most scenes contain two or three meta-actions, and a few scenes with complex driving strategies contain four or more meta-actions.

Annotation of meta-actions.

The meta-action sequence for each driving scene is manually annotated based on the actual driving strategy in the future frames. These meta-actions are designed to encompass a complete driving strategy and are structured to be consistent with the future trajectory of the ego vehicle. They can be divided into three primary classes:

  1. 1.

    Speed-control actions. Discerned from acceleration and braking signals within the ego state data, these actions include These actions can be discerned from acceleration and braking signals within the ego state data. They include speed up, slow down, slow down rapidly, go straight slowly, go straight at a constant speed, stop, wait, and reverse.

  2. 2.

    Turning actions. Deduced from steering wheel signals, these actions consist of turn left, turn right, and turn around.

  3. 3.

    Lane-control actions. Encompassing lane selection decisions, these actions are derived from a combination of steering wheel signals and either map or perception data. They involve change lane to the left, change lane to the right, shift slightly to the left, and shift slightly to the right.

Refer to caption
Figure 5: Distribution of each meta action being the first, second, and third place of the meta action sequence, respectively.
Refer to caption
Figure 6: Distribution of the length of meta actions per scene.
Refer to caption
Figure 7: Diverse driving scenarios in the SUP-AD dataset.

A.2 Scenario Categories

The SUP-AD dataset is comprised of 1,000 video clips of driving scenarios. As illustrated in Figure 7, it encompasses a wide range of driving scenarios, spanning over 40 categories. Below are explanations for some of the scenarios:

AEB Data: Automatic Emergency Braking (AEB) data.

Road Construction: A temporary work zone with caution signs, barriers, and construction equipment ahead.

Close-range Cut-ins: A sudden intrusion into the lane of the ego vehicle by another vehicle.

Roundabout: A type of traffic intersection where vehicles travel in a continuous loop.

Animals Crossing Road: Animals crossing the road in front of the ego vehicle.

Braking: Brake is pressed by human driver of the ego vehicle.

Traffic Police Officers: Traffic police officers managing and guiding traffic.

Blocking Traffic Lights: A massive vehicle obscuring the visibility of the traffic signal.

Cutting into Other Vehicle: Intruding into the lane of another vehicle ahead.

Ramp: A curved roadway that connects the main road to the branch road in highway.

Debris on the Road: Road with different kinds of debris.

Narrow Roads: Narrow roads that require cautious navigation.

Pedestrians Popping Out: Pedestrians popping out in front of the ego vehicle, requiring slowing down or braking.

People on Bus Posters: Buses with posters, which may interfere the perception system.

Merging into High Speed: Driving from a low-speed road into a high-speed road, requiring speeding up.

Barrier Gate: Barrier gate that can be raised obstructing the road.

Fallen Trees: Fallen trees on the road, requiring cautious navigation to avoid potential hazards.

Complex Environments: Complex driving environments that requiring cautious navigation.

Mixed Traffic: A congested scenario where cars, pedestrians, and bicycles appear on the same or adjacent roadway.

Crossing Rivers: Crossing rivers by driving on the bridge.

Screen: Roads with screens on one side, which may interfere the perception system.

Herds of Cattle and Sheep: A rural road with herds of cattle and sheep, requiring careful driving to avoid causing distress to these animals.

Vulnerable Road Users: Road users which are more susceptible to injuries while using roads, such as pedestrians, cyclists, and motorcyclists.

Road with Gallet: A dusty road with gallet scattered across the surface.

The remaining scenario categories are: Motorcycles and Trikes, Intersection, People carrying Umbrella, Vehicles Carrying Cars, Vehicles Carrying Branches, Vehicles with Pipes, Strollers, Children, Tunnel, Down Ramp, Sidewalk Stalls, Rainy Day, Crossing Train Tracks, Unprotected U-turns, Snowfall, Large Vehicles Invading, Falling Leaves, Fireworks, Water Sprinklers, Potholes, Overturned Motorcycles, Self-ignition and Fire, Kites, Agricultural Machinery.

A.3 Annotation Examples

We provide more examples of annotation contents in Figure 89101112, and  13. The scenario categories of these examples are overturned bicycles and motorcycles, herds of cattle and sheep, collapsed trees, crossing rivers, barrier gate, and snowfall respectively.

Refer to caption
Figure 8: An example of overturned bicycles and motorcycles in the SUP-AD dataset. A bicycle has fallen in front of the ego vehicle, requiring the ego vehicle to change lanes.
Refer to caption
Figure 9: An example of herds of cattle and sheep in the SUP-AD dataset. A group of cattle move slowly in front of the ego vehicle, requiring the ego vehicle to proceed slowly and maintain a safe distance from the cattle.
Refer to caption
Figure 10: An example of collapsed trees in the SUP-AD dataset. A tree suddenly falls towards the ego vehicle, requiring the ego vehicle to decelerate immediately.
Refer to caption
Figure 11: An example of crossing rivers in the SUP-AD dataset. The ego vehicle is going across a bridge of which width allows only a single vehicle to pass, requiring the ego vehicle to drive without stopping.
Refer to caption
Figure 12: An example of barrier gate in the SUP-AD dataset. A horizontal barrier blocks the entrance of a park, requiring the ego vehicle to stop and wait for permission to continue.
Refer to caption
Figure 13: An example of snowfall in the SUP-AD dataset. Most of the road is covered by snow, requiring the ego vehicle to move forward cautiously by following the snow-free tire tracks.

Appendix B Evaluation Method

The ability of an autonomous driving system to accurately interpret driving scenes and make logical, suitable decisions is of paramount importance. As presented in this paper, the evaluation of VLMs in autonomous driving concentrates on two primary components: the evaluation of scene description/analysis and the evaluation of meta-actions.

B.1 Scene Description/Analysis Evaluation

In terms of scene description/analysis evaluation, the process of interpreting and articulating driving scenes is subject to inherent subjectivity, as there are numerous valid ways to express similar descriptions textually, which makes it difficult to effectively evaluate the scene description using a fixed metric. To overcome this challenge, we utilize GPT-4 to evaluate the similarity between the scene descriptions generated by the model and the manually annotated ground truth. Initially, we prompt GPT-4 to extract individual pieces of information from each scene description. Subsequently, we score and aggregate the results based on the matching status of each extracted piece of information.

The ground truth labels for scene descriptions encompass both environment descriptions and event summaries. Environmental condition description includes weather conditions, time conditions, road environment, and lane conditions. Event summaries are the characteristics and influence of critical objects. We employ GPT-4 to extract unique key information from both environment descriptions and event summaries. The extracted information is then compared and quantified. Each matched pair is assigned a score, which is estimated based on the extent of the matching, whether complete, partial, or absent. Instances of hallucinated information incur a penalty, detracting from the overall score. The aggregate of these scores constitutes the scene description score.

Score=1.0×nmatched+0.5×npartialngt0.25×nhallucinationngtScore1.0subscript𝑛matched0.5subscript𝑛partialsubscript𝑛gt0.25subscript𝑛hallucinationsubscript𝑛gt\displaystyle\begin{split}\text{Score}=\frac{1.0\times n_{\text{matched}}+0.5% \times n_{\text{partial}}}{n_{\text{gt}}}\\ -\frac{0.25\times n_{\text{hallucination}}}{n_{\text{gt}}}\end{split}start_ROW start_CELL Score = divide start_ARG 1.0 × italic_n start_POSTSUBSCRIPT matched end_POSTSUBSCRIPT + 0.5 × italic_n start_POSTSUBSCRIPT partial end_POSTSUBSCRIPT end_ARG start_ARG italic_n start_POSTSUBSCRIPT gt end_POSTSUBSCRIPT end_ARG end_CELL end_ROW start_ROW start_CELL - divide start_ARG 0.25 × italic_n start_POSTSUBSCRIPT hallucination end_POSTSUBSCRIPT end_ARG start_ARG italic_n start_POSTSUBSCRIPT gt end_POSTSUBSCRIPT end_ARG end_CELL end_ROW (2)

The prompt for GPT-4 in evaluating scene descriptions is carefully designed, as shown in Table 10. Initially, a role prompt is employed to establish as an intelligent and logical evaluator, possessing a comprehensive understanding of appropriate driving styles. This is followed by specifying the input format, which informs GPT-4 that its task involves comparing an output description with a ground truth description. This comparison is based on the extraction and analysis of key information from both descriptions. Lastly, the prompt outlines the criteria for scoring, as well as the format for the evaluation output, ensuring a structured and systematic approach to the evaluation process.

System Prompt:
You are a smart and logical evaluator with extensive driving experience, and you will try your best to give reasonable and logical evaluation result.
Input Prompt:
Given two driving scenario descriptions, one is the reference description, the other is the output description from a model. Please extract key information from the reference description and use it as comparison standards to compare the proximity of the output descriptions. Additionally, evaluate the completeness of each piece of information. Key information extracted can be classified into environmental conditions and critical events. Environmental condition information includes weather, time, road environment, and lane conditions. A critical event describes the event related to ego vehicleś driving decisions, which is usually given in the format of critical object + verb/adj + influence on ego vehicle. An example of a critical event is like this: A construction zone is blocking the ego lane and the right lane in front, so the car needs to go around to avoid it. Now I will give you the reference description and the output description.
The reference description:
Weather: Sunny. —— Time: Day. —— Road Environment: Urban. —— Lane Options: Left Lane, Own Lane, Right Lane. —— Ego Lane Position: Middle Lane.
Critical Events: There are police officers ahead inspecting passing vehicles. The police officer in front of your lane is signaling you to stop for inspection. There are also vehicles queuing in the left lane. The right lane is a merging lane and is currently separated, so you cannot change lanes to the right.
The output description:
Weather: Sunny. —— Time: Day.
There is a vehicle driving in the distance in front of you. There are traffic police on the left and right, and the traffic police signal to stop.
Compare the two descriptions using extracted key information and output a similarities score on the scale from 0 to 1.0 and an explanation to describe the logic of calculating the similarity scale. Each matched key information scores 1.0, and each partially matched key information scores 0.5. Hallucinations will be penalized as well. If any key information exists in the output but not in the reference, a score of -0.25 will be assigned. The final score is the total score divided by the number of key information extracted in the reference description. For example, if the total score is 1.5, and the total number of key information in reference description is 3, then the final score would be 1.5/3 = 0.5. Please output the detailed extraction and evaluation process.
Response:
Reference description key information:
1. Weather: Sunny (environmental conditions)
2. Time: Day (environmental conditions)
3. Road Environment: Urban (environmental conditions)
4. Lane Options: Left Lane, Own Lane, Right Lane (environmental conditions)
5. Ego Lane Position: Middle Lane (environmental conditions)
6. Police officers ahead inspecting vehicles (critical event)
7. The police officer in front of ego vehicle signalling to stop (critical event)
8. Vehicles queuing in the left lane (critical event)
9. Cannot change to the right lane as itś an merging lane and currently separated (critical event)
Output description key information:
1. Weather: Sunny.
2. Time: Day.
3. Vehicle driving in the distance in front of driver
4. Traffic police on the left and right
5. Traffic police signal to stop
By comparing the output to the reference:
- Key information 1, 2, 6, and 7 are matched, each for 1.0 points(weather and time, police officers and their signal to stop)
- Key information 3 in the output is not mentioned in reference, deduction of -0.25 points.
No scores for:
- Key information 3, 4, 5, 8, and 9.
Total score = (1.0 + 1.0 + 1.0 + 1.0 - 0.25) / 9 = 0.417.

Table 10: Evaluation process of scene description by GPT-4. Our proposed evaluation method is not only capable of extracting and evaluating information from structured scene descriptions but is also applicable to unstructured descriptions.

B.2 Meta-action Evaluation

The evaluation process for the meta-action sequence must consider both the quantity and the sequential arrangement of the matched meta-actions. We employ dynamic programming to compare the model’s output and the annotated ground truth. Our dynamic programming approach is similar to the method utilized in identifying the longest common subsequence, albeit with two supplementary considerations.

The first consideration acknowledges the unequal weighting of different meta-actions. For instance, certain meta actions such as “Slow Down”, “Wait”, and “Go Straight Slowly” exhibit a greater emphasis on attitude rather than action. The presence or absence of these actions from a meta-action sequence does not alter the basic semantic essence of driving decisions but rather modifies the driving strategy to be either more assertive or more cautious. For example, a meta action sequence of “Slow Down -> Stop -> Wait” conveys a similar driving decision as a sequence with only the meta action “Stop”. Consequently, these sequences should not incur a penalty comparable to other meta actions such as “Turn Left” or “Change Lane to the Right”. Therefore, these are designated as “conservative actions”, and a reduced penalty is applied when they do not match during sequence evaluation.

The second consideration addresses the potential semantic equality among different meta-action sequences. For example, the sequences “Change Lane to the Left -> Speed Up -> Go Straight At a Constant Speed -> Change Lane to the Right” and “Change Lane to the Left -> Speed Up Rapidly -> Go Straight At a Constant Speed -> Change Lane to the Right” might both represent valid approaches to overtaking a slow-speed vehicle ahead. Recognizing that different meta-action sequences might convey similar meanings, we initially use GPT-4 to generate variant sequences that have comparable semantic meanings, in addition to the unique ground truth meta-action sequence, as shown in Table 11. In the subsequent sequence-matching phase of the evaluation, all these variations, together with the manually annotated ground truth, are taken into consideration. The highest-scoring matching is then adopted as the definitive score for the final decision evaluation.

Prompt:
Given a predefined set of action tokens and a sequence of reference driving actions, find all alternative sequences of actions that have similar meanings but choose slightly different action tokens from the predefined action tokens.
Predefined action tokens: [Speed up, Slow down, Speed up rapidly, Slow down rapidly, Go straight slowly, Go straight at a constant speed, Turn left, Turn right, Change lane to the left, Change lane to the right, Shift slightly to the left, Shift slightly to the right, Stop, Wait, Turn around, Reverse].
For example, if the reference driving action sequence is: Slow down -> Shift slightly to the right -> Go straight at a constant speed, alternative action sequences with similar meanings could be:
1. Slow down -> Change lane to the right -> Go straight at a constant speed;
2. Slow down rapidly -> Shift slightly to the right -> Go straight at a constant speed;
3. Slow down -> Change lane to the right -> Go straight slowly;
4. Slow down -> Shift slightly to the right -> Go straight slowly.
Now the reference action sequence is: Slow down -> Shift slightly to the right -> Go straight at a constant speed. What are the alternative action sequences?
Response:
Given the reference sequence: ”Slow down -> Shift slightly to the right -> Go straight at a constant speed.” Alternative action sequences with similar meanings based on the provided action tokens are: 1. Slow down -> Change lane to the right -> Go straight at a constant speed.
2. Slow down rapidly -> Shift slightly to the right -> Go straight at a constant speed.
3. Slow down -> Change lane to the right -> Go straight slowly.
4. Slow down -> Shift slightly to the right -> Go straight slowly. You’ve provided this example in your question.
Based on the tokens and the context, these are indeed appropriate alternative sequences.
Table 11: Example of generating alternative action sequences by GPT-4.

The state of dynamic programming is saved in a 2D matrix, wherein each row corresponds to a meta action in the ground truth action sequence, and each column corresponds to a meta action in the model output action sequence, noted as Sr,csuperscript𝑆𝑟𝑐S^{\,r,\,c}italic_S start_POSTSUPERSCRIPT italic_r , italic_c end_POSTSUPERSCRIPT. The dynamic programming initiates recursive calculations beginning from the first meta action of both sequences. Each element of the 2D matrix encompasses the optimal total score at the current matching position, as well as the preceding matching condition that yielded the optimal matching. In our dynamic programming algorithm, three transition equations govern distinct cases: Smissingsubscript𝑆missingS_{\text{missing}}italic_S start_POSTSUBSCRIPT missing end_POSTSUBSCRIPT for missing matching, Sredundantsubscript𝑆redundantS_{\text{redundant}}italic_S start_POSTSUBSCRIPT redundant end_POSTSUBSCRIPT for redundant matching, and Smatchingsubscript𝑆matchingS_{\text{matching}}italic_S start_POSTSUBSCRIPT matching end_POSTSUBSCRIPT for successful matching. Successful matching occurs when the meta action is identical at the rthsuperscript𝑟𝑡r^{th}italic_r start_POSTSUPERSCRIPT italic_t italic_h end_POSTSUPERSCRIPT position in the reference sequence and the cthsuperscript𝑐𝑡c^{th}italic_c start_POSTSUPERSCRIPT italic_t italic_h end_POSTSUPERSCRIPT position in the model-generated sequence. In the case of missing matching, the meta action at the rthsuperscript𝑟𝑡r^{th}italic_r start_POSTSUPERSCRIPT italic_t italic_h end_POSTSUPERSCRIPT position in the reference sequence is unmatched, prompting a comparison with the r1th𝑟superscript1𝑡{r-1}^{th}italic_r - 1 start_POSTSUPERSCRIPT italic_t italic_h end_POSTSUPERSCRIPT position in the reference sequence and the cthsuperscript𝑐𝑡c^{th}italic_c start_POSTSUPERSCRIPT italic_t italic_h end_POSTSUPERSCRIPT position in the model-generated sequence. Conversely, redundant matching implies that the meta action at the cthsuperscript𝑐𝑡c^{th}italic_c start_POSTSUPERSCRIPT italic_t italic_h end_POSTSUPERSCRIPT position in the model-generated sequence is unmatched, leading to further examination of the rthsuperscript𝑟𝑡r^{th}italic_r start_POSTSUPERSCRIPT italic_t italic_h end_POSTSUPERSCRIPT position in the reference and the c1th𝑐superscript1𝑡{c-1}^{th}italic_c - 1 start_POSTSUPERSCRIPT italic_t italic_h end_POSTSUPERSCRIPT position in the model-generated sequence. The transformation equations for these cases are as follows:

Smissingr,c=Sr1,cpmissing,Sredundantr,c=Sr,c1predundant,Smatchingr,c=Sr1,c1+smatching,Sr,c=max(Smissingr,c,Sredundantr,c,Smatchingr,c),formulae-sequencesuperscriptsubscript𝑆missing𝑟𝑐superscript𝑆𝑟1𝑐subscript𝑝missingformulae-sequencesuperscriptsubscript𝑆redundant𝑟𝑐superscript𝑆𝑟𝑐1subscript𝑝redundantformulae-sequencesuperscriptsubscript𝑆matching𝑟𝑐superscript𝑆𝑟1𝑐1subscript𝑠matchingsuperscript𝑆𝑟𝑐superscriptsubscript𝑆missing𝑟𝑐superscriptsubscript𝑆redundant𝑟𝑐superscriptsubscript𝑆matching𝑟𝑐\displaystyle\begin{split}S_{\text{missing}}^{\,r,\,c}&=S^{\,r-1,\,c}-p_{\text% {missing}},\\ S_{\text{redundant}}^{\,r,\,c}&=S^{\,r,\,c-1}-p_{\text{redundant}},\\ S_{\text{matching}}^{\,r,\,c}&=S^{\,r-1,\,c-1}+s_{\text{matching}},\\ S^{\,r,\,c}&=\max(S_{\text{missing}}^{\,r,\,c},S_{\text{redundant}}^{\,r,\,c},% S_{\text{matching}}^{\,r,\,c}),\end{split}start_ROW start_CELL italic_S start_POSTSUBSCRIPT missing end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_r , italic_c end_POSTSUPERSCRIPT end_CELL start_CELL = italic_S start_POSTSUPERSCRIPT italic_r - 1 , italic_c end_POSTSUPERSCRIPT - italic_p start_POSTSUBSCRIPT missing end_POSTSUBSCRIPT , end_CELL end_ROW start_ROW start_CELL italic_S start_POSTSUBSCRIPT redundant end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_r , italic_c end_POSTSUPERSCRIPT end_CELL start_CELL = italic_S start_POSTSUPERSCRIPT italic_r , italic_c - 1 end_POSTSUPERSCRIPT - italic_p start_POSTSUBSCRIPT redundant end_POSTSUBSCRIPT , end_CELL end_ROW start_ROW start_CELL italic_S start_POSTSUBSCRIPT matching end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_r , italic_c end_POSTSUPERSCRIPT end_CELL start_CELL = italic_S start_POSTSUPERSCRIPT italic_r - 1 , italic_c - 1 end_POSTSUPERSCRIPT + italic_s start_POSTSUBSCRIPT matching end_POSTSUBSCRIPT , end_CELL end_ROW start_ROW start_CELL italic_S start_POSTSUPERSCRIPT italic_r , italic_c end_POSTSUPERSCRIPT end_CELL start_CELL = roman_max ( italic_S start_POSTSUBSCRIPT missing end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_r , italic_c end_POSTSUPERSCRIPT , italic_S start_POSTSUBSCRIPT redundant end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_r , italic_c end_POSTSUPERSCRIPT , italic_S start_POSTSUBSCRIPT matching end_POSTSUBSCRIPT start_POSTSUPERSCRIPT italic_r , italic_c end_POSTSUPERSCRIPT ) , end_CELL end_ROW (3)

where smatching=1.0subscript𝑠matching1.0s_{\text{matching}}=1.0italic_s start_POSTSUBSCRIPT matching end_POSTSUBSCRIPT = 1.0 represents the reward score after a successful matching. If an action considered missing or redundant is classified as a conservative action, the penalties pmissingsubscript𝑝missingp_{\text{missing}}italic_p start_POSTSUBSCRIPT missing end_POSTSUBSCRIPT and predundantsubscript𝑝redundantp_{\text{redundant}}italic_p start_POSTSUBSCRIPT redundant end_POSTSUBSCRIPT are quantified as half of smatchingsubscript𝑠matchings_{\text{matching}}italic_s start_POSTSUBSCRIPT matching end_POSTSUBSCRIPT, i.e., 0.5. Conversely, if an action is not conservative, both penalties are assigned the same magnitude as smatchingsubscript𝑠matchings_{\text{matching}}italic_s start_POSTSUBSCRIPT matching end_POSTSUBSCRIPT, i.e., 1.0. This approach is based on the premise that omitting a crucial meta action or inaccurately introducing a non-existent one equally hampers the effectiveness of the action sequence. The final score Scoreaction𝑆𝑐𝑜𝑟subscript𝑒actionScore_{\text{{action}}}italic_S italic_c italic_o italic_r italic_e start_POSTSUBSCRIPT action end_POSTSUBSCRIPT should be divided by the length of the selected reference meta-action sequence, formulated as follow:

Scoreaction=Sr,cNr𝑆𝑐𝑜𝑟subscript𝑒actionsuperscript𝑆𝑟𝑐subscript𝑁𝑟\displaystyle Score_{\,\text{action}}=\frac{S^{\,r,\,c}}{N_{r}}italic_S italic_c italic_o italic_r italic_e start_POSTSUBSCRIPT action end_POSTSUBSCRIPT = divide start_ARG italic_S start_POSTSUPERSCRIPT italic_r , italic_c end_POSTSUPERSCRIPT end_ARG start_ARG italic_N start_POSTSUBSCRIPT italic_r end_POSTSUBSCRIPT end_ARG (4)

Appendix C Qualitative Results

To further demonstrate the effectiveness and robustness of our DriveVLM, we provide additional visualization results in Figure 14151617, and 18. In Figure 14, DriveVLM recognizes the slowly moving vehicle ahead and provides a driving decision to change lanes for overtaking. In Figures 15  and 16, DriveVLM accurately identifies the type of unconventional vehicles and a fallen tree, demonstrating its capability in recognizing long-tail objects. In Figure 17, the traffic police signaling to proceed with hand gestures has been accurately captured by DriveVLM. In Figure 18, DriveVLM successfully recognizes the road environment of a roundabout and generates a planned trajectory with a curved path.

Refer to caption
Figure 14: Visualization of DriveVLM’s output. DriveVLM recognizes the slowly moving vehicle ahead and provides a driving decision for changing lanes to overtake.
Refer to caption
Figure 15: Visualization of DriveVLM’s output. DriveVLM has identified the type of the unconventional vehicle ahead.
Refer to caption
Figure 16: Visualization of DriveVLM’s output. DriveVLM precisely detect the fallen tree and its position, subsequently planning an appropriate detour trajectory.
Refer to caption
Figure 17: Visualization of DriveVLM’s output. The traffic police signaling to proceed with hand gestures has been accurately captured by DriveVLM.
Refer to caption
Figure 18: Visualization of DriveVLM’s output. DriveVLM successfully recognizes the road environment of a roundabout and generates a planned trajectory with a curved path.