LLVM-AD @ ITSC 2024
About the Workshop
The 2nd Workshop on Large Language and Vision Models for Autonomous Driving (LLVM-AD) at ITSC 2024 aims to bring together professionals from academia and industry to explore the application of large language and vision models in autonomous driving. As part of this initiative, the 2nd LLVM-AD workshop launches an open dataset challenge for real-world traffic understanding.
Keynote Speakers
Workshop Recording
Organizers
Keynote Talks
Towards Scalable Autonomy (Manmohan Chandraker)
Abstract: While autonomous driving has made large strides over the past decade to be deployed in practice today, scaling it across a diversity of conditions and behaviors remains expensive. This talk explores the possibility of recent advances in large language models (LLMs) and computer vision foundational models allowing the development of more scalable driving, simulation and devops stacks for autonomous mobility. These include perception and planning that leverage language guidance, photorealistic simulation of safety-critical scenarios with neural rendering and controllable diffusion, automated devops based on vision-language understanding and a look ahead to agentic LLMs that automate complex autonomy workflows.
Bio: Manmohan Chandraker is a professor in the CSE department of the University of California, San Diego and leads computer vision research at NEC Labs. His research interests are in vision, learning and graphics, with applications to autonomous driving and augmented reality. His works have been recognized with best paper awards at CVPR, ICCV and ECCV, the NSF CAREER Award, Qualcomm and Google Research Awards. He serves on NSF panels on vision, learning and robotics and on senior program committees at CVPR, ICCV, ECCV, AAAI, NeurIPS and ICLR.
Large Language and Vision Models for Autonomous Driving (Long Chen)
Bio: Long Chen is a distinguished research lead, with a proven track record in developing disruptive AI technologies. He currently holds the position of Staff Scientist at Wayve, where he is at the forefront of building vision-language-action (VLA) models for the next wave of autonomous driving, such as Driving-with-LLMs and LINGO. Previously, he was a research engineer at Lyft Level 5, where he led the data-driven planning models from crowd-sourced data for Lyft’s self-driving cars. His extensive experience also includes applying AI technologies in various domains such as mixed reality, surgical robots, and healthcare.
Schedule
Time | Event |
---|---|
8:30-8:45 | Opening Remarks |
8:45-9:30 | Keynote: Large Language and Vision Models for Autonomous Driving Long Chen, Wayve https://long.ooo/ |
9:30-9:45 | ContextVLM: Zero-Shot and Few-Shot Context Understanding for Autonomous Driving Using Vision Language Models Shounak Sural, CMU ssural@andrew.cmu.edu |
9:45-10:00 | Semantic Trajectory Data Mining with LLM-Informed POI Classification Yifan Liu, UCLA https://mobility-lab.seas.ucla.edu/ |
10:00-10:15 | A LLM-based Multimodal Warning System for Driver Assistance Zixuan Xu, KAIST zixuan.xu@kaist.ac.kr |
10:15-10:30 | Real-Time Data Informed Intelligent ChatBot for Transportation Surveillance and Management Joey Cai, UW zhiyu_cai@berkeley.edu (Zoom) |
10:30-10:45 | Personalized Autonomous Driving with Large Language Models: Field Experiments Ziran Wang, Purdue ziran@purdue.edu |
10:45-11:00 | ScenarioQA: Evaluating Test Scenario Reasoning Capabilities of Large Language Models Ishaan Paranjape, UC Santa Cruz iparanja@ucsc.edu (Zoom) |
11:00-11:15 | Large Language Models for Human-Like Autonomous Driving: A Survey Yun LI, The University of Tokyo li-yun@g.ecc.u-tokyo.ac.jp |
11:15-11:30 | Evaluating and Enhancing Trustworthiness of LLMs in Perception Tasks Malsha Mahawatta, University of Gothenburg malsha.mahawatta@gu.se |
11:30-12:15 | Keynote: Towards Scalable Autonomy Manmohan Chandraker, NEC Labs America https://www.nec-labs.com/research/media-analytics/people/manmohan-chandraker/ |
12:15-12:30 | Closing Remarks |