China Automotive Multimodal Interaction Development Research Report, 2025
Research on Automotive Multimodal Interaction: The Interaction Evolution of L1~L4 Cockpits
ResearchInChina has released the "China Automotive Multimodal Interaction Development Research Report, 2025". This report comprehensively sorts out the installation of Interaction Modalities in automotive cockpits, multimodal interaction patents, mainstream cockpit interaction modes, application of interaction modes in key vehicle models launched in 2025, cockpit interaction solutions of automakers/suppliers, and integration trends of multimodal interaction.
I. Closed-Loop Evolution of Multimodal Interaction: Progressive Evolution of L1~L4 Intelligent Cockpits
According to the "White Paper on Automotive Intelligent Cockpit Levels and Comprehensive Evaluation" jointly released by the China Society of Automotive Engineers (China-SAE), five levels of intelligent cockpits are defined: L0-L4.
As a key driver for cockpit intelligence, multimodal interaction capability relies on the collaboration of AI large models and multiple hardware to achieve the fusion processing of multi-source interaction data. On this basis, it accurately understands the intentions of drivers and passengers and provides scenario-based feedback, ultimately achieving natural, safe, and personalized human-machine interaction. Currently, the automotive intelligent cockpit industry is generally in the L2 stage, with some leading manufacturers exploring and moving towards the L3.
The core feature of L2 intelligent cockpits is "strong perception, weak cognition". In the L2 stage, the multimodal interaction function of cockpits achieves signal-level fusion. Based on multimodal large model technology, it can "understand users' ambiguous intentions" and "simultaneously process multiple commands" to execute users' immediate and explicit commands. At present, most mass-produced intelligent cockpits can enable this.
In the case of Li i6, it is equipped with MindGPT-4o, the latest multimodal model which boasts understanding and response capabilities with ultra-long memory and ultra-low latency, and features more natural language generation. It supports multimodal "see and speak" (voice + vision fusion search: allowing illiterate children to select the cartoons they want to watch by describing the content on the video cover); multimodal referential interaction (voice + gesture: ① Voice reference to objects: while issuing commands, extend the index finger: pointing left can control the window and complete vehicle control. ② Voice reference to personnel: passengers in the same row can achieve voice control over designated personnel through gesture and voice coordination, e.g., pointing right and saying "Turn on the seat heating for him").
The core feature of L3 intelligent cockpits is "strong perception, strong cognition". In the L3 stage, the multimodal interaction function of cockpits achieves cognitive-level fusion. Relying on large model capabilities, the cockpit system can comprehensively understand the complete current scenario and actively initiate reasonable services or suggestions without the user issuing explicit commands.
The core feature of L4 intelligent cockpits is "full-domain cognition and autonomous evolution", creating a "full-domain intelligent manager" for users. In the L4 stage, the application of intelligent cockpits will go far beyond the tool attribute and become a "digital twin partner" that can predict users' unspoken needs, have shared memories, and dispatch all resources for users. Its core experience is: before the user clearly perceives or expresses the need, the system has completed prediction and planning and entered the execution state.
II. Multimodal AI Agent: Understand What You Need and Predict What You Think
AI Agent can be regarded as the core execution unit and key technical architecture for the specific implementation of functions in the evolution of intelligent cockpits from L2 to L4. By integrating voice, vision, touch and situational information, AI Agent can not only "understand" commands, but also "see" the environment and "perceive" the state, thereby integrating the original discrete cockpit functions into a coherent, active and personalized service process.
Agent applications under L2 can be regarded as "enhanced command execution", which is the ultimate extension of L2 cockpit interaction capabilities. Based on large model technology, the cockpit system decomposes a user's complex command into multiple steps and then calls different Agent tools to execute them. For example, a passenger says: "I'm tired, help me buy a cup of coffee." The large model of the L2 cockpit system will understand this complex command and then call in sequence:
1.Voice Agent: Parse user needs in real time;
2.Food Ordering Agent: Recommend the best options according to user preferences, real-time location, and restaurant business status;
3.Payment Agent: Automatically complete unconscious payment;
4.Delivery Agent: Dynamically plan the food delivery time combined with vehicle navigation data (e.g., "food arrives when the car arrives", ensuring that the food is delivered synchronously when the user reaches the destination).
Currently, Agent applications are essentially responses and executions to a user's explicit and complex commands. The cockpit system does not do anything "actively", and it just "completes the tasks assigned by the user" more intelligently.
Case (1): IM Motors released the "IM AIOS Ecological Cockpit" jointly developed with Banma Zhixing. This cockpit is the first to implement Alibaba's ecosystem services in the form of AI Agent, creating a "No Touch & No App" human-vehicle interaction mode. The "AI Food Ordering Agent" and "AI Ticketing Agent" functions launched by the IM AIOS Ecological Cockpit allow users to complete food selection/ticketing and payment only through voice interaction without needing manual operation.
Case (2): On August 4, 2025, Denza officially launched the "Car Life Agent" intelligent service system at its brand press conference, which is first equipped on two flagship models, Denza Z9 and Z9GT. The "Car Life Agent" supports voice food ordering and enables payment by face with face recognition technology. After completing the order, the system will automatically plan the navigation route, forming a seamless experience of "demand-service-closed loop".
In the next level of intelligent cockpits, Agent applications will change from "you say, I do" to "I watch, I guess, I suggest, let's do it together". Users do not need to issue any explicit commands. They just sigh and rub their temples, and the system can comprehensively judge data from "camera" (tired micro-expressions), "biological sensors" (heart rate changes), "navigation data" (continuous driving for 2 hours), and "time" (3 pm (afternoon sleepiness period)) via the large model to know that "the user is in the tired period of long-distance driving and has the need to rest and refresh". Based on this, the system will take the initiative to initiate interaction: "You seem to need a rest. There is a service zone* kilometers ahead with your favorite ** coffee. Do you need me to turn on the navigation? At the same time, I can play refreshing music for you." After the user agrees, the system then calls navigation, entertainment and other Agent tools.
Automotive 4D Radar Industry Research Report 2025
4D radar research: From "optional" to "essential," 4D radar's share will exceed 50% by 2030.
1. 4D imaging radar has transformed from an "optional" to a "must-have" sensor.
4D radar adds the detecti...
China Automotive Multimodal Interaction Development Research Report, 2025
Research on Automotive Multimodal Interaction: The Interaction Evolution of L1~L4 Cockpits
ResearchInChina has released the "China Automotive Multimodal Interaction Development Research Report, 2025"...
Automotive Vision Industry Report, 2025
Automotive Vision Research: Average Camera Installation per Vehicle Reaches 5.2 Units, and Front-View Tricam Installation Exceeds 1.2 Million Sets.
From January to September 2025, the total installa...
Automotive Infrared Night Vision System Research Report, 2025
Automotive night vision research: The rise of infrared AEB, with automotive infrared night vision experiencing a 384.7% year-on-year increase from January to September.
From January to September 2025...
New Energy Vehicle Cross-Domain (Electric Drive System and Powertrain Domain) Integration Trend Report 2025-2026
Electric Drive and Powertrain Domain Research: New technologies such as three-motor four-wheel drive, drive-brake integration, and corner modules are being rapidly installed in vehicles.
Electric dri...
Analysis on Desay SV and Joyson Electronic's Electrification, Connectivity, Intelligence and Sharing, 2025
Research on Desay SV and Joyson Electronic: Who is the No.1 Intelligent Supplier?
Both Desay SV and Joyson Electronic are leading domestic suppliers in automotive intelligence. "Analysis on Desay SV ...
OEMs and Tier 1 Suppliers' Cost Reduction and Efficiency Enhancement Strategy Analysis Report, 2025
ResearchInChina released the "OEMs and Tier 1 Suppliers' Cost Reduction and Efficiency Enhancement Strategy Analysis Report, 2025", summarizing hundreds of cost reduction strategies to provide referen...
Automotive Fixed Panoramic Sunroof and Smart Roof Research Report, 2025
With the intelligent application of car roofs as the core, this report systematically sorts out a series of new products such as fixed panoramic sunroof/openable sunroof, ceiling screen, roof ambient ...
Automotive-Grade Power Semiconductor and Module (SiC, GaN) Industry Research Report, 2025
SiC/GaN Research: Sales volume of 800V+ architecture-based vehicles will increase more than 10 times, and hybrid carbon (SiC+IGBT) power modules are rapidly being deployed in vehicles.
Sales volume o...
Cockpit Agent Engineering Research Report, 2025
Cockpit Agent Engineering Research: Breakthrough from Digital AI to Physical AI
Cockpit Agent Engineering Research Report, 2025 starts with the status quo of cockpit agents, summarizes the technical ...
Prospective Study on L3 Intelligent Driving Technology of OEMs and Tier 1 Suppliers, 2025
L3 Research: The Window of Opportunity Has Arrived - Eight Trends in L3 Layout of OEMs and Tier 1 Suppliers
Through in-depth research on 15 OEMs (including 8 Chinese and 7 foreign OEMs) and 9 Tier 1 ...
China Commercial Vehicle IoV and Intelligent Cockpit Industry Research Report 2025
Commercial Vehicle IoV and Cockpit Research: The Third Wave of Passenger Car/Commercial Vehicle Technology Integration Arrives, and T-Box Integrates e-Call and 15.6-inch for Vehicles
I. The third wav...
Intelligent Vehicle Electronic and Electrical Architecture (EEA) and Technology Supply Chain Construction Strategy Research Report, 2025
E/E Architecture Research: 24 OEMs Deploy Innovative Products from Platform Architectures to Technical Selling Points
According to statistics from ResearchInChina, 802,000 passenger cars with domain...
Research Report on Intelligent Vehicle Cross-Domain Integration Strategies and Innovative Function Scenarios, 2025
Cross-Domain Integration Strategy Research: Automakers' Competition Extends to Cross-Domain Innovative Function Scenarios such as Cockpit-Driving, Powertrain, and Chassis
Cross-domain integration of ...
China Autonomous Driving Data Closed Loop Research Report, 2025
Data Closed-Loop Research: Synthetic Data Accounts for Over 50%, Full-process Automated Toolchain Gradually Implemented
Key Points:From 2023 to 2025, the proportion of synthetic data increased from 2...
Automotive Glass and Smart Glass Research Report, 2025
Automotive Glass Report: Dimmable Glass Offers Active Mode, Penetration Rate Expected to Reach 10% by 2030
ResearchInChina releases the Automotive Glass and Smart Glass Research Report, 2025. This r...
Passenger Car Brake-by-Wire (BBW) Research Report, 2025
Brake-by-Wire: EHB to Be Installed in 12 Million Vehicles in 2025
1. EHB Have Been Installed in over 10 Million Vehicles, A Figure to Hit 12 Million in 2025.
In 2024, the brake-by-wire, Electro-Hydr...
Autonomous Driving Domain Controller and Central Computing Unit (CCU) Industry Report, 2025
Research on Autonomous Driving Domain Controllers: Monthly Penetration Rate Exceeded 30% for the First Time, and 700T+ Ultrahigh-compute Domain Controller Products Are Rapidly Installed in Vehicles
L...