Home >Technology peripherals >AI >End-to-end autonomous driving, multi-modal scene brain, new 3D interaction: SenseTime Jueying large model will be unveiled at the Beijing Auto Show
In 2024, the Beijing International Automobile Exhibition will return with a bang from April 25 to May 4. As the first company in the industry to propose a general model for autonomous driving with integrated perception and decision-making, SenseTime Jueying will bring its first on-car demonstration of UniAD, a true end-to-end autonomous driving solution for mass production. At the same time, the AI large model cockpit product matrix with the multi-modal scene brain as the core and the new cockpit 3D interaction will also be displayed.
Currently, general artificial intelligence is being deeply integrated with the smart car industry. Based on a series of original ecological car models built by the company’s years of technology accumulation, SenseTime Jueying will bring a variety of products covering smart driving and smart cockpits at the Beijing Auto Show. AI large-scale model products and functions demonstrate to the public a future experience that is flexible, adaptive, deeply personalized, safe, trustworthy, and humane. The launch of these products and functions will further accelerate the development of the smart car industry and create a new experience in smart travel.
UniAD, the industry’s first universal model for integrated autonomous driving with perception and decision-makingFirst on-car demonstration
SenseTime and its partner laboratories proposed UniAD, the industry's first general model for autonomous driving with integrated perception and decision-making, won the best paper at the International Conference on Computer Vision and Pattern Recognition (CVPR) in 2023. This model leads the end-to-end autonomous driving trend in academia and industry, achieving smooth end-to-end autonomous driving. In the future, autonomous driving will become more popular, and SenseTime's technology will play an important role in the field of autonomous driving, leading the industry to realize the trend of end-to-end autonomous driving.
Recently, Tesla has begun to push the FSD V12 version based on the end-to-end algorithm to users, which has attracted widespread attention in the industry. At this Beijing Auto Show, SenseTime will also bring a true end-to-end autonomous driving solution with integrated perception and decision-making for mass production - UniAD, and will display UniAD's real-vehicle road demonstration results to the public for the first time.
With the continuous evolution of autonomous driving technology, rule-based smart driving solutions require a large amount of resources invested in multi-sensor fusion sensors to deal with complex road conditions such as unprotected left turns and mixed traffic between people and vehicles. The SenseTime UniAD solution only relies on the visual perception of the camera, without the need for high-precision maps, and can cope with the complex environment of urban areas through data learning and driving. Not only that, UniAD can also drive freely on rural roads without a center line that are difficult to break through with traditional solutions, completing a series of difficult operations including large-angle left turns on bridges, avoiding road-occupying vehicles and construction areas, and detouring running pedestrians. Truly "drive like a human being".
In order to enhance the interpretability and interaction capabilities of the end-to-end autonomous driving solution, relying on the multi-modal large model, SenseTime has developed a new generation of autonomous driving large model DriveAGI, which is applicable to wider scenarios, higher performance, The threshold is lower, and the experience is perceptible, interactive, and trustworthy. As the next generation of autonomous driving technology, DriveAGI will also be previewed during the auto show.
Shangtang Jueying creates a multi-modal scene brain with keen insight, deep thinking, and efficient execution
Xiaomi SU7, which has attracted much attention recently, has started to bring AI large models Enter the cockpit. SenseTime Jueying has created an AI large model cockpit product matrix with the "multimodal scene brain" as the core. Through the combination of multimodal large models, large language models, text generation models and other capabilities, it injects powerful environmental understanding into the smart cockpit. , logical thinking and content generation abilities.
The large AI model installed on SenseTime Jueying will be able to recommend restaurants that match the user's preferences based on geographical location, introduce external natural scenery that the user is interested in, and recommend high-quality scenic spots. Not only that, based on powerful multi-modal sensing capabilities, vehicles equipped with SenseTime's Jueying solution can accurately perceive and identify information outside the vehicle, including the models of surrounding vehicles, landmark buildings, etc., through voice, pictures or videos. Provide more accurate and comprehensive content for cabin personnel and help users understand and master external environment information.
With the multi-modal scene brain as the core, SenseTime Jueying has promoted the mass production of a series of AI large-model cockpit products with panoramic perception, active care and creativity, such as "Travel Doctor", "God" With innovative functions such as "Pen" and "Child Companion", the public will be able to experience a deeply personalized and humanistic new smart travel experience at the SenseTime Jueying booth.
Vision Pro "Get in the car with naked eyes", SenseTime Jueying's world's first cockpit 3D Vision interaction is about to debut
In 2023, the emergence of Apple Vision Pro will allow people to appreciate the charm of 3D interaction. Through its profound technical research and development capabilities and efficient innovation and iteration efficiency, SenseTime will bring two new cockpit 3D interactive demonstrations at the auto show: 3D Gaze high-precision eye interaction and 3D dynamic gesture interaction.
As the world’s first smart cockpit technology that can interact with screen icons through gaze positioning, SenseTime’s 3D Gaze high-precision gaze interaction will allow users to control central control icons through their eyes. 3D dynamic gesture interaction is the industry's leading smart cockpit technology that supports dynamic gestures and hand micro-movement recognition, allowing users to conduct various cockpit interactions "through the air" through gestures.
The two in-cabin 3D interactive functions of SenseTime Jueying cooperate with each other, just like "Vision Pro" naked eyes on the car, to create a natural interactive experience for users that is more in line with human intuition.
As AI large models collide and integrate with the automotive industry, SenseTime Jueying will give full play to its leading technological advantages and product innovation capabilities accumulated over the years, and continue to inject new vitality from AGI into the smart car industry. Starting from April 25, welcome to the SenseTime Jueying booth to explore the future travel methods of smart cars in the AGI era.
The above is the detailed content of End-to-end autonomous driving, multi-modal scene brain, new 3D interaction: SenseTime Jueying large model will be unveiled at the Beijing Auto Show. For more information, please follow other related articles on the PHP Chinese website!