search
HomeTechnology peripheralsAIThis article will give you an easy-to-understand understanding of autonomous driving

The principles of autonomous driving can often be best understood from the architecture of autonomous driving. The public’s most understandable understanding of autonomous driving is perception, decision-making, and execution. All robots have this structure.

This article will give you an easy-to-understand understanding of autonomous driving

  • Perception answers questions about what is around, similar to human eyes and ears. Obtain information about surrounding obstacles and roads through cameras, radar, maps and other means.
  • Decision-making answers the question of what I want to do, similar to the brain. By analyzing the sensed information, the path and vehicle speed are generated.
  • Execution is similar to hands and feet, converting decision-making information into brake, accelerator and steering signals to control the vehicle to drive as expected.

Next we go deeper and the problem becomes a little more complicated.

This article will give you an easy-to-understand understanding of autonomous driving

In our daily lives, we may intuitively think that I decide my next decision based on the information I see with my eyes at every moment. , but this is often not the case. There is always a time delay from the eyes to the head to the hands and feet, and the same is true for autonomous driving. But we don’t feel the impact because our brains automatically handle prediction. Even if it is only a few milliseconds, our decisions are based on predictions of what we see to guide the operation of our hands and feet. This is the basis for us to maintain normal functions. Therefore, we will add a prediction module before autonomous driving decisions.

The perception process also contains the universe, and upon careful consideration, it can be divided into two stages: "sensing" and "perception". "Sensing" obtains raw data from sensors such as pictures, while "perception" processes useful information from pictures (such as how many people are in the picture). As the old saying goes, "Seeing is believing, hearing is believing." The useful information of "perception" can be further divided into self-vehicle perception and external perception. People or self-driving cars often have different strategies when processing these two types of information.

  • Self-vehicle perception - information obtained by receptors at all times (including cameras, radar, GPS, etc.)
  • External perception - information collected and processed by external agents or past memories (including positioning, maps, vehicle-linked information, etc.), which requires the input of self-vehicle positioning perception (GPS).

This article will give you an easy-to-understand understanding of autonomous driving

In addition, the information about obstacles, lanes and other information processed by various sensors through algorithms is often contradictory. If the radar sees an obstacle ahead but the camera tells you there isn't, then you need to add a "fusion" module. Make further connections and judgments about inconsistent information.

Here we often summarize “fusion and prediction” as “world model”. This word is very vivid, whether you are a materialist or an idealist. It is impossible to cram all the "world" into your brain, but what guides our work and life is the "model" of the "world", that is, by processing what we see after birth, and gradually constructing a model in our mind. The Taoist understanding of the world is called "inner view". The core responsibility of the world model is to understand the attributes and relationships of the current environmental elements through "fusion", and to make "predictions" in conjunction with "a priori laws" to provide a more leisurely judgment for decision execution. This time span can range from a few milliseconds to a few seconds. Hour.

Due to the addition of the world model, the entire architecture has become richer, but there is another detail that is often overlooked. That is the flow of information. A simple understanding is that people perceive things through their eyes, process them in the brain, and then hand them over to their hands and feet for execution. However, the actual situation is often more complicated. There are two typical behaviors here that constitute a completely opposite information flow, that is, "plan for goal achievement" and "shift of attention."

How do you understand "plan to achieve goals"? In fact, the beginning of thinking is not perception but "goal". Only when you have a goal can you trigger a meaningful "perception-decision-execution" process. For example, if you want to drive to a destination, you may know several routes, and you will finally choose one of the routes based on the congestion situation. Congestion belongs to the world model, while "getting there" belongs to the decision. This is the process by which decisions are passed to the world model.

How to understand "shift of attention"? Even if it is a picture, neither humans nor machines can obtain all the information hidden inside. Starting from a need and context, we tend to focus on a limited scope and limited categories. This information cannot be obtained from the picture itself, but comes from the "world model" and "goal". It is a process from decision-making to world model and then to perception.

We add some necessary information and rearrange the entire structure. It looks like the following. Is it a little more complicated? It’s not over yet, let’s continue watching.

This article will give you an easy-to-understand understanding of autonomous driving

Autopilot algorithms, like the brain, have a requirement for processing time. The general cycle is between 10ms-100ms, which can meet the response requirements to environmental changes. But the environment is sometimes simple and sometimes very complex. Many algorithm modules cannot meet this time requirement. For example, thinking about the meaning of life may not be something that can be done in 100ms. If you have to think about life every step of the way, it must be a kind of destruction to the brain. The same is true for computers. There are physical limits to computing power and speed. The solution is to introduce a layered framework.

This layered mechanism will generally shorten the processing cycle by 3-10 times as it goes up. Of course, it does not necessarily need to appear completely in the actual framework. In engineering, it is based on on-board resources and algorithms. Usage can be flexibly adjusted. Basically, perception is an upward process that continuously refines specific elements based on attention, providing perceptual information with "depth and direction." Decision-making is a downward process, decomposing actions from the goal to each execution unit layer by layer according to different levels of world models. World models generally have no specific flow direction and are used to construct environmental information at different granular scales.

According to the complexity of the processing tasks, the division of labor and the communication environment will also be appropriately castrated and merged. For example, low-level ADAS functions (ACC) require less computing power and can be designed with only one layer. High-end ADAS functions (AutoPilot) generally have two levels of configuration. As for the autonomous driving function, there are many complex algorithms, and a three-layer design is sometimes necessary. In software architecture design, there are also situations where the world model and the perception or decision-making module at the same layer are merged.

This article will give you an easy-to-understand understanding of autonomous driving

Various autonomous driving companies or industry standards will release their own software architecture designs, but they are often the result of castration based on the status quo and are not It has universal applicability, but in order to facilitate everyone's understanding, I still substitute the current mainstream functional modules. Let's take a look at the comparison relationship, which will be more helpful for understanding the principle.

This article will give you an easy-to-understand understanding of autonomous driving

It needs to be noted in advance that although this already means a bit of software architecture, it is still a description of the principle and the actual software architecture. The design is even more complicated than this. Instead of going into all the details here, we focus on the parts that are easily confused. Let’s focus on it below.

This article will give you an easy-to-understand understanding of autonomous driving

Environment Awareness-ALL IN Deep Learning

In order to ensure that unmanned vehicles understand and grasp the environment, the environment perception part of the unmanned driving system usually needs to obtain a large amount of information about the surrounding environment, including the location of obstacles, speed, the precise shape of the lane ahead, and signs. Card position type, etc. This information is usually obtained by fusing data from multiple sensors such as lidar, peripheral/surround camera, and millimeter wave radar.

The development of deep learning has made it an industry-wide consensus to complete the construction of autonomous driving through neural network algorithms. The algorithm of the perception module is the "pawn" of the entire deep learning process and is the first software module to complete the transformation.

The correlation and difference between positioning map and V2X-self-vehicle perception and external perception

In the traditional sense Understand that external perception is based on GPS positioning signals and converts information in absolute coordinate systems such as high-precision maps and vehicle-to-everything messages (V2X) into the own vehicle coordinate system for use by vehicles. It is similar to the Amap navigator used by people. Combined with the "self-vehicle perception" information originally in the self-vehicle coordinate system, it provides environmental information for autonomous driving.

But the actual design is often more complex. Due to the unreliability of GPS, the IMU needs continuous correction. Mass-produced autonomous driving positioning often uses perceptual map matching to accurately obtain accurate absolute positioning. Position, use the sensing results to correct the IMU to obtain accurate relative position, and form redundancy with the INS system composed of GPS-IMU. Therefore, the positioning signals necessary for "external perception" often rely on "self-vehicle perception" information.

In addition, although the map is strictly part of the "world model", it is limited by the sensitivity of GPS. During the domestic software implementation process, the positioning module and The map module integrates and offsets all GPS data to ensure that no sensitive positioning information is leaked.

Fusion prediction module - the core focus is on the difference between the two

The core of fusion is to solve two problems, one It is a spatio-temporal synchronization problem. Using the coordinate system conversion algorithm and the time synchronization algorithm of software and hardware collaboration, first align the sensing measurement results of lidar, camera and millimeter wave radar to a spatio-temporal point to ensure the unity of the original data of the entire environment perception. The other is to solve the problem of association and anomaly elimination, handle the association of different sensors mapping to the same "world model" element (a person/a lane, etc.), and eliminate anomalies that may be caused by misdetection by a single sensor. However, the fundamental difference between fusion and prediction is that it only processes information from the past and the current moment, and does not process external moments.

The prediction will make a judgment on the future time based on the fusion result. This future time ranges from 10ms to 5 minutes. This includes predicting traffic lights, predicting the driving path of surrounding obstacles, or predicting the cornering position in the distance. Forecasts of different periods will provide planning for corresponding periods and predictions of different granularities, thereby providing greater space for planning adjustments.

Planning Control - Hierarchical Strategy Decomposition

##Planning is when the unmanned vehicle makes some effective decisions to achieve a certain goal. The process of purposeful decision-making. For driverless vehicles, this goal usually refers to reaching the destination from the starting point while avoiding obstacles, and continuously optimizing the driving trajectory and behavior to ensure the safety and comfort of passengers. The structural summary of the planning is to integrate information based on the environment of different granularities, conduct layer-by-layer evaluation and decomposition starting from the external goals, and finally pass it to the executor to form a complete decision-making.

Broken down, the planning module is generally divided into three layers: Mission Planning, Behavioral Planning and Motion Planning. The core is to obtain the global path based on the road network and discrete path search algorithm, which is given a large-scale task type, often with a long cycle. Behavior planning is based on the finite state machine to determine the specific behavior that the vehicle should take in a medium cycle (left change Road, avoidance, E-STOP) and set some boundary parameters and approximate path range. The motion planning layer is often based on sampling or optimization methods to ultimately obtain the only path that meets comfort and safety requirements. Finally, it is handed over to the control module to follow the unique path through feedforward prediction and feedback control algorithms, and manipulate the brake, steering, throttle, body and other actuators to finally execute the command.

I don’t know how far you have understood it, but the above is just an introduction to the principles of autonomous driving. The current theories, algorithms and architecture of autonomous driving are developing very fast, although the above content is relatively basic. The knowledge points will not be outdated for a long time. But the new demands have brought a lot of new understanding to the architecture and principles of autonomous driving.

This article will give you an easy-to-understand understanding of autonomous driving

The above is the detailed content of This article will give you an easy-to-understand understanding of autonomous driving. For more information, please follow other related articles on the PHP Chinese website!

Statement
This article is reproduced at:51CTO.COM. If there is any infringement, please contact admin@php.cn delete
在 CARLA自动驾驶模拟器中添加真实智体行为在 CARLA自动驾驶模拟器中添加真实智体行为Apr 08, 2023 pm 02:11 PM

arXiv论文“Insertion of real agents behaviors in CARLA autonomous driving simulator“,22年6月,西班牙。由于需要快速prototyping和广泛测试,仿真在自动驾驶中的作用变得越来越重要。基于物理的模拟具有多种优势和益处,成本合理,同时消除了prototyping、驾驶员和弱势道路使用者(VRU)的风险。然而,主要有两个局限性。首先,众所周知的现实差距是指现实和模拟之间的差异,阻碍模拟自主驾驶体验去实现有效的现实世界

特斯拉自动驾驶算法和模型解读特斯拉自动驾驶算法和模型解读Apr 11, 2023 pm 12:04 PM

特斯拉是一个典型的AI公司,过去一年训练了75000个神经网络,意味着每8分钟就要出一个新的模型,共有281个模型用到了特斯拉的车上。接下来我们分几个方面来解读特斯拉FSD的算法和模型进展。01 感知 Occupancy Network特斯拉今年在感知方面的一个重点技术是Occupancy Network (占据网络)。研究机器人技术的同学肯定对occupancy grid不会陌生,occupancy表示空间中每个3D体素(voxel)是否被占据,可以是0/1二元表示,也可以是[0, 1]之间的

一文通览自动驾驶三大主流芯片架构一文通览自动驾驶三大主流芯片架构Apr 12, 2023 pm 12:07 PM

当前主流的AI芯片主要分为三类,GPU、FPGA、ASIC。GPU、FPGA均是前期较为成熟的芯片架构,属于通用型芯片。ASIC属于为AI特定场景定制的芯片。行业内已经确认CPU不适用于AI计算,但是在AI应用领域也是必不可少。 GPU方案GPU与CPU的架构对比CPU遵循的是冯·诺依曼架构,其核心是存储程序/数据、串行顺序执行。因此CPU的架构中需要大量的空间去放置存储单元(Cache)和控制单元(Control),相比之下计算单元(ALU)只占据了很小的一部分,所以CPU在进行大规模并行计算

自动驾驶汽车激光雷达如何做到与GPS时间同步?自动驾驶汽车激光雷达如何做到与GPS时间同步?Mar 31, 2023 pm 10:40 PM

gPTP定义的五条报文中,Sync和Follow_UP为一组报文,周期发送,主要用来测量时钟偏差。 01 同步方案激光雷达与GPS时间同步主要有三种方案,即PPS+GPRMC、PTP、gPTPPPS+GPRMCGNSS输出两条信息,一条是时间周期为1s的同步脉冲信号PPS,脉冲宽度5ms~100ms;一条是通过标准串口输出GPRMC标准的时间同步报文。同步脉冲前沿时刻与GPRMC报文的发送在同一时刻,误差为ns级别,误差可以忽略。GPRMC是一条包含UTC时间(精确到秒),经纬度定位数据的标准格

特斯拉自动驾驶硬件 4.0 实物拆解:增加雷达,提供更多摄像头特斯拉自动驾驶硬件 4.0 实物拆解:增加雷达,提供更多摄像头Apr 08, 2023 pm 12:11 PM

2 月 16 日消息,特斯拉的新自动驾驶计算机,即硬件 4.0(HW4)已经泄露,该公司似乎已经在制造一些带有新系统的汽车。我们已经知道,特斯拉准备升级其自动驾驶硬件已有一段时间了。特斯拉此前向联邦通信委员会申请在其车辆上增加一个新的雷达,并称计划在 1 月份开始销售,新的雷达将意味着特斯拉计划更新其 Autopilot 和 FSD 的传感器套件。硬件变化对特斯拉车主来说是一种压力,因为该汽车制造商一直承诺,其自 2016 年以来制造的所有车辆都具备通过软件更新实现自动驾驶所需的所有硬件。事实证

端到端自动驾驶中轨迹引导的控制预测:一个简单有力的基线方法TCP端到端自动驾驶中轨迹引导的控制预测:一个简单有力的基线方法TCPApr 10, 2023 am 09:01 AM

arXiv论文“Trajectory-guided Control Prediction for End-to-end Autonomous Driving: A Simple yet Strong Baseline“, 2022年6月,上海AI实验室和上海交大。当前的端到端自主驾驶方法要么基于规划轨迹运行控制器,要么直接执行控制预测,这跨越了两个研究领域。鉴于二者之间潜在的互利,本文主动探索两个的结合,称为TCP (Trajectory-guided Control Prediction)。具

一文聊聊自动驾驶中交通标志识别系统一文聊聊自动驾驶中交通标志识别系统Apr 12, 2023 pm 12:34 PM

什么是交通标志识别系统?汽车安全系统的交通标志识别系统,英文翻译为:Traffic Sign Recognition,简称TSR,是利用前置摄像头结合模式,可以识别常见的交通标志 《 限速、停车、掉头等)。这一功能会提醒驾驶员注意前面的交通标志,以便驾驶员遵守这些标志。TSR 功能降低了驾驶员不遵守停车标志等交通法规的可能,避免了违法左转或者无意的其他交通违法行为,从而提高了安全性。这些系统需要灵活的软件平台来增强探测算法,根据不同地区的交通标志来进行调整。交通标志识别原理交通标志识别又称为TS

一文聊聊SLAM技术在自动驾驶的应用一文聊聊SLAM技术在自动驾驶的应用Apr 09, 2023 pm 01:11 PM

定位在自动驾驶中占据着不可替代的地位,而且未来有着可期的发展。目前自动驾驶中的定位都是依赖RTK配合高精地图,这给自动驾驶的落地增加了不少成本与难度。试想一下人类开车,并非需要知道自己的全局高精定位及周围的详细环境,有一条全局导航路径并配合车辆在该路径上的位置,也就足够了,而这里牵涉到的,便是SLAM领域的关键技术。什么是SLAMSLAM (Simultaneous Localization and Mapping),也称为CML (Concurrent Mapping and Localiza

See all articles

Hot AI Tools

Undresser.AI Undress

Undresser.AI Undress

AI-powered app for creating realistic nude photos

AI Clothes Remover

AI Clothes Remover

Online AI tool for removing clothes from photos.

Undress AI Tool

Undress AI Tool

Undress images for free

Clothoff.io

Clothoff.io

AI clothes remover

AI Hentai Generator

AI Hentai Generator

Generate AI Hentai for free.

Hot Article

R.E.P.O. Energy Crystals Explained and What They Do (Yellow Crystal)
2 weeks agoBy尊渡假赌尊渡假赌尊渡假赌
R.E.P.O. Best Graphic Settings
2 weeks agoBy尊渡假赌尊渡假赌尊渡假赌
R.E.P.O. How to Fix Audio if You Can't Hear Anyone
3 weeks agoBy尊渡假赌尊渡假赌尊渡假赌

Hot Tools

SublimeText3 Chinese version

SublimeText3 Chinese version

Chinese version, very easy to use

SublimeText3 English version

SublimeText3 English version

Recommended: Win version, supports code prompts!

MantisBT

MantisBT

Mantis is an easy-to-deploy web-based defect tracking tool designed to aid in product defect tracking. It requires PHP, MySQL and a web server. Check out our demo and hosting services.

Dreamweaver CS6

Dreamweaver CS6

Visual web development tools

WebStorm Mac version

WebStorm Mac version

Useful JavaScript development tools