


This week’s opinion
1.1 NVIDIA: Embodied Intelligence Path with Omniverse as the Core
NVIDIA chooses the path of embodied intelligence, which is characterized by development from the edge to the cloud. Its main products are VIMA robotic arms and Isaac AMR mobile robot platforms. NVIDIA released a robotic arm based on the VIMA large model at ITF2023. According to NVIDIA, VIMA can understand, reason and interact with the physical world, such as moving and arranging objects according to visual and text prompts. VIMA can also run in Omniverse to simulate physics and interact with the physical world. Make predictions consistent with the laws of physics. Isaac AMR is a platform for simulating, validating, deploying, optimizing, and managing fleets of autonomous mobile robots. It includes edge-to-cloud software services, computing, and a set of reference sensors and robot hardware to accelerate large-scale deployments by connecting to DeepMap's cloud services. The mapping and semantic understanding of the environment reduces the time it takes for robots to map large facilities from weeks to days and achieves centimeter-level accuracy without the need for a senior technical team. By generating rich 3D voxel maps, it can create occupancy maps and semantic maps for many types of adaptive mesh refinement (AMR).
Embodied intelligence refers to training algorithms by imitating the way the human brain works, thereby achieving self-understanding, self-optimization, and ultimately learning and growing in a human-like manner. NVIDIA VIMA supports multi-modal modes such as text, vision, and voice as the robot's task input, and uses systematic generalization tests including simulation benchmark tests, more than 600,000 expert trajectories, and multiple-level evaluation protocols as the target output set; using pre- The trained T5 model encodes multimodal cues and modulates the robot controller on the cues through a cross-attention layer, predicting motor commands conditioned on cues and interaction history as a set of predicted outputs; predicting outputs and targets through the model Output comparison and self-optimization drive VIMA autonomous learning. In the most difficult zero-sample generalization training task, VIMA's task success rate is up to 2.9 times higher than the previous optimal method, while using 10 times less training data, VIMA The performance is still 2.7 times better than the top competing methods.
With real-world data sets, embodied intelligence has core competitiveness, and NVIDIA Omniverse is the industry benchmark. To conduct embodied intelligence training, special data sets are required. For example, for the VIMA robotic arm, the following data is required:
1) Multi-modal task set: 17 meta-tasks, each meta-task can be split into 1000 separate tasks, including various multi-modal tasks such as semantic understanding and visual segmentation
2) Success Stories: NVIDIA has prepared 650K success trajectories
3) Reward benchmark: Establish VIMA-Bench to use probability theory methods to reasonably evaluate task AI prediction results and give AI correct feedback
NVIDIA Omnivers is the world's leading digital twin platform, consisting of five important components: Nucleus, Connect, Kit, Simulation and RTX renderer. These components as well as interoperable third-party digital content creation (DCC) tools and renderers , plus extensions, applications and microservices built by third parties and NVIDIA form a complete Omniverse ecosystem. Through real-world data sets such as USD, digital twins can accurately comply with physical laws, accurately respond to object movements, and be synchronized with displays. These real-world data are an important foundation for NVIDIA's embodied intelligence to be implemented.
NVIDIA's embodied intelligence path for edge AI's bottom-up breakthrough has been opened up. Combined with cloud computing, it can provide rich technical support in all aspects of edge AI from development to deployment. NVIDIA selected Microsoft Azure as the first cloud service provider for Omniverse Cloud. Microsoft Azure will enable enterprises to access the full set of Omniverse software applications and NVIDIA OVX™ computing systems while enjoying the scale analysis features and security of Azure cloud services, providing customers with a set of tools that can be used to design, develop, deploy and manage industrial elements. Universe Application’s full-stack cloud environment and platform functions connect and use related products from NVIDIA’s partner ecosystem, such as NVIDIA certified workstations, edge computing modules, etc.
NVIDIA Isaac is an acceleration platform specially built for robot development and AI, and the AMR mobile robot platform was the first to be launched. The NVIDIA Isaac platform starts with a pre-trained model, augmented with synthetic data generated in Isaac Replicator, and trained with NVIDIA TAO to achieve target performance; leverage NVIDIA Isaac Sim available on-premises and in the cloud to create physically accurate and realistic environments to develop and test all aspects related to running your robot; save time with Nova Ori and bring AI to NVIDIA Jetson-based robots using hardware acceleration SDKs, such as Isaac ROS GEM for ROS-based robots, Isaac ROS GEM for video stream parsing NVIDIA DeepStream SDK, NVIDIA Riva for natural language processing; manage robot fleets to optimize productivity with EGX Fleet Command and Isaac for AMR including Metropolis, CuOpt and DeepMap. Isaac AMR is a platform for simulating, validating, deploying, optimizing and managing fleets of autonomous mobile robots. It includes edge-to-cloud digital twin training, software services, compute and a set of reference sensors and robotic hardware to accelerate the development and deployment of AMRs. Deployment speed, reducing costs and time to market.
Isaac AMR is built on the NVIDIA Nova Orin reference architecture. Nova Orin integrates multiple sensors and system modules including stereo cameras, fisheye cameras, 2D and 3D lidar, supports advanced AI and hardware acceleration algorithms, and provides 275TOPS of real-time edge computing performance. Synchronized and calibrated sensor suites provide sensor diversity and redundancy for real-time 3D sensing and mapping. Cloud-native tools for recording, uploading and playback, making debugging, map creation, training and analysis easier.
1.2 Qualcomm: Hybrid AI path integrating cloud and edge
Qualcomm announced the launch of a cloud-edge hybrid AI solution, committed to in-depth development of edge AI, while also cooperating with Microsoft and other companies to develop cloud AI. According to the Qualcomm hybrid AI white paper, in hybrid AI scenarios, the edge large model is the sensing organ of the cloud large model. For example, the user speaks to the mobile phone, and the automatic speech recognition (ASR) AI model such as Whisper converts the speech into text on the device and sends it to the cloud. , run large models in the cloud and send back text answers. In the advanced version, device AI further protects privacy, takes on more processing, and provides more personalized prompts to the cloud: through device learning and personal data, the device creates a user profile, collaborates with the scheduler, and provides better context-based Prompts; for example, the user asks for a mobile phone reservation to eat with friends at a favorite restaurant. For simple queries, a smaller large model can be run on the device without cloud interaction. If the user needs complex information, the demand is converted into prompts locally and sent to Large model in the cloud and return detailed answers.
According to Qualcomm Hybrid AI white paper, hybrid AI mainly has the following advantages:
1) Economy: Reduce cloud inference costs, efficiently utilize edge computing power, and lower the threshold for AI application development;
2) Low energy consumption: Edge devices can run large models with lower energy consumption. If factors such as processing and data transmission are taken into account, the energy consumption savings will be more significant;
3) Reliability: Edge AI is not affected by network conditions and runs more stably;
4) Privacy: The data is completely inferred locally, greatly reducing the risk of leaks;
5) Personalization: Edge devices can collect users’ real-life behaviors, values, pain points, needs, concerns and other information to form customized services.
Software side: Qualcomm AI development stack has been released. Qualcomm AI development stack supports mainstream AI frameworks such as TensorFlow, PyTorch, ONNX, and Keras, as well as runtimes such as TensorFlow Lite, TensorFlow Lite Micro, and ONNX runtime. In addition, it includes inference software development kits (SDKs) such as the popular Qualcomm® Neural Processing SDK for Android, Linux and Windows. Developer libraries and services support the latest programming languages, virtual platforms, and compilers. At a lower level, Qualcomm AI development stack system software includes a basic real-time operating system (RTOS), system interfaces and drivers. Qualcomm AI development stack provides rich and diverse operating system support on different product lines, including Android, Windows, Linux and QNX, as well as deployment and monitoring infrastructure such as Prometheus, Kubernetes and Docker. Qualcomm AI development stack also includes AI Studio, which supports everything from A complete large model workflow from design to optimization, deployment and analysis, integrating all tools into a graphical user interface and providing visualization tools to simplify the developer experience and enable them to view model development in real time, including AI model efficiency tools Package (AIMET), AIMET Model Repository, Model Analyzer and Neural Architecture Search (NAS).
Hardware side: Qualcomm Hexagon Processor core. Qualcomm Hexagon Processor adopts a new architecture and a dedicated power supply system. It adds special hardware to improve group convolution and activation function acceleration in AI inference, and doubles the performance of tensor accelerator. It uses a unique method to convert complex The AI model is decomposed into Micro Tiles to accelerate the inference process. Scalar, vector and tensor accelerators can work simultaneously without involving memory each time, thus saving power consumption and time. In addition, seamless multi-IP communication with Hexagon is achieved through physical bridges. This connection can be used to drive those high-bandwidth, low-latency applications, such as cognitive ISPs or boosting low resolutions in gaming scenarios. Qualcomm Hexagon Processor successfully converted multiple deep learning models from FP32 to INT4, reducing energy consumption by up to 60% while increasing performance by 90%.
Qualcomm has already used the stable diffusion model in practice, and its future plans for large-scale artificial intelligence edge deployment are very clear. In February 2023, Qualcomm used the Qualcomm AI Stack to perform full-stack AI optimization and deployed Stable Diffusion on Android smartphones for the first time. Qualcomm released a hybrid artificial intelligence white paper in May 2023, which predicted that by the end of 2023, edge artificial intelligence technology will cover various models with parameters below 1 billion.
1.3 Chuangda: Integrated development of software and hardware for large models and large platforms
On May 18, 2023, Thunderbolt released the Rubik large model, which is the prototype of the first unified operating system for edge AI terminals in China. At the same time, Thunderstar also jointly established an artificial intelligence joint innovation laboratory with Amazon Cloud Technology , and displayed a series of Chuangda Rubik's Cube products such as TurboX modules, smart speaker reference designs, Rubik GeniusCanvas, etc. According to the company's 2022 annual report, since its establishment, the company has been accumulating and accumulating terminal-side, edge-side, and cloud-side technologies, and has now become the world's leading technology manufacturer in the above fields. In addition, the company's robot products have covered almost all currently existing robot scenarios and have been favored by many robot manufacturers around the world. In addition, the company maintains in-depth cooperation with leading companies in the industry chain's technology and products, building an ecological advantage. The company has large models on the software side of edge AI, and on the hardware side, it is deeply involved in the ecology of technology giants such as Qualcomm and Amazon. The development path of integrated software and hardware has great potential.
1.3.1 Horizontal view: Chuangda Rubik builds AI ecosystem around existing business
The core product of the Rubik large model series is the Rubik Language large model. The chairman predicts that it will reach the ChatGPT3.5 level in 2024. Rubik Edge, Rubik Multi-Modal, and the Rubit Robot expected to be launched in 2027 will all serve the fields of smartphones and smart driving, improving the human-computer interaction experience. At the same time, the Rubik series builds an ecosystem around human-computer interaction and existing businesses: RubikStudio, RubikAuto, RubikDevice and Rubik Enterprise. While having a powerful large model, it will also turn the large model into a variety of small and medium-sized models to meet the splitting of various scenarios and knowledge and improve the adaptability with customers.
Similar to Google, Chuangda’s Rubik large model is expected to be the first to be launched as a robot. Among the many intelligent hardware products, Thunderstar's robot products can cover a variety of different application scenarios, and have helped many robot manufacturers around the world achieve mass production of their products. Based on its profound accumulation in the field of robotics, Thunderousas has integrated smart speakers with robots, and through continuous training with the Rubik large model, it has realized an intelligent sales robot that can talk freely and can independently answer customers' questions about the company and products. kind of problem.
1.3.2 Vertical view: Chuangda AI application ecosystem is ready to go
The Smart to Intelligent strategy of Thunderstar is launched to realize a new intelligent world from intelligent application-centered to model-driven machine-machine, machine-human interaction. The Rubik large model will be integrated with the company's existing smart car and IoT businesses and meet the needs of various industries through privatized deployment and system tuning. The company is expected to turn robots into reality through continuous optimization of large models and exert its core competitive advantages in the toBtoC field of the future intelligent computing industry.
RUBIK Auto: A car is essentially a robot. Car manufacturers have three main needs for large car models. The first is terminal-side operation. The terminal-side experience, data, and performance are the best and can best protect customer privacy. , but the hardware requirements are higher; the second is private cloud Plugin, which can be flexibly tuned; the third is the coexistence of multiple open models. The company's RUBIK Auto will support customer privatization deployment (it has cooperated with leading overseas car manufacturers to conduct POC research and development based on the company's model), and also supports model quantification and tailoring, and then adapts to various chips to flexibly interface with car manufacturers.
RUBIK Device: Among smart hardware, as long as it involves large computing products, Chuangda’s leading smart hardware originally deployed operating systems can directly add AI. Once AI settles on the edge, it means that the smart hardware becomes a robot. It will form the center of the scene. Whether it is a home scene, a building scene, or a factory scene, every scene can become an intelligent center through edge deployment.
RUBIK Enterprise (Enterprise Edition): One of Chuangda’s obvious advantages is internationalization. The company has R&D centers and teams in 15 countries and regions around the world. The company will deploy through localization and support customers' privatized deployment; the second obvious advantage of the company is the integration of software and hardware. Whether it is inference or training, the company has a very good understanding of the entire underlying platform, because whether it is the AI framework or the Other open source frameworks are essentially part of the operating system. Whether it is data parallelism or model parallelism, the company believes that they are all middleware. These obvious advantages allow us to ultimately optimize model performance and efficiency, optimize model scale, run it on the edge, and empower thousands of industries.
RUBIK Studio: Based on the company’s deep understanding of the operating system, the company differentiates and models each segment of the operating system, accumulating thousands of successfully released packages in the past, hundreds of billions of lines of code accumulation, and long-term development experience Accumulate, turn Android knowledge into a huge knowledge base, and exert huge value. Rubik Studio will be a big tool and environment that changes the entire future. Users can directly and quickly develop PC applications, mobile applications, and websites with relatively closed functions through Rubik Studio.
1.4 Apple: ChatGPT is first connected to IOS, WWDC 2023 is worth looking forward to
On May 18, 2023, OpenAI released the IOS version of the ChatGPT APP (at the same time, it was officially announced that the Android version of the APP is under development). According to Apple’s official website, the application only supports English, has an age rating of “12 years and above”, and supports network synchronization. Chat records, whisper voice input and other functions. The IOS version of the ChatGPT download application is free, but a "ChatGPT Plus" paid item priced at US$19.99 within the APP will be provided to enable the use of GPT4 large models with stronger performance.
Rewritten sentence: ChatGPT APP may become the prototype of mobile phone super APP, and reconstruct the artificial intelligence application ecosystem based on this. According to Jiupai Finance, the IOS version of ChatGPT APP already has the ability to flexibly solve various daily problems. ChatGPT APP can provide instant answers, and users can obtain accurate information without filtering advertisements or multiple results; it provides customized suggestions, and users Turn to them for guidance on cooking, travel planning, or crafting a thoughtful message; provide creative inspiration, generate gift ideas for users, outline a presentation, or write beautiful poetry. In addition, ChatGPT APP can also help users improve work efficiency through professional information, such as idea feedback, note summaries, and technical topics, and provide users with learning opportunities to help them explore new languages, modern history, etc. at their own pace. We believe that under the premise of the successful implementation of the AI application ecosystem of ChatGPT Plugin, ChatGPT APP is expected to become the core of the mobile AI application ecosystem. Through the training of massive interactive data from mobile phone users, in the future users are expected to use ChatGPT APP to call other applications to complete various rigid tasks. , ChatGPT APP will develop into an unprecedented super APP.
The slogan of Apple WWDC 2023 is "Code new worlds". As an important entrance to human-computer interaction, Siri is expected to become an important starting point for Apple to enter AI. According to IT House, Xinzhiyuan, and Tencent Cloud, Apple will begin developing applications that allow users to use Siri to create programs for MR in February 2023 at the latest. The technology behind this application construction method comes from Apple’s acquisition of MR in 2017. Fabric Software: Users can use Siri to build AR apps, asking the AI assistant to help build apps that allow virtual animals to move around a room, on top of and around real objects, without having to design the animal from scratch, program its animations, and calculate its movements. Applications for movement in 3D space with obstacles, which include “scanning and importing real-world objects into the headset so that they can be accurately represented in 3D and behave as if they existed in real life.”
1.5 Huawei: IEF Gaussian database fully covers edge AI scenarios
Huawei Intelligent Edge Platform IEF is an edge-cloud collaborative operating system built on cloud-native technology. It can run on a large number of heterogeneous edge devices and integrate rich AI, data analysis, middleware and other applications from The cloud is deployed to the edge to meet users' business demands for edge-cloud collaboration of intelligent applications. IEF can extend Huawei Cloud's AI/big data capabilities to the edge, support intelligent video analysis, text recognition, image recognition, big data stream processing and other capabilities, and provide real-time intelligent edge services nearby; it supports two operating modes: container and function. It meets users' demands for lightweight application management; it natively supports kubernetes and docker ecology, and applications can be quickly started and upgraded quickly; it supports function engines such as Python and NodeJS to quickly respond to edge events; in addition, it has excellent compatibility, security and reliability.
Huawei Gauss Database is an enterprise-level distributed relational database launched by Huawei based on the openGaussDB self-research ecosystem. It has enterprise-level complex transaction mixed load capabilities, supports strong consistency of distributed transactions, can be deployed across AZs in the same city, has zero data loss, and supports 1000 Expansion capabilities, PB-level mass storage. At the same time, it has key capabilities such as high availability, high reliability, high security, elastic scaling, one-click deployment, fast backup and recovery, and monitoring and alarming on the cloud. It can provide enterprises with an enterprise-level database with comprehensive functions, stability, reliability, strong scalability, and superior performance. Serve. On June 7, 2023, Huawei Goss will hold a conference on database acceleration of core financial business upgrades.
Investment advice: The general trend of AI from cloud to edge has been established, and our early judgment has been continuously verified: edge AI is an industry trend, and embodied intelligence is the internal logic (AI self-improvement requires the interaction data between humans and the environment to be concentrated on the terminal) , robots are the ultimate application. We released a report on May 13, 2023, "China Science and Technology Thunder: Large Models from Cloud to Edge, Terminal Interaction Revolution Breeds Historical Opportunities", which clearly stated that Google will vigorously enter the terminal large model market, and terminal AI will become the next must-have for military strategists. land; then released "Google's "Empire Strikes Back": The Inflection Point of AI from Cloud to Edge" on May 14th, detailing Google's AI blueprint from cloud to edge, and clearly stated that AI edge deployment has entered reality; through the release AI experience reports on the IOS version of ChatGPT, Windows AI and other AI experience reports, we released "ChatGPT APP marks a new stage of AI market" on May 21, clearly proposing that the large model is the ultimate operating system in the AI era. ChatGPT super APP is only the first step, generative AI's progress from cloud to end is still accelerating. This week, companies such as Nvidia, Qualcomm, and Thunderstar have released related products to speed up the implementation of edge AI. Next week, Huawei and Apple’s press conferences are expected to focus on terminal AI. The general trend of AI from cloud to edge has become extremely clear. , it is recommended to pay attention to leading companies such as Chuangda, iFlytek, and EZVIZ Network.
Source: Selected Brokerage Research Reports
The above is the detailed content of Minsheng Securities: Edge AI is an industry trend. It is recommended to pay attention to these targets.. For more information, please follow other related articles on the PHP Chinese website!

PHP错误处理:最佳实践和建议在编写PHP代码时,错误处理是一项非常重要的任务。如果不正确地处理错误,就会导致应用程序中的漏洞和安全问题。同时,良好的错误处理还有助于提高代码的可维护性和可扩展性。本文将介绍一些PHP错误处理的最佳实践和建议,并提供一些代码示例。使用异常处理在PHP中,异常是一种用于处理运行时错误的机制。通过使用异常,可以将错误

C++语言作为一种通用的高级编程语言,被广泛用于开发各种应用程序和系统。然而,C++的复杂性和灵活性也使得开发人员面临着一些挑战,特别是在大型项目中。在处理大型项目时,模块化的开发方法是至关重要的。本文将介绍如何进行模块化的C++开发,并提供一些建议和最佳实践。模块化开发是指将一个大型项目划分为多个小模块,每个模块有自己的功能和职责,通过模块间的接口进行通信

距离Win11系统发布已经过去了一年的时间,许多人一直想知道2022年是否建议升级至Win11。实际上,如果我们目前使用的系统感觉良好并且没有遇到任何问题,那么升级并不是必须的。答:2022年不建议升级win11,因为现在win11相比win11并没有多少提升。如果我们喜欢Win11的全新界面和设置,不妨下载来试试看看。1、现在win11和win10的软件兼容性没啥区别,能在win11用的win10也能用。2、如果我们习惯了win10操作的话,win11可能还用着不习惯,很多功能找不到。3、例如

深入理解JavaGUI开发的经验与建议作为一种常用的面向对象编程语言,Java在软件开发中扮演着举足轻重的角色。而在Java开发中,GUI(GraphicalUserInterface)即图形用户界面的开发更是日常工作中所需要掌握的重要技能之一。在GUI开发中,丰富的用户界面和交互性能会直接影响到软件的用户体验和用户满意度,因此,深入理解

C++是一种高效、可靠、可扩展性强的编程语言,适合用于开发各种类型的软件。在开发C++应用程序时,测试驱动开发(TDD)是一种可行的方法,可以确保代码的质量并提高开发效率。本文将介绍如何进行C++测试驱动开发。理解测试驱动开发测试驱动开发是一种软件开发方法,其核心理念是在实现功能之前编写测试用例。在开发的过程中,开发人员编写单元测试用例,并使用自动化测试工具

CentOS搭建Web服务器的数据库选择与配置建议概述:在搭建Web服务器时,数据库的选择与配置是非常重要的一环。本文将介绍在CentOS系统上搭建Web服务器时,如何选择合适的数据库,并给出相应的配置建议。同时,还会提供一些代码示例,以帮助读者更好地理解和操作。选择数据库:在选择数据库时,应根据自身需求和项目特点来决定。常见的数据库有MySQL、Postg

PHP(HypertextPreprocessor)是一种广泛应用于web开发的脚本语言,它具有简单易学、灵活高效等特点,因此受到众多开发者的青睐。在PHP开发中,如何实现各种功能是开发者需要面对的重要问题。本文将分享一些PHP开发中的功能实现秘籍与建议,帮助读者更好地进行开发。一、用户管理功能实现用户注册与登录功能实现用户注册与登录功能是PHP开发的基础

近日,英伟达宣布将在2023年上海国际嵌入式展中,展示适用于自主机器和很多其他嵌入式应用的英伟达Jetson边缘计算平台,并带来生态合作伙伴基于相关软硬件在交通、工业、机器人等多个垂直行业领域所构建的解决方案。民生证券在研报中指出,AIGC推动应用端革命,人工智能则赋能千行百业,大模型部署在边缘侧与移动端成为未来趋势,边缘AI正成为科技巨头们抢占的下一块高地。2030年全球潜在市场规模4450亿美元据了解,边缘计算是一种分散式运算的架构,将数据资料、应用程序、服务的运算,由网络中心节点,移至网络


Hot AI Tools

Undresser.AI Undress
AI-powered app for creating realistic nude photos

AI Clothes Remover
Online AI tool for removing clothes from photos.

Undress AI Tool
Undress images for free

Clothoff.io
AI clothes remover

AI Hentai Generator
Generate AI Hentai for free.

Hot Article

Hot Tools

Dreamweaver Mac version
Visual web development tools

SublimeText3 Chinese version
Chinese version, very easy to use

SAP NetWeaver Server Adapter for Eclipse
Integrate Eclipse with SAP NetWeaver application server.

Safe Exam Browser
Safe Exam Browser is a secure browser environment for taking online exams securely. This software turns any computer into a secure workstation. It controls access to any utility and prevents students from using unauthorized resources.

VSCode Windows 64-bit Download
A free and powerful IDE editor launched by Microsoft
