search
HomeTechnology peripheralsAILeCun predicts AGI: Large models and reinforcement learning are both rampant! My 'world model' is the new way

Yann LeCun, one of the most famous contemporary giants in the AI ​​industry and the soul of Meta's AI laboratory, has long been committed to giving machines a basic understanding of the world's operating concepts, that is, allowing AI to acquire common sense. What LeCun did in the past was to use video excerpts to train neural networks and let the AI ​​predict, pixel by pixel, what will appear in the next frame of daily activity videos. Unsurprisingly, he admitted that this approach hit a brick wall. After thinking about it for several months to a year and a half, LeCun had new ideas for the next generation of AI.

LeCun predicts AGI: Large models and reinforcement learning are both rampant! My world model is the new way

New Path of AI

In an interview with "MIT Technology Review", LeCun outlined his new research path, saying This will give the machine a common sense basis for exploring the world. For LeCun, this is the first step in building AGI (Artificial General Intelligence). Machines that can think like humans have been the guiding vision since the birth of the AI ​​industry, and it is also one of the most controversial concepts.

But LeCun’s new path may still be incomplete, raising more questions than answers. The biggest question is that LeCun himself admits that he doesn’t yet know how to build the kind of AI he describes. At the core of this approach is a neural network that can look at and learn from the real world in a different way than before. LeCun finally gave up on letting AI guess the next frame of video pixel by pixel, and only let the new neural network learn the key knowledge necessary to complete the task.

LeCun predicts AGI: Large models and reinforcement learning are both rampant! My world model is the new way

LeCun then plans to pair this neural network with another neural network called a "configurator." The "configurator" is responsible for deciding which details the main neural network must learn and automatically adjust the main system accordingly. For LeCun, AGI is an integral part of human interaction with future technologies. Of course, this outlook coincides with his employer Meta Company, which has invested all his wealth in developing the metaverse.

LeCun said that in 10-15 years, AR glasses will replace the current status of smartphones. AR glasses must have a virtual intelligent assistant that can assist human daily activities. If these assistants are to be most effective, they must more or less keep up with the intelligence of the human brain.

"World model" is the core of AGI

LeCun's recent passion for "world model", according to him, is the basic operating mode of most animal brains: for the real world Run a simulation. From infancy, animals use prediction-trial-and-error methods to develop intelligence. Young children develop the foundations of intelligence in the first few months of life by observing real-world movements and setbacks.

Observing a small ball falling hundreds of times, ordinary babies have a basic understanding of the existence and operation of gravity even if they have never taken a basic physics class or learned Newton's three laws. Therefore, this kind of intuitive/tacit reasoning is called "common sense" by ordinary people. Human beings use common sense to understand most possible futures and impossible fantasies in the real world, to foresee the consequences of their actions and make decisions accordingly. Such human intelligence requires neither pixel-accurate details nor a comprehensive library of physical parameters. Even if someone has no vision or is illiterate, they can still use their intelligence normally.

LeCun predicts AGI: Large models and reinforcement learning are both rampant! My world model is the new way

But it is difficult to teach a machine to learn common sense. Today's neural networks need to be shown thousands of examples before they begin to vaguely discover underlying patterns. LeCun said that the basis of intelligence is the common sense ability to predict the immediate future. However, after giving up on letting AI predict pixel by pixel, LeCun said he wanted to change his mind. LeCun gave an analogy: Imagine you hold a pen in the air and let it go. Common sense tells you that the pen will definitely fall, but the precise location of the fall is beyond the scope of human intelligence prediction. According to the past AI development model, AI has to run complex physics models to predict whether the pen will fall and to obtain the precise location of the fall.

Now LeCun is trying hard to let AI only predict the common sense conclusion that the pen will fall. As for the precise position, it is not within the scope of solution. LeCun said this is the basic pattern of the "world model".

LeCun predicts AGI: Large models and reinforcement learning are both rampant! My world model is the new way

LeCun said that he has built an early version of the "world model" that can complete basic object recognition, and is now working on training it to learn the above-mentioned common sense predictions.

However, LeCun said that he has not yet understood the function of the "configurator". The "configurator" AI in LeCun's imagination is the control component of the entire AGI system. It will determine what common-sense predictions the World Model needs to make at any moment, and adjust the details of the data the World Model should handle to do so. LeCun now firmly believes that a "configurator" is essential, but he doesn't know how to train a neural network to achieve this effect.

"We need to explore a list of feasible technologies, and this list does not exist yet." In LeCun's vision, "configurator" and "world model" are the future AGI The two core parts of the basic cognitive architecture are based on which the cognitive model for perceiving the world, the incentive model that drives AI to adjust behavior, etc. can be developed. LeCun said that this way the neural network can successfully simulate every part of the human brain. For example, the "configurator" and "world model" play the role of the prefrontal lobe, the motivation model is the amygdala of AI, and so on. LeCun predicts AGI: Large models and reinforcement learning are both rampant! My world model is the new way

Cognitive architecture and prediction models at different levels of detail are all views that have been established in the industry for many years. However, when deep learning becomes the mainstream of the AI ​​industry, many of these old ideas become outdated. Now LeCun is returning to traditional wisdom: "The AI ​​research community has forgotten these things a lot."

Large models and reinforcement learning are dead ends

The reason why we go back to the old ways road because LeCun firmly believes that the current mainstream path in the industry has reached a dead end. Regarding how to build AGI, there are currently two mainstream views in the AI ​​industry.

First, many researchers firmly believe in the path that leads to their own mistakes: just like OpenAI’s GPT series and DALL-E series, the bigger the model, the better, so large that it exceeds the critical point. , AI has awakened human intelligence.

The second is reinforcement learning: continuous trial and error, and rewarding and punishing the AI ​​according to the trial and error results. This is DeepMind’s method for making various chess and card AI and game AI. Believers of this path believe that as long as the reward incentives are set correctly, reinforcement learning will eventually create a real AGI.

Lecun said that the two types of people here are rubbish: "Infinitely expanding the magnitude of existing large language models, and finally being able to create human-level AI? This absurd argument, I didn't believe it for a second. These models can only process various text and image data, without any direct experience of the real world." "Reinforcement learning requires a huge amount of data to train the model to perform the simplest tasks. I I don’t think this method has a chance of making AGI.”

LeCun predicts AGI: Large models and reinforcement learning are both rampant! My world model is the new way

People in the industry have both support and opposition to LeCun’s views. If LeCun's vision is realized, AI will become the next generation of basic high-performance technology no less than the Internet. But his announcement did not include the performance, incentive mechanism, control mechanism, etc. of his own model. However, these shortcomings are minor matters, because regardless of praise or criticism, industry insiders agree that facing these shortcomings will be a long time coming. Because even LeCun can't make AGI right now.

Lecun himself also acknowledged this situation. He said that he only hoped to sow seeds for new theoretical paths and let latecomers build results on this basis. "Achieving this goal requires too many people and too much effort. I am bringing this up now just because I think this path is the final right path." Even if this is not possible, LeCun hopes to persuade his colleagues not to just focus on it. With large models and reinforcement learning, it’s best to open your mind. "I hate to see people wasting time."

Industry reaction: both praise and criticism

Yoshua Bengio, another leader in the AI ​​industry and a good friend of LeCun, expressed his joy Seeing old friends come true. "Yann has been talking about this for a while, but I'm quite happy to see him comprehensively summarizing all his remarks in one place. However, these are just applications for research directions rather than report of results. We usually only discuss them privately. Sharing this below, the risk of talking publicly is quite high."

LeCun predicts AGI: Large models and reinforcement learning are both rampant! My world model is the new way

## David Silver, who leads the development of the game AI AlphaZero at DeepMind, disapproves of LeCun's comments on his project Criticism, but welcome to his vision.

"The world model described by LeCun is indeed an exciting new idea." Melanie Mitchell of the Santa Fe Institute in California agreed with LeCun: "The industry really doesn't see this kind of thing in the deep learning community very often. point of view. But the big language model really lacks both memory and the backbone of the internal world model that can play a role."

Natasha Jaques of Google Brain disagrees: "Everyone has seen that big language The model is extremely efficient and incorporates a lot of human knowledge. Without a language model, how can I upgrade the world model proposed by LeCun? Even if humans learn, the way is not only personal experience, but also word of mouth."

The above is the detailed content of LeCun predicts AGI: Large models and reinforcement learning are both rampant! My 'world model' is the new way. For more information, please follow other related articles on the PHP Chinese website!

Statement
This article is reproduced at:51CTO.COM. If there is any infringement, please contact admin@php.cn delete
从VAE到扩散模型:一文解读以文生图新范式从VAE到扩散模型:一文解读以文生图新范式Apr 08, 2023 pm 08:41 PM

1 前言在发布DALL·E的15个月后,OpenAI在今年春天带了续作DALL·E 2,以其更加惊艳的效果和丰富的可玩性迅速占领了各大AI社区的头条。近年来,随着生成对抗网络(GAN)、变分自编码器(VAE)、扩散模型(Diffusion models)的出现,深度学习已向世人展现其强大的图像生成能力;加上GPT-3、BERT等NLP模型的成功,人类正逐步打破文本和图像的信息界限。在DALL·E 2中,只需输入简单的文本(prompt),它就可以生成多张1024*1024的高清图像。这些图像甚至

找不到中文语音预训练模型?中文版 Wav2vec 2.0和HuBERT来了找不到中文语音预训练模型?中文版 Wav2vec 2.0和HuBERT来了Apr 08, 2023 pm 06:21 PM

Wav2vec 2.0 [1],HuBERT [2] 和 WavLM [3] 等语音预训练模型,通过在多达上万小时的无标注语音数据(如 Libri-light )上的自监督学习,显著提升了自动语音识别(Automatic Speech Recognition, ASR),语音合成(Text-to-speech, TTS)和语音转换(Voice Conversation,VC)等语音下游任务的性能。然而这些模型都没有公开的中文版本,不便于应用在中文语音研究场景。 WenetSpeech [4] 是

普林斯顿陈丹琦:如何让「大模型」变小普林斯顿陈丹琦:如何让「大模型」变小Apr 08, 2023 pm 04:01 PM

“Making large models smaller”这是很多语言模型研究人员的学术追求,针对大模型昂贵的环境和训练成本,陈丹琦在智源大会青源学术年会上做了题为“Making large models smaller”的特邀报告。报告中重点提及了基于记忆增强的TRIME算法和基于粗细粒度联合剪枝和逐层蒸馏的CofiPruning算法。前者能够在不改变模型结构的基础上兼顾语言模型困惑度和检索速度方面的优势;而后者可以在保证下游任务准确度的同时实现更快的处理速度,具有更小的模型结构。陈丹琦 普

解锁CNN和Transformer正确结合方法,字节跳动提出有效的下一代视觉Transformer解锁CNN和Transformer正确结合方法,字节跳动提出有效的下一代视觉TransformerApr 09, 2023 pm 02:01 PM

由于复杂的注意力机制和模型设计,大多数现有的视觉 Transformer(ViT)在现实的工业部署场景中不能像卷积神经网络(CNN)那样高效地执行。这就带来了一个问题:视觉神经网络能否像 CNN 一样快速推断并像 ViT 一样强大?近期一些工作试图设计 CNN-Transformer 混合架构来解决这个问题,但这些工作的整体性能远不能令人满意。基于此,来自字节跳动的研究者提出了一种能在现实工业场景中有效部署的下一代视觉 Transformer——Next-ViT。从延迟 / 准确性权衡的角度看,

Stable Diffusion XL 现已推出—有什么新功能,你知道吗?Stable Diffusion XL 现已推出—有什么新功能,你知道吗?Apr 07, 2023 pm 11:21 PM

3月27号,Stability AI的创始人兼首席执行官Emad Mostaque在一条推文中宣布,Stable Diffusion XL 现已可用于公开测试。以下是一些事项:“XL”不是这个新的AI模型的官方名称。一旦发布稳定性AI公司的官方公告,名称将会更改。与先前版本相比,图像质量有所提高与先前版本相比,图像生成速度大大加快。示例图像让我们看看新旧AI模型在结果上的差异。Prompt: Luxury sports car with aerodynamic curves, shot in a

五年后AI所需算力超100万倍!十二家机构联合发表88页长文:「智能计算」是解药五年后AI所需算力超100万倍!十二家机构联合发表88页长文:「智能计算」是解药Apr 09, 2023 pm 07:01 PM

人工智能就是一个「拼财力」的行业,如果没有高性能计算设备,别说开发基础模型,就连微调模型都做不到。但如果只靠拼硬件,单靠当前计算性能的发展速度,迟早有一天无法满足日益膨胀的需求,所以还需要配套的软件来协调统筹计算能力,这时候就需要用到「智能计算」技术。最近,来自之江实验室、中国工程院、国防科技大学、浙江大学等多达十二个国内外研究机构共同发表了一篇论文,首次对智能计算领域进行了全面的调研,涵盖了理论基础、智能与计算的技术融合、重要应用、挑战和未来前景。论文链接:​https://spj.scien

​什么是Transformer机器学习模型?​什么是Transformer机器学习模型?Apr 08, 2023 pm 06:31 PM

译者 | 李睿审校 | 孙淑娟​近年来, Transformer 机器学习模型已经成为深度学习和深度神经网络技术进步的主要亮点之一。它主要用于自然语言处理中的高级应用。谷歌正在使用它来增强其搜索引擎结果。OpenAI 使用 Transformer 创建了著名的 GPT-2和 GPT-3模型。自从2017年首次亮相以来,Transformer 架构不断发展并扩展到多种不同的变体,从语言任务扩展到其他领域。它们已被用于时间序列预测。它们是 DeepMind 的蛋白质结构预测模型 AlphaFold

AI模型告诉你,为啥巴西最可能在今年夺冠!曾精准预测前两届冠军AI模型告诉你,为啥巴西最可能在今年夺冠!曾精准预测前两届冠军Apr 09, 2023 pm 01:51 PM

说起2010年南非世界杯的最大网红,一定非「章鱼保罗」莫属!这只位于德国海洋生物中心的神奇章鱼,不仅成功预测了德国队全部七场比赛的结果,还顺利地选出了最终的总冠军西班牙队。不幸的是,保罗已经永远地离开了我们,但它的「遗产」却在人们预测足球比赛结果的尝试中持续存在。在艾伦图灵研究所(The Alan Turing Institute),随着2022年卡塔尔世界杯的持续进行,三位研究员Nick Barlow、Jack Roberts和Ryan Chan决定用一种AI算法预测今年的冠军归属。预测模型图

See all articles

Hot AI Tools

Undresser.AI Undress

Undresser.AI Undress

AI-powered app for creating realistic nude photos

AI Clothes Remover

AI Clothes Remover

Online AI tool for removing clothes from photos.

Undress AI Tool

Undress AI Tool

Undress images for free

Clothoff.io

Clothoff.io

AI clothes remover

AI Hentai Generator

AI Hentai Generator

Generate AI Hentai for free.

Hot Article

Repo: How To Revive Teammates
1 months agoBy尊渡假赌尊渡假赌尊渡假赌
R.E.P.O. Energy Crystals Explained and What They Do (Yellow Crystal)
2 weeks agoBy尊渡假赌尊渡假赌尊渡假赌
Hello Kitty Island Adventure: How To Get Giant Seeds
1 months agoBy尊渡假赌尊渡假赌尊渡假赌

Hot Tools

Dreamweaver Mac version

Dreamweaver Mac version

Visual web development tools

MantisBT

MantisBT

Mantis is an easy-to-deploy web-based defect tracking tool designed to aid in product defect tracking. It requires PHP, MySQL and a web server. Check out our demo and hosting services.

Notepad++7.3.1

Notepad++7.3.1

Easy-to-use and free code editor

SAP NetWeaver Server Adapter for Eclipse

SAP NetWeaver Server Adapter for Eclipse

Integrate Eclipse with SAP NetWeaver application server.

SublimeText3 Mac version

SublimeText3 Mac version

God-level code editing software (SublimeText3)