


In the past few days, the grand conference of AI - ICLR was held in Vienna.
OpenAI, Meta, Google, Zhipu AI and other world-leading AI technology companies gathered together.
The venue is crowded with celebrities and the stars are dazzling. If you walk a few steps, you can bump into a celebrity who has published a subversive paper.
Unsurprisingly, the ICLR 2024 exhibition hall has also become a star-chasing scene. The lively atmosphere almost blew the roof off.
On-site star chasing Turing giants
LeCun, the famous "e man" among the three Turing giants, was on X in advance He generously announced his schedule and was looking forward to meeting his fans.
In the comment area, not only were fans excited to check in, but some were even ready to submit their resumes on the spot.
The fans' trip was indeed worthwhile. At the scene, LeCun explained eloquently, and the enthusiastic audience formed a dense circle around him.
Getting back to the topic, throughout the ICLR event, the Meta team will share more than 25 papers and two seminars. This time, the LeCun team published the following two papers on ICLR.
Paper address: https://arxiv.org/abs/2305.19523
Paper address: https://arxiv.org/abs/2311.12983
Another Turing giant, Yoshua Bengio, also showed his high popularity.
The audience concluded: "A person really needs to be unique in his field to have such a long queue outside his conference room!"
LeCun and Hinton have both expressed strong opinions on this before. Bengio's attitude seems to have been relatively vague. I can't wait to know what he thinks of AGI. On the coming May 11, he will give a speech at a Workshop on AGI.
It is worth mentioning that the Bengio team also received an honorable mention for outstanding paper at this year’s ICLR.
##Paper address: https://openreview.net/pdf?id=Ouj6p4ca60
Next door to Google Meta, Zhipu AI is also present at, where Google’s open source model Gema, the framework behind robotic agents Robotics Transformers, and other groundbreaking research are presented.
Next to Meta and Google, there is a very eye-catching company in the middle of the exhibition hall - Zhipu AI.
The children’s shoes on site are introducing a series of research results such as GLM-4 and ChatGLM.
This series of displays attracted the attention of many foreign scholars.
The nearly two thousand guests and scholars at the scene listened carefully to the introduction of the GLM large model technical team.
The introduction includes a number of cutting-edge research results on the GLM series of large models, covering fields such as mathematics, Vincentian diagrams, image understanding, visual UI understanding, and Agent intelligence.
At the scene, everyone had a heated discussion about their views on Scaling Law. The GLM team also has unique insights into this -
"Compared to model size or training calculation amount, intelligent emergence and pre-training loss are more closely related."
For example, the famous OpenAI 996 researcher Jason Wei expressed his admiration after carefully reading the Zhipu AI paper on pre-training loss.
In the paper, the team evaluated its performance on 12 Chinese and English data sets by training 30 LLMs with different parameters and data sizes.
Paper address: https://arxiv.org/abs/2403.15796
Results observed , LLM will have emergent ability only when the pre-training loss is lower than a certain threshold.
Moreover, defining "emergent ability" from the perspective of pre-training loss is better than relying solely on model parameters or training volume.
The performance of Zhipu AI has also made more and more foreign netizens realize——
Tanishq, research director of Stability AI who received a PhD at the age of 19, said that the most competitive open source basic models such as CogVLM, which have made significant contributions to the open source ecosystem, come from China.
The former CEO of the game studio started using CogVLM and Stable Diffusion to make a complete open source version last year.
Yes, since CogVLM was released, its powerful capabilities have caused foreign netizens to exclaim.
In the LLM rankings in January this year, someone also discovered——
At that time, Gemini and GPT-4V were far ahead of any open source LLM, with the only exception being CogVLM.
It can be seen that with this wave of large-scale domestic models going overseas, Zhipu AI has quietly established its own huge influence abroad.
Special Speeches
In addition to the wonderful demonstrations in the exhibition hall, this year's ICLR invited a total of seven special speakers to share their insights on AI.
There are Raia Hadsell, a research scientist from Google DeepMind, Devi Parik, associate professor at Georgia Institute of Technology & Chief Scientist of FAIR, and director from the Max Planck Institute for Computer Science (MPI-SWS) Moritz Hardt, the only Chinese team is the GLM large model technical team of Zhipu AI.
Raia Hadsell
The title of Google DeepMind scientist Raia Hadsell's speech is - "Learning during the ups and downs of artificial intelligence development: Unexpected truths on the road to AGI ”.
After decades of steady development and occasional setbacks, AI is at a critical inflection point.
AI products have exploded into the mainstream market, and we have not yet reached the ceiling of scaling dividends, so the entire community is exploring the next step.
In this speech, based on more than 20 years of experience in the field of AI, Raia discussed our assumptions about the development path of AGI, how Change over time.
At the same time, she also revealed the unexpected discoveries we made during this exploration.
From reinforcement learning to distributed architecture to neural networks, they are already playing a potentially revolutionary role in the scientific field.
Raia believes that by learning from past experiences and lessons, important insights can be provided for the future research direction of AI.
Devi Parikh
#On the other side, FAIR chief scientist Devi Parik told everyone the story of her life.
As can be seen from the title of the speech, the content shared by Parik is extraordinary.
At the ICLR conference, when explaining why the technical environment is what it is now, everyone will focus on the development of the Internet, big data and computing power.
However, few people pay attention to those small, but important personal stories.
In fact, everyone’s story can be gathered into an important force to promote technological progress.
In this way, we can learn from each other and inspire each other. This makes us more tenacious and efficient in pursuing our goals.
Moritz Hardt
Moritz Hardt, Director of the German MPI-SWS, brought "Emerging Scientific Benchmarks" ” speech.
Obviously, benchmark testing has become the "core pillar" in the field of machine learning.
Since the 1980s, although humans have made many achievements under this research paradigm, their deep understanding is still limited.
#In this talk, Hardt explores the fundamentals of benchmarking as an emerging science through a series of selected empirical studies and theoretical analyses. principle.
He specifically discussed the impact of annotation errors on data quality, external validation of model rankings, and the prospects for multi-task benchmarking.
At the same time, Hard also presented a number of case studies.
These challenge our conventional wisdom and highlight the importance and benefits of developing scientific benchmarks.
GLM Team
In China, the GLM large model technical team of Zhipu AI has also brought "ChatGLM to AGI" "Road" wonderful speech.
It is worth mentioning that this is also the "first time" in China that a keynote speech related to large models has been presented at a top international conference.
This speech will first introduce the development process of AI in the past few decades from a Chinese perspective.
At the same time, they used ChatGLM as an example to explain the understanding and insights they gained during practice.
##2024 AGI Preview: GLM 4.5, GLM-OS, GLM-zero
At ICLR, the GLM large model team introduced the three major technical trends of GLM for AGI.
Where is the only way to AGI?
The industry has mixed opinions on this. Some people think it is an intelligent agent, some people think it is multi-modal, and some people say that Scaling Law is a necessary but not sufficient condition for AGI.
But LeCun insists that LLM is a wrong road to AGI, and LLM cannot bring AGI.
In this regard, the team also put forward its own unique point of view.
First of all, they talked about the subsequent upgraded version of GLM-4, namely GLM-4.5 and its upgraded model.
The subsequent upgraded version of GLM-4 will be based on SuperIntelligence and SuperAlignment technologies, while making great progress in the field of native multi-modality and AI security. .
The GLM large model team believes that text is the most critical foundation on the road to AGI.
The next step is to mix text, images, video, audio and other modalities together for training to become a true "native multi-modal model".
At the same time, in order to solve more complex problems, they also introduced the concept of GLM-OS, a general computing system centered on large models.
This view coincides with the view of large-model operating systems previously proposed by Karpathy.
At the ICLR site, the GLM large model team introduced the implementation of GLM-OS in detail:
Based on the existing All-Tools capabilities, coupled with memory and self-reflection capabilities, GLM-OS is expected to successfully imitate the human PDCA mechanism, namely Plan-Do-Check-Act cycle.
Specifically, make a plan first, then give it a try to form feedback, adjust the plan, and then take action in order to achieve better results.
Relying on the PDCA cycle mechanism, LLM can self-feedback and evolve independently - just like humans do.
In addition, the GLM large model team also revealed that since 2019, the team has been studying a technology called GLM-zero, aiming to study human "unconscious" learning mechanisms.
"When people are sleeping, the brain is still learning unconsciously."
The GLM large model team said that "unconscious" learning Mechanisms are an important part of human cognitive abilities, including self-learning, self-reflection, and self-criticism.
There are two systems in the human brain, "feedback" and "decision-making", which respectively correspond to the LLM large model and memory.
Therefore, related research on GLM-zero will further expand human understanding of consciousness, knowledge, and learning behavior.
Although it is still in a very early research stage, GLM-zero can be regarded as the only way to AGI.
This is also the first time that the GLM large model team has disclosed this technology trend to the outside world.
Domestic top technical team
At the end of 2020, the GLM large model technical team developed the GLM pre-training architecture.
In 2021, the tens of billions parameter model GLM-10B was trained, and in the same year, the converged trillions sparse model was successfully trained using the MoE architecture.
In 2022, they also collaborated to develop the Chinese-English bilingual 100-billion-level ultra-large-scale pre-training model GLM-130B and open sourced it.
In the past year, the team has completed an upgrade of the large base model almost every 3-4 months, and it has now been updated to the GLM-4 version.
Not only that, as the first domestic LLM company to enter the market, Zhipu AI has set an ambitious goal in 2023 - to benchmark OpenAI across the board.
The GLM large model technical team has built a complete large model product matrix based on the AGI vision.
In addition to the GLM series, there are also CogView grammatical model, CodeGeeX code model, multi-modal understanding model CogVLM, and then GLM-4V multi-modal large model and All-Tools Functions and AI assistant to clear words.
At the same time, the researchers of the GLM large model technology team have a very high influence in the industry.
For example, Li Feifei, who is very popular in the circle, teaches the CS25 course at Stanford University. Every time, she invites experts at the forefront of Transformer research to share her latest breakthroughs.
It has been confirmed that among the guests of the CS25 course, there are researchers from Zhipu AI.
CogVLM
The open source visual language model CogVLM developed by the team, once The release attracted industry attention.
A paper published by Stability AI in March showed that CogVLM was directly used by Stable Diffufion 3 for image annotation due to its excellent performance.
Paper address: https://arxiv.org/abs/2403.03206
CogAgent
On this basis, CogAgent, an open source visual language model improved based on CogVLM, is mainly aimed at the user graphical interface GUI. understand.
The relevant papers of CogAgent have been included in CVPR 2024, the highest-level academic conference in the field of international computer vision.
You must know that CVPR is known for its strict admissions. This year's thesis acceptance rate is only about 2.8%.
Paper address: https://arxiv.org/abs/2312.08914
ChatGLM-Math
#To solve mathematical problems with LLM, the GLM large model team proposed the "Self-Critique" iterative training method.
Through the self-feedback mechanism, it helps LLM improve both language and mathematics abilities.
Paper address: https://arxiv.org/abs/2404.02893
This method , including two key steps:
First train a "Math-Critique" model generated from LLM itself to evaluate the model to generate answers to mathematical questions and provide feedback signals.
Secondly, through rejection sampling fine-tuning and DPO, the new model is used to supervise the generation of LLM itself.
The GLM large model team also designed the MATHUSEREVAL benchmark test set to evaluate the mathematical capabilities of the new model. The results are as follows:
It is obvious that the new method significantly improves LLM’s mathematical problem-solving ability while still improving its language skills. Importantly, it outperforms larger models with twice the number of parameters in some cases.
GLM-4 ranks among the first echelons in the world
In the OpenCompass 2.0 benchmark test, the strength of Zhipu AI’s new generation base large model Not to be underestimated.
In the overall ranking, GLM-4 ranks third and ranks first in the country.
In the "SuperBench Large Model Comprehensive Capability Evaluation Report" released by the SuperBench team not long ago, GLM-4 also ranked among the first tier in the world.
Especially in the most critical semantic understanding and agent capabilities, GLM-4 ranks first in the country, overwhelming all competitors.
In the first year of big models that has just passed, a lively model war has been going on for a year.
If 2024 is to be the first year of AGI, the world’s largest model teams still have a long way to go.
The above is the detailed content of The Turing giant appeared at ICLR and went crazy for stars LeCun and Bengio at the summit! Three major technology trends of Chinese teams set off new imagination of AGI. For more information, please follow other related articles on the PHP Chinese website!

ai合并图层的快捷键是“Ctrl+Shift+E”,它的作用是把目前所有处在显示状态的图层合并,在隐藏状态的图层则不作变动。也可以选中要合并的图层,在菜单栏中依次点击“窗口”-“路径查找器”,点击“合并”按钮。

ai橡皮擦擦不掉东西是因为AI是矢量图软件,用橡皮擦不能擦位图的,其解决办法就是用蒙板工具以及钢笔勾好路径再建立蒙板即可实现擦掉东西。

虽然谷歌早在2020年,就在自家的数据中心上部署了当时最强的AI芯片——TPU v4。但直到今年的4月4日,谷歌才首次公布了这台AI超算的技术细节。论文地址:https://arxiv.org/abs/2304.01433相比于TPU v3,TPU v4的性能要高出2.1倍,而在整合4096个芯片之后,超算的性能更是提升了10倍。另外,谷歌还声称,自家芯片要比英伟达A100更快、更节能。与A100对打,速度快1.7倍论文中,谷歌表示,对于规模相当的系统,TPU v4可以提供比英伟达A100强1.

ai可以转成psd格式。转换方法:1、打开Adobe Illustrator软件,依次点击顶部菜单栏的“文件”-“打开”,选择所需的ai文件;2、点击右侧功能面板中的“图层”,点击三杠图标,在弹出的选项中选择“释放到图层(顺序)”;3、依次点击顶部菜单栏的“文件”-“导出”-“导出为”;4、在弹出的“导出”对话框中,将“保存类型”设置为“PSD格式”,点击“导出”即可;

Yann LeCun 这个观点的确有些大胆。 「从现在起 5 年内,没有哪个头脑正常的人会使用自回归模型。」最近,图灵奖得主 Yann LeCun 给一场辩论做了个特别的开场。而他口中的自回归,正是当前爆红的 GPT 家族模型所依赖的学习范式。当然,被 Yann LeCun 指出问题的不只是自回归模型。在他看来,当前整个的机器学习领域都面临巨大挑战。这场辩论的主题为「Do large language models need sensory grounding for meaning and u

ai顶部属性栏不见了的解决办法:1、开启Ai新建画布,进入绘图页面;2、在Ai顶部菜单栏中点击“窗口”;3、在系统弹出的窗口菜单页面中点击“控制”,然后开启“控制”窗口即可显示出属性栏。

ai移动不了东西的解决办法:1、打开ai软件,打开空白文档;2、选择矩形工具,在文档中绘制矩形;3、点击选择工具,移动文档中的矩形;4、点击图层按钮,弹出图层面板对话框,解锁图层;5、点击选择工具,移动矩形即可。

引入密集强化学习,用 AI 验证 AI。 自动驾驶汽车 (AV) 技术的快速发展,使得我们正处于交通革命的风口浪尖,其规模是自一个世纪前汽车问世以来从未见过的。自动驾驶技术具有显着提高交通安全性、机动性和可持续性的潜力,因此引起了工业界、政府机构、专业组织和学术机构的共同关注。过去 20 年里,自动驾驶汽车的发展取得了长足的进步,尤其是随着深度学习的出现更是如此。到 2015 年,开始有公司宣布他们将在 2020 之前量产 AV。不过到目前为止,并且没有 level 4 级别的 AV 可以在市场


Hot AI Tools

Undresser.AI Undress
AI-powered app for creating realistic nude photos

AI Clothes Remover
Online AI tool for removing clothes from photos.

Undress AI Tool
Undress images for free

Clothoff.io
AI clothes remover

AI Hentai Generator
Generate AI Hentai for free.

Hot Article

Hot Tools

Safe Exam Browser
Safe Exam Browser is a secure browser environment for taking online exams securely. This software turns any computer into a secure workstation. It controls access to any utility and prevents students from using unauthorized resources.

SublimeText3 Linux new version
SublimeText3 Linux latest version

VSCode Windows 64-bit Download
A free and powerful IDE editor launched by Microsoft

Atom editor mac version download
The most popular open source editor

SublimeText3 Mac version
God-level code editing software (SublimeText3)
