Home  >  Article  >  Technology peripherals  >  Marcus posted an article criticizing LeCun: Deep learning alone cannot achieve human-like intelligence

Marcus posted an article criticizing LeCun: Deep learning alone cannot achieve human-like intelligence

WBOY
WBOYforward
2023-04-09 23:01:141051browse

In March of this year, Gary Marcus caused a stir in the artificial intelligence academic community after he proposed the idea that "deep learning has hit a wall."

#At that time, even the three giants of deep learning could not sit still. First, Geoffrey Hinton refuted this view in a podcast.

The following June, Yann LeCun wrote an article in response to this, and pointed out that don’t treat temporary difficulties as hitting a wall.

Now, Marcus has published an article titled "Deep Learning Alone Isn't Getting Us To Human-Like AI" in the US NOEMA magazine .

Marcus posted an article criticizing LeCun: Deep learning alone cannot achieve human-like intelligence

Similarly, he still has not changed his view - deep learning alone cannot achieve human-like intelligence.

# and proposed that the current artificial intelligence research mainly focuses on deep learning, and it is time to reconsider it.

Marcus posted an article criticizing LeCun: Deep learning alone cannot achieve human-like intelligence

##In the past 70 years, the most fundamental debate in the field of artificial intelligence is: whether artificial intelligence systems should be built on "symbols" In terms of operation, it should still be based on a brain-like "neural network" system.

#In fact, there is a third possibility here: hybrid model - combining the deep learning of neural networks with the powerful abstraction capabilities of symbolic operations.

LeCun's recent article "What AI Can Tell Us About Intelligence" published in NOEMA magazine also discussed this issue, but Marcus pointed out that his article It seems clear, but it has obvious flaws, namely logical contradictions.

The article begins with them rejecting the hybrid model, but the article ends by acknowledging the existence of the hybrid model and mentioning it as a possible way forward.

Hybrid Model of Neural Networks and Symbolic Operations

Marcus pointed out that LeCun and Browning’s point is mainly that “if a model learns Symbolic manipulation, it is not mixed".

#But the problem of machine learning is a development problem (how is the system generated?)

How the system operates once it is developed is a computational problem (for example, does it use one mechanism or two mechanisms?), that is, "any system that utilizes both symbols and neural networks is a hybrid model."

Perhaps what they really want to say is that artificial intelligence is likely to be a learned hybrid rather than a born hybrid. But a hybrid of learning is still a hybrid.

#And Marcus’s point of view is, “Symbol operation itself is innate, or there is another thing that is innate, This kind of thing indirectly contributed to the emergence of symbolic operations."

#So our research focus should be on how to discover this medium that indirectly promotes symbolic operations.

That is to put forward the hypothesis that as long as we can figure out what medium allows the system to reach the level of learning symbolic abstraction, we can build and utilize all the knowledge in the world system.

#Next, Marcus cited classics and combed the history of the debate on symbolic manipulation and neural networks in the field of artificial intelligence.

Early artificial intelligence pioneers such as Marvin Minsky and John McCarthy believed that symbolic manipulation was the only reasonable way forward.

Marcus posted an article criticizing LeCun: Deep learning alone cannot achieve human-like intelligence

Frank Rosenblatt, the pioneer of neural networks, believes that AI may perform better based on a structure in which neural nodes superimpose and process digital inputs.

Marcus posted an article criticizing LeCun: Deep learning alone cannot achieve human-like intelligence

In fact, these two possibilities are not mutually exclusive.

The neural network used by AI is not a literal biological neuron network. On the contrary, it is a simplified digital model that has some characteristics of the human brain. features, but with minimal complexity.

In principle, these abstract symbols can be connected in many different ways, some of which can directly implement logical and symbolic operations.

Warren S. McCulloch and Walter Pitts published A Logical Calculus of the Ideas Inmanent in Nervous Activity in 1943, explicitly recognizing this This possibility.

Others, including Frank Rosenblatt in the 1950s and David Rumelhart and Jay McClelland in the 1980s, proposed neural networks as an alternative to symbolic manipulation. Geoffrey Hinton also generally supports this position.

#Then Marcus cueed Turing Award winners such as LeCun, Hinton and Yoshua Bengio one after another.

It means that it doesn’t matter what I say, other big guys said so!

In 2015, LeCun, Bengio and Hinton wrote a manifesto-style paper on deep learning in Nature.

The article ends with an attack on symbols, arguing that "a new paradigm is needed to replace rule-based operations on symbolic expressions by operations on large vectors ”.

Marcus posted an article criticizing LeCun: Deep learning alone cannot achieve human-like intelligence

In fact, Hinton was so convinced that symbols were a dead end that the same year he published a paper at Stanford University called "The Ether Symbols” speech—likening symbols to one of the biggest mistakes in the history of science.

A similar argument was made in the 1980s by two of his former collaborators, Rumelhart and McClelland, in a famous 1986 book Arguing that symbols are not "the essence of human computation" sparked a great debate.

Marcus said that when he wrote an article in 2018 defending symbolic manipulation, LeCun dismissed his argument for hybrid AI without explanation. , dismissed it as "mostly wrong" on Twitter.

Marcus posted an article criticizing LeCun: Deep learning alone cannot achieve human-like intelligence

Then he said that two famous experts in the field of deep learning also expressed support for hybrid AI.

Andrew Ng expressed support for such a system in March. Sepp Hochreiter—the co-creator of LSTMs, one of the leading deep learning architectures for learning sequences—has done the same thing, publicly stating in April that the most promising broad-based AI approach is neurosymbolic AI.

In LeCun and Browning’s new perspective, symbolic manipulation is actually crucial, as Marcus and Steven Pinker have proposed since 1988.

Marcus therefore accused Lecun, "I proposed your point of view decades ago, but your research has gone back decades." .

#And it’s not just me who says this, other big guys think so too.

The rest of LeCun and Browning’s articles can be roughly divided into three parts:

1. My position Error description 2. Efforts to narrow the scope of the hybrid model 3. Symbolic manipulation may be due to learning rather than innate reasons.

Next, Marcus refuted the views in LeCun’s paper:

LeCun and Browning said, “Marcus said , if you don't have symbolic manipulation in the first place, you'll never have it".

#In fact, I explicitly admitted in my 2001 book "Algebraic Thinking" that we are not sure whether symbolic operations are innate.

They criticized my statement that "deep learning cannot make further progress", but my actual point is not that DL will never make progress on any issue. , but that deep learning itself is the wrong tool for certain tasks such as composition and reasoning.

#Similarly, they slandered me by saying that symbolic reasoning is either present or not (1 or 0) in the system.

#This is just nonsense.

It is true that DALL-E does not use symbols for reasoning, but this does not mean that any system that includes symbolic reasoning must be either there or not.

At least as early as the system MYCIN in the 1970s, there was a purely symbolic system that could carry out various quantitative reasoning.

Symbolic manipulation innateness

Can symbolic manipulation ability be learned rather than established from the beginning?

The answer is yes.

# Marcus said that although previous experiments cannot guarantee that the ability to manipulate symbols is innate, they are almost indistinguishable from this view. They do pose a challenge to any theory of learning that relies on large amounts of experience.

and put forward the following 2 main arguments:

1 , Learnability

In his 2001 book The Algebraic Mind, Marcus showed that some systems can learn symbolic operations .

#A system that has some built-in starting points will be more effective at understanding the world than a pure blank slate.

In fact, even LeCun’s own most famous work-On Convolutional Neural Networks-is a good example: the study of neural network learning methods Built-in constraints greatly improve efficiency. When symbolic operations are well integrated, even greater gains may be achieved.

2. Human infants show some ability to manipulate symbols

In a series of oft-cited rule-learning experiments, infants generalized abstract patterns beyond the concrete examples they had been trained on. Subsequent research on the implicit logical reasoning ability of human infants further proved this point.

# Additionally, research has shown that bees, for example, can generalize the solar azimuth function to lighting conditions they have never seen before.

Marcus posted an article criticizing LeCun: Deep learning alone cannot achieve human-like intelligence

In LeCun’s view, learning symbols is equivalent to something acquired in later years, because when you are young, you need to be more precise and more precise. Professional skills.

What is puzzling is that, after objecting to the innate nature of symbolic manipulation, LeCun did not provide strong evidence to prove that symbolic manipulation is an acquired habit. Got it.

If a baby goat can crawl down a hill soon after birth, why can't a nascent neural network incorporate a little symbolic manipulation out of the box?

Marcus posted an article criticizing LeCun: Deep learning alone cannot achieve human-like intelligence

At the same time, LeCun and Browning did not specify, how can the lack of intrinsic mechanisms of symbolic manipulation solve well-known specific problems in language understanding and reasoning?

They only give a weak generalization: since deep learning has overcome the problem from 1 to N, we should have some confidence that it can overcome the N 1 problem confidence.

One should wonder whether deep learning has reached its limits. Given the continued incremental improvements to missions seen recently in DALL-E 2, Gato, and PaLM, it seems wise not to mistake momentary difficulties for "walls." The inevitable failure of deep learning has been predicted before, but it’s not worth betting on.

Optimism is one thing, but reality must be seen clearly.

Deep learning in principle faces some specific challenges, mainly in terms of compositionality, systematization and language understanding, all of which revolve around generalization and "distribution" migrate".

#Now everyone realizes that distribution migration is the Achilles heel of current neural networks. Of course, deep learning has made progress, but not much progress has been made on these basic problems.

In Marcus’ view, the circumstances under which symbolic manipulation might be innate are much the same as before:

1. The current system, 20 years after the emergence of "algebraic thinking", is still unable to reliably extract symbolic operations (such as multiplication) even in the face of massive data sets and training.

2. The example of human infants shows that before formal education, they have the ability to generalize complex aspects of natural language and reasoning. 3. A little inner symbolic meaning can greatly improve learning efficiency. Part of the power of AlphaFold 2 comes from carefully constructed innate representations of molecular biology.

In short, the world may be roughly divided into three bins:

One is a system in which symbol manipulation equipment is fully installed in the factory.

#The second is that the system with innate learning device lacks symbolic operations, but with the correct data and training environment, it has enough ability to obtain it.

#The third is a system that cannot obtain a complete symbol manipulation mechanism even with sufficient training.

#Current deep learning systems seem to fall into the third category: no symbolic manipulation mechanisms at the beginning, and no reliable symbolic manipulation mechanisms in the process.

#Currently, understanding the origins of symbolic operations is our top priority. Even the most ardent proponents of neural networks now recognize the importance of symbolic operations in achieving AI.

And this is what the neurosemiotics community has been paying attention to: How to integrate data-driven learning and symbolic representation into a single, more powerful intelligence Working in harmony?

The above is the detailed content of Marcus posted an article criticizing LeCun: Deep learning alone cannot achieve human-like intelligence. For more information, please follow other related articles on the PHP Chinese website!

Statement:
This article is reproduced at:51cto.com. If there is any infringement, please contact admin@php.cn delete