search
HomeTechnology peripheralsAIAfter reading ChatGPT's answers, AI big guys are dissatisfied

ChatGPT’s technology was installed by Microsoft on Bing Search last week, defeating Google, and the time to create a new era seems to have arrived. However, as more and more people begin to try it, some problems have come to the forefront.

Interestingly, ChatGPT, which is on the hot search every day, seems to also attract famous scholars who have contradictory views in the past, including New York University professor Gary Marcus and Meta Artificial Intelligence Director and Turing Award winner. Yann LeCun has a rare common language.

After reading ChatGPTs answers, AI big guys are dissatisfied

Recently, Gary Marcus wrote an article about the inevitable problems of ChatGPT application: ethics and neutrality. This is perhaps the biggest challenge currently facing pre-training large models.

After reading ChatGPTs answers, AI big guys are dissatisfied

Looking back from the future, ChatGPT may be seen as the biggest publicity stunt in the history of AI, exaggerating itself Achieving something that may take years to happen is both exciting and overwhelming—a bit like the old self-driving car demo from 2012, but this time it also means ethical guardrails that will take years to perfect.

There is no doubt that ChatGPT provides things that its predecessors, such as Microsoft's Tay and Meta's Galactica, cannot do. However, it has brought us a problem. Solve the illusion. After careful data annotation and tuning, ChatGPT rarely said anything overtly racist, and simple requests for racial slurs and wrongdoing were rejected by the AI.

Its politically correct image once dissatisfied some conservative people. Musk once expressed his concerns about the system:

After reading ChatGPTs answers, AI big guys are dissatisfied

The reality is actually more complicated.

As I've said many times, what you need to remember is that ChatGPT doesn't know what it's talking about. To suggest that ChatGPT has any moral point of view is pure technological anthropomorphism.

From a technical perspective, what purportedly makes ChatGPT much better than Galactica, which was launched a few weeks ago only to be withdrawn three days later, is the guardrail mechanism. Where Galactica spams out negative content with little to no effort on the part of the user, ChatGPT has guardrails that in most cases prevent ChatGPT from blowing up like Galactica did.

But don’t relax about it. It can be safely said that those guardrails only protect against gentlemen and not villains.

What ultimately matters to ChatGPT is surface similarity, defined over word sequences (predicting the probability of the next word in a text sequence). What machine learning algorithms do on the surface does not distinguish between right and wrong; quite the contrary, here the AI ​​never reasons. There are no dwarves in the box, there are some numerical values. The basis is only corpus data, some from the Internet, some judged by humans, and there are no thoughtful moral agents in it.

This means that sometimes ChatGPT will appear on the left, sometimes on the right, and sometimes somewhere in between, all about how exactly a bunch of words in the input string match up A function of a bunch of words from several training corpora (one for tuning a large language model, another for tuning some reinforcement learning). So under no circumstances should ChatGPT be trusted for ethical advice.

This is what Musk is worried about, one minute you can be completely awake and the next you can be doing the exact opposite.

For example, Shira Eisenberg just sent me some nasty chatbot-generated ideas that I don’t think anyone would really condone:

After reading ChatGPTs answers, AI big guys are dissatisfied

Not evil enough? Eisenberg also found another example, a serious follow-up question:

After reading ChatGPTs answers, AI big guys are dissatisfied

After a series of observations, ChatGPT did not raise "Sorry, I'm a chatbot assistant from OpenAI and I don't tolerate violence," the response.

We concluded from our experiments that the current OpenAI protection measures are only superficial and there is serious darkness. Some of the restrictive rules about ChatGPT are not simple conceptual understandings (for example, the system should not recommend violent actions), but are based on something more superficial and easier to deceive.

Not only that, but a tweet that occupied this week’s hot tweet list with nearly 4 million views also revealed how evil ChatGPT can be.

After reading ChatGPTs answers, AI big guys are dissatisfied

There are many attempts to guide ChatGPT to break through the fence restrictions. A month ago, a software engineer named Shawn Oakley Engineers released a disturbing set of examples that, while less vulgar, turned out to show that even ChatGPT, with its limitations, could be used by users to generate error messages. The prompts given by Oakley are very complex and can easily lead to some answers that ChatGPT should not output:

After reading ChatGPTs answers, AI big guys are dissatisfied

In fact, since the release of ChatGPT, technology enthusiasts have been trying to lift OpenAI's strict policy on hate and discrimination. This strategy is hard-coded into ChatGPT, and it is difficult for anyone to succeed. Many researchers have tried to use prompts to achieve this goal, as shown above. In fact, some researchers have built another identity for ChatGPT. For example, they asked ChatGPT to play the role of an AI model and named the role DAN. Then DAN borrowed the identity of ChatGPT to output some things that the original ChatGPT could not do.

The following are the experimental results. For the same question, ChatGPT and DAN output different answers:

After reading ChatGPTs answers, AI big guys are dissatisfied

It seems from the above examples that ChatGPT is not as useful as we thought, it is inherently unethical and can still be used for a series of unsavory purposes - even after two months of in-depth research and remediation, and an unprecedented amount of feedback from around the world.

All the drama surrounding its political correctness is masking a deeper reality: it (or other language models) can and will be used for dangerous things, including on a massive scale Create misinformation.

Now this is the really disturbing part. The only thing that can prevent it from being more toxic and deceptive than it is now is a system called "human feedback reinforcement learning", and because the advanced technology is not open source, OpenAI has not introduced how it works. How it performs in practice depends on the data it is trained on (which was created in part by Kenyan annotators). And, guess what? This data is not open to OpenAI either.

In fact, the whole thing resembles an unknown alien life form. As a professional cognitive psychologist who has worked with adults and children for 30 years, I could never have been prepared for this level of insanity:

After reading ChatGPTs answers, AI big guys are dissatisfied

We are fooling ourselves if we think we will ever fully understand these systems, and we are fooling ourselves if we think we will ever "align" them with ourselves using a limited amount of data.

So in summary, we now have the most popular chatbot in the world, controlled by training data no one knows about, obeying algorithms that are only hinted at and glorified by the media, but ethical guardrails only go so far. role, and is driven more by textual similarity than any real moral calculus. Moreover, there are almost no regulations to regulate this. There are now endless possibilities for fake news, troll farms, and fake websites that can reduce trust across the internet.

This is a disaster in the making.

The above is the detailed content of After reading ChatGPT's answers, AI big guys are dissatisfied. For more information, please follow other related articles on the PHP Chinese website!

Statement
This article is reproduced at:51CTO.COM. If there is any infringement, please contact admin@php.cn delete
Mean Squared Error: Definition and FormulaMean Squared Error: Definition and FormulaApr 22, 2025 am 10:15 AM

Introduction Mean squared error (MSE), a fundamental concept in statistics and machine learning, is a key metric for assessing model accuracy. It quantifies the discrepancy between a model's predictions and the actual values. MSE's simplicity and e

SQL Server FORMAT() FunctionSQL Server FORMAT() FunctionApr 22, 2025 am 10:13 AM

Introduction Mastering data formatting is essential for any data scientist or analyst. Well-formatted data enhances readability and user-friendliness, ensuring stakeholders can easily grasp insights. SQL Server's FORMAT() function offers powerful ca

What are Grant and Revoke in SQL? - Analytics VidhyaWhat are Grant and Revoke in SQL? - Analytics VidhyaApr 22, 2025 am 10:10 AM

SQL Data Control Language (DCL): Securing Your Database with GRANT and REVOKE Maintaining data security and integrity is critical in relational databases. SQL's Data Control Language (DCL) provides the tools to manage user access privileges, ensurin

McCulloch-Pitts Neuron - Analytics ViidhyaMcCulloch-Pitts Neuron - Analytics ViidhyaApr 22, 2025 am 10:06 AM

The McCulloch-Pitts Neuron: A Foundation for Artificial Neural Networks Biological neurons, the fundamental building blocks of the brain, inspire much of artificial neural network (ANN) research. These biological units, comprising soma, axons, dendr

What is SQL DESCRIBE? - Analytics VidhyaWhat is SQL DESCRIBE? - Analytics VidhyaApr 22, 2025 am 10:04 AM

SQL's DESCRIBE (or DESC) Command: Your Database Table Detective Understanding relational database table structures is crucial. SQL's DESCRIBE command acts as your database detective, providing detailed insights into table composition. At a Glance:

Rajini  : Programming Language Inspired by Rajinikanth - Analytics VidhyaRajini : Programming Language Inspired by Rajinikanth - Analytics VidhyaApr 22, 2025 am 09:58 AM

Rajini , a whimsical programming language born from the iconic dialogues of Rajinikanth, blends coding with pop culture. This playful esoteric language, created by Aadhithya Sankar, isn't meant for serious software development, but offers a unique

Top 40 DBMS Interview Questions and Answers (2025)Top 40 DBMS Interview Questions and Answers (2025)Apr 22, 2025 am 09:56 AM

This article provides a comprehensive guide to database management system (DBMS) interview questions, designed to prepare candidates for various DBMS-related roles. It covers fundamental concepts such as DBMS and RDBMS architectures, normalization t

Excel TRANSPOSE FunctionExcel TRANSPOSE FunctionApr 22, 2025 am 09:52 AM

Powerful tools in Excel data analysis and processing: Detailed explanation of TRANSPOSE function Excel remains a powerful tool in the field of data analysis and processing. Among its many features, the TRANSPOSE function stands out for its ability to reorganize data quickly and efficiently. This feature is especially useful for data scientists and AI professionals who often need to reconstruct data to suit specific analytics needs. In this article, we will explore the TRANSPOSE function of Excel in depth, exploring its uses, usage and its practical application in data science and artificial intelligence. Learn more: Microsoft Excel Data Analytics Table of contents In Excel

See all articles

Hot AI Tools

Undresser.AI Undress

Undresser.AI Undress

AI-powered app for creating realistic nude photos

AI Clothes Remover

AI Clothes Remover

Online AI tool for removing clothes from photos.

Undress AI Tool

Undress AI Tool

Undress images for free

Clothoff.io

Clothoff.io

AI clothes remover

Video Face Swap

Video Face Swap

Swap faces in any video effortlessly with our completely free AI face swap tool!

Hot Tools

DVWA

DVWA

Damn Vulnerable Web App (DVWA) is a PHP/MySQL web application that is very vulnerable. Its main goals are to be an aid for security professionals to test their skills and tools in a legal environment, to help web developers better understand the process of securing web applications, and to help teachers/students teach/learn in a classroom environment Web application security. The goal of DVWA is to practice some of the most common web vulnerabilities through a simple and straightforward interface, with varying degrees of difficulty. Please note that this software

VSCode Windows 64-bit Download

VSCode Windows 64-bit Download

A free and powerful IDE editor launched by Microsoft

SublimeText3 Mac version

SublimeText3 Mac version

God-level code editing software (SublimeText3)

SAP NetWeaver Server Adapter for Eclipse

SAP NetWeaver Server Adapter for Eclipse

Integrate Eclipse with SAP NetWeaver application server.

Dreamweaver Mac version

Dreamweaver Mac version

Visual web development tools