Home  >  Article  >  Technology peripherals  >  AI reading brain explodes! Scan brain images and Stable Diffusion reproduces images realistically

AI reading brain explodes! Scan brain images and Stable Diffusion reproduces images realistically

WBOY
WBOYforward
2023-04-12 15:52:031690browse

Even without the magic of Hogwarts, you can still see what others are thinking!

The method is very simple. It visualizes brain images based on Stable Diffusion.

For example, the bears, airplanes, and trains you see look like this.

AI reading brain explodes! Scan brain images and Stable Diffusion reproduces images realistically

When the AI ​​sees the brain signal, the image generated is as follows, which shows that all the necessary points are included .

AI reading brain explodes! Scan brain images and Stable Diffusion reproduces images realistically

This AI brain reading technology has just been accepted by CVPR 2023, giving fans an instant "intracranial orgasm."

AI reading brain explodes! Scan brain images and Stable Diffusion reproduces images realistically

Too wild! Forget about prompting projects, now you just need to use your brain to "think" about those pictures.

AI reading brain explodes! Scan brain images and Stable Diffusion reproduces images realistically

Imagine using Stable Diffusion to reconstruct visual images from fMRI data, which may mean the development of non-invasive technologies in the future brain-computer interface.

Let AI skip human language directly and perceive what is thought in the human brain.

AI reading brain explodes! Scan brain images and Stable Diffusion reproduces images realistically

By then, Neuralink developed by Musk will also catch up with this AI ceiling.

No need for fine-tuning, use AI to directly reproduce what you are thinking

So, how does AI realize brain reading?

The latest research comes from a research team at Osaka University in Japan.

AI reading brain explodes! Scan brain images and Stable Diffusion reproduces images realistically

##Paper address: https://www.php.cn/link/0424d20160a6a558e5bf86a7bc9b67f0

Researchers at Osaka University Graduate School of Frontier Biosciences and CiNet at NICT in Japan reconstructed visual experience from fMRI data based on the latent diffusion model (LDM), more specifically via Stable Diffusion .

The framework of the entire operation process is also very simple: 1 image encoder, 1 image decoder, and 1 semantic decoder.

AI reading brain explodes! Scan brain images and Stable Diffusion reproduces images realistically

# By doing this, the team eliminated the need to train and fine-tune complex AI models.

All that needs to be trained is a simple linear model that maps fMRI signals from lower and upper visual brain regions to a single Stable Diffusion component.

Specifically, the researchers mapped brain regions as input to image and text encoders. The lower brain areas are mapped to image encoders, and the upper brain areas are mapped to text encoders. This allows the system to use image composition and semantic content for reconstruction.

The first is decoding analysis. The LDM model used in the study consists of an image encoder ε, an image decoder D, and a text encoder τ.

The researchers decoded the latent representation of the reconstructed image z and the related text c from the fMRI signals of the early and high-level visual cortex respectively, and used them as input to generate the reproduced image Xzc by the autoencoder.

AI reading brain explodes! Scan brain images and Stable Diffusion reproduces images realistically

The researchers then established a coding model to predict the fMRI signals from different components of the LDM, thereby Explore the inner workings of LDM.

AI reading brain explodes! Scan brain images and Stable Diffusion reproduces images realistically

The researchers conducted experiments using fMRI images from the Natural Scenes Dataset (NSD) and tested whether they could Stable Diffusion to reconstruct what the subject saw.

It can be seen that the prediction accuracy of the latent image related to the encoding model and LDM, the last model produces the highest prediction accuracy in the visual cortex at the back of the brain.

AI reading brain explodes! Scan brain images and Stable Diffusion reproduces images realistically

The visual reconstruction results of a subject show that the image reconstructed with only z is visually consistent with the original image, But it cannot capture the semantic content.

While the image reconstructed using only c has better semantic fidelity, but poor visual consistency, the image reconstructed using zc can have both high semantic fidelity and poor visual consistency. high resolution.

AI reading brain explodes! Scan brain images and Stable Diffusion reproduces images realistically

Reconstruction results from all subjects on the same image show that the effect of reconstruction varies between different subjects It is stable and relatively accurate.

The differences in specific details may stem from differences in individual perception experience or data quality, rather than errors in the reconstruction process.

AI reading brain explodes! Scan brain images and Stable Diffusion reproduces images realistically

#Finally, the results of the quantitative assessment were graphed.

Various results show that the method used in the study can not only capture the low-level visual appearance, but also capture the high-level semantic content of the original stimulus.

AI reading brain explodes! Scan brain images and Stable Diffusion reproduces images realistically

From this point of view, the experiments show that the combination of image and text decoding provides accurate reconstruction.

There were differences in accuracy between subjects, but these differences were related to the quality of the fMRI images, the researchers said. According to the team, the quality of the reconstruction is comparable to current SOTA methods, but does not require training of the AI ​​model used in it.

At the same time, the team also used models derived from fMRI data to study the various building blocks of Stable Diffusion, such as how semantic content is generated during the reverse diffusion process, Or what process happens in U-Net.

In the early stages of the denoising process, U-Net's bottleneck layer (orange) produces the highest prediction performance, and as the denoising process proceeds, the early layers (blue) For predictions of early visual cortex activity, the bottleneck layer shifts to higher-level visual cortex.

This means that at the beginning of the diffusion process, the image information is compressed in the bottleneck layer, and with denoising, the separation between U-Net layers appears in the visual cortex .

AI reading brain explodes! Scan brain images and Stable Diffusion reproduces images realistically

# Additionally, the team is developing a quantitative explanation of image transformations at different stages of diffusion. In this way, the researchers aim to contribute to a better understanding of diffusion models from a biological perspective, which are widely used but whose understanding is still limited.

Have human brain images been decoded by AI?

For years, researchers have been using artificial intelligence models to decode information from the human brain.

At the core of most methods, pre-recorded fMRI images are used as input to generative AI models of text or images.

For example, in early 2018, a team of researchers from Japan showed how a neural network could reconstruct images from fMRI recordings.

In 2019, a group reconstructed images from monkey neurons, and Meta's research group, led by Jean-Remi King, published new work such as from fMRI data to get the text.

AI reading brain explodes! Scan brain images and Stable Diffusion reproduces images realistically

In October 2022, a team at the University of Texas at Austin showed that a GPT model can generate data from fMRI scans Text that describes the semantic content one sees in a video is inferred.

In November 2022, researchers from the National University of Singapore, the Chinese University of Hong Kong, and Stanford University used the MinD-Vis diffusion model to reconstruct images from fMRI scans with significantly higher accuracy than methods available at the time.

AI reading brain explodes! Scan brain images and Stable Diffusion reproduces images realistically

Going further back, some netizens pointed out that “generating images based on brain waves has been around since at least 2008. Yes, it is simply ridiculous to imply that Stable Diffusion can read people's minds in some way."

This paper published in Nature by the University of California, Berkeley, stated, A person's brain wave activity can be converted into images using a visual decoder.

AI reading brain explodes! Scan brain images and Stable Diffusion reproduces images realistically

AI reading brain explodes! Scan brain images and Stable Diffusion reproduces images realistically

When it comes to tracing history, there are still people who directly Take out a 1999 study by Stanford Li Feifei on reconstructing images from the cerebral cortex.

AI reading brain explodes! Scan brain images and Stable Diffusion reproduces images realistically

## Li Feifei also commented and forwarded it, saying that she was still a university intern at the time.

AI reading brain explodes! Scan brain images and Stable Diffusion reproduces images realistically

Also in 2011, a UC Berkeley study used functional magnetic resonance imaging (fMRI) and computational models to A preliminary reconstruction of the brain's "dynamic visual image".

AI reading brain explodes! Scan brain images and Stable Diffusion reproduces images realistically

In other words, they recreated clips that people have seen.

But compared with the latest research, this reconstruction cannot be called "high definition" at all and is almost unrecognizable.

about the author

Yu Takagi

Yu Takagi is an assistant professor at Osaka University. His research interests are at the intersection of computational neuroscience and artificial intelligence.

During his PhD, he studied techniques for predicting individual differences from whole-brain functional connectivity using functional magnetic resonance imaging (fMRI) in the ATR Brain Information Communication Research Laboratory.

Most recently, he has used machine learning techniques to understand dynamic computations in complex decision-making tasks at the Oxford Center for Human Brain Activity at the University of Oxford and the Department of Psychology at the University of Tokyo.

AI reading brain explodes! Scan brain images and Stable Diffusion reproduces images realistically

Shinji Nishimoto

Shinji Nishimoto is Professor at Osaka University. His research focuses on the quantitative understanding of visual and cognitive processing in the brain.

AI reading brain explodes! Scan brain images and Stable Diffusion reproduces images realistically

More specifically, the research focus of Professor Nishimoto’s team is to build predictive models of brain activity induced under natural perception and cognitive conditions. Understanding neural processing and representation.

AI reading brain explodes! Scan brain images and Stable Diffusion reproduces images realistically

Some netizens asked the author whether this research can be used to interpret dreams?

"It is possible to apply the same technology to brain activity during sleep, but the accuracy of such an application is unclear."

AI reading brain explodes! Scan brain images and Stable Diffusion reproduces images realistically

#After reading this research: Legilimency is in place.

Reference:

https: //www.php.cn/link/0424d20160a6a558e5bf86a7bc9b67f0

https://www.php.cn/link /02d72b702eed900577b953ef7a9c1182

The above is the detailed content of AI reading brain explodes! Scan brain images and Stable Diffusion reproduces images realistically. For more information, please follow other related articles on the PHP Chinese website!

Statement:
This article is reproduced at:51cto.com. If there is any infringement, please contact admin@php.cn delete