Home >Technology peripherals >AI >Five major models of generative AI: VAEs, GANs, Diffusion, Transformers, NeRFs
Choosing the right GenAI model for the task requires understanding the technology used by each model and its specific capabilities. Please learn about the five GenAI models of VAEs, GANs, Diffusion, Transformers and NerFs below.
Previously, most AI models focused on better processing, analyzing, and interpreting data. Until recently, breakthroughs in so-called generative neural network models have led to a range of new tools for creating all kinds of content, from photos and paintings to poetry, code, screenplays and films.
In the mid-2010s, researchers discovered new prospects for generative artificial intelligence models. At that time, they developed variational autoencoders (VAEs), generative adversarial networks (GAN), and diffusion models (Diffusion). Transformers, introduced in 2017, are groundbreaking neural networks capable of analyzing large data sets at scale to automatically create large language models (LLMs). In 2020, researchers introduced Neural Radiation Field (NeRF) technology, which can generate 2D content from 3D images
The rapid development of these generative models is an ongoing work, because researchers' tweaks often lead to huge improvements, and remarkable progress isn't slowing down. Professor White said at the University of California, Berkeley: "Model architectures are constantly changing, and new model architectures will continue to be developed."
Each model has its special capabilities, and currently, diffusion The model (Diffusion) performs exceptionally well in the field of image and video synthesis, the Transformers model (Transformers) performs well in the field of text, and GAN is good at expanding small data sets with reasonable synthetic samples. But choosing the best model always depends on the specific use case.
All models are different and AI researchers and ML (machine learning) engineers must choose the right one for the appropriate use case and required performance, taking into account the model’s use in computing , possible limitations in memory and capital.
Converter models in particular have contributed to the recent progress and excitement in generative models. Adnan Masood, chief artificial intelligence architect at UST Digital Transformation Consulting, said: "The latest breakthroughs in artificial intelligence models come from pre-training on large amounts of data and using self-supervised learning to train models without explicit labels."
For example, OpenAI’s generative pre-trained converter family of models is one of the largest and most powerful models in the category. Among them, the GPT-3 model contains 17.5 billion parameters
Masood explained that the top Generative AI models use a variety of different techniques and methods to generate entirely new data. The main features and uses of these models include:
#Let’s go over each method in more detail.
VAE was developed in 2014 to use neural networks to encode data more efficiently
Yael Lev, head of AI at Sisense, said that the artificial intelligence analysis platform VAE has learned to express information more effectively. VAE consists of two parts: an encoder that compresses the data, and a decoder that restores the data to its original form. They are ideal for generating new instances from smaller pieces of information, repairing noisy images or data, detecting anomalous content in data and filling in missing information
However, variational autoencoders (VAEs) also tend to produce blurry or low-quality images, according to UST’s Masood. Another problem is that the low-dimensional latent space used to capture the data structure is complex and challenging. These shortcomings may limit the effectiveness of VAE in applications that require high-quality images or a clear understanding of the latent space. The next iteration of VAE will likely focus on improving the quality of generated data, speeding up training, and exploring its applicability to sequence data
GANs were developed in 2014 and are used to generate realistic faces and print figures. GANs pit neural networks that generate real content against neural networks that detect fake content. "Gradually, the two networks merge together to produce generated images that are indistinguishable from the original data," said Anand Rao, global AI leader at PwC.
GAN Commonly used for image generation, image editing, super-resolution, data enhancement, style transfer, music generation and deepfake creation. One problem with GANs is that they can suffer from mode collapse, where the generator produces limited and repetitive outputs, making them difficult to train. Masood said the next generation of GANs will focus on improving the stability and convergence of the training process, extending its applicability to other fields, and developing more effective evaluation metrics. GANs are also difficult to optimize and stabilize, and there is no clear control over the samples generated.
The diffusion model was developed in 2015 by a team of researchers at Stanford University using For simulating and inverting entropy and noise. Diffusion techniques provide a way to simulate phenomena such as how a substance such as salt diffuses into a liquid and then reverses it. This same model also helps generate new content from blank images.
Diffusion models are currently the first choice for image generation, and they are the basic models for popular image generation services, such as Dall-E 2, Stable Diffusion, Midjourney and Imagen. They are also used in pipelines to generate speech, video, and 3D content. Additionally, diffusion techniques can be used for data imputation, where missing data is predicted and generated. Many applications pair diffusion models with LLM for text-to-image or text-to-video generation. For example, Stable Diffusion 2 uses a contrastive language-image pre-trained model as a text encoder, and it also adds models for depth and upscaling.
Masood predicts that further improvements to models such as stable diffusion may focus on improving negative cues, enhancing the ability to generate images in the style of a specific artist, and improving celebrity images.
Transformers
Rewritten content: These techniques Can be applied to text summarization, chatbots, recommendation engines, language translation, knowledge bases, personalized recommendations (via preference models), sentiment analysis, and named entity recognition for identifying people, places, and things. Additionally, they can be used in areas such as speech recognition, such as OpenAI’s Whisper technology, as well as object detection in videos and images, image captioning, text classification, and dialogue generation.
AlthoughTransformers
are versatile, but they do have limitations. They can be expensive to train and require large data sets. The resulting models are also quite large, making it challenging to identify sources of bias or inaccurate results. "Their complexity also makes it difficult to explain their inner workings, hindering their interpretability and transparency," Massoud said.Transformer Model Architecture
NeRF
However, in 2022, researchers at Nvidia discovered a way to generate a new model in about 30 seconds. These models can represent 3D objects in units of a few megabytes with comparable quality while other technologies may require gigabytes. These models promise to lead to more efficient techniques for capturing and generating 3D objects in the Metaverse. Alexander Keller, research director at Nvidia, said NeRFs could eventually be as important to 3D graphics as digital cameras are to modern photography. Masood said NeRFs It shows great potential in robotics, urban mapping, autonomous navigation and virtual reality applications. However, NERF remains computationally expensive and combining multiple NERFs into larger scenes is challenging. The only viable use case for NeRF today is to convert images into 3D objects or scenes. Despite these limitations, Masood predicts that NeRF will find new roles in basic image processing tasks such as denoising, deblurring, upsampling, compression, and image editing within the GenAI ecosystem. It is important to note that these models are a work in progress, and researchers are seeking ways to improve individual models and combine them with other models and processing techniques. Lev predicts that generative models will become more general, applications will expand beyond traditional domains, and users will be able to more effectively guide AI models and understand how they work better.
There is also work in progress on multimodal models that use retrieval methods to call model libraries optimized for specific tasks. He also hopes that the generative model will be able to develop other capabilities, such as making API calls and using external tools. For example, an LLM fine-tuned based on the company's call center knowledge will provide answers to questions and perform troubleshooting, such as resetting the customer modem or when the problem is solved. send email.
Some people predict that the generative AI ecosystem will evolve into a three-layer model. The base layer is a series of basic models based on text, images, speech and code. These models ingest large amounts of data and are built on large deep learning models, combined with human judgment. Next, industry- and function-specific domain models will improve healthcare, legal, or other types of data processing. At the top level, companies will build proprietary models using proprietary data and subject matter expertise. These three layers will disrupt the way teams develop models and usher in a new era of models as a service
How to choose a generative AI model: First considerationsAccording to Sisense’s Lev, top considerations when choosing between models include the following:
The problem you are trying to solve.
Select a model known to be suitable for your specific task. For example, use transformers for language tasks and NeRF for 3D scenes.Quantity and quality of data. Diffusion requires a lot of good data to work properly, while VAE works better with less data.
Quality of results. GAN is better for clear and detailed images, while VAE is better for smoother results.
The difficulty of training the model. GAN can be difficult to train, while VAE and Diffusion are easier.
Computing resource requirements. Both NeRF and Diffusion require a lot of computer power to work properly.
Requires control and understanding. If you want more control over the results or a better understanding of how the model works, VAEs may be better than GANs.
The above is the detailed content of Five major models of generative AI: VAEs, GANs, Diffusion, Transformers, NeRFs. For more information, please follow other related articles on the PHP Chinese website!