Home >Technology peripherals >AI >Google and MIT propose a unified framework MAGE: representation learning surpasses MAE, and unsupervised image generation surpasses Latent Diffusion
Recognition and generation are the two core tasks in the field of artificial intelligence. If they can be merged into a unified system, these two tasks should be complementary. In fact, in natural language processing, models like BERT [1] are not only able to generate high-quality text but also extract features from the text.
However, in the field of computer vision, current image generation models and recognition models are mostly trained separately, without fully utilizing the synergy of these two tasks. This is mainly due to the fact that the models of image generation and image recognition usually have essential structural differences: the input of image generation is low-dimensional features or noise, and the output is a high-dimensional original image; in contrast, the input of image recognition is high-dimensional. dimensional original image, while the output is low-dimensional features.
Recently, researchers from MIT and Google Research proposed a representation learning method based on image semantic masking, which for the first time achieved image generation and representation in a unified framework learned and achieved SOTA performance on multiple data sets. The research paper has been accepted by CVPR 2023, and the relevant code and pre-trained model have been open source.
In CVPR 2022 On, MAE [2] proposed a representation learning method based on image masks (MIM) and achieved very good results on multiple subtasks. At a masking rate of up to 75%, MAE can reconstruct an image that closely matches the semantics of the original image, thereby allowing the network to self-supervisedly learn features in the image. However, as shown in Figure 1, although the image reconstructed by MAE has similar semantic information to the original image, serious blurring and distortion problems occur. Similar issues arise in all MIM-based representation learning methods. At the same time, current generative models, whether diffusion models or GANs, lack the ability to extract high-quality image features.
Figure 1: Comparison of MAE and MAGE reconstruction
Method OverviewIn response to the above problems, the author of this article proposed MAGE (Masked Generative Encoder), which for the first time realized a unified image generation and feature extraction model. Different from the masking method where MIM acts directly on the image, MAGE proposes a masked image token modeling method based on image semantic symbols. As shown in the figure, MAGE first uses the VQGAN [3] encoder to convert the original image into discrete semantic symbols. After that, MAGE randomly masks it and uses the transformer-based encoder-decoder structure to reconstruct the mask. The reconstructed semantic symbols can be used to generate the original image through the VQGAN decoder. By using different masking rates in training, MAGE can train both generative models (nearly 100% masking rate) and representation learning (50%-80% masking rate). As shown in Figure 1, the image reconstructed by MAGE not only has semantic information consistent with the original image, but can also ensure the diversity and authenticity of the generated image at the same time.
##Figure 2: MAGE Structure DiagramExperimental results
MAGE has reached or exceeded SOTA on multiple image generation and image recognition tasks.
In the unsupervised image generation task of ImageNet, the FID of MAGE dropped from the previous > 20 to 7.04, even reaching the level of supervised image generation (the FID of supervised Latent Diffusion on ImageNet is 3.60) :
picture 3: MAGE unsupervised image generation example
MAGE can also perform various image editing tasks, including image inpainting, outpainting, and uncropping:
#Figure 4: MAGE image editing sample
In In terms of representation learning, MAGE has greatly improved compared to the current MIM method in tasks such as ImageNet linear probing, few-shot learning, and transfer learning, and can reach or exceed the level of the current optimal self-supervised learning method.
ConclusionThis article aims to unify image generation and representation learning. To this end, the author of this article proposes MAGE, a self-supervised learning framework based on image semantic masking. This framework is simple and efficient, and for the first time reaches or exceeds SOTA performance in both image generation and representation learning. Interested readers can view the original text of the paper to learn more research details.
The above is the detailed content of Google and MIT propose a unified framework MAGE: representation learning surpasses MAE, and unsupervised image generation surpasses Latent Diffusion. For more information, please follow other related articles on the PHP Chinese website!