search
HomeTechnology peripheralsAIPixelLM, a byte multi-modal large model that efficiently implements pixel-level reasoning without SA dependency

Multi-modal large models have exploded, are you ready to enter practical applications in fine-grained tasks such as image editing, autonomous driving and robotics?

The capabilities of most current models are still limited to generating text descriptions of the overall image or specific areas, and their capabilities in pixel-level understanding (such as object segmentation) are relatively limited.

In response to this problem, some work has begun to explore the use of multi-modal large models to process user segmentation instructions (for example, "Please segment the fruits rich in vitamin C in the picture").

However, the methods on the market all suffer from two major shortcomings:

1) cannot handle tasks involving multiple target objects, which is indispensable in real-world scenarios;

2) Relies on pre-trained image segmentation models like SAM, and the amount of calculation required for one forward propagation of SAM is enough for Llama-7B to generate more than 500 tokens.

In order to solve this problem, the Bytedance Intelligent Creation Team teamed up with researchers from Beijing Jiaotong University and University of Science and Technology Beijing to propose PixelLM, the first large-scale efficient pixel-level inference model that does not rely on SAM.

Before introducing it in detail, let’s experience the actual segmentation effect of several groups of PixelLM:

Compared with previous work, the advantages of PixelLM are:

  • It can Proficiently handle any number of open domain targets and diverse complex inference segmentation tasks.
  • Avoids additional and costly segmentation models, improving efficiency and migration capabilities to different applications.

Further, in order to support model training and evaluation in this research field, the research team built a data set for multi-objective reasoning segmentation scenarios based on the LVIS data set with the help of GPT-4V. MUSE, which contains more than 200,000 question-answer pairs involving more than 900,000 instance segmentation masks.

PixelLM, a byte multi-modal large model that efficiently implements pixel-level reasoning without SA dependency

PixelLM, a byte multi-modal large model that efficiently implements pixel-level reasoning without SA dependency

In order to achieve the above effects, how was this research conducted?

Principle behind

PixelLM, a byte multi-modal large model that efficiently implements pixel-level reasoning without SA dependencyPicture

As shown in the framework diagram in the paper, the PixelLM architecture is very simple and consists of four main parts. The latter two are the core of PixelLM:

  1. Pre-trained CLIP-ViT visual encoder
  2. Large language model
  3. Lightweight pixel decoder
  4. Seg Codebook

The Seg codebook contains learnable tokens, which are used to encode target information at different scales of CLIP-ViT. Then, the pixel decoder generates object segmentation results based on these tokens and the image features of CLIP-ViT. Thanks to this design, PixelLM can generate high-quality segmentation results without an external segmentation model, significantly improving model efficiency.

According to the researcher's description, the tokens in the Seg codebook can be divided into L groups, each group contains N tokens, and each group corresponds to a scale from CLIP-ViT visual features.

For the input image, PixelLM extracts L scale features from the image features produced by the CLIP-ViT visual encoder. The last layer covers the global image information and will be used by LLM to understand the image content. .

The tokens of the Seg codebook will be input into the LLM together with the text instructions and the last layer of image features to produce output in the form of autoregression. The output will also include Seg codebook tokens processed by LLM, which will be input into the pixel decoder together with L scale CLIP-ViT features to produce the final segmentation result.

PixelLM, a byte multi-modal large model that efficiently implements pixel-level reasoning without SA dependencyPicture

PixelLM, a byte multi-modal large model that efficiently implements pixel-level reasoning without SA dependencyPicture

So why do we need to set each group to contain N tokens? ? The researchers explained in conjunction with the following figure:

In scenarios involving multiple targets or targets with very complex semantics, although LLM can provide detailed text responses, it may not be fully captured using only a single token. The entire content of the target semantics.

In order to enhance the model's ability in complex reasoning scenarios, the researchers introduced multiple tokens within each scale group and performed a linear fusion operation of one token. Before the token is passed to the decoder, a linear projection layer is used to merge the tokens within each group.

The following figure shows the effect when there are multiple tokens in each group. The attention map is what each token looks like after being processed by the decoder. This visualization shows that multiple tokens provide unique and complementary information, resulting in more effective segmentation output.

PixelLM, a byte multi-modal large model that efficiently implements pixel-level reasoning without SA dependencyPicture

In addition, in order to enhance the model's ability to distinguish multiple targets, PixelLM also designed an additional Target Refinement Loss.

MUSE Dataset

Although the above solutions have been proposed, in order to fully utilize the capabilities of the model, the model still requires appropriate training data. Reviewing the currently available public data sets, we found that the existing data have the following major limitations:

1) Insufficient description of object details;
2) Lack of problems with complex reasoning and multiple target numbers -The answer is correct.

In order to solve these problems, the research team used GPT-4V to build an automated data annotation pipeline, and thus generated the MUSE data set. The figure below shows an example of the prompts used when generating MUSE and the data generated.

PixelLM, a byte multi-modal large model that efficiently implements pixel-level reasoning without SA dependencyPicture

In MUSE, all instance masks are from the LVIS dataset, and additionally detailed text descriptions generated based on the image content are added. MUSE contains 246,000 question-answer pairs, and each question-answer pair involves an average of 3.7 target objects. In addition, the research team conducted an exhaustive statistical analysis of the dataset:

Category statistics: There are more than 1,000 categories in MUSE from the original LVIS dataset, and 900,000 instances with unique descriptions based on The question-answer pairs vary depending on the context. Figure (a) shows the number of instances of each category across all question-answer pairs.

Token number statistics: Figure (b) shows the distribution of the number of tokens described in the instance. Some instance descriptions contain more than 100 tokens. These descriptions are not limited to simple category names; instead, they are enriched with detailed information about each instance, including appearance, properties, and relationships to other objects, through a GPT-4V-based data generation process. The depth and diversity of information in the dataset enhances the generalization ability of the trained model, allowing it to effectively solve open domain problems.

Statistics of the number of targets: Figure (c) shows the statistics of the number of targets in each question-answer pair. The average number of targets is 3.7, and the maximum number of targets can reach 34. This number can cover most target inference scenarios for a single image.

Algorithm Evaluation

The research team evaluated the performance of PixelLM on three benchmarks, including MUSE benchmark, referring segmentation benchmark, and multi-referring segmentation benchmark. In the multi-referring segmentation benchmark, The research team requires the model to continuously segment multiple targets contained in each image in the referring segmentation benchmark in one problem.

At the same time, since PixelLM is the first model to handle complex pixel reasoning tasks involving multiple targets, the research team established four baselines to conduct comparative analysis of the models.

Three of the baselines are based on LISA, the most relevant work on PixelLM, including:

1) Original LISA;

2) LISA_rec: First enter the question into LLAVA-13B to Get the target's text reply, and then use LISA to split the text;

3) LISA_aug: directly add MUSE to the LISA training data.

4) The other is SEEM, a general segmentation model that does not use LLM.

PixelLM, a byte multi-modal large model that efficiently implements pixel-level reasoning without SA dependencyPicture

In most indicators of the three benchmarks, PixelLM's performance is better than other methods, and because PixelLM does not rely on SAM, Its TFLOPs are far lower than models of the same size.

Interested friends can follow the wave first and wait for the code to be open source~

Reference link:
[1]https://www.php.cn/ link/9271858951e6fe9504d1f05ae8576001
[2]https://www.php.cn/link/f1686b4badcf28d33ed632036c7ab0b8

The above is the detailed content of PixelLM, a byte multi-modal large model that efficiently implements pixel-level reasoning without SA dependency. For more information, please follow other related articles on the PHP Chinese website!

Statement
This article is reproduced at:51CTO.COM. If there is any infringement, please contact admin@php.cn delete
A Comprehensive Guide to ExtrapolationA Comprehensive Guide to ExtrapolationApr 15, 2025 am 11:38 AM

Introduction Suppose there is a farmer who daily observes the progress of crops in several weeks. He looks at the growth rates and begins to ponder about how much more taller his plants could grow in another few weeks. From th

The Rise Of Soft AI And What It Means For Businesses TodayThe Rise Of Soft AI And What It Means For Businesses TodayApr 15, 2025 am 11:36 AM

Soft AI — defined as AI systems designed to perform specific, narrow tasks using approximate reasoning, pattern recognition, and flexible decision-making — seeks to mimic human-like thinking by embracing ambiguity. But what does this mean for busine

Evolving Security Frameworks For The AI FrontierEvolving Security Frameworks For The AI FrontierApr 15, 2025 am 11:34 AM

The answer is clear—just as cloud computing required a shift toward cloud-native security tools, AI demands a new breed of security solutions designed specifically for AI's unique needs. The Rise of Cloud Computing and Security Lessons Learned In th

3 Ways Generative AI Amplifies Entrepreneurs: Beware Of Averages!3 Ways Generative AI Amplifies Entrepreneurs: Beware Of Averages!Apr 15, 2025 am 11:33 AM

Entrepreneurs and using AI and Generative AI to make their businesses better. At the same time, it is important to remember generative AI, like all technologies, is an amplifier – making the good great and the mediocre, worse. A rigorous 2024 study o

New Short Course on Embedding Models by Andrew NgNew Short Course on Embedding Models by Andrew NgApr 15, 2025 am 11:32 AM

Unlock the Power of Embedding Models: A Deep Dive into Andrew Ng's New Course Imagine a future where machines understand and respond to your questions with perfect accuracy. This isn't science fiction; thanks to advancements in AI, it's becoming a r

Is Hallucination in Large Language Models (LLMs) Inevitable?Is Hallucination in Large Language Models (LLMs) Inevitable?Apr 15, 2025 am 11:31 AM

Large Language Models (LLMs) and the Inevitable Problem of Hallucinations You've likely used AI models like ChatGPT, Claude, and Gemini. These are all examples of Large Language Models (LLMs), powerful AI systems trained on massive text datasets to

The 60% Problem — How AI Search Is Draining Your TrafficThe 60% Problem — How AI Search Is Draining Your TrafficApr 15, 2025 am 11:28 AM

Recent research has shown that AI Overviews can cause a whopping 15-64% decline in organic traffic, based on industry and search type. This radical change is causing marketers to reconsider their whole strategy regarding digital visibility. The New

MIT Media Lab To Put Human Flourishing At The Heart Of AI R&DMIT Media Lab To Put Human Flourishing At The Heart Of AI R&DApr 15, 2025 am 11:26 AM

A recent report from Elon University’s Imagining The Digital Future Center surveyed nearly 300 global technology experts. The resulting report, ‘Being Human in 2035’, concluded that most are concerned that the deepening adoption of AI systems over t

See all articles

Hot AI Tools

Undresser.AI Undress

Undresser.AI Undress

AI-powered app for creating realistic nude photos

AI Clothes Remover

AI Clothes Remover

Online AI tool for removing clothes from photos.

Undress AI Tool

Undress AI Tool

Undress images for free

Clothoff.io

Clothoff.io

AI clothes remover

AI Hentai Generator

AI Hentai Generator

Generate AI Hentai for free.

Hot Article

R.E.P.O. Energy Crystals Explained and What They Do (Yellow Crystal)
4 weeks agoBy尊渡假赌尊渡假赌尊渡假赌
R.E.P.O. Best Graphic Settings
4 weeks agoBy尊渡假赌尊渡假赌尊渡假赌
R.E.P.O. How to Fix Audio if You Can't Hear Anyone
4 weeks agoBy尊渡假赌尊渡假赌尊渡假赌
WWE 2K25: How To Unlock Everything In MyRise
1 months agoBy尊渡假赌尊渡假赌尊渡假赌

Hot Tools

MinGW - Minimalist GNU for Windows

MinGW - Minimalist GNU for Windows

This project is in the process of being migrated to osdn.net/projects/mingw, you can continue to follow us there. MinGW: A native Windows port of the GNU Compiler Collection (GCC), freely distributable import libraries and header files for building native Windows applications; includes extensions to the MSVC runtime to support C99 functionality. All MinGW software can run on 64-bit Windows platforms.

mPDF

mPDF

mPDF is a PHP library that can generate PDF files from UTF-8 encoded HTML. The original author, Ian Back, wrote mPDF to output PDF files "on the fly" from his website and handle different languages. It is slower than original scripts like HTML2FPDF and produces larger files when using Unicode fonts, but supports CSS styles etc. and has a lot of enhancements. Supports almost all languages, including RTL (Arabic and Hebrew) and CJK (Chinese, Japanese and Korean). Supports nested block-level elements (such as P, DIV),

Safe Exam Browser

Safe Exam Browser

Safe Exam Browser is a secure browser environment for taking online exams securely. This software turns any computer into a secure workstation. It controls access to any utility and prevents students from using unauthorized resources.

SublimeText3 English version

SublimeText3 English version

Recommended: Win version, supports code prompts!

SublimeText3 Mac version

SublimeText3 Mac version

God-level code editing software (SublimeText3)