search
HomeTechnology peripheralsAIICML 2024 | Signal representation is exponentially stronger, memory saving exceeds 35%, quantum implicit representation network is coming

ICML 2024 | 信号表征指数级强、内存节省超35%,量子隐式表征网络来了
The AIxiv column is a column where academic and technical content is published on this site. In the past few years, the AIxiv column of this site has received more than 2,000 reports, covering top laboratories from major universities and companies around the world, effectively promoting academic exchanges and dissemination. If you have excellent work that you want to share, please feel free to contribute or contact us for reporting. Submission email: liyazhou@jiqizhixin.com; zhaoyunfeng@jiqizhixin.com

The authors of this article are Professor Zhang Peng of the Department of Intelligence and Computing of Tianjin University, his master's student Zhao Jiaming, and doctoral students Qiao Wenbo and Gao Jue. This research work was funded by the National Natural Science Foundation of China and Tianjin University-China Science and Technology Wenge Joint Laboratory.

ICML 2024 | 信号表征指数级强、内存节省超35%,量子隐式表征网络来了

  • Paper title: Quantum Implicit Neural Representations
  • Paper authors: Jiaming Zhao, Wenbo Qiao, Peng Zhang*, Hui Gao
  • Paper link: https://arxiv.org/abs /2406.03873

Introduction

The Tianjin University Quantum Intelligence and Language Understanding team innovatively introduced quantum computing into the field of implicit neural representation and proposed the quantum implicit representation network (Q uantum Implicit Representation Network, QIREN). Compared with the classic neural network method, this method has exponentially stronger signal representation capabilities in theory. Experimental results also confirm that QIREN indeed exhibits excellent performance beyond the SOTA model on signal representation tasks, with fitting errors reduced by up to 35% with fewer parameters. Figure 1 shows the core idea and main conclusions of this paper. Relevant papers have been accepted by ICML 2024, one of the most authoritative conferences in the field of machine learning. ICML 2024 | 信号表征指数级强、内存节省超35%,量子隐式表征网络来了
                                                                                                                                                                                                                                       Figure 1. Classical Fourier neural network and quantum Fourier neural network.

In recent years, implicit neural representation has attracted widespread attention as an emerging signal representation method. Implicit neural representations have a number of unique advantages over traditional discrete grid representations, such as images represented by a grid of pixels. First, it has the capability of "infinite resolution" and can sample at any spatial resolution. Secondly, implicit neural representation has excellent storage space saving and provides convenience for data storage. Because of these unique advantages, implicit neural representation has quickly become a mainstream paradigm for representing signals such as images, objects, and 3D scenes. Most early research on implicit neural representations was built on ReLU-based multilayer perceptrons (MLPs). However, it is difficult for ReLU-based MLP to accurately model the high-frequency part of the signal, as shown in Figure 2. Recent research has begun to explore the use of Fourier neural networks (FNN) to overcome this limitation. However, in the face of increasingly complex fitting tasks in real-world applications, classical Fourier neural networks also require more and more training parameters, which increases the demand for computing resources. The quantum implicit neural representation proposed in this article takes advantage of quantum advantages to reduce parameters and computational consumption. This solution can bring new inspiration to the field of implicit neural representation and even machine learning.像 Figure 2. Different frequency components of the different frequency components (top) and RELU -based MLP -based images (bottom)

ICML 2024 | 信号表征指数级强、内存节省超35%,量子隐式表征网络来了

Model

                                                                                                                                                                        Figure 3. Model architecture

The overall architecture of QIREN is shown in Figure 3. Consists of N mixed layers and linear layers at the end. The model takes coordinates as input and outputs signal values. The data initially enters the mixed layer, starting with the Linear layer and the BatchNorm layer, resulting in:

and is then fed into the data re-upload quantum circuit QC. In Figure 2 (b) and (c), we give the specific implementation of the parameter layer and coding layer quantum circuits. The parameter layer consists of K stacked blocks. Each block contains a spin gate applied to each qubit, as well as CNOT gates connected in a round-robin fashion. The coding layer applies gates on each qubit. Finally, we measure the expected value of a quantum state relative to an observable. The output of a quantum circuit is given by:

ICML 2024 | 信号表征指数级强、内存节省超35%,量子隐式表征网络来了

where O represents any observable. The output of the nth blending layer will be used as the input of the (n+1)th layer. Finally, we add a linear layer to receive and output. We use mean square error (MSE) as the loss function to train the model:

ICML 2024 | 信号表征指数级强、内存节省超35%,量子隐式表征网络来了

Model theoretical analysis

In some previous studies, the mathematical properties of the data re-upload quantum circuit have been revealed. In essence, the data re-upload quantum circuit is to fit the objective function in the form of a Fourier series. However, previous work only explored multi-layer single-qubit circuits or single-layer multi-qubit circuits, and did not compare with classical methods and did not find the advantages of data re-uploading quantum circuits. We extend our research to multi-layer multi-qubit circuits. In addition, we have proven that in the field of implicit neural representation, the hybrid quantum neural network QIREN, which uses data reuploading quantum circuits as its core component, has exponential advantages over classical methods. We analyzed the role of the quantum layer and the classical layer in QIREN and summarized it into the following three points:

1. Under optimal conditions, the ability of data re-uploading quantum circuits to represent Fourier series increases with the circuit's The size grows exponentially.

See sections 4.2 and 4.3 of the paper for specific derivation.

2. The function of the linear layer is to further expand the spectrum and adjust the frequency, thereby improving the fitting performance.

Applying a linear layer before uploading data to a quantum circuit is equivalent to adjusting the eigenvalues ​​of the coding layer Hamiltonian, ultimately affecting the spectrum. This approach has two advantages. First, it can make the spectrum larger. Some redundant terms are produced in the spectrum when encoding only with gates. This redundancy can be reduced by using linear layers. Second, it enables the coverage of the spectrum to be adjusted, aiming to cover frequencies with larger coefficients that are more important. Therefore, adding a linear layer can further improve the fitting performance of QIREN.

3. The role of the Batchnorm layer is to accelerate the convergence of the quantum model.

In feedforward neural networks, data usually passes through the BatchNorm layer before the activation function, which effectively prevents the vanishing gradient problem. Similarly, in QIREN, quantum circuits replace the activation function and play a role in providing nonlinearity (the quantum circuit itself is linear, but the process of uploading classical data to the quantum circuit is nonlinear). Therefore, we added a BatchNorm layer here with the purpose of stabilizing and accelerating the convergence of the model.

Experimental results

We verified QIREN’s superior performance in representing signals, especially high-frequency signals, through image representation and sound representation tasks. The experimental results are shown in Table 1. QIREN and SIREN showed similar performance on the sound representation task. Although the performance of the two models seems to be comparable, it is worth emphasizing that our model achieves 35.1% memory savings with the fewest parameters, and the convergence of SIREN requires setting appropriate hyperparameters, while our model does not This kind of restriction. We then analyzed the model output from a frequency perspective. We visualize the spectrum of the model output in Figure 4 . It is obvious that the low-frequency distributions output by the model are close to the real situation. However, when it comes to high-frequency distributions, both QIREN and SIREN fit well, followed by ReLU-based MLP with random Fourier features (RFF). ReLU-based and Tanh-based MLPs even lack the high-frequency part of the signal.

ICML 2024 | 信号表征指数级强、内存节省超35%,量子隐式表征网络来了

Table 1. MSE() of the model on signal representation and image super-resolution tasks. Models considered SOTA are marked *. params represents the amount of model parameters, and mem represents the memory saved by the model compared with discrete grid representation.任 Figure 4. Sound indicating the spectrum of the model output in the task. Qiren achieves the best performance in the image representation task. Compared with the SOTA model, the error is The maximum reduction was 34.8%. To further explore the signal representation capabilities of the model, we use filters to separate the high-frequency and low-frequency components of its output and compare the fitting errors of these two components respectively, with the results shown in Figure 5. QIREN consistently achieves the lowest errors when fitting high- and low-frequency components.

                                                                                                                                                                                                                                                             Figure 5. Relative error of each model compared to Tanh-based MLP. Shaded areas represent low-frequency errors, while unshaded areas represent high-frequency errors.

Latest research introduces a breakthrough framework to extend implicit neural representations to image generation. More specifically, the framework utilizes a hypernetwork taking random distributions as input to generate parameters that implicitly characterize the network. Subsequently, these generated parameters are assigned to the implicit representation network. Finally, the implicit representation network generates images taking coordinates as input. An adversarial approach is employed to ensure that the generated images are consistent with our desired results. In this task, we adopt such a framework and build on StyleGAN2.

The experimental results are shown in Table 2. We also further explore some exciting features of the QIREN generator, as shown in Figures 6 and 7. F Table 2. FID score of the model on FFHQ and CELEBA-HQ datasets.

ICML 2024 | 信号表征指数级强、内存节省超35%,量子隐式表征网络来了

                                                                                                                                                                                                                                                                                                                                                                                 Figure 7. Meaningful image space interpolation

ICML 2024 | 信号表征指数级强、内存节省超35%,量子隐式表征网络来了

summary

This work not only integrates quantum advantages into implicit neural representation, but also opens up a promising application direction for quantum neural networks - implicit neural representation. It is worth emphasizing that implicit neural representations have many other potential applications, such as representing scenes or 3D objects, time series prediction, and solving differential equations. For a large class of tasks that model continuous signals, we can consider introducing implicit representation networks as a basic component. Based on the theoretical and experimental foundations of this paper, we can extend QIREN to these applications in future work, and QIREN is expected to produce better results with fewer parameters in these fields. At the same time, we found a suitable application scenario for quantum machine learning. Thereby promoting further practical and innovative research within the quantum machine learning community. ICML 2024 | 信号表征指数级强、内存节省超35%,量子隐式表征网络来了

The above is the detailed content of ICML 2024 | Signal representation is exponentially stronger, memory saving exceeds 35%, quantum implicit representation network is coming. For more information, please follow other related articles on the PHP Chinese website!

Statement
The content of this article is voluntarily contributed by netizens, and the copyright belongs to the original author. This site does not assume corresponding legal responsibility. If you find any content suspected of plagiarism or infringement, please contact admin@php.cn
7 Powerful AI Prompts Every Project Manager Needs To Master Now7 Powerful AI Prompts Every Project Manager Needs To Master NowMay 08, 2025 am 11:39 AM

Generative AI, exemplified by chatbots like ChatGPT, offers project managers powerful tools to streamline workflows and ensure projects stay on schedule and within budget. However, effective use hinges on crafting the right prompts. Precise, detail

Defining The Ill-Defined Meaning Of Elusive AGI Via The Helpful Assistance Of AI ItselfDefining The Ill-Defined Meaning Of Elusive AGI Via The Helpful Assistance Of AI ItselfMay 08, 2025 am 11:37 AM

The challenge of defining Artificial General Intelligence (AGI) is significant. Claims of AGI progress often lack a clear benchmark, with definitions tailored to fit pre-determined research directions. This article explores a novel approach to defin

IBM Think 2025 Showcases Watsonx.data's Role In Generative AIIBM Think 2025 Showcases Watsonx.data's Role In Generative AIMay 08, 2025 am 11:32 AM

IBM Watsonx.data: Streamlining the Enterprise AI Data Stack IBM positions watsonx.data as a pivotal platform for enterprises aiming to accelerate the delivery of precise and scalable generative AI solutions. This is achieved by simplifying the compl

The Rise of the Humanoid Robotic Machines Is Nearing.The Rise of the Humanoid Robotic Machines Is Nearing.May 08, 2025 am 11:29 AM

The rapid advancements in robotics, fueled by breakthroughs in AI and materials science, are poised to usher in a new era of humanoid robots. For years, industrial automation has been the primary focus, but the capabilities of robots are rapidly exp

Netflix Revamps Interface — Debuting AI Search Tools And TikTok-Like DesignNetflix Revamps Interface — Debuting AI Search Tools And TikTok-Like DesignMay 08, 2025 am 11:25 AM

The biggest update of Netflix interface in a decade: smarter, more personalized, embracing diverse content Netflix announced its largest revamp of its user interface in a decade, not only a new look, but also adds more information about each show, and introduces smarter AI search tools that can understand vague concepts such as "ambient" and more flexible structures to better demonstrate the company's interest in emerging video games, live events, sports events and other new types of content. To keep up with the trend, the new vertical video component on mobile will make it easier for fans to scroll through trailers and clips, watch the full show or share content with others. This reminds you of the infinite scrolling and very successful short video website Ti

Long Before AGI: Three AI Milestones That Will Challenge YouLong Before AGI: Three AI Milestones That Will Challenge YouMay 08, 2025 am 11:24 AM

The growing discussion of general intelligence (AGI) in artificial intelligence has prompted many to think about what happens when artificial intelligence surpasses human intelligence. Whether this moment is close or far away depends on who you ask, but I don’t think it’s the most important milestone we should focus on. Which earlier AI milestones will affect everyone? What milestones have been achieved? Here are three things I think have happened. Artificial intelligence surpasses human weaknesses In the 2022 movie "Social Dilemma", Tristan Harris of the Center for Humane Technology pointed out that artificial intelligence has surpassed human weaknesses. What does this mean? This means that artificial intelligence has been able to use humans

Venkat Achanta On TransUnion's Platform Transformation And AI AmbitionVenkat Achanta On TransUnion's Platform Transformation And AI AmbitionMay 08, 2025 am 11:23 AM

TransUnion's CTO, Ranganath Achanta, spearheaded a significant technological transformation since joining the company following its Neustar acquisition in late 2021. His leadership of over 7,000 associates across various departments has focused on u

When Trust In AI Leaps Up, Productivity FollowsWhen Trust In AI Leaps Up, Productivity FollowsMay 08, 2025 am 11:11 AM

Building trust is paramount for successful AI adoption in business. This is especially true given the human element within business processes. Employees, like anyone else, harbor concerns about AI and its implementation. Deloitte researchers are sc

See all articles

Hot AI Tools

Undresser.AI Undress

Undresser.AI Undress

AI-powered app for creating realistic nude photos

AI Clothes Remover

AI Clothes Remover

Online AI tool for removing clothes from photos.

Undress AI Tool

Undress AI Tool

Undress images for free

Clothoff.io

Clothoff.io

AI clothes remover

Video Face Swap

Video Face Swap

Swap faces in any video effortlessly with our completely free AI face swap tool!

Hot Tools

SAP NetWeaver Server Adapter for Eclipse

SAP NetWeaver Server Adapter for Eclipse

Integrate Eclipse with SAP NetWeaver application server.

mPDF

mPDF

mPDF is a PHP library that can generate PDF files from UTF-8 encoded HTML. The original author, Ian Back, wrote mPDF to output PDF files "on the fly" from his website and handle different languages. It is slower than original scripts like HTML2FPDF and produces larger files when using Unicode fonts, but supports CSS styles etc. and has a lot of enhancements. Supports almost all languages, including RTL (Arabic and Hebrew) and CJK (Chinese, Japanese and Korean). Supports nested block-level elements (such as P, DIV),

EditPlus Chinese cracked version

EditPlus Chinese cracked version

Small size, syntax highlighting, does not support code prompt function

SublimeText3 English version

SublimeText3 English version

Recommended: Win version, supports code prompts!

Zend Studio 13.0.1

Zend Studio 13.0.1

Powerful PHP integrated development environment