Home  >  Article  >  Technology peripherals  >  7262 papers were submitted, ICLR 2024 became a hit, and two domestic papers were nominated for outstanding papers.

7262 papers were submitted, ICLR 2024 became a hit, and two domestic papers were nominated for outstanding papers.

WBOY
WBOYforward
2024-05-08 20:34:241062browse

A total of 5 outstanding paper awards and 11 honorable mentions were selected this year.

ICLR stands for International Conference on Learning Representations. This year is the 12th conference, held in Vienna, Austria, from May 7th to 11th.

In the machine learning community, ICLR is a relatively "young" top academic conference. It is hosted by deep learning giants and Turing Award winners Yoshua Bengio and Yann LeCun. It just held its first session in 2013. However, ICLR quickly gained wide recognition from academic researchers and is considered the top academic conference on deep learning.

This conference received a total of 7262 submitted papers and accepted 2260 papers. The overall acceptance rate was about 31%, which was the same as last year (31.8%). In addition, the proportion of Spotlights papers is 5% and the proportion of Oral papers is 1.2%.

7262篇提交,ICLR 2024爆火,两篇国内论文获杰出论文提名

7262篇提交,ICLR 2024爆火,两篇国内论文获杰出论文提名

Compared with previous years, whether it is the number of participants or the number of paper submissions, the popularity of ICLR can be said to have greatly improved. .

7262篇提交,ICLR 2024爆火,两篇国内论文获杰出论文提名

##                                                                                                                                                                                                                                           Previous ICLR paper data chart

Among the award-winning papers announced recently, the conference selected 5 Outstanding Paper Awards and 11 Honorable Mention Awards.

5 Outstanding Paper Awards

Outstanding Paper winners

Paper: Generalization in diffusion models arises from geometry -adaptive harmonic representations

7262篇提交,ICLR 2024爆火,两篇国内论文获杰出论文提名

  • Paper address: https://openreview.net/pdf?id=ANvmVS2Yr0

  • Institution: New York University, Collège de France

  • Author: Zahra Kadkhodaie, Florentin Guth, Eero P. Simoncelli, Stéphane Mallat

This article provides an important in-depth analysis of the generalization and memory aspects of the image diffusion model. The authors empirically study when an image generation model switches from memory input to generalization mode, and connect it with the idea of ​​harmonic analysis through geometrically adaptive harmonic representation, further explaining this phenomenon from the perspective of architectural induction bias. This article covers a key missing piece in our understanding of generative models of vision and has great implications for future research.

Thesis: Learning Interactive Real-World Simulators

7262篇提交,ICLR 2024爆火,两篇国内论文获杰出论文提名

  • Thesis address: https://openreview. net/forum?id=sFyTZEqmUY

  • Institutions: UC Berkeley, Google DeepMind, MIT, University of Alberta

  • Author: Sherry Yang, Yilun Du, Kamyar Ghasemipour, Jonathan Tompson, Leslie Kaelbling, Dale Schuurmans, Pieter Abbeel

Aggregating data across multiple sources to train base models for robots is a long-term goal. Since different robots have different sensorimotor interfaces, this poses significant challenges to training across large-scale datasets.

UniSim is an important step in this direction and an engineering feat that leverages a unified interface based on textual descriptions of visual perception and control to aggregate data and Train robot simulators by leveraging recent developments in vision and language.

In summary, this article explores UniSim, a general-purpose simulator that learns real-world interactions through generative models, and takes the first step in building a general-purpose simulator. For example, UniSim can simulate how humans and agents interact with the world by simulating high-level instructions such as "open a drawer" and the visual results of low-level instructions.

This paper combines large amounts of data (including Internet text-image pairs, rich data from navigation, human activities, robot actions, etc., and data from simulations and renderings) into a conditional video generation framework. Then by carefully orchestrating rich data along different axes, this paper shows that UniSim can successfully merge experience from different axes of data and generalize beyond the data to enable rich interactions through fine-grained motion control of static scenes and objects.

As shown in Figure 3 below, UniSim can simulate a series of rich actions, such as washing hands, taking bowls, cutting carrots, and drying hands in a kitchen scene; the upper right of Figure 3 shows pressing different switches; the bottom of Figure 3 are two navigation scenarios.

7262篇提交,ICLR 2024爆火,两篇国内论文获杰出论文提名

##                                                                                                                                                                                               to correspond to the navigation scene in the lower right corner of Figure 3

7262篇提交,ICLR 2024爆火,两篇国内论文获杰出论文提名

## The navigation scenario below 3 right

Thesis: Never Train from Scratch: Fair Comparison of Long-sequence Models Requires Data-Driven Priors

7262篇提交,ICLR 2024爆火,两篇国内论文获杰出论文提名

  • Paper address: https://openreview.net/forum?id=PdaPky8MUn

  • Institution: Tel Aviv University, IBM

  • Authors: Ido Amos, Jonathan Berant, Ankit Gupta

This paper delves into a recently proposed State space models and transformer architectures have the ability to model long-term sequence dependencies.

Surprisingly, the authors found that training a transformer model from scratch results in its performance being underestimated, and that significant performance gains can be achieved with pre-training and fine-tuning settings. The paper excels in its focus on simplicity and systematic insights.

Paper: Protein Discovery with Discrete Walk-Jump Sampling

7262篇提交,ICLR 2024爆火,两篇国内论文获杰出论文提名

  • Paper address: https:// openreview.net/forum?id=zMPHKOmQNb

  • Institution: Genentech, New York University

  • Author: Nathan C. Frey, Dan Berenberg, Karina Zadorozhny, Joseph Kleinhenz, Julien Lafrance-Vanasse, Isidro Hotzel, Yan Wu, Stephen Ra, Richard Bonneau, Kyunghyun Cho, Andreas Loukas, Vladimir Gligorijevic, Saeed Saremi

This paper solves Solving the problem of sequence-based antibody design, this is a timely and important application of protein sequence generation models.

To this end, the author introduces an innovative and effective new modeling method specifically targeted at the problem of processing discrete protein sequence data. In addition to validating the method in silico, the authors performed extensive wet laboratory experiments to measure in vitro antibody binding affinities, demonstrating the effectiveness of their generated method.

Paper: Vision Transformers Need Registers

7262篇提交,ICLR 2024爆火,两篇国内论文获杰出论文提名

    ##Paper address: https://openreview.net/ forum?id=2dnO3LLiJ1
  • Institution: Meta et al
  • Author: Timothée Darcet, Maxime Oquab, Julien Mairal, Piotr Bojanowski
  • This paper identifies artifacts in the feature map of a vision transformer network, which are characterized by high-norm tokens in low-information background regions.

The authors propose key assumptions for why this phenomenon occurs and provide a simple yet elegant solution using additional register tokens to account for these traces, thereby enhancing the model's performance on a variety of tasks. Insights gained from this work could also impact other application areas.

This paper is well written and provides a good example of conducting research: "Identify the problem, understand why it occurs, and then propose a solution."

11 Honorable Mentions

In addition to 5 outstanding papers, ICLR 2024 also selected 11 honorable mentions.

Paper: Amortizing intractable inference in large language models

    Institution: University of Montreal, University of Oxford
  • Authors: Edward J Hu, Moksh Jain, Eric Elmoznino, Younesse Kaddar, Guillaume Lajoie, Yoshua Bengio, Nikolay Malkin
  • Paper address: https://openreview.net/forum? id=Ouj6p4ca60
  • This paper proposes a promising alternative to autoregressive decoding in large language models from a Bayesian inference perspective, which may inspire subsequent research .

Paper: Approximating Nash Equilibria in Normal-Form Games via Stochastic Optimization

##Institution: DeepMind
  • Authors: Ian Gemp, Luke Marris, Georgios Piliouras
  • Paper address: https://openreview.net/forum?id=cc8h3I3V4E
  • This is a very clearly written paper that contributes significantly to solving the important problem of developing efficient and scalable Nash solvers.
  • Paper: Beyond Weisfeiler-Lehman: A Quantitative Framework for GNN Expressiveness

    • ##Institution: Peking University, Beijing Zhiyuan Artificial Intelligence Research Institute

    • Author: Zhang Bohang Gai Jingchu Du Yiheng Ye Qiwei Hedi Wang Liwei

    • ## Paper address: https://openreview.net/forum?id=HSKaGOi7Ar
    • The expressiveness of GNNs is an important topic, and current solutions still have significant limitations. The author proposes a new expressivity theory based on homomorphic counting.

    Paper: Flow Matching on General Geometries

      Institution: Meta
    • Author: Ricky T. Q. Chen, Yaron Lipman
    • Paper address: https://openreview.net/forum?id=g7ohDlTITL
    • This paper discusses the application of general geometry This is a challenging but important problem of generative modeling on manifolds, and a practical and efficient algorithm is proposed. The paper is excellently presented and fully experimentally validated on a wide range of tasks.

    Paper: Is ImageNet worth 1 video? Learning strong image encoders from 1 long unlabelled video

      Institution: University of Central Florida, Google DeepMind, University of Amsterdam, etc.
    • Authors: Shashanka Venkataramanan, Mamshad Nayeem Rizve, Joao Carreira, Yuki M Asano, Yannis Avrithis
    • Paper address: https: //openreview.net/forum?id=Yen1lGns2o
    • This paper proposes a novel self-supervised image pre-training method by learning from continuous videos. This paper contributes both a new type of data and a method for learning from new data.

    Thesis: Meta Continual Learning Revisited: Implicitly Enhancing Online Hessian Approximation via Variance Reduction

      Institutions: City University of Hong Kong, Tencent AI Laboratory, Xi'an Jiaotong University, etc.
    • Authors: Yichen Wu, Long-Kai Huang, Renzhen Wang, Deyu Meng, Wei Ying (Ying Wei)
    • Paper address: https://openreview.net/forum?id=TpD2aG1h0D
    • The author proposes a new meta-continuous learning variance reduction method. The method performs well and not only has practical impact but is also supported by regret analysis.

    Paper: Model Tells You What to Discard: Adaptive KV Cache Compression for LLMs

    ##Institution: University of Illinois at Urbana-Champaign, Microsoft
    • Authors: Suyu Ge, Yunan Zhang, Liyuan Liu, Minjia Zhang, Jiawei Han, Jianfeng Gao
    • Paper address: https:/ /openreview.net/forum?id=uNrFpDPMyo
    • This article aims at the KV cache compression problem (this problem has a great impact on Transformer-based LLM) and uses a simple idea to reduce memory. And it can be deployed without extensive resource-intensive fine-tuning or retraining. This method is very simple and has proven to be very effective.

    Paper: Proving Test Set Contamination in Black-Box Language Models

    ##Institution: Stanford University, Columbia University

    • Authors: Yonatan Oren, Nicole Meister, Niladri S. Chatterji, Faisal Ladhak, Tatsunori Hashimoto

    • Paper address: https://openreview.net/forum?id= KS8mIvetg2

    • This paper uses a simple and elegant method for testing whether a supervised learning dataset has been included in the training of a large language model.

    Paper: Robust agents learn causal world models

    Institution: Google DeepMind

    • Author: Jonathan Richens, Tom Everitt

    • Paper address: https://openreview.net/forum?id=pOoKI3ouv1

    • This paper was laid down Considerable progress has been made in the theoretical foundations for understanding the role of causal reasoning in an agent's ability to generalize to new domains, with implications for a range of related fields.

    Paper: The mechanistic basis of data dependence and abrupt learning in an in-context classification task

    Institution: Princeton University, Harvard University, etc.

    • Author: Gautam Reddy

    • Paper address: https://openreview.net/forum?id=aN4Jf6Cx69

    • This is a timely and extremely systematic study that explores the mechanisms between in-context learning and in-weight learning as we begin to understand these phenomena.

    Thesis: Towards a statistical theory of data selection under weak supervision

    Institution: Granica Computing

    • Authors: Germain Kolossov, Andrea Montanari, Pulkit Tandon

    • Paper address: https://openreview.net/forum?id=HhfcNgQn6p

    • This paper establishes a statistical foundation for data subset selection and identifies the shortcomings of popular data selection methods.

    Reference link: https://blog.iclr.cc/2024/05/06/iclr-2024-outstanding-paper-awards/

The above is the detailed content of 7262 papers were submitted, ICLR 2024 became a hit, and two domestic papers were nominated for outstanding papers.. For more information, please follow other related articles on the PHP Chinese website!

Statement:
This article is reproduced at:jiqizhixin.com. If there is any infringement, please contact admin@php.cn delete