NeurIPS Highlights | ML Contests (2024)

A few weeks before the end of last year, NeurIPS wrapped up its week-long 2023 programme in New Orleans.It was the biggest NeurIPS yet in terms of in-person attendees (13,307)and accepted papers (3,540), and possibly the largest academic AI conference ever1.

Given its scale, it’s an impossible conference to summarise. For some fragments from the invited talks, as well assome of the orals,the exhibit hall, poster sessions, tutorials, workshops, and competitions, see our daily blogs.

Having said that, in this post we attempt to take a step back and highlight themes from the conference that stood outto us — as well as what it might say about AI trends in 2024.

Plenty of room at the bottom

One of the main themes throughout the conference sessions was that many current cutting-edge models are too big.Not that they’re cumbersome to manage, expensive to run, difficult to train, or take up a lot of memory —but that they’re bigger than they need to be, and that equivalent performance can be achieved with smaller models.

Throughout this NeurIPS, there were researchers presenting significant leaps forward on the efficiency front —whether through mathematically-equivalent algorithmic improvements on the implementation of attention, alternativesto attention which improve asymptotic scaling, clever quantisation techniques which reduce memory usage, or morethoughtful data filtering which improves performance. For just a few highlights of these, see our summary of theEfficiency Oral Session,Chris Re’s Invited Talk,and the Beyond Scaling panel.

This thinking was also validated by models released during NeurIPS, such asMixtral andPhi. Both of thesesmall-ish models show benchmark performance that’s equivalent and sometimes superior to larger models.

To quote Björn Ommer (quoting Richard Feynman) during hisinvited talk on Scaling and Generative AI“There’s plenty of room at the bottom”.2

Flavour of the week: LLMs and Diffusion Models

This was the first NeurIPS with submission deadlines after ChatGPT and Stable Diffusion’s release3, and asexpected there was a lot of attention on both LLMS and on Diffusion Models. Many of the best-attended sessions focusedon topics related to these — such as the tutorial onLatent Diffusion Models, and several ofthe Invited Talks.

Fittingly, the Test of Time award went to a paper which set up a lot of the ingredientsfor the LLM revolution (Jeff Dean and Greg Corrado presented; Ilya Sutskever and the other co-authors weren’t there tocollect it in person).

The exhibition hall featured lots of companies with specialised their solutions for effectively pre-training,fine-tuning, and serving LLMs — alongside the usual large tech firms, quant traders, and MLOps solutions.

Better data, please

Alongside the growth of the relatively youngdatasets and benchmarks track,data continues to be a focus at NeurIPS. Many of the speakers referenced the importance of a deep understanding oftraining and evaluation data, with the emphasis shifting from quantity to quality.

One of the runner-ups to the outstanding paper award,Scaling Data-Constrained Language Modelsexamined the effects of multi-epoch training in LLMs, as well as presenting several other interesting empirical resultsaround training data for LLMs.

In one of the conference competitions, the LLM Efficiency Challenge(where participants maximised fine-tuned model performance given only 24h and a single GPU),the winners attributed much of their edge over others to selecting the right subset of training data.

The tutorial on Data-Centric AI made a compelling case for data-centriclearning (as opposed to model-centric learning), and presented several useful resources to help use this approach inbuilding more reliable and responsible AI, including a tool for monitoring performance on subsets of data during modeltraining.

Degrees of openness

In the panel on Beyond Scaling, Percy Liang pointed out that thinking of afoundation model4 as “open” or “not open” isn’t a veryuseful distinction, and that it’s more useful to think about properties such as a model being open-weights,open-training-data, or open-training-code.

See Also
Registration

Many recent models, like Meta’s Llama/Llama2, Microsoft’s Phi,and Mistral’s models, are open-weights — in the sense that anyone can download the model weights for their owninference or fine-tuning. But this doesn’t tell us how the model was trained, or on what data5. And withoutknowing those two things, it’s hard to really know how good a model is, or how to get the most out of it.

Organisations that the panel highlighted for releasing models which are open in more respects than justweights wereEleuther,HuggingFace,BigScience,AI2, andLLM360.

Benchmarking and Goodhart’s Law

As the community shifts to using more foundation models with varying degrees of openness, the benchmarking norms thatwere designed for open models (or closed models fully developed within the organisation using them) are no longersufficient.

One of the key difficulties is: how can we know that a model wasn’t trained on the benchmark dataset it’s beingevaluated on?

Even if a model wasn’t trained directly on a benchmark dataset, over time any publiclyavailable benchmark dataset will leak into other data, especially when web-scraped training data is so pervasive.Without access to the training data, evaluators are unable to examine similarity between the eval/benchmark samples andthe training corpus. This problem is exacerbated by the fact that models are marketed on their benchmark performance,creating incentives that aren’t conducive to thorough cleaning of training data — a clear example of Goodhart’sLaw.

This is an open challenge, though the competitions track has been dealing withthese considerations for some time.

For occasional email updates from ML Contestswith more content like this conference coverage and insights into competitive ML,subscribe to our mailing list.

It was a great NeurIPS, and left us with the feeling that there’s much more to come soon — especially in termsof democratising access to powerful and fast models. We look forward to another year of groundbreaking research!

For more on NeurIPS 2023, read our daily blogs: expo day, tutorials,day 1, day 2, day 3,and the competition track days.

  1. Our World in Data showsrecent data for some of the top conferences, aggregating both virtual and in-person attendees.NeurIPS 2020 and 2021 were fully virtual, and NeurIPS 2022 had 9,835 attendees(source: NeurIPS fact sheet).The only other conferenceslisted there with more than 13,000 attendees are IROS 2020 and ICML 2021, which were both fully virtual. It’s possiblethat there were larger AI conferences a few decades ago; data for those is not as readily available.↩︎

  2. Richard Feynman used this phrase as the title of a lecture which some see as the origin ofnanotechnology. He was referring specifically to smaller-scale mechanical manipulation down to the level of individualatoms; in the machine-learning context it refers to parameter counts or memory usage rather thanphysical dimensions. More onWikipedia.↩︎

  3. Stable Diffusion was released in August 2022, and ChatGPT in November 2022. The NeurIPS 2022 conferencetook place after this, in December 2022, but much of the agenda for that conference had been set muchearlier — with abstract and paper submission deadlines in May 2022.↩︎

  4. Foundation model: “any model that is trained on broad data (generally using self-supervision at scale)that can be adapted(e.g., fine-tuned) to a wide range of downstream tasks” (arXiv)↩︎

  5. There is a bit more info on Phi-2 training data — “Dataset size: 250B tokens, combination of NLPsynthetic data created by AOAI GPT-3.5 and filtered web data from Falcon RefinedWeb and SlimPajama, which was assessedby AOAI GPT-4” — than Llama2 — “Llama 2 was pretrained on 2 trillion tokens of data from publicly availablesources. The fine-tuning data includes publicly available instruction datasets, as well as over one million newhuman-annotated examples. Neither the pretraining nor the fine-tuning datasets include Meta user data”.(source: HuggingFace model cards)↩︎

NeurIPS Highlights | ML Contests (2024)

FAQs

Is there a word limit for NeurIPS rebuttal? ›

You can submit a rebuttal of up to 6000 characters per review, and one global rebuttal of up to 6000 reviews. These are posted by clicking the "Rebuttal" and "Author Rebuttal" buttons. You can additionally add a one-page PDF with Figures and tables.

Is NeurIPS double blind? ›

Researchers submit their papers to NeurIPS, which are then reviewed by a panel of experts in the field. The review process is double-blind, meaning that the reviewers do not know the identity of the authors, and vice versa. This helps ensure a fair evaluation of the research based solely on its merit.

Is NeurIPS a Conference or journal? ›

The Conference and Workshop on Neural Information Processing Systems (abbreviated as NeurIPS and formerly NIPS) is a machine learning and computational neuroscience conference held every December.

Do NeurIPS workshops have proceedings? ›

For example, full-length papers go to proceedings, extended abstracts go to non-proceedings. Note that NeurIPS will not itself publish proceedings for workshops: workshops will need to set up their own proceedings if desired.

What is the acceptance rate for NeurIPS? ›

This year the paper acceptance rate was 26%, making it one of the most selective machine learning conferences in the world, alongside Computer Vision and Pattern Recognition Conference (CVPR) and International Conference of Computer Vision (ICCV).

Can anyone submit to NeurIPS? ›

Anyone who plans to submit a paper as an author or a co-author will need to create (or update) their OpenReview profile by the full paper submission deadline. Your OpenReview profile can be edited by logging in and clicking on your name in https://openreview.net/.

Is NeurIPS prestigious? ›

“NeurIPS is one of the most prestigious and highly ranked conferences in machine learning, artificial intelligence, and data science,” said Makarychev, a professor of computer science at the McCormick School of Engineering.

How many accepted papers are there in NeurIPS? ›

This year's organizers received a record number of paper submissions. Of the 13,300 submitted papers that were reviewed by 968 Area Chairs, 98 senior area chairs, and 396 Ethics reviewers 3,540 were accepted after 502 papers were flagged for ethics reviews.

How many people attend NeurIPS? ›

The thirty-seventh Conference on Neural Information Processing Systems (NeurIPS 2023) took place in New Orleans from Sunday 10 to Saturday 16 December. The event was vast, with over 13,000 people in attendance at the venue, and a further 3,000 tuning in virtually.

Where will NeurIPS 2024 be held? ›

NeurIPS 2024 Meeting Dates

The Thirty-eighth annual conference is held Mon. Dec 9th through Sun the 15th, 2024 at the Vancouver Convention Center.

Are NeurIPS reviews public? ›

After decisions have been made, reviews and meta-reviews will be made public (but reviewer and SAC/AC identities will remain anonymous). Area Chair. Each area chair (AC) oversees around 20 submissions, making sure that the reviewing process goes smoothly.

Is NeurIPS workshop peer-reviewed? ›

The workshop is non-archival. Please note that we do allow the submission of recently published work. However, when selecting papers for oral presentation, preference is given to original works. Submissions will be peer-reviewed by at least 2 reviewers, in addition to an area chair.

Is ICLR better than NeurIPS? ›

In terms of content, NeurIPS has a wider scope than ICLR or ICML, covering a range of topics related to Neural Networks and their applications in areas such as Computer Vision, Natural Language Processing, and Robotics.

What is the citation style for NeurIPS? ›

Citations may be author/year or numeric, as long as you maintain internal consistency. As to the format of the references themselves, any style is acceptable as long as it is used consistently. As submission is double blind, refer to your own published work in the third person.

How do you write a good review on NeurIPS? ›

Best Practices
  • Be thoughtful. ...
  • Be fair. ...
  • Be useful. ...
  • Be specific. ...
  • Be flexible. ...
  • Be timely. ...
  • If someone pressures you into providing a positive or negative review for a submission, please notify program chairs right away (program-chairs@neurips.cc).

What is the page limit for NeurIPS? ›

Papers to be submitted to NeurIPS 2023 must be prepared according to the instructions presented here. Papers may only be up to nine pages long, including figures. Additional pages containing only acknowledgments and references are allowed.

How long is the abstract for NeurIPS? ›

Use 10 point type, with a vertical spacing (leading) of 11 points. precede the abstract. The abstract must be limited to one paragraph. ahead of the submission deadline: https://neurips.cc/public/guides/PaperChecklist.

What is the page limit for ICML? ›

Paper Submissions

Submitted papers are composed of a main body, which can be up to eight pages long, followed by unlimited pages for references and an appendix, all in a single file.

What is rebuttal in conference? ›

Most conferences and journals give you the option to respond to reviews for your paper — this is called "rebuttal". Your goal during the rebuttal phase is to try and convince the reviewers to increse their scores for your paper.

Top Articles
Latest Posts
Article information

Author: Rob Wisoky

Last Updated:

Views: 6055

Rating: 4.8 / 5 (68 voted)

Reviews: 91% of readers found this page helpful

Author information

Name: Rob Wisoky

Birthday: 1994-09-30

Address: 5789 Michel Vista, West Domenic, OR 80464-9452

Phone: +97313824072371

Job: Education Orchestrator

Hobby: Lockpicking, Crocheting, Baton twirling, Video gaming, Jogging, Whittling, Model building

Introduction: My name is Rob Wisoky, I am a smiling, helpful, encouraging, zealous, energetic, faithful, fantastic person who loves writing and wants to share my knowledge and understanding with you.