Skip to yearly menu bar Skip to main content


Timezone: America/Chicago
Filter Events
Registration Desk
7:30 AM - 6:00 PM
Affinity Workshop

Women in ML

Nataša Tagasovska · Eda Okur · Hewitt Tusiime · Megha Srivastava · Jessica Sorrell · Estefany Kelly Buchanan · Shweta Khushu
8:00 AM - 3:30 PM
Affinity Workshop

Latinx in AI

Ignacio G. Lopez-Francos · CJ Barberan · Francisco Zabala · Karla Caballero · Cleber Zanchettin · Ana Maria Quintero · Vítor Lourenço · Walter M Mayor · Vinicius Caridá · Sebastian Caldas · Brayan Ortiz · Gabriela L. Vega Lopez · Abel Reyes-Angulo · Luis G. Sanchez Giraldo · Rocio Athziri Padilla Medina · Aaron Ferber · Marco Sanchez Sorondo · Laura Montoya
8:15 AM - 3:30 PM

The 6th Annual LXAI Research Workshop, held alongside NeurIPS conference, is a one-day event that unites faculty, researchers, practitioners, and students globally to foster collaborations and exchange novel ideas in the AI field. Spotlighting the contributions of the Latinx/Hispanic community, the workshop offers a platform to discuss current research trends and showcase innovative work. The agenda includes sessions with renowned and early-career speakers, oral presentations, industry and mentoring panels, and collaborative poster sessions, culminating in networking social events. While the primary presenters are from the Latinx/Hispanic community, all are welcome to join and enrich the dialogue.

... more
Affinity Workshop

New in ML

Mélisande Teng · Nishanth Anand · Reyhane Askari Hemmat · Subhrajyoti Dasgupta · Diganta Misra · Beheshteh Toloueirakhshan · Zach Xu · Isabelle Guyon · Rohan Sukumaran · Zhimeng Jiang
8:15 AM - 3:30 PM
Affinity Workshop

Black in AI

Marquita Riggins · Gelyn Watkins
8:15 AM - 3:30 PM
Affinity Workshop

Muslims in ML

Sanae Lotfi · Hammaad Adam · Hadeel Al-Negheimish · Sarah Fakhoury · Razan Baltaji · Marzyeh Ghassemi · Shakir Mohamed · Aya Salama · S. M. Ali Eslami · Tasmie Sarker
8:30 AM - 3:30 PM

The Muslims in ML workshop seeks to promote awareness, collaboration, and the development of mitigation strategies to ensure that machine learning and artificial intelligence advancements are implemented fairly and equitably for Muslims worldwide. By bringing together a diverse range of experts and incorporating multiple perspectives and backgrounds, our workshop aims to examine the challenges and opportunities of integrating AI/ML in the lives of Muslims and those in Muslim-majority countries. The workshop's focus extends beyond religious identification, encompassing cultural association and proximity to the Muslim identity. This broad approach acknowledges the complexity and diversity within the Muslim community and emphasizes the importance of inclusivity and understanding in addressing the potential impact of AI/ML technologies.

... more
Affinity Workshop

North Africans in ML

SOFIA BOURHIM · Oumaima Hourrane
8:30 AM - 3:30 PM

The North Africans in ML (NAML) aims to inspire and empower North Africans, enabling them to realize their full potential in the AI field up to publishing in top AI conferences. By fostering collaboration, networking, and skill development, the workshop intends to create a supportive community that celebrates and amplifies the contributions of North Africans to the world of machine learning.

... more
Affinity Workshop

Indigenous in AI/ML

Mason Grimshaw · Andrea M. Delgado-Olson · Michael Running Wolf
9:00 AM - 3:15 PM

Indigenous In AI’s vision is to build an international community of Native, Aboriginal, and First Nations who will collectively transform their home communities with advanced technology. By elevating the voices of Indigenous ML researchers we will inspire future impactful work and break stereotypes. Additionally, this group will strive to educate the broader NeurIPS on contemporary indigenous issues relevant to information technology and practices.

... more
Affinity Workshop

Queer in AI

Jaidev Shriram · Sharvani Jha · Ruchira Ray · Sarthak Arora
9:30 AM - 3:30 PM

Queer in AI’s workshop + socials at NeurIPS 2023 aim to act as a gathering space for queer folks to build community + solidarity while enabling participants to learn about key issues + topics at the intersection of AI and queerness.

... more
Tutorial

Data Contribution Estimation for Machine Learning

Stephanie Schoch · Ruoxi Jia · Yangfeng Ji
9:45 AM - 12:15 PM

Tasks enabled by data contribution estimation (DCE) aid model improvement through data improvement. While benchmark DCE evaluation tasks show application across many ML domains, DCE has limited visibility in other research domains that stand to benefit from its use cases. We propose a tutorial on data contribution for machine learning to address this. This tutorial will provide an overview of DCE for machine learning and natural language processing. Following this tutorial, attendees will have gained an understanding of 1) broadly, what questions data contribution estimation aims to answer; 2) the theory and methods that are widely in use within the DCE community that can be applied to a broad range of domains; 3) DCE from the perspectives of large language models and privacy.

... more
Tutorial

How to Work With Real Humans in Human-AI Systems

Elizabeth Bondi-Kelly · Krishnamurthy Dvijotham · Matthew Taylor
9:45 AM - 12:15 PM

As more and more AI systems are deployed in the real world, it becomes imperative to study these systems with real humans to avoid unexpected negative consequences during deployment. Yet, this can be challenging for researchers with more experience designing algorithms and less experience running human participant experiments, or deploying systems in the real world. In this tutorial, we will discuss the state of the human-AI collaboration field, emphasizing (i) incorporating humans into AI systems, including multi-agent, machine learning, and reinforcement learning systems, (ii) investigating when to rely on human vs. AI strengths, and (iii) designing human-AI studies to evaluate algorithms with real humans.

... more
Tutorial

Language Models Meet World Models

Zhiting Hu · Tianmin Shu
9:45 AM - 12:15 PM

Large language models (LMs) have achieved remarkable success in many language tasks.
Recent works have also shown that knowledge of the world can emerge from large LMs, enabling large LMs to assist decision-making for embodied tasks. However, the world knowledge exhibited by the current large LMs is often not robust and cannot be grounded in physical environments without additional models. This hinders large LMs’ abilities to perform complex reasoning and planning tasks reliably. For example, in creating action plans to move blocks to a target state, GPT-3 achieves a success rate of only 1%, compared to 78% for humans.

On the other hand, humans perform deliberate reasoning and planning based on the mental model of the world (i.e., world model, WMs) that enables us to simulate actions and their effects on the world’s state. WMs encoding the knowledge of the physical world can drastically improve the data efficiency and robustness of intelligent agents. However, WMs were typically studied in reinforcement learning and robotics, which are conceptually distinct from problems studied in language modeling.

This gap indicates enormous new opportunities for connecting WMs and LMs, to enhance LM capabilities of reasoning/planning in both embodied and general settings, and address the aforementioned limitations. Emerging studies on the intersection of WMs and LMs have demonstrated promising results. This tutorial aims to summarize and present a unified view of connecting WMs and LMs and highlight the various opportunities for improved machine reasoning and planning based on (or even beyond) large LMs through world modeling. We will review recent works on learning WMs and on using them to further learn and perform embodied tasks. We will show how LMs can utilize external WMs to compensate for their lack of grounded world knowledge and how LMs themselves can learn world models from embodied experiences that are beyond text data and use the internal WMs to guide complex reasoning.

... more
Tutorial
9:45 AM - 12:15 PM

Large matrices arise in many ML applications, including as representations of datasets, graphs, model weights, first and second-order derivatives, etc. Randomized Numerical Linear Algebra (RandNLA) is an area that uses randomness to develop improved algorithms for ubiquitous matrix problems. The area has reached a certain level of maturity, and current efforts of incorporating RandNLA algorithms into core numerical libraries, as well as recent advances in ML, Statistics, and Random Matrix Theory, have led to new theoretical and practical challenges. This tutorial will provide a self-contained overview of RandNLA in light of these important developments.

... more
Tutorial
9:45 AM - 12:15 PM

The success of the Transformer model has pushed the limits of deep learning to operate on the scale of trillions of parameters. This proliferation of large model size has outpaced the advances in hardware, resulting in an urgent need to distribute enormous models across multiple GPUs. Despite this trend, best practices for choosing an optimal strategy are still lacking due to the breadth of knowledge required across both deep learning and parallel computing.
This drives researchers to question deeply about: How to improve the training and inference efficiency of large models to reduce costs? Can we accommodate larger models with limited resources? What efforts can we make to enable more AI community members to access big models easily? In this tutorial, we investigate the efforts to solving above problems. A diverse set of parallelism is an important tool to improving the efficiency of large model training and inference. Heterogeneous memory management can enhance the model accommodation capacity of processors (e.g. GPUs).Further, deep learning systems for large AI models will significantly reduce the specialized background knowledge required from users, allowing AI users to quickly get started with larger models. We believe that with the benefits of these effective and extensive technologies for AI models, realizing an efficient and democratic big model era has become possible. We will provide participants with a systemic open-source solution and practical demonstrations for big models, in the hope of encouraging more practitioners and helping them apply mentioned technologies to their own practice.

... more
Tutorial
9:45 AM - 12:15 PM

Diffusion models have emerged as a powerful class of generative models and demonstrated astonishing results, in particular in image synthesis. However, training high-resolution diffusion models in pixel space can be highly expensive. Overcoming these limitations, Latent Diffusion Models (LDMs) first map high-resolution data into a compressed, typically lower-dimensional latent space using an autoencoder, and then train a diffusion model in that latent space more efficiently. Thereby, LDMs enable high-quality image synthesis while avoiding excessive compute demands. Furthermore, the LDM paradigm with an autoencoder, which can be tailored to specific problems and data, and a separate diffusion model in latent space offers significant flexibility with respect to architecture and model design. This has allowed LDMs to be successfully extended to various tasks beyond image generation, such as video synthesis, 3D object and scene generation, language modeling, and more. Most prominently, the well-known text-to-image model Stable Diffusion leverages the LDM framework. LDMs have become very popular and widely used in the generative modeling literature.

In this tutorial, we aim to provide an introduction to LDMs. While the literature on diffusion models has become broad, the LDM paradigm stands out as a particularly powerful approach due to its flexibility and excellent trade-off with respect to performance and compute demands. We aim to present a tutorial on LDMs that will benefit researchers interested in efficient and flexible, yet expressive generative modeling frameworks. We will also highlight advanced techniques for accelerated sampling and controllability, and discuss various applications of LDMs beyond image synthesis. Moreover, a panel discussion will provide diverse perspectives on this dynamic field and offer an outlook for future research on LDMs.

... more
Tutorial
9:45 AM - 12:15 PM

The tutorial aims to familiarize the ML community with major existing AI governance frameworks, ongoing AI policy proposals worldwide, and the concrete tools the research community has developed to adhere to standards and regulations applicable to ML systems in socially high-stakes domains. As a concrete governance challenge, we will focus on issues of bias and unfairness and overview pipeline-centric approaches to operationalize algorithmic harm prevention. As we will discuss, this approach is particularly relevant to challenges around leveraging the disparate impact doctrine for algorithmic harm prevention and recent FTC advanced notice of proposed rulemakings (ANPRMs). The concluding expert panel is an opportunity for the ML community to hear diverse perspectives on the key AI governance challenges in the near future and how the ML research community can prepare for and support efforts to address those challenges.

... more
Affinity Workshop

Global South AI

Susanna Raj · Pariya Sarin · Sudha Jamthe
11:00 AM - 2:00 PM

Global South in AI has the mission to add inclusion to Language AI. They focus on training new researchers from Global South Languages and Countries to present posters (peer reviewed selection) and bring them to NeurIPS to collaborate.

... more
Mentorship

Education Outreach

Tristan Naumann · Sanmi Koyejo · Marzyeh Ghassemi · Saadia Gabriel
12:00 PM - 1:40 PM

This event is by invitation only

12:00 - 12:10

  • Welcoming students
  • Give out lunch

12:10 - 12:25

  • Opening remark with Tristan Naumann
  • Welcome to NeurIPS
  • Purpose/history/vision of NeurIPS

12:30 - 12:55

  • Outreach program
  • Purpose of the program
  • Last year’s success
  • How many schools/students participated
  • This year’s scope
  • NeurIPS schedule announcement
  • Affinity workshop & Tutorial
  • Welcome reception
  • Main tracks
  • Keynote speeches
  • Workshops

13:00 - 13:30

  • Fireside Chat with Sanmi Koyejo, Marzyeh Ghassemi, Saadia Gabriel
  • Theme: Becoming a successful AI researcher/engineer as a budding student

13:30 - 13:40

  • Q&A
  • Wrap-up
... more
Mentorship

Would you like to learn how to communicate your AI research to a general audience? In this short tutorial you will learn how to turn your research articles into blog posts, how to use social media to promote your work, and how to avoid hype when writing about your research. You will also hear from AI researchers on how science communication has helped them improve their communication skills, and made their research more visible and impactful.

One of the challenges facing the field of AI is its portrayal in the media, which leads to misconceptions among policy makers, business leaders, and the general public alike. By communicating about AI in a clear, informed, and measured manner we can help to combat the flow of misinformation and convey the reality of today’s technology.

We will guide participants on how to quickly shape the story of their AI research. We’ll focus on how to structure this research story to form a blog post. Participants will learn how to explain their research to a general audience in a clear and concise manner, and how to find suitable images to illustrate their work.

After the session we will host a two-hour drop-in session to work with you one-on-one on your sci-comm questions, ideas and stories.

... more
Tutorial

Modeling and Exploiting Data Heterogeneity under Distribution Shifts

Jiashuo Liu · Tianhui Cai · Peng Cui · Hongseok Namkoong
1:45 PM - 4:15 PM

Data heterogeneity is a key determinant of the performance of ML systems. Standard algorithms that optimize for average-case performance do not consider the presence of diversity within data. As a result, variations in data sources, data generation mechanisms, and sub-populations lead to unreliable decision-making, poor generalization, unfairness, and false scientific discoveries. Carefully modeling data heterogeneity is a necessary step in building reliable data-driven systems. Its rigorous study is a nascent field of research spanning several disciplines, including statistics, causal inference, machine learning, economics, and operations research. In this tutorial, we develop a unified view of the disparate intellectual threads developed by different communities. We aim to foster interdisciplinary research by providing a unified view based on a shared language. Drawing upon several separate literatures, we establish a taxonomy of heterogeneity and present quantitative measures and learning algorithms that consider heterogeneous data. To spur empirical progress, we conclude by discussing validation protocols and benchmarking practices.

... more
Tutorial

Machine Learning for Theorem Proving

Zhangir Azerbayev · Emily First · Albert Q. Jiang · Kaiyu Yang · Anima Anandkumar · Noah Goodman · Alex Sanchez-Stern · Dawn Song · Sean Welleck
1:45 PM - 4:15 PM

Machine learning, especially large language models (LLMs), has shown promise in proving formal theorems using proof assistants such as Coq, Isabelle, and Lean. Theorem proving is an important challenge for machine learning: Formal proofs are computer programs whose correctness can be verified. Therefore, theorem proving is a form of code generation with rigorous evaluation and no room for the model to hallucinate, opening up a new avenue for addressing LLMs’ flaws in factuality. Despite its potential, learning-based theorem proving has significant entry barriers, primarily due to the steep learning curve for proof assistants. This tutorial aims to bridge this gap and make theorem proving accessible to researchers with a general machine learning background. To that end, our presentation will contextualize theorem proving from a machine learning perspective and demonstrate how to develop LLMs for theorem proving, using newly available open-source tools that provide interfaces to proof assistants without requiring in-depth knowledge of their internals. Furthermore, we will cover advanced topics and open problems in learning-based theorem proving, including its synergies with natural language processing and software verification. Throughout the presentation, we will highlight several conceptual themes recurring in theorem proving that are also critical for machine learning, such as mathematical reasoning, code generation, and hallucination prevention. The panel will complement the presentation through a broader discussion of related topics such as trustworthy machine learning, LLMs for code, reasoning, and program synthesis.

... more
Tutorial
1:45 PM - 4:15 PM

The rise of large language models (LLMs) offers a new approach for quickly building AI applications. While LLMs such as ChatGPT, Bard, and Bing chat are widely understood as consumer tools, the best practices for developers to effectively use these models through API calls remain poorly understood. This tutorial will share with the NeurIPS audience best practices for building AI applications using LLMs.
This course will include, but also go significantly beyond, “prompt engineering.” We will share best practices for integrating LLMs into more complex software systems, evaluating and continually improving their performance, and enhancing their safety. We will discuss best practices for using LLMs in common operations such as summarizing, making inferences, transforming text, and expanding text, as well as in-context learning, fine-tuning, and the utilization of both open-source and proprietary cloud-hosted LLMs.
LLMs are transforming the development process of AI applications. For example, a sentiment classifier that used to take weeks to build, via a process of collecting and labeling training examples, tuning a supervised model, and then finally deploying the model to make inferences, can now be built in hours by prompting an LLM API.
Through this tutorial, we hope to connect research and practice, and also inspire researchers to pursue new directions relevant to how LLMs are being used today.

... more
Tutorial

Data-Centric AI for reliable and responsible AI: from theory to practice

Mihaela van der Schaar · Isabelle Guyon · Nabeel Seedat · Jennifer Wortman Vaughan · Kyunghyun Cho · Razvan Pascanu · Jim Weatherall
1:45 PM - 4:15 PM

Data-Centric AI has recently been raised as an important paradigm shift in machine learning and AI — placing the previously undervalued “data work’ at the center of AI development. This tutorial aims to illuminate the fundamentals of Data-Centric AI and articulate its transformative potential. We will explore the motivation behind the data-centric approach, highlighting the power to improve model performance, engender more trustworthy, fair, and unbiased AI systems, as well as discuss benchmarking from a data-centric perspective. Our examination extends to standardized documentation frameworks, exposing how they form the backbone of this new paradigm. The tutorial will cover state-of-the-art methodologies that underscore these areas, which we will contextualize around the high-stakes setting of healthcare. A focus of this tutorial is providing participants with an interactive and hands-on experience. To this end, we provide coding/software tools and resources, thereby enabling practical engagement. The panel discussion, with experts spanning diverse industries, will provide a dynamic platform for discourse, enabling a nuanced understanding of the implications and limitations of Data-Centric AI across different contexts. Ultimately, our goal is that participants gain a practical foundation in data-centric AI, such that they can use or contribute to Data-Centric AI research.

... more
Tutorial

Do You Prefer Learning with Preferences?

Aditya Gopalan · Aadirupa Saha · Yoshua Bengio · Craig Boutilier · Elad Hazan · Robert Nowak · Tobias Schnabel
1:45 PM - 4:15 PM

AI desires to imitate human intelligence for designing efficient decision-making systems, but are we really training them the way humans learn every day or take decisions? Studies have shown humans are inherently more comfortable making decisions on a relative scale or choosing alternatives from a set, which often helps us converge to an optimal decision faster. In recent times, as we are employing more and more AI tools for executing everyday tasks, it’s becoming necessary to align machine behavior with human-like decisions. Another critical challenge in training user-friendly systems lies in the requirement of a huge amount of human feedback, which is often costly and hard to obtain. The solution lies in learning to train our machines through human preferences! Our tutorial aims to address the critical need for educating researchers on different types of preference models by exploring real-world problems and showcasing how training systems through preference feedback can provide cutting-edge solutions. We will equip attendees with a comprehensive understanding of diverse preference models and inference techniques. Another goal of the tutorial is to encourage collaboration among various communities that have significant connections to preference-based learning, including bandits, multiagent games, econometrics, social choice theory, RL, optimization, robotics, and more. We will consider our tutorial a success if it inspires researchers to embark on novel insights in the general area of preference-based learning, bringing attention from different communities to foster dissemination, cross-fertilization, and discussion at scale. Let’s learn to train our machines like humans: Machine Learning meets Human Learning through preference feedback!

Tutorial website: https://sites.google.com/view/pref-learning-tutorial-neurips/home

... more
Tutorial

Reconsidering Overfitting in the Age of Overparameterized Models

Spencer Frei · Vidya Muthukumar · Fanny Yang · Arash Amini · Kamalika Chaudhuri · Daniel Hsu · Nati Srebro · Chiyuan Zhang
1:45 PM - 4:15 PM

Large, overparameterized models such as neural networks are now the workhorses of modern machine learning. These models are often trained to near-zero error on noisy datasets and simultaneously generalize well to unseen data, in contrast to the textbook intuition regarding the perils of overfitting. At the same time, near-perfect data-fitting can have severe issues in the context of robustness, privacy, and fairness. Classical theoretical frameworks provide little guidance for navigating these questions due to overparameterization. It is thus crucial to develop new intuition regarding overfitting and generalization that are reflective of these empirical observations. In this tutorial, we discuss recent work in the learning theory literature that provides theoretical insights into these phenomena.

... more
Tutorial

What can we do about NeurIPS Reviewer #2?Challenges, Solutions, Experiments and Open Problemsin Peer Review

Nihar Shah · Hugo Larochelle · Andrew McCallum · Alice Oh · - Mausam · Charvi Rastogi
1:45 PM - 4:15 PM

Peer review is fundamental to scientific research, impacting scientific progress, grant funding allocation, researcher well-being, career paths, and the public's view of science. This tutorial provides a scientific lens on the systemic issues in peer review. It aims to stimulate discussions and inform policy-making based on scientific evidence (rather than individual opinions or anecdotes), addressing a topic that directly affects us all. To this end, the tutorial will delve into various inherent challenges, drawing on experiments on the peer-review process in diverse scientific disciplines. It will also discuss viable solutions and important open problems. The tutorial material will be available at https://cs.cmu.edu/~nihars/tutorials/NeurIPS2023. Finally, the presenter is excited about two things—peer review and minions—and both of these will be reflected generously in the tutorial.

... more
Affinity Poster Session
3:30 PM - 4:30 PM
Remarks
5:00 PM - 5:25 PM
Invited Talk
Speaker Bio
Björn Ommer
Björn Ommer is a full professor at University of Munich where he is heading the Computer Vision & Learning Group. Before he was a full professor in the department of mathematics and computer science at Heidelberg University and a co-director of its Interdisciplinary Center for Scientific Computing. He received his diploma in computer science from University of Bonn, his PhD from ETH Zurich, and he was a postdoc at UC Berkeley. Björn serves as an associate editor for IEEE T-PAMI. His research interests include semantic scene understanding and retrieval, generative AI and visual synthesis, self-supervised metric and representation learning, and explainable AI. Moreover, he is applying this basic research in interdisciplinary projects within neuroscience and the digital humanities. His group has published a series of generative approaches, including "VQGAN" and "Stable Diffusion", which are now democratizing the creation of visual content and have already opened up an abundance of new directions in research, industry, the media, and beyond.
... more
Reception
6:15 PM - 8:30 PM

Creative AI Performances 1

Jean Oh · Isabelle Guyon
6:30 PM - 8:00 PM
6 Events in this session
Go to Event Page