Timezone: »
Soon after the birth of modern computer science in the 1930s, two fundamental questions arose: 1. How can computers learn useful programs from experience, as opposed to being programmed by human programmers? 2. How to program parallel multiprocessor machines, as opposed to traditional serial architectures? Both questions found natural answers in the field of Recurrent Neural Networks (RNNs), which are brain-inspired general purpose computers that can learn parallel-sequential programs or algorithms encoded as weight matrices.
Our first RNNaissance NIPS workshop dates back to 2003: http://people.idsia.ch/~juergen/rnnaissance.html . Since then, a lot has happened. Some of the most successful applications in machine learning (including deep learning) are now driven by RNNs such as Long Short-Term Memory, e.g., speech recognition, video recognition, natural language processing, image captioning, time series prediction, etc. Through the world's most valuable public companies, billions of people have now access to this technology through their smartphones and other devices, e.g., in the form of Google Voice or on Apple's iOS. Reinforcement-learning and evolutionary RNNs are solving complex control tasks from raw video input. Many RNN-based methods learn sequential attention strategies.
Here we will review the latest developments in all of these fields, and focus not only on RNNs, but also on learning machines in which RNNs interact with external memory such as neural Turing machines, memory networks, and related memory architectures such as fast weight networks and neural stack machines. In this context we will also will discuss asymptotically optimal program search methods and their practical relevance.
Our target audience has heard a bit about recurrent neural networks but will happy to hear again a summary of the basics, and then delve into the latest advanced stuff, to see and understand what has recently become possible. We are hoping for thousands of attendees.
All talks (mostly by famous experts in the field who have already agreed to speak) will be followed by open discussions. We will also have a call for posters. Selected posters will adorn the environment of the lecture hall. We will also have a panel discussion on the bright future of RNNs, and their pros and cons.
Author Information
Jürgen Schmidhuber (Swiss AI Lab, IDSIA (USI & SUPSI); NNAISENSE; KAUST)
Since age 15 or so, the main goal of professor Jürgen Schmidhuber has been to build a self-improving Artificial Intelligence (AI) smarter than himself, then retire. His lab's Deep Learning Neural Networks based on ideas published in the "Annus Mirabilis" 1990-1991 have revolutionised machine learning and AI. By the mid 2010s, they were on 3 billion devices, and used billions of times per day through users of the world's most valuable public companies, e.g., for greatly improved (CTC-LSTM-based) speech recognition on all Android phones, greatly improved machine translation through Google Translate and Facebook (over 4 billion LSTM-based translations per day), Apple's Siri and Quicktype on all iPhones, the answers of Amazon's Alexa, and numerous other applications. In 2011, his team was the first to win official computer vision contests through deep neural nets, with superhuman performance. In 2012, they had the first deep NN to win a medical imaging contest (on cancer detection). All of this attracted enormous interest from industry. His research group also established the fields of mathematically rigorous universal AI and recursive self-improvement in metalearning machines that learn to learn (since 1987). In 1990, he introduced unsupervised adversarial neural networks that fight each other in a minimax game to achieve artificial curiosity (GANs are a special case). In 1991, he introduced very deep learning through unsupervised pre-training, and neural fast weight programmers formally equivalent to what's now called linear Transformers. His formal theory of creativity & curiosity & fun explains art, science, music, and humor. He also generalized algorithmic information theory and the many-worlds theory of physics, and introduced the concept of Low-Complexity Art, the information age's extreme form of minimal art. He is recipient of numerous awards, author of over 350 peer-reviewed papers, and Chief Scientist of the company NNAISENSE, which aims at building the first practical general purpose AI. He is a frequent keynote speaker, and advising various governments on AI strategies.
Sepp Hochreiter (LIT AI Lab / University Linz)
Head of the LIT AI Lab and Professor of bioinformatics at the University of Linz. First to identify and analyze the vanishing gradient problem, the fundamental deep learning problem, in 1991. First author of the main paper on the now widely used LSTM RNNs. He implemented 'learning how to learn' (meta-learning) networks via LSTM RNNs and applied Deep Learning and RNNs to self-driving cars, sentiment analysis, reinforcement learning, bioinformatics, and medicine.
Alex Graves (Google DeepMind)
Main contributions to neural networks include the Connectionist Temporal Classification training algorithm (widely used for speech, handwriting and gesture recognition, e.g. by Google voice search), a type of differentiable attention for RNNs (originally for handwriting generation, now a standard tool in computer vision, machine translation and elsewhere), stochastic gradient variational inference, and Neural Turing Machines. He works at Google Deep Mind.
Rupesh K Srivastava (NNAISENSE)
More from the Same Authors
-
2021 : Learning Adaptive Control Flow in Transformers for Improved Systematic Generalization »
Róbert Csordás · Kazuki Irie · Jürgen Schmidhuber -
2021 : Augmenting Classic Algorithms with Neural Components for Strong Generalisation on Ambiguous and High-Dimensional Data »
Imanol Schlag · Jürgen Schmidhuber -
2021 : Improving Baselines in the Wild »
Kazuki Irie · Imanol Schlag · Róbert Csordás · Jürgen Schmidhuber -
2021 : A Modern Self-Referential Weight Matrix That Learns to Modify Itself »
Kazuki Irie · Imanol Schlag · Róbert Csordás · Jürgen Schmidhuber -
2021 : Exploring through Random Curiosity with General Value Functions »
Aditya Ramesh · Louis Kirsch · Sjoerd van Steenkiste · Jürgen Schmidhuber -
2021 : Unsupervised Learning of Temporal Abstractions using Slot-based Transformers »
Anand Gopalakrishnan · Kazuki Irie · Jürgen Schmidhuber · Sjoerd van Steenkiste -
2021 : Modern Hopfield Networks for Return Decomposition for Delayed Rewards »
Michael Widrich · Markus Hofmarcher · Vihang Patil · Angela Bitto · Sepp Hochreiter -
2021 : Understanding the Effects of Dataset Composition on Offline Reinforcement Learning »
Kajetan Schweighofer · Markus Hofmarcher · Marius-Constantin Dinu · Philipp Renz · Angela Bitto · Vihang Patil · Sepp Hochreiter -
2021 : Unsupervised Learning of Temporal Abstractions using Slot-based Transformers »
Anand Gopalakrishnan · Kazuki Irie · Jürgen Schmidhuber · Sjoerd van Steenkiste -
2022 : Learning to Control Rapidly Changing Synaptic Connections: An Alternative Type of Memory in Sequence Processing Artificial Neural Networks »
Kazuki Irie · Jürgen Schmidhuber -
2022 : On Narrative Information and the Distillation of Stories »
Dylan Ashley · Vincent Herrmann · Zachary Friggstad · Jürgen Schmidhuber -
2022 : The Benefits of Model-Based Generalization in Reinforcement Learning »
Kenny Young · Aditya Ramesh · Louis Kirsch · Jürgen Schmidhuber -
2023 Poster: Contrastive Training of Complex-Valued Autoencoders for Object Discovery »
Aleksandar Stanić · Anand Gopalakrishnan · Kazuki Irie · Jürgen Schmidhuber -
2022 : Learning gaze control, external attention, and internal attention since 1990-91 »
Jürgen Schmidhuber -
2022 Poster: Neural Differential Equations for Learning to Program Neural Nets Through Continuous Learning Rules »
Kazuki Irie · Francesco Faccio · Jürgen Schmidhuber -
2022 Poster: Exploring through Random Curiosity with General Value Functions »
Aditya Ramesh · Louis Kirsch · Sjoerd van Steenkiste · Jürgen Schmidhuber -
2021 : Panel Discussion 1 »
Megan Peters · Jürgen Schmidhuber · Simona Ghetti · Nick Roy · Oiwi Parker Jones · Ingmar Posner -
2021 : Credit Assignment & Meta-Learning in a Single Lifelong Trial »
Jürgen Schmidhuber -
2021 Poster: Going Beyond Linear Transformers with Recurrent Fast Weight Programmers »
Kazuki Irie · Imanol Schlag · Róbert Csordás · Jürgen Schmidhuber -
2021 Poster: Meta Learning Backpropagation And Improving It »
Louis Kirsch · Jürgen Schmidhuber -
2019 : Panel Discussion »
Jacob Andreas · Edward Gibson · Stefan Lee · Noga Zaslavsky · Jason Eisner · Jürgen Schmidhuber -
2019 Poster: Are Disentangled Representations Helpful for Abstract Visual Reasoning? »
Sjoerd van Steenkiste · Francesco Locatello · Jürgen Schmidhuber · Olivier Bachem -
2018 : Invited Speaker #4 Juergen Schmidhuber »
Jürgen Schmidhuber -
2018 Poster: Recurrent World Models Facilitate Policy Evolution »
David Ha · Jürgen Schmidhuber -
2018 Oral: Recurrent World Models Facilitate Policy Evolution »
David Ha · Jürgen Schmidhuber -
2018 Poster: Learning to Reason with Third Order Tensor Products »
Imanol Schlag · Jürgen Schmidhuber -
2017 : Invited Talk 3 »
Sepp Hochreiter -
2017 : Morning panel discussion »
Jürgen Schmidhuber · Noah Goodman · Anca Dragan · Pushmeet Kohli · Dhruv Batra -
2017 : HRL with gradient-based subgoal generators, asymptotically optimal incremental problem solvers, various meta-learners, and PowerPlay (Jürgen Schmidhuber) »
Jürgen Schmidhuber -
2017 : Panel: Machine learning and audio signal processing: State of the art and future perspectives »
Sepp Hochreiter · Bo Li · Karen Livescu · Arindam Mandal · Oriol Nieto · Malcolm Slaney · Hendrik Purwins -
2017 : Invited Talk »
Jürgen Schmidhuber -
2017 Spotlight: Self-Normalizing Neural Networks »
Günter Klambauer · Thomas Unterthiner · Andreas Mayr · Sepp Hochreiter -
2017 Poster: Self-Normalizing Neural Networks »
Günter Klambauer · Thomas Unterthiner · Andreas Mayr · Sepp Hochreiter -
2017 Poster: GANs Trained by a Two Time-Scale Update Rule Converge to a Local Nash Equilibrium »
Martin Heusel · Hubert Ramsauer · Thomas Unterthiner · Bernhard Nessler · Sepp Hochreiter -
2017 Poster: Neural Expectation Maximization »
Klaus Greff · Sjoerd van Steenkiste · Jürgen Schmidhuber -
2016 : Juergen Schmidhuber (Scientific Director of the Swiss AI Lab IDSIA) »
Jürgen Schmidhuber -
2016 Poster: Scaling Memory-Augmented Neural Networks with Sparse Reads and Writes »
Jack Rae · Jonathan J Hunt · Ivo Danihelka · Tim Harley · Andrew Senior · Gregory Wayne · Alex Graves · Timothy Lillicrap -
2016 Poster: Conditional Image Generation with PixelCNN Decoders »
Aaron van den Oord · Nal Kalchbrenner · Lasse Espeholt · koray kavukcuoglu · Oriol Vinyals · Alex Graves -
2016 Poster: Tagger: Deep Unsupervised Perceptual Grouping »
Klaus Greff · Antti Rasmus · Mathias Berglund · Hotloo Xiranood · Harri Valpola · Jürgen Schmidhuber -
2016 Poster: Memory-Efficient Backpropagation Through Time »
Audrunas Gruslys · Remi Munos · Ivo Danihelka · Marc Lanctot · Alex Graves -
2016 Poster: Strategic Attentive Writer for Learning Macro-Actions »
Alexander (Sasha) Vezhnevets · Volodymyr Mnih · Simon Osindero · Alex Graves · Oriol Vinyals · John Agapiou · koray kavukcuoglu -
2015 : Deep Learning RNNaissance »
Jürgen Schmidhuber -
2015 : On General Problem Solving and How to Learn an Algorithm »
Jürgen Schmidhuber -
2015 Poster: Rectified Factor Networks »
Djork-Arné Clevert · Andreas Mayr · Thomas Unterthiner · Sepp Hochreiter -
2015 Poster: Training Very Deep Networks »
Rupesh K Srivastava · Klaus Greff · Jürgen Schmidhuber -
2015 Spotlight: Training Very Deep Networks »
Rupesh K Srivastava · Klaus Greff · Jürgen Schmidhuber -
2015 Poster: Parallel Multi-Dimensional LSTM, With Application to Fast Biomedical Volumetric Image Segmentation »
Marijn F Stollenga · Wonmin Byeon · Marcus Liwicki · Jürgen Schmidhuber -
2014 Poster: Recurrent Models of Visual Attention »
Volodymyr Mnih · Nicolas Heess · Alex Graves · koray kavukcuoglu -
2014 Spotlight: Recurrent Models of Visual Attention »
Volodymyr Mnih · Nicolas Heess · Alex Graves · koray kavukcuoglu -
2014 Poster: Deep Networks with Internal Selective Attention through Feedback Connections »
Marijn F Stollenga · Jonathan Masci · Faustino Gomez · Jürgen Schmidhuber -
2013 Poster: Compete to Compute »
Rupesh K Srivastava · Jonathan Masci · Sohrob Kazerounian · Faustino Gomez · Jürgen Schmidhuber -
2012 Poster: Deep Neural Networks Segment Neuronal Membranes in Electron Microscopy Images »
Dan Ciresan · Alessandro Giusti · luca Maria Gambardella · Jürgen Schmidhuber -
2011 Poster: Practical Variational Inference for Neural Networks »
Alex Graves -
2011 Spotlight: Practical Variational Inference for Neural Networks »
Alex Graves -
2010 Poster: Improving the Asymptotic Performance of Markov Chain Monte-Carlo by Inserting Vortices »
Yi Sun · Faustino Gomez · Jürgen Schmidhuber -
2008 Poster: Offline Handwriting Recognition with Multidimensional Recurrent Neural Networks »
Alex Graves · Jürgen Schmidhuber -
2008 Spotlight: Offline Handwriting Recognition with Multidimensional Recurrent Neural Networks »
Alex Graves · Jürgen Schmidhuber -
2007 Poster: Unconstrained On-line Handwriting Recognition with Recurrent Neural Networks »
Alex Graves · Santiago Fernandez · Marcus Liwicki · Horst Bunke · Jürgen Schmidhuber