• Home
  • Schedule
  • Workshops
  • Papers
  • Sponsor Hall
  • Socials
  • Chat
  • Help
  • Calendar
  • Monday
  • Tuesday
  • Wednesday
  • Thursday
  • Expo
  • Code of Conduct
Calendar denotes when presenters are available for Q&A. All talks are available now. We recommend that you watch prior to live events.
Download Calendar

Monday

AI + Africa = Global Innovation

Dr. Aisha Walcott-Bryant / IBM Research Africa, Nairobi
Artificial Intelligence (AI) has for some time stoked the creative fires of computer scientists and researchers world-wide -- even before the so-called AI winter. After emerging from the winter, with much improved compute, vast amounts of data, and new techniques, AI has ignited our collective imaginations. We have been captivated by its promise while wary of its possible misuse in applications. AI has certainly demonstrated its enormous potential especially in fields such as healthcare. There, it has been used to support radiologists and to further precision medicine; conversely it has been used to generate photorealistic videos which distort our concept of what is real. Hence, we must thoughtfully harness AI to address the myriad of scientific and societal challenges; and open pathways to opportunities in governance, policy, and management. In this talk, I will share innovative solutions which leverage AI for global health with a focus on Africa. I will present a vision for the collaborations in hopes to inspire our community to join on this journey to transform Africa and impact the world.

Doing for Our Robots What Nature Did For Us

Prof. Leslie Kaelbling / MIT
We, as robot engineers, have to think hard about our role in the design of robots and how it interacts with learning, both in 'the factory' (that is, at engineering time) and in 'the wild' (that is, when the robot is delivered to a customer). I will share some general thoughts about the strategies for robot design and then talk in detail about some work I have been involved in, both in the design of an overall architecture for an intelligent robot and in strategies for learning to integrate new skills into the repertoire of an already competent robot.

Orals/Spotlights: Reliable RL

Measuring the Reliability of Reinforcement Learning Algorithms
Stephanie C.Y. Chan, Samuel Fishman, Anoop Korattikara, John Canny, Sergio Guadarrama,
Disagreement-Regularized Imitation Learning
Kiante Brantley, Wen Sun, Mikael Henaff,
Making Sense of Reinforcement Learning and Probabilistic Inference
Brendan O'Donoghue, Ian Osband, Catalin Ionescu,
The Ingredients of Real World Robotic Reinforcement Learning
Henry Zhu, Justin Yu, Abhishek Gupta, Dhruv Shah, Kristian Hartikainen, Avi Singh, Vikash Kumar, Sergey Levine,
Behaviour Suite for Reinforcement Learning
Ian Osband, Yotam Doron, Matteo Hessel, John Aslanides, Eren Sezener, Andre Saraiva, Katrina McKinney, Tor Lattimore, Csaba Szepesvari, Satinder Singh, Benjamin Van Roy, Richard Sutton, David Silver, Hado Van Hasselt,
Doubly Robust Bias Reduction in Infinite Horizon Off-Policy Estimation
Ziyang Tang, Yihao Feng, Lihong Li, Dengyong Zhou, Qiang Liu,
Implementation Matters in Deep RL: A Case Study on PPO and TRPO
Logan Engstrom, Andrew Ilyas, Shibani Santurkar, Dimitris Tsipras, Firdaus Janoos, Larry Rudolph, Aleksander Madry,
A Closer Look at Deep Policy Gradients
Andrew Ilyas, Logan Engstrom, Shibani Santurkar, Dimitris Tsipras, Firdaus Janoos, Larry Rudolph, Aleksander Madry,
Previous Next

Orals/Spotlights: Robustness and Verification

White Noise Analysis of Neural Networks
Ali Borji, Sikun Lin,
On Robustness of Neural Ordinary Differential Equations
Hanshu YAN, Jiawei DU, Vincent TAN, Jiashi FENG,
Skip Connections Matter: On the Transferability of Adversarial Examples Generated with ResNets
Dongxian Wu, Yisen Wang, Shu-Tao Xia, James Bailey, Xingjun Ma,
Enhancing Adversarial Defense by k-Winners-Take-All
Chang Xiao, Peilin Zhong, Changxi Zheng,
Defending Against Physically Realizable Attacks on Image Classification
Tong Wu, Liang Tong, Yevgeniy Vorobeychik,
HOPPITY: LEARNING GRAPH TRANSFORMATIONS TO DETECT AND FIX BUGS IN PROGRAMS
Elizabeth Dinella, Hanjun Dai, Ziyang Li, Mayur Naik, Le Song, Ke Wang,
Adversarial Training and Provable Defenses: Bridging the Gap
Mislav Balunovic, Martin Vechev,
Neural Network Branching for Neural Network Verification
Jingyue Lu, M. Pawan Kumar,
Previous Next

Orals/Spotlights: Optimisation Theory

Gradientless Descent: High-Dimensional Zeroth-Order Optimization
Daniel Golovin, John Karro, Greg Kochanski, Chansoo Lee, Xingyou Song, Qiuyi Zhang,
Generalization of Two-layer Neural Networks: An Asymptotic Viewpoint
Jimmy Ba, Murat Erdogdu, Taiji Suzuki, Denny Wu, Tianzong Zhang,
Depth-Width Trade-offs for ReLU Networks via Sharkovsky's Theorem
Vaggos Chatziafratis, Sai Ganesh Nagarajan, Ioannis Panageas, Xiao Wang,
At Stability's Edge: How to Adjust Hyperparameters to Preserve Minima Selection in Asynchronous Training of Neural Networks?
Niv Giladi, Mor Shpigel Nacson, Elad Hoffer, Daniel Soudry,
Online and stochastic optimization beyond Lipschitz continuity: A Riemannian approach
Kimon Antonakopoulos, E. Veronica Belmega, Panayotis Mertikopoulos,
Asymptotics of Wide Networks from Feynman Diagrams
Ethan Dyer, Guy Gur-Ari,
Why Gradient Clipping Accelerates Training: A Theoretical Justification for Adaptivity
Jingzhao Zhang, Tianxing He, Suvrit Sra, Ali Jadbabaie,
On the Convergence of FedAvg on Non-IID Data
Xiang Li, Kaixuan Huang, Wenhao Yang, Shusen Wang, Zhihua Zhang,
Previous Next

Orals/Spotlights: Multiagent Systems

Influence-Based Multi-Agent Exploration
Tonghan Wang, Jianhao Wang, Yi Wu, Chongjie Zhang,
Emergent Tool Use From Multi-Agent Autocurricula
Bowen Baker, Ingmar Kanitscheider, Todor Markov, Yi Wu, Glenn Powell, Bob McGrew, Igor Mordatch,
Simplified Action Decoder for Deep Multi-Agent Reinforcement Learning
Hengyuan Hu, Jakob N Foerster,
A Generalized Training Approach for Multiagent Learning
Paul Muller, Shayegan Omidshafiei, Mark Rowland, Karl Tuyls, Julien Perolat, Siqi Liu, Daniel Hennes, Luke Marris, Marc Lanctot, Edward Hughes, Zhe Wang, Guy Lever, Nicolas Heess, Thore Graepel, Remi Munos,
Posterior sampling for multi-agent reinforcement learning: solving extensive games with imperfect information
Yichi Zhou, Jialian Li, Jun Zhu,
Previous Next

Orals/Spotlights: Meta-learning

Meta-Learning Acquisition Functions for Transfer Learning in Bayesian Optimization
Michael Volpp, Lukas P. Fröhlich, Kirsten Fischer, Andreas Doerr, Stefan Falkner, Frank Hutter, Christian Daniel,
Meta-Learning without Memorization
Mingzhang Yin, George Tucker, Mingyuan Zhou, Sergey Levine, Chelsea Finn,
Improving Generalization in Meta Reinforcement Learning using Learned Objectives
Louis Kirsch, Sjoerd van Steenkiste, Juergen Schmidhuber,
Meta-Learning with Warped Gradient Descent
Sebastian Flennerhag, Andrei A. Rusu, Razvan Pascanu, Francesco Visin, Hujun Yin, Raia Hadsell,
Meta-Q-Learning
Rasool Fakoor, Pratik Chaudhari, Stefano Soatto, Alexander J. Smola,
Previous Next

Orals/Spotlights: Generative models

Hamiltonian Generative Networks
Peter Toth, Danilo J. Rezende, Andrew Jaegle, Sébastien Racanière, Aleksandar Botev, Irina Higgins,
Stable Rank Normalization for Improved Generalization in Neural Networks and GANs
Amartya Sanyal, Philip H. Torr, Puneet K. Dokania,
Real or Not Real, that is the Question
Yuanbo Xiangli, Yubin Deng, Bo Dai, Chen Change Loy, Dahua Lin,
Your classifier is secretly an energy based model and you should treat it like one
Will Grathwohl, Kuan-Chieh Wang, Joern-Henrik Jacobsen, David Duvenaud, Mohammad Norouzi, Kevin Swersky,
Optimal Strategies Against Generative Attacks
Roy Mor, Erez Peterfreund, Matan Gavish, Amir Globerson,
Previous Next

Orals/Spotlights: Signals and Systems

How much Position Information Do Convolutional Neural Networks Encode?
Md Amirul Islam, Sen Jia, Neil D. B. Bruce,
Learning to Control PDEs with Differentiable Physics
Philipp Holl, Nils Thuerey, Vladlen Koltun,
Scaling Autoregressive Video Models
Dirk Weissenborn, Oscar Täckström, Jakob Uszkoreit,
High Fidelity Speech Synthesis with Adversarial Networks
Mikołaj Bińkowski, Jeff Donahue, Sander Dieleman, Aidan Clark, Erich Elsen, Norman Casagrande, Luis C. Cobo, Karen Simonyan,
Learning Hierarchical Discrete Linguistic Units from Visually-Grounded Speech
David Harwath, Wei-Ning Hsu, James Glass,
Previous Next

Orals/Spotlights: Sequence Representations

DDSP: Differentiable Digital Signal Processing
Jesse Engel, Lamtharn (Hanoi) Hantrakul, Chenjie Gu, Adam Roberts,
word2ket: Space-efficient Word Embeddings inspired by Quantum Entanglement
Aliakbar Panahi, Seyran Saeedi, Tom Arodz,
Previous Next

Poster Sessions

Session 1 1 - (05:00-07:00 GMT)

Session 2 2 - (08:00-10:00 GMT)

Session 3 3 - (12:00-14:00 GMT)

Session 4 4 - (17:00-19:00 GMT)

Session 5 5 - (20:00-22:00 GMT)

Tuesday

2020 Vision: Reimagining the Default Settings of Technology & Society

Prof. Ruha Benjamin / Princeton
From everyday apps to complex algorithms, technology has the potential to hide, speed, and even deepen discrimination, while appearing neutral and even benevolent when compared to racist practices of a previous era. In this talk, I explore a range of discriminatory designs that encode inequity: by explicitly amplifying racial hierarchies, by ignoring but thereby replicating social divisions, or by aiming to fix racial bias but ultimately doing quite the opposite. This presentation takes us into the world of biased bots, altruistic algorithms, and their many entanglements, and provides conceptual tools to decode tech promises with sociologically informed skepticism. In doing so, it challenges us to question not only the technologies we are sold, but also the ones we manufacture ourselves.

Invertible Models and Normalizing Flows

Dr. Laurent Dinh / Google AI
Normalizing flows provide a tool to build an expressive and tractable family of probability distributions. In the last few years, research in this field has successfully harnessed some of the latest advances in deep learning to design flexible invertible models. Recently, these methods have seen wider adoption in the machine learning community for applications such as probabilistic inference, density estimation, and classification. In this talk, I will reflect on the recent progress made by the community on using, expanding, and repurposing this toolset, and describe my perspective on challenges and opportunities in this direction.

Orals/Spotlights: Probabilistic Approaches

A Probabilistic Formulation of Unsupervised Text Style Transfer
Junxian He, Xinyi Wang, Graham Neubig, Taylor Berg-Kirkpatrick,
Unbiased Contrastive Divergence Algorithm for Training Energy-Based Latent Variable Models
Yixuan Qiu, Lingsong Zhang, Xiao Wang,
SUMO: Unbiased Estimation of Log Marginal Probability for Latent Variable Models
Yucen Luo, Alex Beatson, Mohammad Norouzi, Jun Zhu, David Duvenaud, Ryan P. Adams, Ricky T. Q. Chen,
Directional Message Passing for Molecular Graphs
Johannes Klicpera, Janek Groß, Stephan Günnemann,
Intensity-Free Learning of Temporal Point Processes
Oleksandr Shchur, Marin Biloš, Stephan Günnemann,
Neural Tangents: Fast and Easy Infinite Neural Networks in Python
Roman Novak, Lechao Xiao, Jiri Hron, Jaehoon Lee, Alexander A. Alemi, Jascha Sohl-Dickstein, Samuel S. Schoenholz,
Learning to Balance: Bayesian Meta-Learning for Imbalanced and Out-of-distribution Tasks
Hae Beom Lee, Hayeon Lee, Donghyun Na, Saehoon Kim, Minseop Park, Eunho Yang, Sung Ju Hwang,
Cyclical Stochastic Gradient MCMC for Bayesian Deep Learning
Ruqi Zhang, Chunyuan Li, Jianyi Zhang, Changyou Chen, Andrew Gordon Wilson,
Previous Next

Orals/Spotlights: Biology and ML

Emergence of functional and structural properties of the head direction system by optimization of recurrent neural networks
Christopher J. Cueva, Peter Y. Wang, Matthew Chin, Xue-Xin Wei,
Disentangling neural mechanisms for perceptual grouping
Junkyung Kim, Drew Linsley, Kalpit Thakkar, Thomas Serre,
Reconstructing continuous distributions of 3D protein structure from cryo-EM images
Ellen D. Zhong, Tristan Bepler, Joseph H. Davis, Bonnie Berger,
Energy-based models for atomic-resolution protein conformations
Yilun Du, Joshua Meier, Jerry Ma, Rob Fergus, Alexander Rives,
Deep neuroethology of a virtual rodent
Josh Merel, Diego Aldarondo, Jesse Marshall, Yuval Tassa, Greg Wayne, Bence Olveczky,
Rotation-invariant clustering of neuronal responses in primary visual cortex
Ivan Ustyuzhaninov, Santiago A. Cadena, Emmanouil Froudarakis, Paul G. Fahey, Edgar Y. Walker, Erick Cobos, Jacob Reimer, Fabian H. Sinz, Andreas S. Tolias, Matthias Bethge, Alexander S. Ecker,
RNA Secondary Structure Prediction By Learning Unrolled Algorithms
Xinshi Chen, Yu Li, Ramzan Umarov, Xin Gao, Le Song,
Previous Next

Orals/Spotlights: Representation Learning

PROGRESSIVE LEARNING AND DISENTANGLEMENT OF HIERARCHICAL REPRESENTATIONS
Zhiyuan Li, Jaideep Vitthal Murkute, Prashnna Kumar Gyawali, Linwei Wang,
Sparse Coding with Gated Learned ISTA
Kailun Wu, Yiwen Guo, Ziang Li, Changshui Zhang,
Disentanglement by Nonlinear ICA with General Incompressible-flow Networks (GIN)
Peter Sorrenson, Carsten Rother, Ullrich Köthe,
Kaleidoscope: An Efficient, Learnable Representation For All Structured Linear Maps
Tri Dao, Nimit Sohoni, Albert Gu, Matthew Eichhorn, Amit Blonder, Megan Leszczynski, Atri Rudra, Christopher Ré,
Harnessing the Power of Infinitely Wide Deep Nets on Small-data Tasks
Sanjeev Arora, Simon S. Du, Zhiyuan Li, Ruslan Salakhutdinov, Ruosong Wang, Dingli Yu,
Self-labelling via simultaneous clustering and representation learning
Asano YM., Rupprecht C., Vedaldi A.,
Building Deep Equivariant Capsule Networks
Sai Raam Venkataraman, S. Balasubramanian, R. Raghunatha Sarma,
Target-Embedding Autoencoders for Supervised Representation Learning
Daniel Jarrett, Mihaela van der Schaar,
Previous Next

Orals/Spotlights: Feature Discovery for Structured Data

Multi-Scale Representation Learning for Spatial Feature Distributions using Grid Cells
Gengchen Mai, Krzysztof Janowicz, Bo Yan, Rui Zhu, Ling Cai, Ni Lao,
Tranquil Clouds: Neural Networks for Learning Temporally Coherent Features in Point Clouds
Lukas Prantl, Nuttapong Chentanez, Stefan Jeschke, Nils Thuerey,
Spectral Embedding of Regularized Block Models
Nathan De Lara, Thomas Bonald,
Convolutional Conditional Neural Processes
Jonathan Gordon, Wessel P. Bruinsma, Andrew Y. K. Foong, James Requeima, Yann Dubois, Richard E. Turner,
Restricting the Flow: Information Bottlenecks for Attribution
Karl Schulz, Leon Sixt, Federico Tombari, Tim Landgraf,
Previous Next

Orals/Spotlights: RL and Estimation

Estimating Gradients for Discrete Random Variables by Sampling without Replacement
Wouter Kool, Herke van Hoof, Max Welling,
Maximum Likelihood Constraint Inference for Inverse Reinforcement Learning
Dexter R.R. Scobee, S. Shankar Sastry,
Is a Good Representation Sufficient for Sample Efficient Reinforcement Learning?
Simon S. Du, Sham M. Kakade, Ruosong Wang, Lin F. Yang,
SEED RL: Scalable and Efficient Deep-RL with Accelerated Central Inference
Lasse Espeholt, Raphaël Marinier, Piotr Stanczyk, Ke Wang, Marcin Michalski‎,
GenDICE: Generalized Offline Estimation of Stationary Values
Ruiyi Zhang, Bo Dai, Lihong Li, Dale Schuurmans,
Previous Next

Poster Sessions

Session 1 1 - (05:00-07:00 GMT)

Session 2 2 - (08:00-10:00 GMT)

Session 3 3 - (12:00-14:00 GMT)

Session 4 4 - (17:00-19:00 GMT)

Session 5 5 - (20:00-22:00 GMT)

Wednesday

Machine Learning: Changing the future of healthcare

Prof. Mihaela van der Schaar / University of Cambridge, The Alan Turing Institute, UCLA
Medicine stands apart from other areas where machine learning can be applied. While we have seen advances in other fields with lots of data, it is not the volume of data that makes medicine so hard, it is the challenges arising from extracting actionable information from the complexity of the data. It is these challenges that make medicine the most exciting area for anyone who is really interested in the frontiers of machine learning – giving us real-world problems where the solutions are ones that are societally important and which potentially impact on us all. Think Covid 19! In this talk I will show how machine learning is transforming medicine and how medicine is driving new advances in machine learning, including new methodologies in automated machine learning, interpretable and explainable machine learning, dynamic forecasting, and causal inference.

AI Systems That Can See And Talk

Prof. Devi Parikh / Georgia Tech and Facebook AI Research
I will talk about AI systems at the intersection of computer vision and natural language processing. I will give an overview of why problems at the intersection of vision and language are exciting, what capabilities today's AI systems have, and what challenges remain.

Orals/Spotlights: RL and Planning

Dream to Control: Learning Behaviors by Latent Imagination
Danijar Hafner, Timothy Lillicrap, Jimmy Ba, Mohammad Norouzi,
Model Based Reinforcement Learning for Atari
Łukasz Kaiser, Mohammad Babaeizadeh, Piotr Miłos, Błażej Osiński, Roy H Campbell, Konrad Czechowski, Dumitru Erhan, Chelsea Finn, Piotr Kozakowski, Sergey Levine, Afroz Mohiuddin, Ryan Sepassi, George Tucker, Henryk Michalewski,
Learning to Plan in High Dimensions via Neural Exploration-Exploitation Trees
Binghong Chen, Bo Dai, Qinjie Lin, Guo Ye, Han Liu, Le Song,
Program Guided Agent
Shao-Hua Sun, Te-Lin Wu, Joseph J. Lim,
Learning Compositional Koopman Operators for Model-Based Control
Yunzhu Li, Hao He, Jiajun Wu, Dina Katabi, Antonio Torralba,
Watch the Unobserved: A Simple Approach to Parallelizing Monte Carlo Tree Search
Anji Liu, Jianshu Chen, Mingze Yu, Yu Zhai, Xuewen Zhou, Ji Liu,
Harnessing Structures for Value-Based Planning and Reinforcement Learning
Yuzhe Yang, Guo Zhang, Zhi Xu, Dina Katabi,
Previous Next

Orals/Spotlights: Theory

Compression based bound for non-compressed network: unified generalization error analysis of large compressible deep neural network
Taiji Suzuki, Hiroshi Abe, Tomoaki Nishimura,
Graph Neural Networks Exponentially Lose Expressive Power for Node Classification
Kenta Oono, Taiji Suzuki,
What Can Neural Networks Reason About?
Keyulu Xu, Jingling Li, Mozhi Zhang, Simon S. Du, Ken-ichi Kawarabayashi, Stefanie Jegelka,
The intriguing role of module criticality in the generalization of deep networks
Niladri Chatterji, Behnam Neyshabur, Hanie Sedghi,
Truth or backpropaganda? An empirical investigation of deep learning theory
Micah Goldblum, Jonas Geiping, Avi Schwarzschild, Michael Moeller, Tom Goldstein,
Ridge Regression: Structure, Cross-Validation, and Sketching
Sifan Liu, Edgar Dobriban,
Gradient Descent Maximizes the Margin of Homogeneous Neural Networks
Kaifeng Lyu, Jian Li,
A Theory of Usable Information under Computational Constraints
Yilun Xu, Shengjia Zhao, Jiaming Song, Russell Stewart, Stefano Ermon,
Previous Next

Orals/Spotlights: Sequence Representations

Symplectic Recurrent Neural Networks
Zhengdao Chen, Jianyu Zhang, Martin Arjovsky, Léon Bottou,
Encoding word order in complex embeddings
Benyou Wang, Donghao Zhao, Christina Lioma, Qiuchi Li, Peng Zhang, Jakob Grue Simonsen,
A Mutual Information Maximization Perspective of Language Representation Learning
Lingpeng Kong, Cyprien de Masson d'Autume, Lei Yu, Wang Ling, Zihang Dai, Dani Yogatama,
Duration-of-Stay Storage Assignment under Uncertainty
Michael Lingzhi Li, Elliott Wolf, Daniel Wintz,
Reformer: The Efficient Transformer
Nikita Kitaev, Lukasz Kaiser, Anselm Levskaya,
Mogrifier LSTM
Gábor Melis, Tomáš Kočiský, Phil Blunsom,
Previous Next

Orals/Spotlights: Actions and Counterfactuals

Estimating counterfactual treatment outcomes over time through adversarially balanced representations
Ioana Bica, Ahmed M Alaa, James Jordon, Mihaela van der Schaar,
CoPhy: Counterfactual Learning of Physical Dynamics
Fabien Baradel, Natalia Neverova, Julien Mille, Greg Mori, Christian Wolf,
CLEVRER: Collision Events for Video Representation and Reasoning
Kexin Yi, Chuang Gan, Yunzhu Li, Pushmeet Kohli, Jiajun Wu, Antonio Torralba, Joshua B. Tenenbaum,
CATER: A diagnostic dataset for Compositional Actions & TEmporal Reasoning
Rohit Girdhar, Deva Ramanan,
Causal Discovery with Reinforcement Learning
Shengyu Zhu, Ignavier Ng, Zhitang Chen,
Previous Next

Orals/Spotlights: Symbols and Discovery

Learning from Rules Generalizing Labeled Exemplars
Abhijeet Awasthi, Sabyasachi Ghosh, Rasna Goyal, Sunita Sarawagi,
Deep Learning For Symbolic Mathematics
Guillaume Lample, François Charton,
Neural Symbolic Reader: Scalable Integration of Distributed and Symbolic Representations for Reading Comprehension
Xinyun Chen, Chen Liang, Adams Wei Yu, Denny Zhou, Dawn Song, Quoc V. Le,
Mathematical Reasoning in Latent Space
Dennis Lee, Christian Szegedy, Markus Rabe, Sarah Loos, Kshitij Bansal,
Intrinsically Motivated Discovery of Diverse Patterns in Self-Organizing Systems
Chris Reinke, Mayalen Etcheverry, Pierre-Yves Oudeyer,
Previous Next

Orals/Spotlights: Optimisation I

Drawing Early-Bird Tickets: Toward More Efficient Training of Deep Networks
Haoran You, Chaojian Li, Pengfei Xu, Yonggan Fu, Yue Wang, Xiaohan Chen, Richard G. Baraniuk, Zhangyang Wang, Yingyan Lin,
An Exponential Learning Rate Schedule for Deep Learning
Zhiyuan Li, Sanjeev Arora,
Finite Depth and Width Corrections to the Neural Tangent Kernel
Boris Hanin, Mihai Nica,
Understanding Why Neural Networks Generalize Well Through GSNR of Parameters
Jinlong Liu, Yunzhi Bai, Guoqing Jiang, Ting Chen, Huayan Wang,
Deep Batch Active Learning by Diverse, Uncertain Gradient Lower Bounds
Jordan T. Ash, Chicheng Zhang, Akshay Krishnamurthy, John Langford, Alekh Agarwal,
Geometric Analysis of Nonconvex Optimization Landscapes for Overcomplete Learning
Qing Qu, Yuexiang Zhai, Xiao Li, Yuqian Zhang, Zhihui Zhu,
Previous Next

Poster Sessions

Session 1 1 - (05:00-07:00 GMT)

Session 2 2 - (08:00-10:00 GMT)

Session 3 3 - (12:00-14:00 GMT)

Session 4 4 - (17:00-19:00 GMT)

Session 5 5 - (20:00-22:00 GMT)

Thursday

Reflections from the Turing Award Winners

Profs. Yann LeCun and Yoshua Bengio / Facebook AI + NYU / MILA
Yoshua Bengio (Deep Learning Priors Associated with Conscious Processing): Some of the aspects of the world around us are captured in natural language and refer to semantic high-level variables, which often have a causal role (referring to agents, objects, and actions or intentions). These high-level variables also seem to satisfy very peculiar characteristics which low-level data (like images or sounds) do not share, and this work is about characterizing these characteristics in the form of priors which can guide the design of machine learning systems benefitting from these priors. Since these priors are not just about their joint distribution (e.g. it has a sparse factor graph) but also about how the distribution changes (typically by causal interventions), this analysis may also help to build machine learning systems which can generalize better out-of-distribution. There are fascinating connections between these priors and what is hypothesized about conscious processing in the brain, with conscious processing allowing us to reason (i.e., perform chains of inferences about the past and the future, as well as credit assignment) at the level of these high-level variables. This involves attention mechanisms and short-term memory to form a bottleneck of information being broadcast around the brain between different parts of it, as we focus on different high-level variables and some of their interactions. The presentation summarizes a few recent results using some of these ideas for discovering causal structure and modularizing recurrent neural networks with attention mechanisms in order to obtain better out-of-distribution generalization.

Yann LeCun (The Future is Self-Supervised): Humans and animals learn enormous amount of background knowledge about the world in the early months of life with little supervision and almost no interactions. How can we reproduce this learning paradigm in machines? One proposal for doing so is Self-Supervised Learning (SSL) in which a system is trained to predict a part of the input from the rest of the input. SSL, in the form of denoising auto-encoder, has been astonishingly successful for learning task-independent representations of text. But the success has not been translated to images and videos. The main obstacle is how to represent uncertainty in high-dimensional continuous spaces in which probability densities are generally intractable. We propose to use Energy-Based Models (EBM) to represent data manifolds or level-sets of distributions on the variables to be predicted. There are two classes of methods to train EBMs: (1) contrastive methods that push down on the energy of data points and push up elsewhere; (2) architectural and regularizing methods that limit or minimize the volume of space that can take low energies by regularizing the information capacity of a latent variable. While contrastive methods have been somewhat successful to learn image features, they are very expensive computationally. I will propose that the future of self-supervised representation learning lies in regularized latent-variable energy-based models.

The Decision-Making Side of Machine Learning: Dynamical, Statistical and Economic Perspectives

Prof. Michael I. Jordan / University of California, Berkeley
While there has been significant progress at the interface of statistics and computer science in recent years, many fundamental challenges remain. Some are mathematical and algorithmic in nature, such as the challenges associated with optimization and sampling in high-dimensional spaces. Some are statistical, including the challenges associated with multiple decision-making. Others are economic in nature, including the need to cope with scarcity and provide incentives in learning-based two-way markets. I will present recent progress on each of these fronts.

Orals/Spotlights: Natural Language

FreeLB: Enhanced Adversarial Training for Natural Language Understanding
Chen Zhu, Yu Cheng, Zhe Gan, Siqi Sun, Tom Goldstein, Jingjing Liu,
A Latent Morphology Model for Open-Vocabulary Neural Machine Translation
Duygu Ataman, Wilker Aziz, Alexandra Birch,
Neural Machine Translation with Universal Visual Representation
Zhuosheng Zhang, Kehai Chen, Rui Wang, Masao Utiyama, Eiichiro Sumita, Zuchao Li, Hai Zhao,
Sequential Latent Knowledge Selection for Knowledge-Grounded Dialogue
Byeongchang Kim, Jaewoo Ahn, Gunhee Kim,
ALBERT: A Lite BERT for Self-supervised Learning of Language Representations
Zhenzhong Lan, Mingda Chen, Sebastian Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut,
Towards Hierarchical Importance Attribution: Explaining Compositional Semantics for Neural Sequence Models
Xisen Jin, Zhongyu Wei, Junyi Du, Xiangyang Xue, Xiang Ren,
Mirror-Generative Neural Machine Translation
Zaixiang Zheng, Hao Zhou, Shujian Huang, Lei Li, Xin-Yu Dai, Jiajun Chen,
Data-dependent Gaussian Prior Objective for Language Generation
Zuchao Li, Rui Wang, Kehai Chen, Masso Utiyama, Eiichiro Sumita, Zhuosheng Zhang, Hai Zhao,
Previous Next

Orals/Spotlights: Network Architectures

PC-DARTS: Partial Channel Connections for Memory-Efficient Architecture Search
Yuhui Xu, Lingxi Xie, Xiaopeng Zhang, Xin Chen, Guo-Jun Qi, Qi Tian, Hongkai Xiong,
A Signal Propagation Perspective for Pruning Neural Networks at Initialization
Namhoon Lee, Thalaiyasingam Ajanthan, Stephen Gould, Philip H. S. Torr,
Network Deconvolution
Chengxi Ye, Matthew Evanusa, Hua He, Anton Mitrokhin, Tom Goldstein, James A. Yorke, Cornelia Fermuller, Yiannis Aloimonos,
Neural Arithmetic Units
Andreas Madsen, Alexander Rosenberg Johansen,
And the Bit Goes Down: Revisiting the Quantization of Neural Networks
Pierre Stock, Armand Joulin, Rémi Gribonval, Benjamin Graham, Hervé Jégou,
NAS-Bench-201: Extending the Scope of Reproducible Neural Architecture Search
Xuanyi Dong, Yi Yang,
Understanding and Robustifying Differentiable Architecture Search
Arber Zela, Thomas Elsken, Tonmoy Saikia, Yassine Marrakchi, Thomas Brox, Frank Hutter,
Comparing Rewinding and Fine-tuning in Neural Network Pruning
Alex Renda, Jonathan Frankle, Michael Carbin,
Previous Next

Orals/Spotlights: Graphs and Representations

Inductive Matrix Completion Based on Graph Neural Networks
Muhan Zhang, Yixin Chen,
Geom-GCN: Geometric Graph Convolutional Networks
Hongbin Pei, Bingzhe Wei, Kevin Chen-Chuan Chang, Yu Lei, Bo Yang,
DeepSphere: a graph-based spherical CNN
Michaël Defferrard, Martino Milani, Frédérick Gusset, Nathanaël Perraudin,
The Logical Expressiveness of Graph Neural Networks
Pablo Barceló, Egor V. Kostylev, Mikael Monet, Jorge Pérez, Juan Reutter, Juan Pablo Silva,
InfoGraph: Unsupervised and Semi-supervised Graph-Level Representation Learning via Mutual Information Maximization
Fan-Yun Sun, Jordan Hoffman, Vikas Verma, Jian Tang,
Strategies for Pre-training Graph Neural Networks
Weihua Hu, Bowen Liu, Joseph Gomes, Marinka Zitnik, Percy Liang, Vijay Pande, Jure Leskovec,
GraphZoom: A Multi-level Spectral Approach for Accurate and Scalable Graph Embedding
Chenhui Deng, Zhiqiang Zhao, Yongyu Wang, Zhiru Zhang, Zhuo Feng,
Contrastive Learning of Structured World Models
Thomas Kipf, Elise van der Pol, Max Welling,
Previous Next

Orals/Spotlights: Continual Learning and Few Shot Learning

Cross-Domain Few-Shot Classification via Learned Feature-Wise Transformation
Hung-Yu Tseng, Hsin-Ying Lee, Jia-Bin Huang, Ming-Hsuan Yang,
Sliced Cramer Synaptic Consolidation for Preserving Deeply Learned Representations
Soheil Kolouri, Nicholas A. Ketz, Andrea Soltoggio, Praveen K. Pilly,
Continual learning with hypernetworks
Johannes von Oswald, Christian Henning, João Sacramento, Benjamin F. Grewe,
Fast Task Inference with Variational Intrinsic Successor Features
Steven Hansen, Will Dabney, Andre Barreto, David Warde-Farley, Tom Van de Wiele, Volodymyr Mnih,
Dynamics-Aware Unsupervised Skill Discovery
Archit Sharma, Shixiang Gu, Sergey Levine, Vikash Kumar, Karol Hausman,
Continual learning with hypernetworks
Johannes von Oswald, Christian Henning, João Sacramento, Benjamin F. Grewe,
Previous Next

Orals/Spotlights: Optimisation II

Kernelized Wasserstein Natural Gradient
M Arbel, A Gretton, W Li, G Montufar,
The Break-Even Point on Optimization Trajectories of Deep Neural Networks
Stanislaw Jastrzebski, Maciej Szymczak, Stanislav Fort, Devansh Arpit, Jacek Tabor, Kyunghyun Cho, Krzysztof Geras,
Differentiation of Blackbox Combinatorial Solvers
Marin Vlastelica Pogančić, Anselm Paulus, Vit Musil, Georg Martius, Michal Rolinek,
Differentiable Reasoning over a Virtual Knowledge Base
Bhuwan Dhingra, Manzil Zaheer, Vidhisha Balachandran, Graham Neubig, Ruslan Salakhutdinov, William W. Cohen,
Principled Weight Initialization for Hypernetworks
Oscar Chang, Lampros Flokas, Hod Lipson,
Previous Next

Orals/Spotlights: Fairness, Interpretabiity and Deployment

Training individually fair ML models with sensitive subspace robustness
Mikhail Yurochkin, Amanda Bower, Yuekai Sun,
Conditional Learning of Fair Representations
Han Zhao, Amanda Coston, Tameem Adel, Geoffrey J. Gordon,
Explanation by Progressive Exaggeration
Sumedha Singla, Brian Pollack, Junxiang Chen, Kayhan Batmanghelich,
BackPACK: Packing more into Backprop
Felix Dangel, Frederik Kunstner, Philipp Hennig,
Federated Learning with Matched Averaging
Hongyi Wang, Mikhail Yurochkin, Yuekai Sun, Dimitris Papailiopoulos, Yasaman Khazaeni,
Previous Next

Poster Sessions

Session 1 1 - (05:00-07:00 GMT)

Session 2 2 - (08:00-10:00 GMT)

Session 3 3 - (12:00-14:00 GMT)

Session 4 4 - (17:00-19:00 GMT)

Session 5 5 - (20:00-22:00 GMT)

Expos

IBM
Neurosymbolic Hybrid AI

David Cox
Tue (14:00-15:00 GMT)

ByteDance
Learning Deep Latent Models for Text Sequences

Lei Li
Wed (19:00-20:00 GMT)

Element AI
Towards Ecologically Valid Research on Natural Language Interfaces

Harm de Vries
Wed (15:00-16:00 GMT)

Amazon
Reinforcement Learning @ Amazon

Britt Allen, Rui Song
Thurs (15:00-16:00 GMT)

Code of Conduct

The open exchange of ideas, the freedom of thought and expression, and respectful scientific debate are central to the goals of this conference on machine learning; this requires a community and an environment that recognizes and respects the inherent worth of every person.

Who

All participants---attendees, organizers, reviewers, speakers, sponsors, and volunteers at our conference, workshops, and conference-sponsored social events---are required to agree with this Code of Conduct both during the event and on official communication channels, including social media. Organizers will enforce this code, and we expect cooperation from all participants to help ensure a safe and productive environment for everybody.

Scope

The conference commits itself to provide an experience for all participants that is free from harassment, bullying, discrimination, and retaliation. This includes offensive comments related to gender, gender identity and expression, age, sexual orientation, disability, physical appearance, body size, race, ethnicity, religion (or lack thereof), politics, technology choices, or any other personal characteristics. Bullying, intimidation, personal attacks, harassment, sustained disruption of talks or other events, and behavior that interferes with another participant's full participation will not be tolerated. This includes sexual harassment, stalking, following, harassing photography or recording, inappropriate physical contact, unwelcome sexual attention, public vulgar exchanges, and diminutive characterizations, which are all unwelcome in this community.

The expected behaviour in line with the scope above extends to any format of the conference, including any virtual forms, and to the use of any online tools related to the conference. These include comments on OpenReview within or outside of reviewing periods, conference-wide chat tools, Q&A tools, live stream interactions, and any other forms of virtual interaction. Trolling, use of inappropriate imagery or videos, offensive language either written or in-person over video or audio, unwarranted direct messages (DMs), and extensions of such behaviour to tools outside those used by the conference but related to the conference, its program and attendees, are not allowed. In addition, doxxing or revealing any personal information to target any participant will not be tolerated.

Sponsors are equally subject to this Code of Conduct. In particular, sponsors should not use images, activities, or other materials that are of a sexual, racial, or otherwise offensive nature. Sponsor representatives and staff (including volunteers) should not use sexualized clothing/uniforms/costumes or otherwise create a sexualized environment. This code applies both to official sponsors as well as any organization that uses the conference name as branding as part of its activities at or around the conference.

Outcomes

Participants asked by any member of the community to stop any such behavior are expected to comply immediately. If a participant engages in such behavior, the conference organizers may take any action they deem appropriate, including: a formal or informal warning to the offender, expulsion from the conference (either physical expulsion, or termination of access codes) with no refund, barring from participation in future conferences or their organization, reporting the incident to the offender’s local institution or funding agencies, or reporting the incident to local law enforcement. A response of "just joking" will not be accepted; behavior can be harassing without an intent to offend. If action is taken, an appeals process will be made available.

Reporting

If you have concerns related to your inclusion at that conference, or observe someone else's difficulties, or have any other concerns related to inclusion, please email our ICLR Hotline.  For online events and tools, there are options to directly report specific chat/text comments, in addition to the above reporting. Complaints and violations will be handled with discretion. Reports made during the conference will be responded to within 24 hours; those at other times in less than two weeks. We are prepared and eager to help participants contact relevant help services, to escort them to a safe location, or to otherwise assist those experiencing harassment to feel safe for the duration of the conference. We gratefully accept feedback from the community on policy and actions; please contact us.

Back to Top

© 2020 International Conference on Learning Representations