【ICLR2019】Poster 论文汇总

【ICLR2019】Poster 论文汇总

转自:https://blog.csdn.net/u014636245/article/details/88825615

ICLR2019 Poster 文章汇总, 共478 papers

在这里插入图片描述

Convolutional Neural Networks on Non-uniform Geometrical Signals Using Euclidean Spectral Transformation 
Keywords:Non-uniform Fourier Transform, 3D Learning, CNN, surface reconstruction
TL;DR:We use non-Euclidean Fourier Transformation of shapes defined by a simplicial complex for deep learning, achieving significantly better results than point-based sampling techiques used in current 3D learning literature.
Augmented Cyclic Adversarial Learning for Low Resource Domain Adaptation 
Keywords:Domain adaptation, generative adversarial network, cyclic adversarial learning, speech
TL;DR:A new cyclic adversarial learning augmented with auxiliary task model which improves domain adaptation performance in low resource supervised and unsupervised situations
Variance Networks: When Expectation Does Not Meet Your Expectations 
Keywords:deep learning, variational inference, variational dropout
TL;DR:It is possible to learn a zero-centered Gaussian distribution over the weights of a neural network by learning only variances, and it works surprisingly well.
Initialized Equilibrium Propagation for Backprop-Free Training 
Keywords:credit assignment, energy-based models, biologically plausible learning
TL;DR:We train a feedforward network without backprop by using an energy-based model to provide local targets
Explaining Image Classifiers by Counterfactual Generation 
Keywords:Explainability, Interpretability, Generative Models, Saliency Map, Machine Learning, Deep Learning
TL;DR:We compute saliency by using a strong generative model to efficiently marginalize over plausible alternative inputs, revealing concentrated pixel areas that preserve label information.
SNIP: SINGLE-SHOT NETWORK PRUNING BASED ON CONNECTION SENSITIVITY 
Keywords:neural network pruning, connection sensitivity
TL;DR:We present a new approach, SNIP, that is simple, versatile and interpretable; it prunes irrelevant connections for a given task at single-shot prior to training and is applicable to a variety of neural network models without modifications.
Diagnosing and Enhancing VAE Models 
Keywords:variational autoencoder, generative models
TL;DR:We closely analyze the VAE objective function and draw novel conclusions that lead to simple enhancements.
Disjoint Mapping Network for Cross-modal Matching of Voices and Faces 
Keywords:None
TL;DR:None
Automatically Composing Representation Transformations as a Means for Generalization 
Keywords:compositionality, deep learning, metareasoning
TL;DR:We explore the problem of compositional generalization and propose a means for endowing neural network architectures with the ability to compose themselves to solve these problems.
Visual Reasoning by Progressive Module Networks 
Keywords:None
TL;DR:None
Bayesian Deep Convolutional Networks with Many Channels are Gaussian Processes 
Keywords:Deep Convolutional Neural Networks, Gaussian Processes, Bayesian
TL;DR:Finite-width SGD trained CNNs vs. infinitely wide fully Bayesian CNNs. Who wins?
Learning to Learn without Forgetting by Maximizing Transfer and Minimizing Interference 
Keywords:None
TL;DR:None
Sparse Dictionary Learning by Dynamical Neural Networks 
Keywords:None
TL;DR:None
Eidetic 3D LSTM: A Model for Video Prediction and Beyond 
Keywords:None
TL;DR:None
ALISTA: Analytic Weights Are As Good As Learned Weights in LISTA 
Keywords:None
TL;DR:None
Three Mechanisms of Weight Decay Regularization 
Keywords:Generalization, Regularization, Optimization
TL;DR:We investigate weight decay regularization for different optimizers and identify three distinct mechanisms by which weight decay improves generalization.
Learning Multimodal Graph-to-Graph Translation for Molecule Optimization 
Keywords:graph-to-graph translation, graph generation, molecular optimization
TL;DR:We introduce a graph-to-graph encoder-decoder framework for learning diverse graph translations.
A Data-Driven and Distributed Approach to Sparse Signal Representation and Recovery 
Keywords:Sparsity, Compressive Sensing, Convolutional Network
TL;DR:We use deep learning techniques to solve the sparse signal representation and recovery problem.
On the Minimal Supervision for Training Any Binary Classifier from Only Unlabeled Data 
Keywords:learning from only unlabeled data, empirical risk minimization, unbiased risk estimator
TL;DR:Three class priors are all you need to train deep models from only U data, while any two should not be enough.
Neural Logic Machines 
Keywords:Neural-Symbolic Computation, Rule Induction, First-Order Logic
TL;DR:We propose the Neural Logic Machine (NLM), a neural-symbolic architecture for both inductive learning and logic reasoning.
Neural Speed Reading with Structural-Jump-LSTM 
Keywords:natural language processing, speed reading, recurrent neural network, classification
TL;DR:We propose a new model for neural speed reading that utilizes the inherent punctuation structure of a text to define effective jumping and skipping behavior.
Rigorous Agent Evaluation: An Adversarial Approach to Uncover Catastrophic Failures 
Keywords:agent evaluation, adversarial examples, robustness, safety, reinforcement learning
TL;DR:We show that rare but catastrophic failures may be missed entirely by random testing, which poses issues for safe deployment. Our proposed approach for adversarial testing fixes this.
Woulda, Coulda, Shoulda: Counterfactually-Guided Policy Search 
Keywords:None
TL;DR:None
signSGD via Zeroth-Order Oracle 
Keywords:nonconvex optimization, zeroth-order algorithm, black-box adversarial attack
TL;DR:We design and analyze a new zeroth-order stochastic optimization algorithm, ZO-signSGD, and demonstrate its connection and application to black-box adversarial attacks in robust deep learning
Preventing Posterior Collapse with delta-VAEs 
Keywords:Posterior Collapse, VAE, Autoregressive Models
TL;DR: Avoid posterior collapse by lower bounding the rate.
Algorithmic Framework for Model-based Deep Reinforcement Learning with Theoretical Guarantees 
Keywords:model-based reinforcement learning, sample efficiency, deep reinforcement learning
TL;DR:We design model-based reinforcement learning algorithms with theoretical guarantees and achieve state-of-the-art results on Mujuco benchmark tasks when one million or fewer samples are permitted.
Knowledge Flow: Improve Upon Your Teachers 
Keywords:Transfer Learning, Reinforcement Learning
TL;DR:‘Knowledge Flow’ trains a deep net (student) by injecting information from multiple nets (teachers). The student is independent upon training and performs very well on learned tasks irrespective of the setting (reinforcement or supervised learning).
Directed-Info GAIL: Learning Hierarchical Policies from Unsegmented Demonstrations using Directed Information 
Keywords:Imitation Learning, Reinforcement Learning, Deep Learning
TL;DR:Learning Hierarchical Policies from Unsegmented Demonstrations using Directed Information
A Max-Affine Spline Perspective of Recurrent Neural Networks 
Keywords:RNN, max-affine spline operators
TL;DR:We provide new insights and interpretations of RNNs from a max-affine spline operators perspective.
Learning to Navigate the Web 
Keywords:navigating web pages, reinforcement learning, q learning, curriculum learning, meta training
TL;DR:We train reinforcement learning policies using reward augmentation, curriculum learning, and meta-learning to successfully navigate web pages.
Training for Faster Adversarial Robustness Verification via Inducing ReLU Stability 
Keywords:verification, adversarial robustness, adversarial examples, stability, deep learning, regularization
TL;DR:We develop methods to train deep neural models that are both robust to adversarial perturbations and whose robustness is significantly easier to verify.
Learning to Learn with Conditional Class Dependencies 
Keywords:meta-learning, learning to learn, few-shot learning
TL;DR:CAML is an instance of MAML with conditional class dependencies.
Hierarchical Visuomotor Control of Humanoids 
Keywords:hierarchical reinforcement learning, motor control, motion capture
TL;DR:Solve tasks involving vision-guided humanoid locomotion, reusing locomotion behavior from motion capture data.
Unsupervised Adversarial Image Reconstruction 
Keywords:None
TL;DR:None
Max-MIG: an Information Theoretic Approach for Joint Learning from Crowds 
Keywords:None
TL;DR:None
AutoLoss: Learning Discrete Schedule for Alternate Optimization 
Keywords:Meta Learning, AutoML, Optimization Schedule
TL;DR:We propose a unified formulation for iterative alternate optimization and develop AutoLoss, a framework to automatically learn and generate optimization schedules.
Learning what and where to attend 
Keywords:Attention models, human feature importance, object recognition, cognitive science
TL;DR:A large-scale dataset for training attention models for object recognition leads to more accurate, interpretable, and human-like object recognition.
ROBUST ESTIMATION VIA GENERATIVE ADVERSARIAL NETWORKS 
Keywords:robust statistics, neural networks, minimax rate, data depth, contamination model, Tukey median, GAN
TL;DR:GANs are shown to provide us a new effective robust mean estimate against agnostic contaminations with both statistical optimality and practical tractability.
INVASE: Instance-wise Variable Selection using Neural Networks 
Keywords:None
TL;DR:None
Meta-Learning with Latent Embedding Optimization 
Keywords:meta-learning, few-shot, miniImageNet, tieredImageNet, hypernetworks, generative, latent embedding, optimization
TL;DR:Latent Embedding Optimization (LEO) is a novel gradient-based meta-learner with state-of-the-art performance on the challenging 5-way 1-shot and 5-shot miniImageNet and tieredImageNet classification tasks.
Non-vacuous Generalization Bounds at the ImageNet Scale: a PAC-Bayesian Compression Approach 
Keywords:generalization, deep-learning, pac-bayes
TL;DR:We obtain non-vacuous generalization bounds on ImageNet-scale deep neural networks by combining an original PAC-Bayes bound and an off-the-shelf neural network compression method.
Learning to Represent Edits 
Keywords:None
TL;DR:None
Neural Probabilistic Motor Primitives for Humanoid Control 
Keywords:Motor Primitives, Distillation, Reinforcement Learning, Continuous Control, Humanoid Control, Motion Capture, One-Shot Imitation
TL;DR:Neural Probabilistic Motor Primitives compress motion capture tracking policies into one flexible model capable of one-shot imitation and reuse as a low-level controller.
Differentiable Perturb-and-Parse: Semi-Supervised Parsing with a Structured Variational Autoencoder 
Keywords:differentiable dynamic programming, variational auto-encoder, dependency parsing, semi-supervised learning
TL;DR:Differentiable dynamic programming over perturbed input weights with application to semi-supervised VAE
Janossy Pooling: Learning Deep Permutation-Invariant Functions for Variable-Size Inputs 
Keywords:representation learning, permutation invariance, set functions, feature pooling
TL;DR:We propose Janossy pooling, a method for learning deep permutation invariant functions designed to exploit relationships within the input sequence and tractable inference strategies such as a stochastic optimization procedure we call piSGD
An Empirical Study of Example Forgetting during Deep Neural Network Learning 
Keywords:catastrophic forgetting, sample weighting, deep generalization
TL;DR:We show that catastrophic forgetting occurs within what is considered to be a single task and find that examples that are not prone to forgetting can be removed from the training set without loss of generalization.
RNNs implicitly implement tensor-product representations 
Keywords:tensor-product representations, compositionality, neural network interpretability, recurrent neural networks
TL;DR:RNNs implicitly implement tensor-product representations, a principled and interpretable method for representing symbolic structures in continuous space.
Learning To Solve Circuit-SAT: An Unsupervised Differentiable Approach 
Keywords:Neuro-Symbolic Methods, Circuit Satisfiability, Neural SAT Solver, Graph Neural Networks
TL;DR:We propose a neural framework that can learn to solve the Circuit Satisfiability problem from (unlabeled) circuit instances.
Dynamic Channel Pruning: Feature Boosting and Suppression 
Keywords:dynamic network, faster CNNs, channel pruning
TL;DR:We make convolutional layers run faster by dynamically boosting and suppressing channels in feature computation.
signSGD with Majority Vote is Communication Efficient and Fault Tolerant 
Keywords:large-scale learning, distributed systems, communication efficiency, convergence rate analysis, robust optimisation
TL;DR:Workers send gradient signs to the server, and the update is decided by majority vote. We show that this algorithm is convergent, communication efficient and fault tolerant, both in theory and in practice.
Bounce and Learn: Modeling Scene Dynamics with Real-World Bounces 
Keywords:None
TL;DR:None
K for the Price of 1: Parameter-efficient Multi-task and Transfer Learning 
Keywords:deep learning, mobile, transfer learning, multi-task learning, computer vision, small models, imagenet, inception, batch normalization
TL;DR:A novel and practically effective method to adapt pretrained neural networks to new tasks by retraining a minimal (e.g., less than 2%) number of parameters
Towards Metamerism via Foveated Style Transfer 
Keywords:Metamerism, foveation, perception, style transfer, psychophysics
TL;DR:We introduce a novel feed-forward framework to generate visual metamers
Post Selection Inference with Incomplete Maximum Mean Discrepancy Estimator 
Keywords:None
TL;DR:None
Emergent Coordination Through Competition 
Keywords:Multi-agent learning, Reinforcement Learning
TL;DR:We introduce a new MuJoCo soccer environment for continuous multi-agent reinforcement learning research, and show that population-based training of independent reinforcement learners can learn cooperative behaviors
Prior Convictions: Black-box Adversarial Attacks with Bandits and Priors 
Keywords:adversarial examples, gradient estimation, black-box attacks, model-based optimization, bandit optimization
TL;DR:We present a unifying view on black-box adversarial attacks as a gradient estimation problem, and then present a framework (based on bandits optimization) to integrate priors into gradient estimation, leading to significantly increased performance.
Sample Efficient Imitation Learning for Continuous Control 
Keywords:Imitation Learning, Continuous Control, Reinforcement Learning, Inverse Reinforcement Learning, Conditional Generative Adversarial Network
TL;DR:In this paper, we proposed a model-free, off-policy IL algorithm for continuous control. Experimental results showed that our algorithm achieves competitive results with GAIL while significantly reducing the environment interactions.
Generative Code Modeling with Graphs 
Keywords:Generative Model, Source Code, Graph Learning
TL;DR:Representing programs as graphs including semantics helps when generating programs
Critical Learning Periods in Deep Networks 
Keywords:Critical Period, Deep Learning, Information Theory, Artificial Neuroscience, Information Plasticity
TL;DR:Sensory deficits in early training phases can lead to irreversible performance loss in both artificial and neuronal networks, suggesting information phenomena as the common cause, and point to the importance of the initial transient and forgetting.
CEM-RL: Combining evolutionary and gradient-based methods for policy search 
Keywords:evolution strategy, deep reinforcement learning
TL;DR:We propose a new combination of evolution strategy and deep reinforcement learning which takes the best of both worlds
LanczosNet: Multi-Scale Deep Graph Convolutional Networks 
Keywords:None
TL;DR:None
Excessive Invariance Causes Adversarial Vulnerability 
Keywords:Generalization, Adversarial Examples, Invariance, Information Theory, Invertible Networks
TL;DR:We show deep networks are not only too sensitive to task-irrelevant changes of their input, but also too invariant to a wide range of task-relevant changes, thus making vast regions in input space vulnerable to adversarial attacks.
Hindsight policy gradients 
Keywords:reinforcement learning, policy gradients, multi-goal reinforcement learning
TL;DR:We introduce the capacity to exploit information about the degree to which an arbitrary goal has been achieved while another goal was intended to policy gradient methods.
Adaptive Gradient Methods with Dynamic Bound of Learning Rate 
Keywords:Optimization, SGD, Adam, Generalization
TL;DR:Novel variants of optimization methods that combine the benefits of both adaptive and non-adaptive methods.
Decoupled Weight Decay Regularization 
Keywords:None
TL;DR:None
Optimistic mirror descent in saddle-point problems: Going the extra (gradient) mile 
Keywords:Mirror descent, extra-gradient, generative adversarial networks, saddle-point problems
TL;DR:We show how the inclusion of an extra-gradient step in first-order GAN training methods can improve stability and lead to improved convergence results.
DialogWAE: Multimodal Response Generation with Conditional Wasserstein Auto-Encoder 
Keywords:None
TL;DR:None
No Training Required: Exploring Random Encoders for Sentence Classification 
Keywords:None
TL;DR:None
Neural Graph Evolution: Automatic Robot Design 
Keywords:Reinforcement learning, graph neural networks, robotics, deep learning, transfer learning
TL;DR:Automatic robotic design search with graph neural networks
Function Space Particle Optimization for Bayesian Neural Networks 
Keywords:None
TL;DR:None
Structured Adversarial Attack: Towards General Implementation and Better Interpretability 
Keywords:None
TL;DR:None
Spherical CNNs on Unstructured Grids 
Keywords:Spherical CNN, unstructured grid, panoramic, semantic segmentation, parameter efficiency
TL;DR:We present a new CNN kernel for unstructured grids for spherical signals, and show significant accuracy and parameter efficiency gain on tasks such as 3D classfication and omnidirectional image segmentation.
Optimal Transport Maps For Distribution Preserving Operations on Latent Spaces of Generative Models 
Keywords:generative models, optimal transport, distribution preserving operations
TL;DR:We propose a framework for modifying the latent space operations such that the distribution mismatch between the resulting outputs and the prior distribution the generative model was trained on is fully eliminated.
Deep Lagrangian Networks: Using Physics as Model Prior for Deep Learning 
Keywords:Deep Model Learning, Robot Control
TL;DR:This paper introduces a physics prior for Deep Learning and applies the resulting network topology for model-based control.
Accumulation Bit-Width Scaling For Ultra-Low Precision Training Of Deep Networks 
Keywords:reduced precision floating-point, partial sum accumulation bit-width, deep learning, training
TL;DR:We present an analytical framework to determine accumulation bit-width requirements in all three deep learning training GEMMs and verify the validity and tightness of our method via benchmarking experiments.
Deep Convolutional Networks as shallow Gaussian Processes 
Keywords:Gaussian process, CNN, ResNet, Bayesian
TL;DR:We show that CNNs and ResNets with appropriate priors on the parameters are Gaussian processes in the limit of infinitely many convolutional filters.
Unsupervised Domain Adaptation for Distance Metric Learning 
Keywords:domain adaptation, distance metric learning, face recognition
TL;DR:A new theory of unsupervised domain adaptation for distance metric learning and its application to face recognition across diverse ethnicity variations.
A comprehensive, application-oriented study of catastrophic forgetting in DNNs 
Keywords:incremental learning, deep neural networks, catatrophic forgetting, sequential learning
TL;DR:We check DNN models for catastrophic forgetting using a new evaluation scheme that reflects typical application conditions, with surprising results.
Posterior Attention Models for Sequence to Sequence Learning 
Keywords:posterior inference, attention, seq2seq learning, translation
TL;DR:Computing attention based on posterior distribution leads to more meaningful attention and better performance
Generative Question Answering: Learning to Answer the Whole Question 
Keywords:Question answering, question generation, reasoning, squad, clevr
TL;DR:Question answering models that model the joint distribution of questions and answers can learn more than discriminative models
Diversity and Depth in Per-Example Routing Models 
Keywords:conditional computation, routing models, depth
TL;DR:Per-example routing models benefit from architectural diversity, but still struggle to scale to a large number of routing decisions.
Selfless Sequential Learning 
Keywords:Lifelong learning, Continual Learning, Sequential learning, Regularization
TL;DR:A regularization strategy for improving the performance of sequential learning
M^3RL: Mind-aware Multi-agent Management Reinforcement Learning 
Keywords:Multi-agent Reinforcement Learning, Deep Reinforcement Learning
TL;DR:We propose Mind-aware Multi-agent Management Reinforcement Learning (M^3RL) for training a manager to motivate self-interested workers to achieve optimal collaboration by assigning suitable contracts to them.
The Deep Weight Prior 
Keywords:deep learning, variational inference, prior distributions
TL;DR:The deep weight prior learns a generative model for kernels of convolutional neural networks, that acts as a prior distribution while training on new datasets.
Efficient Multi-Objective Neural Architecture Search via Lamarckian Evolution 
Keywords:Neural Architecture Search, AutoML, AutoDL, Deep Learning, Evolutionary Algorithms, Multi-Objective Optimization
TL;DR:We propose a method for efficient Multi-Objective Neural Architecture Search based on Lamarckian inheritance and evolutionary algorithms.
Quaternion Recurrent Neural Networks 
Keywords:None
TL;DR:None
Adversarial Audio Synthesis 
Keywords:audio, waveform, spectrogram, GAN, adversarial, WaveGAN, SpecGAN
TL;DR:Learning to synthesize raw waveform audio with GANs
Preconditioner on Matrix Lie Group for SGD 
Keywords:preconditioner, stochastic gradient descent, Newton method, Fisher information, natural gradient, Lie group
TL;DR:We propose a new framework for preconditioner learning, derive new forms of preconditioners and learning methods, and reveal the relationship to methods like RMSProp, Adam, Adagrad, ESGD, KFAC, batch normalization, etc.
Learning to Screen for Fast Softmax Inference on Large Vocabulary Neural Networks 
Keywords:None
TL;DR:None
Adaptive Posterior Learning: few-shot learning with a surprise-based memory module 
Keywords:metalearning, memory, few-shot, relational, self-attention, classification, sequential, reasoning, working memory, episodic memory
TL;DR:We introduce a model which generalizes quickly from few observations by storing surprising information and attending over the most relevant data at each time point.
Probabilistic Planning with Sequential Monte Carlo methods 
Keywords:control as inference, probabilistic planning, sequential monte carlo, model based reinforcement learning
TL;DR:Leveraging control as inference and Sequential Monte Carlo methods, we proposed a probabilistic planning algorithm.
Plan Online, Learn Offline: Efficient Learning and Exploration via Model-Based Control 
Keywords:deep reinforcement learning, exploration, model-based
TL;DR:We propose a framework that incorporates planning for efficient exploration and learning in complex environments.
DHER: Hindsight Experience Replay for Dynamic Goals 
Keywords:None
TL;DR:None
FlowQA: Grasping Flow in History for Conversational Machine Comprehension 
Keywords:Machine Comprehension, Conversational Agent, Natural Language Processing, Deep Learning
TL;DR:We propose the Flow mechanism and an end-to-end architecture, FlowQA, that achieves SotA on two conversational QA datasets and a sequential instruction understanding task.
Learning to Design RNA 
Keywords:matter engineering, bioinformatics, rna design, reinforcement learning, meta learning, neural architecture search, hyperparameter optimization
TL;DR:We learn to solve the RNA Design problem with reinforcement learning using meta learning and autoML approaches.
Robust Conditional Generative Adversarial Networks 
Keywords:conditional GAN, unsupervised pathway, autoencoder, robustness
TL;DR:We introduce a new type of conditional GAN, which aims to leverage structure in the target space of the generator. We augment the generator with a new, unsupervised pathway to learn the target structure.
Top-Down Neural Model For Formulae 
Keywords:logic, formula, recursive neural networks, recurrent neural networks
TL;DR:A top-down approach how to recursively represent propositional formulae by neural networks is presented.
Cost-Sensitive Robustness against Adversarial Examples 
Keywords:Certified robustness, Adversarial examples, Cost-sensitive learning
TL;DR:A general method for training certified cost-sensitive robust classifier against adversarial perturbations
The role of over-parametrization in generalization of neural networks 
Keywords:Generalization, Over-Parametrization, Neural Networks, Deep Learning
TL;DR:We suggest a generalization bound that could partly explain the improvement in generalization with over-parametrization.
Diffusion Scattering Transforms on Graphs 
Keywords:graph neural networks, deep learning, stability, scattering transforms, convolutional neural networks
TL;DR:Stability of scattering transform representations of graph data to deformations of the underlying graph support.
Capsule Graph Neural Network 
Keywords:CapsNet, Graph embedding, GNN
TL;DR:Inspired by CapsNet, we propose a novel architecture for graph embeddings on the basis of node features extracted from GNN.
Energy-Constrained Compression for Deep Neural Networks via Weighted Sparse Projection and Layer Input Masking 
Keywords:None
TL;DR:None
Emerging Disentanglement in Auto-Encoder Based Unsupervised Image Content Transfer 
Keywords:Image-to-image Translation, Disentanglement, Autoencoders, Faces
TL;DR:An image to image translation method which adds to one image the content of another thereby creating a new image.
SGD Converges to Global Minimum in Deep Learning via Star-convex Path 
Keywords:None
TL;DR:None
Toward Understanding the Impact of Staleness in Distributed Machine Learning 
Keywords:None
TL;DR:None
Transfer Learning for Sequences via Learning to Collocate 
Keywords:transfer learning, recurrent neural network, attention, natural language processing
TL;DR:Transfer learning for sequence via learning to align cell-level information across domains.
Learning Procedural Abstractions and Evaluating Discrete Latent Temporal Structure 
Keywords:None
TL;DR:None
Unsupervised Speech Recognition via Segmental Empirical Output Distribution Matching 
Keywords:None
TL;DR:None
Adversarial Attacks on Graph Neural Networks via Meta Learning 
Keywords:graph mining, adversarial attacks, meta learning, graph neural networks, node classification
TL;DR:We use meta-gradients to attack the training procedure of deep neural networks for graphs.
Maximal Divergence Sequential Autoencoder for Binary Software Vulnerability Detection 
Keywords:Vulnerabilities Detection, Sequential Auto-Encoder, Separable Representation
TL;DR:We propose a novel method named Maximal Divergence Sequential Auto-Encoder that leverages Variational AutoEncoder representation for binary code vulnerability detection.
Neural Program Repair by Jointly Learning to Localize and Repair 
Keywords:neural program repair, neural program embeddings, pointer networks
TL;DR:Multi-headed Pointer Networks for jointly learning to localize and repair Variable Misuse bugs
Information-Directed Exploration for Deep Reinforcement Learning 
Keywords:reinforcement learning, exploration, information directed sampling
TL;DR:We develop a practical extension of Information-Directed Sampling for Reinforcement Learning, which accounts for parametric uncertainty and heteroscedasticity in the return distribution for exploration.
Attention, Learn to Solve Routing Problems! 
Keywords:learning, routing problems, heuristics, attention, reinforce, travelling salesman problem, vehicle routing problem, orienteering problem, prize collecting travelling salesman problem
TL;DR:Attention based model trained with REINFORCE with greedy rollout baseline to learn heuristics with competitive results on TSP and other routing problems
L2-Nonexpansive Neural Networks 
Keywords:None
TL;DR:None
Improving Generalization and Stability of Generative Adversarial Networks 
Keywords:GAN, generalization, gradient penalty, zero centered, convergence
TL;DR:We propose a zero-centered gradient penalty for improving generalization and stability of GANs
Adaptive Input Representations for Neural Language Modeling 
Keywords:Neural language modeling
TL;DR:Variable capacity input word embeddings and SOTA on WikiText-103, Billion Word benchmarks.
Neural Persistence: A Complexity Measure for Deep Neural Networks Using Algebraic Topology 
Keywords:Algebraic topology, persistent homology, network complexity, neural network
TL;DR:We develop a new topological complexity measure for deep neural networks and demonstrate that it captures their salient properties.
Efficient Augmentation via Data Subsampling 
Keywords:data augmentation, invariance, subsampling, influence
TL;DR:Selectively augmenting difficult to classify points results in efficient training.
Neural TTS Stylization with Adversarial and Collaborative Games 
Keywords:Text-To-Speech synthesis, GANs
TL;DR:a generative adversarial network for style modeling in a text-to-speech system
Optimal Control Via Neural Networks: A Convex Approach 
Keywords:None
TL;DR:None
CBOW Is Not All You Need: Combining CBOW with the Compositional Matrix Space Model 
Keywords:Text representation learning, Sentence embedding, Efficient training scheme, word2vec
TL;DR:We present a novel training scheme for efficiently obtaining order-aware sentence representations.
Stochastic Optimization of Sorting Networks via Continuous Relaxations 
Keywords:continuous relaxations, sorting, permutation, stochastic computation graphs, Plackett-Luce
TL;DR:We provide a continuous relaxation to the sorting operator, enabling end-to-end, gradient-based stochastic optimization.
Adaptivity of deep ReLU network for learning in Besov and mixed smooth Besov spaces: optimal rate and curse of dimensionality 
Keywords:None
TL;DR:None
Generating Multiple Objects at Spatially Distinct Locations 
Keywords:controllable image generation, text-to-image synthesis, generative model, generative adversarial network, gan
TL;DR:Extend GAN architecture to obtain control over locations and identities of multiple objects within generated images.
Near-Optimal Representation Learning for Hierarchical Reinforcement Learning 
Keywords:representation hierarchy reinforcement learning
TL;DR:We translate a bound on sub-optimality of representations to a practical training objective in the context of hierarchical reinforcement learning.
Understanding Composition of Word Embeddings via Tensor Decomposition 
Keywords:word embeddings, semantic composition, tensor decomposition
TL;DR:We present a generative model for compositional word embeddings that captures syntactic relations, and provide empirical verification and evaluation.
Structured Neural Summarization 
Keywords:Summarization, Graphs, Source Code
TL;DR:One simple trick to improve sequence models: Compose them with a graph model
Graph Wavelet Neural Network 
Keywords:graph convolution, graph wavelet transform, graph Fourier transform, semi-supervised learning
TL;DR:We present graph wavelet neural network (GWNN), a novel graph convolutional neural network (CNN), leveraging graph wavelet transform to address the shortcoming of previous spectral graph CNN methods that depend on graph Fourier transform.
A rotation-equivariant convolutional neural network model of primary visual cortex 
Keywords:rotation equivariance, equivariance, primary visual cortex, V1, neuroscience, system identification
TL;DR:A rotation-equivariant CNN model of V1 that outperforms previous models and suggest functional groupings of V1 neurons.
Supervised Community Detection with Line Graph Neural Networks 
Keywords:community detection, graph neural networks, belief propagation, energy landscape, non-backtracking matrix
TL;DR:We propose a novel graph neural network architecture based on the non-backtracking matrix defined over the edge adjacencies and demonstrate its effectiveness in community detection tasks on graphs.
Multiple-Attribute Text Rewriting 
Keywords:controllable text generation, generative models, conditional generative models, style transfer
TL;DR:A system for rewriting text conditioned on multiple controllable attributes
Wasserstein Barycenter Model Ensembling 
Keywords:Wasserstein barycenter model ensembling
TL;DR:we propose to use Wasserstein barycenters for semantic model ensembling
Policy Transfer with Strategy Optimization 
Keywords:transfer learning, reinforcement learning, modeling error, strategy optimization
TL;DR:We propose a policy transfer algorithm that can overcome large and challenging discrepancies in the system dynamics such as latency, actuator modeling error, etc.
code2seq: Generating Sequences from Structured Representations of Code 
Keywords:source code, programs, code2seq
TL;DR:We leverage the syntactic structure of source code to generate natural language sequences.
Predict then Propagate: Graph Neural Networks meet Personalized PageRank 
Keywords:Graph, GCN, GNN, Neural network, Graph neural network, Message passing neural network, Semi-supervised classification, Semi-supervised learning, PageRank, Personalized PageRank
TL;DR:Personalized propagation of neural predictions (PPNP) improves graph neural networks by separating them into prediction and propagation via personalized PageRank.
Slimmable Neural Networks 
Keywords:Slimmable neural networks, mobile deep learning, accuracy-efficiency trade-offs
TL;DR:We present a simple and general method to train a single neural network executable at different widths (number of channels in a layer), permitting instant and adaptive accuracy-efficiency trade-offs at runtime.
Analysing Mathematical Reasoning Abilities of Neural Models 
Keywords:mathematics, dataset, algebraic, reasoning
TL;DR:A dataset for testing mathematical reasoning (and algebraic generalization), and results on current sequence-to-sequence models.
RotDCF: Decomposition of Convolutional Filters for Rotation-Equivariant Deep Networks 
Keywords:None
TL;DR:None
Execution-Guided Neural Program Synthesis 
Keywords:None
TL;DR:None
Dynamic Sparse Graph for Efficient Deep Learning 
Keywords:Sparsity, compression, training, acceleration
TL;DR:We construct dynamic sparse graph via dimension-reduction search to reduce compute and memory cost in both DNN training and inference.
Fixup Initialization: Residual Learning Without Normalization 
Keywords:deep learning, residual networks, initialization, batch normalization, layer normalization
TL;DR:All you need to train deep residual networks is a good initialization; normalization layers are not necessary.
ProbGAN: Towards Probabilistic GAN with Theoretical Guarantees 
Keywords:Generative Adversarial Networks, Bayesian Deep Learning, Mode Collapse, Inception Score, Generator, Discriminator, CIFAR-10, STL-10, ImageNet
TL;DR:A novel probabilistic treatment for GAN with theoretical guarantee.
Exploration by random network distillation 
Keywords:reinforcement learning, exploration, curiosity
TL;DR:A simple exploration bonus is introduced and achieves state of the art performance in 3 hard exploration Atari games.
Unsupervised Learning of the Set of Local Maxima 
Keywords:None
TL;DR:None
On the Convergence of A Class of Adam-Type Algorithms for Non-Convex Optimization 
Keywords:nonconvex optimization, Adam, convergence analysis
TL;DR:We analyze convergence of Adam-type algorithms and provide mild sufficient conditions to guarantee their convergence, we also show violating the conditions can makes an algorithm diverge.
Minimum Divergence vs. Maximum Margin: an Empirical Comparison on Seq2Seq Models 
Keywords:None
TL;DR:None
GANSynth: Adversarial Neural Audio Synthesis 
Keywords:GAN, Audio, WaveNet, NSynth, Music
TL;DR:High-quality audio synthesis with GANs
Sliced Wasserstein Auto-Encoders 
Keywords:optimal transport, Wasserstein distances, auto-encoders, unsupervised learning
TL;DR:In this paper we use the sliced-Wasserstein distance to shape the latent distribution of an auto-encoder into any samplable prior distribution.
Learning Two-layer Neural Networks with Symmetric Inputs 
Keywords:Neural Network, Optimization, Symmetric Inputs, Moment-of-moments
TL;DR:We give an algorithm for learning a two-layer neural network with symmetric input distribution.
Learning to Understand Goal Specifications by Modelling Reward 
Keywords:instruction following, reward modelling, language understanding
TL;DR:We propose AGILE, a framework for training agents to perform instructions from examples of respective goal-states.
Do Deep Generative Models Know What They Don't Know? 
Keywords:None
TL;DR:None
Identifying and Controlling Important Neurons in Neural Machine Translation 
Keywords:neural machine translation, individual neurons, unsupervised, analysis, correlation, translation control, distributivity, localization
TL;DR:Unsupervised methods for finding, analyzing, and controlling important neurons in NMT
Representing Formal Languages: A Comparison Between Finite Automata and Recurrent Neural Networks 
Keywords:Language recognition, Recurrent Neural Networks, Representation Learning, deterministic finite automaton, automaton
TL;DR:Finite Automata Can be Linearly decoded from Language-Recognizing RNNs using low coarseness abstraction functions and high accuracy decoders.
Visual Explanation by Interpretation: Improving Visual Feedback Capabilities of Deep Neural Networks 
Keywords:model explanation, model interpretation, explainable ai, evaluation
TL;DR:Interpretation by Identifying model-learned features that serve as indicators for the task of interest. Explain model decisions by highlighting the response of these features in test data. Evaluate explanations objectively with a controlled dataset.
Don't let your Discriminator be fooled 
Keywords:GAN, generative models, computer vision
TL;DR:A discriminator that is not easily fooled by adversarial example makes GAN training more robust and leads to a smoother objective.
Latent Convolutional Models 
Keywords:latent models, convolutional networks, unsupervised learning, deep learning, modeling natural images, image restoration
TL;DR:We present a new deep latent model of natural images that can be trained from unlabeled datasets and can be utilized to solve various image restoration tasks.
A Universal Music Translation Network 
Keywords:None
TL;DR:None
How to train your MAML 
Keywords:meta-learning, deep-learning, few-shot learning, supervised learning, neural-networks, stochastic optimization
TL;DR:MAML is great, but it has many problems, we solve many of those problems and as a result we learn most hyper parameters end to end, speed-up training and inference and set a new SOTA in few-shot learning
Learning a SAT Solver from Single-Bit Supervision 
Keywords:sat, search, graph neural network, theorem proving, proof
TL;DR:We train a graph network to predict boolean satisfiability and show that it learns to search for solutions, and that the solutions it finds can be decoded from its activations.
Learning Representations of Sets through Optimized Permutations 
Keywords:sets, representation learning, permutation invariance
TL;DR:Learn how to permute a set, then encode permuted set with RNN to obtain a set representation.
Big-Little Net: An Efficient Multi-Scale Feature Representation for Visual and Speech Recognition 
Keywords:None
TL;DR:None
Unsupervised Hyper-alignment for Multilingual Word Embeddings 
Keywords:None
TL;DR:None
Visual Semantic Navigation using Scene Priors 
Keywords:None
TL;DR:None
NOODL: Provable Online Dictionary Learning and Sparse Coding 
Keywords:provable dictionary learning, sparse coding, support recovery, iterative hard thresholding, matrix factorization, neural architectures, noodl
TL;DR:We present a provable algorithm for exactly recovering both factors of the dictionary learning model.
Stochastic Gradient/Mirror Descent: Minimax Optimality and Implicit Regularization 
Keywords:None
TL;DR:None
Active Learning with Partial Feedback 
Keywords:None
TL;DR:None
Gradient descent aligns the layers of deep linear networks 
Keywords:None
TL;DR:None
Data-Dependent Coresets for Compressing Neural Networks with Applications to Generalization Bounds 
Keywords:None
TL;DR:None
On the loss landscape of a class of deep neural networks with no bad local valleys 
Keywords:None
TL;DR:None
DOM-Q-NET: Grounded RL on Structured Language 
Keywords:Reinforcement Learning, Web Navigation, Graph Neural Networks
TL;DR:Graph-based Deep Q Network for Web Navigation
Boosting Robustness Certification of Neural Networks 
Keywords:Robustness certification, Adversarial Attacks, Abstract Interpretation, MILP Solvers, Verification of Neural Networks
TL;DR:We refine the over-approximation results from incomplete verifiers using MILP solvers to prove more robustness properties than state-of-the-art.
Learning To Simulate 
Keywords:Simulation in machine learning, reinforcement learning, policy gradients, image rendering
TL;DR:We propose an algorithm that automatically adjusts parameters of a simulation engine to generate training data for a neural network such that validation accuracy is maximized.
Towards Understanding Regularization in Batch Normalization 
Keywords:None
TL;DR:None
The Laplacian in RL: Learning Representations with Efficient Approximations 
Keywords:Laplacian, reinforcement learning, representation
TL;DR:We propose a scalable method to approximate the eigenvectors of the Laplacian in the reinforcement learning context and we show that the learned representations can improve the performance of an RL agent.
Predicting the Generalization Gap in Deep Networks with Margin Distributions 
Keywords:Deep learning, large margin, generalization bounds, generalization gap.
TL;DR:We develop a new scheme to predict the generalization gap in deep networks with high accuracy.
Adversarial Imitation via Variational Inverse Reinforcement Learning 
Keywords:Our method introduces the empowerment-regularized maximum-entropy inverse reinforcement learning to learn near-optimal rewards and policies from expert demonstrations.
TL;DR:Inverse Reinforcement Learning, Imitation learning, Variational lnference, Learning from demonstrations
Reasoning About Physical Interactions with Object-Oriented Prediction and Planning 
Keywords:structured scene representation, predictive models, intuitive physics, self-supervised learning
TL;DR:We present a framework for learning object-centric representations suitable for planning in tasks that require an understanding of physics.
LayoutGAN: Generating Graphic Layouts with Wireframe Discriminators 
Keywords:None
TL;DR:None
Learning Mixed-Curvature Representations in Product Spaces 
Keywords:embeddings, non-Euclidean geometry, manifolds, geometry of data
TL;DR:Product manifold embedding spaces with heterogenous curvature yield improved representations compared to traditional embedding spaces for a variety of structures.
StrokeNet: A Neural Painting Environment 
Keywords:image generation, differentiable model, reinforcement learning, deep learning, model based
TL;DR:StrokeNet is a novel architecture where the agent is trained to draw by strokes on a differentiable simulation of the environment, which could effectively exploit the power of back-propagation.
Harmonizing Maximum Likelihood with GANs for Multimodal Conditional Generation 
Keywords:conditional GANs, conditional image generation, multimodal generation, reconstruction loss, maximum likelihood estimation, moment matching
TL;DR:We prove that the mode collapse in conditional GANs is largely attributed to a mismatch between reconstruction loss and GAN loss and introduce a set of novel loss functions as alternatives for reconstruction loss.
Measuring Compositionality in Representation Learning 
Keywords:compositionality, representation learning, evaluation
TL;DR:This paper proposes a simple procedure for evaluating compositional structure in learned representations, and uses the procedure to explore the role of compositionality in four learning problems.
Benchmarking Neural Network Robustness to Common Corruptions and Perturbations 
Keywords:robustness, benchmark, convnets, perturbations
TL;DR:We propose ImageNet-C to measure classifier corruption robustness and ImageNet-P to measure perturbation robustness
ADef: an Iterative Algorithm to Construct Adversarial Deformations 
Keywords:Adversarial examples, deformations, deep neural networks, computer vision
TL;DR:We propose a new, efficient algorithm to construct adversarial examples by means of deformations, rather than additive perturbations.
Discriminator-Actor-Critic: Addressing Sample Inefficiency and Reward Bias in Adversarial Imitation Learning 
Keywords:deep learning, reinforcement learning, imitation learning, adversarial learning
TL;DR:We address sample inefficiency and reward bias in adversarial imitation learning algorithms such as GAIL and AIRL.
Doubly Reparameterized Gradient Estimators for Monte Carlo Objectives 
Keywords:variational autoencoder, reparameterization trick, IWAE, VAE, RWS, JVI
TL;DR:Doubly reparameterized gradient estimators provide unbiased variance reduction which leads to improved performance.
Learning Recurrent Binary/Ternary Weights 
Keywords:Quantized Recurrent Neural Network, Hardware Implementation, Deep Learning
TL;DR:We propose high-performance LSTMs with binary/ternary weights, that can greatly reduce implementation complexity
Learning concise representations for regression by evolving networks of trees 
Keywords:regression, stochastic optimization, evolutionary compution, feature engineering
TL;DR:Representing the network architecture as a set of syntax trees and optimizing their structure leads to accurate and concise regression models.
Efficient Training on Very Large Corpora via Gramian Estimation 
Keywords:similarity learning, pairwise learning, matrix factorization, Gramian estimation, variance reduction, neural embedding models, recommender systems
TL;DR:We develop efficient methods to train neural embedding models with a dot-product structure, by reformulating the objective function in terms of generalized Gram matrices, and maintaining estimates of those matrices.
MAE: Mutual Posterior-Divergence Regularization for Variational AutoEncoders 
Keywords:None
TL;DR:None
Residual Non-local Attention Networks for Image Restoration 
Keywords:Non-local network, attention network, image restoration, residual learning
TL;DR:New state-of-the-art framework for image restoration
Meta-Learning For Stochastic Gradient MCMC 
Keywords:Meta Learning, MCMC
TL;DR:This paper proposes a method to automate the design of stochastic gradient MCMC proposal using meta learning approach.
Systematic Generalization: What Is Required and Can It Be Learned? 
Keywords:systematic generalization, language understanding, visual questions answering, neural module networks
TL;DR:We show that modular structured models are the best in terms of systematic generalization and that their end-to-end versions don't generalize as well.
Efficient Lifelong Learning with A-GEM 
Keywords:Lifelong Learning, Continual Learning, Catastrophic Forgetting, Few-shot Transfer
TL;DR:An efficient lifelong learning algorithm that provides a better trade-off between accuracy and time/ memory complexity compared to other algorithms.
Multi-step Retriever-Reader Interaction for Scalable Open-domain Question Answering 
Keywords:Open domain Question Answering, Reinforcement Learning, Query reformulation
TL;DR:Paragraph retriever and machine reader interacts with each other via reinforcement learning to yield large improvements on open domain datasets
Double Viterbi: Weight Encoding for High Compression Ratio and Fast On-Chip Reconstruction for Deep Neural Network 
Keywords:quantization, pruning, memory footprint, model compression, sparse matrix
TL;DR:We present a new weight encoding scheme which enables high compression ratio and fast sparse-to-dense matrix conversion.
Overcoming the Disentanglement vs Reconstruction Trade-off via Jacobian Supervision 
Keywords:disentangling, autoencoders, jacobian, face manipulation
TL;DR:A method for learning image representations that are good for both disentangling factors of variation and obtaining faithful reconstructions.
RotatE: Knowledge Graph Embedding by Relational Rotation in Complex Space 
Keywords:knowledge graph embedding, knowledge graph completion, adversarial sampling
TL;DR:A new state-of-the-art approach for knowledge graph embedding.
Guiding Policies with Language via Meta-Learning 
Keywords:meta-learning, language grounding, interactive
TL;DR:We propose a meta-learning method for interactively correcting policies with natural language.
AdaShift: Decorrelation and Convergence of Adaptive Learning Rate Methods 
Keywords:optimizer, Adam, convergence, decorrelation
TL;DR:We analysis and solve the non-convergence issue of Adam.
AD-VAT: An Asymmetric Dueling mechanism for learning Visual Active Tracking 
Keywords:Active tracking, reinforcement learning, adversarial learning, multi agent
TL;DR:We propose AD-VAT, where the tracker and the target object, viewed as two learnable agents, are opponents and can mutually enhance during training.
Marginal Policy Gradients: A Unified Family of Estimators for Bounded Action Spaces with Applications 
Keywords:None
TL;DR:None
On Self Modulation for Generative Adversarial Networks 
Keywords:unsupervised learning, generative adversarial networks, deep generative modelling
TL;DR:A simple GAN modification that improves performance across many losses, architectures, regularization schemes, and datasets.
Off-Policy Evaluation and Learning from Logged Bandit Feedback: Error Reduction via Surrogate Policy 
Keywords:None
TL;DR:None
Subgradient Descent Learns Orthogonal Dictionaries 
Keywords:Dictionary learning, Sparse coding, Non-convex optimization, Theory
TL;DR:Efficient dictionary learning by L1 minimization via a novel analysis of the non-convex non-smooth geometry.
ClariNet: Parallel Wave Generation in End-to-End Text-to-Speech 
Keywords:None
TL;DR:None
MARGINALIZED AVERAGE ATTENTIONAL NETWORK FOR WEAKLY-SUPERVISED LEARNING 
Keywords:feature aggregation, weakly supervised learning, temporal action localization
TL;DR:A novel marginalized average attentional network for weakly-supervised temporal action localization
Towards GAN Benchmarks Which Require Generalization 
Keywords:evaluation, generative adversarial networks, adversarial divergences
TL;DR:We argue that GAN benchmarks must require a large sample from the model to penalize memorization and investigate whether neural network divergences have this property.
A Closer Look at Few-shot Classification 
Keywords:few shot classification, meta-learning
TL;DR: A detailed empirical study in few-shot classification that revealing challenges in standard evaluation setting and showing a new direction.
Meta-Learning Probabilistic Inference for Prediction 
Keywords:probabilistic models, approximate inference, few-shot learning, meta-learning
TL;DR:Novel framework for meta-learning that unifies and extends a broad class of existing few-shot learning methods. Achieves strong performance on few-shot learning benchmarks without requiring iterative test-time inference.
Deep reinforcement learning with relational inductive biases 
Keywords:relational reasoning, reinforcement learning, graph neural networks, starcraft, generalization, inductive bias
TL;DR:Relational inductive biases improve out-of-distribution generalization capacities in model-free reinforcement learning agents
Relaxed Quantization for Discretized Neural Networks 
Keywords:Quantization, Compression, Neural Networks, Efficiency
TL;DR:We introduce a technique that allows for gradient based training of quantized neural networks.
Tree-Structured Recurrent Switching Linear Dynamical Systems for Multi-Scale Modeling 
Keywords:None
TL;DR:None
STCN: Stochastic Temporal Convolutional Networks 
Keywords:latent variables, variational inference, temporal convolutional networks, sequence modeling, auto-regressive modeling
TL;DR:We combine the computational advantages of temporal convolutional architectures with the expressiveness of stochastic latent variables.
Soft Q-Learning with Mutual-Information Regularization 
Keywords:None
TL;DR:None
On the Turing Completeness of Modern Neural Network Architectures 
Keywords:Transformer, NeuralGPU, Turing completeness
TL;DR:We show that the Transformer architecture and the Neural GPU are Turing complete.
Improving Differentiable Neural Computers Through Memory Masking, De-allocation, and Link Distribution Sharpness Control 
Keywords:None
TL;DR:None
Evaluating Robustness of Neural Networks with Mixed Integer Programming 
Keywords:verification, adversarial robustness, adversarial examples, deep learning
TL;DR:We efficiently verify the robustness of deep neural models with over 100,000 ReLUs, certifying more samples than the state-of-the-art and finding more adversarial examples than a strong first-order attack.
Random mesh projectors for inverse problems 
Keywords:imaging, inverse problems, subspace projections, random Delaunay triangulations, CNN, geophysics, regularization
TL;DR:We solve ill-posed inverse problems with scarce ground truth examples by estimating an ensemble of random projections of the model instead of the model itself.
Multi-Agent Dual Learning 
Keywords:None
TL;DR:None
Complement Objective Training 
Keywords:optimization, entropy, image recognition, natural language understanding, adversarial attacks, deep learning
TL;DR:We propose Complement Objective Training (COT), a new training paradigm that optimizes both the primary and complement objectives for effectively learning the parameters of neural networks.
Mode Normalization 
Keywords:Deep Learning, Expert Models, Normalization, Computer Vision
TL;DR:We present a novel normalization method for deep neural networks that is robust to multi-modalities in intermediate feature distributions.
Detecting Egregious Responses in Neural Sequence-to-sequence Models 
Keywords:Deep Learning, Natural Language Processing, Adversarial Attacks, Dialogue Response Generation
TL;DR:This paper aims to provide an empirical answer to the question of whether well-trained dialogue response model can output malicious responses.
Learning Actionable Representations with Goal Conditioned Policies 
Keywords:Representation Learning, Reinforcement Learning
TL;DR:Learning state representations which capture factors necessary for control
Verification of Non-Linear Specifications for Neural Networks 
Keywords:None
TL;DR:None
Generating Liquid Simulations with Deformation-aware Neural Networks 
Keywords:Learning weighting and deformations of space-time data sets for highly efficient approximations of liquid behavior.
TL;DR:deformation learning, spatial transformer networks, fluid simulation
DyRep: Learning Representations over Dynamic Graphs 
Keywords:Dynamic Graphs, Representation Learning, Dynamic Processes, Temporal Point Process, Attention, Latent Representation
TL;DR:Models Representation Learning over dynamic graphs as latent hidden process bridging two observed processes of Topological Evolution of and Interactions on dynamic graphs.
Trellis Networks for Sequence Modeling 
Keywords:sequence modeling, language modeling, recurrent networks, convolutional networks, trellis networks
TL;DR:Trellis networks are a new sequence modeling architecture that bridges recurrent and convolutional models and sets a new state of the art on word- and character-level language modeling.
Scalable Unbalanced Optimal Transport using Generative Adversarial Networks 
Keywords:unbalanced optimal transport, generative adversarial networks, population modeling
TL;DR:We propose new methodology for unbalanced optimal transport using generative adversarial networks.
Solving the Rubik's Cube with Approximate Policy Iteration 
Keywords:reinforcement learning, Rubik's Cube, approximate policy iteration, deep learning, deep reinforcement learning
TL;DR:We solve the Rubik's Cube with pure reinforcement learning
Variance Reduction for Reinforcement Learning in Input-Driven Environments 
Keywords:reinforcement learning, policy gradient, input-driven environments, variance reduction, baseline
TL;DR:For environments dictated partially by external input processes, we derive an input-dependent baseline that provably reduces the variance for policy gradient methods and improves the policy performance in a wide range of RL tasks.
Model-Predictive Policy Learning with Uncertainty Regularization for Driving in Dense Traffic 
Keywords:model-based reinforcement learning, stochastic video prediction, autonomous driving
TL;DR:A model-based RL approach which uses a differentiable uncertainty penalty to learn driving policies from purely observational data.
GAN Dissection: Visualizing and Understanding Generative Adversarial Networks 
Keywords:GAN representations are examined in detail, and sets of representation units are found that control the generation of semantic concepts in the output.
TL;DR:GANs, representation, interpretability, causality
Improving MMD-GAN Training with Repulsive Loss Function 
Keywords:generative adversarial nets, loss function, maximum mean discrepancy, image generation, unsupervised learning
TL;DR:Rearranging the terms in maximum mean discrepancy yields a much better loss function for the discriminator of generative adversarial nets
Deterministic PAC-Bayesian generalization bounds for deep networks via generalizing noise-resilience 
Keywords:generalization, PAC-Bayes, SGD, learning theory, implicit regularization
TL;DR:We provide a PAC-Bayes based generalization guarantee for uncompressed, deterministic deep networks by generalizing noise-resilience of the network on the training data to the test data.
Recall Traces: Backtracking Models for Efficient Reinforcement Learning 
Keywords:Model free RL, Variational Inference
TL;DR:A backward model of previous (state, action) given the next state, i.e. P(s_t, a_t | s_{t+1}), can be used to simulate additional trajectories terminating at states of interest! Improves RL learning efficiency.
Stable Recurrent Models 
Keywords:stability, gradient descent, non-convex optimization, recurrent neural networks
TL;DR:Stable recurrent models can be approximated by feed-forward networks and empirically perform as well as unstable models on benchmark tasks.
The Limitations of Adversarial Training and the Blind-Spot Attack 
Keywords:Adversarial Examples, Adversarial Training, Blind-Spot Attack
TL;DR:We show that even the strongest adversarial training methods cannot defend against adversarial examples crafted on slightly scaled and shifted test images.
Efficiently testing local optimality and escaping saddles for ReLU networks 
Keywords:local optimality, second-order stationary point, escaping saddle points, nondifferentiability, ReLU, empirical risk
TL;DR:A theoretical algorithm for testing local optimality and extracting descent directions at nondifferentiable points of empirical risks of one-hidden-layer ReLU networks.
ProxylessNAS: Direct Neural Architecture Search on Target Task and Hardware 
Keywords:Neural Architecture Search, Efficient Neural Networks
TL;DR:Proxy-less neural architecture search for directly learning architectures on large-scale target task (ImageNet) while reducing the cost to the same level of normal training.
Hierarchical Reinforcement Learning via Advantage-Weighted Information Maximization 
Keywords:Hierarchical reinforcement learning, Representation learning, Continuous control
TL;DR:This paper presents a hierarchical reinforcement learning framework based on deterministic option policies and mutual information maximization.
Generalizable Adversarial Training via Spectral Normalization 
Keywords:None
TL;DR:None
Adversarial Domain Adaptation for Stable Brain-Machine Interfaces 
Keywords:Brain-Machine Interfaces, Domain Adaptation, Adversarial Networks
TL;DR:We implement an adversarial domain adaptation network to stabilize a fixed Brain-Machine Interface against gradual changes in the recorded neural signals.
Deep Online Learning Via Meta-Learning: Continual Adaptation for Model-Based RL 
Keywords:None
TL;DR:None
Deep Anomaly Detection with Outlier Exposure 
Keywords:confidence, uncertainty, anomaly, robustness
TL;DR:OE teaches anomaly detectors to learn heuristics for detecting unseen anomalies; experiments are in classification, density estimation, and calibration in NLP and vision settings; we do not tune on test distribution samples, unlike previous work
Contingency-Aware Exploration in Reinforcement Learning 
Keywords:Reinforcement Learning, Exploration, Contingency-Awareness
TL;DR:We investigate contingency-awareness and controllable aspects in exploration and achieve state-of-the-art performance on Montezuma's Revenge without expert demonstrations.
Context-adaptive Entropy Model for End-to-end Optimized Image Compression 
Keywords:image compression, deep learning, entropy model
TL;DR:Context-adaptive entropy model for use in end-to-end optimized image compression, which significantly improves compression performance
Variational Discriminator Bottleneck: Improving Imitation Learning, Inverse RL, and GANs by Constraining Information Flow 
Keywords:reinforcement learning, generative adversarial networks, imitation learning, inverse reinforcement learning, information bottleneck
TL;DR:Regularizing adversarial learning with an information bottleneck, applied to imitation learning, inverse reinforcement learning, and generative adversarial networks.
Meta-learning with differentiable closed-form solvers 
Keywords:few-shot learning, one-shot learning, meta-learning, deep learning, ridge regression, classification
TL;DR:We propose a meta-learning approach for few-shot classification that achieves strong performance at high-speed by back-propagating through the solution of fast solvers, such as ridge regression or logistic regression.
Learning Self-Imitating Diverse Policies 
Keywords:Reinforcement-learning, Imitation-learning, Ensemble-training
TL;DR:Policy optimization by using past good rollouts from the agent; learning shaped rewards via divergence minimization; SVPG with JS-kernel for population-based exploration.
ProxQuant: Quantized Neural Networks via Proximal Operators 
Keywords:Model quantization, Optimization, Regularization
TL;DR:A principled framework for model quantization using the proximal gradient method, with empirical evaluation and theoretical convergence analyses.
Universal Transformers 
Keywords:sequence-to-sequence, rnn, transformer, machine translation, language understanding, learning to execute
TL;DR:We introduce the Universal Transformer, a self-attentive parallel-in-time recurrent sequence model that outperforms Transformers and LSTMs on a wide range of sequence-to-sequence tasks, including machine translation.
Learning to Adapt in Dynamic, Real-World Environments through Meta-Reinforcement Learning 
Keywords:meta-learning, reinforcement learning, meta reinforcement learning, online adaptation
TL;DR:A model-based meta-RL algorithm that enables a real robot to adapt online in dynamic environments
L-Shapley and C-Shapley: Efficient Model Interpretation for Structured Data 
Keywords:Model Interpretation, Feature Selection
TL;DR:We develop two linear-complexity algorithms for model-agnostic model interpretation based on the Shapley value, in the settings where the contribution of features to the target is well-approximated by a graph-structured factorization.
Discovery of Natural Language Concepts in Individual Units of CNNs 
Keywords:interpretability of deep neural networks, natural language representation
TL;DR:We show that individual units in CNN representations learned in NLP tasks are selectively responsive to natural language concepts.
Towards the first adversarially robust neural network model on MNIST 
Keywords:None
TL;DR:None
Discriminator Rejection Sampling 
Keywords:GANs, rejection sampling
TL;DR:We use a GAN discriminator to perform an approximate rejection sampling scheme on the output of the GAN generator.
Harmonic Unpaired Image-to-image Translation 
Keywords:unpaired image-to-image translation, cyclegan, smoothness constraint
TL;DR:Smooth regularization over sample graph for unpaired image-to-image translation results in significantly improved consistency
Universal Successor Features Approximators 
Keywords:None
TL;DR:None
Gradient Descent Provably Optimizes Over-parameterized Neural Networks 
Keywords:theory, non-convex optimization, overparameterization, gradient descent
TL;DR:We prove gradient descent achieves zero training loss with a linear rate on over-parameterized neural networks.
Opportunistic Learning: Budgeted Cost-Sensitive Learning from Data Streams 
Keywords:Cost-Aware Learning, Feature Acquisition, Reinforcement Learning, Stream Learning, Deep Q-Learning
TL;DR:An online algorithm for cost-aware feature acquisition and prediction
DARTS: Differentiable Architecture Search 
Keywords:deep learning, autoML, neural architecture search, image classification, language modeling
TL;DR:We propose a differentiable architecture search algorithm for both convolutional and recurrent networks, achieving competitive performance with the state of the art using orders of magnitude less computation resources.
Feature-Wise Bias Amplification 
Keywords:None
TL;DR:None
The relativistic discriminator: a key element missing from standard GAN 
Keywords:Improving the quality and stability of GANs using a relativistic discriminator; IPM GANs (such as WGAN-GP) are a special case.
TL;DR:AI, deep learning, generative models, GAN
Understanding and Improving Interpolation in Autoencoders via an Adversarial Regularizer 
Keywords:We propose a regularizer that improves interpolation and autoencoders and show that it also improves the learned representation for downstream tasks.
TL;DR:autoencoders, interpolation, unsupervised learning, representation learning, adversarial learning
Quasi-hyperbolic momentum and Adam for deep learning 
Keywords:sgd, momentum, nesterov, adam, qhm, qhadam, optimization
TL;DR:Mix plain SGD and momentum (or do something similar with Adam) for great profit.
Local SGD Converges Fast and Communicates Little 
Keywords:optimization, communication, theory, stochastic gradient descent, SGD, mini-batch, local SGD, parallel restart SGD, distributed training
TL;DR:We prove that parallel local SGD achieves linear speedup with much lesser communication than parallel mini-batch SGD.
Learning Finite State Representations of Recurrent Policy Networks 
Keywords:recurrent neural networks, finite state machine, quantization, interpretability, autoencoder, moore machine, reinforcement learning, imitation learning, representation, Atari, Tomita
TL;DR:Extracting a finite state machine from a recurrent neural network via quantization for the purpose of interpretability with experiments on Atari.
Multilingual Neural Machine Translation with Knowledge Distillation 
Keywords:NMT, Multilingual NMT, Knowledge Distillation
TL;DR:We proposed a knowledge distillation based method to boost the accuracy of multilingual neural machine translation.
MisGAN: Learning from Incomplete Data with Generative Adversarial Networks 
Keywords:generative models, missing data
TL;DR:This paper presents a GAN-based framework for learning the distribution from high-dimensional incomplete data.
A Direct Approach to Robust Deep Learning Using Adversarial Networks 
Keywords:deep learning, adversarial learning, generative adversarial networks
TL;DR:Jointly train an adversarial noise generating network with a classification network to provide better robustness to adversarial attacks.
Combinatorial Attacks on Binarized Neural Networks 
Keywords:binarized neural networks, combinatorial optimization, integer programming
TL;DR:Gradient-based attacks on binarized neural networks are not effective due to the non-differentiability of such networks; Our IPROP algorithm solves this problem using integer optimization
Exemplar Guided Unsupervised Image-to-Image Translation with Semantic Consistency 
Keywords:image-to-image translation, image generation, domain adaptation
TL;DR:We propose the Exemplar Guided & Semantically Consistent Image-to-image Translation (EGSC-IT) network which conditions the translation process on an exemplar image in the target domain.
ARM: Augment-REINFORCE-Merge Gradient for Stochastic Binary Networks 
Keywords:Antithetic sampling, variable augmentation, deep discrete latent variable models, variance reduction, variational auto-encoder
TL;DR:An unbiased and low-variance gradient estimator for discrete latent variable models
Building Dynamic Knowledge Graphs from Text using Machine Reading Comprehension 
Keywords:None
TL;DR:None
Information asymmetry in KL-regularized RL 
Keywords:Deep Reinforcement Learning, Continuous Control, RL as Inference
TL;DR:Limiting state information for the default policy can improvement performance, in a KL-regularized RL framework where both agent and default policy are optimized together
TimbreTron: A WaveNet(CycleGAN(CQT(Audio))) Pipeline for Musical Timbre Transfer 
Keywords:Generative models, Timbre Transfer, Wavenet, CycleGAN
TL;DR:We present the TimbreTron, a pipeline for perfoming high-quality timbre transfer on musical waveforms using CQT-domain style transfer.
Whitening and Coloring Batch Transform for GANs 
Keywords:None
TL;DR:None
Learnable Embedding Space for Efficient Neural Architecture Compression 
Keywords:Network Compression, Neural Architecture Search, Bayesian Optimization, Architecture Embedding
TL;DR:We propose a method to incrementally learn an embedding space over the domain of network architectures, to enable the careful selection of architectures for evaluation during compressed architecture search.
On the Sensitivity of Adversarial Robustness to Input Data Distributions 
Keywords:adversarial robustness, adversarial training, PGD training, adversarial perturbation, input data distribution
TL;DR:Robustness performance of PGD trained models are sensitive to semantics-preserving transformation of image datasets, which implies the trickiness of evaluation of robust learning algorithms in practice.
Minimal Images in Deep Neural Networks: Fragile Object Recognition in Natural Images 
Keywords:None
TL;DR:None
A Statistical Approach to Assessing Neural Network Robustness 
Keywords:neural network verification, multi-level splitting, formal verification
TL;DR:We introduce a statistical approach to assessing neural network robustness that provides an informative notion of how robust a network is, rather than just the conventional binary assertion of whether or not of property is violated.
Improving Sequence-to-Sequence Learning via Optimal Transport 
Keywords:None
TL;DR:None
PATE-GAN: Generating Synthetic Data with Differential Privacy Guarantees 
Keywords:None
TL;DR:None
Integer Networks for Data Compression with Latent-Variable Models 
Keywords:data compression, variational models, network quantization
TL;DR:We train variational models with quantized networks for computational determinism. This enables using them for cross-platform data compression.
Value Propagation Networks 
Keywords:Reinforcement Learning, Value Iteration, Navigation, Convolutional Neural Networks, Learning to plan
TL;DR:We present planners based on convnets that are sample-efficient and that generalize to larger instances of navigation and pathfinding problems.
Bayesian Policy Optimization for Model Uncertainty 
Keywords:Bayes-Adaptive Markov Decision Process, Model Uncertainty, Bayes Policy Optimization
TL;DR:We formulate model uncertainty in Reinforcement Learning as a continuous Bayes-Adaptive Markov Decision Process and present a method for practical and scalable Bayesian policy optimization.
Variational Bayesian Phylogenetic Inference 
Keywords:Bayesian phylogenetic inference, Variational inference, Subsplit Bayesian networks
TL;DR:The first variational Bayes formulation of phylogenetic inference, a challenging inference problem over structures with intertwined discrete and continuous components
LEARNING FACTORIZED REPRESENTATIONS FOR OPEN-SET DOMAIN ADAPTATION 
Keywords:None
TL;DR:None
On the Universal Approximability and Complexity Bounds of Quantized ReLU Neural Networks 
Keywords:Quantized Neural Networks, Universial Approximability, Complexity Bounds, Optimal Bit-width
TL;DR:This paper proves the universal approximability of quantized ReLU neural networks and puts forward the complexity bound given arbitrary error.
Learning Localized Generative Models for 3D Point Clouds via Graph Convolution 
Keywords:A GAN using graph convolution operations with dynamically computed graphs from hidden features
TL;DR:GAN, graph convolution, point clouds
ACCELERATING NONCONVEX LEARNING VIA REPLICA EXCHANGE LANGEVIN DIFFUSION 
Keywords:None
TL;DR:None
Dynamically Unfolding Recurrent Restorer: A Moving Endpoint Control Method for Image Restoration 
Keywords:image restoration, differential equation
TL;DR:We propose a novel method to handle image degradations of different levels by learning a diffusion terminal time. Our model can generalize to unseen degradation level and different noise statistic.
Bias-Reduced Uncertainty Estimation for Deep Neural Classifiers 
Keywords:Uncertainty estimation, Deep learning
TL;DR:We use snapshots from the training process to improve any uncertainty estimation method of a DNN classifier.
CAMOU: Learning Physical Vehicle Camouflages to Adversarially Attack Detectors in the Wild 
Keywords:Adversarial Attack, Object Detection, Synthetic Simulation
TL;DR:We propose a method to learn physical vehicle camouflage to adversarially attack object detectors in the wild. We find our camouflage effective and transferable.
Learning Latent Superstructures in Variational Autoencoders for Deep Multidimensional Clustering 
Keywords:latent tree model, variational autoencoder, deep learning, latent variable model, bayesian network, structure learning, stepwise em, message passing, graphical model, multidimensional clustering, unsupervised learning
TL;DR:We investigate a variant of variational autoencoders where there is a superstructure of discrete latent variables on top of the latent features.
Learning Programmatically Structured Representations with Perceptor Gradients 
Keywords:None
TL;DR:None
Variational Autoencoders with Jointly Optimized Latent Dependency Structure 
Keywords:deep generative models, structure learning
TL;DR:We propose a method for learning latent dependency structure in variational autoencoders.
The Unusual Effectiveness of Averaging in GAN Training 
Keywords:None
TL;DR:None
Beyond Pixel Norm-Balls: Parametric Adversaries using an Analytically Differentiable Renderer 
Keywords:adversarial examples, norm-balls, differentiable renderer
TL;DR:Enabled by a novel differentiable renderer, we propose a new metric that has real-world implications for evaluating adversarial machine learning algorithms, resolving the lack of realism of the existing metric based on pixel norms.
Diversity is All You Need: Learning Skills without a Reward Function 
Keywords:reinforcement learning, unsupervised learning, skill discovery
TL;DR:We propose an algorithm for learning useful skills without a reward function, and show how these skills can be used to solve downstream tasks.
Supervised Policy Update for Deep Reinforcement Learning 
Keywords:Deep Reinforcement Learning
TL;DR:first posing and solving the sample efficiency optimization problem in the non-parameterized policy space, and then solving a supervised regression problem to find a parameterized policy that is near the optimal non-parameterized policy.
Learning sparse relational transition models 
Keywords:Deictic reference, relational model, rule-based transition model
TL;DR:A new approach that learns a representation for describing transition models in complex uncertaindomains using relational rules.
Learning to Schedule Communication in Multi-agent Reinforcement Learning 
Keywords:None
TL;DR:None
Hierarchical RL Using an Ensemble of Proprioceptive Periodic Policies 
Keywords:None
TL;DR:None
Multi-class classification without multi-class labels 
Keywords:None
TL;DR:None
What do you learn from context? Probing for sentence structure in contextualized word representations 
Keywords:natural language processing, word embeddings, transfer learning, interpretability
TL;DR:We probe for sentence structure in ELMo and related contextual embedding models. We find existing models efficiently encode syntax and show evidence of long-range dependencies, but only offer small improvements on semantic tasks.
Spectral Inference Networks: Unifying Deep and Spectral Learning 
Keywords:spectral learning, unsupervised learning, manifold learning, dimensionality reduction
TL;DR:We show how to learn spectral decompositions of linear operators with deep learning, and use it for unsupervised learning without a generative model.
PeerNets: Exploiting Peer Wisdom Against Adversarial Attacks 
Keywords:None
TL;DR:None
Attentive Neural Processes 
Keywords:Neural Processes, Conditional Neural Processes, Stochastic Processes, Regression, Attention
TL;DR:A model for regression that learns conditional distributions of a stochastic process, by incorporating attention into Neural Processes.
Representation Degeneration Problem in Training Natural Language Generation Models 
Keywords:None
TL;DR:None
Hierarchical interpretations for neural network predictions 
Keywords:interpretability, natural language processing, computer vision
TL;DR:We introduce and validate hierarchical local interpretations, the first technique to automatically search for and display important interactions for individual predictions made by LSTMs and CNNs.
Spreading vectors for similarity search 
Keywords:dimensionality reduction, similarity search, indexing, differential entropy
TL;DR:We learn a neural network that uniformizes the input distribution, which leads to competitive indexing performance in high-dimensional space
A Convergence Analysis of Gradient Descent for Deep Linear Neural Networks 
Keywords:Deep Learning, Learning Theory, Non-Convex Optimization
TL;DR:We analyze gradient descent for deep linear neural networks, providing a guarantee of convergence to global optimum at a linear rate.
Feed-forward Propagation in Probabilistic Neural Networks with Categorical and Max Layers 
Keywords:probabilistic neural network, uncertainty, dropout, bayesian, softmax, argmax, logsumexp
TL;DR:Approximating mean and variance of the NN output over noisy input / dropout / uncertain parameters. Analytic approximations for argmax, softmax and max layers.
Measuring and regularizing networks in function space 
Keywords:function space, Hilbert space, empirical characterization, multitask learning, catastrophic forgetting, optimization, natural gradient
TL;DR:We find movement in function space is not proportional to movement in parameter space during optimization. We propose a new natural-gradient style optimizer to address this.
Fluctuation-dissipation relations for stochastic gradient descent 
Keywords:stochastic gradient descent, adaptive method, loss surface, Hessian
TL;DR:We prove fluctuation-dissipation relations for SGD, which can be used to (i) adaptively set learning rates and (ii) probe loss surfaces.
Poincare Glove: Hyperbolic Word Embeddings 
Keywords:word embeddings, hyperbolic spaces, poincare ball, hypernymy, analogy, similarity, gaussian embeddings
TL;DR:We embed words in the hyperbolic space and make the connection with the Gaussian word embeddings.
Episodic Curiosity through Reachability 
Keywords:deep learning, reinforcement learning, curiosity, exploration, episodic memory
TL;DR:We propose a novel model of curiosity based on episodic memory and the ideas of reachability which allows us to overcome the known "couch-potato" issues of prior work.
Phase-Aware Speech Enhancement with Deep Complex U-Net 
Keywords:speech enhancement, deep learning, complex neural networks, phase estimation
TL;DR:This paper proposes a novel complex masking method for speech enhancement along with a loss function for efficient phase estimation.
Generative predecessor models for sample-efficient imitation learning 
Keywords:None
TL;DR:None
Adaptive Estimators Show Information Compression in Deep Neural Networks 
Keywords:deep neural networks, mutual information, information bottleneck, noise, L2 regularization
TL;DR:We developed robust mutual information estimates for DNNs and used them to observe compression in networks with non-saturating activation functions
Multilingual Neural Machine Translation With Soft Decoupled Encoding 
Keywords:None
TL;DR:None
Approximating CNNs with Bag-of-local-Features models works surprisingly well on ImageNet 
Keywords:interpretability, representation learning, bag of features, deep learning, object recognition
TL;DR:Aggregating class evidence from many small image patches suffices to solve ImageNet, yields more interpretable models and can explain aspects of the decision-making of popular DNNs.
Reward Constrained Policy Optimization 
Keywords:reinforcement learning, markov decision process, constrained markov decision process, deep learning
TL;DR:For complex constraints in which it is not easy to estimate the gradient, we use the discounted penalty as a guiding signal. We prove that under certain assumptions it converges to a feasible solution.
On the Relation Between the Sharpest Directions of DNN Loss and the SGD Step Length 
Keywords:optimization, generalization, theory of deep learning, SGD, hessian
TL;DR:SGD is steered early on in training towards a region in which its step is too large compared to curvature, which impacts the rest of training.
Modeling the Long Term Future in Model-Based Reinforcement Learning 
Keywords:model-based reinforcement learning, variation inference
TL;DR:incorporating, in the model, latent variables that encode future content improves the long-term prediction accuracy, which is critical for better planning in model-based RL.
Understanding Straight-Through Estimator in Training Activation Quantized Neural Nets 
Keywords:straight-through estimator, quantized activation, binary neuron
TL;DR:We make theoretical justification for the concept of straight-through estimator.
DISTRIBUTIONAL CONCAVITY REGULARIZATION FOR GANS 
Keywords:None
TL;DR:None
LeMoNADe: Learned Motif and Neuronal Assembly Detection in calcium imaging videos 
Keywords:VAE, unsupervised learning, neuronal assemblies, calcium imaging analysis
TL;DR:We present LeMoNADe, an end-to-end learned motif detection method directly operating on calcium imaging videos.
Competitive experience replay 
Keywords:reinforcement learning, sparse reward, goal-based learning
TL;DR:a novel method to learn with sparse reward using adversarial reward re-labeling
Multi-Domain Adversarial Learning 
Keywords:multi-domain learning, domain adaptation, adversarial learning, H-divergence, deep representation learning, high-content microscopy
TL;DR:Adversarial Domain adaptation and Multi-domain learning: a new loss to handle multi- and single-domain classes in the semi-supervised setting.
ProMP: Proximal Meta-Policy Search 
Keywords:Meta-Reinforcement Learning, Meta-Learning, Reinforcement-Learning
TL;DR:A novel and theoretically grounded meta-reinforcement learning algorithm
Don't Settle for Average, Go for the Max: Fuzzy Sets and Max-Pooled Word Vectors 
Keywords:word vectors, sentence representations, distributed representations, fuzzy sets, bag-of-words, unsupervised learning, word vector compositionality, max-pooling, Jaccard index
TL;DR:Max-pooled word vectors with fuzzy Jaccard set similarity are an extremely competitive baseline for semantic similarity; we propose a simple dynamic variant that performs even better.
Stable Opponent Shaping in Differentiable Games 
Keywords:multi-agent learning, multiple interacting losses, opponent shaping, exploitation, convergence
TL;DR:Opponent shaping is a powerful approach to multi-agent learning but can prevent convergence; our SOS algorithm fixes this with strong guarantees in all differentiable games.
A Mean Field Theory of Batch Normalization 
Keywords:theory, batch normalization, mean field theory, trainability
TL;DR:Batch normalization causes exploding gradients in vanilla feedforward networks.
Learning Exploration Policies for Navigation 
Keywords:None
TL;DR:None
Distribution-Interpolation Trade off in Generative Models 
Keywords:generative models, latent distribution, Cauchy distribution, interpolations
TL;DR:We theoretically prove that linear interpolations are unsuitable for analysis of trained implicit generative models.
Learning to Describe Scenes with Programs 
Keywords:Structured scene representations, program synthesis
TL;DR:We present scene programs, a structured scene representation that captures both low-level object appearance and high-level regularity in the scene.
Visceral Machines: Risk-Aversion in Reinforcement Learning with Intrinsic Physiological Rewards 
Keywords:Reinforcement Learning, Simulation, Affective Computing
TL;DR:We present a novel approach to reinforcement learning that leverages a task-independent intrinsic reward function trained on peripheral pulse measurements that are correlated with human autonomic nervous system responses.
Deep Frank-Wolfe For Neural Network Optimization 
Keywords:optimization, conditional gradient, Frank-Wolfe, SVM
TL;DR:We train neural networks by locally linearizing them and using a linear SVM solver (Frank-Wolfe) at each iteration.
LEARNING TO PROPAGATE LABELS: TRANSDUCTIVE PROPAGATION NETWORK FOR FEW-SHOT LEARNING 
Keywords:few-shot learning, meta-learning, label propagation, manifold learning
TL;DR:We propose a novel meta-learning framework for transductive inference that classifies the entire test set at once to alleviate the low-data problem.
Improving the Generalization of Adversarial Training with Domain Adaptation 
Keywords:adversarial training, domain adaptation, adversarial example, deep learning
TL;DR:We propose a novel adversarial training with domain adaptation method that significantly improves the generalization ability on adversarial examples from different attacks.
Dimensionality Reduction for Representing the Knowledge of Probabilistic Models 
Keywords:metric learning, distance learning, dimensionality reduction, bound guarantees
TL;DR:dimensionality reduction for cases where examples can be represented as soft probability distributions
Learning protein sequence embeddings using information from structure 
Keywords:sequence embedding, sequence alignment, RNN, LSTM, protein structure, amino acid sequence, contextual embeddings, transmembrane prediction
TL;DR:We present a method for learning protein sequence embedding models using structural information in the form of global structural similarity between proteins and within protein residue-residue contacts.
Variational Smoothing in Recurrent Neural Network Language Models 
Keywords:None
TL;DR:None
Biologically-Plausible Learning Algorithms Can Scale to Large Datasets 
Keywords:biologically plausible learning algorithm, ImageNet, sign-symmetry, feedback alignment
TL;DR:Biologically plausible learning algorithms, particularly sign-symmetry, work well on ImageNet
Coarse-grain Fine-grain Coattention Network for Multi-evidence Question Answering 
Keywords:question answering, reading comprehension, nlp, natural language processing, attention, representation learning
TL;DR:A new state-of-the-art model for multi-evidence question answering using coarse-grain fine-grain hierarchical attention.
Learning a Meta-Solver for Syntax-Guided Program Synthesis 
Keywords:Syntax-guided Synthesis, Context Free Grammar, Logical Specification, Representation Learning, Meta Learning, Reinforcement Learning
TL;DR:We propose a meta-learning framework that learns a transferable policy from only weak supervision to solve synthesis tasks with different logical specifications and grammars.
Towards Robust, Locally Linear Deep Networks 
Keywords:robust derivatives, transparency, interpretability
TL;DR:A scalable algorithm to establish robust derivatives of deep networks w.r.t. the inputs.
How Important is a Neuron 
Keywords:None
TL;DR:None
Learning to Make Analogies by Contrasting Abstract Relational Structure 
Keywords:cognitive science, analogy, psychology, cognitive theory, cognition, abstraction, generalization
TL;DR:The most robust capacity for analogical reasoning is induced when networks learn analogies by contrasting abstract relational structures in their input domains.
Learning what you can do before doing anything 
Keywords:unsupervised learning, vision, motion, action space, video prediction, variational models
TL;DR:We learn a representation of an agent's action space from pure visual observations. We use a recurrent latent variable approach with a novel composability loss.
Learning Grid Cells as Vector Representation of Self-Position Coupled with Matrix Representation of Self-Motion 
Keywords:None
TL;DR:None
Universal Stagewise Learning for Non-Convex Problems with Convergence on Averaged Solutions 
Keywords:None
TL;DR:None
Invariant and Equivariant Graph Networks 
Keywords:graph learning, equivariance, deep learning
TL;DR:The paper provides a full characterization of permutation invariant and equivariant linear layers for graph data.
Robustness May Be at Odds with Accuracy 
Keywords:adversarial examples, robust machine learning, robust optimization, deep feature representations
TL;DR:We show that adversarial robustness might come at the cost of standard classification performance, but also yields unexpected benefits.
Feature Intertwiner for Object Detection 
Keywords:feature learning, computer vision, deep learning
TL;DR:(Camera-ready version) A feature intertwiner module to leverage features from one accurate set to help the learning of another less reliable set.
Adversarial Reprogramming of Neural Networks 
Keywords:Adversarial, Neural Networks, Machine Learning Security
TL;DR:We introduce the first instance of adversarial attacks that reprogram the target model to perform a task chosen by the attacker---without the attacker needing to specify or compute the desired output for each test-time input.
G-SGD: Optimizing ReLU Neural Networks in its Positively Scale-Invariant Space 
Keywords:None
TL;DR:None
From Hard to Soft: Understanding Deep Network Nonlinearities via Vector Quantization and Statistical Inference 
Keywords:Spline, Vector Quantization, Inference, Nonlinearities, Deep Network
TL;DR:Reformulate deep networks nonlinearities from a vector quantization scope and bridge most known nonlinearities together.
Aggregated Momentum: Stability Through Passive Damping 
Keywords:momentum, optimization, deep learning, neural networks
TL;DR:We introduce a simple variant of momentum optimization which is able to outperform classical momentum, Nesterov, and Adam on deep learning tasks with minimal hyperparameter tuning.
Variational Autoencoder with Arbitrary Conditioning 
Keywords:unsupervised learning, generative models, conditional variational autoencoder, variational autoencoder, missing features multiple imputation, inpainting
TL;DR:We propose an extension of conditional variational autoencoder that allows conditioning on an arbitrary subset of the features and sampling the remaining ones.
Time-Agnostic Prediction: Predicting Predictable Video Frames 
Keywords:visual prediction, subgoal generation, bottleneck states, time-agnostic
TL;DR:In visual prediction tasks, letting your predictive model choose which times to predict does two things: (i) improves prediction quality, and (ii) leads to semantically coherent "bottleneck state" predictions, which are useful for planning.
A Closer Look at Deep Learning Heuristics: Learning rate restarts, Warmup and Distillation 
Keywords:deep learning heuristics, learning rate restarts, learning rate warmup, knowledge distillation, mode connectivity, SVCCA
TL;DR:We use empirical tools of mode connectivity and SVCCA to investigate neural network training heuristics of learning rate restarts, warmup and knowledge distillation.
Self-Monitoring Navigation Agent via Auxiliary Progress Estimation 
Keywords:visual grounding, textual grounding, instruction-following, navigation agent
TL;DR:We propose a self-monitoring agent for the Vision-and-Language Navigation task.
Kernel Change-point Detection with Auxiliary Deep Generative Models 
Keywords:deep kernel learning, generative models, kernel two-sample test, time series change-point detection
TL;DR:In this paper, we propose KL-CPD, a novel kernel learning framework for time series CPD that optimizes a lower bound of test power via an auxiliary generative model as a surrogate to the abnormal distribution.
Unsupervised Learning via Meta-Learning 
Keywords:unsupervised learning, meta-learning
TL;DR:An unsupervised learning method that uses meta-learning to enable efficient learning of downstream image classification tasks, outperforming state-of-the-art methods.
Auxiliary Variational MCMC 
Keywords:None
TL;DR:None
Neural network gradient-based learning of black-box function interfaces 
Keywords:neural networks, black box functions, gradient descent
TL;DR:Training DNNs to interface w\ black box functions w\o intermediate labels by using an estimator sub-network that can be replaced with the black box after training
Self-Tuning Networks: Bilevel Optimization of Hyperparameters using Structured Best-Response Functions 
Keywords:hyperparameter optimization, game theory, optimization
TL;DR:We use a hypernetwork to predict optimal weights given hyperparameters, and jointly train everything together.
Unsupervised Control Through Non-Parametric Discriminative Rewards 
Keywords:deep reinforcement learning, goals, UVFA, mutual information
TL;DR:Unsupervised reinforcement learning method for learning a policy to robustly achieve perceptually specified goals.
Interpolation-Prediction Networks for Irregularly Sampled Time Series 
Keywords:irregular sampling, multivariate time series, supervised learning, interpolation, missing data
TL;DR:This paper presents a new deep learning architecture for addressing the problem of supervised learning with sparse and irregularly sampled multivariate time series.
Riemannian Adaptive Optimization Methods 
Keywords:Riemannian optimization, adaptive, hyperbolic, curvature, manifold, adam, amsgrad, adagrad, rsgd, convergence
TL;DR:Adapting Adam, Amsgrad, Adagrad to Riemannian manifolds.
Minimal Random Code Learning: Getting Bits Back from Compressed Model Parameters 
Keywords:compression, neural networks, bits-back argument, Bayesian, Shannon, information theory
TL;DR:This paper proposes an effective method to compress neural networks based on recent results in information theory.
Characterizing Audio Adversarial Examples Using Temporal Dependency 
Keywords:audio adversarial example, mitigation, detection, machine learning
TL;DR:Adversarial audio discrimination using temporal dependency
Equi-normalization of Neural Networks 
Keywords:convolutional neural networks, Normalization, Sinkhorn, Regularization
TL;DR:Fast iterative algorithm to balance the energy of a network while staying in the same functional equivalence class
Generalized Tensor Models for Recurrent Neural Networks 
Keywords:expressive power, recurrent neural networks, Tensor-Train decomposition
TL;DR:Analysis of expressivity and generality of recurrent neural networks with ReLu nonlinearities using Tensor-Train decomposition.
Wizard of Wikipedia: Knowledge-Powered Conversational Agents 
Keywords:dialogue, knowledge, language, conversation
TL;DR:We build knowledgeable conversational agents by conditioning on Wikipedia + a new supervised task.
Are adversarial examples inevitable? 
Keywords:adversarial examples, neural networks, security
TL;DR:This paper identifies classes of problems for which adversarial examples are inescapable, and derives fundamental bounds on the susceptibility of any classifier to adversarial examples.
A Variational Inequality Perspective on Generative Adversarial Networks 
Keywords:optimization, variational inequality, games, saddle point, extrapolation, averaging, extragradient, generative modeling, generative adversarial network
TL;DR:We cast GANs in the variational inequality framework and import techniques from this literature to optimize GANs better; we give algorithmic extensions and empirically test their performance for training GANs.
Learning-Based Frequency Estimation Algorithms 
Keywords:streaming algorithms, heavy-hitters, Count-Min, Count-Sketch
TL;DR:Data stream algorithms can be improved using deep learning, while retaining performance guarantees.
From Language to Goals: Inverse Reinforcement Learning for Vision-Based Instruction Following 
Keywords:inverse reinforcement learning, language grounding, instruction following, language-based learning
TL;DR:We ground language commands in a high-dimensional visual environment by learning language-conditioned rewards using inverse reinforcement learning.
Backpropamine: training self-modifying neural networks with differentiable neuromodulated plasticity 
Keywords:meta-learning, reinforcement learning, plasticity, neuromodulation, Hebbian learning, recurrent neural networks
TL;DR:Neural networks can be trained to modify their own connectivity, improving their online learning performance on challenging tasks.
Recurrent Experience Replay in Distributed Reinforcement Learning 
Keywords:RNN, LSTM, experience replay, distributed training, reinforcement learning
TL;DR:Investigation on combining recurrent neural networks and experience replay leading to state-of-the-art agent on both Atari-57 and DMLab-30 using single set of hyper-parameters.
A Generative Model For Electron Paths 
Keywords:Molecules, Reaction Prediction, Graph Neural Networks, Deep Generative Models
TL;DR:A generative model for reaction prediction that learns the mechanistic electron steps of a reaction directly from raw reaction data.
Modeling Uncertainty with Hedged Instance Embeddings 
Keywords:uncertainty, instance embedding, metric learning, probabilistic embedding
TL;DR:The paper proposes using probability distributions instead of points for instance embeddings tasks such as recognition and verification.
Beyond Greedy Ranking: Slate Optimization via List-CVAE 
Keywords:CVAE, VAE, recommendation system, slate optimization, whole page optimization
TL;DR:We used a CVAE type model structure to learn to directly generate slates/whole pages for recommendation systems.
Stochastic Prediction of Multi-Agent Interactions from Partial Observations 
Keywords:Dynamics modeling, partial observations, multi-agent interactions, predictive models
TL;DR:We present a method which learns to integrate temporal information and ambiguous visual information in the context of interacting agents.
GamePad: A Learning Environment for Theorem Proving 
Keywords:Theorem proving, ITP, systems, neural embeddings
TL;DR:We introduce a system called GamePad to explore the application of machine learning methods to theorem proving in the Coq proof assistant.
GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding 
Keywords:natural language understanding, multi-task learning, evaluation
TL;DR:We present a multi-task benchmark and analysis platform for evaluating generalization in natural language understanding systems.
On Computation and Generalization of Generative Adversarial Networks under Spectrum Control 
Keywords:None
TL;DR:None
Large-Scale Study of Curiosity-Driven Learning 
Keywords:exploration, curiosity, intrinsic reward, no extrinsic reward, unsupervised, no-reward, skills
TL;DR:An agent trained only with curiosity, and no extrinsic reward, does surprisingly well on 54 popular environments, including the suite of Atari games, Mario etc.
Unsupervised Discovery of Parts, Structure, and Dynamics 
Keywords:Self-Supervised Learning, Visual Prediction, Hierarchical Models
TL;DR:Learning object parts, hierarchical structure, and dynamics by watching how they move
Music Transformer: Generating Music with Long-Term Structure 
Keywords:music generation
TL;DR:We show the first successful use of Transformer in generating music that exhibits long-term structure.
BabyAI: A Platform to Study the Sample Efficiency of Grounded Language Learning 
Keywords:language, learning, efficiency, imitation learning, reinforcement learning
TL;DR:We present the BabyAI platform for studying data efficiency of language learning with a human in the loop
Analyzing Inverse Problems with Invertible Neural Networks 
Keywords:Inverse problems, Neural Networks, Uncertainty, Invertible Neural Networks
TL;DR:To analyze inverse problems with Invertible Neural Networks
RelGAN: Relational Generative Adversarial Networks for Text Generation 
Keywords:None
TL;DR:None
The Singular Values of Convolutional Layers 
Keywords:singular values, operator norm, convolutional layers, regularization
TL;DR:We characterize the singular values of the linear transformation associated with a standard 2D multi-channel convolutional layer, enabling their efficient computation.
An Empirical study of Binary Neural Networks' Optimisation 
Keywords:None
TL;DR:None
Approximability of Discriminators Implies Diversity in GANs 
Keywords:Theory, Generative adversarial networks, Mode collapse, Generalization
TL;DR:GANs can in principle learn distributions sample-efficiently, if the discriminator class is compact and has strong distinguishing power against the particular generator class.
Learning Embeddings into Entropic Wasserstein Spaces 
Keywords:Embedding, Wasserstein, Sinkhorn, Optimal Transport
TL;DR:We show that Wasserstein spaces are good targets for embedding data with complex semantic structure.
DeepOBS: A Deep Learning Optimizer Benchmark Suite 
Keywords:deep learning, optimization
TL;DR:We provide a software package that drastically simplifies, automates, and improves the evaluation of deep learning optimizers.
InfoBot: Transfer and Exploration via the Information Bottleneck 
Keywords:Information bottleneck, policy transfer, policy generalization, exploration
TL;DR:Training agents with goal-policy information bottlenecks promotes transfer and yields a powerful exploration bonus
The Comparative Power of ReLU Networks and Polynomial Kernels in the Presence of Sparse Latent Structure 
Keywords:theory, representational power, universal approximators, polynomial kernels, latent sparsity, beyond worst case, separation result
TL;DR:Beyond-worst-case analysis of the representational power of ReLU nets & polynomial kernels -- in particular in the presence of sparse latent structure.
Learning Implicitly Recurrent CNNs Through Parameter Sharing 
Keywords:deep learning, architecture search, computer vision
TL;DR:We propose a method that enables CNN folding to create recurrent connections
Learning Particle Dynamics for Manipulating Rigid Bodies, Deformable Objects, and Fluids 
Keywords:Dynamics modeling, Control, Particle-Based Representation
TL;DR:Learning particle dynamics with dynamic interaction graphs for simulating and control rigid bodies, deformable objects, and fluids.
Regularized Learning for Domain Adaptation under Label Shifts 
Keywords:Deep Learning, Domain Adaptation, Label Shift, Importance Weights, Generalization
TL;DR:A practical and provably guaranteed approach for training efficiently classifiers in the presence of label shifts between Source and Target data sets
Von Mises-Fisher Loss for Training Sequence to Sequence Models with Continuous Outputs 
Keywords:Language Generation, Regression, Word Embeddings, Machine Translation
TL;DR:Language generation using seq2seq models which produce word embeddings instead of a softmax based distribution over the vocabulary at each step enabling much faster training while maintaining generation quality
Relational Forward Models for Multi-Agent Learning 
Keywords:multi-agent reinforcement learning, relational reasoning, forward models
TL;DR:Relational Forward Models for multi-agent learning make accurate predictions of agents' future behavior, they produce intepretable representations and can be used inside agents.
Imposing Category Trees Onto Word-Embeddings Using A Geometric Construction 
Keywords:category tree, word-embeddings, geometry
TL;DR:we show a geometric method to perfectly encode categroy tree information into pre-trained word-embeddings.
Two-Timescale Networks for Nonlinear Value Function Approximation 
Keywords:Reinforcement learning, policy evaluation, nonlinear function approximation
TL;DR:We propose an architecture for learning value functions which allows the use of any linear policy evaluation algorithm in tandem with nonlinear feature learning.
Diversity-Sensitive Conditional Generative Adversarial Networks 
Keywords:Conditional Generative Adversarial Network, mode-collapse, multi-modal generation, image-to-image translation, image in-painting, video prediction
TL;DR:We propose a simple and general approach that avoids a mode collapse problem in various conditional GANs.
Query-Efficient Hard-label Black-box Attack: An Optimization-based Approach 
Keywords:None
TL;DR:None
Rethinking the Value of Network Pruning 
Keywords:In structured network pruning, fine-tuning a pruned model only gives comparable performance with training it from scratch.
TL;DR:network pruning, network compression, architecture search, train from scratch
Hyperbolic Attention Networks 
Keywords:Hyperbolic Geometry, Attention Methods, Reasoning on Graphs, Relation Learning, Scale Free Graphs, Transformers, Power Law
TL;DR:We propose to incorporate inductive biases and operations coming from hyperbolic geometry to improve the attention mechanism of the neural networks.
Learning from Positive and Unlabeled Data with a Selection Bias 
Keywords:None
TL;DR:None
Adv-BNN: Improved Adversarial Defense through Robust Bayesian Neural Network 
Keywords:None
TL;DR:None
Optimal Completion Distillation for Sequence Learning 
Keywords:Sequence Learning, Edit Distance, Speech Recognition, Deep Reinforcement Learning
TL;DR:Optimal Completion Distillation (OCD) is a training procedure for optimizing sequence to sequence models based on edit distance which achieves state-of-the-art on end-to-end Speech Recognition tasks.
Caveats for information bottleneck in deterministic scenarios 
Keywords:Information bottleneck behaves in surprising ways whenever the output is a deterministic function of the input.
TL;DR:information bottleneck, supervised learning, deep learning, information theory
Deep Learning 3D Shapes Using Alt-az Anisotropic 2-Sphere Convolution 
Keywords:Spherical Convolution, Geometric deep learning, 3D shape analysis
TL;DR:A method for applying deep learning to 3D surfaces using their spherical descriptors and alt-az anisotropic convolution on 2-sphere.
Small nonlinearities in activation functions create bad local minima in neural networks 
Keywords:spurious local minima, loss surface, optimization landscape, neural network
TL;DR:We constructively prove that even the slightest nonlinear activation functions introduce spurious local minima, for general datasets and activation functions.
Information Theoretic lower bounds on negative log likelihood 
Keywords:latent variable modeling, rate-distortion theory, log likelihood bounds
TL;DR:Use rate-distortion theory to bound how much a latent variable model can be improved
Preferences Implicit in the State of the World 
Keywords:Preference learning, Inverse reinforcement learning, Inverse optimal stochastic control, Maximum entropy reinforcement learning, Apprenticeship learning
TL;DR:When a robot is deployed in an environment that humans have been acting in, the state of the environment is already optimized for what humans want, and we can use this to infer human preferences.
A Kernel Random Matrix-Based Approach for Sparse PCA 
Keywords:None
TL;DR:None
Bayesian Prediction of Future Street Scenes using Synthetic Likelihoods 
Keywords:bayesian inference, segmentation, anticipation, multi-modality
TL;DR:Dropout based Bayesian inference is extended to deal with multi-modality and is evaluated on scene anticipation tasks.
There Are Many Consistent Explanations of Unlabeled Data: Why You Should Average 
Keywords:semi-supervised learning, computer vision, classification, consistency regularization, flatness, weight averaging, stochastic weight averaging
TL;DR:Consistency-based models for semi-supervised learning do not converge to a single point but continue to explore a diverse set of plausible solutions on the perimeter of a flat region. Weight averaging helps improve generalization performance.
Large-Scale Answerer in Questioner's Mind for Visual Dialog Question Generation 
Keywords:None
TL;DR:None
Graph HyperNetworks for Neural Architecture Search 
Keywords:None
TL;DR:None
DELTA: DEEP LEARNING TRANSFER USING FEATURE MAP WITH ATTENTION FOR CONVOLUTIONAL NETWORKS 
Keywords:transfer learning, deep learning, regularization, attention, cnn
TL;DR:improving deep transfer learning with regularization using attention based feature maps
textTOvec: DEEP CONTEXTUALIZED NEURAL AUTOREGRESSIVE TOPIC MODELS OF LANGUAGE WITH DISTRIBUTED COMPOSITIONAL PRIOR 
Keywords:neural topic model, natural language processing, text representation, language modeling, information retrieval, deep learning
TL;DR:Unified neural model of topic and language modeling to introduce language structure in topic models for contextualized topic vectors
Amortized Bayesian Meta-Learning 
Keywords:variational inference, meta-learning, few-shot learning, uncertainty quantification
TL;DR:We propose a meta-learning method which efficiently amortizes hierarchical variational inference across training episodes.
Probabilistic Recursive Reasoning for Multi-Agent Reinforcement Learning 
Keywords:Multi-agent Reinforcement Learning, Recursive Reasoning
TL;DR:We proposed a novel probabilisitic recursive reasoning (PR2) framework for multi-agent deep reinforcement learning tasks.
Learning Neural PDE Solvers with Convergence Guarantees 
Keywords:Partial differential equation, deep learning
TL;DR:We learn a fast neural solver for PDEs that has convergence guarantees.
A new dog learns old tricks: RL finds classic optimization algorithms 
Keywords:reinforcement learning, algorithms, adwords, knapsack, secretary
TL;DR:By combining ideas from traditional algorithms design and reinforcement learning, we introduce a novel framework for learning algorithms that solve online combinatorial optimization problems.
Deep Graph Infomax 
Keywords:Unsupervised Learning, Graph Neural Networks, Graph Convolutions, Mutual Information, Infomax, Deep Learning
TL;DR:A new method for unsupervised representation learning on graphs, relying on maximizing mutual information between local and global representations in a graph. State-of-the-art results, competitive with supervised learning.
Theoretical Analysis of Auto Rate-Tuning by Batch Normalization 
Keywords:batch normalization, scale invariance, learning rate, stationary point
TL;DR:We give a theoretical analysis of the ability of batch normalization to automatically tune learning rates, in the context of finding stationary points for a deep learning objective.
Per-Tensor Fixed-Point Quantization of the Back-Propagation Algorithm 
Keywords:deep learning, reduced precision, fixed-point, quantization, back-propagation algorithm
TL;DR:We analyze and determine the precision requirements for training neural networks when all tensors, including back-propagated signals and weight accumulators, are quantized to fixed-point format.
FUNCTIONAL VARIATIONAL BAYESIAN NEURAL NETWORKS 
Keywords:functional variational inference, Bayesian neural networks, stochastic processes
TL;DR:We perform functional variational inference on the stochastic processes defined by Bayesian neural networks.
NADPEx: An on-policy temporally consistent exploration method for deep reinforcement learning 
Keywords:None
TL;DR:None
SPIGAN: Privileged Adversarial Learning from Simulation 
Keywords:domain adaptation, GAN, semantic segmentation, simulation, privileged information
TL;DR:An unsupervised sim-to-real domain adaptation method for semantic segmentation using privileged information from a simulator with GAN-based image translation.
Generating Multi-Agent Trajectories using Programmatic Weak Supervision 
Keywords:deep learning, generative models, imitation learning, hierarchical methods, data programming, weak supervision, spatiotemporal
TL;DR:We blend deep generative models with programmatic weak supervision to generate coordinated multi-agent trajectories of significantly higher quality than previous baselines.
Label super-resolution networks 
Keywords:weakly supervised segmentation, land cover mapping, medical imaging
TL;DR:Super-resolving coarse labels into pixel-level labels, applied to aerial imagery and medical scans.
ANYTIME MINIBATCH: EXPLOITING STRAGGLERS IN ONLINE DISTRIBUTED OPTIMIZATION 
Keywords:distributed optimization, gradient descent, minibatch, stragglers
TL;DR:Accelerate distributed optimization by exploiting stragglers.
Sample Efficient Adaptive Text-to-Speech 
Keywords:few shot, meta learning, text to speech, wavenet
TL;DR:Sample efficient algorithms to adapt a text-to-speech model to a new voice style with the state-of-the-art performance.
Practical lossless compression with latent variables using bits back coding 
Keywords:compression, variational auto-encoders, deep latent gaussian models, lossless compression, latent variables, approximate inference, variational inference
TL;DR:We do lossless compression of large image datasets using a VAE, beat existing compression algorithms.
Kernel RNN Learning (KeRNL) 
Keywords:RNNs, Biologically plausible learning rules, Algorithm, Neural Networks, Supervised Learning
TL;DR:A biologically plausible learning rule for training recurrent neural networks
Deep, Skinny Neural Networks are not Universal Approximators 
Keywords:This paper proves that skinny neural networks cannot approximate certain functions, no matter how deep they are.
TL;DR:neural network, universality, expressability
Large Scale Graph Learning From Smooth Signals 
Keywords:None
TL;DR:None
Overcoming Catastrophic Forgetting for Continual Learning via Model Adaptation 
Keywords:None
TL;DR:None
Analysis of Quantized Models 
Keywords:weight quantization, gradient quantization, distributed learning
TL;DR:In this paper, we studied efficient training of loss-aware weight-quantized networks with quantized gradient in a distributed environment, both theoretically and empirically.
Deep learning generalizes because the parameter-function map is biased towards simple functions 
Keywords:generalization, deep learning theory, PAC-Bayes, Gaussian processes, parameter-function map, simplicity bias
TL;DR:The parameter-function map of deep networks is hugely biased; this can explain why they generalize. We use PAC-Bayes and Gaussian processes to obtain nonvacuous bounds.
Learning when to Communicate at Scale in Multiagent Cooperative and Competitive Tasks 
Keywords:multiagent, communication, competitive, cooperative, continuous, emergent, reinforcement learning
TL;DR:We introduce IC3Net, a single network which can be used to train agents in cooperative, competitive and mixed scenarios. We also show that agents can learn when to communicate using our model.
Synthetic Datasets for Neural Program Synthesis 
Keywords:None
TL;DR:None
DPSNet: End-to-end Deep Plane Sweep Stereo 
Keywords:Deep Learning, Stereo, Depth, Geometry
TL;DR:A convolution neural network for multi-view stereo matching whose design is inspired by best practices of traditional geometry-based approaches
Conditional Network Embeddings 
Keywords:Network embedding, graph embedding, learning node representations, link prediction, multi-label classification of nodes
TL;DR:We introduce a network embedding method that accounts for prior information about the network, yielding superior empirical performance.
Defensive Quantization: When Efficiency Meets Robustness 
Keywords:defensive quantization, model quantization, adversarial attack, efficiency, robustness
TL;DR:We designed a novel quantization methodology to jointly optimize the efficiency and robustness of deep learning models.
GO Gradient for Expectation-Based Objectives 
Keywords:generalized reparameterization gradient, variance reduction, non-reparameterizable, discrete random variable, GO gradient, general and one-sample gradient, expectation-based objective, variable nabla, statistical back-propagation, hierarchical, graphical model
TL;DR:a Rep-like gradient for non-reparameterizable continuous/discrete distributions; further generalized to deep probabilistic models, yielding statistical back-propagation
h-detach: Modifying the LSTM Gradient Towards Better Optimization 
Keywords:LSTM, Optimization, Long term dependencies, Back-propagation through time
TL;DR:A simple algorithm to improve optimization and handling of long term dependencies in LSTM
An analytic theory of generalization dynamics and transfer learning in deep linear networks 
Keywords:Generalization, Theory, Transfer, Multi-task, Linear
TL;DR:We provide many insights into neural network generalization from the theoretically tractable linear case.
Differentiable Learning-to-Normalize via Switchable Normalization 
Keywords:None
TL;DR:None
SOM-VAE: Interpretable Discrete Representation Learning on Time Series 
Keywords:deep learning, self-organizing map, variational autoencoder, representation learning, time series, machine learning, interpretability
TL;DR:We present a method to learn interpretable representations on time series using ideas from variational autoencoders, self-organizing maps and probabilistic models.
Hierarchical Generative Modeling for Controllable Speech Synthesis 
Keywords:speech synthesis, representation learning, deep generative model, sequence-to-sequence model
TL;DR:Building a TTS model with Gaussian Mixture VAEs enables fine-grained control of speaking style, noise condition, and more.
Learning Factorized Multimodal Representations 
Keywords:multimodal learning, representation learning
TL;DR:We propose a model to learn factorized multimodal representations that are discriminative, generative, and interpretable.
Composing Complex Skills by Learning Transition Policies 
Keywords:reinforcement learning, hierarchical reinforcement learning, continuous control, modular framework
TL;DR:Transition policies enable agents to compose complex skills by smoothly connecting previously acquired primitive skills.
Human-level Protein Localization with Convolutional Neural Networks 
Keywords:None
TL;DR:None
Environment Probing Interaction Policies 
Keywords:None
TL;DR:None
Lagging Inference Networks and Posterior Collapse in Variational Autoencoders 
Keywords:variational autoencoders, posterior collapse, generative models
TL;DR:To address posterior collapse in VAEs, we propose a novel yet simple training procedure that aggressively optimizes inference network with more updates. This new training procedure mitigates posterior collapse and leads to a better VAE model.
A2BCD: Asynchronous Acceleration with Optimal Complexity 
Keywords:asynchronous, optimization, parallel, accelerated, complexity
TL;DR:We prove the first-ever convergence proof of an asynchronous accelerated algorithm that attains a speedup.
Learning to Infer and Execute 3D Shape Programs 
Keywords:Program Synthesis, 3D Shape Modeling, Self-supervised Learning
TL;DR:We propose 3D shape programs, a structured, compositional shape representation. Our model learns to infer and execute shape programs to explain 3D shapes.
Deep Decoder: Concise Image Representations from Untrained Non-convolutional Networks 
Keywords:natural image model, image prior, under-determined neural networks, untrained network, non-convolutional network, denoising, inverse problem
TL;DR:We introduce an underparameterized, nonconvolutional, and simple deep neural network that can, without training, effectively represent natural images and solve image processing tasks like compression and denoising competitively.
SNAS: stochastic neural architecture search 
Keywords:None
TL;DR:None
Revealing interpretable object representations from human behavior 
Keywords:category representation, sparse coding, representation learning, interpretable representations
TL;DR:Human behavioral judgments are used to obtain sparse and interpretable representations of objects that generalize to other tasks
AntisymmetricRNN: A Dynamical System View on Recurrent Neural Networks 
Keywords:None
TL;DR:None
Global-to-local Memory Pointer Networks for Task-Oriented Dialogue 
Keywords:pointer networks, memory networks, task-oriented dialogue systems, natural language processing
TL;DR:GLMP: Global memory encoder (context RNN, global pointer) and local memory decoder (sketch RNN, local pointer) that share external knowledge (MemNN) are proposed to strengthen response generation in task-oriented dialogue.
InstaGAN: Instance-aware Image-to-Image Translation 
Keywords:Image-to-Image Translation, Generative Adversarial Networks
TL;DR:We propose a novel method to incorporate the set of instance attributes for image-to-image translation.
Deep Layers as Stochastic Solvers 
Keywords:deep networks, optimization
TL;DR:A framework that links deep network layers to stochastic optimization algorithms; can be used to improve model accuracy and inform network design.
Learning Multi-Level Hierarchies with Hindsight 
Keywords:Hierarchical Reinforcement Learning, Reinforcement Learning, Deep Reinforcement Learning
TL;DR:We introduce the first Hierarchical RL approach to successfully learn 3-level hierarchies in parallel in tasks with continuous state and action spaces.
  • 0
    点赞
  • 1
    收藏
    觉得还不错? 一键收藏
  • 0
    评论

“相关推荐”对你有帮助么?

  • 非常没帮助
  • 没帮助
  • 一般
  • 有帮助
  • 非常有帮助
提交
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值