ICLR 2018

Deep Bayesian Bandits Showdown: An Empirical Comparison of Bayesian Deep Networks for Thompson Sampling

ICLR 2018 tensorflow/models

At the same time, advances in approximate Bayesian methods have made posterior approximation for flexible neural network models practical.

DECISION MAKING MULTI-ARMED BANDITS

Parameter Space Noise for Exploration

ICLR 2018 tensorflow/models

Combining parameter noise with traditional RL methods allows to combine the best of both worlds.

CONTINUOUS CONTROL

Ensemble Adversarial Training: Attacks and Defenses

ICLR 2018 tensorflow/models

We show that this form of adversarial training converges to a degenerate global minimum, wherein small curvature artifacts near the data points obfuscate a linear approximation of the loss.

ADVERSARIAL TRAINING

Scalable Private Learning with PATE

ICLR 2018 tensorflow/models

Models and examples built with TensorFlow

Regularizing and Optimizing LSTM Language Models

ICLR 2018 google-research/google-research

Recurrent neural networks (RNNs), such as long short-term memory networks (LSTMs), serve as a fundamental building block for many sequence learning tasks, including machine translation, language modeling, and question answering.

LANGUAGE MODELLING

Discrete Autoencoders for Sequence Models

ICLR 2018 tensorflow/tensor2tensor

We propose to improve the representation in sequence models by augmenting current approaches with an autoencoder that is forced to compress the sequence through an intermediate discrete latent space.

LANGUAGE MODELLING MACHINE TRANSLATION

Depthwise Separable Convolutions for Neural Machine Translation

ICLR 2018 tensorflow/tensor2tensor

In this work, we study how depthwise separable convolutions can be applied to neural machine translation.

MACHINE TRANSLATION

Generating Wikipedia by Summarizing Long Sequences

ICLR 2018 tensorflow/tensor2tensor

We show that generating English Wikipedia articles can be approached as a multi- document summarization of source documents.

DOCUMENT SUMMARIZATION MULTI-DOCUMENT SUMMARIZATION

On Convergence and Stability of GANs

ICLR 2018 eriklindernoren/PyTorch-GAN

We propose studying GAN training dynamics as regret minimization, which is in contrast to the popular view that there is consistent minimization of a divergence between real and generated distributions.