Search Results for author: Martin Müller

Found 18 papers, 8 papers with code

Replay Memory as An Empirical MDP: Combining Conservative Estimation with Experience Replay

1 code implementation ICLR 2023 Hongming Zhang, Chenjun Xiao, Han Wang, Jun Jin, Bo Xu, Martin Müller

In this work, we further exploit the information in the replay memory by treating it as an empirical \emph{Replay Memory MDP (RM-MDP)}.

Supervised and Reinforcement Learning from Observations in Reconnaissance Blind Chess

no code implementations3 Aug 2022 Timo Bertram, Johannes Fürnkranz, Martin Müller

In this work, we adapt a training approach inspired by the original AlphaGo system to play the imperfect information game of Reconnaissance Blind Chess.

reinforcement-learning Reinforcement Learning (RL)

Quantity vs Quality: Investigating the Trade-Off between Sample Size and Label Reliability

no code implementations20 Apr 2022 Timo Bertram, Johannes Fürnkranz, Martin Müller

In this paper, we study learning in probabilistic domains where the learner may receive incorrect labels but can improve the reliability of labels by repeatedly sampling them.

Cedille: A large autoregressive French language model

1 code implementation7 Feb 2022 Martin Müller, Florian Laurent

Scaling up the size and training of autoregressive language models has enabled novel ways of solving Natural Language Processing tasks using zero-shot and few-shot learning.

Few-Shot Learning Language Modelling +1

Efficient Sampling-Based Bayesian Active Learning for synaptic characterization

no code implementations19 Jan 2022 Camille Gontier, Simone Carlo Surace, Igor Delvendahl, Martin Müller, Jean-Pascal Pfister

Bayesian Active Learning (BAL) is an efficient framework for learning the parameters of a model, in which input stimuli are selected to maximize the mutual information between the observations and the unknown parameters.

Active Learning

A Simple Unified Framework for Anomaly Detection in Deep Reinforcement Learning

no code implementations21 Sep 2021 Hongming Zhang, Ke Sun, Bo Xu, Linglong Kong, Martin Müller

In this paper, we propose a simple yet effective anomaly detection framework for deep RL algorithms that simultaneously considers random, adversarial and out-of-distribution~(OOD) state outliers.

Anomaly Detection Atari Games +3

Addressing materials' microstructure diversity using transfer learning

no code implementations29 Jul 2021 Aurèle Goetz, Ali Riza Durmaz, Martin Müller, Akhil Thomas, Dominik Britz, Pierre Kerfriden, Chris Eberl

We show that a state-of-the-art UDA approach surpasses the na\"ive application of source domain trained models on the target domain (generalization baseline) to a large extent.

Data Augmentation Domain Generalization +2

A Comparison of Contextual and Non-Contextual Preference Ranking for Set Addition Problems

no code implementations9 Jul 2021 Timo Bertram, Johannes Fürnkranz, Martin Müller

We discuss and compare two different Siamese network architectures for this task: a twin network that compares the two sets resulting after the addition, and a triplet network that models the contribution of each candidate to the existing set.

Predicting Human Card Selection in Magic: The Gathering with Contextual Preference Ranking

1 code implementation25 May 2021 Timo Bertram, Johannes Fürnkranz, Martin Müller

Drafting, i. e., the selection of a subset of items from a larger candidate set, is a key element of many games and related problems.

Card Games

A Deep Dive into Conflict Generating Decisions

1 code implementation10 May 2021 Md Solimul Chowdhury, Martin Müller, Jia You

We also show an important connection between consecutive clauses learned within the same mc decision, where one learned clause triggers the learning of the next one forming a chain of clauses.

Addressing machine learning concept drift reveals declining vaccine sentiment during the COVID-19 pandemic

1 code implementation3 Dec 2020 Martin Müller, Marcel Salathé

We show that while vaccine sentiment has declined considerably during the COVID-19 pandemic in 2020, algorithms trained on pre-pandemic data would have largely missed this decline due to concept drift.

BIG-bench Machine Learning

COVID-Twitter-BERT: A Natural Language Processing Model to Analyse COVID-19 Content on Twitter

1 code implementation15 May 2020 Martin Müller, Marcel Salathé, Per E Kummervold

In this work, we release COVID-Twitter-BERT (CT-BERT), a transformer-based model, pretrained on a large corpus of Twitter messages on the topic of COVID-19.

Classification General Classification +1

Learning to Combat Compounding-Error in Model-Based Reinforcement Learning

no code implementations24 Dec 2019 Chenjun Xiao, Yifan Wu, Chen Ma, Dale Schuurmans, Martin Müller

Despite its potential to improve sample complexity versus model-free approaches, model-based reinforcement learning can fail catastrophically if the model is inaccurate.

Model-based Reinforcement Learning reinforcement-learning +1

Maximum Entropy Monte-Carlo Planning

no code implementations NeurIPS 2019 Chenjun Xiao, Ruitong Huang, Jincheng Mei, Dale Schuurmans, Martin Müller

We then extend this approach to general sequential decision making by developing a general MCTS algorithm, Maximum Entropy for Tree Search (MENTS).

Atari Games Decision Making

Characterization of Glue Variables in CDCL SAT Solving

no code implementations25 Apr 2019 Md Solimul Chowdhury, Martin Müller, Jia-Huai You

We first show experimentally, by running the state-of-the-art CDCL SAT solver MapleLCMDist on benchmarks from SAT Competition-2017 and 2018, that branching decisions with glue variables are categorically more inference and conflict efficient than nonglue variables.

Structured Best Arm Identification with Fixed Confidence

no code implementations16 Jun 2017 Ruitong Huang, Mohammad M. Ajallooeian, Csaba Szepesvári, Martin Müller

We study the problem of identifying the best action among a set of possible options when the value of each action is given by a mapping from a number of noisy micro-observables in the so-called fixed confidence setting.

Cannot find the paper you are looking for? You can Submit a new open access paper.