Search Results for author: Luca D'Amico-Wong

Found 4 papers, 1 papers with code

Easy as ABCs: Unifying Boltzmann Q-Learning and Counterfactual Regret Minimization

no code implementations19 Feb 2024 Luca D'Amico-Wong, Hugh Zhang, Marc Lanctot, David C. Parkes

We propose ABCs (Adaptive Branching through Child stationarity), a best-of-both-worlds algorithm combining Boltzmann Q-learning (BQL), a classic reinforcement learning algorithm for single-agent domains, and counterfactual regret minimization (CFR), a central algorithm for learning in multi-agent domains.

counterfactual OpenAI Gym +1

American Stories: A Large-Scale Structured Text Dataset of Historical U.S. Newspapers

no code implementations NeurIPS 2023 Melissa Dell, Jacob Carlson, Tom Bryan, Emily Silcock, Abhishek Arora, Zejiang Shen, Luca D'Amico-Wong, Quan Le, Pablo Querubin, Leander Heldring

The resulting American Stories dataset provides high quality data that could be used for pre-training a large language model to achieve better understanding of historical English and historical world knowledge.

Language Modelling Large Language Model +3

Understanding and Mitigating Extrapolation Failures in Physics-Informed Neural Networks

1 code implementation15 Jun 2023 Lukas Fesser, Luca D'Amico-Wong, Richard Qiu

Physics-informed Neural Networks (PINNs) have recently gained popularity due to their effective approximation of partial differential equations (PDEs) using deep neural networks (DNNs).

Transfer Learning

Noise-Robust De-Duplication at Scale

no code implementations9 Oct 2022 Emily Silcock, Luca D'Amico-Wong, Jinglin Yang, Melissa Dell

Identifying near duplicates within large, noisy text corpora has a myriad of applications that range from de-duplicating training datasets, reducing privacy risk, and evaluating test set leakage, to identifying reproduced news articles and literature within large corpora.

Cannot find the paper you are looking for? You can Submit a new open access paper.