1 code implementation • 15 Oct 2023 • Jake Grigsby, Linxi Fan, Yuke Zhu
We introduce AMAGO, an in-context Reinforcement Learning (RL) agent that uses sequence models to tackle the challenges of generalization, long-term memory, and meta-learning.
1 code implementation • 2 May 2023 • Zhe Wang, Jake Grigsby, Yanjun Qi
In this work, we develop a novel DG training strategy, we call PGrad, to learn a robust gradient direction, improving models' generalization ability on unseen domains.
no code implementations • 1 Dec 2022 • Vanamala Venkataswamy, Jake Grigsby, Andrew Grimshaw, Yanjun Qi
We utilize Offline RL as a launchpad to learn effective scheduling policies from prior experience collected using Oracle or heuristic policies.
no code implementations • 10 Nov 2022 • Vanamala Venkataswamy, Jake Grigsby, Andrew Grimshaw, Yanjun Qi
Finally, we demonstrate that the DRL scheduler can learn from and improve upon existing heuristic policies using Offline Learning.
3 code implementations • 10 Oct 2021 • Jake Grigsby, Yanjun Qi
A thorough investigation on a custom benchmark helps identify several key challenges involved in learning from high-noise datasets.
no code implementations • 27 Sep 2021 • Zhe Wang, Jake Grigsby, Arshdeep Sekhon, Yanjun Qi
This paper proposes a novel method, ST-MAML, that empowers model-agnostic meta-learning (MAML) to learn from multiple task distributions.
2 code implementations • 24 Sep 2021 • Jake Grigsby, Zhe Wang, Nam Nguyen, Yanjun Qi
Multivariate time series forecasting focuses on predicting future values based on historical context.
1 code implementation • 16 Jun 2021 • Jake Grigsby, Jin Yong Yoo, Yanjun Qi
Model-free off-policy actor-critic methods are an efficient solution to complex continuous control tasks.
2 code implementations • 13 Oct 2020 • Jake Grigsby, Yanjun Qi
Self-supervised learning and data augmentation have significantly reduced the performance gap between state and image-based reinforcement learning agents in continuous control tasks.
2 code implementations • EMNLP 2020 • John X. Morris, Eli Lifland, Jin Yong Yoo, Jake Grigsby, Di Jin, Yanjun Qi
TextAttack also includes data augmentation and adversarial training modules for using components of adversarial attacks to improve model accuracy and robustness.