Search Results for author: Jerry Zhu

Found 14 papers, 0 papers with code

The Delusional Hedge Algorithm as a Model of Human Learning from Diverse Opinions

no code implementations21 Feb 2024 Yun-Shiuan Chuang, Jerry Zhu, Timothy T. Rogers

Whereas cognitive models of learning often assume direct experience with both the features of an event and with a true label or outcome, much of everyday learning arises from hearing the opinions of others, without direct access to either the experience or the ground truth outcome.

Corruption-Robust Offline Reinforcement Learning

no code implementations11 Jun 2021 Xuezhou Zhang, Yiding Chen, Jerry Zhu, Wen Sun

Surprisingly, in this case, the knowledge of $\epsilon$ is necessary, as we show that being adaptive to unknown $\epsilon$ is impossible. This again contrasts with recent results on corruption-robust online RL and implies that robust offline RL is a strictly harder problem.

Adversarial Robustness Offline RL +2

Policy Gradient Bayesian Robust Optimization for Imitation Learning

no code implementations11 Jun 2021 Zaynah Javed, Daniel S. Brown, Satvik Sharma, Jerry Zhu, Ashwin Balakrishna, Marek Petrik, Anca D. Dragan, Ken Goldberg

Results suggest that PG-BROIL can produce a family of behaviors ranging from risk-neutral to risk-averse and outperforms state-of-the-art imitation learning algorithms when learning from ambiguous demonstrations by hedging against uncertainty, rather than seeking to uniquely identify the demonstrator's reward function.

Imitation Learning

Active Learning with Oracle Epiphany

no code implementations NeurIPS 2016 Tzu-Kuo Huang, Lihong Li, Ara Vartanian, Saleema Amershi, Jerry Zhu

We present a theoretical analysis of active learning with more realistic interactions with human oracles.

Active Learning

Human Memory Search as Initial-Visit Emitting Random Walk

no code implementations NeurIPS 2015 Kwang-Sung Jun, Jerry Zhu, Timothy T. Rogers, Zhuoran Yang, Ming Yuan

In this paper, we propose the first efficient maximum likelihood estimate (MLE) for INVITE by decomposing the censored output into a series of absorbing random walks.

Optimal Teaching for Limited-Capacity Human Learners

no code implementations NeurIPS 2014 Kaustubh R. Patil, Jerry Zhu, Łukasz Kopeć, Bradley C. Love

We apply a machine teaching procedure to a cognitive model that is either limited capacity (as humans are) or unlimited capacity (as most machine learning systems are).

Retrieval

Machine Teaching for Bayesian Learners in the Exponential Family

no code implementations NeurIPS 2013 Jerry Zhu

What if there is a teacher who knows the learning goal and wants to design good training data for a machine learner?

Learning Higher-Order Graph Structure with Features by Structure Penalty

no code implementations NeurIPS 2011 Shilin Ding, Grace Wahba, Jerry Zhu

In discrete undirected graphical models, the conditional independence of node labels Y is specified by the graph structure.

How Do Humans Teach: On Curriculum Learning and Teaching Dimension

no code implementations NeurIPS 2011 Faisal Khan, Bilge Mutlu, Jerry Zhu

We study the empirical strategies that humans follow as they teach a target concept with a simple 1D threshold to a robot.

Humans Learn Using Manifolds, Reluctantly

no code implementations NeurIPS 2010 Tim Rogers, Chuck Kalish, Joseph Harrison, Jerry Zhu, Bryan R. Gibson

When the distribution of unlabeled data in feature space lies along a manifold, the information it provides may be used by a learner to assist classification in a semi-supervised setting.

BIG-bench Machine Learning General Classification

Human Rademacher Complexity

no code implementations NeurIPS 2009 Jerry Zhu, Bryan R. Gibson, Timothy T. Rogers

We propose to use Rademacher complexity, originally developed in computational learning theory, as a measure of human learning capacity.

Generalization Bounds Learning Theory

Unlabeled data: Now it helps, now it doesn't

no code implementations NeurIPS 2008 Aarti Singh, Robert Nowak, Jerry Zhu

We show that there are large classes of problems for which SSL can significantly outperform supervised learning, in finite sample regimes and sometimes also in terms of error convergence rates.

Cannot find the paper you are looking for? You can Submit a new open access paper.