Search Results for author: Chaoqi Wang

Found 9 papers, 5 papers with code

Beyond Reverse KL: Generalizing Direct Preference Optimization with Diverse Divergence Constraints

no code implementations28 Sep 2023 Chaoqi Wang, Yibo Jiang, Chenghao Yang, Han Liu, Yuxin Chen

The increasing capabilities of large language models (LLMs) raise opportunities for artificial general intelligence but concurrently amplify safety concerns, such as potential misuse of AI systems, necessitating effective AI alignment.

Active Policy Improvement from Multiple Black-box Oracles

1 code implementation17 Jun 2023 Xuefeng Liu, Takuma Yoneda, Chaoqi Wang, Matthew R. Walter, Yuxin Chen

We introduce MAPS and MAPS-SE, a class of policy improvement algorithms that perform imitation learning from multiple suboptimal oracles.

Imitation Learning Reinforcement Learning (RL)

Teaching an Active Learner with Contrastive Examples

no code implementations NeurIPS 2021 Chaoqi Wang, Adish Singla, Yuxin Chen

Our focus is to design a teaching algorithm that can provide an informative sequence of contrastive examples to the learner to speed up the learning process.

Active Learning

Beyond Marginal Uncertainty: How Accurately can Bayesian Regression Models Estimate Posterior Predictive Correlations?

1 code implementation6 Nov 2020 Chaoqi Wang, Shengyang Sun, Roger Grosse

While uncertainty estimation is a well-studied topic in deep learning, most such work focuses on marginal uncertainty estimates, i. e. the predictive mean and variance at individual input locations.

Active Learning Benchmarking +1

Picking Winning Tickets Before Training by Preserving Gradient Flow

3 code implementations ICLR 2020 Chaoqi Wang, Guodong Zhang, Roger Grosse

Overparameterization has been shown to benefit both the optimization and generalization of neural networks, but large networks are resource hungry at both training and test time.

Network Pruning

EigenDamage: Structured Pruning in the Kronecker-Factored Eigenbasis

1 code implementation15 May 2019 Chaoqi Wang, Roger Grosse, Sanja Fidler, Guodong Zhang

Reducing the test time resource requirements of a neural network while preserving test accuracy is crucial for running inference on resource-constrained devices.

Network Pruning

Three Mechanisms of Weight Decay Regularization

no code implementations ICLR 2019 Guodong Zhang, Chaoqi Wang, Bowen Xu, Roger Grosse

Weight decay is one of the standard tricks in the neural network toolbox, but the reasons for its regularization effect are poorly understood, and recent results have cast doubt on the traditional interpretation in terms of $L_2$ regularization.

A Revisit on Deep Hashings for Large-scale Content Based Image Retrieval

no code implementations16 Nov 2017 Deng Cai, Xiuye Gu, Chaoqi Wang

However, there are serious flaws in the evaluations of existing deep hashing papers: (1) The datasets they used are too small and simple to simulate the real CBIR situation.

Content-Based Image Retrieval Deep Hashing

Cannot find the paper you are looking for? You can Submit a new open access paper.