Search Results for author: Guoxi Zhang

Found 6 papers, 2 papers with code

INSIGHT: End-to-End Neuro-Symbolic Visual Reinforcement Learning with Language Explanations

no code implementations19 Mar 2024 Lirui Luo, Guoxi Zhang, Hongming Xu, Yaodong Yang, Cong Fang, Qing Li

In this paper, we present a framework that is capable of learning structured states and symbolic policies simultaneously, whose key idea is to overcome the efficiency bottleneck by distilling vision foundation models into a scalable perception module.

Decision Making

Online Policy Learning from Offline Preferences

no code implementations15 Mar 2024 Guoxi Zhang, Han Bao, Hisashi Kashima

To address this problem, the present study introduces a framework that consolidates offline preferences and \emph{virtual preferences} for PbRL, which are comparisons between the agent's behaviors and the offline data.

Continuous Control

Estimating Treatment Effects Under Heterogeneous Interference

1 code implementation25 Sep 2023 Xiaofeng Lin, Guoxi Zhang, Xiaotian Lu, Han Bao, Koh Takeuchi, Hisashi Kashima

One popular application of this estimation lies in the prediction of the impact of a treatment (e. g., a promotion) on an outcome (e. g., sales) of a particular unit (e. g., an item), known as the individual treatment effect (ITE).

Decision Making

Behavior Estimation from Multi-Source Data for Offline Reinforcement Learning

1 code implementation29 Nov 2022 Guoxi Zhang, Hisashi Kashima

To overcome this drawback, the present study proposes a latent variable model to infer a set of policies from data, which allows an agent to use as behavior policy the policy that best describes a particular trajectory.

Offline RL reinforcement-learning +1

Batch Reinforcement Learning from Crowds

no code implementations8 Nov 2021 Guoxi Zhang, Hisashi Kashima

This paper addresses the lack of reward in a batch reinforcement learning setting by learning a reward function from preferences.

reinforcement-learning Reinforcement Learning (RL)

Cannot find the paper you are looking for? You can Submit a new open access paper.