Search Results for author: Xishan Zhang

Found 15 papers, 4 papers with code

Pushing the Limits of Machine Design: Automated CPU Design with AI

1 code implementation21 Jun 2023 Shuyao Cheng, Pengwei Jin, Qi Guo, Zidong Du, Rui Zhang, Yunhao Tian, Xing Hu, Yongwei Zhao, Yifan Hao, Xiangtao Guan, Husheng Han, Zhengyue Zhao, Ximing Liu, Ling Li, Xishan Zhang, Yuejie Chu, Weilong Mao, Tianshi Chen, Yunji Chen

By efficiently exploring a search space of unprecedented size 10^{10^{540}}, which is the largest one of all machine-designed objects to our best knowledge, and thus pushing the limits of machine design, our approach generates an industrial-scale RISC-V CPU within only 5 hours.

Online Prototype Alignment for Few-shot Policy Transfer

1 code implementation12 Jun 2023 Qi Yi, Rui Zhang, Shaohui Peng, Jiaming Guo, Yunkai Gao, Kaizhao Yuan, Ruizhi Chen, Siming Lan, Xing Hu, Zidong Du, Xishan Zhang, Qi Guo, Yunji Chen

Domain adaptation in reinforcement learning (RL) mainly deals with the changes of observation when transferring the policy to a new environment.

Domain Adaptation Reinforcement Learning (RL)

Ultra-low Precision Multiplication-free Training for Deep Neural Networks

no code implementations28 Feb 2023 Chang Liu, Rui Zhang, Xishan Zhang, Yifan Hao, Zidong Du, Xing Hu, Ling Li, Qi Guo

The energy-efficient works try to decrease the precision of multiplication or replace the multiplication with energy-efficient operations such as addition or bitwise shift, to reduce the energy consumption of FP32 multiplications.

Quantization

Causality-driven Hierarchical Structure Discovery for Reinforcement Learning

no code implementations13 Oct 2022 Shaohui Peng, Xing Hu, Rui Zhang, Ke Tang, Jiaming Guo, Qi Yi, Ruizhi Chen, Xishan Zhang, Zidong Du, Ling Li, Qi Guo, Yunji Chen

To address this issue, we propose CDHRL, a causality-driven hierarchical reinforcement learning framework, leveraging a causality-driven discovery instead of a randomness-driven exploration to effectively build high-quality hierarchical structures in complicated environments.

Hierarchical Reinforcement Learning reinforcement-learning +1

Object-Category Aware Reinforcement Learning

no code implementations13 Oct 2022 Qi Yi, Rui Zhang, Shaohui Peng, Jiaming Guo, Xing Hu, Zidong Du, Xishan Zhang, Qi Guo, Yunji Chen

Object-oriented reinforcement learning (OORL) is a promising way to improve the sample efficiency and generalization ability over standard RL.

Feature Engineering Object +3

Neural Program Synthesis with Query

no code implementations ICLR 2022 Di Huang, Rui Zhang, Xing Hu, Xishan Zhang, Pengwei Jin, Nan Li, Zidong Du, Qi Guo, Yunji Chen

In this work, we propose a query-based framework that trains a query neural network to generate informative input-output examples automatically and interactively from a large query space.

Program Synthesis

Hindsight Value Function for Variance Reduction in Stochastic Dynamic Environment

1 code implementation26 Jul 2021 Jiaming Guo, Rui Zhang, Xishan Zhang, Shaohui Peng, Qi Yi, Zidong Du, Xing Hu, Qi Guo, Yunji Chen

In this paper, we propose to replace the state value function with a novel hindsight value function, which leverages the information from the future to reduce the variance of the gradient estimate for stochastic dynamic environments.

Policy Gradient Methods

Domain-Specific Suppression for Adaptive Object Detection

no code implementations CVPR 2021 Yu Wang, Rui Zhang, Shuo Zhang, Miao Li, Yangyang Xia, Xishan Zhang, Shaoli Liu

The directions of weights, and the gradients, can be divided into domain-specific and domain-invariant parts, and the goal of domain adaptation is to concentrate on the domain-invariant direction while eliminating the disturbance from domain-specific one.

Domain Adaptation Object +2

Fixed-Point Back-Propagation Training

no code implementations CVPR 2020 Xishan Zhang, Shaoli Liu, Rui Zhang, Chang Liu, Di Huang, Shiyi Zhou, Jiaming Guo, Qi Guo, Zidong Du, Tian Zhi, Yunji Chen

Recent emerged quantization technique (i. e., using low bit-width fixed-point data instead of high bit-width floating-point data) has been applied to inference of deep neural networks for fast and efficient execution.

Image Classification Machine Translation +4

DWM: A Decomposable Winograd Method for Convolution Acceleration

no code implementations3 Feb 2020 Di Huang, Xishan Zhang, Rui Zhang, Tian Zhi, Deyuan He, Jiaming Guo, Chang Liu, Qi Guo, Zidong Du, Shaoli Liu, Tianshi Chen, Yunji Chen

In this paper, we propose a novel Decomposable Winograd Method (DWM), which breaks through the limitation of original Winograd's minimal filtering algorithm to a wide and general convolutions.

Task-Driven Dynamic Fusion: Reducing Ambiguity in Video Description

no code implementations CVPR 2017 Xishan Zhang, Ke Gao, Yongdong Zhang, Dongming Zhang, Jintao Li, Qi Tian

This paper contributes to: 1)The first in-depth study of the weakness inherent in data-driven static fusion methods for video captioning.

Video Captioning Video Description

Cannot find the paper you are looking for? You can Submit a new open access paper.