no code implementations • 24 Apr 2024 • Jialong Wu, Chaoyi Deng, Jianmin Wang, Mingsheng Long
Effective code optimization in compilers plays a central role in computer and software engineering.
1 code implementation • 2 Mar 2024 • Jialong Wu, Linhai Zhang, Deyu Zhou, Guoqiang Xu
However, most of the present debiasing methods focus on single-variable causal inference, which is not suitable for ABSA with two input variables (the target aspect and the review).
Aspect-Based Sentiment Analysis Aspect-Based Sentiment Analysis (ABSA) +3
1 code implementation • 2 Mar 2024 • Linhai Zhang, Jialong Wu, Deyu Zhou, Guoqiang Xu
For poor model calibration, we incorporate the regularization method during LoRA training to keep the model from being over-confident, and the Monte-Carlo dropout mechanism is employed to enhance the uncertainty estimation.
no code implementations • 30 Jan 2024 • Tiannan Wang, Jiamin Chen, Qingrui Jia, Shuai Wang, Ruoyu Fang, Huilin Wang, Zhaowei Gao, Chunzhao Xie, Chuou Xu, Jihong Dai, Yibin Liu, Jialong Wu, Shengwei Ding, Long Li, Zhiwei Huang, Xinle Deng, Teng Yu, Gangan Ma, Han Xiao, Zixin Chen, Danjun Xiang, Yunxia Wang, Yuanyuan Zhu, Yi Xiao, Jing Wang, Yiru Wang, Siran Ding, Jiayang Huang, Jiayi Xu, Yilihamu Tayier, Zhenyu Hu, Yuan Gao, Chengfeng Zheng, Yueshu Ye, Yihang Li, Lei Wan, Xinyue Jiang, Yujie Wang, Siyu Cheng, Zhule Song, Xiangru Tang, Xiaohua Xu, Ningyu Zhang, Huajun Chen, Yuchen Eleanor Jiang, Wangchunshu Zhou
Weaver is pre-trained on a carefully selected corpus that focuses on improving the writing capabilities of large language models.
no code implementations • 30 Oct 2023 • Xuefeng Bai, Jialong Wu, Yulong Chen, Zhongqing Wang, Yue Zhang
Constituency parsing is a fundamental yet unsolved natural language processing task.
no code implementations • 30 Sep 2023 • Haoyu Ma, Jialong Wu, Ningya Feng, Chenjun Xiao, Dong Li, Jianye Hao, Jianmin Wang, Mingsheng Long
Model-based reinforcement learning (MBRL) holds the promise of sample-efficient learning by utilizing a world model, which models how the environment works and typically encompasses components for two tasks: observation modeling and reward modeling.
Ranked #4 on Atari Games 100k on Atari 100k
1 code implementation • 14 Sep 2023 • Wangchunshu Zhou, Yuchen Eleanor Jiang, Long Li, Jialong Wu, Tiannan Wang, Shi Qiu, Jintian Zhang, Jing Chen, Ruipu Wu, Shuai Wang, Shiding Zhu, Jiyu Chen, Wentao Zhang, Xiangru Tang, Ningyu Zhang, Huajun Chen, Peng Cui, Mrinmaya Sachan
Recent advances on large language models (LLMs) enable researchers and developers to build autonomous language agents that can automatically solve various tasks and interact with environments, humans, and other agents using natural language interfaces.
1 code implementation • NeurIPS 2023 • Jialong Wu, Haoyu Ma, Chaoyi Deng, Mingsheng Long
To tackle this issue, we introduce Contextualized World Models (ContextWM) that explicitly separate context and dynamics modeling to overcome the complexity and diversity of in-the-wild videos and facilitate knowledge transfer between distinct scenes.
1 code implementation • 2 Feb 2023 • Yang Shu, Xingzhuo Guo, Jialong Wu, Ximei Wang, Jianmin Wang, Mingsheng Long
This paper aims at generalizing CLIP to out-of-distribution test data on downstream tasks.
1 code implementation • 13 Nov 2022 • Yiwen Qiu, Jialong Wu, Zhangjie Cao, Mingsheng Long
Existing imitation learning works mainly assume that the demonstrator who collects demonstrations shares the same dynamics as the imitator.
no code implementations • 11 Jul 2022 • Walter Zimmer, Jialong Wu, Xingcheng Zhou, Alois C. Knoll
This work aims to address the challenges in autonomous driving by focusing on the 3D perception of the environment using roadside LiDARs.
1 code implementation • 13 Feb 2022 • Haixu Wu, Jialong Wu, Jiehui Xu, Jianmin Wang, Mingsheng Long
By respectively conserving the incoming flow of sinks for source competition and the outgoing flow of sources for sink allocation, Flow-Attention inherently generates informative attentions without using specific inductive biases.
Ranked #4 on D4RL on D4RL
3 code implementations • 13 Feb 2022 • Jialong Wu, Haixu Wu, Zihan Qiu, Jianmin Wang, Mingsheng Long
Policy constraint methods to offline reinforcement learning (RL) typically utilize parameterization or regularization that constrains the policy to perform actions within the support set of the behavior policy.