Search Results for author: Minjie Zhu

Found 6 papers, 2 papers with code

Language-Conditioned Robotic Manipulation with Fast and Slow Thinking

no code implementations8 Jan 2024 Minjie Zhu, Yichen Zhu, Jinming Li, Junjie Wen, Zhiyuan Xu, Zhengping Che, Chaomin Shen, Yaxin Peng, Dong Liu, Feifei Feng, Jian Tang

The language-conditioned robotic manipulation aims to transfer natural language instructions into executable actions, from simple pick-and-place to tasks requiring intent recognition and visual reasoning.

Decision Making Intent Recognition +2

LLaVA-Phi: Efficient Multi-Modal Assistant with Small Language Model

1 code implementation4 Jan 2024 Yichen Zhu, Minjie Zhu, Ning Liu, Zhicai Ou, Xiaofeng Mou, Jian Tang

In this paper, we introduce LLaVA-$\phi$ (LLaVA-Phi), an efficient multi-modal assistant that harnesses the power of the recently advanced small language model, Phi-2, to facilitate multi-modal dialogues.

Language Modelling Visual Question Answering

SpeechAct: Towards Generating Whole-body Motion from Speech

no code implementations29 Nov 2023 Jinsong Zhang, Minjie Zhu, Yuxiang Zhang, Yebin Liu, Kun Li

Then, we regress the motion representation from the audio signal by a translation model employing our contrastive motion learning method.

Cannot find the paper you are looking for? You can Submit a new open access paper.