1 code implementation • 17 Feb 2024 • Shanshan Zhong, Zhongzhan Huang, Daifeng Li, Wushao Wen, Jinghui Qin, Liang Lin
This strategy can implicitly enhance the model's robustness during the optimization process, mitigating instability risks arising from multimodal information inputs.
1 code implementation • 5 Dec 2023 • Shanshan Zhong, Zhongzhan Huang, ShangHua Gao, Wushao Wen, Liang Lin, Marinka Zitnik, Pan Zhou
To this end, we study LLMs on the popular Oogiri game which needs participants to have good creativity and strong associative thinking for responding unexpectedly and humorously to the given image, text, or both, and thus is suitable for LoT study.
no code implementations • ICCV 2023 • Zhongzhan Huang, Mingfu Liang, Jinghui Qin, Shanshan Zhong, Liang Lin
The self-attention mechanism (SAM) is widely used in various fields of artificial intelligence and has successfully boosted the performance of different models.
1 code implementation • 9 May 2023 • Shanshan Zhong, Wushao Wen, Jinghui Qin, Qiangpu Chen, Zhongzhan Huang
In computer vision, the performance of deep neural networks (DNNs) is highly related to the feature extraction ability, i. e., the ability to recognize and focus on key pixel regions in an image.
1 code implementation • 9 May 2023 • Shanshan Zhong, Zhongzhan Huang, Wushao Wen, Jinghui Qin, Liang Lin
Our approach can make text-to-image diffusion models easier to use with better user experience, which demonstrates our approach has the potential for further advancing the development of user-friendly text-to-image generation models by bridging the semantic gap between simple narrative prompts and complex keyword-based prompts.
no code implementations • 13 Apr 2023 • Shanshan Zhong, Zhongzhan Huang, Wushao Wen, Jinghui Qin, Liang Lin
This technique enables the mitigation of the extra costs for performance improvement during training, such as parameter size and inference time, through these transformations during inference, and therefore SRP has great potential for industrial and practical applications.
no code implementations • 27 Oct 2022 • Shanshan Zhong, Wushao Wen, Jinghui Qin, Zhongzhan Huang
More and more empirical and theoretical evidence shows that deepening neural networks can effectively improve their performance under suitable training settings.
1 code implementation • 6 Oct 2022 • Shanshan Zhong, Jinghui Qin, Zhongzhan Huang, Daifeng Li
However, most existing methods mainly focus on the dialogue context or assist with global satisfaction prediction based on multi-task learning, which ignore the grounded relationships among the causal variables, like the user state and labor cost.
1 code implementation • 13 Sep 2022 • Shanshan Zhong, Wushao Wen, Jinghui Qin
Attention mechanism has gained great success in vision recognition.
1 code implementation • 22 Aug 2022 • Shanshan Zhong, Wushao Wen, Jinghui Qin
Recently many effective attention modules are proposed to boot the model performance by exploiting the internal information of convolutional neural networks in computer vision.