Search Results for author: Guangming Zhu

Found 16 papers, 7 papers with code

Language Model Guided Interpretable Video Action Reasoning

no code implementations2 Apr 2024 Ning Wang, Guangming Zhu, HS Li, Liang Zhang, Syed Afaq Ali Shah, Mohammed Bennamoun

Extensive experiments on two complex video action datasets, Charades & CAD-120, validates the improved performance and interpretability of our LaIAR framework.

Action Recognition Decision Making +3

The two-way knowledge interaction interface between humans and neural networks

no code implementations10 Jan 2024 Zhanliang He, Nuoye Xiong, Hongsheng Li, Peiyi Shen, Guangming Zhu, Liang Zhang

Through experimental validation, based on this interaction interface, NN can provide humans with easily understandable explanations of the reasoning process.

Content-Conditioned Generation of Stylized Free hand Sketches

no code implementations9 Jan 2024 Jiajun Liu, Siyuan Wang, Guangming Zhu, Liang Zhang, Ning li, Eryang Gao

We explore the performance of the model, including using styles randomly sampled from a prior normal distribution to generate images with various free-hand sketching styles, disentangling the painters' styles from known free-hand sketches to generate images with specific styles, and generating images of unknown classes that are not in the training set.

Data Augmentation Image Generation

A multimodal gesture recognition dataset for desktop human-computer interaction

no code implementations8 Jan 2024 Qi Wang, Fengchao Zhu, Guangming Zhu, Liang Zhang, Ning li, Eryang Gao

Gesture recognition is an indispensable component of natural and efficient human-computer interaction technology, particularly in desktop-level applications, where it can significantly enhance people's productivity.

Gesture Recognition

Enhance Sketch Recognition's Explainability via Semantic Component-Level Parsing

1 code implementation13 Dec 2023 Guangming Zhu, Siyuan Wang, Tianci Wu, Liang Zhang

Humans can recognize varied sketches of a category easily by identifying the concurrence and layout of the intrinsic semantic components of the category, since humans draw free-hand sketches based a common consensus that which types of semantic components constitute each sketch category.

Sketch Recognition

Sketch Input Method Editor: A Comprehensive Dataset and Methodology for Systematic Input Recognition

1 code implementation30 Nov 2023 Guangming Zhu, Siyuan Wang, Qing Cheng, Kelong Wu, Hao Li, Liang Zhang

With the recent surge in the use of touchscreen devices, free-hand sketching has emerged as a promising modality for human-computer interaction.

Class Incremental Learning Domain Adaptation +2

Spatio-Temporal Interaction Graph Parsing Networks for Human-Object Interaction Recognition

no code implementations19 Aug 2021 Ning Wang, Guangming Zhu, Liang Zhang, Peiyi Shen, Hongsheng Li, Cong Hua

With the effective spatio-temporal relationship modeling, it is possible not only to uncover contextual information in each frame but also to directly capture inter-time dependencies.

Human-Object Interaction Detection Object

A Systematic Collection of Medical Image Datasets for Deep Learning

1 code implementation24 Jun 2021 Johann Li, Guangming Zhu, Cong Hua, Mingtao Feng, BasheerBennamoun, Ping Li, Xiaoyuan Lu, Juan Song, Peiyi Shen, Xu Xu, Lin Mei, Liang Zhang, Syed Afaq Ali Shah, Mohammed Bennamoun

Thus, as comprehensive as possible, this paper provides a collection of medical image datasets with their associated challenges for deep learning research.

Free-form Description Guided 3D Visual Graph Network for Object Grounding in Point Cloud

1 code implementation ICCV 2021 Mingtao Feng, Zhen Li, Qi Li, Liang Zhang, Xiangdong Zhang, Guangming Zhu, HUI ZHANG, Yaonan Wang, Ajmal Mian

There are three main challenges in 3D object grounding: to find the main focus in the complex and diverse description; to understand the point cloud scene; and to locate the target object.

Object

MeDaS: An open-source platform as service to help break the walls between medicine and informatics

no code implementations12 Jul 2020 Liang Zhang, Johann Li, Ping Li, Xiaoyuan Lu, Peiyi Shen, Guangming Zhu, Syed Afaq Shah, Mohammed Bennarmoun, Kun Qian, Björn W. Schuller

To the best of our knowledge, MeDaS is the first open-source platform proving a collaborative and interactive service for researchers from a medical background easily using DL related toolkits, and at the same time for scientists or engineers from information sciences to understand the medical knowledge side.

Structure-Feature based Graph Self-adaptive Pooling

1 code implementation30 Jan 2020 Liang Zhang, Xudong Wang, Hongsheng Li, Guangming Zhu, Peiyi Shen, Ping Li, Xiaoyuan Lu, Syed Afaq Ali Shah, Mohammed Bennamoun

To solve these problems mentioned above, we propose a novel graph self-adaptive pooling method with the following objectives: (1) to construct a reasonable pooled graph topology, structure and feature information of the graph are considered simultaneously, which provide additional veracity and objectivity in node selection; and (2) to make the pooled nodes contain sufficiently effective graph information, node feature information is aggregated before discarding the unimportant nodes; thus, the selected nodes contain information from neighbor nodes, which can enhance the use of features of the unselected nodes.

Graph Classification

Efficient Scene Text Detection with Textual Attention Tower

no code implementations30 Jan 2020 Liang Zhang, Yufei Liu, Hang Xiao, Lu Yang, Guangming Zhu, Syed Afaq Shah, Mohammed Bennamoun, Peiyi Shen

Scene text detection has received attention for years and achieved an impressive performance across various benchmarks.

Scene Text Detection Text Detection

Attention in Convolutional LSTM for Gesture Recognition

1 code implementation NeurIPS 2018 Liang Zhang, Guangming Zhu, Lin Mei, Peiyi Shen, Syed Afaq Ali Shah, Mohammed Bennamoun

On this basis, a new variant of LSTM is derived, in which the convolutional structures are only embedded into the input-to-state transition of LSTM.

Gesture Recognition

Cannot find the paper you are looking for? You can Submit a new open access paper.