Search Results for author: Ge Li

Found 132 papers, 64 papers with code

DreamPBR: Text-driven Generation of High-resolution SVBRDF with Multi-modal Guidance

no code implementations23 Apr 2024 Linxuan Xin, Zheng Zhang, Jinfu Wei, Ge Li, Duan Gao

Prior material creation methods had limitations in producing diverse results mainly because reconstruction-based methods relied on real-world measurements and generation-based methods were trained on relatively small material datasets.

Exploring and Unleashing the Power of Large Language Models in Automated Code Translation

1 code implementation23 Apr 2024 Zhen Yang, Fang Liu, Zhongxing Yu, Jacky Wai Keung, Jia Li, Shuo Liu, Yifan Hong, Xiaoxue Ma, Zhi Jin, Ge Li

Specifically, UniTrans first craft a series of test cases for target programs with the assistance of source programs.

LLM-Powered Test Case Generation for Detecting Tricky Bugs

no code implementations16 Apr 2024 Kaibo Liu, Yiyang Liu, Zhenpeng Chen, Jie M. Zhang, Yudong Han, Yun Ma, Ge Li, Gang Huang

Conventional automated test generation tools struggle to generate test oracles and tricky bug-revealing test inputs.

Peer-aided Repairer: Empowering Large Language Models to Repair Advanced Student Assignments

no code implementations2 Apr 2024 Qianhui Zhao, Fang Liu, Li Zhang, Yang Liu, Zhen Yan, Zhenghao Chen, Yufei Zhou, Jing Jiang, Ge Li

Automated generation of feedback on programming assignments holds significant benefits for programming education, especially when it comes to advanced assignments.

Language Modelling Large Language Model +1

EvoCodeBench: An Evolving Code Generation Benchmark Aligned with Real-World Code Repositories

1 code implementation31 Mar 2024 Jia Li, Ge Li, Xuanming Zhang, Yihong Dong, Zhi Jin

Existing benchmarks demonstrate poor alignment with real-world code repositories and are insufficient to evaluate the coding abilities of LLMs.

Code Generation

ST-LLM: Large Language Models Are Effective Temporal Learners

1 code implementation30 Mar 2024 Ruyang Liu, Chen Li, Haoran Tang, Yixiao Ge, Ying Shan, Ge Li

In this paper, we investigate a straightforward yet unexplored question: Can we feed all spatial-temporal tokens into the LLM, thus delegating the task of video sequence modeling to the LLMs?

Reading Comprehension Video Understanding

Evaluating Large Language Models with Runtime Behavior of Program Execution

no code implementations25 Mar 2024 Junkai Chen, Zhiyuan Pan, Xing Hu, Zhenhao Li, Ge Li, Xin Xia

Typically, they focus on predicting the input and output of a program, ignoring the evaluation of the intermediate behavior during program execution, as well as the logical consistency (e. g., the model should not give the correct output if the prediction of execution path is wrong) when performing the reasoning.

SEED: Customize Large Language Models with Sample-Efficient Adaptation for Code Generation

no code implementations29 Feb 2024 Xue Jiang, Yihong Dong, Zhi Jin, Ge Li

Specifically, SEED involves identifying error code generated by LLMs, employing Self-revise for code revision, optimizing the model with revised code, and iteratively adapting the process for continuous improvement.

Code Generation

Generalization or Memorization: Data Contamination and Trustworthy Evaluation for Large Language Models

1 code implementation24 Feb 2024 Yihong Dong, Xue Jiang, Huanyu Liu, Zhi Jin, Ge Li

CDD necessitates only the sampled texts to detect data contamination, by identifying the peakedness of LLM's output distribution.

Memorization

Hierarchical Prior-based Super Resolution for Point Cloud Geometry Compression

1 code implementation17 Feb 2024 Dingquan Li, Kede Ma, Jing Wang, Ge Li

The content-dependent hierarchical prior is constructed at the encoder side, which enables coarse-to-fine super resolution of the point cloud geometry at the decoder side.

Quantization Super-Resolution

Open the Black Box: Step-based Policy Updates for Temporally-Correlated Episodic Reinforcement Learning

1 code implementation21 Jan 2024 Ge Li, Hongyi Zhou, Dominik Roth, Serge Thilges, Fabian Otto, Rudolf Lioutikov, Gerhard Neumann

Current advancements in reinforcement learning (RL) have predominantly focused on learning step-based policies that generate actions for each perceived state.

Reinforcement Learning (RL)

Uncertainty-aware No-Reference Point Cloud Quality Assessment

no code implementations17 Jan 2024 Songlin Fan, Zixuan Guo, Wei Gao, Ge Li

The evolution of compression and enhancement algorithms necessitates an accurate quality assessment for point clouds.

Point Cloud Quality Assessment

DevEval: Evaluating Code Generation in Practical Software Projects

no code implementations12 Jan 2024 Jia Li, Ge Li, YunFei Zhao, Yongmin Li, Zhi Jin, Hao Zhu, Huanyu Liu, Kaibo Liu, Lecheng Wang, Zheng Fang, Lanshen Wang, Jiazheng Ding, Xuanming Zhang, Yihong Dong, Yuqi Zhu, Bin Gu, Mengfei Yang

Compared to previous benchmarks, DevEval aligns to practical projects in multiple dimensions, e. g., real program distributions, sufficient dependencies, and enough-scale project contexts.

Code Generation

TACO: Topics in Algorithmic COde generation dataset

1 code implementation22 Dec 2023 Rongao Li, Jie Fu, Bo-Wen Zhang, Tao Huang, Zhihong Sun, Chen Lyu, Guang Liu, Zhi Jin, Ge Li

Moreover, each TACO problem includes several fine-grained labels such as task topics, algorithms, programming skills, and difficulty levels, providing a more precise reference for the training and evaluation of code generation models.

Code Generation

Adaptive Annotation Distribution for Weakly Supervised Point Cloud Semantic Segmentation

no code implementations11 Dec 2023 Zhiyi Pan, Nan Zhang, Wei Gao, Shan Liu, Ge Li

Based on our analysis, we propose a label-aware point cloud downsampling strategy to increase the proportion of annotations involved in the training stage.

Semantic Segmentation

Mug-STAN: Adapting Image-Language Pretrained Models for General Video Understanding

1 code implementation25 Nov 2023 Ruyang Liu, Jingjia Huang, Wei Gao, Thomas H. Li, Ge Li

Large-scale image-language pretrained models, e. g., CLIP, have demonstrated remarkable proficiency in acquiring general multi-modal knowledge through web-scale image-text data.

Video Understanding

Lightweight super resolution network for point cloud geometry compression

1 code implementation2 Nov 2023 Wei zhang, Dingquan Li, Ge Li, Wen Gao

This paper presents an approach for compressing point cloud geometry by leveraging a lightweight super-resolution network.

Point cloud reconstruction Point Cloud Super Resolution +1

ChatCoder: Chat-based Refine Requirement Improves LLMs' Code Generation

no code implementations1 Nov 2023 Zejun Wang, Jia Li, Ge Li, Zhi Jin

To help human users refine their requirements and improve large language models' code generation performances, we propose ChatCoder: a method to refine the requirements via chatting with large language models.

Code Generation

Large Language Model-Aware In-Context Learning for Code Generation

no code implementations15 Oct 2023 Ge Li, Chongyang Tao, Jia Li, Huangzhao Zhang, Fang Liu, Zhi Jin

Large language models (LLMs) have shown impressive in-context learning (ICL) ability in code generation.

Code Generation Contrastive Learning +3

Hot or Cold? Adaptive Temperature Sampling for Code Generation with Large Language Models

1 code implementation6 Sep 2023 Yuqi Zhu, Ge Li, YunFei Zhao, Jia Li, Zhi Jin, Hong Mei

With an analysis of loss distributions of code tokens, we find that code tokens can be divided into two categories: challenging tokens that are difficult to predict and confident tokens that can be easily inferred.

Code Generation

ZC3: Zero-Shot Cross-Language Code Clone Detection

1 code implementation26 Aug 2023 Chongyang Tao, Zhi Jin, Fang Liu, Jia Li, Ge Li

In this paper, we propose a novel method named ZC3 for Zero-shot Cross-language Code Clone detection.

Clone Detection Language Modelling

EditSum: A Retrieve-and-Edit Framework for Source Code Summarization

no code implementations26 Aug 2023 Jia Li, Yongmin Li, Ge Li, Xing Hu, Xin Xia, Zhi Jin

Besides the patternized words, a code summary also contains important keywords, which are the key to reflecting the functionality of the code.

Code Summarization Informativeness +1

PACE: Improving Prompt with Actor-Critic Editing for Large Language Model

no code implementations19 Aug 2023 Yihong Dong, Kangcheng Luo, Xue Jiang, Zhi Jin, Ge Li

Large language models (LLMs) have showcased remarkable potential across various tasks by conditioning on prompts.

Language Modelling Large Language Model

MP3: Movement Primitive-Based (Re-)Planning Policy

no code implementations22 Jun 2023 Fabian Otto, Hongyi Zhou, Onur Celik, Ge Li, Rudolf Lioutikov, Gerhard Neumann

We introduce a novel deep reinforcement learning (RL) approach called Movement Primitive-based Planning Policy (MP3).

Reinforcement Learning (RL)

LIVABLE: Exploring Long-Tailed Classification of Software Vulnerability Types

1 code implementation12 Jun 2023 Xin-Cheng Wen, Cuiyun Gao, Feng Luo, Haoyu Wang, Ge Li, Qing Liao

(2) adaptive re-weighting module, which adjusts the learning weights for different types according to the training epochs and numbers of associated samples by a novel training loss.

Classification Representation Learning +1

Structured Chain-of-Thought Prompting for Code Generation

no code implementations11 May 2023 Jia Li, Ge Li, Yongmin Li, Zhi Jin

In this paper, we propose Structured CoTs (SCoTs) and present a novel prompting technique for code generation, named SCoT prompting.

Code Generation Text Generation

Self-Edit: Fault-Aware Code Editor for Code Generation

no code implementations6 May 2023 Kechi Zhang, Zhuo Li, Jia Li, Ge Li, Zhi Jin

Inspired by the process of human programming, we propose a generate-and-edit approach named Self-Edit that utilizes execution results of the generated code from LLMs to improve the code quality on the competitive programming task.

Code Generation

Implant Global and Local Hierarchy Information to Sequence based Code Representation Models

1 code implementation14 Mar 2023 Kechi Zhang, Zhuo Li, Zhi Jin, Ge Li

Furthermore, we propose the Hierarchy Transformer (HiT), a simple but effective sequence model to incorporate the complete hierarchical embeddings of source code into a Transformer model.

A Provably Secure Strong PUF based on LWE: Construction and Implementation

no code implementations5 Mar 2023 Xiaodan Xi, Ge Li, Ye Wang, Yeonsoo Jeon, Michael Orshansky

We construct lattice PUF with a physically obfuscated key and an LWE decryption function block.

Revisiting Temporal Modeling for CLIP-based Image-to-Video Knowledge Transferring

1 code implementation CVPR 2023 Ruyang Liu, Jingjia Huang, Ge Li, Jiashi Feng, Xinglong Wu, Thomas H. Li

In this paper, based on the CLIP model, we revisit temporal modeling in the context of image-to-video knowledge transferring, which is the key point for extending image-text pretrained models to the video domain.

Ranked #7 on Video Retrieval on MSR-VTT-1kA (using extra training data)

Representation Learning Retrieval +3

Improving Graph Representation for Point Cloud Segmentation via Attentive Filtering

no code implementations CVPR 2023 Nan Zhang, Zhiyi Pan, Thomas H. Li, Wei Gao, Ge Li

Recently, self-attention networks achieve impressive performance in point cloud segmentation due to their superiority in modeling long-range dependencies.

Point Cloud Segmentation

Efficient Hierarchical Entropy Model for Learned Point Cloud Compression

no code implementations CVPR 2023 Rui Song, Chunyang Fu, Shan Liu, Ge Li

Learning an accurate entropy model is a fundamental way to remove the redundancy in point cloud compression.

AdaNIC: Towards Practical Neural Image Compression via Dynamic Transform Routing

1 code implementation ICCV 2023 Lvfang Tao, Wei Gao, Ge Li, Chenhao Zhang

Compressive autoencoders (CAEs) play an important role in deep learning-based image compression, but large-scale CAEs are computationally expensive.

Image Compression

Fine-Tuning Pre-Trained Language Models Effectively by Optimizing Subnetworks Adaptively

1 code implementation3 Nov 2022 Haojie Zhang, Ge Li, Jia Li, Zhongjin Zhang, Yuqi Zhu, Zhi Jin

Large-scale pre-trained language models have achieved impressive results on a wide range of downstream tasks recently.

Language Modelling

CodePAD: Sequence-based Code Generation with Pushdown Automaton

1 code implementation2 Nov 2022 Yihong Dong, Xue Jiang, Yuchen Liu, Ge Li, Zhi Jin

CodePAD can leverage existing sequence-based models, and we show that it can achieve 100\% grammatical correctness percentage on these benchmark datasets.

Code Generation Text Generation

Poison Attack and Defense on Deep Source Code Processing Models

no code implementations31 Oct 2022 Jia Li, Zhuo Li, Huangzhao Zhang, Ge Li, Zhi Jin, Xing Hu, Xin Xia

The attackers aim to inject insidious backdoors into models by poisoning the training data with poison samples.

Clone Detection Code Repair +1

CodeEditor: Learning to Edit Source Code with Pre-trained Models

1 code implementation31 Oct 2022 Jia Li, Ge Li, Zhuo Li, Zhi Jin, Xing Hu, Kechi Zhang, Zhiyi Fu

Pre-trained models are first pre-trained with pre-training tasks and fine-tuned with the code editing task.

Language Modelling Masked Language Modeling

Frequency-Aware Self-Supervised Monocular Depth Estimation

1 code implementation11 Oct 2022 Xingyu Chen, Thomas H. Li, Ruonan Zhang, Ge Li

We present two versatile methods to generally enhance self-supervised monocular depth estimation (MDE) models.

Depth Prediction Monocular Depth Estimation +1

ProDMPs: A Unified Perspective on Dynamic and Probabilistic Movement Primitives

no code implementations4 Oct 2022 Ge Li, Zeqi Jin, Michael Volpp, Fabian Otto, Rudolf Lioutikov, Gerhard Neumann

MPs can be broadly categorized into two types: (a) dynamics-based approaches that generate smooth trajectories from any initial state, e. g., Dynamic Movement Primitives (DMPs), and (b) probabilistic approaches that capture higher-order statistics of the motion, e. g., Probabilistic Movement Primitives (ProMPs).

Numerical Integration

Antecedent Predictions Are More Important Than You Think: An Effective Method for Tree-Based Code Generation

no code implementations22 Aug 2022 Yihong Dong, Ge Li, Xue Jiang, Zhi Jin

To evaluate the effectiveness of our proposed loss, we implement and train an Antecedent Prioritized Tree-based code generation model called APT.

Code Generation Position

Incorporating Domain Knowledge through Task Augmentation for Front-End JavaScript Code Generation

no code implementations22 Aug 2022 Sijie Shen, Xiang Zhu, Yihong Dong, Qizhi Guo, Yankun Zhen, Ge Li

However, in some domain-specific scenarios, building such a large paired corpus for code generation is difficult because there is no directly available pairing data, and a lot of effort is required to manually write the code descriptions to construct a high-quality training dataset.

Code Generation

Learning Program Representations with a Tree-Structured Transformer

1 code implementation18 Aug 2022 Wenhan Wang, Kechi Zhang, Ge Li, Shangqing Liu, Anran Li, Zhi Jin, Yang Liu

Learning vector representations for programs is a critical step in applying deep learning techniques for program understanding tasks.

Representation Learning

OpenMedIA: Open-Source Medical Image Analysis Toolbox and Benchmark under Heterogeneous AI Computing Platforms

no code implementations11 Aug 2022 Jia-Xin Zhuang, Xiansong Huang, Yang Yang, Jiancong Chen, Yue Yu, Wei Gao, Ge Li, Jie Chen, Tong Zhang

In this paper, we present OpenMedIA, an open-source toolbox library containing a rich set of deep learning methods for medical image analysis under heterogeneous Artificial Intelligence (AI) computing platforms.

Image Classification Medical Image Classification +2

Salient Object Detection for Point Clouds

1 code implementation25 Jul 2022 Songlin Fan, Wei Gao, Ge Li

This paper researches the unexplored task-point cloud salient object detection (SOD).

Object object-detection +2

What does Transformer learn about source code?

no code implementations18 Jul 2022 Kechi Zhang, Ge Li, Zhi Jin

In the field of source code processing, the transformer-based representation models have shown great powerfulness and have achieved state-of-the-art (SOTA) performance in many tasks.

Variable misuse

SKFlow: Learning Optical Flow with Super Kernels

1 code implementation29 May 2022 Shangkun Sun, Yuanqi Chen, Yu Zhu, Guodong Guo, Ge Li

In this paper, we propose the Super Kernel Flow Network (SKFlow), a CNN architecture to ameliorate the impacts of occlusions on optical flow estimation.

Optical Flow Estimation

Nebula-I: A General Framework for Collaboratively Training Deep Learning Models on Low-Bandwidth Cloud Clusters

1 code implementation19 May 2022 Yang Xiang, Zhihua Wu, Weibao Gong, Siyu Ding, Xianjie Mo, Yuang Liu, Shuohuan Wang, Peng Liu, Yongshuai Hou, Long Li, Bin Wang, Shaohuai Shi, Yaqian Han, Yue Yu, Ge Li, Yu Sun, Yanjun Ma, dianhai yu

We took natural language processing (NLP) as an example to show how Nebula-I works in different training phases that include: a) pre-training a multilingual language model using two remote clusters; and b) fine-tuning a machine translation model using knowledge distilled from pre-trained models, which run through the most popular paradigm of recent deep learning.

Cross-Lingual Natural Language Inference Distributed Computing +2

Deep Geometry Post-Processing for Decompressed Point Clouds

1 code implementation29 Apr 2022 Xiaoqing Fan, Ge Li, Dingquan Li, Yurui Ren, Wei Gao, Thomas H. Li

Point cloud compression plays a crucial role in reducing the huge cost of data storage and transmission.

Quantization

Self-Supervised Arbitrary-Scale Point Clouds Upsampling via Implicit Neural Representation

1 code implementation CVPR 2022 Wenbo Zhao, Xianming Liu, Zhiwei Zhong, Junjun Jiang, Wei Gao, Ge Li, Xiangyang Ji

Most existing methods either take the end-to-end supervised learning based manner, where large amounts of pairs of sparse input and dense ground-truth are exploited as supervision information; or treat up-scaling of different scale factors as independent tasks, and have to build multiple networks to handle upsampling with varying factors.

Self-Supervised Learning

Neural Texture Extraction and Distribution for Controllable Person Image Synthesis

1 code implementation CVPR 2022 Yurui Ren, Xiaoqing Fan, Ge Li, Shan Liu, Thomas H. Li

Our model is trained to predict human images in arbitrary poses, which encourages it to extract disentangled and expressive neural textures representing the appearance of different semantic entities.

Image Generation

Multi-direction and Multi-scale Pyramid in Transformer for Video-based Pedestrian Retrieval

1 code implementation12 Feb 2022 Xianghao Zang, Ge Li, Wei Gao

To fuse multi-scale feature representation, this paper presents a pyramid structure containing global-level information and many pieces of local-level information from different scales.

Person Re-Identification Retrieval

OctAttention: Octree-Based Large-Scale Contexts Model for Point Cloud Compression

1 code implementation12 Feb 2022 Chunyang Fu, Ge Li, Rui Song, Wei Gao, Shan Liu

In point cloud compression, sufficient contexts are significant for modeling the point cloud distribution.

Multimodal data matters: language model pre-training over structured and unstructured electronic health records

1 code implementation25 Jan 2022 Sicen Liu, Xiaolong Wang, Yongshuai Hou, Ge Li, Hui Wang, Hui Xu, Yang Xiang, Buzhou Tang

As two important textual modalities in electronic health records (EHR), both structured data (clinical codes) and unstructured data (clinical narratives) have recently been increasingly applied to the healthcare domain.

Decision Making Language Modelling +1

Contextual Debiasing for Visual Recognition With Causal Mechanisms

1 code implementation CVPR 2022 Ruyang Liu, Hao liu, Ge Li, Haodi Hou, TingHao Yu, Tao Yang

As a common problem in the visual world, contextual bias means the recognition may depend on the co-occurrence context rather than the objects themselves, which is even more severe in multi-label tasks due to multiple targets and the absence of location.

Causal Inference counterfactual +2

Learning to Share in Multi-Agent Reinforcement Learning

2 code implementations16 Dec 2021 Yuxuan Yi, Ge Li, YaoWei Wang, Zongqing Lu

Inspired by the fact that sharing plays a key role in human's learning of cooperation, we propose LToS, a hierarchically decentralized MARL framework that enables agents to learn to dynamically share reward with neighbors so as to encourage agents to cooperate on the global objective through collectives.

Multi-agent Reinforcement Learning reinforcement-learning +1

Specializing Versatile Skill Libraries using Local Mixture of Experts

1 code implementation8 Dec 2021 Onur Celik, Dongzhuoran Zhou, Ge Li, Philipp Becker, Gerhard Neumann

This local and incremental learning results in a modular MoE model of high accuracy and versatility, where both properties can be scaled by adding more components on the fly.

Incremental Learning Reinforcement Learning (RL)

Precise Learning of Source Code Contextual Semantics via Hierarchical Dependence Structure and Graph Attention Networks

no code implementations20 Nov 2021 Zhehao Zhao, Bo Yang, Ge Li, Huai Liu, Zhi Jin

Based on that, we also designed a neural network that depends on the graph attention mechanism. Specifically, we introduced the syntactic structural of the basic block, i. e., its corresponding AST, in source code model to provide sufficient information and fill the gap.

Feature Engineering Graph Attention

Learning to Disentangle Scenes for Person Re-identification

1 code implementation10 Nov 2021 Xianghao Zang, Ge Li, Wei Gao, Xiujun Shu

In this way, the complex scenes in the ReID task are effectively disentangled, and the burden of each branch is relieved.

Person Re-Identification

Machine Learning for Multimodal Electronic Health Records-based Research: Challenges and Perspectives

no code implementations9 Nov 2021 Ziyi Liu, JiaQi Zhang, Yongshuai Hou, Xinran Zhang, Ge Li, Yang Xiang

Background: Electronic Health Records (EHRs) contain rich information of patients' health history, which usually include both structured and unstructured data.

BIG-bench Machine Learning

PIRenderer: Controllable Portrait Image Generation via Semantic Neural Rendering

1 code implementation ICCV 2021 Yurui Ren, Ge Li, Yuanqi Chen, Thomas H. Li, Shan Liu

The proposed model can generate photo-realistic portrait images with accurate movements according to intuitive modifications.

Image Generation Neural Rendering

Power-Based Attacks on Spatial DNN Accelerators

no code implementations28 Aug 2021 Ge Li, Mohit Tiwari, Michael Orshansky

Spatial accelerators, that parallelize matrix/vector operations, are utilized for enhancing energy efficiency of DNN computation.

Model extraction

Combining Attention with Flow for Person Image Synthesis

no code implementations4 Aug 2021 Yurui Ren, Yubo Wu, Thomas H. Li, Shan Liu, Ge Li

Pose-guided person image synthesis aims to synthesize person images by transforming reference images into target poses.

Image Generation

Large-Scale Spatio-Temporal Person Re-identification: Algorithms and Benchmark

2 code implementations31 May 2021 Xiujun Shu, Xiao Wang, Xianghao Zang, Shiliang Zhang, Yuanqi Chen, Ge Li, Qi Tian

We also verified that models pre-trained on LaST can generalize well on existing datasets with short-term and cloth-changing scenarios.

Person Re-Identification

Low Pass Filter for Anti-aliasing in Temporal Action Localization

no code implementations23 Apr 2021 Cece Jin, Yuanqi Chen, Ge Li, Tao Zhang, Thomas Li

This paper aims to verify the existence of aliasing in TAL methods and investigate utilizing low pass filters to solve this problem by inhibiting the high-frequency band.

Temporal Action Localization

SSD-GAN: Measuring the Realness in the Spatial and Spectral Domains

1 code implementation10 Dec 2020 Yuanqi Chen, Ge Li, Cece Jin, Shan Liu, Thomas Li

This issue makes the generator lack the incentive from the discriminator to learn high-frequency content of data, resulting in a significant spectrum discrepancy between generated images and real images.

Learning to Represent Programs with Heterogeneous Graphs

no code implementations8 Dec 2020 Kechi Zhang, Wenhan Wang, Huangzhao Zhang, Ge Li, Zhi Jin

To address the information of node and edge types, we bring the idea of heterogeneous graphs to learning on source code and present a new formula of building heterogeneous program graphs from ASTs with additional type information for nodes and edges.

Code Comment Generation Comment Generation

Point Cloud Attribute Compression via Successive Subspace Graph Transform

no code implementations29 Oct 2020 Yueru Chen, Yiting shao, Jing Wang, Ge Li, C. -C. Jay Kuo

Inspired by the recently proposed successive subspace learning (SSL) principles, we develop a successive subspace graph transform (SSGT) to address point cloud attribute compression in this work.

Attribute

Retrieve and Refine: Exemplar-based Neural Comment Generation

1 code implementation9 Oct 2020 Bolin Wei, Yongmin Li, Ge Li, Xin Xia, Zhi Jin

Inspired by the IR-based and template-based approaches, in this paper, we propose a neural comment generation approach where we use the existing comments of similar code snippets as exemplars to guide comment generation.

Code Comment Generation Comment Generation +4

Towards Full-line Code Completion with Neural Language Models

no code implementations18 Sep 2020 Wenhan Wang, Sijie Shen, Ge Li, Zhi Jin

In this paper, we take a further step and discuss the probability of directly completing a whole line of code instead of a single token.

Code Completion

Toward Zero-Shot Unsupervised Image-to-Image Translation

1 code implementation28 Jul 2020 Yuanqi Chen, Xiaoming Yu, Shan Liu, Ge Li

Recent studies have shown remarkable success in unsupervised image-to-image translation.

Attribute Translation +2

A Technical Report for VIPriors Image Classification Challenge

no code implementations17 Jul 2020 Zhipeng Luo, Ge Li, Zhiguang Zhang

This paper is a brief report to our submission to the VIPriors Image Classification Challenge.

Classification Ensemble Learning +3

GID-Net: Detecting Human-Object Interaction with Global and Instance Dependency

no code implementations11 Mar 2020 Dongming Yang, Yuexian Zou, Jian Zhang, Ge Li

GID block breaks through the local neighborhoods and captures long-range dependency of pixels both in global-level and instance-level from the scene to help detecting interactions between instances.

Human-Object Interaction Detection Object

Deep Image Spatial Transformation for Person Image Generation

2 code implementations CVPR 2020 Yurui Ren, Xiaoming Yu, Junming Chen, Thomas H. Li, Ge Li

Finally, we warp the source features using a content-aware sampling method with the obtained local attention coefficients.

Image Generation

Detecting Code Clones with Graph Neural Networkand Flow-Augmented Abstract Syntax Tree

1 code implementation20 Feb 2020 Wenhan Wang, Ge Li, Bo Ma, Xin Xia, Zhi Jin

As far as we have concerned, we are the first to apply graph neural networks on the domain of code clone detection.

Clone Detection

NLocalSAT: Boosting Local Search with Solution Prediction

1 code implementation26 Jan 2020 Wenjie Zhang, Zeyu Sun, Qihao Zhu, Ge Li, Shaowei Cai, Yingfei Xiong, Lu Zhang

However, in this method, the initialization is assigned in a random manner, which impacts the effectiveness of SLS solvers.

C3DVQA: Full-Reference Video Quality Assessment with 3D Convolutional Neural Network

no code implementations30 Oct 2019 Munan Xu, Junming Chen, Haiqiang Wang, Shan Liu, Ge Li, Zhiqiang Bai

However, video quality exhibits different characteristics from static image quality due to the existence of temporal masking effects.

Video Quality Assessment Visual Question Answering (VQA)

Code Generation as a Dual Task of Code Summarization

2 code implementations NeurIPS 2019 Bolin Wei, Ge Li, Xin Xia, Zhiyi Fu, Zhi Jin

Code summarization (CS) and code generation (CG) are two crucial tasks in the field of automatic software development.

Code Generation Code Summarization +1

Multi-mapping Image-to-Image Translation via Learning Disentanglement

1 code implementation NeurIPS 2019 Xiaoming Yu, Yuanqi Chen, Thomas Li, Shan Liu, Ge Li

Recent advances of image-to-image translation focus on learning the one-to-many mapping from two aspects: multi-modal translation and multi-domain translation.

Disentanglement Image-to-Image Translation +1

A Self-Attentional Neural Architecture for Code Completion with Multi-Task Learning

no code implementations16 Sep 2019 Fang Liu, Ge Li, Bolin Wei, Xin Xia, Zhiyi Fu, Zhi Jin

To enable the knowledge sharing between related tasks, we creatively propose a Multi-Task Learning (MTL) framework to learn two related tasks in code completion jointly.

Code Completion Language Modelling +1

C-RPNs: Promoting Object Detection in real world via a Cascade Structure of Region Proposal Networks

no code implementations19 Aug 2019 Dongming Yang, Yuexian Zou, Jian Zhang, Ge Li

Although two-stage detectors like Faster R-CNN achieved big successes in object detection due to the strategy of extracting region proposals by region proposal network, they show their poor adaption in real-world object detection as a result of without considering mining hard samples during extracting region proposals.

Object object-detection +2

ARMIN: Towards a More Efficient and Light-weight Recurrent Memory Network

1 code implementation28 Jun 2019 Zhangheng Li, Jia-Xing Zhong, Jingjia Huang, Tao Zhang, Thomas Li, Ge Li

In recent years, memory-augmented neural networks(MANNs) have shown promising power to enhance the memory ability of neural networks for sequential processing tasks.

Deep AutoEncoder-based Lossy Geometry Compression for Point Clouds

no code implementations18 Apr 2019 Wei Yan, Yiting shao, Shan Liu, Thomas H. Li, Zhu Li, Ge Li

Point cloud is a fundamental 3D representation which is widely used in real world applications such as autonomous driving.

Autonomous Driving Image Compression

Bi-Skip: A Motion Deblurring Network Using Self-paced Learning

no code implementations24 Feb 2019 Yiwei Zhang, Chunbiao Zhu, Ge Li, Yuan Zhao, Haifeng Shen

A fast and effective motion deblurring method has great application values in real life.

Deblurring

A Grammar-Based Structural CNN Decoder for Code Generation

1 code implementation14 Nov 2018 Zeyu Sun, Qihao Zhu, Lili Mou, Yingfei Xiong, Ge Li, Lu Zhang

In this paper, we propose a grammar-based structural convolutional neural network (CNN) for code generation.

Code Generation Semantic Parsing +1

BLP -- Boundary Likelihood Pinpointing Networks for Accurate Temporal Action Localization

no code implementations6 Nov 2018 Weijie Kong, Nannan Li, Shan Liu, Thomas Li, Ge Li

Despite tremendous progress achieved in temporal action detection, state-of-the-art methods still suffer from the sharp performance deterioration when localizing the starting and ending temporal action boundaries.

Action Detection regression +1

SingleGAN: Image-to-Image Translation by a Single-Generator Network using Multiple Generative Adversarial Learning

1 code implementation11 Oct 2018 Xiaoming Yu, Xing Cai, Zhenqiang Ying, Thomas Li, Ge Li

Besides, we explore variants of SingleGAN for different tasks, including one-to-many domain translation, many-to-many domain translation and one-to-one domain translation with multimodality.

Image-to-Image Translation Translation

Mini-batch Serialization: CNN Training with Inter-layer Data Reuse

1 code implementation30 Sep 2018 Sangkug Lym, Armand Behroozi, Wei Wen, Ge Li, Yongkee Kwon, Mattan Erez

Training convolutional neural networks (CNNs) requires intense computations and high memory bandwidth.

SEQUENCE MODELLING WITH AUTO-ADDRESSING AND RECURRENT MEMORY INTEGRATING NETWORKS

no code implementations27 Sep 2018 Zhangheng Li, Jia-Xing Zhong, Jingjia Huang, Tao Zhang, Thomas Li, Ge Li

Processing sequential data with long term dependencies and learn complex transitions are two major challenges in many deep learning applications.

Step-by-step Erasion, One-by-one Collection: A Weakly Supervised Temporal Action Detector

no code implementations9 Jul 2018 Jia-Xing Zhong, Nannan Li, Weijie Kong, Tao Zhang, Thomas H. Li, Ge Li

Weakly supervised temporal action detection is a Herculean task in understanding untrimmed videos, since no supervisory signal except the video-level category label is available on training data.

Action Detection Temporal Localization

Multi-Mapping Image-to-Image Translation with Central Biasing Normalization

no code implementations26 Jun 2018 Xiaoming Yu, Zhenqiang Ying, Thomas Li, Shan Liu, Ge Li

Recent advances in image-to-image translation have seen a rise in approaches generating diverse images through a single network.

Image-to-Image Translation Translation

Exploiting the Value of the Center-dark Channel Prior for Salient Object Detection

no code implementations14 May 2018 Chunbiao Zhu, Wen-Hao Zhang, Thomas H. Li, Ge Li

In this paper, we propose a novel salient object detection algorithm for RGB-D images using center-dark channel priors.

object-detection RGB Salient Object Detection +2

Deep Learning in Software Engineering

no code implementations13 May 2018 Xiaochen Li, He Jiang, Zhilei Ren, Ge Li, Jing-Xuan Zhang

To answer these questions, we conduct a bibliography analysis on 98 research papers in SE that use deep learning techniques.

Software Engineering

Hybrid Point Cloud Attribute Compression Using Slice-based Layered Structure and Block-based Intra Prediction

no code implementations28 Apr 2018 Yiting Shao, Qi Zhang, Ge Li, Zhu Li

In intra-frame compression of point cloud color attributes, results demonstrate that our method performs better than the state-of-the-art region-adaptive hierarchical transform (RAHT) system, and on average a 29. 37$\%$ BD-rate gain is achieved.

Multimedia

A multilayer backpropagation saliency detection algorithm and its applications

no code implementations26 Mar 2018 Chunbiao Zhu, Ge Li

In this paper, we propose a multilayer backpropagation saliency detection algorithm based on depth mining by which we exploit depth cue from three different layers of images.

object-detection Object Detection +1

A Bio-Inspired Multi-Exposure Fusion Framework for Low-light Image Enhancement

no code implementations2 Nov 2017 Zhenqiang Ying, Ge Li, Wen Gao

Inspired by human visual system, we design a multi-exposure fusion framework for low-light image enhancement.

Low-Light Image Enhancement

Robust Saliency Detection via Fusing Foreground and Background Priors

no code implementations1 Nov 2017 Kan Huang, Chunbiao Zhu, Ge Li

Automatic Salient object detection has received tremendous attention from research community and has been an increasingly important tool in many computer vision tasks.

object-detection RGB Salient Object Detection +2

ORGB: Offset Correction in RGB Color Space for Illumination-Robust Image Processing

no code implementations3 Aug 2017 Zhenqiang Ying, Ge Li, Sixin Wen, Guozhen Tan

This paper is concerned with the detection and correction of the offset between the intersection and origin.

A Self-Adaptive Proposal Model for Temporal Action Detection based on Reinforcement Learning

1 code implementation22 Jun 2017 Jingjia Huang, Nannan Li, Tao Zhang, Ge Li

Existing action detection algorithms usually generate action proposals through an extensive search over the video at multiple temporal scales, which brings about huge computational overhead and deviates from the human perception procedure.

Action Detection Position +2

Compressing Neural Language Models by Sparse Word Representations

1 code implementation ACL 2016 Yunchuan Chen, Lili Mou, Yan Xu, Ge Li, Zhi Jin

Such approaches are time- and memory-intensive because of the large numbers of parameters for word embeddings and the output layer.

Language Modelling Word Embeddings

Searching Action Proposals via Spatial Actionness Estimation and Temporal Path Inference and Tracking

no code implementations23 Aug 2016 Nannan Li, Dan Xu, Zhenqiang Ying, Zhihao LI, Ge Li

In this paper, we address the problem of searching action proposals in unconstrained video clips.

How Transferable are Neural Networks in NLP Applications?

no code implementations EMNLP 2016 Lili Mou, Zhao Meng, Rui Yan, Ge Li, Yan Xu, Lu Zhang, Zhi Jin

Transfer learning is aimed to make use of valuable knowledge in a source domain to help model performance in a target domain.

Transfer Learning

Improved Relation Classification by Deep Recurrent Neural Networks with Data Augmentation

no code implementations COLING 2016 Yan Xu, Ran Jia, Lili Mou, Ge Li, Yunchuan Chen, Yangyang Lu, Zhi Jin

However, existing neural networks for relation classification are usually of shallow architectures (e. g., one-layer convolutional neural networks or recurrent networks).

Classification Data Augmentation +3

Backward and Forward Language Modeling for Constrained Sentence Generation

no code implementations21 Dec 2015 Lili Mou, Rui Yan, Ge Li, Lu Zhang, Zhi Jin

Provided a specific word, we use RNNs to generate previous words and future words, either simultaneously or asynchronously, resulting in two model variants.

Language Modelling Machine Translation +4

On End-to-End Program Generation from User Intention by Deep Neural Networks

no code implementations25 Oct 2015 Lili Mou, Rui Men, Ge Li, Lu Zhang, Zhi Jin

This paper envisions an end-to-end program generation scenario using recurrent neural networks (RNNs): Users can express their intention in natural language; an RNN then automatically generates corresponding code in a characterby-by-character fashion.

A Comparative Study on Regularization Strategies for Embedding-based Neural Networks

no code implementations EMNLP 2015 Hao Peng, Lili Mou, Ge Li, Yunchuan Chen, Yangyang Lu, Zhi Jin

This paper aims to compare different regularization strategies to address a common phenomenon, severe overfitting, in embedding-based neural networks for NLP.

Distilling Word Embeddings: An Encoding Approach

no code implementations15 Jun 2015 Lili Mou, Ran Jia, Yan Xu, Ge Li, Lu Zhang, Zhi Jin

Distilling knowledge from a well-trained cumbersome network to a small one has recently become a new research topic, as lightweight neural networks with high performance are particularly in need in various resource-restricted systems.

Word Embeddings

Convolutional Neural Networks over Tree Structures for Programming Language Processing

8 code implementations18 Sep 2014 Lili Mou, Ge Li, Lu Zhang, Tao Wang, Zhi Jin

Programming language processing (similar to natural language processing) is a hot research topic in the field of software engineering; it has also aroused growing interest in the artificial intelligence community.

Sentence

Building Program Vector Representations for Deep Learning

1 code implementation11 Sep 2014 Lili Mou, Ge Li, Yuxuan Liu, Hao Peng, Zhi Jin, Yan Xu, Lu Zhang

In this pioneering paper, we propose the "coding criterion" to build program vector representations, which are the premise of deep learning for program analysis.

Representation Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.