Search Results for author: Chaofan Tao

Found 18 papers, 4 papers with code

Rethinking Kullback-Leibler Divergence in Knowledge Distillation for Large Language Models

no code implementations3 Apr 2024 Taiqiang Wu, Chaofan Tao, Jiahao Wang, Zhe Zhao, Ngai Wong

Kullback-Leiber divergence has been widely used in Knowledge Distillation (KD) to compress Large Language Models (LLMs).

Knowledge Distillation

Electrocardiogram Instruction Tuning for Report Generation

no code implementations7 Mar 2024 Zhongwei Wan, Che Liu, Xin Wang, Chaofan Tao, Hui Shen, Zhenwu Peng, Jie Fu, Rossella Arcucci, Huaxiu Yao, Mi Zhang

Electrocardiogram (ECG) serves as the primary non-invasive diagnostic tool for cardiac conditions monitoring, are crucial in assisting clinicians.

A Spectral Perspective towards Understanding and Improving Adversarial Robustness

no code implementations25 Jun 2023 Binxiao Huang, Rui Lin, Chaofan Tao, Ngai Wong

Deep neural networks (DNNs) are incredibly vulnerable to crafted, imperceptible adversarial perturbations.

Adversarial Robustness

CrossGET: Cross-Guided Ensemble of Tokens for Accelerating Vision-Language Transformers

1 code implementation27 May 2023 Dachuan Shi, Chaofan Tao, Anyi Rao, Zhendong Yang, Chun Yuan, Jiaqi Wang

Although extensively studied for unimodal models, the acceleration for multimodal models, especially the vision-language Transformers, is relatively under-explored.

Image Captioning Image Retrieval +5

UPop: Unified and Progressive Pruning for Compressing Vision-Language Transformers

1 code implementation31 Jan 2023 Dachuan Shi, Chaofan Tao, Ying Jin, Zhendong Yang, Chun Yuan, Jiaqi Wang

Real-world data contains a vast amount of multimodal information, among which vision and language are the two most representative modalities.

Image Captioning Image Classification +7

Frequency Regularization for Improving Adversarial Robustness

no code implementations24 Dec 2022 Binxiao Huang, Chaofan Tao, Rui Lin, Ngai Wong

Deep neural networks are incredibly vulnerable to crafted, human-imperceptible adversarial perturbations.

Adversarial Robustness

ODG-Q: Robust Quantization via Online Domain Generalization

no code implementations17 Oct 2022 Chaofan Tao, Ngai Wong

To our best knowledge, this work is the first work that trains both quantized and binary neural networks on ImageNet that consistently improve robustness under different attacks.

Domain Generalization Quantization

Compression of Generative Pre-trained Language Models via Quantization

no code implementations ACL 2022 Chaofan Tao, Lu Hou, Wei zhang, Lifeng Shang, Xin Jiang, Qun Liu, Ping Luo, Ngai Wong

We find that previous quantization methods fail on generative tasks due to the \textit{homogeneous word embeddings} caused by reduced capacity, and \textit{varied distribution of weights}.

Model Compression Quantization +1

Interpretable Mammographic Image Classification using Case-Based Reasoning and Deep Learning

no code implementations12 Jul 2021 Alina Jade Barnett, Fides Regina Schwartz, Chaofan Tao, Chaofan Chen, Yinhao Ren, Joseph Y. Lo, Cynthia Rudin

Compared to other methods, our model detects clinical features (mass margins) with equal or higher accuracy, provides a more detailed explanation of its prediction, and is better able to differentiate the classification-relevant parts of the image.

Image Classification

IAIA-BL: A Case-based Interpretable Deep Learning Model for Classification of Mass Lesions in Digital Mammography

no code implementations23 Mar 2021 Alina Jade Barnett, Fides Regina Schwartz, Chaofan Tao, Chaofan Chen, Yinhao Ren, Joseph Y. Lo, Cynthia Rudin

Mammography poses important challenges that are not present in other computer vision tasks: datasets are small, confounding information is present, and it can be difficult even for a radiologist to decide between watchful waiting and biopsy based on a mammogram alone.

BIG-bench Machine Learning Interpretable Machine Learning

FAT: Learning Low-Bitwidth Parametric Representation via Frequency-Aware Transformation

1 code implementation15 Feb 2021 Chaofan Tao, Rui Lin, Quan Chen, Zhaoyang Zhang, Ping Luo, Ngai Wong

Prior arts often discretize the network weights by carefully tuning hyper-parameters of quantization (e. g. non-uniform stepsize and layer-wise bitwidths), which are complicated and sub-optimal because the full-precision and low-precision models have a large discrepancy.

Neural Network Compression Quantization

Dynamic and Static Context-aware LSTM for Multi-agent Motion Prediction

no code implementations ECCV 2020 Chaofan Tao, Qinhong Jiang, Lixin Duan, Ping Luo

Existing work addressed this challenge by either learning social spatial interactions represented by the positions of a group of pedestrians, while ignoring their temporal coherence (\textit{i. e.} dependencies between different long trajectories), or by understanding the complicated scene layout (\textit{e. g.} scene segmentation) to ensure safe navigation.

motion prediction Trajectory Prediction

MiniMax Entropy Network: Learning Category-Invariant Features for Domain Adaptation

no code implementations21 Apr 2019 Chaofan Tao, Fengmao Lv, Lixin Duan, Min Wu

Unlike most existing approaches which employ a generator to deal with domain difference, MMEN focuses on learning the categorical information from unlabeled target samples with the help of labeled source samples.

Domain Adaptation

This Looks Like That: Deep Learning for Interpretable Image Recognition

3 code implementations NeurIPS 2019 Chaofan Chen, Oscar Li, Chaofan Tao, Alina Jade Barnett, Jonathan Su, Cynthia Rudin

In this work, we introduce a deep network architecture -- prototypical part network (ProtoPNet), that reasons in a similar way: the network dissects the image by finding prototypical parts, and combines evidence from the prototypes to make a final classification.

General Classification Image Classification

Cannot find the paper you are looking for? You can Submit a new open access paper.