Search Results for author: Lijie Fan

Found 17 papers, 8 papers with code

Learning Vision from Models Rivals Learning Vision from Data

1 code implementation28 Dec 2023 Yonglong Tian, Lijie Fan, KaiFeng Chen, Dina Katabi, Dilip Krishnan, Phillip Isola

We introduce SynCLR, a novel approach for learning visual representations exclusively from synthetic images and synthetic captions, without any real data.

Contrastive Learning Image Captioning +3

Scaling Laws of Synthetic Images for Model Training ... for Now

1 code implementation7 Dec 2023 Lijie Fan, KaiFeng Chen, Dilip Krishnan, Dina Katabi, Phillip Isola, Yonglong Tian

Our findings also suggest that scaling synthetic data can be particularly effective in scenarios such as: (1) when there is a limited supply of real images for a supervised problem (e. g., fewer than 0. 5 million images in ImageNet), (2) when the evaluation dataset diverges significantly from the training data, indicating the out-of-distribution scenario, or (3) when synthetic data is used in conjunction with real images, as demonstrated in the training of CLIP models.

Improving CLIP Training with Language Rewrites

1 code implementation NeurIPS 2023 Lijie Fan, Dilip Krishnan, Phillip Isola, Dina Katabi, Yonglong Tian

During training, LaCLIP randomly selects either the original texts or the rewritten versions as text augmentations for each image.

In-Context Learning Sentence

Visual Dependency Transformers: Dependency Tree Emerges from Reversed Attention

1 code implementation CVPR 2023 Mingyu Ding, Yikang Shen, Lijie Fan, Zhenfang Chen, Zitian Chen, Ping Luo, Joshua B. Tenenbaum, Chuang Gan

When looking at an image, we can decompose the scene into entities and their parts as well as obtain the dependencies between them.

Unsupervised Learning for Human Sensing Using Radio Signals

no code implementations6 Jul 2022 Tianhong Li, Lijie Fan, Yuan Yuan, Dina Katabi

Thus, in this paper, we explore the feasibility of adapting RGB-based unsupervised representation learning to RF signals.

Action Recognition Contrastive Learning +3

Targeted Supervised Contrastive Learning for Long-Tailed Recognition

1 code implementation CVPR 2022 Tianhong Li, Peng Cao, Yuan Yuan, Lijie Fan, Yuzhe Yang, Rogerio Feris, Piotr Indyk, Dina Katabi

This forces all classes, including minority classes, to maintain a uniform distribution in the feature space, improves class boundaries, and provides better generalization even in the presence of long-tail data.

Contrastive Learning Long-tail Learning

Addressing Feature Suppression in Unsupervised Visual Representations

no code implementations17 Dec 2020 Tianhong Li, Lijie Fan, Yuan Yuan, Hao He, Yonglong Tian, Rogerio Feris, Piotr Indyk, Dina Katabi

However, contrastive learning is susceptible to feature suppression, i. e., it may discard important information relevant to the task of interest, and learn irrelevant features.

Attribute Contrastive Learning +1

In-Home Daily-Life Captioning Using Radio Signals

no code implementations ECCV 2020 Lijie Fan, Tianhong Li, Yuan Yuan, Dina Katabi

This paper aims to caption daily life --i. e., to create a textual description of people's activities and interactions with objects in their homes.

Privacy Preserving Video Captioning

Learning Longterm Representations for Person Re-Identification Using Radio Signals

no code implementations CVPR 2020 Lijie Fan, Tianhong Li, Rongyao Fang, Rumen Hristov, Yuan Yuan, Dina Katabi

RF signals traverse clothes and reflect off the human body; thus they can be used to extract more persistent human-identifying features like body size and shape.

Person Re-Identification Privacy Preserving

Controllable Image-to-Video Translation: A Case Study on Facial Expression Generation

no code implementations9 Aug 2018 Lijie Fan, Wenbing Huang, Chuang Gan, Junzhou Huang, Boqing Gong

The recent advances in deep learning have made it possible to generate photo-realistic images by using neural networks and even to extrapolate video frames from an input video clip.

Facial expression generation Image-to-Image Translation +2

End-to-End Learning of Motion Representation for Video Understanding

1 code implementation CVPR 2018 Lijie Fan, Wenbing Huang, Chuang Gan, Stefano Ermon, Boqing Gong, Junzhou Huang

Despite the recent success of end-to-end learned representations, hand-crafted optical flow features are still widely used in video analysis tasks.

Action Recognition Optical Flow Estimation +1

Spatiotemporal Networks for Video Emotion Recognition

no code implementations3 Apr 2017 Lijie Fan, Yunjie Ke

Our experiment adapts several popular deep learning methods as well as some traditional methods on the problem of video emotion recognition.

Classification General Classification +1

Cannot find the paper you are looking for? You can Submit a new open access paper.