Search Results for author: Mingsheng Long

Found 97 papers, 59 papers with code

depyf: Open the Opaque Box of PyTorch Compiler for Machine Learning Researchers

1 code implementation14 Mar 2024 Kaichao You, Runsheng Bai, Meng Cao, Jianmin Wang, Ion Stoica, Mingsheng Long

PyTorch \texttt{2. x} introduces a compiler designed to accelerate deep learning programs.

TimeXer: Empowering Transformers for Time Series Forecasting with Exogenous Variables

no code implementations29 Feb 2024 Yuxuan Wang, Haixu Wu, Jiaxiang Dong, Yong liu, Yunzhong Qiu, Haoran Zhang, Jianmin Wang, Mingsheng Long

Experimentally, TimeXer significantly improves time series forecasting with exogenous variables and achieves consistent state-of-the-art performance in twelve real-world forecasting benchmarks.

Time Series Time Series Forecasting

EuLagNet: Eulerian Fluid Prediction with Lagrangian Dynamics

no code implementations4 Feb 2024 Qilong Ma, Haixu Wu, Lanxiang Xing, Jianmin Wang, Mingsheng Long

Accurately predicting the future fluid is important to extensive areas, such as meteorology, oceanology and aerodynamics.

Future prediction

TimeSiam: A Pre-Training Framework for Siamese Time-Series Modeling

no code implementations4 Feb 2024 Jiaxiang Dong, Haixu Wu, Yuxuan Wang, Yunzhong Qiu, Li Zhang, Jianmin Wang, Mingsheng Long

To emphasize temporal correlation modeling, this paper proposes TimeSiam as a simple but effective self-supervised pre-training framework for Time series based on Siamese networks.

Contrastive Learning Data Augmentation +1

Transolver: A Fast Transformer Solver for PDEs on General Geometries

no code implementations4 Feb 2024 Haixu Wu, Huakun Luo, Haowen Wang, Jianmin Wang, Mingsheng Long

Transformers have empowered many milestones across various fields and have recently been applied to solve partial differential equations (PDEs).

Timer: Transformers for Time Series Analysis at Scale

1 code implementation4 Feb 2024 Yong liu, Haoran Zhang, Chenyu Li, Xiangdong Huang, Jianmin Wang, Mingsheng Long

Continuous progresses have been achieved as the emergence of large language models, exhibiting unprecedented ability in few-shot generalization, scalability, and task generality, which is however absent in time series models.

Anomaly Detection Imputation +2

AutoTimes: Autoregressive Time Series Forecasters via Large Language Models

1 code implementation4 Feb 2024 Yong liu, Guo Qin, Xiangdong Huang, Jianmin Wang, Mingsheng Long

Foundation models of time series have not been fully developed due to the limited availability of large-scale time series and the underexploration of scalable pre-training.

In-Context Learning Language Modelling +1

HelmFluid: Learning Helmholtz Dynamics for Interpretable Fluid Prediction

no code implementations16 Oct 2023 Lanxiang Xing, Haixu Wu, Yuezhou Ma, Jianmin Wang, Mingsheng Long

Compared with previous velocity estimating methods, HelmFluid is faithfully derived from Helmholtz theorem and ravels out complex fluid dynamics with physically interpretable evidence.

Future prediction

iTransformer: Inverted Transformers Are Effective for Time Series Forecasting

4 code implementations10 Oct 2023 Yong liu, Tengge Hu, Haoran Zhang, Haixu Wu, Shiyu Wang, Lintao Ma, Mingsheng Long

These forecasters leverage Transformers to model the global dependencies over temporal tokens of time series, with each token formed by multiple variates of the same timestamp.

Time Series Time Series Forecasting

On the Embedding Collapse when Scaling up Recommendation Models

no code implementations6 Oct 2023 Xingzhuo Guo, Junwei Pan, Ximei Wang, Baixu Chen, Jie Jiang, Mingsheng Long

Recent advances in deep foundation models have led to a promising trend of developing large recommendation models to leverage vast amounts of available data.

HarmonyDream: Task Harmonization Inside World Models

no code implementations30 Sep 2023 Haoyu Ma, Jialong Wu, Ningya Feng, Chenjun Xiao, Dong Li, Jianye Hao, Jianmin Wang, Mingsheng Long

Model-based reinforcement learning (MBRL) holds the promise of sample-efficient learning by utilizing a world model, which models how the environment works and typically encompasses components for two tasks: observation modeling and reward modeling.

Atari Games 100k Model-based Reinforcement Learning +1

Koopa: Learning Non-stationary Time Series Dynamics with Koopman Predictors

1 code implementation NeurIPS 2023 Yong liu, Chenyu Li, Jianmin Wang, Mingsheng Long

While previous models suffer from complicated series variations induced by changing temporal distribution, we tackle non-stationary time series with modern Koopman theory that fundamentally considers the underlying time-variant dynamics.

Time Series

Pre-training Contextualized World Models with In-the-wild Videos for Reinforcement Learning

1 code implementation NeurIPS 2023 Jialong Wu, Haoyu Ma, Chaoyi Deng, Mingsheng Long

To tackle this issue, we introduce Contextualized World Models (ContextWM) that explicitly separate context and dynamics modeling to overcome the complexity and diversity of in-the-wild videos and facilitate knowledge transfer between distinct scenes.

Autonomous Driving Model-based Reinforcement Learning +3

CLIPood: Generalizing CLIP to Out-of-Distributions

1 code implementation2 Feb 2023 Yang Shu, Xingzhuo Guo, Jialong Wu, Ximei Wang, Jianmin Wang, Mingsheng Long

This paper aims at generalizing CLIP to out-of-distribution test data on downstream tasks.

SimMTM: A Simple Pre-Training Framework for Masked Time-Series Modeling

1 code implementation NeurIPS 2023 Jiaxiang Dong, Haixu Wu, Haoran Zhang, Li Zhang, Jianmin Wang, Mingsheng Long

By relating masked modeling to manifold learning, SimMTM proposes to recover masked time points by the weighted aggregation of multiple neighbors outside the manifold, which eases the reconstruction task by assembling ruined but complementary temporal variations from multiple masked series.

Representation Learning Time Series +1

Solving High-Dimensional PDEs with Latent Spectral Models

1 code implementation30 Jan 2023 Haixu Wu, Tengge Hu, Huakun Luo, Jianmin Wang, Mingsheng Long

A burgeoning paradigm is learning neural operators to approximate the input-output mappings of PDEs.

Vocal Bursts Intensity Prediction

ForkMerge: Mitigating Negative Transfer in Auxiliary-Task Learning

1 code implementation NeurIPS 2023 Junguang Jiang, Baixu Chen, Junwei Pan, Ximei Wang, Liu Dapeng, Jie Jiang, Mingsheng Long

Auxiliary-Task Learning (ATL) aims to improve the performance of the target task by leveraging the knowledge obtained from related tasks.

Out-of-Dynamics Imitation Learning from Multimodal Demonstrations

1 code implementation13 Nov 2022 Yiwen Qiu, Jialong Wu, Zhangjie Cao, Mingsheng Long

Existing imitation learning works mainly assume that the demonstrator who collects demonstrations shares the same dynamics as the imitator.

Imitation Learning

TimesNet: Temporal 2D-Variation Modeling for General Time Series Analysis

3 code implementations5 Oct 2022 Haixu Wu, Tengge Hu, Yong liu, Hang Zhou, Jianmin Wang, Mingsheng Long

TimesBlock can discover the multi-periodicity adaptively and extract the complex temporal variations from transformed 2D tensors by a parameter-efficient inception block.

Action Recognition Anomaly Detection +4

Recommender Transformers with Behavior Pathways

no code implementations13 Jun 2022 Zhiyu Yao, Xinyang Chen, Sinan Wang, Qinyan Dai, Yumeng Li, Tanchao Zhu, Mingsheng Long

We conclude this characteristic for sequential behaviors of each user as the Behavior Pathway.

Sequential Recommendation

Hub-Pathway: Transfer Learning from A Hub of Pre-trained Models

no code implementations8 Jun 2022 Yang Shu, Zhangjie Cao, Ziyang Zhang, Jianmin Wang, Mingsheng Long

The proposed framework can be trained end-to-end with the target task-specific loss, where it learns to explore better pathway configurations and exploit the knowledge in pre-trained models for each target datum.

Transfer Learning

Non-stationary Transformers: Exploring the Stationarity in Time Series Forecasting

1 code implementation28 May 2022 Yong liu, Haixu Wu, Jianmin Wang, Mingsheng Long

However, their performance can degenerate terribly on non-stationary real-world data in which the joint distribution changes over time.

Time Series Time Series Forecasting

MetaSets: Meta-Learning on Point Sets for Generalizable Representations

no code implementations CVPR 2021 Chao Huang, Zhangjie Cao, Yunbo Wang, Jianmin Wang, Mingsheng Long

It is a challenging problem due to the substantial geometry shift from simulated to real data, such that most existing 3D models underperform due to overfitting the complete geometries in the source domain.

Domain Generalization Meta-Learning

Continual Predictive Learning from Videos

1 code implementation CVPR 2022 Geng Chen, Wendong Zhang, Han Lu, Siyu Gao, Yunbo Wang, Mingsheng Long, Xiaokang Yang

Can we develop predictive learning algorithms that can deal with more realistic, non-stationary physical environments?

Continual Learning Test-time Adaptation +1

From Big to Small: Adaptive Learning to Partial-Set Domains

1 code implementation14 Mar 2022 Zhangjie Cao, Kaichao You, Ziyang Zhang, Jianmin Wang, Mingsheng Long

Still, the common requirement of identical class space shared across domains hinders applications of domain adaptation to partial-set domains.

Partial Domain Adaptation

Flowformer: Linearizing Transformers with Conservation Flows

1 code implementation13 Feb 2022 Haixu Wu, Jialong Wu, Jiehui Xu, Jianmin Wang, Mingsheng Long

By respectively conserving the incoming flow of sinks for source competition and the outgoing flow of sources for sink allocation, Flow-Attention inherently generates informative attentions without using specific inductive biases.

Ranked #4 on D4RL on D4RL

D4RL Offline RL +2

Supported Policy Optimization for Offline Reinforcement Learning

3 code implementations13 Feb 2022 Jialong Wu, Haixu Wu, Zihan Qiu, Jianmin Wang, Mingsheng Long

Policy constraint methods to offline reinforcement learning (RL) typically utilize parameterization or regularization that constrains the policy to perform actions within the support set of the behavior policy.

Offline RL reinforcement-learning +1

Transferability in Deep Learning: A Survey

1 code implementation15 Jan 2022 Junguang Jiang, Yang Shu, Jianmin Wang, Mingsheng Long

The success of deep learning algorithms generally depends on large-scale data, while humans appear to have inherent ability of knowledge transfer, by recognizing and applying relevant knowledge from previous learning experiences when encountering and solving unseen tasks.

Domain Adaptation Transfer Learning

Ranking and Tuning Pre-trained Models: A New Paradigm for Exploiting Model Hubs

1 code implementation20 Oct 2021 Kaichao You, Yong liu, Ziyang Zhang, Jianmin Wang, Michael I. Jordan, Mingsheng Long

(2) The best ranked PTM can either be fine-tuned and deployed if we have no preference for the model's architecture or the target PTM can be tuned by the top $K$ ranked PTMs via a Bayesian procedure that we propose.

Omni-Training: Bridging Pre-Training and Meta-Training for Few-Shot Learning

no code implementations14 Oct 2021 Yang Shu, Zhangjie Cao, Jinghan Gao, Jianmin Wang, Philip S. Yu, Mingsheng Long

While pre-training and meta-training can create deep models powerful for few-shot generalization, we find that pre-training and meta-training focuses respectively on cross-domain transferability and cross-task transferability, which restricts their data efficiency in the entangled settings of domain shift and task shift.

Few-Shot Learning Transfer Learning

X-model: Improving Data Efficiency in Deep Learning with A Minimax Model

no code implementations ICLR 2022 Ximei Wang, Xinyang Chen, Jianmin Wang, Mingsheng Long

To take the power of both worlds, we propose a novel X-model by simultaneously encouraging the invariance to {data stochasticity} and {model stochasticity}.

Age Estimation Object Recognition +2

ModeRNN: Harnessing Spatiotemporal Mode Collapse in Unsupervised Predictive Learning

1 code implementation8 Oct 2021 Zhiyu Yao, Yunbo Wang, Haixu Wu, Jianmin Wang, Mingsheng Long

To this end, we propose ModeRNN, which introduces a novel method to learn structured hidden representations between recurrent states.

Inductive Bias

Anomaly Transformer: Time Series Anomaly Detection with Association Discrepancy

3 code implementations ICLR 2022 Jiehui Xu, Haixu Wu, Jianmin Wang, Mingsheng Long

Unsupervised detection of anomaly points in time series is a challenging problem, which requires the model to derive a distinguishable criterion.

Anomaly Detection Time Series +1

Decoupled Adaptation for Cross-Domain Object Detection

2 code implementations ICLR 2022 Junguang Jiang, Baixu Chen, Jianmin Wang, Mingsheng Long

Besides, previous methods focused on category adaptation but ignored another important part for object detection, i. e., the adaptation on bounding box regression.

Domain Adaptation Object +3

Zoo-Tuning: Adaptive Transfer from a Zoo of Models

no code implementations29 Jun 2021 Yang Shu, Zhi Kou, Zhangjie Cao, Jianmin Wang, Mingsheng Long

We propose \emph{Zoo-Tuning} to address these challenges, which learns to adaptively transfer the parameters of pretrained models to the target task.

Facial Landmark Detection Image Classification +1

Transferable Query Selection for Active Domain Adaptation

no code implementations CVPR 2021 Bo Fu, Zhangjie Cao, Jianmin Wang, Mingsheng Long

Due to the domain shift, the query selection criteria of prior active learning methods may be ineffective to select the most informative target samples for annotation.

Active Learning Unsupervised Domain Adaptation

MetaSets:Meta-Learning on Point Sets for Generalizable Representations

no code implementations CVPR 2021 Chao Huang, Zhangjie Cao, Yunbo Wang, Jianmin Wang, Mingsheng Long

It is a challenging problem due to the substantial geometry shift from simulated to real data, such that most existing 3D models underperform due to overfitting the complete geometries in the source domain.

Domain Generalization

Open Domain Generalization with Domain-Augmented Meta-Learning

no code implementations CVPR 2021 Yang Shu, Zhangjie Cao, Chenyu Wang, Jianmin Wang, Mingsheng Long

Leveraging datasets available to learn a model with high generalization ability to unseen domains is important for computer vision, especially when the unseen domain's annotated data are unavailable.

Domain Generalization Meta-Learning

PredRNN: A Recurrent Neural Network for Spatiotemporal Predictive Learning

3 code implementations17 Mar 2021 Yunbo Wang, Haixu Wu, Jianjin Zhang, Zhifeng Gao, Jianmin Wang, Philip S. Yu, Mingsheng Long

This paper models these structures by presenting PredRNN, a new recurrent network, in which a pair of memory cells are explicitly decoupled, operate in nearly independent transition manners, and finally form unified representations of the complex environment.

 Ranked #1 on Video Prediction on KTH (Cond metric)

Video Prediction Weather Forecasting

Regressive Domain Adaptation for Unsupervised Keypoint Detection

2 code implementations CVPR 2021 Junguang Jiang, Yifei Ji, Ximei Wang, Yufeng Liu, Jianmin Wang, Mingsheng Long

First, based on our observation that the probability density of the output space is sparse, we introduce a spatial probability distribution to describe this sparsity and then use it to guide the learning of the adversarial regressor.

Domain Adaptation Keypoint Detection

MotionRNN: A Flexible Model for Video Prediction with Spacetime-Varying Motions

1 code implementation CVPR 2021 Haixu Wu, Zhiyu Yao, Jianmin Wang, Mingsheng Long

With high flexibility, this framework can adapt to a series of models for deterministic spatiotemporal prediction.

Video Prediction

Self-Tuning for Data-Efficient Deep Learning

2 code implementations25 Feb 2021 Ximei Wang, Jinghan Gao, Mingsheng Long, Jianmin Wang

Deep learning has made revolutionary advances to diverse applications in the presence of large-scale labeled datasets.

Transfer Learning

LogME: Practical Assessment of Pre-trained Models for Transfer Learning

1 code implementation22 Feb 2021 Kaichao You, Yong liu, Jianmin Wang, Mingsheng Long

In pursuit of a practical assessment method, we propose to estimate the maximum value of label evidence given features extracted by pre-trained models.

Model Selection regression +2

Stochastic Normalization

2 code implementations NeurIPS 2020 Zhi Kou, Kaichao You, Mingsheng Long, Jianmin Wang

During training, two branches are stochastically selected to avoid over-depending on some sample statistics, resulting in a strong regularization effect, which we interpret as ``architecture regularization.''

Co-Tuning for Transfer Learning

2 code implementations NeurIPS 2020 Kaichao You, Zhi Kou, Mingsheng Long, Jianmin Wang

Fine-tuning pre-trained deep neural networks (DNNs) to a target dataset, also known as transfer learning, is widely used in computer vision and NLP.

Image Classification Transfer Learning +1

Learning to Adapt to Evolving Domains

1 code implementation NeurIPS 2020 Hong Liu, Mingsheng Long, Jianmin Wang, Yu Wang

(2) Since the target data arrive online, the agent should also maintain competence on previous target domains, i. e. to adapt without forgetting.

Meta-Learning Transfer Learning +1

Bi-tuning of Pre-trained Representations

no code implementations12 Nov 2020 Jincheng Zhong, Ximei Wang, Zhi Kou, Jianmin Wang, Mingsheng Long

It is common within the deep learning community to first pre-train a deep neural network from a large-scale dataset and then fine-tune the pre-trained model to a specific downstream task.

Contrastive Learning Unsupervised Pre-training

Unsupervised Transfer Learning for Spatiotemporal Predictive Networks

1 code implementation ICML 2020 Zhiyu Yao, Yunbo Wang, Mingsheng Long, Jian-Min Wang

This paper explores a new research problem of unsupervised transfer learning across multiple spatiotemporal prediction tasks.

Transfer Learning

On Localized Discrepancy for Domain Adaptation

no code implementations14 Aug 2020 Yuchen Zhang, Mingsheng Long, Jian-Min Wang, Michael. I. Jordan

Finally, we further extend the localized discrepancies for achieving super transfer and derive generalization bounds that could be even more sample-efficient on source domain.

Generalization Bounds Unsupervised Domain Adaptation

Transferable Calibration with Lower Bias and Variance in Domain Adaptation

no code implementations NeurIPS 2020 Ximei Wang, Mingsheng Long, Jian-Min Wang, Michael. I. Jordan

In this paper, we delve into the open problem of Calibration in DA, which is extremely challenging due to the coexistence of domain shift and the lack of target labels.

Decision Making Domain Adaptation

VideoDG: Generalizing Temporal Relations in Videos to Novel Domains

1 code implementation8 Dec 2019 Zhiyu Yao, Yunbo Wang, Jianmin Wang, Philip S. Yu, Mingsheng Long

This paper introduces video domain generalization where most video classification networks degenerate due to the lack of exposure to the target domains of divergent distributions.

Action Recognition Data Augmentation +5

Minimum Class Confusion for Versatile Domain Adaptation

3 code implementations ECCV 2020 Ying Jin, Ximei Wang, Mingsheng Long, Jian-Min Wang

It can be characterized as (1) a non-adversarial DA method without explicitly deploying domain alignment, enjoying faster convergence speed; (2) a versatile approach that can handle four existing scenarios: Closed-Set, Partial-Set, Multi-Source, and Multi-Target DA, outperforming the state-of-the-art methods in these scenarios, especially on one of the largest and hardest datasets to date (7. 3% on DomainNet).

Inductive Bias Multi-target Domain Adaptation +1

Catastrophic Forgetting Meets Negative Transfer: Batch Spectral Shrinkage for Safe Transfer Learning

2 code implementations NeurIPS 2019 Xinyang Chen, Sinan Wang, Bo Fu, Mingsheng Long, Jian-Min Wang

Before sufficient training data is available, fine-tuning neural networks pre-trained on large-scale datasets substantially outperforms training from random initialization.

Transfer Learning

Towards Understanding the Transferability of Deep Representations

no code implementations26 Sep 2019 Hong Liu, Mingsheng Long, Jian-Min Wang, Michael. I. Jordan

3) The feasibility of transferability is related to the similarity of both input and label.

How Does Learning Rate Decay Help Modern Neural Networks?

no code implementations ICLR 2020 Kaichao You, Mingsheng Long, Jian-Min Wang, Michael. I. Jordan

Despite the popularity of these common beliefs, experiments suggest that they are insufficient in explaining the general effectiveness of lrDecay in training modern neural networks that are deep, wide, and nonconvex.

Z-Order Recurrent Neural Networks for Video Prediction

no code implementations IEEE International Conference on Multimedia and Expo (ICME) 2019 Jianjin Zhang, Yunbo Wang, Mingsheng Long, Wang Jianmin, Philip S Yu

First, we propose a new RNN architecture for modeling the deterministic dynamics, which updates hidden states along a z-order curve to enhance the consistency of the features of mirrored layers.

 Ranked #1 on Video Prediction on KTH (Cond metric)

Video Prediction

Towards Accurate Model Selection in Deep Unsupervised Domain Adaptation

2 code implementations International Conference on Machine Learning 2019 Kaichao You, Ximei Wang, Mingsheng Long, Michael Jordan

Deep unsupervised domain adaptation (Deep UDA) methods successfully leverage rich labeled data in a source domain to boost the performance on related but unlabeled data in a target domain.

Model Selection Unsupervised Domain Adaptation

Transferability vs. Discriminability: Batch Spectral Penalization for Adversarial Domain Adaptation

2 code implementations International Conference on Machine Learning 2019 Xinyang Chen, Sinan Wang, Mingsheng Long, Jianmin Wang

In this paper, a series of experiments based on spectral analysis of the feature representations have been conducted, revealing an unexpected deterioration of the discriminability while learning transferable features adversarially.

Domain Adaptation Transfer Learning

Eidetic 3D LSTM: A Model for Video Prediction and Beyond

3 code implementations ICLR 2019 Yunbo Wang, Lu Jiang, Ming-Hsuan Yang, Li-Jia Li, Mingsheng Long, Li Fei-Fei

We first evaluate the E3D-LSTM network on widely-used future video prediction datasets and achieve the state-of-the-art performance.

 Ranked #1 on Video Prediction on KTH (Cond metric)

Activity Recognition Video Prediction +1

Bridging Theory and Algorithm for Domain Adaptation

5 code implementations11 Apr 2019 Yuchen Zhang, Tianle Liu, Mingsheng Long, Michael. I. Jordan

We introduce Margin Disparity Discrepancy, a novel measurement with rigorous generalization bounds, tailored to the distribution comparison with the asymmetric margin loss, and to the minimax optimization for easier training.

Domain Adaptation Generalization Bounds

Learning to Transfer Examples for Partial Domain Adaptation

1 code implementation CVPR 2019 Zhangjie Cao, Kaichao You, Mingsheng Long, Jian-Min Wang, Qiang Yang

Under the condition that target labels are unknown, the key challenge of PDA is how to transfer relevant examples in the shared classes to promote positive transfer, and ignore irrelevant ones in the specific classes to mitigate negative transfer.

Partial Domain Adaptation Transfer Learning

Deep Triplet Quantization

1 code implementation1 Feb 2019 Bin Liu, Yue Cao, Mingsheng Long, Jian-Min Wang, Jingdong Wang

We propose Deep Triplet Quantization (DTQ), a novel approach to learning deep quantization models from the similarity triplets.

Deep Hashing Image Retrieval +1

Memory In Memory: A Predictive Neural Network for Learning Higher-Order Non-Stationarity from Spatiotemporal Dynamics

4 code implementations CVPR 2019 Yunbo Wang, Jianjin Zhang, Hongyu Zhu, Mingsheng Long, Jian-Min Wang, Philip S. Yu

Natural spatiotemporal processes can be highly non-stationary in many ways, e. g. the low-level non-stationarity such as spatial correlations or temporal dependencies of local pixel values; and the high-level variations such as the accumulation, deformation or dissipation of radar echoes in precipitation forecasting.

Precipitation Forecasting Time Series Forecasting +1

Deep Priority Hashing

1 code implementation4 Sep 2018 Zhangjie Cao, Ziping Sun, Mingsheng Long, Jian-Min Wang, Philip S. Yu

Deep hashing enables image retrieval by end-to-end learning of deep representations and hash codes from training data with pairwise similarity information.

Deep Hashing Image Retrieval +1

Multi-Adversarial Domain Adaptation

4 code implementations4 Sep 2018 Zhongyi Pei, Zhangjie Cao, Mingsheng Long, Jian-Min Wang

Recent advances in deep domain adaptation reveal that adversarial learning can be embedded into deep networks to learn transferable features that reduce distribution discrepancy between the source and target domains.

Domain Adaptation

Cross-Modal Hamming Hashing

no code implementations ECCV 2018 Yue Cao , Bin Liu, Mingsheng Long, Jian-Min Wang

Extensive experiments demonstrate that CMHH can generate highly concentrated hash codes and achieve state-of-the-art cross-modal retrieval performance for both hash lookups and linear scan scenarios on three benchmark datasets, NUS-WIDE, MIRFlickr-25K, and IAPR TC-12.

Cross-Modal Retrieval Retrieval

Partial Adversarial Domain Adaptation

2 code implementations ECCV 2018 Zhangjie Cao, Lijia Ma, Mingsheng Long, Jian-Min Wang

We present Partial Adversarial Domain Adaptation (PADA), which simultaneously alleviates negative transfer by down-weighing the data of outlier source classes for training both source classifier and domain adversary, and promotes positive transfer by matching the feature distributions in the shared label space.

Partial Domain Adaptation

HashGAN: Deep Learning to Hash With Pair Conditional Wasserstein GAN

no code implementations CVPR 2018 Yue Cao, Bin Liu, Mingsheng Long, Jian-Min Wang

The main idea is to augment the training data with nearly real images synthesized from a new Pair Conditional Wasserstein GAN (PC-WGAN) conditioned on the pairwise similarity information.

Image Retrieval Representation Learning +1

Deep Cauchy Hashing for Hamming Space Retrieval

no code implementations CVPR 2018 Yue Cao, Mingsheng Long, Bin Liu, Jian-Min Wang

Due to its computation efficiency and retrieval quality, hashing has been widely applied to approximate nearest neighbor search for large-scale image retrieval, while deep hashing further improves the retrieval quality by end-to-end representation learning and hash coding.

Deep Hashing Image Retrieval +1

Transfer Adversarial Hashing for Hamming Space Retrieval

no code implementations13 Dec 2017 Zhangjie Cao, Mingsheng Long, Chao Huang, Jian-Min Wang

Existing work on deep hashing assumes that the database in the target domain is identically distributed with the training set in the source domain.

Deep Hashing Image Retrieval

PredRNN: Recurrent Neural Networks for Predictive Learning using Spatiotemporal LSTMs

no code implementations NeurIPS 2017 Yunbo Wang, Mingsheng Long, Jian-Min Wang, Zhifeng Gao, Philip S. Yu

The core of this network is a new Spatiotemporal LSTM (ST-LSTM) unit that extracts and memorizes spatial and temporal representations simultaneously.

Video Prediction

Deep Visual-Semantic Quantization for Efficient Image Retrieval

no code implementations CVPR 2017 Yue Cao, Mingsheng Long, Jian-Min Wang, Shichen Liu

This paper presents a compact coding solution with a focus on the deep learning to quantization approach, which improves retrieval quality by end-to-end representation learning and compact encoding and has already shown the superior performance over the hashing solutions for similarity retrieval.

Image Retrieval Quantization +2

Conditional Adversarial Domain Adaptation

5 code implementations NeurIPS 2018 Mingsheng Long, Zhangjie Cao, Jian-Min Wang, Michael. I. Jordan

Adversarial learning has been embedded into deep networks to learn disentangled and transferable representations for domain adaptation.

Domain Adaptation General Classification

HashNet: Deep Learning to Hash by Continuation

2 code implementations ICCV 2017 Zhangjie Cao, Mingsheng Long, Jian-Min Wang, Philip S. Yu

Learning to hash has been widely applied to approximate nearest neighbor search for large-scale multimedia retrieval, due to its computation efficiency and retrieval quality.

Binarization Representation Learning +1

Transitive Hashing Network for Heterogeneous Multimedia Retrieval

no code implementations15 Aug 2016 Zhangjie Cao, Mingsheng Long, Qiang Yang

Hashing has been widely applied to large-scale multimedia retrieval due to the storage and retrieval efficiency.

Retrieval

Deep Transfer Learning with Joint Adaptation Networks

4 code implementations ICML 2017 Mingsheng Long, Han Zhu, Jian-Min Wang, Michael. I. Jordan

Deep networks have been successfully applied to learn transferable features for adapting models from a source domain to a different target domain.

Multi-Source Unsupervised Domain Adaptation Transfer Learning

Correlation Hashing Network for Efficient Cross-Modal Retrieval

no code implementations22 Feb 2016 Yue Cao, Mingsheng Long, Jian-Min Wang, Philip S. Yu

This paper presents a Correlation Hashing Network (CHN) approach to cross-modal hashing, which jointly learns good data representation tailored to hash coding and formally controls the quantization error.

Cross-Modal Retrieval Quantization +1

Unsupervised Domain Adaptation with Residual Transfer Networks

2 code implementations NeurIPS 2016 Mingsheng Long, Han Zhu, Jian-Min Wang, Michael. I. Jordan

In this paper, we propose a new approach to domain adaptation in deep networks that can jointly learn adaptive classifiers and transferable features from labeled data in the source domain and unlabeled data in the target domain.

Unsupervised Domain Adaptation

Learning Multiple Tasks with Multilinear Relationship Networks

no code implementations NeurIPS 2017 Mingsheng Long, Zhangjie Cao, Jian-Min Wang, Philip S. Yu

Deep networks trained on large-scale data can learn transferable features to promote learning multiple tasks.

Multi-Task Learning

Learning Transferable Features with Deep Adaptation Networks

5 code implementations10 Feb 2015 Mingsheng Long, Yue Cao, Jian-Min Wang, Michael. I. Jordan

Recent studies reveal that a deep neural network can learn transferable features which generalize well to novel tasks for domain adaptation.

Domain Adaptation Image Classification +1

Transfer Joint Matching for Unsupervised Domain Adaptation

no code implementations CVPR 2014 Mingsheng Long, Jian-Min Wang, Guiguang Ding, Jiaguang Sun, Philip S. Yu

Visual domain adaptation, which learns an accurate classifier for a new domain using labeled images from an old domain, has shown promising value in computer vision yet still been a challenging problem.

Dimensionality Reduction Unsupervised Domain Adaptation

Transfer Sparse Coding for Robust Image Representation

no code implementations CVPR 2013 Mingsheng Long, Guiguang Ding, Jian-Min Wang, Jiaguang Sun, Yuchen Guo, Philip S. Yu

In this paper, we propose a Transfer Sparse Coding (TSC) approach to construct robust sparse representations for classifying cross-distribution images accurately.

Cannot find the paper you are looking for? You can Submit a new open access paper.