no code implementations • 15 Dec 2023 • Ziliang Chen, Yongsen Zheng, Zhao-Rong Lai, Quanlong Guan, Liang Lin
Invariant representation learning (IRL) encourages the prediction from invariant causal features to labels de-confounded from the environments, advancing the technical roadmap of out-of-distribution (OOD) generalization.
no code implementations • ICCV 2023 • BinBin Yang, Yi Luo, Ziliang Chen, Guangrun Wang, Xiaodan Liang, Liang Lin
Thanks to the rapid development of diffusion models, unprecedented progress has been witnessed in image synthesis.
no code implementations • ICCV 2023 • Ziliang Chen, Xin Huang, Quanlong Guan, Liang Lin, Weiqi Luo
The vision community is undergoing the unprecedented progress with the emergence of Vision-Language Pretraining Models (VLMs).
no code implementations • 7 Mar 2022 • Jingyu Zhuang, Ziliang Chen, Pengxu Wei, Guanbin Li, Liang Lin
In Open Set Domain Adaptation (OSDA), large amounts of target samples are drawn from the implicit categories that never appear in the source domain.
no code implementations • 1 Jan 2021 • Junfan Lin, Lin Xu, Ziliang Chen, Liang Lin
To this end, we propose a novel DSMAD agent, INS-DS (Introspective Diagnosis System) comprising of two separate yet cooperative modules, i. e., an inquiry module for proposing symptom-inquiries and an introspective module for deciding when to inform a disease.
1 code implementation • 22 Dec 2020 • Shuai Lin, Pan Zhou, Xiaodan Liang, Jianheng Tang, Ruihui Zhao, Ziliang Chen, Liang Lin
Besides, we develop a Graph-Evolving Meta-Learning (GEML) framework that learns to evolve the commonsense graph for reasoning disease-symptom correlations in a new disease, which effectively alleviates the needs of a large number of dialogues.
1 code implementation • 14 Mar 2020 • Junfan Lin, Keze Wang, Ziliang Chen, Xiaodan Liang, Liang Lin
To eliminate this bias and inspired by the propensity score matching technique with causal diagram, we propose a propensity-based patient simulator to effectively answer unrecorded inquiry by drawing knowledge from the other records; Bias (ii) inherently comes along with the passively collected data, and is one of the key obstacles for training the agent towards "learning how" rather than "remembering what".
no code implementations • 28 Sep 2019 • Xiaopeng Yan, Ziliang Chen, Anni Xu, Xiaoxi Wang, Xiaodan Liang, Liang Lin
Resembling the rapid learning capability of human, few-shot learning empowers vision systems to understand new concepts by training with few samples.
Ranked #19 on Few-Shot Object Detection on MS-COCO (30-shot)
no code implementations • 25 Sep 2019 • Ziliang Chen, Zhanfu Yang
It is feasible and practically-valuable to bridge the characteristics between graph neural networks (GNNs) and logical reasoning.
no code implementations • 25 Sep 2019 • Fei Wang, Zhanfu Yang, Ziliang Chen, Guannan Wei, Tiark Rompf
In this paper, we target the QBF (Quantified Boolean Formula) satisfiability problem, the complexity of which is in-between propositional logic and predicate logic, and investigate the feasibility of learning GNN-based solvers and GNN-based heuristics for the cases with a universal-existential quantifier alternation (so-called 2QBF problems).
1 code implementation • CVPR 2019 • Ziliang Chen, Jingyu Zhuang, Xiaodan Liang, Liang Lin
(Unsupervised) Domain Adaptation (DA) seeks for classifying target instances when solely provided with source labeled and target unlabeled examples for training.
Ranked #3 on Multi-target Domain Adaptation on Office-Home
1 code implementation • 8 Jul 2019 • Ziliang Chen, Zhanfu Yang, Xiaoxi Wang, Xiaodan Liang, Xiaopeng Yan, Guanbin Li, Liang Lin
A broad range of cross-$m$-domain generation researches boil down to matching a joint distribution by deep generative models (DGMs).
no code implementations • 6 May 2019 • Xiao Wang, Ziliang Chen, Rui Yang, Bin Luo, Jin Tang
In this paper, we propose Hard Person Identity Mining (HPIM) that attempts to refine the hard example mining to improve the exploration efficacy in person re-identification.
no code implementations • 27 Apr 2019 • Zhanfu Yang, Fei Wang, Ziliang Chen, Guannan Wei, Tiark Rompf
In this paper, we investigate the feasibility of learning GNN (Graph Neural Network) based solvers and GNN-based heuristics for specified QBF (Quantified Boolean Formula) problems.
no code implementations • 4 Dec 2018 • Xu Cai, Yang Wu, Guanbin Li, Ziliang Chen, Liang Lin
FRAME (Filters, Random fields, And Maximum Entropy) is an energy-based descriptive model that synthesizes visual realism by capturing mutual patterns from structural input signals.
no code implementations • 27 Sep 2018 • Ziliang Chen, Keze Wang, Liang Lin
We evaluate T2T across different learners, teachers, and tasks, which significantly demonstrates that structured knowledge can be inherited by the teachers to further benefit learners' training.
1 code implementation • 30 Jun 2018 • Keze Wang, Liang Lin, Xiaopeng Yan, Ziliang Chen, Dongyu Zhang, Lei Zhang
The proposed process can be compatible with mini-batch based training (i. e., using a batch of unlabeled or partially labeled data as a one-time input) for object detection.
1 code implementation • CVPR 2018 • Ruijia Xu, Ziliang Chen, WangMeng Zuo, Junjie Yan, Liang Lin
Motivated by the theoretical results in \cite{mansour2009domain}, the target distribution can be represented as the weighted combination of source distributions, and, the multi-source unsupervised domain adaptation via DCTN is then performed as two alternating steps: i) It deploys multi-way adversarial learning to minimize the discrepancy between the target and each of the multiple source domains, which also obtains the source-specific perplexity scores to denote the possibilities that a target sample belongs to different source domains.
Multi-Source Unsupervised Domain Adaptation Unsupervised Domain Adaptation
no code implementations • 28 Jul 2017 • Ziliang Chen, Keze Wang, Xiao Wang, Pai Peng, Ebroul Izquierdo, Liang Lin
Aiming at improving performance of visual classification in a cost-effective manner, this paper proposes an incremental semi-supervised learning paradigm called Deep Co-Space (DCS).