no code implementations • 7 Jun 2023 • Jingyi Cui, Weiran Huang, Yifei Wang, Yisen Wang
Therefore, to explore the mechanical differences between semi-supervised and noisy-labeled information in helping contrastive learning, we establish a unified theoretical framework of contrastive learning under weak supervision.
no code implementations • 10 Jun 2021 • Jingyi Cui, Hanyuan Hang, Yisen Wang, Zhouchen Lin
In this paper, we propose a density estimation algorithm called \textit{Gradient Boosting Histogram Transform} (GBHT), where we adopt the \textit{Negative Log Likelihood} as the loss function to make the boosting procedure available for the unsupervised tasks.
1 code implementation • 10 Jun 2021 • Hongwei Wen, Jingyi Cui, Hanyuan Hang, Jiabin Liu, Yisen Wang, Zhouchen Lin
As an important branch of weakly supervised learning, partial label learning deals with data where each instance is assigned with a set of candidate labels, whereas only one of them is true.