no code implementations • 26 Oct 2022 • Carey E. Priebe, Ningyuan Huang, Soledad Villar, Cong Mu, Li Chen
We conjecture that for general label noise, mitigation strategies that make use of the noisy data will outperform those that ignore the noisy data.
no code implementations • 29 Aug 2022 • Cong Mu, Youngser Park, Carey E. Priebe
We propose a dynamic network sampling scheme to optimize block recovery for stochastic blockmodel (SBM) in the case where it is prohibitively expensive to observe the entire graph.
no code implementations • 28 May 2022 • Li Chen, Ningyuan Huang, Cong Mu, Hayden S. Helm, Kate Lytvynets, Weiwei Yang, Carey E. Priebe
Our hierarchical approach improves upon regular deep neural networks in learning with label noise.
1 code implementation • 4 Jul 2020 • Cong Mu, Angelo Mele, Lingxin Hao, Joshua Cape, Avanti Athreya, Carey E. Priebe
In network inference applications, it is often desirable to detect community structure, namely to cluster vertices into groups, or blocks, according to some measure of similarity.