2 code implementations • 24 Jun 2023 • Yanai Elazar, Jiayao Zhang, David Wadden, Bo Zhang, Noah A. Smith
However, since quality is a challenging construct to estimate, we use the negative outcome control method, using paper citation count as a control variable to debias the quality confounding effect.
1 code implementation • 9 May 2023 • Zhaowei Wang, Quyet V. Do, Hongming Zhang, Jiayao Zhang, Weiqi Wang, Tianqing Fang, Yangqiu Song, Ginny Y. Wong, Simon See
This paper proposes a new task to detect commonsense causation between two events in an event sequence (i. e., context), called contextualized commonsense causal reasoning.
1 code implementation • 7 Nov 2022 • Jiayao Zhang, Hongming Zhang, Zhun Deng, Dan Roth
We distill several insights from our analysis on study the peer review process with the help of large LMs.
no code implementations • 19 Oct 2022 • Chi-Han Peng, Jiayao Zhang
We propose a novel approach to compute high-resolution (2048x1024 and higher) depths for panoramas that is significantly faster and qualitatively and qualitatively more accurate than the current state-of-the-art method (360MonoDepth).
no code implementations • 6 Jun 2022 • Zhun Deng, Jiayao Zhang, Linjun Zhang, Ting Ye, Yates Coley, Weijie J. Su, James Zou
Specifically, FIFA encourages both classification and fairness generalization and can be flexibly combined with many existing fair learning methods with logits-based losses.
no code implementations • 2 Feb 2022 • Bo Zhang, Jiayao Zhang
We examine the role of textual data as study units when conducting causal inference by drawing parallels between human subjects and organized texts.
1 code implementation • 31 Jan 2022 • Jiayao Zhang, Hongming Zhang, Weijie J. Su, Dan Roth
Commonsense causality reasoning (CCR) aims at identifying plausible causes and effects in natural language descriptions that are deemed reasonable by an average person.
1 code implementation • NeurIPS 2021 • Jiayao Zhang, Hua Wang, Weijie J. Su
Our main finding uncovers a sharp phase transition phenomenon regarding the {intra-class impact: if the SDEs are locally elastic in the sense that the impact is more significant on samples from the same class as the input, the features of the training data become linearly separable, meaning vanishing training loss; otherwise, the features are not separable, regardless of how long the training time is.
1 code implementation • 26 Apr 2021 • Stanislava Fedorova, Alberto Tono, Meher Shashwat Nigam, Jiayao Zhang, Amirhossein Ahmadnia, Cecilia Bolognesi, Dominik L. Michels
The variety of annotations, the flexibility to customize the generated building and dataset parameters make this framework suitable for multiple deep learning tasks, including geometric deep learning that requires direct 3D supervision.
1 code implementation • 7 Aug 2018 • Jiayao Zhang, Guangxu Zhu, Robert W. Heath Jr., Kaibin Huang
We hope to inspire practitioners in different fields to adopt the powerful tool of Grassmannian learning in their research.