1 code implementation • 12 Feb 2024 • Steven Jecmen, Nihar B. Shah, Fei Fang, Leman Akoglu
A major threat to the peer-review systems of computer science conferences is the existence of "collusion rings" between reviewers.
no code implementations • 1 Jun 2023 • Ryan Liu, Nihar B. Shah
We find that across 119 {checklist question, paper} pairs, the LLM had an 86. 6% accuracy.
1 code implementation • NeurIPS 2023 • Martin Saveski, Steven Jecmen, Nihar B. Shah, Johan Ugander
We consider estimates of (i) the effect on review quality when changing weights in the assignment algorithm, e. g., weighting reviewers' bids vs. textual similarity (between the review's past papers and the submission), and (ii) the "cost of randomization", capturing the difference in expected quality between the perturbed and unperturbed optimal match.
2 code implementations • 23 Mar 2023 • Ivan Stelmakh, John Wieting, Graham Neubig, Nihar B. Shah
We address this challenge by collecting a novel dataset of similarity scores that we release to the research community.
1 code implementation • 16 Feb 2023 • Joon Sik Kim, Valerie Chen, Danish Pruthi, Nihar B. Shah, Ameet Talwalkar
Many practical applications, ranging from paper-reviewer assignment in peer review to job-applicant matching for hiring, require human decision makers to identify relevant matches by combining their expertise with predictions from machine learning models.
no code implementations • 22 Nov 2022 • Charvi Rastogi, Ivan Stelmakh, Alina Beygelzimer, Yann N. Dauphin, Percy Liang, Jennifer Wortman Vaughan, Zhenyu Xue, Hal Daumé III, Emma Pierson, Nihar B. Shah
In a top-tier computer science conference (NeurIPS 2021) with more than 23, 000 submitting authors and 9, 000 submitted papers, we survey the authors on three questions: (i) their predicted probability of acceptance for each of their papers, (ii) their perceived ranking of their own papers based on scientific contribution, and (iii) the change in their perception about their own papers after seeing the reviews.
1 code implementation • 18 Sep 2022 • Jingyan Wang, Carmel Baharav, Nihar B. Shah, Anita Williams Woolley, R Ravi
Specifically, in the often-used holistic allocation, each evaluator is assigned a subset of the applicants, and is asked to assess all relevant information for their assigned applicants.
no code implementations • 22 Jul 2022 • Steven Jecmen, Nihar B. Shah, Fei Fang, Vincent Conitzer
Many conferences rely on paper bidding as a key component of their reviewer assignment procedure.
1 code implementation • 24 Jun 2022 • Steven Jecmen, Minji Yoon, Vincent Conitzer, Nihar B. Shah, Fei Fang
The performance of these detection algorithms can be taken as a baseline for future research on detecting malicious bidding.
1 code implementation • 5 Apr 2022 • Yusha Liu, Yichong Xu, Nihar B. Shah, Aarti Singh
Our approach addresses the two aforementioned challenges by: (i) ensuring that rankings are incorporated into the updates scores in the same manner for all papers, thereby mitigating arbitrariness, and (ii) allowing to seamlessly use existing interfaces and workflows designed for scores.
1 code implementation • 25 Jan 2022 • Komal Dhull, Steven Jecmen, Pravesh Kothari, Nihar B. Shah
Finally, we evaluate the methods on a dataset from conference peer review.
1 code implementation • 13 Aug 2021 • Steven Jecmen, Hanrui Zhang, Ryan Liu, Fei Fang, Vincent Conitzer, Nihar B. Shah
Many scientific conferences employ a two-phase paper review process, where some papers are assigned additional reviewers after the initial reviews are submitted.
1 code implementation • 1 Dec 2020 • Jingyan Wang, Ivan Stelmakh, Yuting Wei, Nihar B. Shah
For example, universities ask students to rate the teaching quality of their instructors, and conference organizers ask authors of submissions to evaluate the quality of the reviews.
no code implementations • 30 Nov 2020 • Ivan Stelmakh, Nihar B. Shah, Aarti Singh, Hal Daumé III
Conference peer review constitutes a human-computation process whose importance cannot be overstated: not only it identifies the best submissions for acceptance, but, ultimately, it impacts the future of the whole research area by promoting some ideas and restraining others.
no code implementations • 30 Nov 2020 • Ivan Stelmakh, Nihar B. Shah, Aarti Singh, Hal Daumé III
Modern machine learning and computer science conferences are experiencing a surge in the number of submissions that challenges the quality of peer review as the number of competent reviewers is growing at a much slower rate.
no code implementations • 30 Nov 2020 • Ivan Stelmakh, Charvi Rastogi, Nihar B. Shah, Aarti Singh, Hal Daumé III
Peer review is the backbone of academia and humans constitute a cornerstone of this process, being responsible for reviewing papers and making the final acceptance/rejection decisions.
no code implementations • 29 Oct 2020 • Emaad Manzoor, Nihar B. Shah
In this work, we introduce a novel framework to quantify bias in text caused by the visibility of subgroup membership indicators.
no code implementations • 8 Oct 2020 • Ivan Stelmakh, Nihar B. Shah, Aarti Singh
We consider the issue of strategic behaviour in various peer-assessment tasks, including peer grading of exams or homeworks and peer review in hiring or promotions.
2 code implementations • NeurIPS 2020 • Steven Jecmen, Hanrui Zhang, Ryan Liu, Nihar B. Shah, Vincent Conitzer, Fei Fang
We further consider the problem of restricting the joint probability that certain suspect pairs of reviewers are assigned to certain papers, and show that this problem is NP-hard for arbitrary constraints on these joint probabilities but efficiently solvable for a practical special case.
no code implementations • 29 Jun 2020 • Wenxin Ding, Nihar B. Shah, Weina Wang
The crux of the framework lies in recognizing that a part of the data pertaining to the reviews is already available in public, and we use this information to post-process the data released by any privacy mechanism in a manner that improves the accuracy (utility) of the data while retaining the privacy guarantees.
1 code implementation • 27 Jun 2020 • Tanner Fiez, Nihar B. Shah, Lillian Ratliff
Theoretically, we show a local optimality guarantee of our algorithm and prove that popular baselines are considerably suboptimal.
no code implementations • 21 Jun 2020 • Charvi Rastogi, Sivaraman Balakrishnan, Nihar B. Shah, Aarti Singh
We also provide testing algorithms and associated sample complexity bounds for the problem of two-sample testing with partial (or total) ranking data. Furthermore, we empirically evaluate our results via extensive simulations as well as two real-world datasets consisting of pairwise comparisons.
no code implementations • 10 Jun 2019 • Jing-Yan Wang, Nihar B. Shah, R. Ravi
We show that the MLE incurs a suboptimal rate in terms of bias.
no code implementations • 27 Aug 2018 • Ritesh Noothigattu, Nihar B. Shah, Ariel D. Procaccia
The key challenge that arises is the specification of a loss function for ERM.
no code implementations • 16 Jun 2018 • Ivan Stelmakh, Nihar B. Shah, Aarti Singh
Our fairness objective is to maximize the review quality of the most disadvantaged paper, in contrast to the commonly used objective of maximizing the total quality over all papers.
1 code implementation • 16 Jun 2018 • Yichong Xu, Han Zhao, Xiaofei Shi, Jeremy Zhang, Nihar B. Shah
We then empirically show that the requisite property on the authorship graph is indeed satisfied in the submission data from the ICLR conference, and further demonstrate a simple trick to make the partitioning method more practically appealing for conference peer review.
no code implementations • 13 Jun 2018 • Jing-Yan Wang, Nihar B. Shah
A popular approach to address this issue is to assume simplistic models of miscalibration (such as linear biases) to de-bias the scores.
no code implementations • 1 Sep 2017 • Nihar B. Shah, Sivaraman Balakrishnan, Martin J. Wainwright
We consider the problem of noisy matrix completion, in which the goal is to reconstruct a structured matrix whose entries are partially observed in noise.
1 code implementation • 31 Aug 2017 • Nihar B. Shah, Behzad Tabibian, Krikamol Muandet, Isabelle Guyon, Ulrike Von Luxburg
Neural Information Processing Systems (NIPS) is a top-tier annual conference in machine learning.
no code implementations • 30 Jun 2016 • Nihar B. Shah, Sivaraman Balakrishnan, Martin J. Wainwright
The task of aggregating and denoising crowd-labeled data has gained increased significance with the advent of crowdsourcing platforms and massive datasets.
no code implementations • 28 Jun 2016 • Reinhard Heckel, Nihar B. Shah, Kannan Ramchandran, Martin J. Wainwright
We first analyze a sequential ranking algorithm that counts the number of comparisons won, and uses these counts to decide whether to stop, or to compare another pair of items, chosen based on confidence intervals specified by the data collected up to that point.
no code implementations • 22 Mar 2016 • Nihar B. Shah, Sivaraman Balakrishnan, Martin J. Wainwright
Second, we show that a regularized least squares estimator can achieve a poly-logarithmic adaptivity index, thereby demonstrating a $\sqrt{n}$-gap between optimal and computationally achievable adaptivity.
no code implementations • 24 Feb 2016 • Yuan Luo, Nihar B. Shah, Jianwei Huang, Jean Walrand
In order to elicit heterogeneous agents' private information and incentivize agents with different capabilities to act in the principal's best interest, we design an optimal joint incentive mechanism and prediction algorithm called COPE (COst and Prediction Elicitation), the analysis of which offers several valuable engineering insights.
no code implementations • 30 Dec 2015 • Nihar B. Shah, Martin J. Wainwright
We consider data in the form of pairwise comparisons of n items, with the goal of precisely identifying the top k items for some value of k < n, or alternatively, recovering a ranking of all the items.
no code implementations • 19 Oct 2015 • Nihar B. Shah, Sivaraman Balakrishnan, Adityanand Guntuboyina, Martin J. Wainwright
On the other hand, unlike in the BTL and Thurstone models, computing the minimax-optimal estimator in the stochastically transitive model is non-trivial, and we explore various computationally tractable alternatives.
no code implementations • 6 May 2015 • Nihar B. Shah, Sivaraman Balakrishnan, Joseph Bradley, Abhay Parekh, Kannan Ramchandran, Martin J. Wainwright
Data in the form of pairwise comparisons arises in many domains, including preference elicitation, sporting competitions, and peer grading among others.
no code implementations • 25 Mar 2015 • Dengyong Zhou, Qiang Liu, John C. Platt, Christopher Meek, Nihar B. Shah
There is a rapidly increasing interest in crowdsourcing for data labeling.
no code implementations • 19 Feb 2015 • Nihar B. Shah, Dengyong Zhou, Yuval Peres
The growing need for labeled training data has made crowdsourcing an important part of machine learning.
no code implementations • 21 Nov 2014 • Nihar B. Shah, Dengyong Zhou
Human computation or crowdsourcing involves joint inference of the ground-truth-answers and the worker-abilities by optimizing an objective function, for instance, by maximizing the data likelihood based on an assumed underlying model.
no code implementations • NeurIPS 2015 • Nihar B. Shah, Dengyong Zhou
To address this fundamental challenge in crowdsourcing, we propose a simple payment mechanism to incentivize workers to answer only the questions that they are sure of and skip the rest.
no code implementations • 25 Jun 2014 • Nihar B. Shah, Sivaraman Balakrishnan, Joseph Bradley, Abhay Parekh, Kannan Ramchandran, Martin Wainwright
When eliciting judgements from humans for an unknown quantity, one often has the choice of making direct-scoring (cardinal) or comparative (ordinal) measurements.