Search Results for author: Nihar B. Shah

Found 41 papers, 14 papers with code

On the Detection of Reviewer-Author Collusion Rings From Paper Bidding

1 code implementation12 Feb 2024 Steven Jecmen, Nihar B. Shah, Fei Fang, Leman Akoglu

A major threat to the peer-review systems of computer science conferences is the existence of "collusion rings" between reviewers.

Fraud Detection text similarity

ReviewerGPT? An Exploratory Study on Using Large Language Models for Paper Reviewing

no code implementations1 Jun 2023 Ryan Liu, Nihar B. Shah

We find that across 119 {checklist question, paper} pairs, the LLM had an 86. 6% accuracy.

Counterfactual Evaluation of Peer-Review Assignment Policies

1 code implementation NeurIPS 2023 Martin Saveski, Steven Jecmen, Nihar B. Shah, Johan Ugander

We consider estimates of (i) the effect on review quality when changing weights in the assignment algorithm, e. g., weighting reviewers' bids vs. textual similarity (between the review's past papers and the submission), and (ii) the "cost of randomization", capturing the difference in expected quality between the perturbed and unperturbed optimal match.

counterfactual Off-policy evaluation +1

A Gold Standard Dataset for the Reviewer Assignment Problem

2 code implementations23 Mar 2023 Ivan Stelmakh, John Wieting, Graham Neubig, Nihar B. Shah

We address this challenge by collecting a novel dataset of similarity scores that we release to the research community.

Assisting Human Decisions in Document Matching

1 code implementation16 Feb 2023 Joon Sik Kim, Valerie Chen, Danish Pruthi, Nihar B. Shah, Ameet Talwalkar

Many practical applications, ranging from paper-reviewer assignment in peer review to job-applicant matching for hiring, require human decision makers to identify relevant matches by combining their expertise with predictions from machine learning models.

How do Authors' Perceptions of their Papers Compare with Co-authors' Perceptions and Peer-review Decisions?

no code implementations22 Nov 2022 Charvi Rastogi, Ivan Stelmakh, Alina Beygelzimer, Yann N. Dauphin, Percy Liang, Jennifer Wortman Vaughan, Zhenyu Xue, Hal Daumé III, Emma Pierson, Nihar B. Shah

In a top-tier computer science conference (NeurIPS 2021) with more than 23, 000 submitting authors and 9, 000 submitted papers, we survey the authors on three questions: (i) their predicted probability of acceptance for each of their papers, (ii) their perceived ranking of their own papers based on scientific contribution, and (iii) the change in their perception about their own papers after seeing the reviews.

Allocation Schemes in Analytic Evaluation: Applicant-Centric Holistic or Attribute-Centric Segmented?

1 code implementation18 Sep 2022 Jingyan Wang, Carmel Baharav, Nihar B. Shah, Anita Williams Woolley, R Ravi

Specifically, in the often-used holistic allocation, each evaluator is assigned a subset of the applicants, and is asked to assess all relevant information for their assigned applicants.

Attribute

Tradeoffs in Preventing Manipulation in Paper Bidding for Reviewer Assignment

no code implementations22 Jul 2022 Steven Jecmen, Nihar B. Shah, Fei Fang, Vincent Conitzer

Many conferences rely on paper bidding as a key component of their reviewer assignment procedure.

A Dataset on Malicious Paper Bidding in Peer Review

1 code implementation24 Jun 2022 Steven Jecmen, Minji Yoon, Vincent Conitzer, Nihar B. Shah, Fei Fang

The performance of these detection algorithms can be taken as a baseline for future research on detecting malicious bidding.

Descriptive

Integrating Rankings into Quantized Scores in Peer Review

1 code implementation5 Apr 2022 Yusha Liu, Yichong Xu, Nihar B. Shah, Aarti Singh

Our approach addresses the two aforementioned challenges by: (i) ensuring that rankings are incorporated into the updates scores in the same manner for all papers, thereby mitigating arbitrariness, and (ii) allowing to seamlessly use existing interfaces and workflows designed for scores.

Decision Making

Near-Optimal Reviewer Splitting in Two-Phase Paper Reviewing and Conference Experiment Design

1 code implementation13 Aug 2021 Steven Jecmen, Hanrui Zhang, Ryan Liu, Fei Fang, Vincent Conitzer, Nihar B. Shah

Many scientific conferences employ a two-phase paper review process, where some papers are assigned additional reviewers after the initial reviews are submitted.

Debiasing Evaluations That are Biased by Evaluations

1 code implementation1 Dec 2020 Jingyan Wang, Ivan Stelmakh, Yuting Wei, Nihar B. Shah

For example, universities ask students to rate the teaching quality of their instructors, and conference organizers ask authors of submissions to evaluate the quality of the reviews.

A Novice-Reviewer Experiment to Address Scarcity of Qualified Reviewers in Large Conferences

no code implementations30 Nov 2020 Ivan Stelmakh, Nihar B. Shah, Aarti Singh, Hal Daumé III

Conference peer review constitutes a human-computation process whose importance cannot be overstated: not only it identifies the best submissions for acceptance, but, ultimately, it impacts the future of the whole research area by promoting some ideas and restraining others.

Prior and Prejudice: The Novice Reviewers' Bias against Resubmissions in Conference Peer Review

no code implementations30 Nov 2020 Ivan Stelmakh, Nihar B. Shah, Aarti Singh, Hal Daumé III

Modern machine learning and computer science conferences are experiencing a surge in the number of submissions that challenges the quality of peer review as the number of competent reviewers is growing at a much slower rate.

BIG-bench Machine Learning

A Large Scale Randomized Controlled Trial on Herding in Peer-Review Discussions

no code implementations30 Nov 2020 Ivan Stelmakh, Charvi Rastogi, Nihar B. Shah, Aarti Singh, Hal Daumé III

Peer review is the backbone of academia and humans constitute a cornerstone of this process, being responsible for reviewing papers and making the final acceptance/rejection decisions.

Decision Making

Uncovering Latent Biases in Text: Method and Application to Peer Review

no code implementations29 Oct 2020 Emaad Manzoor, Nihar B. Shah

In this work, we introduce a novel framework to quantify bias in text caused by the visibility of subgroup membership indicators.

Catch Me if I Can: Detecting Strategic Behaviour in Peer Assessment

no code implementations8 Oct 2020 Ivan Stelmakh, Nihar B. Shah, Aarti Singh

We consider the issue of strategic behaviour in various peer-assessment tasks, including peer grading of exams or homeworks and peer review in hiring or promotions.

Mitigating Manipulation in Peer Review via Randomized Reviewer Assignments

2 code implementations NeurIPS 2020 Steven Jecmen, Hanrui Zhang, Ryan Liu, Nihar B. Shah, Vincent Conitzer, Fei Fang

We further consider the problem of restricting the joint probability that certain suspect pairs of reviewers are assigned to certain papers, and show that this problem is NP-hard for arbitrary constraints on these joint probabilities but efficiently solvable for a practical special case.

On the Privacy-Utility Tradeoff in Peer-Review Data Analysis

no code implementations29 Jun 2020 Wenxin Ding, Nihar B. Shah, Weina Wang

The crux of the framework lies in recognizing that a part of the data pertaining to the reviews is already available in public, and we use this information to post-process the data released by any privacy mechanism in a manner that improves the accuracy (utility) of the data while retaining the privacy guarantees.

Privacy Preserving

A SUPER* Algorithm to Optimize Paper Bidding in Peer Review

1 code implementation27 Jun 2020 Tanner Fiez, Nihar B. Shah, Lillian Ratliff

Theoretically, we show a local optimality guarantee of our algorithm and prove that popular baselines are considerably suboptimal.

Two-Sample Testing on Ranked Preference Data and the Role of Modeling Assumptions

no code implementations21 Jun 2020 Charvi Rastogi, Sivaraman Balakrishnan, Nihar B. Shah, Aarti Singh

We also provide testing algorithms and associated sample complexity bounds for the problem of two-sample testing with partial (or total) ranking data. Furthermore, we empirically evaluate our results via extensive simulations as well as two real-world datasets consisting of pairwise comparisons.

Two-sample testing

Loss Functions, Axioms, and Peer Review

no code implementations27 Aug 2018 Ritesh Noothigattu, Nihar B. Shah, Ariel D. Procaccia

The key challenge that arises is the specification of a loss function for ERM.

PeerReview4All: Fair and Accurate Reviewer Assignment in Peer Review

no code implementations16 Jun 2018 Ivan Stelmakh, Nihar B. Shah, Aarti Singh

Our fairness objective is to maximize the review quality of the most disadvantaged paper, in contrast to the commonly used objective of maximizing the total quality over all papers.

Fairness

On Strategyproof Conference Peer Review

1 code implementation16 Jun 2018 Yichong Xu, Han Zhao, Xiaofei Shi, Jeremy Zhang, Nihar B. Shah

We then empirically show that the requisite property on the authorship graph is indeed satisfied in the submission data from the ICLR conference, and further demonstrate a simple trick to make the partitioning method more practically appealing for conference peer review.

Your 2 is My 1, Your 3 is My 9: Handling Arbitrary Miscalibrations in Ratings

no code implementations13 Jun 2018 Jing-Yan Wang, Nihar B. Shah

A popular approach to address this issue is to assume simplistic models of miscalibration (such as linear biases) to de-bias the scores.

Low Permutation-rank Matrices: Structural Properties and Noisy Completion

no code implementations1 Sep 2017 Nihar B. Shah, Sivaraman Balakrishnan, Martin J. Wainwright

We consider the problem of noisy matrix completion, in which the goal is to reconstruct a structured matrix whose entries are partially observed in noise.

Matrix Completion

Design and Analysis of the NIPS 2016 Review Process

1 code implementation31 Aug 2017 Nihar B. Shah, Behzad Tabibian, Krikamol Muandet, Isabelle Guyon, Ulrike Von Luxburg

Neural Information Processing Systems (NIPS) is a top-tier annual conference in machine learning.

A Permutation-based Model for Crowd Labeling: Optimal Estimation and Robustness

no code implementations30 Jun 2016 Nihar B. Shah, Sivaraman Balakrishnan, Martin J. Wainwright

The task of aggregating and denoising crowd-labeled data has gained increased significance with the advent of crowdsourcing platforms and massive datasets.

Denoising

Active Ranking from Pairwise Comparisons and when Parametric Assumptions Don't Help

no code implementations28 Jun 2016 Reinhard Heckel, Nihar B. Shah, Kannan Ramchandran, Martin J. Wainwright

We first analyze a sequential ranking algorithm that counts the number of comparisons won, and uses these counts to decide whether to stop, or to compare another pair of items, chosen based on confidence intervals specified by the data collected up to that point.

Open-Ended Question Answering

Feeling the Bern: Adaptive Estimators for Bernoulli Probabilities of Pairwise Comparisons

no code implementations22 Mar 2016 Nihar B. Shah, Sivaraman Balakrishnan, Martin J. Wainwright

Second, we show that a regularized least squares estimator can achieve a poly-logarithmic adaptivity index, thereby demonstrating a $\sqrt{n}$-gap between optimal and computationally achievable adaptivity.

Parametric Prediction from Parametric Agents

no code implementations24 Feb 2016 Yuan Luo, Nihar B. Shah, Jianwei Huang, Jean Walrand

In order to elicit heterogeneous agents' private information and incentivize agents with different capabilities to act in the principal's best interest, we design an optimal joint incentive mechanism and prediction algorithm called COPE (COst and Prediction Elicitation), the analysis of which offers several valuable engineering insights.

Learning Theory

Simple, Robust and Optimal Ranking from Pairwise Comparisons

no code implementations30 Dec 2015 Nihar B. Shah, Martin J. Wainwright

We consider data in the form of pairwise comparisons of n items, with the goal of precisely identifying the top k items for some value of k < n, or alternatively, recovering a ranking of all the items.

Computational Efficiency

Stochastically Transitive Models for Pairwise Comparisons: Statistical and Computational Issues

no code implementations19 Oct 2015 Nihar B. Shah, Sivaraman Balakrishnan, Adityanand Guntuboyina, Martin J. Wainwright

On the other hand, unlike in the BTL and Thurstone models, computing the minimax-optimal estimator in the stochastically transitive model is non-trivial, and we explore various computationally tractable alternatives.

Estimation from Pairwise Comparisons: Sharp Minimax Bounds with Topology Dependence

no code implementations6 May 2015 Nihar B. Shah, Sivaraman Balakrishnan, Joseph Bradley, Abhay Parekh, Kannan Ramchandran, Martin J. Wainwright

Data in the form of pairwise comparisons arises in many domains, including preference elicitation, sporting competitions, and peer grading among others.

Approval Voting and Incentives in Crowdsourcing

no code implementations19 Feb 2015 Nihar B. Shah, Dengyong Zhou, Yuval Peres

The growing need for labeled training data has made crowdsourcing an important part of machine learning.

On the Impossibility of Convex Inference in Human Computation

no code implementations21 Nov 2014 Nihar B. Shah, Dengyong Zhou

Human computation or crowdsourcing involves joint inference of the ground-truth-answers and the worker-abilities by optimizing an objective function, for instance, by maximizing the data likelihood based on an assumed underlying model.

Double or Nothing: Multiplicative Incentive Mechanisms for Crowdsourcing

no code implementations NeurIPS 2015 Nihar B. Shah, Dengyong Zhou

To address this fundamental challenge in crowdsourcing, we propose a simple payment mechanism to incentivize workers to answer only the questions that they are sure of and skip the rest.

When is it Better to Compare than to Score?

no code implementations25 Jun 2014 Nihar B. Shah, Sivaraman Balakrishnan, Joseph Bradley, Abhay Parekh, Kannan Ramchandran, Martin Wainwright

When eliciting judgements from humans for an unknown quantity, one often has the choice of making direct-scoring (cardinal) or comparative (ordinal) measurements.

Cannot find the paper you are looking for? You can Submit a new open access paper.