no code implementations • 15 Nov 2023 • Yahan Yang, Soham Dan, Dan Roth, Insup Lee
We also conduct several ablation experiments to study the effect of language distances, language corpus size, and model size on calibration, and how multilingual models compare with their monolingual counterparts for diverse tasks and languages.
no code implementations • 21 Feb 2023 • Ramneet Kaur, Xiayan Ji, Souradeep Dutta, Michele Caprio, Yahan Yang, Elena Bernardis, Oleg Sokolsky, Insup Lee
This can render the current OOD detectors impermeable to inputs lying outside the training distribution but with the same semantic information (e. g. training class labels).
no code implementations • 20 Dec 2022 • Yahan Yang, Soham Dan, Dan Roth, Insup Lee
Recently it has been shown that state-of-the-art NLP models are vulnerable to adversarial attacks, where the predictions of a model can be drastically altered by slight modifications to the input (such as synonym substitutions).
no code implementations • 10 Jun 2022 • Souradeep Dutta, Yahan Yang, Elena Bernardis, Edgar Dobriban, Insup Lee
We propose a new method for classification which can improve robustness to distribution shifts, by combining expert knowledge about the ``high-level" structure of the data with standard classifiers.
no code implementations • 23 Feb 2020 • Yiannis Kantaros, Taylor Carpenter, Kaustubh Sridhar, Yahan Yang, Insup Lee, James Weimer
To highlight this, we demonstrate the efficiency of the proposed detector on ImageNet, a task that is computationally challenging for the majority of relevant defenses, and on physically attacked traffic signs that may be encountered in real-time autonomy applications.