HellaSwag

Introduced by Zellers et al. in HellaSwag: Can a Machine Really Finish Your Sentence?

HellaSwag is a challenge dataset for evaluating commonsense NLI that is specially hard for state-of-the-art models, though its questions are trivial for humans (>95% accuracy).

Papers


Paper Code Results Date Stars

Tasks


Similar Datasets


License


Modalities


Languages