Search Results for author: Benjamin Feuer

Found 9 papers, 6 papers with code

TuneTables: Context Optimization for Scalable Prior-Data Fitted Networks

2 code implementations17 Feb 2024 Benjamin Feuer, Robin Tibor Schirrmeister, Valeriia Cherepanova, Chinmay Hegde, Frank Hutter, Micah Goldblum, Niv Cohen, Colin White

Similar to large language models, PFNs make use of pretraining and in-context learning to achieve strong performance on new tasks in a single forward pass.

Fairness In-Context Learning +1

Scaling TabPFN: Sketching and Feature Selection for Tabular Prior-Data Fitted Networks

no code implementations17 Nov 2023 Benjamin Feuer, Chinmay Hegde, Niv Cohen

Tabular classification has traditionally relied on supervised algorithms, which estimate the parameters of a prediction model using its training data.

feature selection tabular-classification

Exploring Dataset-Scale Indicators of Data Quality

no code implementations7 Nov 2023 Benjamin Feuer, Chinmay Hegde

Modern computer vision foundation models are trained on massive amounts of data, incurring large economic and environmental costs.

ArcheType: A Novel Framework for Open-Source Column Type Annotation using Large Language Models

1 code implementation27 Oct 2023 Benjamin Feuer, Yurong Liu, Chinmay Hegde, Juliana Freire

We introduce ArcheType, a simple, practical method for context sampling, prompt serialization, model querying, and label remapping, which enables large language models to solve CTA problems in a fully zero-shot manner.

 Ranked #1 on Column Type Annotation on WDC SOTAB (Weighted F1 metric)

Column Type Annotation Zero-Shot Learning

Distributionally Robust Classification on a Data Budget

1 code implementation7 Aug 2023 Benjamin Feuer, Ameya Joshi, Minh Pham, Chinmay Hegde

To our knowledge, this is the first result showing (near) state-of-the-art distributional robustness on limited data budgets.

Classification Image Classification +1

When Do Neural Nets Outperform Boosted Trees on Tabular Data?

1 code implementation NeurIPS 2023 Duncan McElfresh, Sujay Khandagale, Jonathan Valverde, Vishak Prasad C, Benjamin Feuer, Chinmay Hegde, Ganesh Ramakrishnan, Micah Goldblum, Colin White

To this end, we conduct the largest tabular data analysis to date, comparing 19 algorithms across 176 datasets, and we find that the 'NN vs. GBDT' debate is overemphasized: for a surprisingly high number of datasets, either the performance difference between GBDTs and NNs is negligible, or light hyperparameter tuning on a GBDT is more important than choosing between NNs and GBDTs.

LiT Tuned Models for Efficient Species Detection

1 code implementation12 Feb 2023 Andre Nakkab, Benjamin Feuer, Chinmay Hegde

Recent advances in training vision-language models have demonstrated unprecedented robustness and transfer learning effectiveness; however, standard computer vision datasets are image-only, and therefore not well adapted to such training methods.

Fine-Grained Image Classification Transfer Learning +1

Caption supervision enables robust learners

1 code implementation13 Oct 2022 Benjamin Feuer, Ameya Joshi, Chinmay Hegde

Vision language (VL) models like CLIP are robust to natural distribution shifts, in part because CLIP learns on unstructured data using a technique called caption supervision; the model inteprets image-linked texts as ground-truth labels.

A Meta-Analysis of Distributionally-Robust Models

no code implementations15 Jun 2022 Benjamin Feuer, Ameya Joshi, Chinmay Hegde

State-of-the-art image classifiers trained on massive datasets (such as ImageNet) have been shown to be vulnerable to a range of both intentional and incidental distribution shifts.

Cannot find the paper you are looking for? You can Submit a new open access paper.