Search Results for author: Jacob R. Stevens

Found 4 papers, 1 papers with code

X-Former: In-Memory Acceleration of Transformers

no code implementations13 Mar 2023 Shrihari Sridharan, Jacob R. Stevens, Kaushik Roy, Anand Raghunathan

Transformers have achieved great success in a wide variety of natural language processing (NLP) tasks due to the attention mechanism, which assigns an importance score for every word relative to other words in a sequence.

Blocking

Efficient Ensembles of Graph Neural Networks

no code implementations29 Sep 2021 Amrit Nagarajan, Jacob R. Stevens, Anand Raghunathan

In this work, we leverage the unique characteristics of GNNs to overcome these overheads, creating efficient ensemble GNNs that are faster than even single models at inference time.

Ensemble Learning Network Pruning +2

Specialized Transformers: Faster, Smaller and more Accurate NLP Models

no code implementations29 Sep 2021 Amrit Nagarajan, Sanchari Sen, Jacob R. Stevens, Anand Raghunathan

We propose a Specialization framework to create optimized transformer models for a given downstream task.

Hard Attention Quantization

AxFormer: Accuracy-driven Approximation of Transformers for Faster, Smaller and more Accurate NLP Models

1 code implementation7 Oct 2020 Amrit Nagarajan, Sanchari Sen, Jacob R. Stevens, Anand Raghunathan

We propose AxFormer, a systematic framework that applies accuracy-driven approximations to create optimized transformer models for a given downstream task.

Hard Attention Quantization +1

Cannot find the paper you are looking for? You can Submit a new open access paper.