Search Results for author: Arnav Chavan

Found 13 papers, 9 papers with code

Beyond Uniform Scaling: Exploring Depth Heterogeneity in Neural Architectures

no code implementations19 Feb 2024 Akash Guna R. T, Arnav Chavan, Deepak Gupta

Our method is flexible towards skip connections a mainstay in modern vision transformers.

Faster and Lighter LLMs: A Survey on Current Challenges and Way Forward

1 code implementation2 Feb 2024 Arnav Chavan, Raghav Magazine, Shubham Kushwaha, Mérouane Debbah, Deepak Gupta

Despite the impressive performance of LLMs, their widespread adoption faces challenges due to substantial computational and memory requirements during inference.

Model Compression

Rethinking Compression: Reduced Order Modelling of Latent Features in Large Language Models

1 code implementation12 Dec 2023 Arnav Chavan, Nahush Lele, Deepak Gupta

Due to the substantial scale of Large Language Models (LLMs), the direct application of conventional compression methodologies proves impractical.

Model Compression

One-for-All: Generalized LoRA for Parameter-Efficient Fine-tuning

1 code implementation13 Jun 2023 Arnav Chavan, Zhuang Liu, Deepak Gupta, Eric Xing, Zhiqiang Shen

We present Generalized LoRA (GLoRA), an advanced approach for universal parameter-efficient fine-tuning tasks.

Domain Generalization Few-Shot Learning +1

Patch Gradient Descent: Training Neural Networks on Very Large Images

no code implementations31 Jan 2023 Deepak K. Gupta, Gowreesh Mago, Arnav Chavan, Dilip K. Prasad

Traditional CNN models are trained and tested on relatively low resolution images (<300 px), and cannot be directly operated on large-scale images due to compute and memory constraints.

On Designing Light-Weight Object Trackers through Network Pruning: Use CNNs or Transformers?

1 code implementation24 Nov 2022 Saksham Aggarwal, Taneesh Gupta, Pawan Kumar Sahu, Arnav Chavan, Rishabh Tiwari, Dilip K. Prasad, Deepak K. Gupta

A comparison between SOTA trackers using CNNs, transformers as well as the combination of the two is presented to study their stability at various compression ratios.

Network Pruning Object +1

Dynamic Kernel Selection for Improved Generalization and Memory Efficiency in Meta-learning

1 code implementation CVPR 2022 Arnav Chavan, Rishabh Tiwari, Udbhav Bamba, Deepak K. Gupta

MetaDOCK compresses the meta-model as well as the task-specific inner models, thus providing significant reduction in model size for each task, and through constraining the number of active kernels for every task, it implicitly mitigates the issue of meta-overfitting.

Meta-Learning

Vision Transformer Slimming: Multi-Dimension Searching in Continuous Optimization Space

1 code implementation CVPR 2022 Arnav Chavan, Zhiqiang Shen, Zhuang Liu, Zechun Liu, Kwang-Ting Cheng, Eric Xing

This paper explores the feasibility of finding an optimal sub-model from a vision transformer and introduces a pure vision transformer slimming (ViT-Slim) framework.

Transfer Learning Gaussian Anomaly Detection by Fine-tuning Representations

no code implementations9 Aug 2021 Oliver Rippel, Arnav Chavan, Chucai Lei, Dorit Merhof

In our work, we propose a new method to overcome catastrophic forgetting and thus successfully fine-tune pre-trained representations for AD in the transfer learning setting.

Anomaly Detection Transfer Learning

ChipNet: Budget-Aware Pruning with Heaviside Continuous Approximations

1 code implementation ICLR 2021 Rishabh Tiwari, Udbhav Bamba, Arnav Chavan, Deepak K. Gupta

Structured pruning methods are among the effective strategies for extracting small resource-efficient convolutional neural networks from their dense counterparts with minimal loss in accuracy.

Rescaling CNN through Learnable Repetition of Network Parameters

1 code implementation14 Jan 2021 Arnav Chavan, Udbhav Bamba, Rishabh Tiwari, Deepak Gupta

We show that small base networks when rescaled, can provide performance comparable to deeper networks with as low as 6% of optimization parameters of the deeper one.

Multi-Plateau Ensemble for Endoscopic Artefact Segmentation and Detection

1 code implementation23 Mar 2020 Suyog Jadhav, Udbhav Bamba, Arnav Chavan, Rishabh Tiwari, Aryan Raj

Endoscopic artefact detection challenge consists of 1) Artefact detection, 2) Semantic segmentation, and 3) Out-of-sample generalisation.

object-detection Object Detection +2

Cannot find the paper you are looking for? You can Submit a new open access paper.