Search Results for author: Dan Busbridge

Found 16 papers, 7 papers with code

Poly-View Contrastive Learning

no code implementations8 Mar 2024 Amitis Shidani, Devon Hjelm, Jason Ramapuram, Russ Webb, Eeshan Gunesh Dhekane, Dan Busbridge

Contrastive learning typically matches pairs of related views among a number of unrelated negative views.

Contrastive Learning Representation Learning

REALM: Robust Entropy Adaptive Loss Minimization for Improved Single-Sample Test-Time Adaptation

no code implementations7 Sep 2023 Skyler Seto, Barry-John Theobald, Federico Danieli, Navdeep Jaitly, Dan Busbridge

In online F-TTA, a pre-trained model is adapted using a stream of test samples by minimizing a self-supervised objective, such as entropy minimization.

Test-time Adaptation

The Role of Entropy and Reconstruction in Multi-View Self-Supervised Learning

1 code implementation20 Jul 2023 Borja Rodríguez-Gálvez, Arno Blaas, Pau Rodríguez, Adam Goliński, Xavier Suau, Jason Ramapuram, Dan Busbridge, Luca Zappella

We consider a different lower bound on the MI consisting of an entropy and a reconstruction term (ER), and analyze the main MVSSL families through its lens.

Self-Supervised Learning

DUET: 2D Structured and Approximately Equivariant Representations

1 code implementation28 Jun 2023 Xavier Suau, Federico Danieli, T. Anderson Keller, Arno Blaas, Chen Huang, Jason Ramapuram, Dan Busbridge, Luca Zappella

We propose 2D strUctured and EquivarianT representations (coined DUET), which are 2d representations organized in a matrix structure, and equivariant with respect to transformations acting on the input data.

Self-Supervised Learning Transfer Learning

Stabilizing Transformer Training by Preventing Attention Entropy Collapse

1 code implementation11 Mar 2023 Shuangfei Zhai, Tatiana Likhomanenko, Etai Littwin, Dan Busbridge, Jason Ramapuram, Yizhe Zhang, Jiatao Gu, Josh Susskind

We show that $\sigma$Reparam provides stability and robustness with respect to the choice of hyperparameters, going so far as enabling training (a) a Vision Transformer {to competitive performance} without warmup, weight decay, layer normalization or adaptive optimizers; (b) deep architectures in machine translation and (c) speech recognition to competitive performance without warmup and adaptive optimizers.

Automatic Speech Recognition Image Classification +6

Position Prediction as an Effective Pretraining Strategy

1 code implementation15 Jul 2022 Shuangfei Zhai, Navdeep Jaitly, Jason Ramapuram, Dan Busbridge, Tatiana Likhomanenko, Joseph Yitan Cheng, Walter Talbott, Chen Huang, Hanlin Goh, Joshua Susskind

This pretraining strategy which has been used in BERT models in NLP, Wav2Vec models in Speech and, recently, in MAE models in Vision, forces the model to learn about relationships between the content in different parts of the input using autoencoding related objectives.

Position speech-recognition +1

Do Self-Supervised and Supervised Methods Learn Similar Visual Representations?

no code implementations1 Oct 2021 Tom George Grigg, Dan Busbridge, Jason Ramapuram, Russ Webb

Despite the success of a number of recent techniques for visual self-supervised deep learning, there has been limited investigation into the representations that are ultimately learned.

Stochastic Contrastive Learning

no code implementations1 Oct 2021 Jason Ramapuram, Dan Busbridge, Xavier Suau, Russ Webb

While state-of-the-art contrastive Self-Supervised Learning (SSL) models produce results competitive with their supervised counterparts, they lack the ability to infer latent variables.

Contrastive Learning regression +1

Evaluating the fairness of fine-tuning strategies in self-supervised learning

no code implementations1 Oct 2021 Jason Ramapuram, Dan Busbridge, Russ Webb

In this work we examine how fine-tuning impacts the fairness of contrastive Self-Supervised Learning (SSL) models.

Fairness Self-Supervised Learning

Neural Temporal Point Processes For Modelling Electronic Health Records

1 code implementation27 Jul 2020 Joseph Enguehard, Dan Busbridge, Adam Bozson, Claire Woodcock, Nils Y. Hammerla

The modelling of Electronic Health Records (EHRs) has the potential to drive more efficient allocation of healthcare resources, enabling early intervention strategies and advancing personalised healthcare.

Point Processes

Neural Language Priors

no code implementations4 Oct 2019 Joseph Enguehard, Dan Busbridge, Vitalii Zhelezniak, Nils Hammerla

The choice of sentence encoder architecture reflects assumptions about how a sentence's meaning is composed from its constituent words.

Sentence

Relational Graph Attention Networks

2 code implementations ICLR 2019 Dan Busbridge, Dane Sherburn, Pietro Cavallo, Nils Y. Hammerla

We investigate Relational Graph Attention Networks, a class of models that extends non-relational graph attention mechanisms to incorporate relational information, opening up these methods to a wider variety of problems.

Graph Attention

Cannot find the paper you are looking for? You can Submit a new open access paper.