Search Results for author: Kai Sheng Tai

Found 8 papers, 6 papers with code

Multi-entity Video Transformers for Fine-Grained Video Representation Learning

1 code implementation17 Nov 2023 Matthew Walmer, Rose Kanjirathinkal, Kai Sheng Tai, Keyur Muzumdar, Taipeng Tian, Abhinav Shrivastava

In this work, we advance the state-of-the-art for this area by re-examining the design of transformer architectures for video representation learning.

Representation Learning

Analyzing Populations of Neural Networks via Dynamical Model Embedding

no code implementations27 Feb 2023 Jordan Cotler, Kai Sheng Tai, Felipe Hernández, Blake Elias, David Sussillo

The specific model to be emulated is determined by a model embedding vector that the meta-model takes as input; these model embedding vectors constitute a manifold corresponding to the given population of models.

Spartan: Differentiable Sparsity via Regularized Transportation

1 code implementation27 May 2022 Kai Sheng Tai, Taipeng Tian, Ser-Nam Lim

We present Spartan, a method for training sparse neural network models with a predetermined level of sparsity.

Network Pruning

Sinkhorn Label Allocation: Semi-Supervised Classification via Annealed Self-Training

1 code implementation17 Feb 2021 Kai Sheng Tai, Peter Bailis, Gregory Valiant

Self-training is a standard approach to semi-supervised learning where the learner's own predictions on unlabeled data are used as supervision during training.

Classification General Classification +1

Equivariant Transformer Networks

3 code implementations25 Jan 2019 Kai Sheng Tai, Peter Bailis, Gregory Valiant

How can prior knowledge on the transformation invariances of a domain be incorporated into the architecture of a neural network?

General Classification Image Classification

Sketching Linear Classifiers over Data Streams

1 code implementation7 Nov 2017 Kai Sheng Tai, Vatsal Sharan, Peter Bailis, Gregory Valiant

We introduce a new sub-linear space sketch---the Weight-Median Sketch---for learning compressed linear classifiers over data streams while supporting the efficient recovery of large-magnitude weights in the model.

feature selection

Improved Semantic Representations From Tree-Structured Long Short-Term Memory Networks

16 code implementations IJCNLP 2015 Kai Sheng Tai, Richard Socher, Christopher D. Manning

Because of their superior ability to preserve sequence information over time, Long Short-Term Memory (LSTM) networks, a type of recurrent neural network with a more complex computational unit, have obtained strong results on a variety of sequence modeling tasks.

General Classification Semantic Similarity +2

Cannot find the paper you are looking for? You can Submit a new open access paper.