no code implementations • 24 Apr 2024 • Eric Slyman, Stefan Lee, Scott Cohen, Kushal Kafle
Recent dataset deduplication techniques have demonstrated that content-aware dataset pruning can dramatically reduce the cost of training Vision-Language Pretrained (VLP) models without significant performance losses compared to training on the original dataset.
no code implementations • 9 Nov 2023 • Daniel Claborne, Eric Slyman, Karl Pazdernik
We train an identity verification architecture and evaluate modifications to the part of the model that combines audio and visual representations, including in scenarios where one input is missing in either of two examples to be compared.
1 code implementation • ICCV 2023 • Eric Slyman, Minsuk Kahng, Stefan Lee
Recent work in vision-and-language demonstrates that large-scale pretraining can learn generalizable models that are efficiently transferable to downstream tasks.
1 code implementation • 29 Jul 2021 • Eric Slyman, Chris Daw, Morgan Skrabut, Ana Usenko, Brian Hutchinson
We obtain strong results on the new fine-grained task and state-of-the-art on the 4-way task: our best model obtains frame-level error rates of 6. 2%, 7. 7% and 28. 0% when generalizing to unseen instructors for the 4-way, 5-way, and 9-way classification tasks, respectively (relative reductions of 35. 4%, 48. 3% and 21. 6% over a strong baseline).