1 code implementation • 6 Dec 2023 • Cody Tipton, Elizabeth Coda, Davis Brown, Alyson Bittner, Jung Lee, Grayson Jorgenson, Tegan Emerson, Henry Kvinge
Characteristic classes, which are abstract topological invariants associated with vector bundles, have become an important notion in modern physics with surprising real-world consequences.
no code implementations • 23 Oct 2023 • Davis Brown, Charles Godfrey, Nicholas Konz, Jonathan Tu, Henry Kvinge
As language models are applied to an increasing number of real-world applications, understanding their inner workings has become an important issue in model trust, interpretability, and transparency.
no code implementations • 4 Oct 2023 • Nicholas Konz, Charles Godfrey, Madelyn Shapiro, Jonathan Tu, Henry Kvinge, Davis Brown
By now there is substantial evidence that deep learning models learn certain human-interpretable features as part of their internal representations of data.
no code implementations • 24 Jul 2023 • Davis Brown, Nikhil Vyas, Yamini Bansal
Our findings give evidence that while Linear Mode Connectivity improves with increased network width, this improvement is not due to an increase in basis correlation.
no code implementations • 24 Mar 2023 • Charles Godfrey, Henry Kvinge, Elise Bishoff, Myles Mckay, Davis Brown, Tim Doster, Eleanor Byler
Past work exploring adversarial vulnerability have focused on situations where an adversary can perturb all dimensions of model input.
no code implementations • 10 Mar 2023 • Charles Godfrey, Michael G. Rawson, Davis Brown, Henry Kvinge
The space of permutation equivariant linear layers is a generalization of the partition algebra, an object first discovered in statistical physics with deep connections to the representation theory of the symmetric group, and the basis described above generalizes the so-called orbit basis of the partition algebra.
no code implementations • 28 Feb 2023 • Davis Brown, Charles Godfrey, Cody Nizinski, Jonathan Tu, Henry Kvinge
The current trend toward ever-larger models makes standard retraining procedures an ever-more expensive burden.
no code implementations • 16 Feb 2023 • Henry Kvinge, Davis Brown, Charles Godfrey
We find that choice of prompt has a substantial impact on the intrinsic dimension of representations at both layers of the model which we explored, but that the nature of this impact depends on the layer being considered.
no code implementations • 1 Dec 2022 • Emilie Purvine, Davis Brown, Brett Jefferson, Cliff Joslyn, Brenda Praggastis, Archit Rathore, Madelyn Shapiro, Bei Wang, Youjia Zhou
Topological data analysis (TDA) is a branch of computational mathematics, bridging algebraic topology and data science, that provides compact, noise-robust representations of complex structures.
no code implementations • 19 Nov 2022 • Henry Kvinge, Grayson Jorgenson, Davis Brown, Charles Godfrey, Tegan Emerson
While the last five years have seen considerable progress in understanding the internal representations of deep learning models, many questions remain.
1 code implementation • 3 Oct 2022 • Charles Godfrey, Elise Bishoff, Myles Mckay, Davis Brown, Grayson Jorgenson, Henry Kvinge, Eleanor Byler
It is widely acknowledged that trained convolutional neural networks (CNNs) have different levels of sensitivity to signals of different frequency.
no code implementations • 14 Aug 2022 • Brenda Praggastis, Davis Brown, Carlos Ortiz Marrero, Emilie Purvine, Madelyn Shapiro, Bei Wang
Fully connected layers can be studied by decomposing their weight matrices using a singular value decomposition, in effect studying the correlations between the rows in each matrix to discover the dynamics of the map.
2 code implementations • 27 May 2022 • Charles Godfrey, Davis Brown, Tegan Emerson, Henry Kvinge
In this paper we seek to connect the symmetries arising from the architecture of a family of models with the symmetries of that family's internal representation of data.
no code implementations • 14 Oct 2021 • Davis Brown, Henry Kvinge
Methods for model explainability have become increasingly critical for testing the fairness and soundness of deep learning.