1 code implementation • 2 Feb 2024 • Zach Nussbaum, John X. Morris, Brandon Duderstadt, Andriy Mulyar
This technical report describes the training of nomic-embed-text-v1, the first fully reproducible, open-source, open-weights, open-data, 8192 context length English text embedding model that outperforms both OpenAI Ada-002 and OpenAI text-embedding-3-small on short and long-context tasks.
1 code implementation • 6 Nov 2023 • Yuvanesh Anand, Zach Nussbaum, Adam Treat, Aaron Miller, Richard Guo, Ben Schmidt, GPT4All Community, Brandon Duderstadt, Andriy Mulyar
It is our hope that this paper acts as both a technical overview of the original GPT4All models as well as a case study on the subsequent growth of the GPT4All open source ecosystem.
no code implementations • 9 May 2023 • Brandon Duderstadt, Hayden S. Helm, Carey E. Priebe
Further, we demonstrate how our methodology can be extended to facilitate population level model comparison.
no code implementations • 12 Nov 2020 • Hayden S. Helm, Ronak D. Mehta, Brandon Duderstadt, Weiwei Yang, Christoper M. White, Ali Geisa, Joshua T. Vogelstein, Carey E. Priebe
Herein we define a measure of similarity between classification distributions that is both principled from the perspective of statistical pattern recognition and useful from the perspective of machine learning practitioners.
no code implementations • 9 Mar 2018 • Gregory Kiar, Robert J. Anderson, Alex Baden, Alexandra Badea, Eric W. Bridgeford, Andrew Champion, Vikram Chandrashekhar, Forrest Collman, Brandon Duderstadt, Alan C. Evans, Florian Engert, Benjamin Falk, Tristan Glatard, William R. Gray Roncal, David N. Kennedy, Jeremy Maitin-Shepard, Ryan A. Marren, Onyeka Nnaemeka, Eric Perlman, Sharmishtaas Seshamani, Eric T. Trautman, Daniel J. Tward, Pedro Antonio Valdés-Sosa, Qing Wang, Michael I. Miller, Randal Burns, Joshua T. Vogelstein
Neuroscientists are now able to acquire data at staggering rates across spatiotemporal scales.