Search Results for author: Adam S. Charles

Found 10 papers, 5 papers with code

SiBBlInGS: Similarity-driven Building-Block Inference using Graphs across States

1 code implementation7 Jun 2023 Noga Mudrik, Gal Mishne, Adam S. Charles

Time series data across scientific domains are often collected under distinct states (e. g., tasks), wherein latent processes (e. g., biological factors) create complex inter- and intra-state variability.

Dictionary Learning Time Series

Multi-Lingual DALL-E Storytime

2 code implementations22 Dec 2022 Noga Mudrik, Adam S. Charles

While DALL-E is a promising tool for many applications, its decreased performance when given input in a different language, limits its audience and deepens the gap between populations.

Decomposed Linear Dynamical Systems (dLDS) for learning the latent components of neural dynamics

1 code implementation7 Jun 2022 Noga Mudrik, Yenho Chen, Eva Yezerets, Christopher J. Rozell, Adam S. Charles

Learning interpretable representations of neural dynamics at a population level is a crucial first step to understanding how observed neural activity relates to perception and behavior.

Dictionary Learning Time Series Analysis

Data Processing of Functional Optical Microscopy for Neuroscience

no code implementations10 Jan 2022 Hadas Benisty, Alexander Song, Gal Mishne, Adam S. Charles

Functional optical imaging in neuroscience is rapidly growing with the development of new optical systems and fluorescence indicators.

Efficient Tracking of Sparse Signals via an Earth Mover's Distance Dynamics Regularizer

no code implementations12 Jun 2018 Nicholas P. Bertrand, Adam S. Charles, John Lee, Pavel B. Dunn, Christopher J. Rozell

Tracking algorithms such as the Kalman filter aim to improve inference performance by leveraging the temporal dynamics in streaming observations.

Interpreting Deep Learning: The Machine Learning Rorschach Test?

no code implementations1 Jun 2018 Adam S. Charles

Theoretical understanding of deep learning is one of the most important tasks facing the statistics and machine learning communities.

BIG-bench Machine Learning Learning Theory

An Efficient and Flexible Spike Train Model via Empirical Bayes

no code implementations10 May 2016 Qi She, Xiaoli Wu, Beth Jelfs, Adam S. Charles, Rosa H. M. Chan

Our method integrates both Generalized Linear Models (GLMs) and empirical Bayes theory, which aims to (1) improve the accuracy and reliability of parameter estimation, compared to the maximum likelihood-based method for NB-GLM and Poisson-GLM; (2) effectively capture the over-dispersion nature of spike counts from both simulated data and experimental data; and (3) provide insight into both neural interactions and spiking behaviours of the neuronal populations.

Bayesian Inference

Short Term Memory Capacity in Networks via the Restricted Isometry Property

no code implementations1 Jul 2013 Adam S. Charles, Han Lun Yap, Christopher J. Rozell

Cortical networks are hypothesized to rely on transient network activity to support short term memory (STM).

Cannot find the paper you are looking for? You can Submit a new open access paper.