1 code implementation • 24 Oct 2023 • Minh Nguyen, Alan Q. Wang, Heejong Kim, Mert R. Sabuncu
CoPA assumes that (1) generation mechanism is stable, i. e. label Y and confounding variable(s) Z generate X, and (2) the unstable conditional prevalence in each site E fully accounts for the unstable correlations between X and Y .
no code implementations • 2 Oct 2023 • Alan Q. Wang, Batuhan K. Karaman, Heejong Kim, Jacob Rosenthal, Rachit Saluja, Sean I. Young, Mert R. Sabuncu
To answer these questions, we identify a need to formalize the goals and elements of interpretability in MLMI.
2 code implementations • 19 Apr 2023 • Alan Q. Wang, Evan M. Yu, Adrian V. Dalca, Mert R. Sabuncu
Our core insight which addresses these shortcomings is that corresponding keypoints between images can be used to obtain the optimal transformation via a differentiable closed-form expression.
1 code implementation • 7 Dec 2022 • Alan Q. Wang, Mert R. Sabuncu
The first is the label weights, and the second is our novel concept of the ``support influence function,'' which is an easy-to-compute metric that quantifies the influence of a support element on the prediction for a given query.
2 code implementations • 22 Feb 2022 • Alan Q. Wang, Adrian V. Dalca, Mert R. Sabuncu
The typical approach is to train the model for a hyperparameter setting determined with some empirical or theoretical justification.
1 code implementation • 6 Feb 2022 • Tianyu Ma, Alan Q. Wang, Adrian V. Dalca, Mert R. Sabuncu
The key building block of a CNN is the convolutional kernel that aggregates information from the pixel neighborhood and shares weights across all pixels.
1 code implementation • 17 May 2021 • Alan Q. Wang, Aaron K. LaViolette, Leo Moon, Chris Xu, Mert R. Sabuncu
Compressed sensing fluorescence microscopy (CS-FM) proposes a scheme whereby less measurements are collected during sensing and reconstruction is performed to recover the image.
2 code implementations • 6 Jan 2021 • Alan Q. Wang, Adrian V. Dalca, Mert R. Sabuncu
In this paper, we explore a novel strategy of using a hypernetwork to generate the parameters of a separate reconstruction network as a function of the regularization weight(s), resulting in a regularization-agnostic reconstruction model.
1 code implementation • 29 Jul 2020 • Alan Q. Wang, Adrian V. Dalca, Mert R. Sabuncu
In this paper, we explore a novel strategy to train an unrolled reconstruction network in an unsupervised fashion by adopting a loss function widely-used in classical optimization schemes.
1 code implementation • 26 Jul 2019 • Cagla D. Bahadir, Alan Q. Wang, Adrian V. Dalca, Mert R. Sabuncu
In our experiments, we demonstrate that LOUPE-optimized under-sampling masks are data-dependent, varying significantly with the imaged anatomy, and perform well with different reconstruction methods.