1 code implementation • NeurIPS 2023 • Harry Bendekgey, Gabriel Hope, Erik B. Sudderth
By composing graphical models with deep learning architectures, we learn generative models with the strengths of both frameworks.
no code implementations • pproximateinference AABI Symposium 2022 • Gabriel Hope, Madina Abdrakhmanova, Xiaoyin Chen, Michael C Hughes, Erik B. Sudderth
We consider training deep generative models toward two simultaneous goals: discriminative classification and generative modeling using an explicit likelihood.
no code implementations • 12 Dec 2020 • Gabriel Hope, Madina Abdrakhmanova, Xiaoyin Chen, Michael C. Hughes, Erik B. Sudderth
We develop a new framework for learning variational autoencoders and other deep generative models that balances generative and discriminative goals.
no code implementations • 1 Dec 2017 • Michael C. Hughes, Gabriel Hope, Leah Weiner, Thomas H. McCoy, Roy H. Perlis, Erik B. Sudderth, Finale Doshi-Velez
Supervisory signals can help topic models discover low-dimensional data representations that are more interpretable for clinical tasks.
no code implementations • 23 Jul 2017 • Michael C. Hughes, Leah Weiner, Gabriel Hope, Thomas H. McCoy Jr., Roy H. Perlis, Erik B. Sudderth, Finale Doshi-Velez
Supervisory signals have the potential to make low-dimensional data representations, like those learned by mixture and topic models, more interpretable and useful.