Search Results for author: Mete Kemertas

Found 7 papers, 3 papers with code

Maximum Entropy Model Correction in Reinforcement Learning

no code implementations29 Nov 2023 Amin Rakhsha, Mete Kemertas, Mohammad Ghavamzadeh, Amir-Massoud Farahmand

We propose and theoretically analyze an approach for planning with an approximate model in reinforcement learning that can reduce the adverse impact of model error.

Density Estimation reinforcement-learning

Efficient and Accurate Optimal Transport with Mirror Descent and Conjugate Gradients

1 code implementation17 Jul 2023 Mete Kemertas, Allan D. Jepson, Amir-Massoud Farahmand

We design a novel algorithm for optimal transport by drawing from the entropic optimal transport, mirror descent and conjugate gradients literatures.

Benchmarking

Approximate Policy Iteration with Bisimulation Metrics

1 code implementation6 Feb 2022 Mete Kemertas, Allan Jepson

Based on these results, we design an API($\alpha$) procedure that employs conservative policy updates and enjoys better performance bounds than the naive API approach.

Continuous Control Representation Learning

Towards Robust Bisimulation Metric Learning

1 code implementation NeurIPS 2021 Mete Kemertas, Tristan Aumentado-Armstrong

Learned representations in deep reinforcement learning (DRL) have to extract task-relevant information from complex observations, balancing between robustness to distraction and informativeness to the policy.

Continuous Control Informativeness +2

Dependency parsing with structure preserving embeddings

no code implementations EACL 2021 {\'A}kos K{\'a}d{\'a}r, Lan Xiao, Mete Kemertas, Federico Fancellu, Allan Jepson, Afsaneh Fazly

We do so by casting dependency parsing as a tree embedding problem where we incorporate geometric properties of dependency trees in the form of training losses within a graph-based parser.

Dependency Parsing Sentence

RankMI: A Mutual Information Maximizing Ranking Loss

no code implementations CVPR 2020 Mete Kemertas, Leila Pishdad, Konstantinos G. Derpanis, Afsaneh Fazly

We introduce an information-theoretic loss function, RankMI, and an associated training algorithm for deep representation learning for image retrieval.

Image Retrieval Representation Learning +1

Dynamic Scheduling of MPI-based Distributed Deep Learning Training Jobs

no code implementations21 Aug 2019 Tim Capes, Vishal Raheja, Mete Kemertas, Iqbal Mohomed

In this paper, we analyze the math behind ring architectures and make an informed adaptation of dynamic scheduling to ring architectures.

Math Scheduling

Cannot find the paper you are looking for? You can Submit a new open access paper.