Search Results for author: Davit Soselia

Found 5 papers, 1 papers with code

ODIN: Disentangled Reward Mitigates Hacking in RLHF

no code implementations11 Feb 2024 Lichang Chen, Chen Zhu, Davit Soselia, Jiuhai Chen, Tianyi Zhou, Tom Goldstein, Heng Huang, Mohammad Shoeybi, Bryan Catanzaro

In this work, we study the issue of reward hacking on the response length, a challenge emerging in Reinforcement Learning from Human Feedback (RLHF) on LLMs.

Reviving Shift Equivariance in Vision Transformers

no code implementations13 Jun 2023 Peijian Ding, Davit Soselia, Thomas Armstrong, Jiahao Su, Furong Huang

While the self-attention operator in vision transformers (ViT) is permutation-equivariant and thus shift-equivariant, patch embedding, positional encoding, and subsampled attention in ViT variants can disrupt this property, resulting in inconsistent predictions even under small shift perturbations.

Inductive Bias

Learning UI-to-Code Reverse Generator Using Visual Critic Without Rendering

no code implementations24 May 2023 Davit Soselia, Khalid Saifullah, Tianyi Zhou

We evaluate the UI-to-Code performance using a combination of automated metrics such as MSE, BLEU, IoU, and a novel htmlBLEU score.

Code Generation reinforcement-learning

RNN-based Online Handwritten Character Recognition Using Accelerometer and Gyroscope Data

no code implementations24 Jul 2019 Davit Soselia, Shota Amashukeli, Irakli Koberidze, Levan Shugliashvili

We have built a dataset of timestamped gyroscope and accelerometer data gathered during the manual process of handwriting Latin characters, labeled with the character being written; in total, the dataset con-sists of 1500 gyroscope and accelerometer data sequenc-es for 8 characters of the Latin alphabet from 6 different people, and 20 characters, each 1500 samples from Georgian alphabet from 5 different people.

Reproduction Report on "Learn to Pay Attention"

1 code implementation11 Dec 2018 Levan Shugliashvili, Davit Soselia, Shota Amashukeli, Irakli Koberidze

We have successfully implemented the "Learn to Pay Attention" model of attention mechanism in convolutional neural networks, and have replicated the results of the original paper in the categories of image classification and fine-grained recognition.

General Classification Image Classification

Cannot find the paper you are looking for? You can Submit a new open access paper.