Search Results for author: Yoav Levine

Found 21 papers, 8 papers with code

Rationality Report Cards: Assessing the Economic Rationality of Large Language Models

no code implementations14 Feb 2024 Narun Raman, Taylor Lundy, Samuel Amouyal, Yoav Levine, Kevin Leyton-Brown, Moshe Tennenholtz

We begin by surveying the economic literature on rational decision making, taxonomizing a large set of fine-grained "elements" that an agent should exhibit, along with dependencies between them.

Decision Making

Tradeoffs Between Alignment and Helpfulness in Language Models

no code implementations29 Jan 2024 Yotam Wolf, Noam Wies, Dorin Shteyman, Binyamin Rothberg, Yoav Levine, Amnon Shashua

Representation engineering yields gains in alignment oriented tasks such as resistance to adversarial attacks and reduction of social biases, but was also shown to cause a decrease in the ability of the model to perform basic tasks.

Language Modelling

Generating Benchmarks for Factuality Evaluation of Language Models

2 code implementations13 Jul 2023 Dor Muhlgay, Ori Ram, Inbal Magar, Yoav Levine, Nir Ratner, Yonatan Belinkov, Omri Abend, Kevin Leyton-Brown, Amnon Shashua, Yoav Shoham

FACTOR automatically transforms a factual corpus of interest into a benchmark evaluating an LM's propensity to generate true facts from the corpus vs. similar but incorrect statements.

Language Modelling Retrieval

Human or Not? A Gamified Approach to the Turing Test

no code implementations31 May 2023 Daniel Jannai, Amos Meron, Barak Lenz, Yoav Levine, Yoav Shoham

Over the course of a month, the game was played by over 1. 5 million users who engaged in anonymous two-minute chat sessions with either another human or an AI language model which was prompted to behave like humans.

Language Modelling

Fundamental Limitations of Alignment in Large Language Models

no code implementations19 Apr 2023 Yotam Wolf, Noam Wies, Oshri Avnery, Yoav Levine, Amnon Shashua

An important aspect in developing language models that interact with humans is aligning their behavior to be useful and unharmful for their human users.

In-Context Retrieval-Augmented Language Models

1 code implementation31 Jan 2023 Ori Ram, Yoav Levine, Itay Dalmedigos, Dor Muhlgay, Amnon Shashua, Kevin Leyton-Brown, Yoav Shoham

Retrieval-Augmented Language Modeling (RALM) methods, which condition a language model (LM) on relevant documents from a grounding corpus during generation, were shown to significantly improve language modeling performance.

Language Modelling Retrieval +1

Parallel Context Windows for Large Language Models

1 code implementation21 Dec 2022 Nir Ratner, Yoav Levine, Yonatan Belinkov, Ori Ram, Inbal Magar, Omri Abend, Ehud Karpas, Amnon Shashua, Kevin Leyton-Brown, Yoav Shoham

We present Parallel Context Windows (PCW), a method that alleviates the context window restriction for any off-the-shelf LLM without further training.

In-Context Learning Playing the Game of 2048 +2

Standing on the Shoulders of Giant Frozen Language Models

no code implementations21 Apr 2022 Yoav Levine, Itay Dalmedigos, Ori Ram, Yoel Zeldes, Daniel Jannai, Dor Muhlgay, Yoni Osin, Opher Lieber, Barak Lenz, Shai Shalev-Shwartz, Amnon Shashua, Kevin Leyton-Brown, Yoav Shoham

To demonstrate this, we introduce three novel methods for leveraging frozen models: input-dependent prompt tuning, frozen readers, and recursive LMs, each of which vastly improves on current frozen-model approaches.

Sub-Task Decomposition Enables Learning in Sequence to Sequence Tasks

1 code implementation6 Apr 2022 Noam Wies, Yoav Levine, Amnon Shashua

Recently, several works have demonstrated high gains by taking a straightforward approach for incorporating intermediate supervision in compounded natural language problems: the sequence-to-sequence LM is fed with an augmented input, in which the decomposed tasks' labels are simply concatenated to the original input.

The Inductive Bias of In-Context Learning: Rethinking Pretraining Example Design

no code implementations ICLR 2022 Yoav Levine, Noam Wies, Daniel Jannai, Dan Navon, Yedid Hoshen, Amnon Shashua

We highlight a bias introduced by this common practice: we prove that the pretrained NLM can model much stronger dependencies between text segments that appeared in the same training example, than it can between text segments that appeared in different training examples.

Chunking In-Context Learning +4

Which transformer architecture fits my data? A vocabulary bottleneck in self-attention

no code implementations9 May 2021 Noam Wies, Yoav Levine, Daniel Jannai, Amnon Shashua

After their successful debut in natural language processing, Transformer architectures are now becoming the de-facto standard in many domains.

PMI-Masking: Principled masking of correlated spans

1 code implementation ICLR 2021 Yoav Levine, Barak Lenz, Opher Lieber, Omri Abend, Kevin Leyton-Brown, Moshe Tennenholtz, Yoav Shoham

Specifically, we show experimentally that PMI-Masking reaches the performance of prior masking approaches in half the training time, and consistently improves performance at the end of training.

The Depth-to-Width Interplay in Self-Attention

1 code implementation NeurIPS 2020 Yoav Levine, Noam Wies, Or Sharir, Hofit Bata, Amnon Shashua

Our guidelines elucidate the depth-to-width trade-off in self-attention networks of sizes up to the scale of GPT3 (which we project to be too deep for its size), and beyond, marking an unprecedented width of 30K as optimal for a 1-Trillion parameter network.

Deep autoregressive models for the efficient variational simulation of many-body quantum systems

2 code implementations11 Feb 2019 Or Sharir, Yoav Levine, Noam Wies, Giuseppe Carleo, Amnon Shashua

Artificial Neural Networks were recently shown to be an efficient representation of highly-entangled many-body quantum states.

Variational Monte Carlo

Benefits of Depth for Long-Term Memory of Recurrent Networks

no code implementations ICLR 2018 Yoav Levine, Or Sharir, Amnon Shashua

We prove that deep recurrent networks support Start-End separation ranks which are exponentially higher than those supported by their shallow counterparts.

Attribute Time Series Analysis

On the Long-Term Memory of Deep Recurrent Networks

1 code implementation25 Oct 2017 Yoav Levine, Or Sharir, Alon Ziv, Amnon Shashua

A key attribute that drives the unprecedented success of modern Recurrent Neural Networks (RNNs) on learning tasks which involve sequential data, is their ability to model intricate long-term temporal dependencies.

Attribute Tensor Networks

Analysis and Design of Convolutional Networks via Hierarchical Tensor Decompositions

no code implementations5 May 2017 Nadav Cohen, Or Sharir, Yoav Levine, Ronen Tamari, David Yakira, Amnon Shashua

Expressive efficiency refers to the ability of a network architecture to realize functions that require an alternative architecture to be much larger.

Inductive Bias

Deep Learning and Quantum Entanglement: Fundamental Connections with Implications to Network Design

no code implementations ICLR 2018 Yoav Levine, David Yakira, Nadav Cohen, Amnon Shashua

This description enables us to carry a graph-theoretic analysis of a convolutional network, with which we demonstrate a direct control over the inductive bias of the deep network via its channel numbers, that are related to the min-cut in the underlying graph.

Inductive Bias

Cannot find the paper you are looking for? You can Submit a new open access paper.