Search Results for author: Ivan Titov

Found 100 papers, 61 papers with code

Unlearning Reveals the Influential Training Data of Language Models

no code implementations26 Jan 2024 Masaru Isonuma, Ivan Titov

This paper presents UnTrac, which estimates the influence of a training dataset by unlearning it from the trained model.

Compositional Generalization for Data-to-Text Generation

no code implementations5 Dec 2023 Xinnuo Xu, Ivan Titov, Mirella Lapata

Data-to-text generation involves transforming structured data, often represented as predicate-argument tuples, into coherent textual descriptions.

Data-to-Text Generation Sentence

Latent Feature-based Data Splits to Improve Generalisation Evaluation: A Hate Speech Detection Case Study

1 code implementation16 Nov 2023 Maike Züfle, Verna Dankers, Ivan Titov

We challenge hate speech models via new train-test splits of existing datasets that rely on the clustering of models' hidden representations.

Hate Speech Detection

Memorisation Cartography: Mapping out the Memorisation-Generalisation Continuum in Neural Machine Translation

no code implementations9 Nov 2023 Verna Dankers, Ivan Titov, Dieuwke Hupkes

When training a neural network, it will quickly memorise some source-target mappings from your dataset but never learn some others.

counterfactual Machine Translation +2

Subspace Chronicles: How Linguistic Information Emerges, Shifts and Interacts during Language Model Training

no code implementations25 Oct 2023 Max Müller-Eberstein, Rob van der Goot, Barbara Plank, Ivan Titov

We identify critical learning phases across tasks and time, during which subspaces emerge, share information, and later disentangle to specialize.

Language Modelling Multi-Task Learning

Cross-Modal Conceptualization in Bottleneck Models

2 code implementations23 Oct 2023 Danis Alukaev, Semen Kiselev, Ilya Pershin, Bulat Ibragimov, Vladimir Ivanov, Alexey Kornaev, Ivan Titov

Concept Bottleneck Models (CBMs) assume that training examples (e. g., x-ray images) are annotated with high-level concepts (e. g., types of abnormalities), and perform classification by first predicting the concepts, followed by predicting the label relying on these concepts.

Disentanglement

On the Transferability of Visually Grounded PCFGs

1 code implementation21 Oct 2023 Yanpeng Zhao, Ivan Titov

We consider a zero-shot transfer learning setting where a model is trained on the source domain and is directly applied to target domains, without any further training.

Transfer Learning

Cache & Distil: Optimising API Calls to Large Language Models

no code implementations20 Oct 2023 Guillem Ramírez, Matthias Lindemann, Alexandra Birch, Ivan Titov

To curtail the frequency of these calls, one can employ a smaller language model -- a student -- which is continuously trained on the responses of the LLM.

Active Learning Language Modelling +1

Injecting a Structural Inductive Bias into a Seq2Seq Model by Simulation

no code implementations1 Oct 2023 Matthias Lindemann, Alexander Koller, Ivan Titov

Strong inductive biases enable learning from little data and help generalization outside of the training distribution.

Few-Shot Learning Inductive Bias +1

Autoencoding Conditional Neural Processes for Representation Learning

1 code implementation29 May 2023 Victor Prokhorov, Ivan Titov, N. Siddharth

Conditional neural processes (CNPs) are a flexible and efficient family of models that learn to learn a stochastic process from data.

Representation Learning

Compositional Generalization without Trees using Multiset Tagging and Latent Permutations

1 code implementation26 May 2023 Matthias Lindemann, Alexander Koller, Ivan Titov

Our model outperforms pretrained seq2seq models and prior work on realistic semantic parsing tasks that require generalization to longer examples.

Inductive Bias Semantic Parsing +1

Recursive Neural Networks with Bottlenecks Diagnose (Non-)Compositionality

1 code implementation31 Jan 2023 Verna Dankers, Ivan Titov

We illustrate that comparing data's representations in models with and without the bottleneck can be used to produce a compositionality metric.

Sentiment Analysis Sentiment Classification

Hierarchical Phrase-based Sequence-to-Sequence Learning

1 code implementation15 Nov 2022 Bailin Wang, Ivan Titov, Jacob Andreas, Yoon Kim

We describe a neural transducer that maintains the flexibility of standard sequence-to-sequence (seq2seq) models while incorporating hierarchical phrases as a source of inductive bias during training and as explicit constraints during inference.

Inductive Bias Machine Translation +2

Compositional Generalisation with Structured Reordering and Fertility Layers

1 code implementation6 Oct 2022 Matthias Lindemann, Alexander Koller, Ivan Titov

Seq2seq models have been shown to struggle with compositional generalisation, i. e. generalising to new and potentially more complex structures than seen during training.

Semantic Parsing

Can Transformer be Too Compositional? Analysing Idiom Processing in Neural Machine Translation

1 code implementation ACL 2022 Verna Dankers, Christopher G. Lucas, Ivan Titov

In this work, we investigate whether the non-compositionality of idioms is reflected in the mechanics of the dominant NMT model, Transformer, by analysing the hidden states and attention patterns for models with English as source language and one of seven European languages as target language.

Machine Translation NMT +1

Sparse Interventions in Language Models with Differentiable Masking

no code implementations13 Dec 2021 Nicola De Cao, Leon Schmid, Dieuwke Hupkes, Ivan Titov

Typically, interpretation methods i) do not guarantee that the model actually uses the encoded information, and ii) do not discover small subsets of neurons responsible for a considered phenomenon.

Bias Detection Gender Bias Detection

Learning Opinion Summarizers by Selecting Informative Reviews

1 code implementation EMNLP 2021 Arthur Bražinskas, Mirella Lapata, Ivan Titov

Opinion summarization has been traditionally approached with unsupervised, weakly-supervised and few-shot learning techniques.

Few-Shot Learning Opinion Summarization +2

Highly Parallel Autoregressive Entity Linking with Discriminative Correction

1 code implementation EMNLP 2021 Nicola De Cao, Wilker Aziz, Ivan Titov

Generative approaches have been recently shown to be effective for both Entity Disambiguation and Entity Linking (i. e., joint mention detection and disambiguation).

Entity Disambiguation Entity Linking

Language Modeling, Lexical Translation, Reordering: The Training Process of NMT through the Lens of Classical SMT

no code implementations EMNLP 2021 Elena Voita, Rico Sennrich, Ivan Titov

Differently from the traditional statistical MT that decomposes the translation task into distinct separately learned components, neural machine translation uses a single neural network to model the entire translation process.

Language Modelling Machine Translation +4

Beyond Sentence-Level End-to-End Speech Translation: Context Helps

1 code implementation ACL 2021 Biao Zhang, Ivan Titov, Barry Haddow, Rico Sennrich

Document-level contextual information has shown benefits to text-based machine translation, but whether and how context helps end-to-end (E2E) speech translation (ST) is still under-studied.

Computational Efficiency feature selection +3

Exploring Unsupervised Pretraining Objectives for Machine Translation

1 code implementation Findings (ACL) 2021 Christos Baziotis, Ivan Titov, Alexandra Birch, Barry Haddow

Unsupervised cross-lingual pretraining has achieved strong results in neural machine translation (NMT), by drastically reducing the need for large parallel data.

Language Modelling Machine Translation +3

Structured Reordering for Modeling Latent Alignments in Sequence Transduction

1 code implementation NeurIPS 2021 Bailin Wang, Mirella Lapata, Ivan Titov

Despite success in many domains, neural models struggle in settings where train and test examples are drawn from different distributions.

Machine Translation Semantic Parsing +2

Editing Factual Knowledge in Language Models

3 code implementations EMNLP 2021 Nicola De Cao, Wilker Aziz, Ivan Titov

We present KnowledgeEditor, a method which can be used to edit this knowledge and, thus, fix 'bugs' or unexpected predictions without the need for expensive re-training or fine-tuning.

Fact Checking Meta-Learning +1

Sparse Attention with Linear Units

3 code implementations EMNLP 2021 Biao Zhang, Ivan Titov, Rico Sennrich

Recently, it has been argued that encoder-decoder models can be made more interpretable by replacing the softmax function in the attention with its sparse variants.

Machine Translation Translation +1

Learning from Executions for Semantic Parsing

1 code implementation NAACL 2021 Bailin Wang, Mirella Lapata, Ivan Titov

Based on the observation that programs which correspond to NL utterances must be always executable, we propose to encourage a parser to generate executable programs for unlabeled utterances.

Semantic Parsing

An Empirical Study of Compound PCFGs

2 code implementations EACL (AdaptNLP) 2021 Yanpeng Zhao, Ivan Titov

Compound probabilistic context-free grammars (C-PCFGs) have recently established a new state of the art for unsupervised phrase-structure grammar induction.

Sentence

Neutron study of magnetic correlations in rare-earth-free Mn-Bi magnets

no code implementations23 Nov 2020 Artem Malyeyev, Ivan Titov, Philipp Bender, Mathias Bersweiler, Vitaliy Pipich, Sebastian Mühlbauer, Semih Ener, Oliver Gutfleisch, Andreas Michels

We report the results of an unpolarized small-angle neutron scattering (SANS) study on Mn-Bi-based rare-earth-free permanent magnets.

Materials Science

Fast Interleaved Bidirectional Sequence Generation

1 code implementation WMT (EMNLP) 2020 Biao Zhang, Ivan Titov, Rico Sennrich

Instead of assuming independence between neighbouring tokens (semi-autoregressive decoding, SA), we take inspiration from bidirectional sequence generation and introduce a decoder that generates target words from the left-to-right and right-to-left directions simultaneously.

Document Summarization Machine Translation

A Differentiable Relaxation of Graph Segmentation and Alignment for AMR Parsing

no code implementations EMNLP 2021 Chunchuan Lyu, Shay B. Cohen, Ivan Titov

In contrast, we treat both alignment and segmentation as latent variables in our model and induce them as part of end-to-end training.

AMR Parsing Segmentation +1

Meta-Learning for Domain Generalization in Semantic Parsing

no code implementations NAACL 2021 Bailin Wang, Mirella Lapata, Ivan Titov

The importance of building semantic parsers which can be applied to new domains and generate programs unseen at training has long been acknowledged, and datasets testing out-of-domain performance are becoming increasingly available.

Domain Generalization Meta-Learning +1

Analyzing the Source and Target Contributions to Predictions in Neural Machine Translation

1 code implementation ACL 2021 Elena Voita, Rico Sennrich, Ivan Titov

We find that models trained with more data tend to rely on source information more and to have more sharp token contributions; the training process is non-monotonic with several stages of different nature.

Language Modelling Machine Translation +2

Adaptive Feature Selection for End-to-End Speech Translation

1 code implementation Findings of the Association for Computational Linguistics 2020 Biao Zhang, Ivan Titov, Barry Haddow, Rico Sennrich

Information in speech signals is not evenly distributed, making it an additional challenge for end-to-end (E2E) speech translation (ST) to learn to focus on informative features.

Data Augmentation feature selection +1

Visually Grounded Compound PCFGs

1 code implementation EMNLP 2020 Yanpeng Zhao, Ivan Titov

In this work, we study visually grounded grammar induction and learn a constituency parser from both unlabeled text and its visual groundings.

Constituency Grammar Induction Language Modelling

Unsupervised Transfer of Semantic Role Models from Verbal to Nominal Domain

1 code implementation1 May 2020 Yanpeng Zhao, Ivan Titov

Nominal roles are not labeled in the training data, and the learning objective instead pushes the labeler to assign roles predictive of the arguments.

Semantic Role Labeling Sentence

Preventing Posterior Collapse with Levenshtein Variational Autoencoder

no code implementations30 Apr 2020 Serhii Havrylov, Ivan Titov

Variational autoencoders (VAEs) are a standard framework for inducing latent variable models that have been shown effective in learning text representations as well as in text generation.

Sentence Text Generation

Few-Shot Learning for Opinion Summarization

1 code implementation EMNLP 2020 Arthur Bražinskas, Mirella Lapata, Ivan Titov

In this work, we show that even a handful of summaries is sufficient to bootstrap generation of the summary text with all expected properties, such as writing style, informativeness, fluency, and sentiment preservation.

Few-Shot Learning Informativeness +2

Improving Massively Multilingual Neural Machine Translation and Zero-Shot Translation

3 code implementations ACL 2020 Biao Zhang, Philip Williams, Ivan Titov, Rico Sennrich

Massively multilingual models for neural machine translation (NMT) are theoretically attractive, but often underperform bilingual models and deliver poor zero-shot translations.

Machine Translation NMT +1

On Sparsifying Encoder Outputs in Sequence-to-Sequence Models

1 code implementation Findings (ACL) 2021 Biao Zhang, Ivan Titov, Rico Sennrich

Inspired by these observations, we explore the feasibility of specifying rule-based patterns that mask out encoder outputs based on information such as part-of-speech tags, word frequency and word position.

Document Summarization Machine Translation

Information-Theoretic Probing with Minimum Description Length

2 code implementations EMNLP 2020 Elena Voita, Ivan Titov

Instead, we propose an alternative to the standard probes, information-theoretic probing with minimum description length (MDL).

Graph Convolutions over Constituent Trees for Syntax-Aware Semantic Role Labeling

1 code implementation EMNLP 2020 Diego Marcheggiani, Ivan Titov

Semantic role labeling (SRL) is the task of identifying predicates and labeling argument spans with semantic roles.

Semantic Role Labeling

Learning Semantic Parsers from Denotations with Latent Structured Alignments and Abstract Programs

1 code implementation IJCNLP 2019 Bailin Wang, Ivan Titov, Mirella Lapata

Semantic parsing aims to map natural language utterances onto machine interpretable meaning representations, aka programs whose execution against a real-world environment produces a denotation.

Inductive Bias Semantic Parsing

Semantic Role Labeling with Iterative Structure Refinement

1 code implementation IJCNLP 2019 Chunchuan Lyu, Shay B. Cohen, Ivan Titov

Modern state-of-the-art Semantic Role Labeling (SRL) methods rely on expressive sentence encoders (e. g., multi-layer LSTMs) but tend to model only local (if any) interactions between individual argument labeling decisions.

Semantic Role Labeling Sentence

The Bottom-up Evolution of Representations in the Transformer: A Study with Machine Translation and Language Modeling Objectives

no code implementations IJCNLP 2019 Elena Voita, Rico Sennrich, Ivan Titov

In this work, we use canonical correlation analysis and mutual information estimators to study how information flows across Transformer layers and how this process depends on the choice of learning objective.

Language Modelling Machine Translation +2

Widening the Representation Bottleneck in Neural Machine Translation with Lexical Shortcuts

1 code implementation WS 2019 Denis Emelin, Ivan Titov, Rico Sennrich

The transformer is a state-of-the-art neural translation model that uses attention to iteratively refine lexical representations with information drawn from the surrounding context.

Machine Translation Translation

Learning Latent Trees with Stochastic Perturbations and Differentiable Dynamic Programming

1 code implementation ACL 2019 Caio Corro, Ivan Titov

We treat projective dependency trees as latent variables in our probabilistic model and induce them in such a way as to be beneficial for a downstream task, without relying on any direct tree supervision.

Natural Language Inference Sentiment Analysis

Interpretable Neural Predictions with Differentiable Binary Variables

1 code implementation ACL 2019 Jasmijn Bastings, Wilker Aziz, Ivan Titov

The success of neural networks comes hand in hand with a desire for more interpretability.

Distant Learning for Entity Linking with Automatic Noise Detection

1 code implementation ACL 2019 Phong Le, Ivan Titov

As the learning signal is weak and our surrogate labels are noisy, we introduce a noise detection component in our model: it lets the model detect and disregard examples which are likely to be noisy.

Entity Linking

When a Good Translation is Wrong in Context: Context-Aware Machine Translation Improves on Deixis, Ellipsis, and Lexical Cohesion

1 code implementation ACL 2019 Elena Voita, Rico Sennrich, Ivan Titov

Though machine translation errors caused by the lack of context beyond one sentence have long been acknowledged, the development of context-aware NMT systems is hampered by several problems.

Machine Translation NMT +2

Obfuscation for Privacy-preserving Syntactic Parsing

1 code implementation WS 2020 Zhifeng Hu, Serhii Havrylov, Ivan Titov, Shay B. Cohen

We introduce an idea for a privacy-preserving transformation on natural language data, inspired by homomorphic encryption.

Privacy Preserving Sentence

Block Neural Autoregressive Flow

4 code implementations9 Apr 2019 Nicola De Cao, Ivan Titov, Wilker Aziz

Recently, as an alternative to hand-crafted bijections, Huang et al. (2018) proposed neural autoregressive flow (NAF) which is a universal approximator for density functions.

Density Estimation Normalising Flows

Modeling Latent Sentence Structure in Neural Machine Translation

no code implementations18 Jan 2019 Jasmijn Bastings, Wilker Aziz, Ivan Titov, Khalil Sima'an

Recently it was shown that linguistic structure predicted by a supervised parser can be beneficial for neural machine translation (NMT).

Machine Translation NMT +2

Question Answering by Reasoning Across Documents with Graph Convolutional Networks

1 code implementation NAACL 2019 Nicola De Cao, Wilker Aziz, Ivan Titov

Most research in reading comprehension has focused on answering questions based on individual documents or even single paragraphs.

Question Answering Reading Comprehension

Context-Aware Neural Machine Translation Learns Anaphora Resolution

no code implementations ACL 2018 Elena Voita, Pavel Serdyukov, Rico Sennrich, Ivan Titov

Standard machine translation systems process sentences in isolation and hence ignore extra-sentential information, even though extended context can both prevent mistakes in ambiguous cases and improve translation coherence.

Machine Translation Translation

AMR Parsing as Graph Prediction with Latent Alignment

2 code implementations ACL 2018 Chunchuan Lyu, Ivan Titov

AMR parsing is challenging partly due to the lack of annotated alignments between nodes in the graphs and words in the corresponding sentences.

AMR Parsing Sentence

Improving Entity Linking by Modeling Latent Relations between Mentions

2 code implementations ACL 2018 Phong Le, Ivan Titov

Entity linking involves aligning textual mentions of named entities to their corresponding entries in a knowledge base.

Entity Linking

Exploiting Semantics in Neural Machine Translation with Graph Convolutional Networks

no code implementations NAACL 2018 Diego Marcheggiani, Jasmijn Bastings, Ivan Titov

Semantic representations have long been argued as potentially useful for enforcing meaning preservation and improving generalization performance of machine translation methods.

Machine Translation Sentence +1

Embedding Words as Distributions with a Bayesian Skip-gram Model

1 code implementation COLING 2018 Arthur Bražinskas, Serhii Havrylov, Ivan Titov

Rather than assuming that a word embedding is fixed across the entire text collection, as in standard word embedding methods, in our Bayesian model we generate it from a word-specific prior density for each occurrence of a given word.

Emergence of Language with Multi-agent Games: Learning to Communicate with Sequences of Symbols

no code implementations NeurIPS 2017 Serhii Havrylov, Ivan Titov

Learning to communicate through interaction, rather than relying on explicit supervision, is often considered a prerequisite for developing a general AI.

Graph Convolutional Encoders for Syntax-aware Neural Machine Translation

no code implementations EMNLP 2017 Jasmijn Bastings, Ivan Titov, Wilker Aziz, Diego Marcheggiani, Khalil Sima'an

We present a simple and effective approach to incorporating syntactic structure into neural attention-based encoder-decoder models for machine translation.

Machine Translation Translation

Optimizing Differentiable Relaxations of Coreference Evaluation Metrics

1 code implementation CONLL 2017 Phong Le, Ivan Titov

Coreference evaluation metrics are hard to optimize directly as they are non-differentiable functions, not easily decomposable into elementary decisions.

Imitation Learning reinforcement-learning +1

A Simple and Accurate Syntax-Agnostic Neural Model for Dependency-based Semantic Role Labeling

2 code implementations CONLL 2017 Diego Marcheggiani, Anton Frolov, Ivan Titov

However, when automatically predicted part-of-speech tags are provided as input, it substantially outperforms all previous local models and approaches the best reported results on the English CoNLL-2009 dataset.

Semantic Role Labeling

Bilingual Learning of Multi-sense Embeddings with Discrete Autoencoders

1 code implementation NAACL 2016 Simon Šuster, Ivan Titov, Gertjan van Noord

We present an approach to learning multi-sense word embeddings relying both on monolingual and bilingual information.

Sentence Word Embeddings

Adapting to All Domains at Once: Rewarding Domain Invariance in SMT

1 code implementation TACL 2016 Hoang Cuong, Khalil Sima{'}an, Ivan Titov

Existing work on domain adaptation for statistical machine translation has consistently assumed access to a small sample from the test distribution (target domain) at training time.

Domain Adaptation Machine Translation +1

Word Representations, Tree Models and Syntactic Functions

1 code implementation31 Aug 2015 Simon Šuster, Gertjan van Noord, Ivan Titov

Word representations induced from models with discrete latent variables (e. g.\ HMMs) have been shown to be beneficial in many NLP applications.

named-entity-recognition Named Entity Recognition +2

Inducing Semantic Representation from Text by Jointly Predicting and Factorizing Relations

no code implementations19 Dec 2014 Ivan Titov, Ehsan Khoddam

In this work, we propose a new method to integrate two recent lines of work: unsupervised induction of shallow semantics (e. g., semantic roles) and factorization of relations in text and knowledge bases.

Semantic Role Labeling

Learning Semantic Script Knowledge with Event Embeddings

no code implementations18 Dec 2013 Ashutosh Modi, Ivan Titov

Induction of common sense knowledge about prototypical sequences of events has recently received much attention.

Common Sense Reasoning

Cannot find the paper you are looking for? You can Submit a new open access paper.