Search Results for author: Romain Robbes

Found 10 papers, 6 papers with code

Out of Context: How important is Local Context in Neural Program Repair?

1 code implementation8 Dec 2023 Julian Aron Prenner, Romain Robbes

Our results indicate that overall repair success increases with the size of the local context (albeit not for all bug types) and confirm the common practice that roughly 50-60% of the input window should be used for context leading the bug.

Program Repair

INSPECT: Intrinsic and Systematic Probing Evaluation for Code Transformers

1 code implementation8 Dec 2023 Anjan Karmakar, Romain Robbes

We find that models that incorporate some structural information (such as GraphCodeBERT) have a better representation of source code characteristics.

Code Completion Language Modelling

RunBugRun -- An Executable Dataset for Automated Program Repair

no code implementations3 Apr 2023 Julian Aron Prenner, Romain Robbes

With this dataset we follow several goals: we want to lift Neural Program Repair beyond fully static code representations, foster the use of execution-based features and, by including several different languages, counterbalance the predominance of Java in the current landscape of APR datasets and benchmarks.

Program Repair

JEMMA: An Extensible Java Dataset for ML4Code Applications

1 code implementation18 Dec 2022 Anjan Karmakar, Miltiadis Allamanis, Romain Robbes

To demonstrate the utility of the dataset, we also report results from two empirical studies on our data, ultimately showing that significant work lies ahead in the design of context-aware source code models that can reason over a broader network of source code entities in a software project, the very task that JEMMA is designed to help with.

Codex Hacks HackerRank: Memorization Issues and a Framework for Code Synthesis Evaluation

no code implementations6 Dec 2022 Anjan Karmakar, Julian Aron Prenner, Marco D'Ambros, Romain Robbes

In this work, we evaluate the code synthesis capabilities of the Codex model based on a set of 115 Python problem statements from a popular competitive programming portal: HackerRank.

Memorization

What do pre-trained code models know about code?

1 code implementation IEEE/ACM International Conference on Automated Software Engineering (ASE) 2021 Anjan Karmakar, Romain Robbes

Pre-trained models of code built on the transformer architecture have performed well on software engineering (SE) tasks such as predictive code generation, code summarization, among others.

Open-Ended Question Answering

GLUECode: A Benchmark for Source Code Machine Learning Models

no code implementations1 Jan 2021 Anjan Karmakar, Julian Aron Prenner, Miltiadis Allamanis, Romain Robbes

To address this, we present GLUECode, Global and Local Understanding Evaluation of Code, a benchmark of diverse tasks to evaluate machine learning models of source code.

BIG-bench Machine Learning

Big Code != Big Vocabulary: Open-Vocabulary Models for Source Code

2 code implementations17 Mar 2020 Rafael-Michael Karampatsis, Hlib Babii, Romain Robbes, Charles Sutton, Andrea Janes

Statistical language modeling techniques have successfully been applied to large source code corpora, yielding a variety of new software development tools, such as tools for code suggestion, improving readability, and API migration.

Software Engineering

Modeling Vocabulary for Big Code Machine Learning

no code implementations3 Apr 2019 Hlib Babii, Andrea Janes, Romain Robbes

We show that a subset of decisions have decisive characteristics, allowing to train accurate Neural Language Models quickly on a large corpus of 10, 106 projects.

BIG-bench Machine Learning

Cannot find the paper you are looking for? You can Submit a new open access paper.