1 code implementation • 8 Dec 2023 • Julian Aron Prenner, Romain Robbes
Our results indicate that overall repair success increases with the size of the local context (albeit not for all bug types) and confirm the common practice that roughly 50-60% of the input window should be used for context leading the bug.
1 code implementation • 8 Dec 2023 • Anjan Karmakar, Romain Robbes
We find that models that incorporate some structural information (such as GraphCodeBERT) have a better representation of source code characteristics.
no code implementations • 3 Apr 2023 • Julian Aron Prenner, Romain Robbes
With this dataset we follow several goals: we want to lift Neural Program Repair beyond fully static code representations, foster the use of execution-based features and, by including several different languages, counterbalance the predominance of Java in the current landscape of APR datasets and benchmarks.
1 code implementation • 18 Dec 2022 • Anjan Karmakar, Miltiadis Allamanis, Romain Robbes
To demonstrate the utility of the dataset, we also report results from two empirical studies on our data, ultimately showing that significant work lies ahead in the design of context-aware source code models that can reason over a broader network of source code entities in a software project, the very task that JEMMA is designed to help with.
no code implementations • 6 Dec 2022 • Anjan Karmakar, Julian Aron Prenner, Marco D'Ambros, Romain Robbes
In this work, we evaluate the code synthesis capabilities of the Codex model based on a set of 115 Python problem statements from a popular competitive programming portal: HackerRank.
1 code implementation • IEEE/ACM International Conference on Automated Software Engineering (ASE) 2021 • Anjan Karmakar, Romain Robbes
Pre-trained models of code built on the transformer architecture have performed well on software engineering (SE) tasks such as predictive code generation, code summarization, among others.
no code implementations • 1 Jan 2021 • Anjan Karmakar, Julian Aron Prenner, Miltiadis Allamanis, Romain Robbes
To address this, we present GLUECode, Global and Local Understanding Evaluation of Code, a benchmark of diverse tasks to evaluate machine learning models of source code.
1 code implementation • 7 Oct 2020 • Paul Ralph, Nauman bin Ali, Sebastian Baltes, Domenico Bianculli, Jessica Diaz, Yvonne Dittrich, Neil Ernst, Michael Felderer, Robert Feldt, Antonio Filieri, Breno Bernard Nicolau de França, Carlo Alberto Furia, Greg Gay, Nicolas Gold, Daniel Graziotin, Pinjia He, Rashina Hoda, Natalia Juristo, Barbara Kitchenham, Valentina Lenarduzzi, Jorge Martínez, Jorge Melegati, Daniel Mendez, Tim Menzies, Jefferson Molleri, Dietmar Pfahl, Romain Robbes, Daniel Russo, Nyyti Saarimäki, Federica Sarro, Janet Siegmund, Diomidis Spinellis, Miroslaw Staron, Klaas Stol, Margaret-Anne Storey, Davide Taibi, Damian Tamburri, Marco Torchiano, Christoph Treude, Burak Turhan, XiaoFeng Wang, Sira Vegas
Empirical Standards are natural-language models of a scientific community's expectations for a specific kind of study (e. g. a questionnaire survey).
Software Engineering General Literature
2 code implementations • 17 Mar 2020 • Rafael-Michael Karampatsis, Hlib Babii, Romain Robbes, Charles Sutton, Andrea Janes
Statistical language modeling techniques have successfully been applied to large source code corpora, yielding a variety of new software development tools, such as tools for code suggestion, improving readability, and API migration.
Software Engineering
no code implementations • 3 Apr 2019 • Hlib Babii, Andrea Janes, Romain Robbes
We show that a subset of decisions have decisive characteristics, allowing to train accurate Neural Language Models quickly on a large corpus of 10, 106 projects.