1 code implementation • 17 Apr 2024 • Thomas Vecchiato, Claudio Lucchese, Franco Maria Nardini, Sebastian Bruch
Its objective is to return a set of $k$ data points that are closest to a query point, with its accuracy measured by the proportion of exact nearest neighbors captured in the returned set.
no code implementations • 22 Feb 2024 • Stefano Calzavara, Lorenzo Cazzaro, Claudio Lucchese, Giulio Ermanno Pibiri
Verifiable learning advocates for training machine learning models amenable to efficient security verification.
no code implementations • 15 May 2023 • Simon Koschel, Sebastian Buschjäger, Claudio Lucchese, Katharina Morik
Second, we extend our implementation from ranking models to classification models such as Random Forests.
no code implementations • 15 May 2023 • Sebastian Bruch, Claudio Lucchese, Franco Maria Nardini
We believe that by understanding the fundamentals underpinning these algorithmic and data structure solutions for containing the contentious relationship between efficiency and effectiveness, one can better identify future directions and more efficiently determine the merits of ideas.
no code implementations • 29 Dec 2022 • Matteo Rizzo, Alberto Veneri, Andrea Albarelli, Claudio Lucchese, Marco Nobile, Cristina Conati
EXplainable Artificial Intelligence (XAI) is a vibrant research topic in the artificial intelligence community, with growing interest across methods and domains.
1 code implementation • 27 Sep 2022 • Stefano Calzavara, Lorenzo Cazzaro, Claudio Lucchese, Federico Marcuzzi
We present a new approach to the global fairness verification of tree-based classifiers.
1 code implementation • 1 Jun 2022 • Claudio Lucchese, Franco Maria Nardini, Salvatore Orlando, Raffaele Perego, Alberto Veneri
Interpretable Learning to Rank (LtR) is an emerging field within the research area of explainable AI, aiming at developing intelligible and accurate predictive models.
no code implementations • 29 Dec 2021 • Seyum Assefa Abebe, Claudio Lucchese, Salvatore Orlando
Nowadays Machine Learning (ML) techniques are extensively adopted in many socially sensitive systems, thus requiring to carefully study the fairness of the decisions taken by such systems.
no code implementations • 5 Dec 2021 • Stefano Calzavara, Lorenzo Cazzaro, Claudio Lucchese, Federico Marcuzzi, Salvatore Orlando
In this paper we criticize the robustness measure traditionally employed to assess the performance of machine learning models deployed in adversarial settings.
1 code implementation • 6 May 2021 • Francesco Busolin, Claudio Lucchese, Franco Maria Nardini, Salvatore Orlando, Raffaele Perego, Salvatore Trani
Modern search engine ranking pipelines are commonly based on large machine-learned ensembles of regression trees.
no code implementations • 6 Jul 2020 • Stefano Calzavara, Pietro Ferrara, Claudio Lucchese
Machine learning has proved invaluable for a range of different tasks, yet it also proved vulnerable to evasion attacks, i. e., maliciously crafted perturbations of input data designed to force mispredictions.
no code implementations • 30 Apr 2020 • Claudio Lucchese, Franco Maria Nardini, Salvatore Orlando, Raffaele Perego, Salvatore Trani
In this paper, we investigate the novel problem of \textit{query-level early exiting}, aimed at deciding the profitability of early stopping the traversal of the ranking ensemble for all the candidate documents to be scored for a query, by simply returning a ranking based on the additive scores computed by a limited portion of the ensemble.
no code implementations • 7 Apr 2020 • Stefano Calzavara, Claudio Lucchese, Federico Marcuzzi, Salvatore Orlando
The attacker aims at finding a minimal perturbation of a test instance that changes the model outcome.
1 code implementation • 2 Jul 2019 • Stefano Calzavara, Claudio Lucchese, Gabriele Tolomei, Seyum Assefa Abebe, Salvatore Orlando
Despite its success and popularity, machine learning is now recognized as vulnerable to evasion attacks, i. e., carefully crafted perturbations of test inputs designed to force prediction errors.