no code implementations • 5 Feb 2024 • Simone Bombari, Marco Mondelli
Unveiling the reasons behind the exceptional success of transformers requires a better understanding of why attention layers are suitable for NLP tasks.
no code implementations • 20 May 2023 • Simone Bombari, Marco Mondelli
Deep learning models can be vulnerable to recovery attacks, raising privacy concerns to users, and widespread algorithms such as empirical risk minimization (ERM) often do not directly enforce safety guarantees.
1 code implementation • 3 Feb 2023 • Simone Bombari, Shayan Kiyani, Marco Mondelli
However, this "universal" law provides only a necessary condition for robustness, and it is unable to discriminate between models.
no code implementations • 20 May 2022 • Simone Bombari, Mohammad Hossein Amani, Marco Mondelli
The Neural Tangent Kernel (NTK) has emerged as a powerful tool to provide memorization, optimization and generalization guarantees in deep neural networks.
no code implementations • 17 May 2022 • Mohammad Hossein Amani, Simone Bombari, Marco Mondelli, Rattana Pukdee, Stefano Rini
In this paper, we study the compression of a target two-layer neural network with N nodes into a compressed network with M<N nodes.
no code implementations • 30 Mar 2022 • Simone Bombari, Alessandro Achille, Zijian Wang, Yu-Xiang Wang, Yusheng Xie, Kunwar Yashraj Singh, Srikar Appalaraju, Vijay Mahadevan, Stefano Soatto
While bounding general memorization can have detrimental effects on the performance of a trained model, bounding RM does not prevent effective learning.