Search Results for author: Miguel A. Gutiérrez-Naranjo

Found 7 papers, 2 papers with code

SIMAP: A simplicial-map layer for neural networks

no code implementations22 Mar 2024 Rocio Gonzalez-Diaz, Miguel A. Gutiérrez-Naranjo, Eduardo Paluzo-Hidalgo

In this paper, we present SIMAP, a novel layer integrated into deep learning models, aimed at enhancing the interpretability of the output.

Trainable and Explainable Simplicial Map Neural Networks

1 code implementation29 May 2023 Eduardo Paluzo-Hidalgo, Miguel A. Gutiérrez-Naranjo, Rocio Gonzalez-Diaz

In this paper, we overcome these issues by proposing an SMNN training procedure based on a support subset of the given dataset and replacing the construction of the convex polytope by a method based on projections to a hypersphere.

Dynamic Price of Parking Service based on Deep Learning

no code implementations11 Jan 2022 Alejandro Luque-Cerpa, Miguel A. Gutiérrez-Naranjo, Miguel Cárdenas-Montes

Major efforts from government bodies in this area include monitoring and forecasting systems, banning more pollutant motor vehicles, and traffic limitations during the periods of low-quality air.

Summary and Distance between Sets of Texts based on Topological Data Analysis

no code implementations19 Dec 2019 Eduardo Paluzo-Hidalgo, Rocio Gonzalez-Diaz, Miguel A. Gutiérrez-Naranjo

In this paper, we use topological data analysis (TDA) tools such as persistent homology, persistent entropy and bottleneck distance, to provide a {\it TDA-based summary} of any given set of texts and a general method for computing a distance between any two literary styles, authors or periods.

Topological Data Analysis

Topology-based Representative Datasets to Reduce Neural Network Training Resources

1 code implementation20 Mar 2019 Rocio Gonzalez-Diaz, Miguel A. Gutiérrez-Naranjo, Eduardo Paluzo-Hidalgo

We prove that the accuracy of the learning process of a neural network on a representative dataset is "similar" to the accuracy on the original dataset when the neural network architecture is a perceptron and the loss function is the mean squared error.

Computational Efficiency

Cannot find the paper you are looking for? You can Submit a new open access paper.