Alessandro Sperduti
University of Pisa
Network
Latest external collaboration on country level. Dive into details by clicking on the dots.
Publication
Featured researches published by Alessandro Sperduti.
Neural Networks | 1993
Alessandro Sperduti; Antonina Starita
Abstract Methods to speed up learning in back propagation and to optimize the network architecture have been recently studied. This paper shows how adaptation of the steepness of the sigmoids during learning treats these two topics in a common framework. The adaptation of the steepness of the sigmoids is obtained by gradient descent. The resulting learning dynamics can be simulated by a standard network with fixed sigmoids and a learning rule whose main component is a gradient descent with adaptive learning parameters. A law linking variation on the weights to variation on the steepness of the sigmoids is discovered. Optimization of units is obtained by introducing a tendency to decay to zero in the steepness values. This decay corresponds to a decay of the sensitivity of the units. Units with low final sensitivity can be removed after a given transformation of the biases of the network. A decreasing initial distribution of the steepness values is suggested to obtain a good compromise between speed of learning and network optimization. Simulation of the proposed procedure has shown an improvement of the mean convergence rate with respect to the standard back propagation and good optimization performance. Several 4-3-1 networks for the four bits parity problem were discovered.
conference on information and knowledge management | 2000
Fabrizio Sebastiani; Alessandro Sperduti; Nicola Valdambrini
We describe an improved boosting algorithm, called {\sc AdaBoost.MH
Applied Intelligence | 2000
Anna Maria Bianucci; Alessandro Sperduti; Antonina Starita
^{KR}
graphics recognition | 1997
Enrico Francesconi; Paolo Frasconi; Marco Gori; Simone Marinai; Jianqing Sheng; Giovanni Soda; Alessandro Sperduti
}, and its application to text categorization. Boosting is a method for supervised learning which has successfully been applied to many different domains, and that has proven one of the best performers in text categorization exercises so far. Boosting is based on the idea of relying on the collective judgment of a committee of classifiers that are trained sequentially. In training the
Neural Networks | 1997
Alessandro Sperduti
i
Archive | 2003
Anna Maria Bianucci; Alessandro Sperduti; Antonina Starita
-th classifier special emphasis is placed on the correct categorization of the training documents which have proven harder for the previously trained classifiers. {\sc AdaBoost.MH
SSPR '96 Proceedings of the 6th International Workshop on Advances in Structural and Syntactical Pattern Recognition | 1996
Alessandro Sperduti; Darya Majidi; Antonina Starita
^{KR}
international conference on artificial neural networks | 2001
Alessandro Sperduti
} is based on the idea to build, at every iteration of the learning phase, not a single classifier but a sub-committee of the
Archive | 2001
Markus Hagenbuchner; Ah Chung Tsoi; Alessandro Sperduti
K
Neural Computation | 2000
Diego Sona; Alessandro Sperduti; Antonina Starita
classifiers which, at that iteration, look the most promising. We report the results of systematic experimentation of this method performed on the standard {\sf Reuters-21578} benchmark. These experiments have shown that {\sc AdaBoost.MH