Generalization ability and information gain of clock-model perceptrons |
| |
Authors: | B. Schottky F. Gerl U. Krey |
| |
Affiliation: | (1) Institut für Physik III, Universität Regensburg, Universitätsstrasse 31, D-93040 Regensburg, Germany |
| |
Abstract: | ![]() We study the generalization abilitygQofQ-state Clock-model perceptrons for (i) Hebbian and for certain Non-Hebbian learning procedures, namely (ii) learning with maximal stability, (iii) zero stability and (iv) optimal generalization, for the case of random training sets. Among other results we find thatgQbehaves quite different in the Hebbian and in the Non-Hebbian cases in the limitQ . E.g. in the Hebbian case for finite ,gQvanishes always 1/Q, whereas in the Non-Hebbian cases considered,gQconverges forQ to a non-trivial continuous functiong ( ), which vanishes for <2, but increases rapidly for >2. This means that for (ii), (iii) and (iv), as a function of atQ= , there is a 2nd-order phase transition from a non-generalizing phase for  2 to a generalizing phase for >2. Different behaviour of the Hebbian and Non-Hebbian cases, respectively, is also observed for the information gain obtained through learning. For the particular case of AdaTron Learning, which is identical to case (ii), we find a geometrical formulation forgQ( ), which is applicable to more general models. |
| |
Keywords: | 87.10.+e 02.50.+s 05.20.-y |
本文献已被 SpringerLink 等数据库收录! |
|