Role of biases in on-line learning of two-layer networks

Abstract
The influence of biases on the learning dynamics of a two-layer neural network, a normalized soft-committee machine, is studied for on-line gradient descent learning. Within a statistical mechanics framework, numerical studies show that the inclusion of adjustable biases dramatically alters the learning dynamics found previously. The symmetric phase that has often been predominant in the original model all but disappears for a nondegenerate bias task. The extended model furthermore exhibits a much richer dynamical behavior, e.g., attractive suboptimal symmetric phases even for realizable cases and noiseless data.

This publication has 17 references indexed in Scilit: