Part of Advances in Neural Information Processing Systems 4 (NIPS 1991)
The localized linear discriminant network (LLDN) has been designed to address classification problems containing relatively closely spaced data from different classes (encounter zones , the accuracy problem ). Locally trained hyper(cid:173) plane segments are an effective way to define the decision boundaries for these regions . The LLD uses a modified perceptron training algorithm for effective discovery of separating hyperplane/sigmoid units within narrow boundaries. The basic unit of the network is the discriminant receptive field (DRF) which combines the LLD function with Gaussians representing the dispersion of the local training data with respect to the hyperplane. The DRF implements a local distance mea(cid:173) sure , and obtains the benefits of networks oflocalized units . A constructive algorithm for the two-class case is described which incorporates DRF's into the hidden layer to solve local discrimination problems. The output unit produces a smoothed, piecewise linear decision boundary. Preliminary results indicate the ability of the LLDN to efficiently achieve separation when boundaries are narrow and complex, in cases where both the "standard" multilayer perceptron (MLP) and k-nearest neighbor (KNN) yield high error rates on training data.
1 The LLD Training Algorithm and DRF Generation
The LLD is defined by the hyperplane normal vector V and its "midpoint" M (a translated origin  near the center of gravity of the training data in feature space). Incremental corrections to V and M accrue for each training token feature vector Y j in the training set, as iIlustrated in figure 1 (exaggerated magnitudes). The surface of the hyperplane is appropriately moved either towards or away from Yj by rotating V, and shifting M along