public class RBFNetwork<T> extends Object implements Classifier<T>
In its basic form, radial basis function network is in the form
y(x) = Σ wi φ(||x-ci||)
where the approximating function y(x) is represented as a sum of N radial basis functions φ, each associated with a different center ci, and weighted by an appropriate coefficient wi. For distance, one usually chooses Euclidean distance. The weights wi can be estimated using the matrix methods of linear least squares, because the approximating function is linear in the weights.
The centers ci can be randomly selected from training data, or learned by some clustering method (e.g. k-means), or learned together with weight parameters undergo a supervised learning processing (e.g. error-correction learning).
The popular choices for φ comprise the Gaussian function and the so called thin plate splines. The advantage of the thin plate splines is that their conditioning is invariant under scalings. Gaussian, multi-quadric and inverse multi-quadric are infinitely smooth and and involve a scale or shape parameter, r0 > 0. Decreasing r0 tends to flatten the basis function. For a given function, the quality of approximation may strongly depend on this parameter. In particular, increasing r0 has the effect of better conditioning (the separation distance of the scaled points increases).
A variant on RBF networks is normalized radial basis function (NRBF) networks, in which we require the sum of the basis functions to be unity. NRBF arises more naturally from a Bayesian statistical perspective. However, there is no evidence that either the NRBF method is consistently superior to the RBF method, or vice versa.
SVMs with Gaussian kernel have similar structure as RBF networks with Gaussian radial basis functions. However, the SVM approach "automatically" solves the network complexity problem since the size of the hidden layer is obtained as the result of the QP procedure. Hidden neurons and support vectors correspond to each other, so the center problems of the RBF network is also solved, as the support vectors serve as the basis function centers. It was reported that with similar number of support vectors/centers, SVM shows better generalization performance than RBF network when the training data size is relatively small. On the other hand, RBF network gives better generalization performance than SVM on large training data.
RadialBasisFunction
,
SVM
,
NeuralNetwork
Modifier and Type | Class and Description |
---|---|
static class |
RBFNetwork.Trainer<T>
Trainer for RBF networks.
|
Constructor and Description |
---|
RBFNetwork(T[] x,
int[] y,
Metric<T> distance,
RadialBasisFunction[] rbf,
T[] centers)
Constructor.
|
RBFNetwork(T[] x,
int[] y,
Metric<T> distance,
RadialBasisFunction[] rbf,
T[] centers,
boolean normalized)
Constructor.
|
RBFNetwork(T[] x,
int[] y,
Metric<T> distance,
RadialBasisFunction rbf,
T[] centers)
Constructor.
|
RBFNetwork(T[] x,
int[] y,
Metric<T> distance,
RadialBasisFunction rbf,
T[] centers,
boolean normalized)
Constructor.
|
public RBFNetwork(T[] x, int[] y, Metric<T> distance, RadialBasisFunction rbf, T[] centers)
x
- training samples.y
- training labels in [0, k), where k is the number of classes.distance
- the distance metric functor.rbf
- the radial basis function.centers
- the centers of RBF functions.public RBFNetwork(T[] x, int[] y, Metric<T> distance, RadialBasisFunction[] rbf, T[] centers)
x
- training samples.y
- training labels in [0, k), where k is the number of classes.distance
- the distance metric functor.rbf
- the radial basis function.centers
- the centers of RBF functions.public RBFNetwork(T[] x, int[] y, Metric<T> distance, RadialBasisFunction rbf, T[] centers, boolean normalized)
x
- training samples.y
- training labels in [0, k), where k is the number of classes.distance
- the distance metric functor.rbf
- the radial basis functions.centers
- the centers of RBF functions.normalized
- true for the normalized RBF network.public RBFNetwork(T[] x, int[] y, Metric<T> distance, RadialBasisFunction[] rbf, T[] centers, boolean normalized)
x
- training samples.y
- training labels in [0, k), where k is the number of classes.distance
- the distance metric functor.rbf
- the radial basis functions.centers
- the centers of RBF functions.normalized
- true for the normalized RBF network.public int predict(T x)
Classifier
predict
in interface Classifier<T>
x
- the instance to be classified.public int predict(T x, double[] posteriori)
predict
in interface Classifier<T>
x
- the instance to be classified.posteriori
- the array to store a posteriori probabilities on output.Copyright © 2015. All rights reserved.