Professional Documents
Culture Documents
0 .5
2
When mapped into the feature space (z1, z2) , the two
classes become linearly separable.
Training RBF nets
Typically, the weights of the two layers are determined
separately, i.e. find RBF weights, and then find output layer
weights
Hidden layer
– estimate parameters for each hidden unit k (whose output depends
on distance between input and a stored prototype)
e.g. for Gaussian activation function, estimate parameters: µk, σk2
– This stage involves an Unsupervised training process (no targets available)
Output layer
– set the weights (including bias weights)
– the same as training a single layer perceptron: each unit’s output depends on
weighted sum of inputs,
– using for example, the Delta Rule (DR)
– This stage involves a Supervised training process
Clustering
K-Means Approach
1. Select k multidimensional points to be the
“seeds” or initial centroids for the k clusters to be
formed. Seeds usually selected at random
2. Assign each observation to the cluster with the
nearest seed.
3. Update cluster centroids once all observations
have been assigned.
4. Repeat steps 2 and 3 until changes in cluster
centroids small.
5. Repeat steps 1-5 with new starting seeds. Do this
step 3 to 5 times.
K-Means Illustration – two dimensions
Fine Tuning
Computing the Output Weights
2 5
Type of Activation Function: Gaussian
Input Range: x = [-10:10]
Centers = [-8 -5 -2 0 2 5 8]
Function to be Approximated
30
25
20
15
10
Output
-5
-10
-15
-20
-10 -5 0 5 10
Input
Case 1: Width is chosen to equal 6. This way, the
receptive fields overlap, but no neuron function
covers the entire input space. For proper overlap,
the width parameter needs to be at least equal to the
distance between the input parameters.
Testing the RBF Network, w=6
100
50
0
Output
-50
-100
-150
-15 -10 -5 0 5 10 15
Input
Case 2: Width is chosen to equal 0.2 (too
small). This causes poor generalization
inside the training space
50
0
Output
-50
-100
-150
-15 -10 -5 0 5 10 15
Input
Case 3: Width is chosen to be 200 (too large). This
causes each radial basis function to cover the entire
input space, and is being activated for each input
value. Thus, the network cannot properly learn the
desired mapping.
Testing the RBF Network, w=200
100
50
0
Output
-50
-100
-150
-15 -10 -5 0 5 10 15
Input
XOR Problem
The relationship between the input and the
output of the network can be given by