Monday, August 15, 2011

Neural Network (Part 7) : Cut and Paste Code

Ok - so you don't like tutorials, and would rather just cut and paste some code.
This is for you.
http://www.openprocessing.org/visuals/?visualID=33991

Make sure to select "Source code" when you get there, otherwise it will be quite boring.
Here is an animated gif which shows the program in action.



See BELOW for the WHOLE screenshot so that you don't have to speed read.



If you want to know how it works, then you will have to go back and read part 1 to 6.



  • Neural Network



  • But as you can see from the example above:
    Before the neural network is trained, the outputs are not even close to the expected outputs. After training, the neural network produces the desired results (or very close to it).

    Please note, this neural network also allows more than one output neuron, so you are not limited to single yes no decisions. You can use this neural network to make classifications. You will soon see this with my LED colour sensor.

    Feel free to use this Neural Network in your own projects, but please let me know if you do, just for curiosity sake.

    Update: See below for the Processing Sketch (much easier than going to the open processing site). It is a bit long - but saves you from having to navigate to another site.

    Processing sketch

      1
    2
    3
    4
    5
    6
    7
    8
    9
    10
    11
    12
    13
    14
    15
    16
    17
    18
    19
    20
    21
    22
    23
    24
    25
    26
    27
    28
    29
    30
    31
    32
    33
    34
    35
    36
    37
    38
    39
    40
    41
    42
    43
    44
    45
    46
    47
    48
    49
    50
    51
    52
    53
    54
    55
    56
    57
    58
    59
    60
    61
    62
    63
    64
    65
    66
    67
    68
    69
    70
    71
    72
    73
    74
    75
    76
    77
    78
    79
    80
    81
    82
    83
    84
    85
    86
    87
    88
    89
    90
    91
    92
    93
    94
    95
    96
    97
    98
    99
    100
    101
    102
    103
    104
    105
    106
    107
    108
    109
    110
    111
    112
    113
    114
    115
    116
    117
    118
    119
    120
    121
    122
    123
    124
    125
    126
    127
    128
    129
    130
    131
    132
    133
    134
    135
    136
    137
    138
    139
    140
    141
    142
    143
    144
    145
    146
    147
    148
    149
    150
    151
    152
    153
    154
    155
    156
    157
    158
    159
    160
    161
    162
    163
    164
    165
    166
    167
    168
    169
    170
    171
    172
    173
    174
    175
    176
    177
    178
    179
    180
    181
    182
    183
    184
    185
    186
    187
    188
    189
    190
    191
    192
    193
    194
    195
    196
    197
    198
    199
    200
    201
    202
    203
    204
    205
    206
    207
    208
    209
    210
    211
    212
    213
    214
    215
    216
    217
    218
    219
    220
    221
    222
    223
    224
    225
    226
    227
    228
    229
    230
    231
    232
    233
    234
    235
    236
    237
    238
    239
    240
    241
    242
    243
    244
    245
    246
    247
    248
    249
    250
    251
    252
    253
    254
    255
    256
    257
    258
    259
    260
    261
    262
    263
    264
    265
    266
    267
    268
    269
    270
    271
    272
    273
    274
    275
    276
    277
    278
    279
    280
    281
    282
    283
    284
    285
    286
    287
    288
    289
    290
    291
    292
    293
    294
    295
    296
    297
    298
    299
    300
    301
    302
    303
    304
    305
    306
    307
    308
    309
    310
    311
    312
    313
    314
    315
    316
    317
    318
    319
    320
    321
    322
    323
    324
    325
    326
    327
    328
    329
    330
    331
    332
    333
    334
    335
    336
    337
    338
    339
    340
    341
    342
    343
    344
    345
    346
    347
    348
    349
    350
    351
    352
    353
    354
    355
    356
    357
    358
    359
    360
    361
    362
    363
    364
    365
    366
    367
    368
    369
    370
    371
    372
    373
    374
    375
    376
    377
    378
    379
    380
    381
    382
    383
    384
    385
    386
    387
    388
    389
    390
    391
    392
    393
    394
    395
    396
    397
    398
    399
    400
    401
    402
    403
    404
    405
    406
    407
    408
    409
    410
    411
    412
    413
    414
    415
    416
    417
    418
    419
    420
    421
    422
    423
    424
    425
    426
    427
    428
    429
    430
    431
    432
    433
    434
    435
    436
    437
    438
    439
    440
    441
    442
    443
    444
    445
    446
    447
    448
    449
    450
    451
    452
    453
    454
    455
    456
    457
    458
    459
    460
    461
    462
    463
    464
    465
    466
    467
    468
    469
    470
    471
    472
    473
    474
    475
    476
    477
    478
    479
    480
    481
    482
    483
    484
    485
    486
    487
    488
    489
    490
    491
    492
    493
    494
    495
    496
    497
    498
    499
    500
    501
    502
    503
    504
    505
    506
    507
    508
    509
    510
    511
    512
    513
    514
    515
    516
    517
    518
    519
    520
    521
    522
    523
    524
    525
    526
    527
    528
    529
    530
    531
    532
    533
    534
    535
    536
    537
    538
    539
    540
    541
    542
    543
    544
    545
    546
    547
    548
    549
    550
    551
    552
    553
    554
    555
    556
    557
    558
    559
    560
    561
    562
    563
    564
    565
    566
    567
    568
    569
    570
    571
    572
    573
    /*Neural Network created by ScottC on 15th Aug 2011

    Please visit my blog for a detailed explanation of my Neural Network
    http://arduinobasics.blogspot.com/p/arduinoprojects.html

    */



    void setup(){
    ArrayList myTrainingInputs = new ArrayList();
    ArrayList myTrainingOutputs = new ArrayList();

    float[] myInputsA={0,0};
    float[] myInputsB={0,1};
    float[] myInputsC={1,0};
    float[] myInputsD={1,1};
    float[] myOutputsA={1};
    float[] myOutputsB={0};


    println("TRAINING DATA");
    println("--------------------------------------------");
    myTrainingInputs.add(myInputsA);
    myTrainingOutputs.add(myOutputsA);
    println("INPUTS= " + myInputsA[0] + ", " + myInputsA[1] + "; Expected output = " + myOutputsA[0]);
    myTrainingInputs.add(myInputsB);
    myTrainingOutputs.add(myOutputsB);
    println("INPUTS= " + myInputsB[0] + ", " + myInputsB[1] + "; Expected output = " + myOutputsB[0]);
    myTrainingInputs.add(myInputsC);
    myTrainingOutputs.add(myOutputsB);
    println("INPUTS= " + myInputsC[0] + ", " + myInputsC[1] + "; Expected output = " + myOutputsB[0]);
    myTrainingInputs.add(myInputsD);
    myTrainingOutputs.add(myOutputsA);
    println("INPUTS= " + myInputsD[0] + ", " + myInputsD[1] + "; Expected output = " + myOutputsA[0]);
    println("--------------------------------------------");

    NeuralNetwork NN = new NeuralNetwork();
    NN.addLayer(2,2);
    NN.addLayer(2,1);

    println("Before Training");
    float[] myInputDataA1={0,0};
    NN.processInputsToOutputs(myInputDataA1);
    float[] myOutputDataA1={};
    myOutputDataA1=NN.getOutputs();
    println("Feed Forward: INPUT = 0,0; OUTPUT=" + myOutputDataA1[0]);

    float[] myInputDataB1={0,1};
    NN.processInputsToOutputs(myInputDataB1);
    float[] myOutputDataB1={};
    myOutputDataB1=NN.getOutputs();
    println("Feed Forward: INPUT = 0,1; OUTPUT=" + myOutputDataB1[0]);

    float[] myInputDataC1={1,0};
    NN.processInputsToOutputs(myInputDataC1);
    float[] myOutputDataC1={};
    myOutputDataC1=NN.getOutputs();
    println("Feed Forward: INPUT = 1,0; OUTPUT=" + myOutputDataC1[0]);

    float[] myInputDataD1={1,1};
    NN.processInputsToOutputs(myInputDataD1);
    float[] myOutputDataD1={};
    myOutputDataD1=NN.getOutputs();
    println("Feed Forward: INPUT = 1,1; OUTPUT=" + myOutputDataD1[0]);

    println("");
    println("--------------------------------------------");

    println("Begin Training");
    NN.autoTrainNetwork(myTrainingInputs,myTrainingOutputs,0.0001,500000);
    println("");
    println("End Training");
    println("");
    println("--------------------------------------------");
    println("Test the neural network");
    float[] myInputDataA2={0,0};
    NN.processInputsToOutputs(myInputDataA2);
    float[] myOutputDataA2={};
    myOutputDataA2=NN.getOutputs();
    println("Feed Forward: INPUT = 0,0; OUTPUT=" + myOutputDataA2[0]);

    float[] myInputDataB2={0,1};
    NN.processInputsToOutputs(myInputDataB2);
    float[] myOutputDataB2={};
    myOutputDataB2=NN.getOutputs();
    println("Feed Forward: INPUT = 0,1; OUTPUT=" + myOutputDataB2[0]);

    float[] myInputDataC2={1,0};
    NN.processInputsToOutputs(myInputDataC2);
    float[] myOutputDataC2={};
    myOutputDataC2=NN.getOutputs();
    println("Feed Forward: INPUT = 1,0; OUTPUT=" + myOutputDataC2[0]);

    float[] myInputDataD2={1,1};
    NN.processInputsToOutputs(myInputDataD2);
    float[] myOutputDataD2={};
    myOutputDataD2=NN.getOutputs();
    println("Feed Forward: INPUT = 1,1; OUTPUT=" + myOutputDataD2[0]);


    }


    /* ---------------------------------------------------------------------
    A connection determines how much of a signal is passed through to the neuron.
    -------------------------------------------------------------------- */

    class Connection{
    float connEntry;
    float weight;
    float connExit;

    //This is the default constructor for an Connection
    Connection(){
    randomiseWeight();
    }

    //A custom weight for this Connection constructor
    Connection(float tempWeight){
    setWeight(tempWeight);
    }

    //Function to set the weight of this connection
    void setWeight(float tempWeight){
    weight=tempWeight;
    }

    //Function to randomise the weight of this connection
    void randomiseWeight(){
    setWeight(random(2)-1);
    }

    //Function to calculate and store the output of this Connection
    float calcConnExit(float tempInput){
    connEntry = tempInput;
    connExit = connEntry * weight;
    return connExit;
    }
    }


    /* -----------------------------------------------------------------
    A neuron does all the processing and calculation to convert an input into an output
    --------------------------------------------------------------------- */

    class Neuron{
    Connection[] connections={};
    float bias;
    float neuronInputValue;
    float neuronOutputValue;
    float deltaError;

    //The default constructor for a Neuron
    Neuron(){
    }

    /*The typical constructor of a Neuron - with random Bias and Connection weights */
    Neuron(int numOfConnections){
    randomiseBias();
    for(int i=0; i<numOfConnections; i++){
    Connection conn = new Connection();
    addConnection(conn);
    }
    }

    //Function to add a Connection to this neuron
    void addConnection(Connection conn){
    connections = (Connection[]) append(connections, conn);
    }

    /* Function to return the number of connections associated with this neuron.*/
    int getConnectionCount(){
    return connections.length;
    }

    //Function to set the bias of this Neron
    void setBias(float tempBias){
    bias = tempBias;
    }

    //Function to randomise the bias of this Neuron
    void randomiseBias(){
    setBias(random(1));
    }

    /*Function to convert the inputValue to an outputValue
    Make sure that the number of connEntryValues matches the number of connections */

    float getNeuronOutput(float[] connEntryValues){
    if(connEntryValues.length!=getConnectionCount()){
    println("Neuron Error: getNeuronOutput() : Wrong number of connEntryValues");
    exit();
    }

    neuronInputValue=0;

    /* First SUM all of the weighted connection values (connExit) attached to this neuron. This becomes the neuronInputValue. */
    for(int i=0; i<getConnectionCount(); i++){
    neuronInputValue+=connections[i].calcConnExit(connEntryValues[i]);
    }

    //Add the bias to the Neuron's inputValue
    neuronInputValue+=bias;

    /* Send the inputValue through the activation function to produce the Neuron's outputValue */
    neuronOutputValue=Activation(neuronInputValue);

    //Return the outputValue
    return neuronOutputValue;
    }

    //Activation function
    float Activation(float x){
    float activatedValue = 1 / (1 + exp(-1 * x));
    return activatedValue;
    }

    }

    class Layer{
    Neuron[] neurons = {};
    float[] layerINPUTs={};
    float[] actualOUTPUTs={};
    float[] expectedOUTPUTs={};
    float layerError;
    float learningRate;


    /* This is the default constructor for the Layer */
    Layer(int numberConnections, int numberNeurons){
    /* Add all the neurons and actualOUTPUTs to the layer */
    for(int i=0; i<numberNeurons; i++){
    Neuron tempNeuron = new Neuron(numberConnections);
    addNeuron(tempNeuron);
    addActualOUTPUT();
    }
    }


    /* Function to add an input or output Neuron to this Layer */
    void addNeuron(Neuron xNeuron){
    neurons = (Neuron[]) append(neurons, xNeuron);
    }


    /* Function to get the number of neurons in this layer */
    int getNeuronCount(){
    return neurons.length;
    }


    /* Function to increment the size of the actualOUTPUTs array by one. */
    void addActualOUTPUT(){
    actualOUTPUTs = (float[]) expand(actualOUTPUTs,(actualOUTPUTs.length+1));
    }


    /* Function to set the ENTIRE expectedOUTPUTs array in one go. */
    void setExpectedOUTPUTs(float[] tempExpectedOUTPUTs){
    expectedOUTPUTs=tempExpectedOUTPUTs;
    }


    /* Function to clear ALL values from the expectedOUTPUTs array */
    void clearExpectedOUTPUT(){
    expectedOUTPUTs = (float[]) expand(expectedOUTPUTs, 0);
    }


    /* Function to set the learning rate of the layer */
    void setLearningRate(float tempLearningRate){
    learningRate=tempLearningRate;
    }


    /* Function to set the inputs of this layer */
    void setInputs(float[] tempInputs){
    layerINPUTs=tempInputs;
    }



    /* Function to convert ALL the Neuron input values into Neuron output values in this layer, through a special activation function. */
    void processInputsToOutputs(){

    /* neuronCount is used a couple of times in this function. */
    int neuronCount = getNeuronCount();

    /* Check to make sure that there are neurons in this layer to process the inputs */
    if(neuronCount>0) {
    /* Check to make sure that the number of inputs matches the number of Neuron Connections. */
    if(layerINPUTs.length!=neurons[0].getConnectionCount()){
    println("Error in Layer: processInputsToOutputs: The number of inputs do NOT match the number of Neuron connections in this layer");
    exit();
    } else {
    /* The number of inputs are fine : continue
    Calculate the actualOUTPUT of each neuron in this layer,
    based on their layerINPUTs (which were previously calculated).
    Add the value to the layer's actualOUTPUTs array. */
    for(int i=0; i<neuronCount;i++){
    actualOUTPUTs[i]=neurons[i].getNeuronOutput(layerINPUTs);
    }
    }
    }else{
    println("Error in Layer: processInputsToOutputs: There are no Neurons in this layer");
    exit();
    }
    }


    /* Function to get the error of this layer */
    float getLayerError(){
    return layerError;
    }


    /* Function to set the error of this layer */
    void setLayerError(float tempLayerError){
    layerError=tempLayerError;
    }


    /* Function to increase the layerError by a certain amount */
    void increaseLayerErrorBy(float tempLayerError){
    layerError+=tempLayerError;
    }


    /* Function to calculate and set the deltaError of each neuron in the layer */
    void setDeltaError(float[] expectedOutputData){
    setExpectedOUTPUTs(expectedOutputData);
    int neuronCount = getNeuronCount();
    /* Reset the layer error to 0 before cycling through each neuron */
    setLayerError(0);
    for(int i=0; i<neuronCount;i++){
    neurons[i].deltaError = actualOUTPUTs[i]*(1-actualOUTPUTs[i])*(expectedOUTPUTs[i]-actualOUTPUTs[i]);

    /* Increase the layer Error by the absolute difference between the calculated value (actualOUTPUT) and the expected value (expectedOUTPUT). */
    increaseLayerErrorBy(abs(expectedOUTPUTs[i]-actualOUTPUTs[i]));
    }
    }


    /* Function to train the layer : which uses a training set to adjust the connection weights and biases of the neurons in this layer */
    void trainLayer(float tempLearningRate){
    setLearningRate(tempLearningRate);

    int neuronCount = getNeuronCount();

    for(int i=0; i<neuronCount;i++){
    /* update the bias for neuron[i] */
    neurons[i].bias += (learningRate * 1 * neurons[i].deltaError);

    /* update the weight of each connection for this neuron[i] */
    for(int j=0; j<neurons[i].getConnectionCount(); j++){
    neurons[i].connections[j].weight += (learningRate * neurons[i].connections[j].connEntry * neurons[i].deltaError);
    }
    }
    }
    }

    /* -------------------------------------------------------------
    The Neural Network class is a container to hold and manage all the layers
    ---------------------------------------------------------------- */

    class NeuralNetwork{
    Layer[] layers = {};
    float[] arrayOfInputs={};
    float[] arrayOfOutputs={};
    float learningRate;
    float networkError;
    float trainingError;
    int retrainChances=0;

    NeuralNetwork(){
    /* the default learning rate of a neural network is set to 0.1, which can changed by the setLearningRate(lR) function. */
    learningRate=0.1;
    }



    /* Function to add a Layer to the Neural Network */
    void addLayer(int numConnections, int numNeurons){
    layers = (Layer[]) append(layers, new Layer(numConnections,numNeurons));
    }



    /* Function to return the number of layers in the neural network */
    int getLayerCount(){
    return layers.length;
    }



    /* Function to set the learningRate of the Neural Network */
    void setLearningRate(float tempLearningRate){
    learningRate=tempLearningRate;
    }



    /* Function to set the inputs of the neural network */
    void setInputs(float[] tempInputs){
    arrayOfInputs=tempInputs;
    }



    /* Function to set the inputs of a specified layer */
    void setLayerInputs(float[] tempInputs, int layerIndex){
    if(layerIndex>getLayerCount()-1){
    println("NN Error: setLayerInputs: layerIndex=" + layerIndex + " exceeded limits= " + (getLayerCount()-1));
    } else {
    layers[layerIndex].setInputs(tempInputs);
    }
    }



    /* Function to set the outputs of the neural network */
    void setOutputs(float[] tempOutputs){
    arrayOfOutputs=tempOutputs;
    }



    /* Function to return the outputs of the Neural Network */
    float[] getOutputs(){
    return arrayOfOutputs;
    }



    /* Function to process the Neural Network's input values and convert them to an output pattern using ALL layers in the network */
    void processInputsToOutputs(float[] tempInputs){
    setInputs(tempInputs);

    /* Check to make sure that the number of NeuralNetwork inputs matches the Neuron Connection Count in the first layer. */
    if(getLayerCount()>0){
    if(arrayOfInputs.length!=layers[0].neurons[0].getConnectionCount()){
    println("NN Error: processInputsToOutputs: The number of inputs do NOT match the NN");
    exit();
    } else {
    /* The number of inputs are fine : continue */
    for(int i=0; i<getLayerCount(); i++){

    /*Set the INPUTs for each layer: The first layer gets it's input data from the NN, whereas the 2nd and subsequent layers get their input data from the previous layer's actual output. */
    if(i==0){
    setLayerInputs(arrayOfInputs,i);
    } else {
    setLayerInputs(layers[i-1].actualOUTPUTs, i);
    }

    /* Now that the layer has had it's input values set, it can now process this data, and convert them into an output using the layer's neurons. The outputs will be used as inputs in the next layer (if available). */
    layers[i].processInputsToOutputs();
    }
    /* Once all the data has filtered through to the end of network, we can grab the actualOUTPUTs of the LAST layer
    These values become or will be set to the NN output values (arrayOfOutputs), through the setOutputs function call. */
    setOutputs(layers[getLayerCount()-1].actualOUTPUTs);
    }
    }else{
    println("Error: There are no layers in this Neural Network");
    exit();
    }
    }




    /* Function to train the entire network using an array. */
    void trainNetwork(float[] inputData, float[] expectedOutputData){
    /* Populate the ENTIRE network by processing the inputData. */
    processInputsToOutputs(inputData);

    /* train each layer - from back to front (back propagation) */
    for(int i=getLayerCount()-1; i>-1; i--){
    if(i==getLayerCount()-1){
    layers[i].setDeltaError(expectedOutputData);
    layers[i].trainLayer(learningRate);
    networkError=layers[i].getLayerError();
    } else {
    /* Calculate the expected value for each neuron in this layer (eg. HIDDEN LAYER) */
    for(int j=0; j<layers[i].getNeuronCount(); j++){
    /* Reset the delta error of this neuron to zero. */
    layers[i].neurons[j].deltaError=0;
    /* The delta error of a hidden layer neuron is equal to the SUM of [the PRODUCT of the connection.weight and error of the neurons in the next layer(eg OUTPUT Layer)]. */
    /* Connection#1 of each neuron in the output layer connect with Neuron#1 in the hidden layer */
    for(int k=0; k<layers[i+1].getNeuronCount(); k++){
    layers[i].neurons[j].deltaError += (layers[i+1].neurons[k].connections[j].weight * layers[i+1].neurons[k].deltaError);
    }
    /* Now that we have the sum of Errors x weights attached to this neuron. We must multiply it by the derivative of the activation function. */
    layers[i].neurons[j].deltaError *= (layers[i].neurons[j].neuronOutputValue * (1-layers[i].neurons[j].neuronOutputValue));
    }
    /* Now that you have all the necessary fields populated, you can now Train this hidden layer and then clear the Expected outputs, ready for the next round. */
    layers[i].trainLayer(learningRate);
    layers[i].clearExpectedOUTPUT();
    }
    }
    }





    /* Function to train the entire network, using an array of input and expected data within an ArrayList */
    void trainingCycle(ArrayList trainingInputData, ArrayList trainingExpectedData, Boolean trainRandomly){
    int dataIndex;

    /* re-initialise the training Error with every cycle */
    trainingError=0;

    /* Cycle through the training data either randomly or sequentially */
    for(int i=0; i<trainingInputData.size(); i++){
    if(trainRandomly){
    dataIndex=(int) (random(trainingInputData.size()));
    } else {
    dataIndex=i;
    }

    trainNetwork((float[]) trainingInputData.get(dataIndex),(float[]) trainingExpectedData.get(dataIndex));

    /* Use the networkError variable which is calculated at the end of each individual training session to calculate the entire trainingError. */
    trainingError+=abs(networkError);
    }
    }





    /* Function to train the network until the Error is below a specific threshold */
    void autoTrainNetwork(ArrayList trainingInputData, ArrayList trainingExpectedData, float trainingErrorTarget, int cycleLimit){
    trainingError=9999;
    int trainingCounter=0;


    /* cycle through the training data until the trainingError gets below trainingErrorTarget (eg. 0.0005) or the training cycles have exceeded the cycleLimit

    variable (eg. 10000). */
    while(trainingError>trainingErrorTarget && trainingCounter<cycleLimit){

    /* re-initialise the training Error with every cycle */
    trainingError=0;

    /* Cycle through the training data randomly */
    trainingCycle(trainingInputData, trainingExpectedData, true);

    /* increment the training counter to prevent endless loop */
    trainingCounter++;
    }

    /* Due to the random nature in which this neural network is trained. There may be occasions when the training error may drop below the threshold
    To check if this is the case, we will go through one more cycle (but sequentially this time), and check the trainingError for that cycle
    If the training error is still below the trainingErrorTarget, then we will end the training session.
    If the training error is above the trainingErrorTarget, we will continue to train. It will do this check a Maximum of 9 times. */
    if(trainingCounter<cycleLimit){
    trainingCycle(trainingInputData, trainingExpectedData, false);
    trainingCounter++;

    if(trainingError>trainingErrorTarget){
    if (retrainChances<10){
    retrainChances++;
    autoTrainNetwork(trainingInputData, trainingExpectedData,trainingErrorTarget, cycleLimit);
    }
    }

    } else {
    println("CycleLimit has been reached. Has been retrained " + retrainChances + " times. Error is = " + trainingError);
    }
    }
    }
    The above code was formatted using hilite.me

    No comments:

    Post a Comment