Skip to content

lewismoten/artificial-neural-network

Repository files navigation

Artificial Neural Network (ANN)

A project of curiosity learning about artificial neural networks.

API

Creating a network

let algorithm = create(4, 5, 3);
// input layer has 4 nodes
// hidden layer has 5 nodes
// output layer has 3 nodes

Must provide at least 3 size parameters for an input, hidden, and output layer.

You may create deep networks (multiple hidden layers) by providing more sizes. create(4, 5, 5, 3)

Initial weights and bias will be pre-populated.

Reading values

Get a list of all weights/biases used to define the algorithm used for mutations, crossovers, and duplication.

let values = reader(algorithm);
// [
// 0.4987250047589621,
// 0.5012749952410379,
// ...
// ]

Writing values

Replace the weights/biases used to define the target algorithm.

 writer(algorithm, values);

Copying an algorithm

let a = create(4, 5, 3),
  b = create(4, 5, 3),
  values = reader(a);
writer(b, values);
// b has same values as a!

Running the algorithm

Moves the values through each layer and applies weights, biases, and activates each layer.

The values of the first layer must be set first. The last layers values will contain the results.

let algorithm = create(4, 5, 3);
algorithm[0].values = [3.4, 0.1, 1, 0.333];
let output = runner(algorithm);
// [0.1, 0.3, 0.6]

Algorithm

An algorithm is made of multiple layers. It is represented as an array of layers. Calculations are done between two layers in sequential order.

[
  inputLayer,
  hiddenLayer1,
  // ...
  hiddenLayerN,
  outputLayer
]

Layer

A layer represents ... a layer. Each node within the layer has a weight that is applied to each node in the next layer. In addition, each node in the layer has a bias that is applied to all incoming values.

{
  values: [nodeCount],
  weights: [nodeCount][nextNodeCount],
  biases: [nodeCount],
  activation: 'htan|softmax|sigmoid|heviside'
}
  • values: Represents the values coming into the layer.
  • weights: A weight to be applied to the next layers value. Each value has a weight applied to each of the next layers values. The last layer does not use weights.
  • biases: A bias applied to the incoming values before weights are applied. The first layer does not use biases.
  • activation: The function applied to each value in the layer after the previous layer has applied its bias and weights. (The first layer does not get activated.)
    • softmax: The softmax function is applied to all values as a whole. (Typically for the last layer.)
    • htan: The Hyperbolic Tangent is applied to each individual value. (Typically for all layers except the last)
    • heaviside: Applied to individual values.
    • sigmoide: Applied to individual values.
    • [empty]: No threshold is applied.

About

Artificial Neural Network

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published