123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426 |
- package neuralnetworkbase
- import (
- "encoding/binary"
- "errors"
- "fmt"
- "io"
- teach "../teach"
- mat "gonum.org/v1/gonum/mat"
- )
- type NeuralNetwork struct {
- Count int
- Sizes []int
- Biases []*mat.Dense
- Weights []*mat.Dense
- A []*mat.Dense
- Z []*mat.Dense
- WGradient []Gradient
- BGradient []Gradient
- trainingCycles int
- }
- func NewNeuralNetwork(sizes []int, nu float64, trainingCycles int) (nn *NeuralNetwork, err error) {
- err = nil
- if len(sizes) < 3 {
- fmt.Printf("Invalid network configuration: %v\n", sizes)
- return nil, errors.New("Invalid network configuration: %v\n")
- }
- for i := 0; i < len(sizes); i++ {
- if sizes[i] < 2 {
- fmt.Printf("Invalid network configuration: %v\n", sizes)
- return nil, errors.New("Invalid network configuration: %v\n")
- }
- }
- if nu <= 0.0 || nu > 1.0 {
- fmt.Printf("Invalid η value: %v\n", nu)
- return nil, errors.New("Invalid η value: %v\n")
- }
- if trainingCycles <= 0 {
- fmt.Printf("Invalid training cycles number: %v\n", trainingCycles)
- return nil, errors.New("Invalid training cycles number: %v\n")
- }
- if trainingCycles < 100 {
- fmt.Println("Training cycles number probably is too small")
- }
- nn = &NeuralNetwork{}
- nn.Sizes = sizes
- nn.Count = len(sizes)
- nn.Weights = make([]*mat.Dense, nn.Count)
- nn.Biases = make([]*mat.Dense, nn.Count)
- nn.WGradient = make([]Gradient, nn.Count)
- nn.BGradient = make([]Gradient, nn.Count)
- nn.A = make([]*mat.Dense, nn.Count)
- nn.Z = make([]*mat.Dense, nn.Count)
- nn.trainingCycles = trainingCycles
- alpha := nu / float64(nn.Sizes[0])
- for i := 1; i < nn.Count; i++ {
- nn.Weights[i] = generateRandomDense(nn.Sizes[i], nn.Sizes[i-1])
- nn.Biases[i] = generateRandomDense(nn.Sizes[i], 1)
- nn.WGradient[i] = &BackPropGradient{alpha}
- nn.BGradient[i] = &BackPropGradient{alpha}
-
-
- }
- return
- }
- func (nn *NeuralNetwork) Copy() (out *NeuralNetwork) {
- out = &NeuralNetwork{}
- out.Sizes = nn.Sizes
- out.Count = nn.Count
- out.Weights = make([]*mat.Dense, nn.Count)
- out.Biases = make([]*mat.Dense, nn.Count)
- out.A = make([]*mat.Dense, nn.Count)
- out.Z = make([]*mat.Dense, nn.Count)
- out.trainingCycles = nn.trainingCycles
- for i := 1; i < out.Count; i++ {
- out.Weights[i] = mat.DenseCopyOf(nn.Weights[i])
- out.Biases[i] = mat.DenseCopyOf(nn.Biases[i])
- }
- return
- }
- func (nn *NeuralNetwork) Predict(aIn mat.Matrix) (maxIndex int, max float64) {
- r, _ := aIn.Dims()
- if r != nn.Sizes[0] {
- fmt.Printf("Invalid rows number of input matrix size: %v\n", r)
- return -1, 0.0
- }
- nn.forward(aIn)
- result := nn.result()
- r, _ = result.Dims()
- max = 0.0
- maxIndex = 0
- for i := 0; i < r; i++ {
- if result.At(i, 0) > max {
- max = result.At(i, 0)
- maxIndex = i
- }
- }
- return
- }
- func (nn *NeuralNetwork) Teach(teacher teach.Teacher) {
- for i := 0; i < nn.trainingCycles; i++ {
- for teacher.NextData() {
- nn.backward(teacher.GetData())
- }
- }
- }
- func (nn *NeuralNetwork) SaveState(writer io.Writer) {
-
- bufferSize := make([]byte, 4)
- binary.LittleEndian.PutUint32(bufferSize[0:], uint32(nn.Count))
- _, err := writer.Write(bufferSize)
- check(err)
- fmt.Printf("wrote value %d\n", uint32(nn.Count))
-
- buffer := make([]byte, nn.Count*4)
- for i := 0; i < nn.Count; i++ {
- binary.LittleEndian.PutUint32(buffer[i*4:], uint32(nn.Sizes[i]))
- }
- _, err = writer.Write(buffer)
- check(err)
-
-
-
- for i := 1; i < nn.Count; i++ {
- saveDense(writer, nn.Biases[i])
- }
-
-
- for i := 1; i < nn.Count; i++ {
- saveDense(writer, nn.Weights[i])
- }
- }
- func (nn *NeuralNetwork) LoadState(reader io.Reader) {
-
- nn.Count = readInt(reader)
-
- sizeBuffer := readByteArray(reader, nn.Count*4)
- nn.Sizes = make([]int, nn.Count)
- for i := 0; i < nn.Count; i++ {
- nn.Sizes[i] = int(binary.LittleEndian.Uint32(sizeBuffer[i*4:]))
-
- }
- nn.Weights = []*mat.Dense{&mat.Dense{}}
- nn.Biases = []*mat.Dense{&mat.Dense{}}
-
- nn.Biases[0] = &mat.Dense{}
- for i := 1; i < nn.Count; i++ {
- nn.Biases = append(nn.Biases, &mat.Dense{})
- nn.Biases[i] = readDense(reader, nn.Biases[i])
- }
-
- nn.Weights[0] = &mat.Dense{}
- for i := 1; i < nn.Count; i++ {
- nn.Weights = append(nn.Weights, &mat.Dense{})
- nn.Weights[i] = readDense(reader, nn.Weights[i])
- }
- nn.A = make([]*mat.Dense, nn.Count)
- nn.Z = make([]*mat.Dense, nn.Count)
-
- }
- func (nn *NeuralNetwork) forward(aIn mat.Matrix) {
- nn.A[0] = mat.DenseCopyOf(aIn)
- for i := 1; i < nn.Count; i++ {
- nn.A[i] = mat.NewDense(nn.Sizes[i], 1, nil)
- aSrc := nn.A[i-1]
- aDst := nn.A[i]
-
-
-
- aDst.Mul(nn.Weights[i], aSrc)
-
- aDst.Add(aDst, nn.Biases[i])
-
- nn.Z[i] = mat.DenseCopyOf(aDst)
-
- aDst.Apply(applySigmoid, aDst)
- }
- }
- func (nn *NeuralNetwork) backward(aIn, aOut mat.Matrix) {
- nn.forward(aIn)
- lastLayerNum := nn.Count - 1
-
-
-
-
-
-
-
-
-
- err := &mat.Dense{}
- err.Sub(nn.result(), aOut)
-
- sigmoidsPrime := &mat.Dense{}
- sigmoidsPrime.Apply(applySigmoidPrime, nn.Z[lastLayerNum])
-
- delta := &mat.Dense{}
- delta.MulElem(err, sigmoidsPrime)
-
- biases := mat.DenseCopyOf(delta)
-
- weights := &mat.Dense{}
- weights.Mul(delta, nn.A[lastLayerNum-1].T())
-
-
-
-
-
-
-
-
-
-
- newBiases := []*mat.Dense{nn.BGradient[lastLayerNum].ApplyDelta(nn.Biases[lastLayerNum], biases)}
- newWeights := []*mat.Dense{nn.WGradient[lastLayerNum].ApplyDelta(nn.Weights[lastLayerNum], weights)}
-
-
-
-
-
-
-
- for l := nn.Count - 2; l > 0; l-- {
-
- sigmoidsPrime := &mat.Dense{}
- sigmoidsPrime.Apply(applySigmoidPrime, nn.Z[l])
-
-
- wdelta := &mat.Dense{}
- wdelta.Mul(nn.Weights[l+1].T(), delta)
-
-
- delta = &mat.Dense{}
- delta.MulElem(wdelta, sigmoidsPrime)
-
- biases := mat.DenseCopyOf(delta)
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
-
- weights := &mat.Dense{}
- weights.Mul(delta, nn.A[l-1].T())
-
-
-
- newBiases = append([]*mat.Dense{nn.BGradient[l].ApplyDelta(nn.Biases[l], biases)}, newBiases...)
- newWeights = append([]*mat.Dense{nn.WGradient[l].ApplyDelta(nn.Weights[l], weights)}, newWeights...)
- }
- newBiases = append([]*mat.Dense{&mat.Dense{}}, newBiases...)
- newWeights = append([]*mat.Dense{&mat.Dense{}}, newWeights...)
- nn.Biases = newBiases
- nn.Weights = newWeights
- }
- func (nn *NeuralNetwork) result() *mat.Dense {
- return nn.A[nn.Count-1]
- }
|