User:Jeblad/Standard notation (neural net)
Appearance
Standard notation as it is used within deep learning, has changed a lot since the first published works. It is undergoing some standardization, but mostly at an informal level.
Notation
Indexes
- training
- Superscript like denotes the iᵗʰ training example in a trainingset
- layer
- Superscript like denotes the lᵗʰ layer in a set of layers
- sequence
- Superscript like denotes the tᵗʰ item in a sequence of items
- 1D node
- Subscript like denotes the iᵗʰ node in a one-dimensional layer
- 2D node
- Subscript or like or denotes the node at iᵗʰ row and jᵗʰ column in a two-dimensional layer[note 1]
- 1D weight
- Subscript or like or denotes the weight between node iᵗʰ at previous layer and jᵗʰ at following layer[note 2]
Sizes
- number of samples
- is the number of samples in the dataset
- input size
- is the size of input (or number of features)
- output size
- is the size of output (or number of classes)
- hidden units
- is the number of units in hidden layer
- number of layers
- is the number of layers in the network
- input sequence size
- is the size of the input sequence
- output sequence size
- is the size of the output sequence
- input training sequence size
- is the size of the input training sequence (each sample training sequence)
- output training sequence size
- is the size of the output training sequence (each sample training sequence)
Other
- cross entropy
- elementwise sequence loss
- and by using cross entropy that is the sum would be over for classification in and out of a single class
References
Notes
- ^ This can easily be confused with a weight index.
- ^ Michael Nielson defines as weight from kᵗʰ neuron to jᵗʰ, while Andrew Ng defines it in opposite direction.