The Course Notes (by Andrej Karpathy) are well written and they worth a look.
That course notes have inspired me to create a picture for summarising some concepts.
An interesting summary (adapted from here ) is the following:
Input Layer
- Size: W_1 \times H_1 \times D_1
- Hyperparameters:
- Number of filters K
- Dimension of the filter F \times F \times D_1
- Stride: S
- Amount of Zero Padding: P
Output Layer
The parameter sharing introduces F \times F \times D_1 per filter, for a total of (F \times F \times D_1) \times K weights and K biases- Size: W_2 \times H_2 \times D_2
- W_2 = \frac{W_1 - F + 2P}{S} + 1
- H_2 = \frac{H_1 - F + 2P}{S} + 1
- D_2 = K
In the output volume, the d-th depth slice (of size W_2 \times H2) is the result of performing a valid convolution of the d-th filter over the input volume with a stride of S, and then offset by d-th bias.
Another interesting post on the Convolutional Neural Network is here
No comments:
Post a Comment