Neural Network with Backward Propagation
Date : 2022.10.11
*The contents of this book is heavily based on Stanford University’s CS231n course.
In the previous post, we programmed forward and backward propagations for each layer (Affine, ReLU, Sigmoid, Softmax) as separate classes. Now we only need to import and build the CNN.
The benefit of coding each layer as separate classes is that we can build as many layers as we wish by simply importing each layer.
We established two methods for optimizing the weight variable. First, the numerical gradient method, and second, the backward propagation method. The mathematical simplicity of the latter method has resulted in higher efficiency compared to the former. But how good is the backward propagation method in terms of accuracy? We shall do a gradient check for both methods and compare the results.
The difference is close to negligible. (For some reason the training file and two layer net file is not outputting the expected results. Going to create a branch to try find some errors..)
I decided to delete the numerical gradient method due to its comparatively lower efficiency and the need to change my main branch.
After creating a branch solemnly for perfecting the code to work for the backpropagation, I successfully had the code running.
'Tech Development > Deep Learning (CNN)' 카테고리의 다른 글
Weight Initialization, Xavier Weights, Dropout, and Setting Hyperparameters (0) | 2022.12.16 |
---|---|
SGD, Momentum, AdaGrad, and Adam (0) | 2022.12.16 |
Computational Graphs & Backward Propagation (0) | 2022.10.24 |
SGD, Epochs, and Accuracy Testing (0) | 2022.10.22 |
Loss Function and Stochastic Gradient Descent (0) | 2022.10.20 |
댓글