Skip to content

Commit 55cfb16

Browse files
committed
adding equations from chapter 13
1 parent ee91b42 commit 55cfb16

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

docs/equations/pymle-equations.tex

+1-1
Original file line numberDiff line numberDiff line change
@@ -2206,7 +2206,7 @@ \section{Debugging neural networks with gradient checking}
22062206
Remember that we are updating the weights by taking an opposite step towards the direction of the gradient. In gradient checking, we compare this analytical solution to a numerically approximated gradient:
22072207

22082208
\[
2209-
\frac{\partial}{\partial w_{i, j}^{l}} J(\mathbf{W}) \approx \frac{J\big( w_{i, j}^{(l)} + \epsilon \big) - J \big( w_{i, j}^{(l)}\big)}{\epsilon}
2209+
\frac{\partial}{\partial w_{i, j}^{(l)}} J(\mathbf{W}) \approx \frac{J\big( w_{i, j}^{(l)} + \epsilon \big) - J \big( w_{i, j}^{(l)}\big)}{\epsilon}
22102210
\]
22112211

22122212
Here, $\epsilon$ is typically a very small number, for example 1e-5 (note that 1e-5 is justa more convenient notation for 0.00001). Intuitively, we can think of this finite difference approximation as the slope of the secant line connecting the points of the cost function for the two weights w and $w + \epsilon$ (both are scalar values), as shown in the following figure. We are omitting the superscripts and subscripts for simplicity.

0 commit comments

Comments
 (0)