Skip to content

Commit 45fb6e5

Browse files
committed
update week 6
1 parent ec5e1ee commit 45fb6e5

File tree

8 files changed

+537
-155
lines changed

8 files changed

+537
-155
lines changed

doc/pub/week6/html/week6-bs.html

Lines changed: 46 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -46,6 +46,8 @@
4646
2,
4747
None,
4848
'from-nns-and-cnns-to-recurrent-neural-networks-rnns'),
49+
('What is a recurrent NN?', 2, None, 'what-is-a-recurrent-nn'),
50+
('Why RNNs?', 2, None, 'why-rnns'),
4951
('Feedback connections', 2, None, 'feedback-connections'),
5052
('Vanishing gradients', 2, None, 'vanishing-gradients'),
5153
('Recurrent neural networks (RNNs): Overarching view',
@@ -152,6 +154,10 @@
152154
2,
153155
None,
154156
'summary-of-a-typical-rnn'),
157+
('The mathematics of RNNs, the basic architecture',
158+
2,
159+
None,
160+
'the-mathematics-of-rnns-the-basic-architecture'),
155161
('Four effective ways to learn an RNN and preparing for next '
156162
'week',
157163
2,
@@ -195,6 +201,8 @@
195201
<!-- navigation toc: --> <li><a href="#reading-recommendations" style="font-size: 80%;">Reading recommendations</a></li>
196202
<!-- navigation toc: --> <li><a href="#tensorflow-examples" style="font-size: 80%;">TensorFlow examples</a></li>
197203
<!-- navigation toc: --> <li><a href="#from-nns-and-cnns-to-recurrent-neural-networks-rnns" style="font-size: 80%;">From NNs and CNNs to recurrent neural networks (RNNs)</a></li>
204+
<!-- navigation toc: --> <li><a href="#what-is-a-recurrent-nn" style="font-size: 80%;">What is a recurrent NN?</a></li>
205+
<!-- navigation toc: --> <li><a href="#why-rnns" style="font-size: 80%;">Why RNNs?</a></li>
198206
<!-- navigation toc: --> <li><a href="#feedback-connections" style="font-size: 80%;">Feedback connections</a></li>
199207
<!-- navigation toc: --> <li><a href="#vanishing-gradients" style="font-size: 80%;">Vanishing gradients</a></li>
200208
<!-- navigation toc: --> <li><a href="#recurrent-neural-networks-rnns-overarching-view" style="font-size: 80%;">Recurrent neural networks (RNNs): Overarching view</a></li>
@@ -230,6 +238,7 @@
230238
<!-- navigation toc: --> <li><a href="#gradients-of-loss-functions" style="font-size: 80%;">Gradients of loss functions</a></li>
231239
<!-- navigation toc: --> <li><a href="#summary-of-rnns" style="font-size: 80%;">Summary of RNNs</a></li>
232240
<!-- navigation toc: --> <li><a href="#summary-of-a-typical-rnn" style="font-size: 80%;">Summary of a typical RNN</a></li>
241+
<!-- navigation toc: --> <li><a href="#the-mathematics-of-rnns-the-basic-architecture" style="font-size: 80%;">The mathematics of RNNs, the basic architecture</a></li>
233242
<!-- navigation toc: --> <li><a href="#four-effective-ways-to-learn-an-rnn-and-preparing-for-next-week" style="font-size: 80%;">Four effective ways to learn an RNN and preparing for next week</a></li>
234243

235244
</ul>
@@ -312,6 +321,40 @@ <h2 id="from-nns-and-cnns-to-recurrent-neural-networks-rnns" class="anchor">From
312321
inputs.
313322
</p>
314323

324+
<!-- !split -->
325+
<h2 id="what-is-a-recurrent-nn" class="anchor">What is a recurrent NN? </h2>
326+
327+
<p>A recurrent neural network (RNN), as opposed to a regular fully
328+
connected neural network (FCNN) or just neural network (NN), has
329+
layers that are connected to themselves.
330+
</p>
331+
332+
<p>In an FCNN there are no connections between nodes in a single
333+
layer. For instance, \( (h_1^1 \) is not connected to \( (h_2^1 \). In
334+
addition, the input and output are always of a fixed length.
335+
</p>
336+
337+
<p>In an RNN, however, this is no longer the case. Nodes in the hidden
338+
layers are connected to themselves.
339+
</p>
340+
341+
<!-- !split -->
342+
<h2 id="why-rnns" class="anchor">Why RNNs? </h2>
343+
344+
<p>Recurrent neural networks work very well when working with
345+
sequential data, that is data where the order matters. In a regular
346+
fully connected network, the order of input doesn't really matter.
347+
</p>
348+
349+
<p>Another property of RNNs is that they can handle variable input
350+
and output. Consider again the simplified breast cancer dataset. If you
351+
have trained a regular FCNN on the dataset with the two features, it
352+
makes no sense to suddenly add a third feature. The network would not
353+
know what to do with it, and would reject such inputs with three
354+
features (or any other number of features that isn't two, for that
355+
matter).
356+
</p>
357+
315358
<!-- !split -->
316359
<h2 id="feedback-connections" class="anchor">Feedback connections </h2>
317360

@@ -1019,6 +1062,9 @@ <h2 id="summary-of-a-typical-rnn" class="anchor">Summary of a typical RNN </h2>
10191062
The parameters are trained through the so-called back-propagation through time (BPTT) algorithm.
10201063
</p>
10211064

1065+
<!-- !split -->
1066+
<h2 id="the-mathematics-of-rnns-the-basic-architecture" class="anchor">The mathematics of RNNs, the basic architecture </h2>
1067+
10221068
<!-- !split -->
10231069
<h2 id="four-effective-ways-to-learn-an-rnn-and-preparing-for-next-week" class="anchor">Four effective ways to learn an RNN and preparing for next week </h2>
10241070
<ol>

doc/pub/week6/html/week6-reveal.html

Lines changed: 40 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -241,6 +241,42 @@ <h2 id="from-nns-and-cnns-to-recurrent-neural-networks-rnns">From NNs and CNNs t
241241
</p>
242242
</section>
243243

244+
<section>
245+
<h2 id="what-is-a-recurrent-nn">What is a recurrent NN? </h2>
246+
247+
<p>A recurrent neural network (RNN), as opposed to a regular fully
248+
connected neural network (FCNN) or just neural network (NN), has
249+
layers that are connected to themselves.
250+
</p>
251+
252+
<p>In an FCNN there are no connections between nodes in a single
253+
layer. For instance, \( (h_1^1 \) is not connected to \( (h_2^1 \). In
254+
addition, the input and output are always of a fixed length.
255+
</p>
256+
257+
<p>In an RNN, however, this is no longer the case. Nodes in the hidden
258+
layers are connected to themselves.
259+
</p>
260+
</section>
261+
262+
<section>
263+
<h2 id="why-rnns">Why RNNs? </h2>
264+
265+
<p>Recurrent neural networks work very well when working with
266+
sequential data, that is data where the order matters. In a regular
267+
fully connected network, the order of input doesn't really matter.
268+
</p>
269+
270+
<p>Another property of RNNs is that they can handle variable input
271+
and output. Consider again the simplified breast cancer dataset. If you
272+
have trained a regular FCNN on the dataset with the two features, it
273+
makes no sense to suddenly add a third feature. The network would not
274+
know what to do with it, and would reject such inputs with three
275+
features (or any other number of features that isn't two, for that
276+
matter).
277+
</p>
278+
</section>
279+
244280
<section>
245281
<h2 id="feedback-connections">Feedback connections </h2>
246282

@@ -1012,6 +1048,10 @@ <h2 id="summary-of-a-typical-rnn">Summary of a typical RNN </h2>
10121048
</p>
10131049
</section>
10141050

1051+
<section>
1052+
<h2 id="the-mathematics-of-rnns-the-basic-architecture">The mathematics of RNNs, the basic architecture </h2>
1053+
</section>
1054+
10151055
<section>
10161056
<h2 id="four-effective-ways-to-learn-an-rnn-and-preparing-for-next-week">Four effective ways to learn an RNN and preparing for next week </h2>
10171057
<ol>

doc/pub/week6/html/week6-solarized.html

Lines changed: 43 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -73,6 +73,8 @@
7373
2,
7474
None,
7575
'from-nns-and-cnns-to-recurrent-neural-networks-rnns'),
76+
('What is a recurrent NN?', 2, None, 'what-is-a-recurrent-nn'),
77+
('Why RNNs?', 2, None, 'why-rnns'),
7678
('Feedback connections', 2, None, 'feedback-connections'),
7779
('Vanishing gradients', 2, None, 'vanishing-gradients'),
7880
('Recurrent neural networks (RNNs): Overarching view',
@@ -179,6 +181,10 @@
179181
2,
180182
None,
181183
'summary-of-a-typical-rnn'),
184+
('The mathematics of RNNs, the basic architecture',
185+
2,
186+
None,
187+
'the-mathematics-of-rnns-the-basic-architecture'),
182188
('Four effective ways to learn an RNN and preparing for next '
183189
'week',
184190
2,
@@ -269,6 +275,40 @@ <h2 id="from-nns-and-cnns-to-recurrent-neural-networks-rnns">From NNs and CNNs t
269275
inputs.
270276
</p>
271277

278+
<!-- !split --><br><br><br><br><br><br><br><br><br><br>
279+
<h2 id="what-is-a-recurrent-nn">What is a recurrent NN? </h2>
280+
281+
<p>A recurrent neural network (RNN), as opposed to a regular fully
282+
connected neural network (FCNN) or just neural network (NN), has
283+
layers that are connected to themselves.
284+
</p>
285+
286+
<p>In an FCNN there are no connections between nodes in a single
287+
layer. For instance, \( (h_1^1 \) is not connected to \( (h_2^1 \). In
288+
addition, the input and output are always of a fixed length.
289+
</p>
290+
291+
<p>In an RNN, however, this is no longer the case. Nodes in the hidden
292+
layers are connected to themselves.
293+
</p>
294+
295+
<!-- !split --><br><br><br><br><br><br><br><br><br><br>
296+
<h2 id="why-rnns">Why RNNs? </h2>
297+
298+
<p>Recurrent neural networks work very well when working with
299+
sequential data, that is data where the order matters. In a regular
300+
fully connected network, the order of input doesn't really matter.
301+
</p>
302+
303+
<p>Another property of RNNs is that they can handle variable input
304+
and output. Consider again the simplified breast cancer dataset. If you
305+
have trained a regular FCNN on the dataset with the two features, it
306+
makes no sense to suddenly add a third feature. The network would not
307+
know what to do with it, and would reject such inputs with three
308+
features (or any other number of features that isn't two, for that
309+
matter).
310+
</p>
311+
272312
<!-- !split --><br><br><br><br><br><br><br><br><br><br>
273313
<h2 id="feedback-connections">Feedback connections </h2>
274314

@@ -975,6 +1015,9 @@ <h2 id="summary-of-a-typical-rnn">Summary of a typical RNN </h2>
9751015
The parameters are trained through the so-called back-propagation through time (BPTT) algorithm.
9761016
</p>
9771017

1018+
<!-- !split --><br><br><br><br><br><br><br><br><br><br>
1019+
<h2 id="the-mathematics-of-rnns-the-basic-architecture">The mathematics of RNNs, the basic architecture </h2>
1020+
9781021
<!-- !split --><br><br><br><br><br><br><br><br><br><br>
9791022
<h2 id="four-effective-ways-to-learn-an-rnn-and-preparing-for-next-week">Four effective ways to learn an RNN and preparing for next week </h2>
9801023
<ol>

doc/pub/week6/html/week6.html

Lines changed: 43 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -150,6 +150,8 @@
150150
2,
151151
None,
152152
'from-nns-and-cnns-to-recurrent-neural-networks-rnns'),
153+
('What is a recurrent NN?', 2, None, 'what-is-a-recurrent-nn'),
154+
('Why RNNs?', 2, None, 'why-rnns'),
153155
('Feedback connections', 2, None, 'feedback-connections'),
154156
('Vanishing gradients', 2, None, 'vanishing-gradients'),
155157
('Recurrent neural networks (RNNs): Overarching view',
@@ -256,6 +258,10 @@
256258
2,
257259
None,
258260
'summary-of-a-typical-rnn'),
261+
('The mathematics of RNNs, the basic architecture',
262+
2,
263+
None,
264+
'the-mathematics-of-rnns-the-basic-architecture'),
259265
('Four effective ways to learn an RNN and preparing for next '
260266
'week',
261267
2,
@@ -346,6 +352,40 @@ <h2 id="from-nns-and-cnns-to-recurrent-neural-networks-rnns">From NNs and CNNs t
346352
inputs.
347353
</p>
348354

355+
<!-- !split --><br><br><br><br><br><br><br><br><br><br>
356+
<h2 id="what-is-a-recurrent-nn">What is a recurrent NN? </h2>
357+
358+
<p>A recurrent neural network (RNN), as opposed to a regular fully
359+
connected neural network (FCNN) or just neural network (NN), has
360+
layers that are connected to themselves.
361+
</p>
362+
363+
<p>In an FCNN there are no connections between nodes in a single
364+
layer. For instance, \( (h_1^1 \) is not connected to \( (h_2^1 \). In
365+
addition, the input and output are always of a fixed length.
366+
</p>
367+
368+
<p>In an RNN, however, this is no longer the case. Nodes in the hidden
369+
layers are connected to themselves.
370+
</p>
371+
372+
<!-- !split --><br><br><br><br><br><br><br><br><br><br>
373+
<h2 id="why-rnns">Why RNNs? </h2>
374+
375+
<p>Recurrent neural networks work very well when working with
376+
sequential data, that is data where the order matters. In a regular
377+
fully connected network, the order of input doesn't really matter.
378+
</p>
379+
380+
<p>Another property of RNNs is that they can handle variable input
381+
and output. Consider again the simplified breast cancer dataset. If you
382+
have trained a regular FCNN on the dataset with the two features, it
383+
makes no sense to suddenly add a third feature. The network would not
384+
know what to do with it, and would reject such inputs with three
385+
features (or any other number of features that isn't two, for that
386+
matter).
387+
</p>
388+
349389
<!-- !split --><br><br><br><br><br><br><br><br><br><br>
350390
<h2 id="feedback-connections">Feedback connections </h2>
351391

@@ -1052,6 +1092,9 @@ <h2 id="summary-of-a-typical-rnn">Summary of a typical RNN </h2>
10521092
The parameters are trained through the so-called back-propagation through time (BPTT) algorithm.
10531093
</p>
10541094

1095+
<!-- !split --><br><br><br><br><br><br><br><br><br><br>
1096+
<h2 id="the-mathematics-of-rnns-the-basic-architecture">The mathematics of RNNs, the basic architecture </h2>
1097+
10551098
<!-- !split --><br><br><br><br><br><br><br><br><br><br>
10561099
<h2 id="four-effective-ways-to-learn-an-rnn-and-preparing-for-next-week">Four effective ways to learn an RNN and preparing for next week </h2>
10571100
<ol>
0 Bytes
Binary file not shown.

0 commit comments

Comments
 (0)