|
91 | 91 | "cell_type": "markdown", |
92 | 92 | "metadata": {}, |
93 | 93 | "source": [ |
94 | | - "The hypothesis space for Candidate-Elimination spreads between the most general and most specific hypotheses. The other hypotheses are made up by conjunction of features which biases the learner and makes it impossible to find a disjunctive solution." |
| 94 | + "The hypothesis space for Candidate-Elimination spreads between the most general and most specific hypotheses. The other hypotheses are made up by conjunction of features which biases the learner and makes it impossible to find a disjunctive solution.\n", |
| 95 | + "\n", |
| 96 | + "The version space on the other hand is a subset of the hypotheses space. It is the set of all hypotheses between and including the general and the specific boundary.\n", |
| 97 | + "\n" |
95 | 98 | ] |
96 | 99 | }, |
97 | 100 | { |
|
389 | 392 | "Which of the following formulae describes the backpropagation of the error through hidden layers in a Multilayer Perceptron?\n", |
390 | 393 | "Assume they are calculated for each $k=L_H \\dots 1$ and $i=1\\dots N(k)$.\n", |
391 | 394 | "\n", |
392 | | - "1. $\\delta_i(k) = f^\\prime(o_i(k-1)) \\sum\\limits_{j=1}^{N(k+1)} w_{ji}(k+1, k)o_i(k)$\n", |
393 | | - "2. $\\delta_i(k) = f^\\prime(o_i(k-1)) \\sum\\limits_{j=1}^{N(k+1)} w_{ji}(k+1, k)\\delta_i(k+1)$\n", |
394 | | - "3. $\\delta_i(k) = f^\\prime(o_i(k-1)) \\sum\\limits_{j=1}^{N(k+1)} w_{ji}(k, k-1)\\delta_i(k+1)$" |
| 395 | + "1. $\\delta_i(k) = f^\\prime(o_i(k)) \\sum\\limits_{j=1}^{N(k+1)} w_{ji}(k+1, k)o_i(k)$\n", |
| 396 | + "2. $\\delta_i(k) = f^\\prime(o_i(k)) \\sum\\limits_{j=1}^{N(k+1)} w_{ji}(k+1, k)\\delta_i(k+1)$\n", |
| 397 | + "3. $\\delta_i(k) = f^\\prime(o_i(k)) \\sum\\limits_{j=1}^{N(k+1)} w_{ji}(k, k-1)\\delta_i(k+1)$" |
395 | 398 | ] |
396 | 399 | }, |
397 | 400 | { |
|
584 | 587 | "cell_type": "markdown", |
585 | 588 | "metadata": {}, |
586 | 589 | "source": [ |
587 | | - "The (first-order) Markov assumption means that state $s_{t+1}$ only depends on its predecessor state $s_t$ and the action $a_t$ performed then, i.e.: $s_{t+1} = \\delta(s_t, a_t)$. This allows to specify a $Q$-function of the form $Q(s_t,a_t)$, instead of $Q(s_0,a_0,\\ldots,s_t,a_t)$. The Markov assumption does not hold in situations where, e.g. the state does contain full information." |
| 590 | + "The (first-order) Markov assumption means that state $s_{t+1}$ only depends on its predecessor state $s_t$ and the action $a_t$ performed then, i.e.: $s_{t+1} = \\delta(s_t, a_t)$. This allows to specify a $Q$-function of the form $Q(s_t,a_t)$, instead of $Q(s_0,a_0,\\ldots,s_t,a_t)$. The Markov assumption does not hold in situations where more information is needed than provided by the previous state. For example for sentence parsing with each word being a state the Markov assumption does not hold." |
588 | 591 | ] |
589 | 592 | }, |
590 | 593 | { |
|
693 | 696 | "name": "python", |
694 | 697 | "nbconvert_exporter": "python", |
695 | 698 | "pygments_lexer": "ipython3", |
696 | | - "version": "3.5.1" |
| 699 | + "version": "3.5.2" |
697 | 700 | } |
698 | 701 | }, |
699 | 702 | "nbformat": 4, |
|
0 commit comments