-
Notifications
You must be signed in to change notification settings - Fork 574
/
Copy pathsearch_data.json
5103 lines (3920 loc) · 767 KB
/
search_data.json
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
{
"_":{
"title":"",
"content":"",
"url":""
}
,
"dbn-exercises-ex-10": {
"title": "Exercise 15.10",
"breadcrumb": "15-Probabilistic-Reasoning-Over-Time",
"content" : "This exercise is concerned with filtering in an environment with nolandmarks. Consider a vacuum robot in an empty room, represented by an$n times m$ rectangular grid. The robot’s location is hidden; the onlyevidence available to the observer is a noisy location sensor that givesan approximation to the robot’s location. If the robot is at location$(x, y)$ then with probability .1 the sensor gives the correct location,with probability .05 each it reports one of the 8 locations immediatelysurrounding $(x, y)$, with probability .025 each it reports one of the16 locations that surround those 8, and with the remaining probabilityof .1 it reports “no reading.” The robot’s policy is to pick a directionand follow it with probability .8 on each step; the robot switches to arandomly selected new heading with probability .2 (or with probability 1if it encounters a wall). Implement this as an HMM and do filtering totrack the robot. How accurately can we track the robot’s path?",
"url": " /dbn-exercises/ex_10/"
}
,
"dbn-exercises-ex-3": {
"title": "Exercise 15.3",
"breadcrumb": "15-Probabilistic-Reasoning-Over-Time",
"content" : "This exercise develops a space-efficient variant ofthe forward–backward algorithm described inFigure forward-backward-algorithm (page forward-backward-algorithm).We wish to compute $$textbf{P} (textbf{X}_k|textbf{e}_{1:t})$$ for$$k=1,ldots ,t$$. This will be done with a divide-and-conquerapproach.1. Suppose, for simplicity, that $t$ is odd, and let the halfway point be $h=(t+1)/2$. Show that $$textbf{P} (textbf{X}_k|textbf{e}_{1:t}) $$ can be computed for $k=1,ldots ,h$ given just the initial forward message $$textbf{f}_{1:0}$$, the backward message $$textbf{b}_{h+1:t}$$, and the evidence $$textbf{e}_{1:h}$$.2. Show a similar result for the second half of the sequence.3. Given the results of (a) and (b), a recursive divide-and-conquer algorithm can be constructed by first running forward along the sequence and then backward from the end, storing just the required messages at the middle and the ends. Then the algorithm is called on each half. Write out the algorithm in detail.4. Compute the time and space complexity of the algorithm as a function of $t$, the length of the sequence. How does this change if we divide the input into more than two pieces?",
"url": " /dbn-exercises/ex_3/"
}
,
"dbn-exercises-ex-4": {
"title": "Exercise 15.4",
"breadcrumb": "15-Probabilistic-Reasoning-Over-Time",
"content" : "On page flawed-viterbi-page, we outlined a flawedprocedure for finding the most likely state sequence, given anobservation sequence. The procedure involves finding the most likelystate at each time step, using smoothing, and returning the sequencecomposed of these states. Show that, for some temporal probabilitymodels and observation sequences, this procedure returns an impossiblestate sequence (i.e., the posterior probability of the sequence iszero).",
"url": " /dbn-exercises/ex_4/"
}
,
"dbn-exercises-ex-5": {
"title": "Exercise 15.5",
"breadcrumb": "15-Probabilistic-Reasoning-Over-Time",
"content" : "Equation (matrix-filtering-equation) describes thefiltering process for the matrix formulation of HMMs. Give a similarequation for the calculation of likelihoods, which was describedgenerically in Equation (forward-likelihood-equation).",
"url": " /dbn-exercises/ex_5/"
}
,
"dbn-exercises-ex-2": {
"title": "Exercise 15.2",
"breadcrumb": "15-Probabilistic-Reasoning-Over-Time",
"content" : "In this exercise, we examine whathappens to the probabilities in the umbrella world in the limit of longtime sequences.1. Suppose we observe an unending sequence of days on which the umbrella appears. Show that, as the days go by, the probability of rain on the current day increases monotonically toward a fixed point. Calculate this fixed point.2. Now consider forecasting further and further into the future, given just the first two umbrella observations. First, compute the probability $P(r_{2+k}|u_1,u_2)$ for $k=1 ldots 20$ and plot the results. You should see that the probability converges towards a fixed point. Prove that the exact value of this fixed point is 0.5.",
"url": " /dbn-exercises/ex_2/"
}
,
"dbn-exercises-ex-15": {
"title": "Exercise 15.15",
"breadcrumb": "15-Probabilistic-Reasoning-Over-Time",
"content" : "A professor wants to know if students are gettingenough sleep. Each day, the professor observes whether the studentssleep in class, and whether they have red eyes. The professor has thefollowing domain theory:- The prior probability of getting enough sleep, with no observations, is 0.7.- The probability of getting enough sleep on night $t$ is 0.8 given that the student got enough sleep the previous night, and 0.3 if not.- The probability of having red eyes is 0.2 if the student got enough sleep, and 0.7 if not.- The probability of sleeping in class is 0.1 if the student got enough sleep, and 0.3 if not.Formulate this information as a dynamic Bayesian network that theprofessor could use to filter or predict from a sequence ofobservations. Then reformulate it as a hidden Markov model that has onlya single observation variable. Give the complete probability tables forthe model.",
"url": " /dbn-exercises/ex_15/"
}
,
"dbn-exercises-ex-12": {
"title": "Exercise 15.12",
"breadcrumb": "15-Probabilistic-Reasoning-Over-Time",
"content" : "Often, we wish to monitor a continuous-statesystem whose behavior switches unpredictably among a set of $k$ distinct“modes.” For example, an aircraft trying to evade a missile can executea series of distinct maneuvers that the missile may attempt to track. ABayesian network representation of such a switching Kalmanfilter model is shown inFigure switching-kf-figure.1. Suppose that the discrete state $S_t$ has $k$ possible values and that the prior continuous state estimate $${textbf{P}}(textbf{X}_0)$$ is a multivariate Gaussian distribution. Show that the prediction $${textbf{P}}(textbf{X}_1)$$ is a mixture of Gaussians—that is, a weighted sum of Gaussians such that the weights sum to 1.2. Show that if the current continuous state estimate $${textbf{P}}(textbf{X}_t|textbf{e}_{1:t})$$ is a mixture of $m$ Gaussians, then in the general case the updated state estimate $${textbf{P}}(textbf{X}_{t+1}|textbf{e}_{1:t+1})$$ will be a mixture of $km$ Gaussians.3. What aspect of the temporal process do the weights in the Gaussian mixture represent?The results in (a) and (b) show that the representation of the posteriorgrows without limit even for switching Kalman filters, which are amongthe simplest hybrid dynamic models.",
"url": " /dbn-exercises/ex_12/"
}
,
"dbn-exercises-ex-13": {
"title": "Exercise 15.13",
"breadcrumb": "15-Probabilistic-Reasoning-Over-Time",
"content" : "Complete the missing step in the derivationof Equation (kalman-one-step-equation) onpage kalman-one-step-equation, the first update step for the one-dimensional Kalmanfilter.",
"url": " /dbn-exercises/ex_13/"
}
,
"dbn-exercises-ex-14": {
"title": "Exercise 15.14",
"breadcrumb": "15-Probabilistic-Reasoning-Over-Time",
"content" : "Let us examine the behavior of the varianceupdate in Equation (kalman-univariate-equation)(page kalman-univariate-equation).1. Plot the value of $sigma_t^2$ as a function of $t$, given various values for $sigma_x^2$ and $sigma_z^2$.2. Show that the update has a fixed point $sigma^2$ such that $sigma_t^2 rightarrow sigma^2$ as $t rightarrow infty$, and calculate the value of $sigma^2$.3. Give a qualitative explanation for what happens as $sigma_x^2rightarrow 0$ and as $sigma_z^2rightarrow 0$.",
"url": " /dbn-exercises/ex_14/"
}
,
"dbn-exercises-ex-9": {
"title": "Exercise 15.9",
"breadcrumb": "15-Probabilistic-Reasoning-Over-Time",
"content" : "We have described three policies for the vacuum robot: (1) a uniformrandom walk, (2) a bias for wandering southeast, as described inExercise hmm-robust-exercise, and (3) the policydescribed in Exercise roomba-viterbi-exercise. Supposean observer is given the observation sequence from a vacuum robot, butis not sure which of the three policies the robot is following. Whatapproach should the observer use to find the most likely path, given theobservations? Implement the approach and test it. How much does thelocalization accuracy suffer, compared to the case in which the observerknows which policy the robot is following?",
"url": " /dbn-exercises/ex_9/"
}
,
"dbn-exercises-ex-7": {
"title": "Exercise 15.7",
"breadcrumb": "15-Probabilistic-Reasoning-Over-Time",
"content" : "In Section hmm-localization-section, the priordistribution over locations is uniform and the transition model assumesan equal probability of moving to any neighboring square. What if thoseassumptions are wrong? Suppose that the initial location is actuallychosen uniformly from the northwest quadrant of the room and the actionactually tends to move southeast[hmm-robot-southeast-page]. Keepingthe HMM model fixed, explore the effect on localization and pathaccuracy as the southeasterly tendency increases, for different valuesof $epsilon$.",
"url": " /dbn-exercises/ex_7/"
}
,
"dbn-exercises-ex-1": {
"title": "Exercise 15.1",
"breadcrumb": "15-Probabilistic-Reasoning-Over-Time",
"content" : "Show that any second-order Markovprocess can be rewritten as a first-order Markov process with anaugmented set of state variables. Can this always be doneparsimoniously, i.e., without increasing the number ofparameters needed to specify the transition model?",
"url": " /dbn-exercises/ex_1/"
}
,
"dbn-exercises-ex-6": {
"title": "Exercise 15.6",
"breadcrumb": "15-Probabilistic-Reasoning-Over-Time",
"content" : "Consider the vacuum worlds ofFigure vacuum-maze-ch4-figure (perfect sensing) andFigure vacuum-maze-hmm2-figure (noisy sensing). Supposethat the robot receives an observation sequence such that, with perfectsensing, there is exactly one possible location it could be in. Is thislocation necessarily the most probable location under noisy sensing forsufficiently small noise probability $epsilon$? Prove your claim orfind a counterexample.",
"url": " /dbn-exercises/ex_6/"
}
,
"dbn-exercises-ex-8": {
"title": "Exercise 15.8",
"breadcrumb": "15-Probabilistic-Reasoning-Over-Time",
"content" : "Consider a version of the vacuum robot(page vacuum-maze-hmm2-figure) that has the policy of going straight for as longas it can; only when it encounters an obstacle does it change to a new(randomly selected) heading. To model this robot, each state in themodel consists of a (location, heading) pair. Implementthis model and see how well the Viterbi algorithm can track a robot withthis model. The robot’s policy is more constrained than the random-walkrobot; does that mean that predictions of the most likely path are moreaccurate?",
"url": " /dbn-exercises/ex_8/"
}
,
"philosophy-exercises-ex-11": {
"title": "Exercise 26.11",
"breadcrumb": "26-Philosophical-Foundations",
"content" : "How do the potential threats from AI technology compare with those fromother computer science technologies, and to bio-, nano-, and nucleartechnologies?",
"url": " /philosophy-exercises/ex_11/"
}
,
"philosophy-exercises-ex-10": {
"title": "Exercise 26.10",
"breadcrumb": "26-Philosophical-Foundations",
"content" : "Analyze the potential threats from AI technology to society. Whatthreats are most serious, and how might they be combated? How do theycompare to the potential benefits?",
"url": " /philosophy-exercises/ex_10/"
}
,
"philosophy-exercises-ex-3": {
"title": "Exercise 26.3",
"breadcrumb": "26-Philosophical-Foundations",
"content" : "Attempt to write definitions of the terms “intelligence,” “thinking,”and “consciousness.” Suggest some possible objections to yourdefinitions.",
"url": " /philosophy-exercises/ex_3/"
}
,
"philosophy-exercises-ex-4": {
"title": "Exercise 26.4",
"breadcrumb": "26-Philosophical-Foundations",
"content" : "Does a refutation of the Chinese room argument necessarily prove thatappropriately programmed computers have mental states? Does anacceptance of the argument necessarily mean that computers cannot havemental states?",
"url": " /philosophy-exercises/ex_4/"
}
,
"philosophy-exercises-ex-5": {
"title": "Exercise 26.5",
"breadcrumb": "26-Philosophical-Foundations",
"content" : "In the brain replacement argument, it isimportant to be able to restore the subject’s brain to normal, such thatits external behavior is as it would have been if the operation had nottaken place. Can the skeptic reasonably object that this would requireupdating those neurophysiological properties of the neurons relating toconscious experience, as distinct from those involved in the functionalbehavior of the neurons?",
"url": " /philosophy-exercises/ex_5/"
}
,
"philosophy-exercises-ex-2": {
"title": "Exercise 26.2",
"breadcrumb": "26-Philosophical-Foundations",
"content" : "Find and analyze an account in the popular media of one or more of thearguments to the effect that AI is impossible.",
"url": " /philosophy-exercises/ex_2/"
}
,
"philosophy-exercises-ex-12": {
"title": "Exercise 26.12",
"breadcrumb": "26-Philosophical-Foundations",
"content" : "Some critics object that AI is impossible, while others object that itis *too* possible and that ultraintelligent machines pose athreat. Which of these objections do you think is more likely? Would itbe a contradiction for someone to hold both positions?",
"url": " /philosophy-exercises/ex_12/"
}
,
"philosophy-exercises-ex-9": {
"title": "Exercise 26.9",
"breadcrumb": "26-Philosophical-Foundations",
"content" : "I. J. Good claims that intelligence is the most important quality, andthat building ultraintelligent machines will change everything. Asentient cheetah counters that “Actually speed is more important; if wecould build ultrafast machines, that would change everything,” and asentient elephant claims “You’re both wrong; what we need is ultrastrongmachines.” What do you think of these arguments?",
"url": " /philosophy-exercises/ex_9/"
}
,
"philosophy-exercises-ex-7": {
"title": "Exercise 26.7",
"breadcrumb": "26-Philosophical-Foundations",
"content" : "Alan Perlis [Perlis:1982] wrote, “A year spent in artificialintelligence is enough to make one believe in God”. He also wrote, in aletter to Philip Davis, that one of the central dreams of computerscience is that “through the performance of computers and their programswe will remove all doubt that there is only a chemical distinctionbetween the living and nonliving world.” To what extent does theprogress made so far in artificial intelligence shed light on theseissues? Suppose that at some future date, the AI endeavor has beencompletely successful; that is, we have build intelligent agents capableof carrying out any human cognitive task at human levels of ability. Towhat extent would that shed light on these issues?",
"url": " /philosophy-exercises/ex_7/"
}
,
"philosophy-exercises-ex-1": {
"title": "Exercise 26.1",
"breadcrumb": "26-Philosophical-Foundations",
"content" : "Go through Turing’s list of alleged“disabilities” of machines, identifying which have been achieved, whichare achievable in principle by a program, and which are stillproblematic because they require conscious mental states.",
"url": " /philosophy-exercises/ex_1/"
}
,
"philosophy-exercises-ex-6": {
"title": "Exercise 26.6",
"breadcrumb": "26-Philosophical-Foundations",
"content" : "Suppose that a Prolog program containing many clauses about the rules ofBritish citizenship is compiled and run on an ordinary computer. Analyzethe “brain states” of the computer under wide and narrow content.",
"url": " /philosophy-exercises/ex_6/"
}
,
"philosophy-exercises-ex-8": {
"title": "Exercise 26.8",
"breadcrumb": "26-Philosophical-Foundations",
"content" : "Compare the social impact of artificial intelligence in the last fiftyyears with the social impact of the introduction of electric appliancesand the internal combustion engine in the fifty years between 1890 and1940.",
"url": " /philosophy-exercises/ex_8/"
}
,
"concept-learning-exercises-ex-11": {
"title": "Exercise 18.11",
"breadcrumb": "18-Learning-From-Examples",
"content" : "This exercise considers $chi^2$ pruning ofdecision trees (Section chi-squared-section.1. Create a data set with two input attributes, such that the information gain at the root of the tree for both attributes is zero, but there is a decision tree of depth 2 that is consistent with all the data. What would $chi^2$ pruning do on this data set if applied bottom up? If applied top down?2. Modify DECISION-TREE-LEARNING to include $chi^2$-pruning. You might wish to consult Quinlan [Quinlan:1986] or [Kearns+Mansour:1998] for details.",
"url": " /concept-learning-exercises/ex_11/"
}
,
"concept-learning-exercises-ex-16": {
"title": "Exercise 18.16",
"breadcrumb": "18-Learning-From-Examples",
"content" : "Construct a decision list to classify the data below.Select tests to be as small as possible (in terms of attributes),breaking ties among tests with the same number of attributes byselecting the one that classifies the greatest number of examplescorrectly. If multiple tests have the same number of attributes andclassify the same number of examples, then break the tie usingattributes with lower index numbers (e.g., select $A_1$ over $A_2$).| | $quad A_1quad$ | $quad A_2quad$ | $quad A_3quad$ | $quad A_yquad$ | $quad yquad$ || --- | --- | --- | --- | --- | --- || $textbf{x}_1$ | 1 | 0 | 0 | 0 | 1 || $textbf{x}_2$ | 1 | 0 | 1 | 1 | 1 || $textbf{x}_3$ | 0 | 1 | 0 | 0 | 1 || $textbf{x}_4$ | 0 | 1 | 1 | 0 | 0 || $textbf{x}_5$ | 1 | 1 | 0 | 1 | 1 || $textbf{x}_6$ | 0 | 1 | 0 | 1 | 0 || $textbf{x}_7$ | 0 | 0 | 1 | 1 | 1 || $textbf{x}_8$ | 0 | 0 | 1 | 0 | 0 |",
"url": " /concept-learning-exercises/ex_16/"
}
,
"concept-learning-exercises-ex-29": {
"title": "Exercise 18.29",
"breadcrumb": "18-Learning-From-Examples",
"content" : "Suppose you had a neural network with linearactivation functions. That is, for each unit the output is some constant$c$ times the weighted sum of the inputs.1. Assume that the network has one hidden layer. For a given assignment to the weights $textbf{w}$, write down equations for the value of the units in the output layer as a function of $textbf{w}$ and the input layer $textbf{x}$, without any explicit mention of the output of the hidden layer. Show that there is a network with no hidden units that computes the same function.2. Repeat the calculation in part (a), but this time do it for a network with any number of hidden layers.3. Suppose a network with one hidden layer and linear activation functions has $n$ input and output nodes and $h$ hidden nodes. What effect does the transformation in part (a) to a network with no hidden layers have on the total number of weights? Discuss in particular the case $h ll n$.",
"url": " /concept-learning-exercises/ex_29/"
}
,
"concept-learning-exercises-ex-20": {
"title": "Exercise 18.20",
"breadcrumb": "18-Learning-From-Examples",
"content" : "Suppose a $7$-nearest-neighbors regression searchreturns $ {4, 2, 8, 4, 9, 11, 100} $ as the 7 nearest $y$ values for agiven $x$ value. What is the value of $hat{y}$ that minimizes the $L_1$loss function on this data? There is a common name in statistics forthis value as a function of the $y$ values; what is it? Answer the sametwo questions for the $L_2$ loss function.",
"url": " /concept-learning-exercises/ex_20/"
}
,
"concept-learning-exercises-ex-27": {
"title": "Exercise 18.27",
"breadcrumb": "18-Learning-From-Examples",
"content" : "Consider the following set of examples, each with six inputs and onetarget output:| | | | | | | | | | | | | | | || --- | --- | --- | --- | --- | --- || $textbf{x}_1$ | 1 | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 0 | 0 | 0 | 0 | 0 | 0 || $textbf{x}_2$ | 0 | 0 | 0 | 1 | 1 | 0 | 0 | 1 | 1 | 0 | 1 | 0 | 1 | 1 || $textbf{x}_3$ | 1 | 1 | 1 | 0 | 1 | 0 | 0 | 1 | 1 | 0 | 0 | 0 | 1 | 1 || $textbf{x}_4$ | 0 | 1 | 0 | 0 | 1 | 0 | 0 | 1 | 0 | 1 | 1 | 1 | 0 | 1 || $textbf{x}_5$ | 0 | 0 | 1 | 1 | 0 | 1 | 1 | 0 | 1 | 1 | 0 | 0 | 1 | 0 || $textbf{x}_6$ | 0 | 0 | 0 | 1 | 0 | 1 | 0 | 1 | 1 | 0 | 1 | 1 | 1 | 0 || $textbf{T}$ | 1 | 1 | 1 | 1 | 1 | 1 | 0 | 1 | 0 | 0 | 0 | 0 | 0 | 0 |1. Run the perceptron learning rule on these data and show the final weights.2. Run the decision tree learning rule, and show the resulting decision tree.3. Comment on your results.",
"url": " /concept-learning-exercises/ex_27/"
}
,
"concept-learning-exercises-ex-18": {
"title": "Exercise 18.18",
"breadcrumb": "18-Learning-From-Examples",
"content" : "This exercise concerns the expressiveness ofdecision lists (Section learning-theory-section).1. Show that decision lists can represent any Boolean function, if the size of the tests is not limited.2. Show that if the tests can contain at most $k$ literals each, then decision lists can represent any function that can be represented by a decision tree of depth $k$.",
"url": " /concept-learning-exercises/ex_18/"
}
,
"concept-learning-exercises-ex-26": {
"title": "Exercise 18.26",
"breadcrumb": "18-Learning-From-Examples",
"content" : "Recall fromChapter concept-learning-chapter that there are$2^{2^n}$ distinct Boolean functions of $n$ inputs. How many ofthese are representable by a threshold perceptron?",
"url": " /concept-learning-exercises/ex_26/"
}
,
"concept-learning-exercises-ex-19": {
"title": "Exercise 18.19",
"breadcrumb": "18-Learning-From-Examples",
"content" : "Suppose a $7$-nearest-neighbors regression searchreturns $ {7, 6, 8, 4, 7, 11, 100} $ as the 7 nearest $y$ values for agiven $x$ value. What is the value of $hat{y}$ that minimizes the $L_1$loss function on this data? There is a common name in statistics forthis value as a function of the $y$ values; what is it? Answer the sametwo questions for the $L_2$ loss function.",
"url": " /concept-learning-exercises/ex_19/"
}
,
"concept-learning-exercises-ex-21": {
"title": "Exercise 18.21",
"breadcrumb": "18-Learning-From-Examples",
"content" : "Figure <ahref=""#">kernel-machine-figure</a>showed how a circle at the origin can be linearly separated by mappingfrom the features $(x_1, x_2)$ to the two dimensions $(x_1^2, x_2^2)$.But what if the circle is not located at the origin? What if it is anellipse, not a circle? The general equation for a circle (and hence thedecision boundary) is $(x_1-a)^2 +(x_2-b)^2 - r^20$, and the general equation for an ellipse is$c(x_1-a)^2 + d(x_2-b)^2 - 1 0$.1. Expand out the equation for the circle and show what the weights $w_i$ would be for the decision boundary in the four-dimensional feature space $(x_1, x_2, x_1^2, x_2^2)$. Explain why this means that any circle is linearly separable in this space.2. Do the same for ellipses in the five-dimensional feature space $(x_1, x_2, x_1^2, x_2^2, x_1 x_2)$.",
"url": " /concept-learning-exercises/ex_21/"
}
,
"concept-learning-exercises-ex-17": {
"title": "Exercise 18.17",
"breadcrumb": "18-Learning-From-Examples",
"content" : "Prove that a decision list can represent the same function as a decisiontree while using at most as many rules as there are leaves in thedecision tree for that function. Give an example of a functionrepresented by a decision list using strictly fewer rules than thenumber of leaves in a minimal-sized decision tree for that samefunction.",
"url": " /concept-learning-exercises/ex_17/"
}
,
"concept-learning-exercises-ex-28": {
"title": "Exercise 18.28",
"breadcrumb": "18-Learning-From-Examples",
"content" : "Section logistic-regression-section(page logistic-regression-section) noted that the output of the logistic functioncould be interpreted as a probability $p$ assigned by themodel to the proposition that $f(textbf{x})1$; the probability that$f(textbf{x})0$ is therefore $1-p$. Write down the probability $p$as a function of $textbf{x}$ and calculate the derivative of $log p$ withrespect to each weight $w_i$. Repeat the process for $log (1-p)$. Thesecalculations give a learning rule for minimizing thenegative-log-likelihood loss function for a probabilistic hypothesis.Comment on any resemblance to other learning rules in the chapter.",
"url": " /concept-learning-exercises/ex_28/"
}
,
"concept-learning-exercises-ex-10": {
"title": "Exercise 18.10",
"breadcrumb": "18-Learning-From-Examples",
"content" : "A decision graph is a generalization of a decision treethat allows nodes (i.e., attributes used for splits) to have multipleparents, rather than just a single parent. The resulting graph muststill be acyclic. Now, consider the XOR function of threebinary input attributes, which produces the value 1 if and only if anodd number of the three input attributes has value 1.1. Draw a minimal-sized decision tree for the three-input XOR function.2. Draw a minimal-sized decision graph for the three-input XOR function.",
"url": " /concept-learning-exercises/ex_10/"
}
,
"concept-learning-exercises-ex-32": {
"title": "Exercise 18.32",
"breadcrumb": "18-Learning-From-Examples",
"content" : "The neural network whose learning performance is measured inFigure restaurant-back-prop-figure has four hiddennodes. This number was chosen somewhat arbitrarily. Use across-validation method to find the best number of hidden nodes.",
"url": " /concept-learning-exercises/ex_32/"
}
,
"concept-learning-exercises-ex-3": {
"title": "Exercise 18.3",
"breadcrumb": "18-Learning-From-Examples",
"content" : "Draw a decision tree for the problem of deciding whether to move forwardat a road intersection, given that the light has just turned green.",
"url": " /concept-learning-exercises/ex_3/"
}
,
"concept-learning-exercises-ex-4": {
"title": "Exercise 18.4",
"breadcrumb": "18-Learning-From-Examples",
"content" : "We never test the same attribute twice along one path in a decisiontree. Why not?",
"url": " /concept-learning-exercises/ex_4/"
}
,
"concept-learning-exercises-ex-33": {
"title": "Exercise 18.33",
"breadcrumb": "18-Learning-From-Examples",
"content" : "Consider the problem of separating$N$ data points into positive and negative examples using a linearseparator. Clearly, this can always be done for $N2$ pointson a line of dimension $d1$, regardless of how the points arelabeled or where they are located (unless the points are in the sameplace).1. Show that it can always be done for $N3$ points on a plane of dimension $d2$, unless they are collinear.2. Show that it cannot always be done for $N4$ points on a plane of dimension $d2$.3. Show that it can always be done for $N4$ points in a space of dimension $d3$, unless they are coplanar.4. Show that it cannot always be done for $N5$ points in a space of dimension $d3$.5. The ambitious student may wish to prove that $N$ points in general position (but not $N+1$) are linearly separable in a space of dimension $N-1$.",
"url": " /concept-learning-exercises/ex_33/"
}
,
"concept-learning-exercises-ex-5": {
"title": "Exercise 18.5",
"breadcrumb": "18-Learning-From-Examples",
"content" : "Suppose we generate a training set from a decision tree and then applydecision-tree learning to that training set. Is it the case that thelearning algorithm will eventually return the correct tree as thetraining-set size goes to infinity? Why or why not?",
"url": " /concept-learning-exercises/ex_5/"
}
,
"concept-learning-exercises-ex-2": {
"title": "Exercise 18.2",
"breadcrumb": "18-Learning-From-Examples",
"content" : "Repeat Exercise infant-language-exercise for the caseof learning to play tennis (or some other sport with which you arefamiliar). Is this supervised learning or reinforcement learning?",
"url": " /concept-learning-exercises/ex_2/"
}
,
"concept-learning-exercises-ex-15": {
"title": "Exercise 18.15",
"breadcrumb": "18-Learning-From-Examples",
"content" : "Suppose that a learning algorithm is trying to find a consistenthypothesis when the classifications of examples are actually random.There are $n$ Boolean attributes, and examples are drawn uniformly fromthe set of $2^n$ possible examples. Calculate the number of examplesrequired before the probability of finding a contradiction in the datareaches 0.5.",
"url": " /concept-learning-exercises/ex_15/"
}
,
"concept-learning-exercises-ex-12": {
"title": "Exercise 18.12",
"breadcrumb": "18-Learning-From-Examples",
"content" : "The standard DECISION-TREE-LEARNING algorithm described in thechapter does not handle cases in which some examples have missingattribute values.1. First, we need to find a way to classify such examples, given a decision tree that includes tests on the attributes for which values can be missing. Suppose that an example $textbf{x}$ has a missing value for attribute $A$ and that the decision tree tests for $A$ at a node that $textbf{x}$ reaches. One way to handle this case is to pretend that the example has all possible values for the attribute, but to weight each value according to its frequency among all of the examples that reach that node in the decision tree. The classification algorithm should follow all branches at any node for which a value is missing and should multiply the weights along each path. Write a modified classification algorithm for decision trees that has this behavior.2. Now modify the information-gain calculation so that in any given collection of examples $C$ at a given node in the tree during the construction process, the examples with missing values for any of the remaining attributes are given “as-if” values according to the frequencies of those values in the set $C$.",
"url": " /concept-learning-exercises/ex_12/"
}
,
"concept-learning-exercises-ex-24": {
"title": "Exercise 18.24",
"breadcrumb": "18-Learning-From-Examples",
"content" : "Construct by hand a neural network that computes the xorfunction of two inputs. Make sure to specify what sort of units you areusing.",
"url": " /concept-learning-exercises/ex_24/"
}
,
"concept-learning-exercises-ex-23": {
"title": "Exercise 18.23",
"breadcrumb": "18-Learning-From-Examples",
"content" : "Consider an ensemble learning algorithm thatuses simple majority voting among $K$ learned hypotheses.Suppose that each hypothesis has error $epsilon$ and that the errorsmade by each hypothesis are independent of the others’. Calculate aformula for the error of the ensemble algorithm in terms of $K$and $epsilon$, and evaluate it for the cases where$K=5$, 10, and 20 and $epsilon={0.1}$, 0.2,and 0.4. If the independence assumption is removed, is it possible forthe ensemble error to be worse than $epsilon$?",
"url": " /concept-learning-exercises/ex_23/"
}
,
"concept-learning-exercises-ex-22": {
"title": "Exercise 18.22",
"breadcrumb": "18-Learning-From-Examples",
"content" : "Construct a support vector machine that computes thexor function. Use values of +1 and –1 (instead of 1 and 0)for both inputs and outputs, so that an example looks like $([-1, 1],1)$ or $([-1, -1], -1)$. Map the input $[x_1,x_2]$ into a spaceconsisting of $x_1$ and $x_1,x_2$. Draw the four input points in thisspace, and the maximal margin separator. What is the margin? Now drawthe separating line back in the original Euclidean input space.",
"url": " /concept-learning-exercises/ex_22/"
}
,
"concept-learning-exercises-ex-25": {
"title": "Exercise 18.25",
"breadcrumb": "18-Learning-From-Examples",
"content" : "A simple perceptron cannot represent xor (or, generally,the parity function of its inputs). Describe what happens to the weightsof a four-input, hard-threshold perceptron, beginning with all weightsset to 0.1, as examples of the parity function arrive.",
"url": " /concept-learning-exercises/ex_25/"
}
,
"concept-learning-exercises-ex-13": {
"title": "Exercise 18.13",
"breadcrumb": "18-Learning-From-Examples",
"content" : "InSection broadening-decision-tree-section, we noted thatattributes with many different possible values can cause problems withthe gain measure. Such attributes tend to split the examples intonumerous small classes or even singleton classes, thereby appearing tobe highly relevant according to the gain measure. Thegain-ratio criterion selects attributesaccording to the ratio between their gain and their intrinsicinformation content—that is, the amount of information contained in theanswer to the question, “What is the value of this attribute?” Thegain-ratio criterion therefore tries to measure how efficiently anattribute provides information on the correct classification of anexample. Write a mathematical expression for the information content ofan attribute, and implement the gain ratio criterion in DECISION-TREE-LEARNING.",
"url": " /concept-learning-exercises/ex_13/"
}
,
"concept-learning-exercises-ex-14": {
"title": "Exercise 18.14",
"breadcrumb": "18-Learning-From-Examples",
"content" : "Suppose you are running a learning experiment on a new algorithm forBoolean classification. You have a data set consisting of 100 positiveand 100 negative examples. You plan to use leave-one-outcross-validation and compare your algorithm to a baseline function, asimple majority classifier. (A majority classifier is given a set oftraining data and then always outputs the class that is in the majorityin the training set, regardless of the input.) You expect the majorityclassifier to score about 50% on leave-one-out cross-validation, but toyour surprise, it scores zero every time. Can you explain why?",
"url": " /concept-learning-exercises/ex_14/"
}
,
"concept-learning-exercises-ex-9": {
"title": "Exercise 18.9",
"breadcrumb": "18-Learning-From-Examples",
"content" : "Construct a data set (set of examples with attributes andclassifications) that would cause the decision-tree learning algorithmto find a non-minimal-sized tree. Show the tree constructed by thealgorithm and the minimal-sized tree that you can generate by hand.",
"url": " /concept-learning-exercises/ex_9/"
}
,
"concept-learning-exercises-ex-7": {
"title": "Exercise 18.7",
"breadcrumb": "18-Learning-From-Examples",
"content" : "[nonnegative-gain-exercise]Suppose that an attribute splits the set ofexamples $E$ into subsets $E_k$ and that each subset has $p_k$positive examples and $n_k$ negative examples. Show that theattribute has strictly positive information gain unless the ratio$p_k/(p_k+n_k)$ is the same for all $k$.",
"url": " /concept-learning-exercises/ex_7/"
}
,
"concept-learning-exercises-ex-31": {
"title": "Exercise 18.31",
"breadcrumb": "18-Learning-From-Examples",
"content" : "Suppose that a training set contains only a single example, repeated 100times. In 80 of the 100 cases, the single output value is 1; in theother 20, it is 0. What will a back-propagation network predict for thisexample, assuming that it has been trained and reaches a global optimum?(Hint: to find the global optimum, differentiate theerror function and set it to zero.)",
"url": " /concept-learning-exercises/ex_31/"
}
,
"concept-learning-exercises-ex-1": {
"title": "Exercise 18.1",
"breadcrumb": "18-Learning-From-Examples",
"content" : "Consider the problem faced by an infantlearning to speak and understand a language. Explain how this processfits into the general learning model. Describe the percepts and actionsof the infant, and the types of learning the infant must do. Describethe subfunctions the infant is trying to learn in terms of inputs andoutputs, and available example data.",
"url": " /concept-learning-exercises/ex_1/"
}
,
"concept-learning-exercises-ex-6": {
"title": "Exercise 18.6",
"breadcrumb": "18-Learning-From-Examples",
"content" : "In the recursive construction ofdecision trees, it sometimes happens that a mixed set of positive andnegative examples remains at a leaf node, even after all the attributeshave been used. Suppose that we have $p$ positive examples and $n$negative examples.1. Show that the solution used by DECISION-TREE-LEARNING, which picks the majority classification, minimizes the absolute error over the set of examples at the leaf.2. Show that the class probability $p/(p+n)$ minimizes the sum of squared errors.",
"url": " /concept-learning-exercises/ex_6/"
}
,
"concept-learning-exercises-ex-8": {
"title": "Exercise 18.8",
"breadcrumb": "18-Learning-From-Examples",
"content" : "Consider the following data set comprised of three binary inputattributes ($A_1, A_2$, and $A_3$) and one binary output:| $quad textbf{Example}$ | $quad A_1quad$ | $quad A_2quad$ | $quad A_3quad$ | $quad Outputspace y$ || --- | --- | --- | --- | --- || $textbf{x}_1$ | 1 | 0 | 0 | 0 || $textbf{x}_2$ | 1 | 0 | 1 | 0 || $textbf{x}_3$ | 0 | 1 | 0 | 0 || $textbf{x}_4$ | 1 | 1 | 1 | 1 || $textbf{x}_5$ | 1 | 1 | 0 | 1 |Use the algorithm in Figure DTL-algorithm(page DTL-algorithm) to learn a decision tree for these data. Show thecomputations made to determine the attribute to split at each node.",
"url": " /concept-learning-exercises/ex_8/"
}
,
"concept-learning-exercises-ex-30": {
"title": "Exercise 18.30",
"breadcrumb": "18-Learning-From-Examples",
"content" : "Implement a data structure for layered, feed-forward neural networks,remembering to provide the information needed for both forwardevaluation and backward propagation. Using this data structure, write afunction NEURAL-NETWORK-OUTPUT that takes an example and a network and computes theappropriate output values.",
"url": " /concept-learning-exercises/ex_30/"
}
,
"ilp-exercises-ex-3": {
"title": "Exercise 19.3",
"breadcrumb": "19-Knowledge-In-Learning",
"content" : "For each of the following determinations, write down the logicalrepresentation and explain why the determination is true (if it is):1. Zip code determines the state (U.S.).2. Design and denomination determine the mass of a coin.3. Climate, food intake, exercise, and metabolism determine weight gain and loss.4. Baldness is determined by the baldness (or lack thereof) of one’s maternal grandfather.",
"url": " /ilp-exercises/ex_3/"
}
,
"ilp-exercises-ex-4": {
"title": "Exercise 19.4",
"breadcrumb": "19-Knowledge-In-Learning",
"content" : "Would a probabilistic version of determinations be useful? Suggest adefinition.",
"url": " /ilp-exercises/ex_4/"
}
,
"ilp-exercises-ex-5": {
"title": "Exercise 19.5",
"breadcrumb": "19-Knowledge-In-Learning",
"content" : "Fill in the missing values for the clauses $C_1$ or$C_2$ (or both) in the following sets of clauses, given that $C$ is theresolvent of $C_1$ and $C_2$:1. $C = {True} Rightarrow P(A,B)$, $C_1 = P(x,y) Rightarrow Q(x,y)$, $C_2 = ??$.2. $C = {True} Rightarrow P(A,B)$, $C_1 = ??$, $C_2 = ??$.3. $C = P(x,y) Rightarrow P(x,f(y))$, $C_1 = ??$, $C_2 = ??$.If there is more than one possible solution, provide one example of eachdifferent kind.",
"url": " /ilp-exercises/ex_5/"
}
,
"ilp-exercises-ex-2": {
"title": "Exercise 19.2",
"breadcrumb": "19-Knowledge-In-Learning",
"content" : "For each of the following determinations, write down the logicalrepresentation and explain why the determination is true (if it is):1. Design and denomination determine the mass of a coin.2. For a given program, input determines output.3. Climate, food intake, exercise, and metabolism determine weight gain and loss.4. Baldness is determined by the baldness (or lack thereof) of one’s maternal grandfather. ",
"url": " /ilp-exercises/ex_2/"
}
,
"ilp-exercises-ex-7": {
"title": "Exercise 19.7",
"breadcrumb": "19-Knowledge-In-Learning",
"content" : "Suppose that is considering adding a literalto a clause using a binary predicate $P$ and that previous literals(including the head of the clause) contain five different variables.1. How many functionally different literals can be generated? Two literals are functionally identical if they differ only in the names of the *new* variables that they contain.2. Can you find a general formula for the number of different literals with a predicate of arity $r$ when there are $n$ variables previously used?3. Why does not allow literals that contain no previously used variables?",
"url": " /ilp-exercises/ex_7/"
}
,
"ilp-exercises-ex-1": {
"title": "Exercise 19.1",
"breadcrumb": "19-Knowledge-In-Learning",
"content" : "Show, by translating into conjunctive normal form andapplying resolution, that the conclusion drawn on page dbsig-pageconcerning Brazilians is sound.",
"url": " /ilp-exercises/ex_1/"
}
,
"ilp-exercises-ex-6": {
"title": "Exercise 19.6",
"breadcrumb": "19-Knowledge-In-Learning",
"content" : "Suppose one writes a logic program that carriesout a resolution inference step. That is, let ${Resolve}(c_1,c_2,c)$succeed if $c$ is the result of resolving $c_1$ and $c_2$. Normally,${Resolve}$ would be used as part of a theorem prover by calling itwith $c_1$ and $c_2$ instantiated to particular clauses, therebygenerating the resolvent $c$. Now suppose instead that we call it with$c$ instantiated and $c_1$ and $c_2$ uninstantiated. Will this succeedin generating the appropriate results of an inverse resolution step?Would you need any special modifications to the logic programming systemfor this to work?",
"url": " /ilp-exercises/ex_6/"
}
,
"ilp-exercises-ex-8": {
"title": "Exercise 19.8",
"breadcrumb": "19-Knowledge-In-Learning",
"content" : "Using the data from the family tree inFigure family2-figure, or a subset thereof, apply thealgorithm to learn a definition for the ${Ancestor}$ predicate.",
"url": " /ilp-exercises/ex_8/"
}
,
"nlp-english-exercises-ex-11": {
"title": "Exercise 23.11",
"breadcrumb": "23-Natural-Language-For-Communication",
"content" : "Consider the following toy grammar:> $S rightarrow NPspace VP$> $NP rightarrow Noun$> $NP rightarrow NPspace andspace NP$> $NP rightarrow NPspace PP$> $VP rightarrow Verb$> $VP rightarrow VPspace and space VP$> $VP rightarrow VPspace PP$> $PP rightarrow Prepspace NP$> $Noun rightarrow Sallyspace; poolsspace; streamsspace; swims$> $Prep rightarrow in$> $Verb rightarrow poolsspace; streamsspace; swims$1. Show all the parse trees in this grammar for the sentence “Sally swims in streams and pools.”2. Show all the table entries that would be made by a (non-probabalistic) CYK parser on this sentence.",
"url": " /nlp-english-exercises/ex_11/"
}
,
"nlp-english-exercises-ex-16": {
"title": "Exercise 23.16",
"breadcrumb": "23-Natural-Language-For-Communication",
"content" : "Consider the following sentence (from The New York Times,July 28, 2008):> Banks struggling to recover from multibillion-dollar loans on real> estate are curtailing loans to American businesses, depriving even> healthy companies of money for expansion and hiring.1. Which of the words in this sentence are lexically ambiguous?2. Find two cases of syntactic ambiguity in this sentence (there are more than two.)3. Give an instance of metaphor in this sentence.4. Can you find semantic ambiguity?",
"url": " /nlp-english-exercises/ex_16/"
}
,
"nlp-english-exercises-ex-20": {
"title": "Exercise 23.20",
"breadcrumb": "23-Natural-Language-For-Communication",
"content" : "(Adapted from [Knight:1999].) Our translation model assumes that, after the phrasetranslation model selects phrases and the distortion model permutesthem, the language model can unscramble the permutation. This exerciseinvestigates how sensible that assumption is. Try to unscramble theseproposed lists of phrases into the correct order:1. have, programming, a, seen, never, I, language, better2. loves, john, mary3. is the, communication, exchange of, intentional, information brought, by, about, the production, perception of, and signs, from, drawn, a, of, system, signs, conventional, shared4. created, that, we hold these, to be, all men, truths, are, equal, self-evidentWhich ones could you do? What type of knowledge did you draw upon? Traina bigram model from a training corpus, and use it to find thehighest-probability permutation of some sentences from a test corpus.Report on the accuracy of this model.",
"url": " /nlp-english-exercises/ex_20/"
}
,
"nlp-english-exercises-ex-18": {
"title": "Exercise 23.18",
"breadcrumb": "23-Natural-Language-For-Communication",
"content" : "Select five sentences and submit them to an online translation service.Translate them from English to another language and back to English.Rate the resulting sentences for grammaticality and preservation ofmeaning. Repeat the process; does the second round of iteration giveworse results or the same results? Does the choice of intermediatelanguage make a difference to the quality of the results? If you know aforeign language, look at the translation of one paragraph into thatlanguage. Count and describe the errors made, and conjecture why theseerrors were made.",
"url": " /nlp-english-exercises/ex_18/"
}
,
"nlp-english-exercises-ex-19": {
"title": "Exercise 23.19",
"breadcrumb": "23-Natural-Language-For-Communication",
"content" : "The $D_i$ values for the sentence inFigure mt-alignment-figure sum to 0. Will that be trueof every translation pair? Prove it or give a counterexample.",
"url": " /nlp-english-exercises/ex_19/"
}
,
"nlp-english-exercises-ex-21": {
"title": "Exercise 23.21",
"breadcrumb": "23-Natural-Language-For-Communication",
"content" : "Calculate the most probable path through the HMM inFigure sr-hmm-figure for the output sequence$[C_1,C_2,C_3,C_4,C_4,C_6,C_7]$. Also give its probability.",
"url": " /nlp-english-exercises/ex_21/"
}
,
"nlp-english-exercises-ex-17": {
"title": "Exercise 23.17",
"breadcrumb": "23-Natural-Language-For-Communication",
"content" : "Without looking back atExercise washing-clothes-exercise, answer the followingquestions:1. What are the four steps that are mentioned?2. What step is left out?3. What is “the material” that is mentioned in the text?4. What kind of mistake would be expensive?5. Is it better to do too few things or too many? Why?",
"url": " /nlp-english-exercises/ex_17/"
}
,
"nlp-english-exercises-ex-10": {
"title": "Exercise 23.10",
"breadcrumb": "23-Natural-Language-For-Communication",
"content" : "In this exercise you will transform $large varepsilon_0$ intoChomsky Normal Form (CNF). There are five steps: (a) Add a new startsymbol, (b) Eliminate $epsilon$ rules, (c) Eliminate multiple words onright-hand sides, (d) Eliminate rules of the form(${it X}$$rightarrow$${it Y}$),(e) Convert long right-hand sides into binary rules.1. The start symbol, $S$, can occur only on the left-hand side in CNF. Replace ${it S}$ everywhere by a new symbol ${it S'}$ and add a rule of the form ${it S}$ $rightarrow$${it S'}$.2. The empty string, $epsilon$ cannot appear on the right-hand side in CNF. $large varepsilon_0$ does not have any rules with $epsilon$, so this is not an issue.3. A word can appear on the right-hand side in a rule only of the form (${it X}$ $rightarrow$*word*). Replace each rule of the form (${it X}$ $rightarrow$…*word* …) with (${it X}$ $rightarrow$…${it W'}$ …) and (${it W'}$ $rightarrow$*word*), using a new symbol ${it W'}$.4. A rule (${it X}$ $rightarrow$${it Y}$) is not allowed in CNF; it must be (${it X}$ $rightarrow$${it Y}$ ${it Z}$) or (${it X}$ $rightarrow$*word*). Replace each rule of the form (${it X}$ $rightarrow$${it Y}$) with a set of rules of the form (${it X}$ $rightarrow$…), one for each rule (${it Y}$ $rightarrow$…), where (…) indicates one or more symbols.5. Replace each rule of the form (${it X}$ $rightarrow$${it Y}$ ${it Z}$ …) with two rules, (${it X}$ $rightarrow$${it Y}$ ${it Z'}$) and (${it Z'}$ $rightarrow$${it Z}$ …), where ${it Z'}$ is a new symbol.Show each step of the process and the final set of rules.",
"url": " /nlp-english-exercises/ex_10/"
}
,
"nlp-english-exercises-ex-3": {
"title": "Exercise 23.3",
"breadcrumb": "23-Natural-Language-For-Communication",
"content" : "Consider the following PCFG for simple verb phrases:> 0.1: VP $rightarrow$ Verb> 0.2: VP $rightarrow$ Copula Adjective> 0.5: VP $rightarrow$ Verb the Noun> 0.2: VP $rightarrow$ VP Adverb> 0.5: Verb $rightarrow$ is> 0.5: Verb $rightarrow$ shoots> 0.8: Copula $rightarrow$ is> 0.2: Copula $rightarrow$ seems> 0.5: Adjective $rightarrow$ unwell> 0.5: Adjective $rightarrow$ well> 0.5: Adverb $rightarrow$ well> 0.5: Adverb $rightarrow$ badly> 0.6: Noun $rightarrow$ duck> 0.4: Noun $rightarrow$ well1. Which of the following have a nonzero probability as a VP? (i) shoots the duck well well well(ii) seems the well well(iii) shoots the unwell well badly2. What is the probability of generating “is well well”?3. What types of ambiguity are exhibited by the phrase in (b)?4. Given any PCFG, is it possible to calculate the probability that the PCFG generates a string of exactly 10 words?",
"url": " /nlp-english-exercises/ex_3/"
}
,
"nlp-english-exercises-ex-4": {
"title": "Exercise 23.4",
"breadcrumb": "23-Natural-Language-For-Communication",
"content" : "Consider the following simple PCFG for noun phrases:> 0.6: NP $rightarrow$ Det AdjString Noun> 0.4: NP $rightarrow$ Det NounNounCompound> 0.5: AdjString $rightarrow$ Adj AdjString> 0.5: AdjString $rightarrow$ $Lambda$> 1.0: NounNounCompound $rightarrow$ Noun> 0.8: Det $rightarrow$ the> 0.2: Det $rightarrow$ a> 0.5: Adj $rightarrow$ small> 0.5: Adj $rightarrow$ green> 0.6: Noun $rightarrow$ village> 0.4: Noun $rightarrow$ greenwhere $Lambda$ denotes the empty string.1. What is the longest NP that can be generated by this grammar? (i) three words(ii) four words(iii) infinitely many words2. Which of the following have a nonzero probability of being generated as complete NPs? (i) a small green village(ii) a green green green(iii) a small village green3. What is the probability of generating “the green green”?4. What types of ambiguity are exhibited by the phrase in (c)?5. Given any PCFG and any finite word sequence, is it possible to calculate the probability that the sequence was generated by the PCFG?",
"url": " /nlp-english-exercises/ex_4/"
}
,
"nlp-english-exercises-ex-5": {
"title": "Exercise 23.5",
"breadcrumb": "23-Natural-Language-For-Communication",
"content" : "Outline the major differences between Java (or any other computerlanguage with which you are familiar) and English, commenting on the“understanding” problem in each case. Think about such things asgrammar, syntax, semantics, pragmatics, compositionality,context-dependence, lexical ambiguity, syntactic ambiguity, referencefinding (including pronouns), background knowledge, and what it means to“understand” in the first place.",
"url": " /nlp-english-exercises/ex_5/"
}
,
"nlp-english-exercises-ex-2": {
"title": "Exercise 23.2",
"breadcrumb": "23-Natural-Language-For-Communication",
"content" : "An HMM grammar is essentially a standard HMM whose statevariable is $N$ (nonterminal, with values such as $Det$, $Adjective$,$Noun$ and so on) and whose evidence variable is $W$ (word, with valuessuch as $is$, $duck$, and so on). The HMM model includes a prior${textbf{P}}(N_0)$, a transition model${textbf{P}}(N_{t+1}|N_t)$, and a sensor model${textbf{P}}(W_t|N_t)$. Show that every HMM grammar can bewritten as a PCFG. [Hint: start by thinking about how the HMM prior canbe represented by PCFG rules for the sentence symbol. You may find ithelpful to illustrate for the particular HMM with values $A$, $B$ for$N$ and values $x$, $y$ for $W$.]",
"url": " /nlp-english-exercises/ex_2/"
}
,
"nlp-english-exercises-ex-15": {
"title": "Exercise 23.15",
"breadcrumb": "23-Natural-Language-For-Communication",
"content" : "Augment the $large varepsilon_1$ grammar so that it handles article–noun agreement. That is,make sure that “agents” and “an agent” are ${it NP}$s, but“agent” and “an agents” are not.",
"url": " /nlp-english-exercises/ex_15/"
}
,
"nlp-english-exercises-ex-12": {
"title": "Exercise 23.12",
"breadcrumb": "23-Natural-Language-For-Communication",
"content" : "Using DCG notation, write a grammar for alanguage that is just like $large varepsilon_1$, except that it enforces agreement betweenthe subject and verb of a sentence and thus does not generateungrammatical sentences such as “I smells the wumpus.”",
"url": " /nlp-english-exercises/ex_12/"
}
,
"nlp-english-exercises-ex-22": {
"title": "Exercise 23.22",
"breadcrumb": "23-Natural-Language-For-Communication",
"content" : "We forgot to mention that the text inExercise washing-clothes-exercise is entitled “WashingClothes.” Reread the text and answer the questions inExercise washing-clothes2-exercise. Did you do betterthis time? Bransford and Johnson [Bransford+Johnson:1973] used thistext in a controlled experiment and found that the title helpedsignificantly. What does this tell you about how language and memoryworks?",
"url": " /nlp-english-exercises/ex_22/"
}
,
"nlp-english-exercises-ex-13": {
"title": "Exercise 23.13",
"breadcrumb": "23-Natural-Language-For-Communication",
"content" : "Consider the following PCFG:> $S rightarrow NP space VP[1.0] $> $NP rightarrow textit{Noun}[0.6] space|space textit{Pronoun}[0.4] $> $VP rightarrow textit{Verb} space NP[0.8] space|space textit{Modal}space textit{Verb}[0.2]$> $textit{Noun} rightarrow textbf{can}[0.1] space|space textbf{fish}[0.3] space|space ...$> $textit{Pronoun} rightarrow textbf{I}[0.4] space|space ...$> $textit{Verb} rightarrow textbf{can}[0.01] space|space textbf{fish}[0.1] space|space ...$> $textit{Modal} rightarrow textbf{can}[0.3] space|space ...$The sentence “I can fish” has two parse trees with this grammar. Showthe two trees, their prior probabilities, and their conditionalprobabilities, given the sentence.",
"url": " /nlp-english-exercises/ex_13/"
}
,
"nlp-english-exercises-ex-14": {
"title": "Exercise 23.14",
"breadcrumb": "23-Natural-Language-For-Communication",
"content" : "An augmented context-free grammar can represent languages that a regularcontext-free grammar cannot. Show an augmented context-free grammar forthe language $a^nb^nc^n$. The allowable values for augmentationvariables are 1 and $SUCCESSOR(n)$, where $n$ is a value. The rule for a sentencein this language is$$S(n) rightarrow}}A(n) B(n) C(n) .$$Show the rule(s) for each of ${it A}$,${it B}$, and ${it C}$.",
"url": " /nlp-english-exercises/ex_14/"
}
,
"nlp-english-exercises-ex-9": {
"title": "Exercise 23.9",
"breadcrumb": "23-Natural-Language-For-Communication",
"content" : "Some linguists have argued as follows: Children learning a language hear only positive examples of the language and no negative examples. Therefore, the hypothesis that “every possible sentence is in the language” is consistent with all the observed examples. Moreover, this is the simplest consistent hypothesis. Furthermore, all grammars for languages that are supersets of the true language are also consistent with the observed data. Yet children do induce (more or less) the right grammar. It follows that they begin with very strong innate grammatical constraints that rule out all of these more general hypotheses a priori.Comment on the weak point(s) in this argument from a statisticallearning viewpoint.",
"url": " /nlp-english-exercises/ex_9/"
}
,
"nlp-english-exercises-ex-7": {
"title": "Exercise 23.7",
"breadcrumb": "23-Natural-Language-For-Communication",
"content" : "Consider the sentence “Someone walked slowly to the supermarket” and alexicon consisting of the following words:$Pronoun rightarrow textbf{someone} quad Verb rightarrow textbf{walked}$$Adv rightarrow textbf{slowly} quad Prep rightarrow textbf{to}$$Article rightarrow textbf{the} quad Noun rightarrow textbf{supermarket}$Which of the following three grammars, combined with the lexicon,generates the given sentence? Show the corresponding parse tree(s).| $quadquadquadquad (A):quadquadquadquad$ | $quadquadquadquad(B):quadquadquadquad$ | $quadquadquadquad(C):quadquadquadquad$ || --- | --- | --- || $Srightarrow NPspace VP$ | $Srightarrow NPspace VP$ | $Srightarrow NPspace VP$ || $NPrightarrow Pronoun$ | $NPrightarrow Pronoun$ | $NPrightarrow Pronoun$ || $NPrightarrow Articlespace Noun $ | $NPrightarrow Noun$ | $NPrightarrow Articlespace NP$ || $VPrightarrow VPspace PP$ | $NPrightarrow Articlespace NP$ | $VPrightarrow Verbspace Adv$ || $VPrightarrow VPspace Advspace Adv$ | $VPrightarrow Verbspace Vmod$ | $Advrightarrow Advspace Adv$ || $VPrightarrow Verb$ | $Vmodrightarrow Advspace Vmod$ | $Advrightarrow PP$ || $PPrightarrow Prepspace NP$ | $Vmodrightarrow Adv$ | $PPrightarrow Prepspace NP$ || $NPrightarrow Noun$ | $Advrightarrow PP$ | $NPrightarrow Noun$ || $quad$ | $PPrightarrow Prepspace NP$ | $quad$ |For each of the preceding three grammars, write down three sentences ofEnglish and three sentences of non-English generated by the grammar.Each sentence should be significantly different, should be at least sixwords long, and should include some new lexical entries (which youshould define). Suggest ways to improve each grammar to avoid generatingthe non-English sentences.",
"url": " /nlp-english-exercises/ex_7/"
}
,
"nlp-english-exercises-ex-1": {
"title": "Exercise 23.1",
"breadcrumb": "23-Natural-Language-For-Communication",
"content" : "Read the following text once forunderstanding, and remember as much of it as you can. There will be atest later.> The procedure is actually quite simple. First you arrange things intodifferent groups. Of course, one pile may be sufficient depending on howmuch there is to do. If you have to go somewhere else due to lack offacilities that is the next step, otherwise you are pretty well set. Itis important not to overdo things. That is, it is better to do too fewthings at once than too many. In the short run this may not seemimportant but complications can easily arise. A mistake is expensive aswell. At first the whole procedure will seem complicated. Soon, however,it will become just another facet of life. It is difficult to foreseeany end to the necessity for this task in the immediate future, but thenone can never tell. After the procedure is completed one arranges thematerial into different groups again. Then they can be put into theirappropriate places. Eventually they will be used once more and the wholecycle will have to be repeated. However, this is part of life.",
"url": " /nlp-english-exercises/ex_1/"
}
,
"nlp-english-exercises-ex-6": {
"title": "Exercise 23.6",
"breadcrumb": "23-Natural-Language-For-Communication",
"content" : "This exercise concerns grammars for very simple languages.1. Write a context-free grammar for the language $a^n b^n$.2. Write a context-free grammar for the palindrome language: the set of all strings whose second half is the reverse of the first half.3. Write a context-sensitive grammar for the duplicate language: the set of all strings whose second half is the same as the first half.",
"url": " /nlp-english-exercises/ex_6/"
}
,
"nlp-english-exercises-ex-8": {
"title": "Exercise 23.8",
"breadcrumb": "23-Natural-Language-For-Communication",
"content" : "Collect some examples of time expressions, such as “two o’clock,”“midnight,” and “12:46.” Also think up some examples that areungrammatical, such as “thirteen o’clock” or “half past two fifteen.”Write a grammar for the time language.",
"url": " /nlp-english-exercises/ex_8/"
}
,
"probability-exercises-ex-11": {
"title": "Exercise 13.11",
"breadcrumb": "13-Quantifying-Uncertainity",
"content" : "Deciding to put probability theory to good use, we encounter a slotmachine with three independent wheels, each producing one of the foursymbols bar, bell, lemon, orcherry with equal probability. The slot machine has thefollowing payout scheme for a bet of 1 coin (where “?” denotes that wedon’t care what comes up for that wheel): > bar/bar/bar pays 20 coins> bell/bell/bell pays 15 coins> lemon/lemon/lemon pays 5 coins> cherry/cherry/cherry pays 3 coins> cherry/cherry/? pays 2 coins> cherry/?/? pays 1 coin1. Compute the expected “payback” percentage of the machine. In other words, for each coin played, what is the expected coin return?2. Compute the probability that playing the slot machine once will result in a win.3. Estimate the mean and median number of plays you can expect to make until you go broke, if you start with 10 coins. You can run a simulation to estimate this, rather than trying to compute an exact answer.",
"url": " /probability-exercises/ex_11/"
}
,
"probability-exercises-ex-16": {
"title": "Exercise 13.16",
"breadcrumb": "13-Quantifying-Uncertainity",
"content" : "Consider two medical tests, A and B, for a virus. Test A is 95%effective at recognizing the virus when it is present, but has a 10%false positive rate (indicating that the virus is present, when it isnot). Test B is 90% effective at recognizing the virus, but has a 5%false positive rate. The two tests use independent methods ofidentifying the virus. The virus is carried by 1% of all people. Saythat a person is tested for the virus using only one of the tests, andthat test comes back positive for carrying the virus. Which testreturning positive is more indicative of someone really carrying thevirus? Justify your answer mathematically.",
"url": " /probability-exercises/ex_16/"
}
,
"probability-exercises-ex-29": {
"title": "Exercise 13.29",
"breadcrumb": "13-Quantifying-Uncertainity",
"content" : "In our analysis of the wumpus world, we used the fact thateach square contains a pit with probability 0.2, independently of thecontents of the other squares. Suppose instead that exactly $N/5$ pitsare scattered at random among the $N$ squares other than [1,1]. Arethe variables $P_{i,j}$ and $P_{k,l}$ still independent? What is thejoint distribution ${textbf{P}}(P_{1,1},ldots,P_{4,4})$ now?Redo the calculation for the probabilities of pits in [1,3] and[2,2].",
"url": " /probability-exercises/ex_29/"
}
,
"probability-exercises-ex-20": {
"title": "Exercise 13.20",
"breadcrumb": "13-Quantifying-Uncertainity",
"content" : "It is quite often useful to consider theeffect of some specific propositions in the context of some generalbackground evidence that remains fixed, rather than in the completeabsence of information. The following questions ask you to prove moregeneral versions of the product rule and Bayes’ rule, with respect tosome background evidence $textbf{e}$: 1. Prove the conditionalized version of the general product rule: $${textbf{P}}(X,Y textbf{e}) = {textbf{P}}(XY,textbf{e}) {textbf{P}}(Ytextbf{e}) .$$ 2. Prove the conditionalized version of Bayes’ rule in Equation (conditional-bayes-equation). ",
"url": " /probability-exercises/ex_20/"
}
,
"probability-exercises-ex-27": {
"title": "Exercise 13.27",
"breadcrumb": "13-Quantifying-Uncertainity",
"content" : "Write out a general algorithm for answering queries of the form${textbf{P}}({Cause}textbf{e})$, using a naive Bayesdistribution. Assume that the evidence $textbf{e}$ may assign values toany subset of the effect variables.",
"url": " /probability-exercises/ex_27/"
}
,
"probability-exercises-ex-18": {
"title": "Exercise 13.18",
"breadcrumb": "13-Quantifying-Uncertainity",
"content" : "After your yearly checkup, the doctor has bad news and good news. Thebad news is that you tested positive for a serious disease and that thetest is 99% accurate (i.e., the probability of testing positive when youdo have the disease is 0.99, as is the probability of testing negativewhen you don’t have the disease). The good news is that this is a raredisease, striking only 1 in 10,000 people of your age. Why is it goodnews that the disease is rare? What are the chances that you actuallyhave the disease?",
"url": " /probability-exercises/ex_18/"
}
,
"probability-exercises-ex-26": {
"title": "Exercise 13.26",
"breadcrumb": "13-Quantifying-Uncertainity",
"content" : "(Adapted from Pearl [Pearl:1988].) Suppose you are a witness to anighttime hit-and-run accident involving a taxi in Athens. All taxis inAthens are blue or green. You swear, under oath, that the taxi was blue.Extensive testing shows that, under the dim lighting conditions,discrimination between blue and green is 75% reliable. 1. Is it possible to calculate the most likely color for the taxi? (*Hint:* distinguish carefully between the proposition that the taxi *is* blue and the proposition that it *appears* blue.) 2. What if you know that 9 out of 10 Athenian taxis are green?",
"url": " /probability-exercises/ex_26/"
}
,
"probability-exercises-ex-19": {
"title": "Exercise 13.19",
"breadcrumb": "13-Quantifying-Uncertainity",
"content" : "After your yearly checkup, the doctor has bad news and good news. Thebad news is that you tested positive for a serious disease and that thetest is 99% accurate (i.e., the probability of testing positive when youdo have the disease is 0.99, as is the probability of testing negativewhen you don’t have the disease). The good news is that this is a raredisease, striking only 1 in 100,000 people of your age. Why is it goodnews that the disease is rare? What are the chances that you actuallyhave the disease?",
"url": " /probability-exercises/ex_19/"
}
,
"probability-exercises-ex-21": {
"title": "Exercise 13.21",
"breadcrumb": "13-Quantifying-Uncertainity",
"content" : "Show that the statement of conditional independence$${textbf{P}}(X,Y Z) = {textbf{P}}(XZ) {textbf{P}}(YZ)$$is equivalent to each of the statements$${textbf{P}}(XY,Z) = {textbf{P}}(XZ) quadmbox{and}quad {textbf{P}}(YX,Z) = {textbf{P}}(YZ) .$$",
"url": " /probability-exercises/ex_21/"
}
,
"probability-exercises-ex-17": {
"title": "Exercise 13.17",
"breadcrumb": "13-Quantifying-Uncertainity",
"content" : "Suppose you are given a coin that lands ${heads}$ with probability $x$and ${tails}$ with probability $1 - x$. Are the outcomes of successiveflips of the coin independent of each other given that you know thevalue of $x$? Are the outcomes of successive flips of the coinindependent of each other if you do not know the value of$x$? Justify your answer.",
"url": " /probability-exercises/ex_17/"
}
,
"probability-exercises-ex-28": {
"title": "Exercise 13.28",
"breadcrumb": "13-Quantifying-Uncertainity",
"content" : "Text categorization is the task ofassigning a given document to one of a fixed set of categories on thebasis of the text it contains. Naive Bayes models are often used forthis task. In these models, the query variable is the document category,and the “effect” variables are the presence or absence of each word inthe language; the assumption is that words occur independently indocuments, with frequencies determined by the document category.1. Explain precisely how such a model can be constructed, given as “training data” a set of documents that have been assigned to categories.2. Explain precisely how to categorize a new document.3. Is the conditional independence assumption reasonable? Discuss.",
"url": " /probability-exercises/ex_28/"
}
,
"probability-exercises-ex-10": {
"title": "Exercise 13.10",
"breadcrumb": "13-Quantifying-Uncertainity",
"content" : "In his letter of August 24, 1654, Pascalwas trying to show how a pot of money should be allocated when agambling game must end prematurely. Imagine a game where each turnconsists of the roll of a die, player E gets a point whenthe die is even, and player O gets a point when the dieis odd. The first player to get 7 points wins the pot. Suppose the gameis interrupted with E leading 4–2. How should the moneybe fairly split in this case? What is the general formula? (Fermat andPascal made several errors before solving the problem, but you should beable to get it right the first time.)",
"url": " /probability-exercises/ex_10/"
}
,
"probability-exercises-ex-3": {
"title": "Exercise 13.3",
"breadcrumb": "13-Quantifying-Uncertainity",
"content" : "For each of the following statements, either prove it is true or give acounterexample.1. If $P(a b, c) = P(b a, c)$, then $P(a c) = P(b c)$ 2. If $P(a b, c) = P(a)$, then $P(b c) = P(b)$ 3. If $P(a b) = P(a)$, then $P(a b, c) = P(a c)$",
"url": " /probability-exercises/ex_3/"
}
,
"probability-exercises-ex-4": {
"title": "Exercise 13.4",
"breadcrumb": "13-Quantifying-Uncertainity",
"content" : "Would it be rational for an agent to hold the three beliefs$P(A) {0.4}$, $P(B) {0.3}$, and$P(A lor B) {0.5}$? If so, what range of probabilities wouldbe rational for the agent to hold for $A land B$? Make up a table likethe one in Figure de-finetti-table, and show how itsupports your argument about rationality. Then draw another version ofthe table where $P(A lor B){0.7}$. Explain why it is rational to have this probability,even though the table shows one case that is a loss and three that justbreak even. (Hint: what is Agent 1 committed to about theprobability of each of the four cases, especially the case that is aloss?)",
"url": " /probability-exercises/ex_4/"
}
,
"probability-exercises-ex-5": {
"title": "Exercise 13.5",
"breadcrumb": "13-Quantifying-Uncertainity",
"content" : "This question deals with the propertiesof possible worlds, defined on page possible-worlds-page as assignments to allrandom variables. We will work with propositions that correspond toexactly one possible world because they pin down the assignments of allthe variables. In probability theory, such propositions are called atomic event. Forexample, with Boolean variables $X_1$, $X_2$, $X_3$, the proposition$x_1land lnot x_2 land lnot x_3$ fixes the assignment of thevariables; in the language of propositional logic, we would say it hasexactly one model.1. Prove, for the case of $n$ Boolean variables, that any two distinct atomic events are mutually exclusive; that is, their conjunction is equivalent to ${false}$.2. Prove that the disjunction of all possible atomic events is logically equivalent to ${true}$.3. Prove that any proposition is logically equivalent to the disjunction of the atomic events that entail its truth.",
"url": " /probability-exercises/ex_5/"
}
,
"probability-exercises-ex-2": {
"title": "Exercise 13.2",
"breadcrumb": "13-Quantifying-Uncertainity",
"content" : "Using the axioms of probability, prove that anyprobability distribution on a discrete random variable must sum to 1.",
"url": " /probability-exercises/ex_2/"
}
,
"probability-exercises-ex-15": {
"title": "Exercise 13.15",
"breadcrumb": "13-Quantifying-Uncertainity",
"content" : "Show that the three forms of independence inEquation (independence-equation) are equivalent.",
"url": " /probability-exercises/ex_15/"
}
,
"probability-exercises-ex-12": {
"title": "Exercise 13.12",
"breadcrumb": "13-Quantifying-Uncertainity",
"content" : "Deciding to put probability theory to good use, we encounter a slotmachine with three independent wheels, each producing one of the foursymbols bar, bell, lemon, orcherry with equal probability. The slot machine has thefollowing payout scheme for a bet of 1 coin (where “?” denotes that wedon’t care what comes up for that wheel): > bar/bar/bar pays 20 coins> bell/bell/bell pays 15 coins> lemon/lemon/lemon pays 5 coins> cherry/cherry/cherry pays 3 coins> cherry/cherry/? pays 2 coins> cherry/?/? pays 1 coin1. Compute the expected “payback” percentage of the machine. In other words, for each coin played, what is the expected coin return?2. Compute the probability that playing the slot machine once will result in a win.3. Estimate the mean and median number of plays you can expect to make until you go broke, if you start with 10 coins. You can run a simulation to estimate this, rather than trying to compute an exact answer.",
"url": " /probability-exercises/ex_12/"
}
,
"probability-exercises-ex-24": {
"title": "Exercise 13.24",
"breadcrumb": "13-Quantifying-Uncertainity",
"content" : "This exercise investigates the way in which conditional independencerelationships affect the amount of information needed for probabilisticcalculations.1. Suppose we wish to calculate $P(he_1,e_2)$ and we have no conditional independence information. Which of the following sets of numbers are sufficient for the calculation? 1. ${textbf{P}}(E_1,E_2)$, ${textbf{P}}(H)$, ${textbf{P}}(E_1H)$, ${textbf{P}}(E_2H)$ 2. ${textbf{P}}(E_1,E_2)$, ${textbf{P}}(H)$, ${textbf{P}}(E_1,E_2H)$ 3. ${textbf{P}}(H)$, ${textbf{P}}(E_1H)$, ${textbf{P}}(E_2H)$2. Suppose we know that ${textbf{P}}(E_1H,E_2)={textbf{P}}(E_1H)$ for all values of $H$, $E_1$, $E_2$. Now which of the three sets are sufficient?",
"url": " /probability-exercises/ex_24/"
}