Neural Networks Multiple Choice Questions on “Learning – 1″.
1. On what parameters can change in weight vector depend?
A. learning parameters
B. input vector
C. learning signal
D. all of the mentioned
Answer: D
Clarification: Change in weight vector corresponding to jth input at time (t+1) depends on all of these parameters.
2. If the change in weight vector is represented by ∆wij, what does it mean?
A. describes the change in weight vector for ith processing unit, taking input vector jth into account
B. describes the change in weight vector for jth processing unit, taking input vector ith into account
C. describes the change in weight vector for jth & ith processing unit.
D. none of the mentioned
Answer: A
Clarification: ∆wij= µf(wi A.aj, where a is the input vector.
3. What is learning signal in this equation ∆wij= µf(wi A.aj?
A. µ
B. wi a
C. aj
D. f(wi A.
Answer: D
Clarification: This the non linear representation of output of the network.
4. State whether Hebb’s law is supervised learning or of unsupervised type?
A. supervised
B. unsupervised
C. either supervised or unsupervised
D. can be both supervised & unsupervised
Answer: B
Clarification: No desired output is required for it’s implementation.
5. Hebb’s law can be represented by equation?
A. ∆wij= µf(wi A.aj
B. ∆wij= µ(si) aj, where (si) is output signal of ith input
C. both way
D. none of the mentioned
Answer: C
Clarification: (si)= f(wi A., in Hebb’s law.
6. State which of the following statements hold foe perceptron learning law?
A. it is supervised type of learning law
B. it requires desired output for each input
C. ∆wij= µ(bi – si) aj
D. all of the mentioned
Answer: D
Clarification: all statements follow from ∆wij= µ(bi – si) aj, where bi is the target output & hence supervised learning.
7. Delta learning is of unsupervised type?
A. yes
B. no
Answer: B
Clarification: Change in weight is based on the error between the desired & the actual output values for a given input.
8. widrow & hoff learning law is special case of?
A. hebb learning law
B. perceptron learning law
C. delta learning law
D. none of the mentioned
Answer: C
Clarification: Output function in this law is assumed to be linear , all other things same.
9. What’s the other name of widrow & hoff learning law?
A. Hebb
B. LMS
C. MMS
D. None of the mentioned
Answer: B
Clarification: LMS, least mean square. Change in weight is made proportional to negative gradient of error & due to linearity of output function.
10. Which of the following equation represent perceptron learning law?
A. ∆wij= µ(si) aj
B. ∆wij= µ(bi – si) aj
C. ∆wij= µ(bi – si) aj Á(xi),wher Á(xi) is derivative of xi
D. ∆wij= µ(bi – (wi A.) aj
Answer: B
Clarification: Perceptron learning law is supervised, nonlinear type of learning.