MCQOPTIONS
Saved Bookmarks
This section includes 12 Mcqs, each offering curated multiple-choice questions to sharpen your Neural Networks knowledge and support exam preparation. Choose a topic below to get started.
| 1. |
How can learning process be stopped in backpropagation rule? |
| A. | there is convergence involved |
| B. | no heuristic criteria exist |
| C. | on basis of average gradient value |
| D. | none of the mentioned |
| Answer» D. none of the mentioned | |
| 2. |
Does backpropagaion learning is based on gradient descent along error surface? |
| A. | yes |
| B. | no |
| C. | cannot be said |
| D. | it depends on gradient descent but not error surface |
| Answer» B. no | |
| 3. |
What is meant by generalized in statement “backpropagation is a generalized delta rule” ? |
| A. | because delta rule can be extended to hidden layer units |
| B. | because delta is applied to only input and output layers, thus making it more simple and generalized |
| C. | it has no significance |
| D. | none of the mentioned |
| Answer» B. because delta is applied to only input and output layers, thus making it more simple and generalized | |
| 4. |
DOES_BACKPROPAGAION_LEARNING_IS_BASED_ON_GRADIENT_DESCENT_ALONG_ERROR_SURFACE??$ |
| A. | yes |
| B. | no |
| C. | cannot be said |
| D. | it depends on gradient descent but not error surface |
| Answer» B. no | |
| 5. |
How_can_learning_process_be_stopped_in_backpropagation_rule?$ |
| A. | there is convergence involved |
| B. | no heuristic criteria exist |
| C. | on basis of average gradient value |
| D. | none of the mentioned |
| Answer» D. none of the mentioned | |
| 6. |
What are the general tasks that are performed with backpropagation algorithm? |
| A. | pattern mapping |
| B. | function approximation |
| C. | prediction |
| D. | all of the mentioned |
| Answer» E. | |
| 7. |
What are general limitations of back propagation rule? |
| A. | local minima problem |
| B. | slow convergence |
| C. | scaling |
| D. | all of the mentioned |
| Answer» E. | |
| 8. |
What is meant by generalized in statement “backpropagation is a generalized delta rule” ?$ |
| A. | because delta rule can be extended to hidden layer units |
| B. | because delta is applied to only input and output layers, thus making it more simple and generalized |
| C. | it has no significance |
| D. | none of the mentioned |
| Answer» B. because delta is applied to only input and output layers, thus making it more simple and generalized | |
| 9. |
There is feedback in final stage of backpropagation algorithm? |
| A. | yes |
| B. | no |
| Answer» C. | |
| 10. |
What is true regarding backpropagation rule? |
| A. | it is also called generalized delta rule |
| B. | error in output is propagated backwards only to determine weight updates |
| C. | there is no feedback of signal at nay stage |
| D. | all of the mentioned |
| Answer» E. | |
| 11. |
The backpropagation law is also known as generalized delta rule, is it true? |
| A. | yes |
| B. | no |
| Answer» B. no | |
| 12. |
What is the objective of backpropagation algorithm? |
| A. | to develop learning algorithm for multilayer feedforward neural network |
| B. | to develop learning algorithm for single layer feedforward neural network |
| C. | to develop learning algorithm for multilayer feedforward neural network, so that network can be trained to capture the mapping implicitly |
| D. | none of the mentioned |
| Answer» D. none of the mentioned | |