Rr420507 Neural Networks

  • Uploaded by: Justin Cook
  • 0
  • 0
  • June 2020
  • PDF

This document was uploaded by user and they confirmed that they have the permission to share it. If you are author or own the copyright of this book, please report to us by using this DMCA report form. Report DMCA


Overview

Download & View Rr420507 Neural Networks as PDF for free.

More details

  • Words: 967
  • Pages: 4
Set No. 1

Code No: RR420507

IV B.Tech II Semester Regular Examinations, Apr/May 2007 NEURAL NETWORKS ( Common to Computer Science & Engineering and Electronics & Computer Engineering) Time: 3 hours Max Marks: 80 Answer any FIVE Questions All Questions carry equal marks ⋆⋆⋆⋆⋆ 1. (a) Give a brief description of neural networks as optimizing networks. (b) Explain the use of ANNs for clustering and feature detection.

[8] [4+4]

2. Briefly discuss about linear separability and the solution for EX-OR problem.Also suggest a network that can solve EX-OR problem. [4+6+6] 3. Explain about the generalized delta- rule and derive the weight updatation for a multi layer feed forward neural network. [8+8] 4. Construct an energy function for a discrete Hopfield neural network of size N×N neurons. Show that the energy function decreases every time the neuron output is changed. [8+8] 5. Discuss how the “Winner-Take-All” in the Kohonen’s layer is implemented and explain the architecture, Also explain the training algorithm. [16] 6. Derive expressions for the weight updation involved in counter propagation. [16] 7. Give a detailed note on the following: (a) ART1 data structures.

[8]

(b) ART2 simulation.

[8]

8. Describe how a neural network may be trained for a pattern recognition task. Illustrate with an example [16] ⋆⋆⋆⋆⋆

1 of 1

Set No. 2

Code No: RR420507

IV B.Tech II Semester Regular Examinations, Apr/May 2007 NEURAL NETWORKS ( Common to Computer Science & Engineering and Electronics & Computer Engineering) Time: 3 hours Max Marks: 80 Answer any FIVE Questions All Questions carry equal marks ⋆⋆⋆⋆⋆ 1. (a) What is the Hebbian-learning rule for training neural networks? Explain with the help of an illustration. [4+4] (b) What is the delta learning rule in neural networks? Explain with the help of an Illustration. [4+4] 2. State and prove the perceptron convergence theorem.

[2+14]

3. Explain about the generalized delta- rule and derive the weight updatation for a multi layer feed forward neural network. [8+8] 4. The truncated energy function, E(v), of a certain two-neuron network is specified as E (v) = − 12 (v12 + 2v1 v2 + 4v22 + v1 ) ,Assuming high-gain neurons, (a) find the weight matrix W and the bias current vector i.

[8]

(b) Determine whether single-layer feedback neural network postulates (symmetry and lack of self-feedback) are fulfilled for W and i computed in part (a). [8] 5. (a) What is the Kohonen layer architure and explain its features. (b) Explain the Kohonen’s learning algorithm. 6. (a) Explain briefly about the counter propagation-training algorithm. (b) Explain the various applications of counter propagation.

[4+4] [4+4] [10] [6]

7. (a) What are the advantages of ART network. Discuss about gain control in ART network. [3+5] (b) Discuss in detail about orienting subsystem in an ART network.

[8]

8. Describe how a neural network may be trained for a pattern recognition task. Illustrate with an example [16] ⋆⋆⋆⋆⋆

1 of 1

Set No. 3

Code No: RR420507

IV B.Tech II Semester Regular Examinations, Apr/May 2007 NEURAL NETWORKS ( Common to Computer Science & Engineering and Electronics & Computer Engineering) Time: 3 hours Max Marks: 80 Answer any FIVE Questions All Questions carry equal marks ⋆⋆⋆⋆⋆ 1. (a) Explain in detail about the single layer artificial neural network with diagram. (b) Explain in detail about the Multi layer artificial neural network with neat diagram. [8+8] 2. Compare the similarities and differences between single layer and multi layer perceptrons and also discuss in what aspects multi layer perceptrons are advantageous over single layer perceptrons. [6+6+4] 3. Describe how a feed forward multi layer neural network may be trained for a function approximation task. Illustrate with an example. [6+10] 4. (a) What are the limitations of Hopfield network? Suggest methods that may overcome these limitations. [4+4] (b) A Hopfield network made up of five neurons, which is required to store the following three fundamental memories: [8] ξ1 = [+1, +1, +1, +1, +1]T ξ2 = [+1, −1, −1, +1, −1]T ξ3 = [−1, +1, −1, +1, +1]T Evaluate the 5-by-5 synaptic weight matrix of the network. 5. Explain the Kohonen’s method of unsupervised learning. Discuss any example as its application. [8+8] 6. Describe the following: (a) Grossberg layer.

[8]

(b) Counter propagation network.

[8]

7. (a) ART network exploits in full one of the inherent advantages of neural computing technique, namely parallel processing ? Explain. [8] (b) Describe the architecture and operation of ART2 network.

[3+5]

8. What are the applications of Kohonen?s networks in image processing and pattern recognition? [16] ⋆⋆⋆⋆⋆

1 of 1

Set No. 4

Code No: RR420507

IV B.Tech II Semester Regular Examinations, Apr/May 2007 NEURAL NETWORKS ( Common to Computer Science & Engineering and Electronics & Computer Engineering) Time: 3 hours Max Marks: 80 Answer any FIVE Questions All Questions carry equal marks ⋆⋆⋆⋆⋆ 1. What are the modes of operation of a Hopfield network? Explain the algorithm for storage of information in a Hopfield network. Similarly explain the recall algorithm. [4+8+4] 2. Briefly discuss about linear separability and the solution for EX-OR problem.Also suggest a network that can solve EX-OR problem. [4+6+6] 3. Implement a backpropagation algorithm to solve EX-OR problem and try the architecture in which there is a hidden layer with three hidden units and the network is fully connected. [8+8] 4. Show how the traveling salesman problem can be solved using the Hopfield model. [16] 5. Discuss how the “Winner-Take-All” in the Kohonen’s layer is implemented and explain the architecture, Also explain the training algorithm. [16] 6. Explain the operation of counter propagation with suitable network model and give the equations for training. [16] 7. Explain the major phases involved in the ART classification process.

[16]

8. Describe how a neural network may be trained for a pattern recognition task. Illustrate with an example [16] ⋆⋆⋆⋆⋆

1 of 1

Related Documents

Neural Networks
November 2019 56
Neural Networks
June 2020 27
Neural Networks
November 2019 42
Neural Networks
November 2019 42
55206-mt----neural Networks
October 2019 23

More Documents from "SRINIVASA RAO GANTA"