Neural Networks are complex ______________ with many parameters.
a) Linear Functions
b) Nonlinear Functions
c) Discrete Functions
d) Exponential Functions
Neural Networks are complex ______________ with many parameters. a) Linear Functions b) Nonlinear Functions c) Discrete Functions d) Exponential Functions
What are the advantages of neural networks over conventional computers? (i) They have the ability to learn by example (ii) They are more fault tolerant (iii)They are more suited for real time operation due to their high ‘computational’ rates a) (i) and (ii) are true b) (i) and (iii) are true c) Only (i) d) All of the mentioned
Explain in detail Neural Networks?
What are batch, incremental, on-line, off-line, deterministic, stochastic, adaptive, instantaneous, pattern, constructive, and sequential learning?
Which of the following is true for neural networks? (i) The training time depends on the size of the network. (ii) Neural networks can be simulated on a conventional computer. (iii) Artificial neurons are identical in operation to biological ones. a) All of the mentioned b) (ii) is true c) (i) and (ii) are true d) None of the mentioned
What is a Neural Network?
Describe the structure of artificial neural networks?
Which of the following is true? (i) On average, neural networks have higher computational rates than conventional computers. (ii) Neural networks learn by example. (iii) Neural networks mimic the way the human brain works. a) All of the mentioned are true b) (ii) and (iii) are true c) (i), (ii) and (iii) are true d) None of the mentioned
What is Pooling in CNN and how does it work?
Who is concerned with nns?
A 4-input neuron has weights 1, 2, 3 and 4. The transfer function is linear with the constant of proportionality being equal to 2. The inputs are 4, 10, 5 and 20 respectively. The output will be: a) 238 b) 76 c) 119 d) 123
Having multiple perceptrons can actually solve the XOR problem satisfactorily: this is because each perceptron can partition off a linear part of the space itself, and they can then combine their results. a) True – this works always, and these multiple perceptrons learn to classify even complex problems. b) False – perceptrons are mathematically incapable of solving linearly inseparable functions, no matter what you do c) True – perceptrons can do this but are unable to learn to do it – they have to be explicitly hand-coded d) False – just having a single perceptron is enough