Is it normal that one neuron in the hidden layer is sufficient for a neural network with 56 inputs and one output?
1 view (last 30 days)
Show older comments
I am training a neural network with the following specifications:
Number of inputs: 56
Number of outputs: 1
Number of samples in training data: 2000
Training function: trainbr
Training goal: 1e4
As I see, when I start with one neuron in the hidden layer, the neural network works well even for the data which was not used for training. When I start increasing the number of neurons to 5, 10, etc.. and also number of hidden layers like [5 5], [5 5 5] and so on. The result does not improve. Infact, the neural network tries to overfit if I do not restrict my goal to 1e4. By overfitting, I mean that the training curve and the testing curve starting moving apart from each other around mse of 1e4. I have read a lot online where some sources say, that having just one neuron in the hidden layer might be dangerous when the network is implemented as a generalized model. Does it mean that the network may not function well when exposed to more and more data? Any inputs on this will be appreciated. Thanks.
0 Comments
Accepted Answer
Greg Heath
on 17 Nov 2018
If your training data is sufficient, the most stable designs occur when the training goal is satisfied wit h a minimum number of hidden nodes.
Hope this helps
Thank you for formally accepting my answer
Greg
0 Comments
More Answers (0)
See Also
Categories
Find more on Sequence and Numeric Feature Data Workflows in Help Center and File Exchange
Products
Community Treasure Hunt
Find the treasures in MATLAB Central and discover how the community can help you!
Start Hunting!