Hello world.! I have recently been studying neural networks, so I may ask something obvious, but I figured out that when I replicate my inputs and outputs and then train the network for pattern recognition,it has far more accuracy than with the original data. I thought of that in order to replicate some of the extreme values I have. Can that make my network overfit? Thank you everyone
Neeta Dsouza answered .
2025-11-20
a. Do not overfit: Nw < Ntrneq. Preferrably,
Ntrneq >> Nw which yields design Stability and
robustness w.r.t. noise and measurement error.
For example:
i. Increase the number of training examples
ii. Reduce the number of hidden nodes
b. Use VALIDATION STOPPING to prevent overtraining
c. Use the BAYESIAN REGULARIZATION
training function TRAINBR with MSEREG
as a default.
d. Replace the default performance function
MSE with the regularized
modification MSEREG