Abstract
The difficulties that a neural network faces when trying to learn from a quasi-periodic time series are studied analytically using a teacher-student scenario where the random input is divided into two macroscopic regions with different variances, 1 and 1/γ2 (γ ≫ 1). The generalization error is found to decrease as ∈g ∝ exp(-α/γ2), where α is the number of examples per input dimension. In contradiction to this very slow vanishing generalization error, the next output prediction is found to be almost free of mistakes. This picture is consistent with learning quasi-periodic time series produced by feedforward neural networks, which is dominated by enhanced components of the Fourier spectrum of the input. Simulation results are in good agreement with the analytical results.
| Original language | English |
|---|---|
| Pages (from-to) | 4543-4550 |
| Number of pages | 8 |
| Journal | Journal of Physics A: Mathematical and General |
| Volume | 36 |
| Issue number | 16 |
| DOIs | |
| State | Published - 25 Apr 2003 |
Fingerprint
Dive into the research topics of 'Time series prediction by feedforward neural networks - Is it difficult?'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver