Adaptive learning algorithm and its convergence analysis with complex-valued error loss network
In machine learning, the initial task is to construct a model that is capable of predicting the outcomes of new samples with the help of training samples. The loss function plays a key role in this task, as it acts as an important indicator to evaluate the overall model prediction performance. Build...
Uložené v:
| Vydané v: | Neural networks Ročník 190; s. 107677 |
|---|---|
| Hlavní autori: | , , , , |
| Médium: | Journal Article |
| Jazyk: | English |
| Vydavateľské údaje: |
United States
Elsevier Ltd
01.10.2025
|
| Predmet: | |
| ISSN: | 0893-6080, 1879-2782, 1879-2782 |
| On-line prístup: | Získať plný text |
| Tagy: |
Pridať tag
Žiadne tagy, Buďte prvý, kto otaguje tento záznam!
|
| Shrnutí: | In machine learning, the initial task is to construct a model that is capable of predicting the outcomes of new samples with the help of training samples. The loss function plays a key role in this task, as it acts as an important indicator to evaluate the overall model prediction performance. Building upon the work of Chen et al., this study introduces a novel model named as the Complex Error Loss Network (CELN). The CELN is designed to address the scenarios involving complex-valued signals and parameters within the context of supervised learning. Leveraging the contraction mapping theorem, this study investigates the convergence of the corresponding adaptive learning algorithm, underscoring the inherent capability of CELN to consistently approach and potentially reach the global minimum or optimal solution through iterative methods. CELN reduces the error by at least 4.1% compared to benchmark methods while maintaining stability in non-Gaussian noise scenarios. |
|---|---|
| Bibliografia: | ObjectType-Article-1 SourceType-Scholarly Journals-1 ObjectType-Feature-2 content type line 23 |
| ISSN: | 0893-6080 1879-2782 1879-2782 |
| DOI: | 10.1016/j.neunet.2025.107677 |