Khennour, M. E., Bouchachia, A., Kherfi, M. L. and Bouanane, K., 2023. Randomising the Simple Recurrent Network: a lightweight, energy-efficient RNN model with application to forecasting problems. Neural Computing and Applications, 35, 19707-19718.
Full text available as:
|
PDF
RSRN.pdf - Accepted Version Available under License Creative Commons Attribution Non-commercial. 587kB | |
Copyright to original material in this document is with the original owner(s). Access to this content through BURO is granted on condition that you use it only for research, scholarly or other non-commercial purposes. If you wish to use it for any other purposes, you must contact BU via BURO@bournemouth.ac.uk. Any third party copyright material in this document remains the property of its respective owner(s). BU grants no licence for further use of that third party material. |
DOI: 10.1007/s00521-023-08775-8
Abstract
Multi-variate time-series (MTS) forecasting is the prediction of future for a sequence of data. The process of analysing obtained data can benefit the community financially and securely, for instance observing stock exchange trends and predicting malicious attacks whenabout. MTS forecasting models face many problems including data and model complexity, energy constraints and computational cost. These problems could affect budget allocation, latency and carbon emission. Recurrent neural networks are one of these models, which are known for their computational complexity due to slow learning process which requires more energy to train. Contributing to green AI, in this paper, we propose a competitive and energy-efficient lightweight recurrent neural network based on a hybrid neural architecture that combines Random Neural Network (RaNN) and Simple Recurrent Network (SRN), namely Random Simple Recurrent Network (RSRN). We consider RaNN for its distinctive probabilistic properties and SRN for adding lightweight recurrent ability to the RaNN to process sequential data. The paper shows how RSRN is trained using adapted and optimised versions of back propagation (BP), back propagation through time (BPTT) and truncated BPTT (TBPTT). The latter two algorithms use penalised gradient descent to prevent gradient explosion problems by employing the average of total gradient over time. Evaluated on several datasets, RSRN achieves best performance when using TBPTT. Moreover, we performed a comparative study against well-known recurrent models showing its superiority compared to the state-of-the-art models, while requiring much less computational time and training parameters. In addition, we investigated the multi-layer architecture and its properties.
Item Type: | Article |
---|---|
ISSN: | 0941-0643 |
Uncontrolled Keywords: | Random Neural Network (RaNN); Simple Recurrent Network; Random Simple Recurrent Network (RSRN); Time-series data |
Group: | Faculty of Science & Technology |
ID Code: | 38885 |
Deposited By: | Symplectic RT2 |
Deposited On: | 14 Aug 2023 09:47 |
Last Modified: | 01 Jul 2024 01:08 |
Downloads
Downloads per month over past year
Repository Staff Only - |