transformerForecasting: Transformer Deep Learning Model for Time Series Forecasting
Time series forecasting faces challenges due to the non-stationarity, nonlinearity, and chaotic nature of the data. Traditional deep learning models like Recurrent Neural Network (RNN), Long Short-Term Memory (LSTM), and Gated Recurrent Unit (GRU) process data sequentially but are inefficient for long sequences. To overcome the limitations of these models, we proposed a transformer-based deep learning architecture utilizing an attention mechanism for parallel processing, enhancing prediction accuracy and efficiency. This paper presents user-friendly code for the implementation of the proposed transformer-based deep learning architecture utilizing an attention mechanism for parallel processing. References: Nayak et al. (2024) <doi:10.1007/s40808-023-01944-7> and Nayak et al. (2024) <doi:10.1016/j.simpa.2024.100716>.
Version: |
0.1.0 |
Depends: |
R (≥ 4.0.0) |
Imports: |
ggplot2, keras, tensorflow, magrittr, reticulate (≥ 1.20) |
Suggests: |
dplyr, knitr, lubridate, readr, rmarkdown, utils |
Published: |
2025-03-07 |
DOI: |
10.32614/CRAN.package.transformerForecasting |
Author: |
G H Harish Nayak [aut, cre],
Md Wasi Alam [ths],
B Samuel Naik [ctb],
G Avinash [ctb],
Kabilan S [ctb],
Varshini B S [ctb],
Mrinmoy Ray [ths],
Rajeev Ranjan Kumar [ths] |
Maintainer: |
G H Harish Nayak <harishnayak626 at gmail.com> |
License: |
GPL-3 |
NeedsCompilation: |
no |
CRAN checks: |
transformerForecasting results |
Documentation:
Downloads:
Linking:
Please use the canonical form
https://CRAN.R-project.org/package=transformerForecasting
to link to this page.