Case study of time-series regression task for 24h wind forecasting that I did for an interview. Check the notebook on kaggle for the full results.
Best & worst training and test cases (MLP)
The notebook is comparted into three chapters.
- Data Exploration
- Outlier Detection
- Augmented Dickey-Fulley (ADF) for stationarity
- Normalization
- Model Types & Hyperparamter Optimization
- Ordinary Least Square (OLS) Regression
- Extreme Gradient Boosting (XGB)
- Multi-Layer Perceptron (MLP)
- Summary & Discussion
Random search is used to sample the high dimensional hyperparameter space of the models. In succession Bayesian Optimization might be a good option to further tune the models, despite the costly exploration. A example of the hyperparameter's influence on the XGB model performance can be seen below.
Hyperparameter influence on XGB model performance
Just clone the notebook on kaggle or run it on a local Jupyter Notebook server.