diff --git a/Result-Reports/November_1.md b/Result-Reports/November_1.md index 8271bba..6fd55fe 100644 --- a/Result-Reports/November_1.md +++ b/Result-Reports/November_1.md @@ -18,14 +18,14 @@ - [x] Fix debug plots for quantile regression -> predict quantiles and look if true value is below a quantile, if so 1 else 0 and average these over all samples - [ ] Full day debug plots for quantile regression - [x] CPRS Metrics -- [ ] Time as input parameter: +- [x] Time as input parameter: - [x] Probabilistic Baseline -> Quantiles on Training Data -> Breedte bekijken -> Gebruiken voor CPRS en plotjes - [ ] Day-ahead implicit net position - [x] Faster sampling for quantile regression - [ ] Quantile plots for other model (Linear, GRU) (Check if better) - [ ] Check example plots to see if metrics correspond with what seen on plots -- [ ] Time step (96 values) to embedding layer +- [x] Time step (96 values) to embedding layer - [x] Mean of nrv per time step plotten (done for probabilistic baseline) - [x] Convert back to MW on plots diff --git a/requirements.txt b/requirements.txt index cd67fc0..1107a55 100644 --- a/requirements.txt +++ b/requirements.txt @@ -8,4 +8,5 @@ statsmodels lightgbm prettytable clearml -properscoring \ No newline at end of file +properscoring +nbconvert \ No newline at end of file diff --git a/src/data/dataset.py b/src/data/dataset.py index 0b45a09..a09497b 100644 --- a/src/data/dataset.py +++ b/src/data/dataset.py @@ -33,6 +33,14 @@ class NrvDataset(Dataset): dataframe["wind_forecast"].to_numpy(), dtype=torch.float32 ) + self.quarter = torch.tensor( + dataframe["quarter"].to_numpy(), dtype=torch.float32 + ) + + self.day_of_week = torch.tensor( + dataframe["day_of_week"].to_numpy(), dtype=torch.float32 + ) + self.sequence_length = sequence_length self.predict_sequence_length = predict_sequence_length @@ -125,6 +133,20 @@ class NrvDataset(Dataset): ] features.append(wind_forecast.view(-1)) + ### Time Features ### + time_feature = 0 + if self.data_config.QUARTER: + time_feature += self.quarter[actual_idx].item() + + if self.data_config.DAY_OF_WEEK: + d_w = self.day_of_week[actual_idx].item() + if self.data_config.QUARTER: + d_w *= 96 + time_feature += d_w + + if time_feature is not None: + features.append(torch.tensor([time_feature])) + if not features: raise ValueError( "No features are configured to be included in the dataset." @@ -193,6 +215,20 @@ class NrvDataset(Dataset): ] features.append(wind_forecast.view(-1)) + ### Time Features ### + time_feature = 0 + if self.data_config.QUARTER: + time_feature += self.quarter[idx] + + if self.data_config.DAY_OF_WEEK: + d_w = self.day_of_week[idx].item() + if self.data_config.QUARTER: + d_w *= 96 + time_feature += d_w + + if time_feature is not None: + features.append(torch.tensor([time_feature])) + target = self.nrv[ idx + self.sequence_length : idx diff --git a/src/data/preprocessing.py b/src/data/preprocessing.py index 8a07199..d5c113b 100644 --- a/src/data/preprocessing.py +++ b/src/data/preprocessing.py @@ -31,13 +31,14 @@ class DataConfig: ### TIME ### self.YEAR: bool = False - self.DAY: bool = False + self.DAY_OF_WEEK: bool = False self.QUARTER: bool = False class DataProcessor: - def __init__(self, data_config: DataConfig): + def __init__(self, data_config: DataConfig, path:str="./"): self.batch_size = 2048 + self.path = path self.train_range = ( -np.inf, @@ -62,6 +63,13 @@ class DataProcessor: self.wind_forecast, on="datetime", how="left" ) + self.all_features["quarter"] = ( + self.all_features["datetime"].dt.hour * 4 + + self.all_features["datetime"].dt.minute / 15 + ) + + self.all_features["day_of_week"] = self.all_features["datetime"].dt.dayofweek + self.output_size = 96 self.data_config = data_config @@ -103,7 +111,7 @@ class DataProcessor: ) def get_nrv_history(self): - df = pd.read_csv(history_data_path, delimiter=";") + df = pd.read_csv(self.path + history_data_path, delimiter=";") df = df[["datetime", "netregulationvolume"]] df = df.rename(columns={"netregulationvolume": "nrv"}) df["datetime"] = pd.to_datetime(df["datetime"]) @@ -114,7 +122,7 @@ class DataProcessor: return df def get_load_forecast(self): - df = pd.read_csv(forecast_data_path, delimiter=";") + df = pd.read_csv(self.path + forecast_data_path, delimiter=";") df = df.rename( columns={ "Day-ahead 6PM forecast": "load_forecast", @@ -129,7 +137,7 @@ class DataProcessor: return df def get_pv_forecast(self): - df = pd.read_csv(pv_forecast_data_path, delimiter=";") + df = pd.read_csv(self.path + pv_forecast_data_path, delimiter=";") df = df.rename( columns={"dayahead11hforecast": "pv_forecast", "Datetime": "datetime"} @@ -142,7 +150,7 @@ class DataProcessor: return df def get_wind_forecast(self): - df = pd.read_csv(wind_forecast_data_path, delimiter=";") + df = pd.read_csv(self.path + wind_forecast_data_path, delimiter=";") df = df.rename( columns={"dayaheadforecast": "wind_forecast", "datetime": "datetime"} @@ -267,3 +275,14 @@ class DataProcessor: ) input, _, _ = next(iter(data_loader)) return input.shape[-1] + + def get_time_feature_size(self): + time_feature_size = 1 + if self.data_config.QUARTER: + time_feature_size *= 96 + if self.data_config.DAY_OF_WEEK: + time_feature_size *= 7 + + if time_feature_size == 1: + return 0 + return time_feature_size diff --git a/src/models/time_embedding_layer.py b/src/models/time_embedding_layer.py new file mode 100644 index 0000000..a2fea62 --- /dev/null +++ b/src/models/time_embedding_layer.py @@ -0,0 +1,28 @@ +from torch import nn +import torch + +class TimeEmbedding(nn.Module): + def __init__(self, time_features: int, embedding_dim: int): + super().__init__() + self.time_features = time_features + print(time_features) + self.embedding = nn.Embedding(time_features, embedding_dim) + + def forward(self, x): + # Extract the last 'time_features' from the input + time_feature = x[:, -1] + # convert to int + time_feature = time_feature.int() + # Embed these time features + # print max value of time_feature + if time_feature.max() > self.time_features: + # print the row from x that includes the max value in the last column + print(x[time_feature == time_feature.max()]) + print("time feature max value is greater than time features") + + embedded_time = self.embedding(time_feature) + # Concatenate the embedded features with the original input (minus the last 'time feature') + return torch.cat((x[:, :-1], embedded_time), dim=1) + + def output_dim(self, input_dim): + return input_dim + self.embedding.embedding_dim - 1 diff --git a/src/notebooks/training.ipynb b/src/notebooks/training.ipynb index cd8a587..c035fb8 100644 --- a/src/notebooks/training.ipynb +++ b/src/notebooks/training.ipynb @@ -7,21 +7,22 @@ "outputs": [], "source": [ "import sys\n", - "sys.path.append('..')\n", - "from data import DataProcessor, DataConfig\n", - "from trainers.quantile_trainer import AutoRegressiveQuantileTrainer, NonAutoRegressiveQuantileRegression\n", - "from trainers.probabilistic_baseline import ProbabilisticBaselineTrainer\n", - "from trainers.autoregressive_trainer import AutoRegressiveTrainer\n", - "from trainers.trainer import Trainer\n", - "from utils.clearml import ClearMLHelper\n", - "from models import *\n", - "from losses import *\n", + "sys.path.append('../..')\n", + "from src.data import DataProcessor, DataConfig\n", + "from src.trainers.quantile_trainer import AutoRegressiveQuantileTrainer, NonAutoRegressiveQuantileRegression\n", + "from src.trainers.probabilistic_baseline import ProbabilisticBaselineTrainer\n", + "from src.trainers.autoregressive_trainer import AutoRegressiveTrainer\n", + "from src.trainers.trainer import Trainer\n", + "from src.utils.clearml import ClearMLHelper\n", + "from src.models import *\n", + "from src.losses import *\n", "import torch\n", "import numpy as np\n", "from torch.nn import MSELoss, L1Loss\n", "from datetime import datetime\n", - "import pytz\n", "import torch.nn as nn\n", + "from src.models.time_embedding_layer import TimeEmbedding\n", + "\n", "\n", "# auto reload\n", "%load_ext autoreload\n", @@ -33,7 +34,7 @@ }, { "cell_type": "code", - "execution_count": 2, + "execution_count": 9, "metadata": {}, "outputs": [], "source": [ @@ -46,7 +47,10 @@ "data_config.WIND_FORECAST = False\n", "data_config.WIND_HISTORY = False\n", "\n", - "data_processor = DataProcessor(data_config)\n", + "data_config.QUARTER = True\n", + "data_config.DAY_OF_WEEK = False\n", + "\n", + "data_processor = DataProcessor(data_config, path=\"../../\")\n", "data_processor.set_batch_size(1024)\n", "data_processor.set_full_day_skip(False)" ] @@ -60,27 +64,9 @@ }, { "cell_type": "code", - "execution_count": 3, + "execution_count": null, "metadata": {}, - "outputs": [ - { - "name": "stderr", - "output_type": "stream", - "text": [ - "InsecureRequestWarning: Certificate verification is disabled! Adding certificate verification is strongly advised. See: https://urllib3.readthedocs.io/en/latest/advanced-usage.html#ssl-warnings\n" - ] - }, - { - "name": "stdout", - "output_type": "stream", - "text": [ - "ClearML Task: created new task id=8423d146953041eba8d7b4c27d7ed6a5\n", - "ClearML results page: http://192.168.1.182:8080/projects/2e46d4af6f1e4c399cf9f5aa30bc8795/experiments/8423d146953041eba8d7b4c27d7ed6a5/output/log\n", - "2023-11-23 23:07:35,461 - clearml.Task - INFO - Storing jupyter notebook directly as code\n", - "2023-11-23 23:07:39,250 - clearml - WARNING - JSON serialization of artifact 'dictionary' failed, reverting to pickle\n" - ] - } - ], + "outputs": [], "source": [ "data_processor.set_full_day_skip(True)\n", "quantiles = [0.01, 0.05, 0.1, 0.15, 0.4, 0.5, 0.6, 0.85, 0.9, 0.95, 0.99]\n", @@ -100,37 +86,9 @@ }, { "cell_type": "code", - "execution_count": 4, + "execution_count": null, "metadata": {}, - "outputs": [ - { - "name": "stderr", - "output_type": "stream", - "text": [ - "InsecureRequestWarning: Certificate verification is disabled! Adding certificate verification is strongly advised. See: https://urllib3.readthedocs.io/en/latest/advanced-usage.html#ssl-warnings\n" - ] - }, - { - "name": "stdout", - "output_type": "stream", - "text": [ - "ClearML Task: created new task id=11553d672a2744479de07c9ac0a9dbde\n", - "ClearML results page: http://192.168.1.182:8080/projects/2e46d4af6f1e4c399cf9f5aa30bc8795/experiments/11553d672a2744479de07c9ac0a9dbde/output/log\n", - "2023-11-19 18:06:57,539 - clearml.Task - INFO - Storing jupyter notebook directly as code\n", - "2023-11-19 18:06:57,543 - clearml.Repository Detection - WARNING - Can't get url information for git repo in /workspaces/Thesis/src/notebooks\n", - "2023-11-19 18:07:14,402 - clearml.model - WARNING - 500 model found when searching for `file:///workspaces/Thesis/src/notebooks/checkpoint.pt`\n", - "2023-11-19 18:07:14,403 - clearml.model - WARNING - Selected model `Non Autoregressive Quantile Regression` (id=bc0cb0d7fc614e2e8b0edf5b85348646)\n", - "2023-11-19 18:07:14,412 - clearml.frameworks - INFO - Found existing registered model id=bc0cb0d7fc614e2e8b0edf5b85348646 [/workspaces/Thesis/src/notebooks/checkpoint.pt] reusing it.\n", - "2023-11-19 18:07:14,974 - clearml.Task - INFO - Completed model upload to http://192.168.1.182:8081/Thesis/NrvForecast/Non%20Autoregressive%20Model%20%28Non%20Linear%29%20using%20full%20day%20skip%20for%20training%20samples.11553d672a2744479de07c9ac0a9dbde/models/checkpoint.pt\n", - "2023-11-19 18:07:16,827 - clearml.Task - INFO - Completed model upload to http://192.168.1.182:8081/Thesis/NrvForecast/Non%20Autoregressive%20Model%20%28Non%20Linear%29%20using%20full%20day%20skip%20for%20training%20samples.11553d672a2744479de07c9ac0a9dbde/models/checkpoint.pt\n", - "2023-11-19 18:07:18,465 - clearml.Task - INFO - Completed model upload to http://192.168.1.182:8081/Thesis/NrvForecast/Non%20Autoregressive%20Model%20%28Non%20Linear%29%20using%20full%20day%20skip%20for%20training%20samples.11553d672a2744479de07c9ac0a9dbde/models/checkpoint.pt\n", - "2023-11-19 18:07:20,045 - clearml.Task - INFO - Completed model upload to http://192.168.1.182:8081/Thesis/NrvForecast/Non%20Autoregressive%20Model%20%28Non%20Linear%29%20using%20full%20day%20skip%20for%20training%20samples.11553d672a2744479de07c9ac0a9dbde/models/checkpoint.pt\n", - "2023-11-19 18:07:21,843 - clearml.Task - INFO - Completed model upload to http://192.168.1.182:8081/Thesis/NrvForecast/Non%20Autoregressive%20Model%20%28Non%20Linear%29%20using%20full%20day%20skip%20for%20training%20samples.11553d672a2744479de07c9ac0a9dbde/models/checkpoint.pt\n", - "2023-11-19 18:07:28,812 - clearml.Task - INFO - Completed model upload to http://192.168.1.182:8081/Thesis/NrvForecast/Non%20Autoregressive%20Model%20%28Non%20Linear%29%20using%20full%20day%20skip%20for%20training%20samples.11553d672a2744479de07c9ac0a9dbde/models/checkpoint.pt\n", - "Early stopping triggered\n" - ] - } - ], + "outputs": [], "source": [ "#### Hyperparameters ####\n", "inputDim = data_processor.get_input_size()\n", @@ -158,32 +116,9 @@ }, { "cell_type": "code", - "execution_count": 5, + "execution_count": null, "metadata": {}, - "outputs": [ - { - "ename": "ParserError", - "evalue": "Error tokenizing data. C error: Calling read(nbytes) on source failed. Try engine='python'.", - "output_type": "error", - "traceback": [ - "\u001b[0;31m---------------------------------------------------------------------------\u001b[0m", - "\u001b[0;31mParserError\u001b[0m Traceback (most recent call last)", - "\u001b[1;32m/workspaces/Thesis/src/notebooks/training.ipynb Cell 8\u001b[0m line \u001b[0;36m1\n\u001b[1;32m 11\u001b[0m data_config \u001b[39m=\u001b[39m DataConfig()\n\u001b[1;32m 12\u001b[0m data_config\u001b[39m.\u001b[39mLOAD_FORECAST \u001b[39m=\u001b[39m \u001b[39mFalse\u001b[39;00m\n\u001b[0;32m---> 13\u001b[0m data_processor \u001b[39m=\u001b[39m DataProcessor(data_config)\n\u001b[1;32m 14\u001b[0m data_processor\u001b[39m.\u001b[39mset_batch_size(\u001b[39m1024\u001b[39m)\n\u001b[1;32m 17\u001b[0m data_processor\u001b[39m.\u001b[39mset_train_range((datetime(year\u001b[39m=\u001b[39m\u001b[39m2015\u001b[39m, month\u001b[39m=\u001b[39m\u001b[39m1\u001b[39m, day\u001b[39m=\u001b[39m\u001b[39m1\u001b[39m, tzinfo\u001b[39m=\u001b[39mpytz\u001b[39m.\u001b[39mUTC), datetime(year\u001b[39m=\u001b[39m\u001b[39m2022\u001b[39m, month\u001b[39m=\u001b[39m\u001b[39m11\u001b[39m, day\u001b[39m=\u001b[39m\u001b[39m30\u001b[39m, tzinfo\u001b[39m=\u001b[39mpytz\u001b[39m.\u001b[39mUTC)))\n", - "File \u001b[0;32m/workspaces/Thesis/src/notebooks/../data/preprocessing.py:52\u001b[0m, in \u001b[0;36mDataProcessor.__init__\u001b[0;34m(self, data_config)\u001b[0m\n\u001b[1;32m 50\u001b[0m \u001b[39mself\u001b[39m\u001b[39m.\u001b[39mhistory_features \u001b[39m=\u001b[39m \u001b[39mself\u001b[39m\u001b[39m.\u001b[39mget_nrv_history()\n\u001b[1;32m 51\u001b[0m \u001b[39mself\u001b[39m\u001b[39m.\u001b[39mfuture_features \u001b[39m=\u001b[39m \u001b[39mself\u001b[39m\u001b[39m.\u001b[39mget_load_forecast()\n\u001b[0;32m---> 52\u001b[0m \u001b[39mself\u001b[39m\u001b[39m.\u001b[39mpv_forecast \u001b[39m=\u001b[39m \u001b[39mself\u001b[39;49m\u001b[39m.\u001b[39;49mget_pv_forecast()\n\u001b[1;32m 53\u001b[0m \u001b[39mself\u001b[39m\u001b[39m.\u001b[39mwind_forecast \u001b[39m=\u001b[39m \u001b[39mself\u001b[39m\u001b[39m.\u001b[39mget_wind_forecast()\n\u001b[1;32m 55\u001b[0m \u001b[39mself\u001b[39m\u001b[39m.\u001b[39mall_features \u001b[39m=\u001b[39m \u001b[39mself\u001b[39m\u001b[39m.\u001b[39mhistory_features\u001b[39m.\u001b[39mmerge(\n\u001b[1;32m 56\u001b[0m \u001b[39mself\u001b[39m\u001b[39m.\u001b[39mfuture_features, on\u001b[39m=\u001b[39m\u001b[39m\"\u001b[39m\u001b[39mdatetime\u001b[39m\u001b[39m\"\u001b[39m, how\u001b[39m=\u001b[39m\u001b[39m\"\u001b[39m\u001b[39mleft\u001b[39m\u001b[39m\"\u001b[39m\n\u001b[1;32m 57\u001b[0m )\n", - "File \u001b[0;32m/workspaces/Thesis/src/notebooks/../data/preprocessing.py:132\u001b[0m, in \u001b[0;36mDataProcessor.get_pv_forecast\u001b[0;34m(self)\u001b[0m\n\u001b[1;32m 131\u001b[0m \u001b[39mdef\u001b[39;00m \u001b[39mget_pv_forecast\u001b[39m(\u001b[39mself\u001b[39m):\n\u001b[0;32m--> 132\u001b[0m df \u001b[39m=\u001b[39m pd\u001b[39m.\u001b[39;49mread_csv(pv_forecast_data_path, delimiter\u001b[39m=\u001b[39;49m\u001b[39m\"\u001b[39;49m\u001b[39m;\u001b[39;49m\u001b[39m\"\u001b[39;49m)\n\u001b[1;32m 134\u001b[0m df \u001b[39m=\u001b[39m df\u001b[39m.\u001b[39mrename(\n\u001b[1;32m 135\u001b[0m columns\u001b[39m=\u001b[39m{\u001b[39m\"\u001b[39m\u001b[39mdayahead11hforecast\u001b[39m\u001b[39m\"\u001b[39m: \u001b[39m\"\u001b[39m\u001b[39mpv_forecast\u001b[39m\u001b[39m\"\u001b[39m, \u001b[39m\"\u001b[39m\u001b[39mDatetime\u001b[39m\u001b[39m\"\u001b[39m: \u001b[39m\"\u001b[39m\u001b[39mdatetime\u001b[39m\u001b[39m\"\u001b[39m}\n\u001b[1;32m 136\u001b[0m )\n\u001b[1;32m 137\u001b[0m df \u001b[39m=\u001b[39m df[[\u001b[39m\"\u001b[39m\u001b[39mdatetime\u001b[39m\u001b[39m\"\u001b[39m, \u001b[39m\"\u001b[39m\u001b[39mpv_forecast\u001b[39m\u001b[39m\"\u001b[39m]]\n", - "File \u001b[0;32m/opt/conda/lib/python3.10/site-packages/pandas/io/parsers/readers.py:912\u001b[0m, in \u001b[0;36mread_csv\u001b[0;34m(filepath_or_buffer, sep, delimiter, header, names, index_col, usecols, dtype, engine, converters, true_values, false_values, skipinitialspace, skiprows, skipfooter, nrows, na_values, keep_default_na, na_filter, verbose, skip_blank_lines, parse_dates, infer_datetime_format, keep_date_col, date_parser, date_format, dayfirst, cache_dates, iterator, chunksize, compression, thousands, decimal, lineterminator, quotechar, quoting, doublequote, escapechar, comment, encoding, encoding_errors, dialect, on_bad_lines, delim_whitespace, low_memory, memory_map, float_precision, storage_options, dtype_backend)\u001b[0m\n\u001b[1;32m 899\u001b[0m kwds_defaults \u001b[39m=\u001b[39m _refine_defaults_read(\n\u001b[1;32m 900\u001b[0m dialect,\n\u001b[1;32m 901\u001b[0m delimiter,\n\u001b[0;32m (...)\u001b[0m\n\u001b[1;32m 908\u001b[0m dtype_backend\u001b[39m=\u001b[39mdtype_backend,\n\u001b[1;32m 909\u001b[0m )\n\u001b[1;32m 910\u001b[0m kwds\u001b[39m.\u001b[39mupdate(kwds_defaults)\n\u001b[0;32m--> 912\u001b[0m \u001b[39mreturn\u001b[39;00m _read(filepath_or_buffer, kwds)\n", - "File \u001b[0;32m/opt/conda/lib/python3.10/site-packages/pandas/io/parsers/readers.py:583\u001b[0m, in \u001b[0;36m_read\u001b[0;34m(filepath_or_buffer, kwds)\u001b[0m\n\u001b[1;32m 580\u001b[0m \u001b[39mreturn\u001b[39;00m parser\n\u001b[1;32m 582\u001b[0m \u001b[39mwith\u001b[39;00m parser:\n\u001b[0;32m--> 583\u001b[0m \u001b[39mreturn\u001b[39;00m parser\u001b[39m.\u001b[39;49mread(nrows)\n", - "File \u001b[0;32m/opt/conda/lib/python3.10/site-packages/pandas/io/parsers/readers.py:1704\u001b[0m, in \u001b[0;36mTextFileReader.read\u001b[0;34m(self, nrows)\u001b[0m\n\u001b[1;32m 1697\u001b[0m nrows \u001b[39m=\u001b[39m validate_integer(\u001b[39m\"\u001b[39m\u001b[39mnrows\u001b[39m\u001b[39m\"\u001b[39m, nrows)\n\u001b[1;32m 1698\u001b[0m \u001b[39mtry\u001b[39;00m:\n\u001b[1;32m 1699\u001b[0m \u001b[39m# error: \"ParserBase\" has no attribute \"read\"\u001b[39;00m\n\u001b[1;32m 1700\u001b[0m (\n\u001b[1;32m 1701\u001b[0m index,\n\u001b[1;32m 1702\u001b[0m columns,\n\u001b[1;32m 1703\u001b[0m col_dict,\n\u001b[0;32m-> 1704\u001b[0m ) \u001b[39m=\u001b[39m \u001b[39mself\u001b[39;49m\u001b[39m.\u001b[39;49m_engine\u001b[39m.\u001b[39;49mread( \u001b[39m# type: ignore[attr-defined]\u001b[39;49;00m\n\u001b[1;32m 1705\u001b[0m nrows\n\u001b[1;32m 1706\u001b[0m )\n\u001b[1;32m 1707\u001b[0m \u001b[39mexcept\u001b[39;00m \u001b[39mException\u001b[39;00m:\n\u001b[1;32m 1708\u001b[0m \u001b[39mself\u001b[39m\u001b[39m.\u001b[39mclose()\n", - "File \u001b[0;32m/opt/conda/lib/python3.10/site-packages/pandas/io/parsers/c_parser_wrapper.py:234\u001b[0m, in \u001b[0;36mCParserWrapper.read\u001b[0;34m(self, nrows)\u001b[0m\n\u001b[1;32m 232\u001b[0m \u001b[39mtry\u001b[39;00m:\n\u001b[1;32m 233\u001b[0m \u001b[39mif\u001b[39;00m \u001b[39mself\u001b[39m\u001b[39m.\u001b[39mlow_memory:\n\u001b[0;32m--> 234\u001b[0m chunks \u001b[39m=\u001b[39m \u001b[39mself\u001b[39;49m\u001b[39m.\u001b[39;49m_reader\u001b[39m.\u001b[39;49mread_low_memory(nrows)\n\u001b[1;32m 235\u001b[0m \u001b[39m# destructive to chunks\u001b[39;00m\n\u001b[1;32m 236\u001b[0m data \u001b[39m=\u001b[39m _concatenate_chunks(chunks)\n", - "File \u001b[0;32m/opt/conda/lib/python3.10/site-packages/pandas/_libs/parsers.pyx:814\u001b[0m, in \u001b[0;36mpandas._libs.parsers.TextReader.read_low_memory\u001b[0;34m()\u001b[0m\n", - "File \u001b[0;32m/opt/conda/lib/python3.10/site-packages/pandas/_libs/parsers.pyx:875\u001b[0m, in \u001b[0;36mpandas._libs.parsers.TextReader._read_rows\u001b[0;34m()\u001b[0m\n", - "File \u001b[0;32m/opt/conda/lib/python3.10/site-packages/pandas/_libs/parsers.pyx:850\u001b[0m, in \u001b[0;36mpandas._libs.parsers.TextReader._tokenize_rows\u001b[0;34m()\u001b[0m\n", - "File \u001b[0;32m/opt/conda/lib/python3.10/site-packages/pandas/_libs/parsers.pyx:861\u001b[0m, in \u001b[0;36mpandas._libs.parsers.TextReader._check_tokenize_status\u001b[0;34m()\u001b[0m\n", - "File \u001b[0;32m/opt/conda/lib/python3.10/site-packages/pandas/_libs/parsers.pyx:2029\u001b[0m, in \u001b[0;36mpandas._libs.parsers.raise_parser_error\u001b[0;34m()\u001b[0m\n", - "\u001b[0;31mParserError\u001b[0m: Error tokenizing data. C error: Calling read(nbytes) on source failed. Try engine='python'." - ] - } - ], + "outputs": [], "source": [ "#### Hyperparameters ####\n", "inputDim = 96\n", @@ -219,23 +154,30 @@ "cell_type": "markdown", "metadata": {}, "source": [ - "# Quantile Regression" + "# Autoregressive Quantile Regression" ] }, { "cell_type": "code", - "execution_count": 12, + "execution_count": 10, "metadata": {}, "outputs": [ + { + "name": "stdout", + "output_type": "stream", + "text": [ + "96\n" + ] + }, { "name": "stderr", "output_type": "stream", "text": [ - "/workspaces/Thesis/src/notebooks/../trainers/quantile_trainer.py:70: UserWarning:\n", + "/workspaces/Thesis/src/notebooks/../../src/trainers/quantile_trainer.py:68: UserWarning:\n", "\n", "To copy construct from a tensor, it is recommended to use sourceTensor.clone().detach() or sourceTensor.clone().detach().requires_grad_(True), rather than torch.tensor(sourceTensor).\n", "\n", - "/workspaces/Thesis/src/notebooks/../losses/pinball_loss.py:8: UserWarning:\n", + "/workspaces/Thesis/src/notebooks/../../src/losses/pinball_loss.py:8: UserWarning:\n", "\n", "To copy construct from a tensor, it is recommended to use sourceTensor.clone().detach() or sourceTensor.clone().detach().requires_grad_(True), rather than torch.tensor(sourceTensor).\n", "\n" @@ -245,8 +187,8 @@ "name": "stdout", "output_type": "stream", "text": [ - "ClearML Task: created new task id=215dd7634cf2475693ea6081e2ab7559\n", - "ClearML results page: http://192.168.1.182:8080/projects/2e46d4af6f1e4c399cf9f5aa30bc8795/experiments/215dd7634cf2475693ea6081e2ab7559/output/log\n", + "ClearML Task: created new task id=cbf4a5162c604d6ea8f14e71e2d27410\n", + "ClearML results page: http://192.168.1.182:8080/projects/2e46d4af6f1e4c399cf9f5aa30bc8795/experiments/cbf4a5162c604d6ea8f14e71e2d27410/output/log\n", "Early stopping triggered\n" ] } @@ -264,7 +206,9 @@ ").to(\"cuda\")\n", "\n", "# model = LinearRegression(inputDim, len(quantiles))\n", - "model = NonLinearRegression(inputDim, len(quantiles), hiddenSize=1024, numLayers=5)\n", + "time_embedding = TimeEmbedding(data_processor.get_time_feature_size(), 4)\n", + "non_linear_regression_model = NonLinearRegression(time_embedding.output_dim(inputDim), len(quantiles), hiddenSize=1024, numLayers=5)\n", + "model = nn.Sequential(time_embedding, non_linear_regression_model)\n", "optimizer = torch.optim.Adam(model.parameters(), lr=learningRate)\n", "\n", "#### Trainer ####\n", @@ -294,29 +238,35 @@ }, { "cell_type": "code", - "execution_count": 13, + "execution_count": 3, "metadata": {}, "outputs": [ { "name": "stderr", "output_type": "stream", "text": [ - "/workspaces/Thesis/src/notebooks/../trainers/quantile_trainer.py:335: UserWarning:\n", - "\n", - "To copy construct from a tensor, it is recommended to use sourceTensor.clone().detach() or sourceTensor.clone().detach().requires_grad_(True), rather than torch.tensor(sourceTensor).\n", - "\n", - "/workspaces/Thesis/src/notebooks/../losses/pinball_loss.py:23: UserWarning:\n", - "\n", - "To copy construct from a tensor, it is recommended to use sourceTensor.clone().detach() or sourceTensor.clone().detach().requires_grad_(True), rather than torch.tensor(sourceTensor).\n", - "\n" + "/workspaces/Thesis/src/notebooks/../../src/trainers/quantile_trainer.py:335: UserWarning: To copy construct from a tensor, it is recommended to use sourceTensor.clone().detach() or sourceTensor.clone().detach().requires_grad_(True), rather than torch.tensor(sourceTensor).\n", + " quantiles_tensor = torch.tensor(quantiles)\n", + "/workspaces/Thesis/src/notebooks/../../src/losses/pinball_loss.py:22: UserWarning: To copy construct from a tensor, it is recommended to use sourceTensor.clone().detach() or sourceTensor.clone().detach().requires_grad_(True), rather than torch.tensor(sourceTensor).\n", + " self.quantiles_tensor = torch.tensor(quantiles, dtype=torch.float32)\n", + "InsecureRequestWarning: Certificate verification is disabled! Adding certificate verification is strongly advised. See: https://urllib3.readthedocs.io/en/latest/advanced-usage.html#ssl-warnings\n" ] }, { "name": "stdout", "output_type": "stream", "text": [ - "ClearML Task: created new task id=160b4938ae3145db9ef8b55e71452987\n", - "ClearML results page: http://192.168.1.182:8080/projects/2e46d4af6f1e4c399cf9f5aa30bc8795/experiments/160b4938ae3145db9ef8b55e71452987/output/log\n", + "ClearML Task: created new task id=0c748cf6ec0f4c748cc35be78ae4c6c1\n", + "ClearML results page: http://192.168.1.182:8080/projects/2e46d4af6f1e4c399cf9f5aa30bc8795/experiments/0c748cf6ec0f4c748cc35be78ae4c6c1/output/log\n", + "2023-11-26 16:15:07,490 - clearml.Task - INFO - Storing jupyter notebook directly as code\n", + "2023-11-26 16:15:09,255 - clearml.model - WARNING - 500 model found when searching for `file:///workspaces/Thesis/src/notebooks/checkpoint.pt`\n", + "2023-11-26 16:15:09,256 - clearml.model - WARNING - Selected model `Autoregressive Quantile Regression (quarter + day of week)` (id=bc0cb0d7fc614e2e8b0edf5b85348646)\n", + "2023-11-26 16:15:09,265 - clearml.frameworks - INFO - Found existing registered model id=bc0cb0d7fc614e2e8b0edf5b85348646 [/workspaces/Thesis/src/notebooks/checkpoint.pt] reusing it.\n", + "2023-11-26 16:15:09,958 - clearml.Task - INFO - Completed model upload to http://192.168.1.182:8081/Thesis/NrvForecast/Untitled%20Task.0c748cf6ec0f4c748cc35be78ae4c6c1/models/checkpoint.pt\n", + "2023-11-26 16:15:10,998 - clearml.Task - INFO - Completed model upload to http://192.168.1.182:8081/Thesis/NrvForecast/Untitled%20Task.0c748cf6ec0f4c748cc35be78ae4c6c1/models/checkpoint.pt\n", + "2023-11-26 16:15:12,118 - clearml.Task - INFO - Completed model upload to http://192.168.1.182:8081/Thesis/NrvForecast/Untitled%20Task.0c748cf6ec0f4c748cc35be78ae4c6c1/models/checkpoint.pt\n", + "2023-11-26 16:15:13,152 - clearml.Task - INFO - Completed model upload to http://192.168.1.182:8081/Thesis/NrvForecast/Untitled%20Task.0c748cf6ec0f4c748cc35be78ae4c6c1/models/checkpoint.pt\n", + "2023-11-26 16:15:14,540 - clearml.Task - INFO - Completed model upload to http://192.168.1.182:8081/Thesis/NrvForecast/Untitled%20Task.0c748cf6ec0f4c748cc35be78ae4c6c1/models/checkpoint.pt\n", "Early stopping triggered\n" ] }, @@ -324,7 +274,7 @@ "name": "stderr", "output_type": "stream", "text": [ - "/workspaces/Thesis/src/notebooks/../trainers/quantile_trainer.py:366: UserWarning:\n", + "/workspaces/Thesis/src/notebooks/../../src/trainers/quantile_trainer.py:366: UserWarning:\n", "\n", "Creating a tensor from a list of numpy.ndarrays is extremely slow. Please consider converting the list to a single numpy.ndarray with numpy.array() before converting to a tensor. (Triggered internally at /opt/conda/conda-bld/pytorch_1682343967769/work/torch/csrc/utils/tensor_new.cpp:245.)\n", "\n" @@ -368,18 +318,9 @@ }, { "cell_type": "code", - "execution_count": 18, + "execution_count": null, "metadata": {}, - "outputs": [ - { - "name": "stdout", - "output_type": "stream", - "text": [ - "torch.Size([3, 192])\n", - "torch.Size([3, 96])\n" - ] - } - ], + "outputs": [], "source": [ "\n", "\n", diff --git a/src/trainers/quantile_trainer.py b/src/trainers/quantile_trainer.py index fc931ff..ca6c542 100644 --- a/src/trainers/quantile_trainer.py +++ b/src/trainers/quantile_trainer.py @@ -218,7 +218,7 @@ class AutoRegressiveQuantileTrainer(AutoRegressiveTrainer): if other_features is not None: prev_features = torch.cat( - new_features, other_features, dim=1 + (new_features.to(self.device), other_features.to(self.device)), dim=1 ) # (batch_size, 96 + new_features) else: prev_features = new_features @@ -252,36 +252,39 @@ class AutoRegressiveQuantileTrainer(AutoRegressiveTrainer): def plot_quantile_percentages( self, task, data_loader, train: bool = True, iteration: int = None ): + quantiles = self.quantiles.cpu().numpy() total = 0 - quantile_counter = {q: 0 for q in self.quantiles.cpu().numpy()} + quantile_counter = {q: 0 for q in quantiles} + self.model.eval() with torch.no_grad(): for inputs, targets, _ in data_loader: - inputs = inputs.to("cuda") - output = self.model(inputs) + inputs = inputs.to(self.device) + output = self.model(inputs).cpu().numpy() + targets = targets.squeeze(-1).cpu().numpy() # output shape: (batch_size, num_quantiles) # target shape: (batch_size, 1) - for i, q in enumerate(self.quantiles.cpu().numpy()): + for i, q in enumerate(quantiles): quantile_counter[q] += np.sum( - targets.squeeze(-1).cpu().numpy() < output[:, i].cpu().numpy() + targets < output[:, i] ) total += len(targets) # to numpy array of length len(quantiles) percentages = np.array( - [quantile_counter[q] / total for q in self.quantiles.cpu().numpy()] + [quantile_counter[q] / total for q in quantiles] ) bar_width = 0.35 - index = np.arange(len(self.quantiles.cpu().numpy())) + index = np.arange(len(quantiles)) # Plotting the bars fig, ax = plt.subplots(figsize=(15, 10)) bar1 = ax.bar( - index, self.quantiles.cpu().numpy(), bar_width, label="Ideal", color="brown" + index, quantiles, bar_width, label="Ideal", color="brown" ) bar2 = ax.bar( index + bar_width, percentages, bar_width, label="NN model", color="blue" @@ -305,7 +308,7 @@ class AutoRegressiveQuantileTrainer(AutoRegressiveTrainer): ax.set_ylabel("Fraction of data under quantile forecast") ax.set_title(f"Quantile Performance Comparison ({series_name})") ax.set_xticks(index + bar_width / 2) - ax.set_xticklabels(self.quantiles.cpu().numpy()) + ax.set_xticklabels(quantiles) ax.legend() task.get_logger().report_matplotlib_figure( diff --git a/src/trainers/trainer.py b/src/trainers/trainer.py index 014bd37..dc388cf 100644 --- a/src/trainers/trainer.py +++ b/src/trainers/trainer.py @@ -41,7 +41,7 @@ class Trainer: self.patience = patience self.delta = delta - def add_metrics_to_track(self, loss: torch.nn.Module | list[torch.nn.Module]): + def add_metrics_to_track(self, loss): if isinstance(loss, list): self.metrics_to_track.extend(loss) else: @@ -59,7 +59,8 @@ class Trainer: if self.debug: task.add_tags("Debug") - change_description = input("Enter a change description: ") + # change_description = input("Enter a change description: ") + change_description = "" if change_description: task.set_comment(change_description) diff --git a/src/utils/clearml.py b/src/utils/clearml.py index af2c623..825c283 100644 --- a/src/utils/clearml.py +++ b/src/utils/clearml.py @@ -5,6 +5,7 @@ class ClearMLHelper: self.project_name = project_name def get_task(self, task_name: str = "Model Training"): + Task.add_requirements("requirements.txt") Task.ignore_requirements("torch") Task.ignore_requirements("torchvision") Task.ignore_requirements("tensorboard")