Updated thesis
This commit is contained in:
@@ -242,30 +242,31 @@ class DiffusionTrainer:
|
||||
|
||||
_, generated_sampels = self.test(test_loader, -1, task)
|
||||
# self.policy_evaluator.plot_profits_table()
|
||||
if self.policy_evaluator:
|
||||
|
||||
optimal_penalty, profit, charge_cycles = (
|
||||
self.policy_evaluator.optimize_penalty_for_target_charge_cycles(
|
||||
idx_samples=generated_sampels,
|
||||
test_loader=test_loader,
|
||||
initial_penalty=900,
|
||||
target_charge_cycles=283,
|
||||
initial_learning_rate=1,
|
||||
max_iterations=50,
|
||||
tolerance=1,
|
||||
optimal_penalty, profit, charge_cycles = (
|
||||
self.policy_evaluator.optimize_penalty_for_target_charge_cycles(
|
||||
idx_samples=generated_sampels,
|
||||
test_loader=test_loader,
|
||||
initial_penalty=900,
|
||||
target_charge_cycles=283,
|
||||
initial_learning_rate=1,
|
||||
max_iterations=50,
|
||||
tolerance=1,
|
||||
)
|
||||
)
|
||||
)
|
||||
|
||||
print(
|
||||
f"Optimal Penalty: {optimal_penalty}, Profit: {profit}, Charge Cycles: {charge_cycles}"
|
||||
)
|
||||
print(
|
||||
f"Optimal Penalty: {optimal_penalty}, Profit: {profit}, Charge Cycles: {charge_cycles}"
|
||||
)
|
||||
|
||||
task.get_logger().report_single_value(
|
||||
name="Optimal Penalty", value=optimal_penalty
|
||||
)
|
||||
task.get_logger().report_single_value(name="Optimal Profit", value=profit)
|
||||
task.get_logger().report_single_value(
|
||||
name="Optimal Charge Cycles", value=charge_cycles
|
||||
)
|
||||
task.get_logger().report_single_value(
|
||||
name="Optimal Penalty", value=optimal_penalty
|
||||
)
|
||||
task.get_logger().report_single_value(name="Optimal Profit", value=profit)
|
||||
task.get_logger().report_single_value(
|
||||
name="Optimal Charge Cycles", value=charge_cycles
|
||||
)
|
||||
|
||||
if task:
|
||||
task.close()
|
||||
@@ -436,13 +437,13 @@ class DiffusionTrainer:
|
||||
inversed_samples_batched - inversed_expanded_targets
|
||||
)
|
||||
inversed_mae_mean = inversed_mae.mean()
|
||||
all_inversed_mae.extend(inversed_mae_mean.tolist())
|
||||
all_inversed_mae.append(inversed_mae_mean)
|
||||
|
||||
inversed_mse = np.square(
|
||||
inversed_samples_batched - inversed_expanded_targets
|
||||
)
|
||||
inversed_mse_mean = inversed_mse.mean()
|
||||
all_inversed_mse.extend(inversed_mse_mean.tolist())
|
||||
all_inversed_mse.append(inversed_mse_mean)
|
||||
|
||||
# add all values from crps_mean to all_crps
|
||||
all_crps.extend(crps_mean.tolist())
|
||||
@@ -460,12 +461,12 @@ class DiffusionTrainer:
|
||||
|
||||
mean_inversed_mae = np.array(all_inversed_mae).mean()
|
||||
task.get_logger().report_single_value(
|
||||
name="test_MSELoss", value=mean_inversed_mae
|
||||
name="test_L1Loss", value=mean_inversed_mae
|
||||
)
|
||||
|
||||
mean_inversed_mse = np.array(all_inversed_mse).mean()
|
||||
task.get_logger().report_single_value(
|
||||
name="test_L1Loss", value=mean_inversed_mse
|
||||
name="test_MSELoss", value=mean_inversed_mse
|
||||
)
|
||||
|
||||
if self.best_score is None or mean_crps < self.best_score:
|
||||
|
||||
Reference in New Issue
Block a user