From aa9d9a2d46258c8a1985b5d3b2433f39fed9fbe8 Mon Sep 17 00:00:00 2001 From: smit Date: Fri, 19 Jul 2024 12:31:32 +0200 Subject: [PATCH] remove stray comments --- qadence/ml_tools/train_grad.py | 4 ---- qadence/ml_tools/train_no_grad.py | 4 ---- 2 files changed, 8 deletions(-) diff --git a/qadence/ml_tools/train_grad.py b/qadence/ml_tools/train_grad.py index 5f512386..89ac6383 100644 --- a/qadence/ml_tools/train_grad.py +++ b/qadence/ml_tools/train_grad.py @@ -137,10 +137,6 @@ def loss_fn(model: torch.nn.Module, data: torch.Tensor) -> tuple[torch.Tensor, d else: writer = importlib.import_module("mlflow") - # writer.mlflow.pytorch.autolog( - # log_every_n_step=config.write_every, log_models=False, log_datasets=False - # ) - perform_val = isinstance(config.val_every, int) if perform_val: if not isinstance(dataloader, DictDataLoader): diff --git a/qadence/ml_tools/train_no_grad.py b/qadence/ml_tools/train_no_grad.py index 79684371..43a62dfe 100644 --- a/qadence/ml_tools/train_no_grad.py +++ b/qadence/ml_tools/train_no_grad.py @@ -84,10 +84,6 @@ def _update_parameters( else: writer = importlib.import_module("mlflow") - # writer.mlflow.pytorch.autolog( - # log_every_n_step=config.write_every, log_models=False, log_datasets=False - # ) - # set optimizer configuration and initial parameters optimizer.budget = config.max_iter optimizer.enable_pickling()