From 62aca8cbc72b756f61b71cf5f80980a876475392 Mon Sep 17 00:00:00 2001 From: Sebastian Fischer Date: Fri, 31 Jan 2025 05:59:35 +0100 Subject: [PATCH] ... --- NEWS.md | 10 +++++----- R/LearnerTorch.R | 2 +- 2 files changed, 6 insertions(+), 6 deletions(-) diff --git a/NEWS.md b/NEWS.md index 1dc84b36..f2ae78f6 100644 --- a/NEWS.md +++ b/NEWS.md @@ -12,12 +12,12 @@ * feat: Added `n_layers` parameter to MLP * BREAKING_CHANGE: Early stopping now not uses `epochs - patience` for the internally tuned values instead of the trained number of `epochs` as it was before. - Also, the improvement is calcualted as the difference between the current and the best score, + Also, the improvement is calculated as the difference between the current and the best score, not the current and the previous score. -* feat: Added multimodal melanoma example task -* feat: Added a callback to iteratively unfreeze parameters for finetuning -* fix: torch learners can now be used with `AutoTuner` -* feat: Added different learning rate schedulers as callbacks +* feat: Added multimodal melanoma and cifar{10, 100} example tasks. +* feat: Added a callback to iteratively unfreeze parameters for finetuning. +* fix: torch learners can now be used with `AutoTuner`. +* feat: Added different learning rate schedulers as callbacks. # mlr3torch 0.1.2 diff --git a/R/LearnerTorch.R b/R/LearnerTorch.R index 921d3fb4..8d166451 100644 --- a/R/LearnerTorch.R +++ b/R/LearnerTorch.R @@ -567,7 +567,7 @@ marshal_model.learner_torch_model = function(model, inplace = FALSE, ...) { #' @export unmarshal_model.learner_torch_model_marshaled = function(model, inplace = FALSE, device = "cpu", ...) { model = model$marshaled - model$network = if (model$jitted) { + model$network = if (isTRUE(model$jitted)) { deser = jit_unserialize(model$network) deser$to(device = device) deser