From 9be5f6c0e05068f9c29954b218c7ef729db7eb1b Mon Sep 17 00:00:00 2001 From: _Ayaka <33724737+Octopus136@users.noreply.github.com> Date: Thu, 3 Jul 2025 16:50:45 +0800 Subject: [PATCH 1/2] Update Sparse Training Params --- README.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/README.md b/README.md index 9d03404..1047106 100644 --- a/README.md +++ b/README.md @@ -273,7 +273,7 @@ for epoch in range(epochs): out = model(data) loss = F.cross_entropy(out, target) loss.backward() # after loss.backward() - pruner.regularize(model) # <== for sparse training + pruner.regularize(model, loss) # <== for sparse training optimizer.step() # before optimizer.step() ``` From 099d40a49e08663d2268fd53cfe9c82002ef286c Mon Sep 17 00:00:00 2001 From: _Ayaka <33724737+Octopus136@users.noreply.github.com> Date: Thu, 3 Jul 2025 16:51:34 +0800 Subject: [PATCH 2/2] Update Sparse Training Params --- README_CN.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/README_CN.md b/README_CN.md index e2b71f0..aceeb81 100644 --- a/README_CN.md +++ b/README_CN.md @@ -225,7 +225,7 @@ for epoch in range(epochs): out = model(data) loss = F.cross_entropy(out, target) loss.backward() # after loss.backward() - pruner.regularize(model) # <== for sparse training + pruner.regularize(model, loss) # <== for sparse training optimizer.step() # before optimizer.step() ```