Skip to content

Added Project: NN Parameter Sparsity#8

Open
RV2005 wants to merge 2 commits into
Atcold:masterfrom
RV2005:Ray_Verma
Open

Added Project: NN Parameter Sparsity#8
RV2005 wants to merge 2 commits into
Atcold:masterfrom
RV2005:Ray_Verma

Conversation

@RV2005
Copy link
Copy Markdown

@RV2005 RV2005 commented Jan 29, 2026

No description provided.

Copy link
Copy Markdown

@smdzvz smdzvz left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Great project. I think you should remove the (≈ 30 seconds) and (≈ 2 minutes) markers in the section titles of the report.

I notice that you prune the dense network, use the same initialization, and retrain the sparse network. It would be interesting to also test whether the same pruned network can still train successfully when re-initialized with different random weights. We can see whether performance depends on the specific initialization or only on the position of the remaining weights.

Also, although you first perform a search to find a small dense model that performs well, I think you should have a baseline where you randomly remove 80% of the weights and train the network. If this random sparse network performs similarly, it would suggest that the task may be easy enough that many sparse subnetworks work well; if not, it would support your claim that the identified weights (potentially paired with initialization) are meaningfully structured.

@RV2005
Copy link
Copy Markdown
Author

RV2005 commented Jan 30, 2026

Thanks a lot for the review! I've made the changes, and added these in the Sanity Check section.
The new results show that:

  1. Magnitude-based masking outperforms random masking
  2. Original initialization (lottery ticket) outperforms random initialization
  3. This confirms that BOTH aspects are important:
    • Which weights to keep (magnitude-based selection)
    • Starting from the right initialization (lottery ticket)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants