Lottery ticket hypothesis
In machine learning, the lottery ticket hypothesis is that artificial neural networks with random weights can contain a subnetwork which (entirely by chance) can be tuned to a similar performance as tuning the whole network.[1]
The original statement of the hypothesis is:
A randomly-initialized, dense neural network contains a subnetwork that is initialized such that—when trained in isolation—it can match the test accuracy of the original network after training for at most the same number of iterations.
The original algorithm is:
- Randomily initialize dense network: weights .
 - Train for iterations → weights .
 - Magnitude pruning: build a bitmask by setting to 0 the lowest-magnitude weights in each layer (one-shot), or repeat train → prune across rounds to reach higher sparsity (iterative).
 - Reset surviving weights to initialization: use as the starting point. Here is elementwise multiplication.
 - Train the masked network with the same optimizer, schedule, and data.
 - Declare "winning ticket" if it achieves comparable test accuracy in iterations.
 
The term derived from considering the probability of a tunable subnetwork as the equivalent to a winning lottery ticket; the chance of any given ticket winning is tiny, but if you buy enough of them you are certain to win, and the number of possible subnetworks increases exponentially as the power set of the set of connections, making the number of possible subnetworks astronomical for any reasonably large network.
Such networks are a priori difficult to find, since before training, one does not know which of the exponentially many subnetwork would be "lottery tickets". However, after training, these lottery tickets can be discovered by the pruning algorithm.
It was found that during the original training of the initial dense network, the weights that would end up in the winning ticket change a lot, but the other weights change little.
It was found that if instead of , they re-sampled a different random initialization , and used instead, the trained would perform much worse. This indicates that what makes a ticket winning is both its connection graph and the precise initial weights of the connections.
Subsequent work
[edit]Malach et. al. proved a stronger version of the hypothesis, namely that a sufficiently overparameterized untuned network will typically contain a subnetwork that is already an approximation to the given goal, even before tuning.[2] A similar result has been proven for the special case of convolutional neural networks.[3]
See also
[edit]References
[edit]- ^ Frankle, Jonathan; Carbin, Michael (2019-03-04). "The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networks". arXiv:1803.03635 [cs.LG]., published as a conference paper at ICLR 2019.
 - ^ Malach, Eran; Yehudai, Gilad; Shalev-Shwartz, Shai; Shamir, Ohad (2020-02-03). "Proving the Lottery Ticket Hypothesis: Pruning is All You Need". arXiv:2002.00585 [cs.LG]. published in Proceedings of the 37th International Conference on Machine Learning, Online, PMLR 119, 2020
 - ^ da Cunha, Arthur; Natale, Emanuele; Viennot, Laurent (2022). "Proving the Strong Lottery Ticket Hypothesis for Convolutional Neural Networks". ICLR 2022 - 10th International Conference on Learning Representations. Virtual, France.