arXiv:2012.04643v2 [cs.CV] 19 Apr 2021

17
The Lottery Ticket Hypothesis for Object Recognition Sharath Girish * [email protected] Shishira R. Maiya * [email protected] Kamal Gupta [email protected] Hao Chen [email protected] Larry Davis [email protected] Abhinav Shrivastava [email protected] University of Maryland, College Park Abstract Recognition tasks, such as object recognition and key- point estimation, have seen widespread adoption in recent years. Most state-of-the-art methods for these tasks use deep networks that are computationally expensive and have huge memory footprints. This makes it exceedingly difficult to deploy these systems on low power embedded devices. Hence, the importance of decreasing the storage require- ments and the amount of computation in such models is paramount. The recently proposed Lottery Ticket Hypothe- sis (LTH) states that deep neural networks trained on large datasets contain smaller subnetworks that achieve on par performance as the dense networks. In this work, we per- form the first empirical study investigating LTH for model pruning in the context of object detection, instance segmen- tation, and keypoint estimation. Our studies reveal that lottery tickets obtained from Imagenet pretraining do not transfer well to the downstream tasks. We provide guidance on how to find lottery tickets with up to 80% overall spar- sity on different sub-tasks without incurring any drop in the performance. Finally, we analyse the behavior of trained tickets with respect to various task attributes such as ob- ject size, frequency, and difficulty of detection. Our code is made public at: https://github.com/Sharath-girish/LTH- ObjectRecognition. 1. Introduction Recognition tasks, such as object detection, instance segmentation, and keypoint estimation, have emerged as canonical tasks in visual recognition because of their intu- itive appeal and pertinence in a wide variety of real-world problems. The modus operandi followed in nearly all state- of-the-art visual recognition methods is the following: (i) Pre-train a large neural network on a very large and di- * First two authors contributed equally To appear at CVPR 2021 ResNet-18 on COCO 0 20 40 60 80 26 28 30 32 mAP Object Detection 0 20 40 60 80 Network Sparsity (% pruned) 24 26 28 30 Instance Segmentation 0 20 40 60 80 55 57 59 61 Keypoint Estimation ResNet-50 on COCO 0 20 40 60 80 31 33 35 37 39 mAP Object Detection 0 20 40 60 80 Network Sparsity (% pruned) 29 31 33 35 Instance Segmentation 0 20 40 60 80 58 60 62 64 Keypoint Estimation Unpruned Transfer Ticket Direct Pruning via LTH Figure 1: Performance of lottery tickets discovered using direct pruning for various object recognition tasks. Here we have used a Mask R-CNN model with ResNet-18 backbone (top) and ResNet-50 backbone (bottom) to train models for object detection, segmentation and human keypoint es- timation on the COCO dataset. We show the performance of the baseline dense network, the sparse subnetwork obtained by transferring ImageNet pre-trained “universal” lottery tickets, as well as the subnetwork obtained by task-specific pruning. Task-specific pruning outperforms the universal tickets by a wide margin. For each of the tasks, we can obtain the same performance as the original dense networks with only 20% of the weights. verse image classification dataset, (ii) Append a small task- specific network to the pre-trained model and fine-tune the weights jointly on a much smaller dataset for the task. The introduction of ResNets by He et al.[23] made the training of very deep networks possible, helping in scaling up model capacity, both in terms of depth and width, and became a well-established instrument for improving the performance of deep learning models even with smaller datasets [26]. As a result, the past few years have seen increasingly large neu- ral network architectures [36, 57, 49, 24], with sizes often exceeding the memory limits of a single hardware accel- erator. In recent years, efforts towards reducing the mem- ory and computation footprint of deep networks have fol- lowed three seemingly parallel tracks with common objec- 1 arXiv:2012.04643v2 [cs.CV] 19 Apr 2021

Transcript of arXiv:2012.04643v2 [cs.CV] 19 Apr 2021

The Lottery Ticket Hypothesis for Object Recognition

Sharath Girish*

[email protected]

Shishira R. Maiya*

[email protected]

Kamal [email protected]

Hao [email protected]

Larry [email protected]

Abhinav [email protected]

University of Maryland, College Park

Abstract

Recognition tasks, such as object recognition and key-point estimation, have seen widespread adoption in recentyears. Most state-of-the-art methods for these tasks usedeep networks that are computationally expensive and havehuge memory footprints. This makes it exceedingly difficultto deploy these systems on low power embedded devices.Hence, the importance of decreasing the storage require-ments and the amount of computation in such models isparamount. The recently proposed Lottery Ticket Hypothe-sis (LTH) states that deep neural networks trained on largedatasets contain smaller subnetworks that achieve on parperformance as the dense networks. In this work, we per-form the first empirical study investigating LTH for modelpruning in the context of object detection, instance segmen-tation, and keypoint estimation. Our studies reveal thatlottery tickets obtained from Imagenet pretraining do nottransfer well to the downstream tasks. We provide guidanceon how to find lottery tickets with up to 80% overall spar-sity on different sub-tasks without incurring any drop in theperformance. Finally, we analyse the behavior of trainedtickets with respect to various task attributes such as ob-ject size, frequency, and difficulty of detection. Our codeis made public at: https://github.com/Sharath-girish/LTH-ObjectRecognition.

1. Introduction

Recognition tasks, such as object detection, instancesegmentation, and keypoint estimation, have emerged ascanonical tasks in visual recognition because of their intu-itive appeal and pertinence in a wide variety of real-worldproblems. The modus operandi followed in nearly all state-of-the-art visual recognition methods is the following: (i)Pre-train a large neural network on a very large and di-

*First two authors contributed equallyTo appear at CVPR 2021

ResNet-18 on COCO

0 20 40 60 80

26

28

30

32

mA

P→

Object Detection

0 20 40 60 80Network Sparsity (% pruned)

24

26

28

30Instance Segmentation

0 20 40 60 80

55

57

59

61Keypoint Estimation

Unpruned Transfer Ticket Direct Pruning via LTHResNet-50 on COCO

0 20 40 60 80

31

33

35

37

39

mA

P→

Object Detection

0 20 40 60 80Network Sparsity (% pruned)

29

31

33

35

Instance Segmentation

0 20 40 60 80

58

60

62

64

Keypoint Estimation

Unpruned Transfer Ticket Direct Pruning via LTH

Figure 1: Performance of lottery tickets discovered using direct pruningfor various object recognition tasks. Here we have used a Mask R-CNNmodel with ResNet-18 backbone (top) and ResNet-50 backbone (bottom)to train models for object detection, segmentation and human keypoint es-timation on the COCO dataset. We show the performance of the baselinedense network, the sparse subnetwork obtained by transferring ImageNetpre-trained “universal” lottery tickets, as well as the subnetwork obtainedby task-specific pruning. Task-specific pruning outperforms the universaltickets by a wide margin. For each of the tasks, we can obtain the sameperformance as the original dense networks with only 20% of the weights.

verse image classification dataset, (ii) Append a small task-specific network to the pre-trained model and fine-tune theweights jointly on a much smaller dataset for the task. Theintroduction of ResNets by He et al. [23] made the trainingof very deep networks possible, helping in scaling up modelcapacity, both in terms of depth and width, and became awell-established instrument for improving the performanceof deep learning models even with smaller datasets [26]. Asa result, the past few years have seen increasingly large neu-ral network architectures [36, 57, 49, 24], with sizes oftenexceeding the memory limits of a single hardware accel-erator. In recent years, efforts towards reducing the mem-ory and computation footprint of deep networks have fol-lowed three seemingly parallel tracks with common objec-

1

arX

iv:2

012.

0464

3v2

[cs

.CV

] 1

9 A

pr 2

021

tives: weight quantization, sparsity via regularization, andnetwork pruning; Weight Quantization [25, 17, 46, 7, 31]methods either replace weights of a trained neural networkwith lower precision or arithmetic operations with bit-wiseoperations to reduce the memory up to an order of magni-tude. Regularization approaches, such as dropout [47, 2] orLASSO [50], attempt to discourage an over-parameterizednetwork from relying on a large number of features and en-courage learning a sparse and robust predictor. Both quanti-zation and regularization approaches are effective in reduc-ing the number of weights in a network or the memory foot-print, but usually at the cost of increased error rates [21, 31].In comparison, pruning approaches [29, 20] disentangle thelearning task from pruning by alternating between weightoptimization and weight deletion. The recently proposedLottery Ticket Hypothesis ·(LTH) [13] falls in this category.

According to LTH, an over-parameterized network con-tains sparse sub-networks which not only match but some-times even exceed the performance of the original network,all by virtue of a “lucky” random initialization before train-ing. The original paper was followed up with tips and tricksto train large-scale models under the same paradigm [16].Since then, there has been a large, growing body of litera-ture exploring its nuances. Although some of these recentworks have tried to answer the question – how well do thetickets transfer across domains [39, 38], when it comes tovision tasks – the buck stops at image classification.

In this work, we aim to extend and explore the analysisof lottery tickets to fundamental visual recognition tasks ofobject detection, instance segmentation, and keypoint de-tection. Popular methods for such recognition tasks use atwo-stage detection pipeline, with a supervised pre-trainedconvolutional neural network (ConvNet) backbone, a re-gion proposal network (RPN), and one or more region-wisetask-specific neural network branches. Loosely speaking,a ConvNet backbone is the most computationally intensivepart of the architecture, and pre-training is the most time-consuming part. Therefore, as part of this study, we ex-plore the following questions: (a) Are there universal sub-networks within the ConvNet backbone that can be trans-ferred to the downstream object recognition tasks? (b) Canwe train sparser and more accurate sub-networks for eachof the downstream tasks? And, (c) How does the behav-ior or properties of these sub-networks change with respectto the corresponding dense network? We investigate thesequestions under the dominant settings used in object recog-nition frameworks. Specifically, we use ImageNet [8] pre-trained ResNet-18 and ResNet-50 [23] backbones, Faster R-CNN [42] and Mask R-CNN [22] modules for object recog-nition on Pascal VOC [12] and COCO [32] datasets. Ourcontributions are as follows:

• We show that tickets obtained from ImageNet trainingdon’t transfer to object recognition in case of COCO,

i.e., there are no universal tickets in pre-trained Ima-geNet models that can be used for downstream recog-nition tasks without a drop in performance. This is incontrast with previous works related to ticket transferin vision models [38, 39]. In case of smaller datasetssuch as Pascal VOC, we are able to find winning ticketsfrom ImageNet pre-training with upto 40% sparsity.

• With direct pruning, we can find “task-specific” tick-ets with up to 80% sparsity for each of the datasets andbackbones. We also investigate the efficacy of methodsintroduced by [13, 39, 14, 43] such as iterative mag-nitude pruning, late resetting, early bird training, andlayerwise pruning in the context of object recognition.

• Finally we analyse the behavior of tickets obtained forobject recognition tasks, with respect to various taskattributes such as object size, frequency, and difficultyof detection, to make some expected (and some sur-prising) observations.

2. Related WorkModel Compression: Ever since deep neural networksstarted gaining traction in real-world applications, therehave been serious attempts made to reduce their parame-ters, intending to attain lower memory footprints [17, 54,25, 46, 7, 31], higher inference speeds [51, 9, 19] and po-tentially better generalization [1]. Amongst the various pro-posed techniques, model pruning approaches are predomi-nant mainly due to their simplicity and effectiveness. Oneline of methods follow an unstructured process where in-significant weights are set to zero and are frozen for therest of the training. The significance of weights are quanti-fied either by magnitude [20] or gradients during trainingtime [30]. In structured pruning methods, relationships be-tween pruned weights are taken into consideration, leadingto pruning them in groups. Methods like [53] utilize GroupLasso regularization to prune redundant filter weights to en-able structural sparsity, [34] uses explicit L0 regularizationto make weights within structures have exact zero values,and network slimming [33] learns an efficient network bymodelling the scaling factor of batch normalization layer.

The Lottery Ticket Hypothesis: The introduction of Lot-tery Ticket Hypothesis by [16] opened a pandora’s box ofimmense possibilities in the field of pruning and sparsemodels. The original paper was followed by [15] where theauthors introduce the concept of ”late resetting” which en-abled the application of the hypothesis to larger and deepermodels. [58] followed up by proposing an extensive, in-depth analysis where they show that the resetting of theweights need not be to the exact initialization, but just needto the initial signs. [14] probes the aspect of resetting fur-ther to show that the reason why LTH works is because

2

of its ability to make the subnetwork stable to SGD noise.As far as theoretical guarantees are considered, [37] offersstrong theoretical proofs for the experimental evidence ofLTH. [18] probed an orthogonal question about the num-ber of possible tickets from a network. They showed thata single initialization had multiple winning tickets with lowoverlap and empirically conclude that there exists an entire“distribution” of winning lottery tickets.

Complementary to LTH[16], [30] and [52] offer algo-rithms that can pick the winning ticket without the need fortraining. But they do not match the performance of the orig-inal procedure. The problem of longer training using LTHwas effectively tackled by [55] which introduced the con-cept of “early bird tickets” where the authors show that thewinning tickets and their masks are obtained in the first fewepochs of training, foregoing the need to train the originalinitialization till convergence. The intriguing properties ofLTH led to a glut of works which investigated its eclec-tic aspects. [39] scrutinize the generalization properties ofwinning tickets and offer empirical evidence that winningtickets can be transferred across datasets and optimizers, inthe realm of image classification. The authors also discussthe learnt “inductive biases” of the tickets which may leadto worse performance of transferred tickets when comparedwith a ticket obtained from the same dataset. [38] thenproposed a variation of the theory titled “transfer ticket hy-pothesis” where they investigate the effectiveness of trans-ferring a mask generated from source dataset to a targetdataset. [10] shows that the winning tickets do not performsimple overfitting to any domain and carry forward certaininherent biases which can prove useful in other domains too.There have been many applications of LTH in the fields ofNLP [6] [40] [10][4] and Reinforcement Learning [56] [55]as well.

The work of [45] briefly analyzes LTH on single stagedetectors such as YOLOv3 [41] and achieves 90% win-ning tickets, while maintaining the mAP on the Pascal VOC2007 dataset. However, as they evaluate on light-weight andfast detectors, their mAP (∼ 56) is much lower comparedto networks like Faster R-CNN [42] which reach mAP of∼ 69 with just a ResNet-18 backbone. Their work is alsolimited to object detection and does not provide a detailedanalysis of LTH for the task. The idea for transferring sub-networks obtained from ImageNet to object detection taskswas concurrently discussed by [5]. For small datasets suchas Pascal VOC, [5] observes that ImageNet tickets transferfor detection and segmentation tasks. However, we extendthe analysis to the larger COCO dataset and show that thisobservation doesn’t hold. We further build upon these re-sults, to test out the generalization and transfer capabilitiesof winning lottery tickets across different object recognitiondatasets and tasks in computer vision.

Algorithm 1 Iterative Pruning for LTH

1: Randomly initialize network f with initial weights w0,mask m0 = 1, prune target percentage p, and T prun-ing rounds to achieve it.

2: while i < T do3: Train network for N iterations f(x;mi � w0) →

f(x;mi � wi)

4: Prune bottom p1k% of mi � wi and update mi.

5: Reset to initial weights w0

6: i← i+ 1 . next round

3. Background: Lottery Ticket Hypothesis

LTH states that dense randomly-initialized neural net-works contain sparse sub-networks which can be trainedin isolation and can match the test accuracy of the origi-nal network. These sub-networks are called winning tick-ets and can be identified using an algorithm called Itera-tive Magnitude Pruning (IMP). Suppose the number of it-erations for pruning is T and we wish to prune p% of thenetwork weights. The weights/parameters are representedby w ∈ Rn and the pruning mask by m ∈ {0, 1}n where nis the total number of weights in the network. The completealgorithm is presented in 1.

This pruning method can be one-shot when it proceedsfor only a single iteration or it can proceed for multiple it-erations, k, pruning p

1k% each round. The authors also use

other techniques such as learning rate warmup and showthat finding winning tickets is sensitive to the learning rate.While this method obtains winning tickets for smallerdatasets, like MNIST [28], CIFAR10 [27], they fail to gen-eralize to deeper networks, such as ResNets, and larger vi-sion benchmarks, such as ImageNet [8]. [15] shows thatIMP fails when resetting to the original initialization. Theyclaim that resetting instead to the network weights after afew iterations of training provides greater stability and en-ables them to find winning tickets in these larger networks.They show that rewinding/late resetting to 3−7% into train-ing yields subnetworks which are 70% smaller in the caseof ResNet-50, without any drop in accuracy.

4. LTH for Object Recognition

In this section, we extend the Lottery Ticket Hypothesisto several object recognition tasks, such as Object Detec-tion, Instance Segmentation, and Keypoint Detection. In§4.1, we describe the datasets, models, and metrics we usein our paper. §4.2 examines the transfer of the lottery ticketsobtained from ImageNet training to the downstream recog-nition tasks. §4.3 investigates direct pruning on the down-stream tasks. §4.4 analyzes the various properties of win-ning tickets obtained using direct pruning.

3

Table 1: Performance on the COCO dataset for ImageNet transferred tickets for ResNet-18 backbone at varying sparsity. The results for VOC are averagedover 5 runs with the standard deviation in parantheses. We obtain winning tickets at higher sparsity for smaller datasets like VOC compared to COCO.

Prune %COCO Detection COCO segmentation COCO Keypoint VOC Detection

Networksparsity

mAP AP50Networksparsity

mAP AP50Networksparsity

mAP AP50Networksparsity

mAP

90% 31.61% 25.59 43.69 31.61% 24.03 40.89 21.47% 55.30 79.30 79.49% 63.91(±0.41)

80% 28.10% 27.70 46.50 28.10% 25.90 43.70 19.09% 56.70 81.10 70.66% 65.82(±0.23)

50% 17.57% 28.52 47.54 17.57% 26.60 44.66 11.94% 56.96 80.83 44.16% 68.06(±0.11)

0% 0% 29.91 49.05 0% 27.64 46.00 0% 58.59 82.04 0% 68.53(±0.29)

Table 2: Performance on the COCO dataset for ImageNet transferred tickets for ResNet-50 backbone at various levels of pruning. The results for VOC areaveraged over 5 runs with the standard deviation in parantheses. We obtain higher levels of sparsity compared to ResNet-18 transferred tickets which can beexpected as it has fewer redundant parameters. Additionally, tickets for VOC have much higher sparsity with no drop in mAP compared to unpruned model.

Prune %COCO Detection COCO segmentation COCO Keypoint VOC Detection

Networksparsity

mAP AP50Networksparsity

mAP AP50Networksparsity

mAP AP50Networksparsity

mAP

90% 41.99% 30.66 50.75 41.99% 28.68 47.76 31.49% 57.78 82.25 65.37% 71.20(±0.21)

80% 37.33% 31.01 50.98 37.33% 29.04 47.90 28% 58.55 83.06 58.11% 71.08(±0.20)

0% 0% 38.5 59.29 0% 35.13 56.39 0% 64.59 86.48 0% 71.21(±0.32)

4.1. Experimental setup

We evaluate LTH primarily on the 2 datasets - PascalVOC 2007 and COCO. We deal with the 3 tasks of objectdetection, instance segmentation, and keypoint detection forCOCO and only object detection for VOC. We use Mask-RCNN for object detection and segmentation for COCO,Keypoint-RCNN for keypoint detection, and Faster-RCNNfor object detection on VOC. We compare the results forResNet-18 and ResNet-50 backbones. Note that while weuse the term mean Average Precision (mAP) as a perfor-mance metric for all the tasks and datasets, the actual calcu-lation of mAP is done using code provided in the respectivedatasets (and cannot be compared across the tasks).

4.2. Transfer of ImageNet Tickets

Many object recognition tasks utilize pre-trained net-works whose backbones are trained on the ImageNetdataset. This is because ImageNet features and weightshave shown the ability [22] to generalize well to severaldownstream vision tasks. A plethora of works exists whichperform LTH for the ImageNet classification task and ob-tain winning tickets. Therefore, tickets for standard con-volutional backbones, such as ResNets, are readily avail-able. This raises the pertinent question of whether prunedImageNet trained models transfer directly to object recogni-tion tasks. In order to answer this question, we transfer thepruned model to the backbone of the RCNN-based networkand fine-tune the full network while ensuring the prunedweights in the backbone remain as zeros.

We perform experiments for the two architectures:

ResNet-18 and ResNet-50, where we obtain 10%, 20%, and50% tickets on ImageNet by following the approach of [15],and then transfer the model to the backbones of the three R-CNN based networks. All models were trained on COCO,encompassing the three recognition tasks and the resultsare summarized in Tables 1, 2. Additionally, we also per-form similar experiments on the smaller Pascal VOC 2007dataset for object detection to verify whether ImageNettickets transfer without a significant drop in mAP. The re-sults are shown in the last column of Tables 1, 2. Note thatpruning percentage of the ImageNet ticket is not equal tothe actual network sparsity of the various networks as onlythe backbone of the networks are transferred and they makeup a fraction of the total weights.

We see that ImageNet tickets transferred to COCO showa noticeable drop in mAP even with low levels of sparsity.We also note that another drawback of training transferredtickets on COCO is that they require careful tuning of learn-ing rate and batch size for different tasks. On the other hand,for smaller datasets such as Pascal VOC, winning tickets areeasily obtained at higher levels of sparsity which is ∼ 45%for ResNet-18 and ∼ 65% for ResNet-50. The larger net-works can be pruned to a greater extent for both datasets.

ImageNet transferred tickets offer very little sparsity asthe backbone of the networks usually do not make up mostof the weights. For example, the ResNet-18 based Mask-RCNN for COCO Detection and Segmentation has only44% of the parameters in the backbone, and hence, the over-all network sparsity reaches 31% when pruning 90% of thebackbone weights, as shown in Table 1. The rest of theweights are usually from the fully-connected layers of the

4

network. It is therefore imperative to prune layers in addi-tion to the backbone to increase network sparsity withoutdecrease in mAP. As a consequence, we look into directlypruning the full network using LTH.

4.3. Direct Pruning for Downstream Task

In this section, we analyze the effect of various hyper-parameters and pruning strategies for detection networksin order to obtain winning tickets. We primarily use theResNet-18 backbone for Faster-RCNN trained on VOC forall our experiments in this section, unless mentioned other-wise. Even though the ResNet-18 backbone is smaller thanother backbone networks such as ResNet-50, we find thatsimilar conclusions hold for the larger networks as well.

The Faster RCNN network consists of parameters whichwe group into 4 main modules: Base Convolutions, Classi-fication Network Convolutions (Top), Region Proposal Net-work (RPN), Classification network box and classificationfully connected heads (Box and Cls Head). We provide adetailed analysis of pruning these 4 groups and their effecton winning tickets. Additionally, we also analyze differentpruning strategies and the role played by hyperparameters.Varying Pruning Percentage: We evaluate the networkperformance at varying levels of sparsity. We prune differ-ent percentages of parameters in the Base and Top moduleswhich include 88% of the total network parameters. The re-sults are plotted in Fig.2a. We achieve performance withinone standard deviation of the baseline, with 70% sparsity.Our models outperform the baseline mean, thereby provingthat we can indeed obtain high performance winning ticketsat much higher levels of sparsity for detection. Additionally,we see that at any given sparsity, direct pruning yields muchbetter results compared to ImageNet transferred tickets. Wealso show that these observations pan other vision tasks byobtaining winning tickets for Mask-RCNN and Keypoint-RCNN with both ResNet-18 and ResNet-50 backbones onthe COCO dataset. We additionally prune FC layers in theseset of experiments in order to achieve desired sparsity levelsas they take up ∼ 50% of the total weights. The results forResNet-18 are shown in Fig. 1. We obtain winning ticketswith 80% sparsity on all the three tasks while outperform-ing the unpruned network for lower levels of sparsity. Addi-tionally, we consistently outperform the different ImageNettransferred tickets (50%, 80%, 90%) by a large margin sup-porting our claim that direct training of tickets on down-stream tasks yield better results than ImageNet tickets.Effect of Early/Late Resetting: [15] states that resettingthe network to a few iterations through training instead ofthe initialization stabilizes the winning ticket training. Weevaluate whether this holds true for detection tasks as well.We show the performance of winning tickets as a functionof resetting at various stages of training in Fig. 2b and ob-serve that resetting during the earlier or even mid stages of

Table 3: Performance on Pascal VOC by pruning different modules of aResNet-18 Faster-RCNN network. The results are averaged over 5 runswith the standard deviation in parantheses. X represents the module beingpruned, while Param % represents the percentage of parameters occupiedby the modules being pruned.

Base Top RPNBox,

Cls HeadParam%

NetworkSparsity mAP

- - - - 0 0% 69.74 (±0.16)- - - X 0.65 0.52% 70.30 (±0.14)- - X - 9.71 7.77% 70.02 (±0.19)- - X X 10.36 8.29% 70.08 (±0.10)X - - - 21.93 17.55% 69.32 (±0.07)X - - X 22.59 18.07% 69.60 (±0.19)X - X - 31.64 25.31% 69.39 (±0.25)X - X X 32.29 25.83% 69.47 (±0.15)- X - - 66.39 53.11% 69.02 (±0.19)- X - X 67.04 53.63% 68.74 (±0.21)- X X - 76.09 60.88% 68.88 (±0.25)- X X X 76.75 61.40% 68.93 (±0.26)X X - - 88.32 70.66% 68.45 (±0.21)X X - X 88.97 71.18% 68.54 (±0.23)X X X - 98.03 78.42% 68.51 (±0.23)X X X X 98.68 78.94% 68.47 (±0.10)

training does not have a very strong effect on the final mAP.This is likely because the backbones of detection networksare initialized with ImageNet weights and are not randomas is the case with other papers dealing with LTH in theclassification setting. Therefore, the weights are more sta-ble and late resetting is not necessary. We also additionallyanalyze effects of resetting towards the end of training andnotice that there is a sharp drop in the performance after8k iterations. This is because the learning rate is decayedat this stage of training and the parameters change signif-icantly right after. A similar case holds when we performlearning rate warmup but do late resetting before the learn-ing rate is fully warmed up. The performance drops signif-icantly as the learning rate keeps fluctuating showing thatlate resetting is quite sensitive to learning rate.

Pruning different Faster-RCNN modules: We prune 20%of the parameters of the various modules within the Faster-RCNN network and analyze their effects on the mAP. Wealso try different combinations of pruning with the mod-ules and report the results in Table 3. Pruning the Box andClassification head (which takes up only 65% weights) out-performs the baseline case of no pruning, but does not al-ways improve performance when other modules are beingpruned. Additionally, pruning the RPN module increasesthe performance slightly even though it comprises of only10% of the network weights. Next, pruning the Base mod-ule and/or the Top module of the backbone leads to a drop inperformance, which is expected as they consist of 22% and66% of the weights respectively. Pruning the Base alone,excluding the Top, performs nearly as well as the baseline,while including the Top yields a lower mAP.

5

0 20 40 60 80Sparsity %

60

62

64

66

68

70

mAP

Effect of prune percentage

0 2000 4000 6000 8000 10000Reset iteration

60

62

64

66

68

70

mAP

Effect of late resetting

5 10 15Layer number

40

50

60

70

80

90

100

Prun

e %

Prune % across layers

(a) (b) (c)

1 2 3 4 5 6Number of rounds

60

62

64

66

68

70

mAP

Effect of number of pruning rounds

20 40 60 80Sparsity %

60

62

64

66

68

70m

AP

Effect of layerwise pruning

Global pruningLayerwise pruning

2000 4000 6000 8000 10000Early bird iteration

65

66

67

68

69

mAP

mAPIoU

0.90

0.92

0.94

0.96

0.98

1.00

IoU

Effect of early bird training

(d) (e) (f)

Figure 2: Effect of varying different hyperparameters for pruning Faster RCNN with ResNet-18 backbone on the Pascal VOC 2007 [12] dataset. All solidlines reported are the values averaged over 5 runs and the error bands are within 3 times the standard deviation.

Performance of Early-bird tickets: [55] showed that tick-ets can be found at early stages of training. We visualize thisby obtaining masks at various stages in training and evalu-ating their performance. We also plot each masks’ Inter-section over Union (IoU) with the default mask obtained atthe end of training. This IoU shows the overlap in the pa-rameters being pruned. The results are visualized in Fig. 2f.We see that within 50% of network training we find tick-ets whose performance is within a standard deviation of theperformance of the default ticket (obtained at the end oftraining). This is because the IoU becomes more or less sta-ble at around 0.96 during the middle stages of training andthe mask is unchanged as training advances. This allowsus to cut down on the number of training iterations signifi-cantly with very little cost to the network performance.Effect of number of rounds of pruning: [13] states thatiterative pruning performs better than one-shot pruning onthe classification task with small datasets and networks. Weshow that this does not necessarily hold true for detectionand larger backbones. We plot the network’s performanceagainst various rounds of pruning and observe that one-shotpruning outperforms iterative methods in Fig. 2d.Layer-wise vs. global pruning: [15] performs global prun-ing for larger datasets and networks and claims that pruningat the same rate in lower layers as compared to higher lay-ers, is detrimental to the network performance. We evaluatethe two methods of pruning on the detection task and showthe results in Fig. 2e. Additionally, for global pruning, weplot the percentage of parameters pruned in each layer ofthe backbone network in Fig. 2c. Layer-wise pruning does

as good as global pruning for lower levels of sparsity. How-ever, there is a noticeable performance gap for sparsity lev-els above 60%. This is because layer-wise pruning forceslower layers with very few parameters to have high sparsitypercentages. But as per Fig. 2c, for global pruning, we seethat lower layers are pruned less as they are crucial to boththe RPN and Classification stages of the network.

4.4. Properties of Winning Tickets

In Section 4.3, we showed that we can discover sparsernetworks within our two-stage Mask-RCNN detector if wedirectly prune on the task itself. We build upon those resultsto further probe the properties of winning tickets.

Effect of backbone architecture: In Fig. 3, we show howwinning tickets behave for 2 different backbones, ResNet-18 and ResNet-50, at different sparsity levels (50%, 80%,90%). We make two observations: First, the breaking pointfor both networks is ∼ 80% sparsity. However, perfor-mance of ResNet-18 drops more sharply than ResNet-50afterwards. This is intuitive since ResNet-18 has fewer re-dundant parameters and over-pruning leads to drop in theperformance. Second, as we gradually increase the spar-sity of the networks, mAP increases for all tasks in case ofboth networks. However, gains for ResNet-18 models areconsistently more than ResNet-50.

Do winning tickets behave differently for varying objectsizes? Using the definition from [32], we categorize bound-ing boxes into small (area < 322), medium (322 < area< 962), and large (area > 322). To understand how sparse

6

10 20 30 40 50 60 70 80 90Sparsity (% pruned)

10

5

0

5

% c

hang

e in

mA

P

(a) Bounding Box Detection

BackboneResNet18 (unpruned mAP: 29.92)ResNet50 (unpruned mAP: 38.52)

10 20 30 40 50 60 70 80 90Sparsity (% pruned)

10

5

0

5(b) Instance Segmentation

BackboneResNet18 (unpruned mAP: 27.64)ResNet50 (unpruned mAP: 35.13)

10 20 30 40 50 60 70 80 90Sparsity (% pruned)

1

0

1

2

(c) Keypoint Estimation

BackboneResNet18 (unpruned mAP: 58.60)ResNet50 (unpruned mAP: 64.59)

Figure 3: ResNet-18 vs. ResNet-50. We analyse change in mAP by using LTH on Mask R-CNN with different backbones.

(a) Bounding Box Detection (b) Instance Segmentation (c) Keypoint Estimation

Figure 4: Comparison of Mean Average Precision (mAP) of pruned model for different object sizes in case of Object Detection, Instance Segmentation,Keypoint Estimation. x-axis shows the sparsity of the subnetwork (or the percentage of weights removed). y-axis shows the percentage drop in mAP ascompared to the unpruned network. For all tasks, and object sizes, performance doesn’t drop till about 80% sparsity. After which, small objects are hitslightly harder as compared to medium and large objects.

networks behave for different sized objects, we plot the per-centage gain or drop from the mAP of a dense network.Figure 4 shows the percentage change in mAP for differ-ent levels of sparsity in the Mask R-CNN model. We canobserve that in each case, the model performance increaseswith sparsity, until sparsity reaches 80%, after which, mAPsharply declines. We note that the percentage drop for smallboxes is more, with winning tickets (10% of weights) show-ing a drop of over 17% in case of detection and segmenta-tion tasks while medium sized objects show smaller dropsthan large objects for all tasks.

How does the performance of the pruned network varyfor rare vs. frequent categories? We sort the 80 objectcategories in COCO by their frequency of occurrence intraining data. We consider networks with 80% and 90%of their weights pruned and observe the percentage changein the bounding box mAP of the model with respect to theunpruned network for each of the categories. Figure 5(a)depicts the behavior with a bar graph. While for most cat-egories, winning tickets are obtained at 80% sparsity, per-formance drops sharply with more pruning in case of rarecategories (such as toaster, parking meter, and bear) as com-pared to common categories (such as person, car, and chair).

Do the winning tickets behave differently on easy vs

hard categories? For a machine learning model, an ob-ject can be easy or hard to recognize because of a varietyof reasons. We have already discussed two reasons that in-fluence the performance — number of instances availablein the training data, and size of the object. There can alsobe other causes that can render an object unrecognizable ingiven surroundings. Camouflage or occlusion, poor cam-era quality, light conditions, distance from the camera, orjust variations within different instances or views of the ob-ject are few of them. Since exhaustive analyses of thesecauses is intractable, we rank object categories based onperformance of an unpruned Mask R-CNN model. We dothis categorization for detection and segmentation modelsas shown in Figure 5(b) and (c). Note that ‘easy’ and ‘hard’categories from these two definitions have an overlap butthey are not the same. For example, knife, handbag, andspoon are the categories with lowest bounding box mAP,and giraffe, zebra, and stop signs are one with the high-est (excluding ‘hair drier’ which has 0 mAP). On the otherhand, skis, knife, and spoon have the lowest segmentationmAP, while stop sign, bear, and fire hydrant have the high-est. From the Figure 5(b) and (c), we make the followingobservations — (i) tickets with 80% sparsity can actuallyincrease mAP for certain categories like snowboard by asmuch as 38%, (ii) Going from 80% to 90% sparsity, mAP

7

Low High Easy Hard Easy Hard

Figure 5: Comparison of Mean Average Precision (mAP) of pruned model for 80 COCO object categories. x-axis in each of the plot is a list of categories(sorted using different criteria). y-axis shows the percentage drop in mAP as compared to the unpruned network.

Table 4: Effect of ticket transfer across tasks. Transferred tickets do worsethan direct training as expected, but still do not result in drastic drops inthe mAP or AP50. Here we do task transfer using the 80% pruned model.

Targettask

Sourcetask

Networksparsity

mAP AP50

DetDet/Seg 78.4% 30.04 49.40Keypoint 50.11% 23.94 41.08

SegDet/Seg 78.4% 27.90 46.68Keypoint 50.11% 23.02 39.01

KeypointDet/Seg 76.98% 58.31 81.53Keypoint 79.4% 59.34 82.36

drops significantly for easy categories compared to hard cat-egories, (iii) categories that are hit the hardest such as skis,hot dog, spoon, fork, handbags usually have long, thin ap-pearance in images.

Do winning tickets transfer across tasks? We showed thatImageNet tickets transfer to a limited extent to downstreamtasks. We further study whether the tickets obtained fromthe downstream task of detection/segmentation transfer tokeypoint estimation and vice-versa. We train Mask-RCNNand Keypoint-RCNN respectively for the two tasks on theCOCO dataset while maintaining a sparsity level of 80%.For both the tasks we transfer all values till box head mod-ules, after which the model structures differ. The results areshown in Table 4. We can observe that the drop is marginalfor the transfer of tickets between detection-segmentation tokeypoint task, as compared with the reverse case which reg-isters a significant drop. This might be because the ticket isobtained on the keypoint task which is trained only on ‘hu-man’ class and it fails to transfer well for the detection taskwhich uses the entire COCO dataset.

5. Discussion

[38, 39] show that winning tickets transfer well acrossdatasets. However, the study in [38] was limited to smallerdatasets, like CIFAR-10 and FashionMNIST, and both [38,39] are limited to classification tasks. We obtain contrast-

ing results when transferring tickets across tasks as shownin Sec. 4.2. ImageNet tickets transfer with approximately40% sparsity to fall within one standard deviation of thebaseline network. This is likely due to the fact that win-ning tickets retain inductive biases from the source datasetwhich are less likely to transfer to a new domain and task.Additionally, we show that unlike prior LTH works, itera-tive pruning degrades the performance of subnetworks ondetection and one-shot pruning provides the best networks.We also observe that due to the use of pre-trained weightsfrom ImageNet for the backbone of detection networks, lateresetting is not necessary for finding winning tickets. This isin contrast to the [15], which is restricted to the classifica-tion task involving random initialization for the networks.Like previous works, in our experiments as well, we findthat sparse lottery tickets often outperform the dense net-works themselves. However, we make another interestingobservation — in each of object recognition tasks, ticketswith fewer parameters such as ResNet-18 show more gainsin performance as compared to tickets with more parame-ters (ResNet-50). We also find that small and infrequent ob-jects face higher performance drop as the sparsity increases.

6. Conclusion

We investigate the Lottery Ticket Hypothesis in the con-text of various object recognition tasks. Our study re-veals that the main points of original LTH hold for differentrecognition tasks, i.e., we can find subnetworks or winningtickets in object recognition pipelines with up to 80% spar-sity, without any drop in performance on the task. Thesetickets are task-specific, and pre-trained ImageNet modeltickets don’t perform as well on the downstream recogni-tion tasks. We also analyse claims made in recent literatureregarding training and transfer of winning tickets from anobject recognition perspective. Finally, we analyse how thebehavior of sparse tickets differ from their dense counter-parts. In the future, we would like to investigate how muchspeed up can be achieved using these sparse models withvarious hardware [35] and software modifications [11]. Ex-

8

tending this analyses for even bigger datasets such as JFT-300M [48] or IG-1B [36] and for self-supervised learningtechniques is another direction to pursue.

Acknowledgements. This work was partially supported byDARPA GARD #HR00112020007 and a gift from Facebook AI.

References[1] Sanjeev Arora, Rong Ge, Behnam Neyshabur, and Yi Zhang.

Stronger generalization bounds for deep nets via a compres-sion approach. arXiv preprint arXiv:1802.05296, 2018. 2

[2] Jimmy Ba and Brendan Frey. Adaptive dropout for trainingdeep neural networks. In Advances in neural informationprocessing systems, pages 3084–3092, 2013. 2

[3] Daniel Bolya, Sean Foley, James Hays, and Judy Hoffman.Tide: A general toolbox for identifying object detection er-rors. In ECCV, 2020. 11

[4] C Brix, P Bahar, and H Ney. Successfully applying the stabi-lized lottery ticket hypothesis to the transformer architecture.ACL, 2020. 3

[5] Tianlong Chen, Jonathan Frankle, Shiyu Chang, Sijia Liu,Yang Zhang, Michael Carbin, and Zhangyang Wang. Thelottery tickets hypothesis for supervised and self-supervisedpre-training in computer vision models. arXiv preprintarXiv:2012.06908, 2020. 3

[6] T Chen, J Frankle, S Chang, S Liu, Y Zhang, Z Wang, andM Carbin. The lottery ticket hypothesis for pre-trained bertnetworks. In NeurIPS, 2020. 3

[7] Zhiyong Cheng, Daniel Soudry, Zexi Mao, and ZhenzhongLan. Training binary multilayer neural networks for im-age classification using expectation backpropagation. arXivpreprint arXiv:1503.03562, 2015. 2

[8] Jia Deng, Wei Dong, Richard Socher, Li-Jia Li, Kai Li,and Li Fei-Fei. Imagenet: A large-scale hierarchical imagedatabase. In 2009 IEEE conference on computer vision andpattern recognition, pages 248–255. Ieee, 2009. 2, 3

[9] Emily L Denton, Wojciech Zaremba, Joan Bruna, Yann Le-Cun, and Rob Fergus. Exploiting linear structure within con-volutional networks for efficient evaluation. In Advancesin neural information processing systems, pages 1269–1277,2014. 2

[10] Shrey Desai, Hongyuan Zhan, and Ahmed Aly. Evaluat-ing lottery tickets under distributional shifts. arXiv preprintarXiv:1910.12708, 2019. 3

[11] Erich Elsen, Marat Dukhan, Trevor Gale, and Karen Si-monyan. Fast sparse convnets. In Proceedings of theIEEE/CVF Conference on Computer Vision and PatternRecognition, pages 14629–14638, 2020. 8

[12] Mark Everingham, Luc Van Gool, Christopher KI Williams,John Winn, and Andrew Zisserman. The pascal visual objectclasses (voc) challenge. International journal of computervision, 88(2):303–338, 2010. 2, 6

[13] J Frankle and M Carbin. The lottery ticket hypothesis: Find-ing sparse, trainable neural networks. In ICLR, 2019. 2, 6,12

[14] Jonathan Frankle, Gintare Karolina Dziugaite, Daniel MRoy, and Michael Carbin. Linear mode connectivity and the

lottery ticket hypothesis. arXiv preprint arXiv:1912.05671,2019. 2

[15] Jonathan Frankle, Gintare Karolina Dziugaite, Daniel MRoy, and Michael Carbin. Stabilizing the lottery ticket hy-pothesis. arXiv preprint arXiv:1903.01611, 2019. 2, 3, 4, 5,6, 8

[16] J Frankle, G K Dziugaite, D M Roy, and M Carbin. Stabiliz-ing the lottery ticket hypothesis. In ICML, 2020. 2, 3

[17] Yunchao Gong, Liu Liu, Ming Yang, and Lubomir Bour-dev. Compressing deep convolutional networks using vectorquantization. arXiv preprint arXiv:1412.6115, 2014. 2

[18] Kathrin Grosse and Michael Backes. How many win-ning tickets are there in one dnn? arXiv preprintarXiv:2006.07014, 2020. 3

[19] Song Han, Huizi Mao, and William J Dally. Deep com-pression: Compressing deep neural networks with pruning,trained quantization and huffman coding. arXiv preprintarXiv:1510.00149, 2015. 2

[20] Song Han, Jeff Pool, John Tran, and William Dally. Learningboth weights and connections for efficient neural network. InAdvances in neural information processing systems, pages1135–1143, 2015. 2

[21] Trevor Hastie, Robert Tibshirani, and Martin Wainwright.Statistical learning with sparsity: the lasso and generaliza-tions. CRC press, 2015. 2

[22] Kaiming He, Georgia Gkioxari, Piotr Dollar, and Ross Gir-shick. Mask r-cnn. In Proceedings of the IEEE internationalconference on computer vision, pages 2961–2969, 2017. 2,4

[23] Kaiming He, Xiangyu Zhang, Shaoqing Ren, and Jian Sun.Deep residual learning for image recognition. In Proceed-ings of the IEEE conference on computer vision and patternrecognition, pages 770–778, 2016. 1, 2

[24] Yanping Huang, Yonglong Cheng, Dehao Chen, Hy-oukJoong Lee, Jiquan Ngiam, Quoc V. Le, and ZhifengChen. Gpipe: Efficient training of giant neural networks us-ing pipeline parallelism. CoRR, abs/1811.06965, 2018. 1

[25] Itay Hubara, Matthieu Courbariaux, Daniel Soudry, Ran El-Yaniv, and Yoshua Bengio. Quantized neural networks:Training neural networks with low precision weights andactivations. The Journal of Machine Learning Research,18(1):6869–6898, 2017. 2

[26] Simon Kornblith, Jonathon Shlens, and Quoc V Le. Do bet-ter imagenet models transfer better? In Proceedings of theIEEE conference on computer vision and pattern recogni-tion, pages 2661–2671, 2019. 1

[27] Alex Krizhevsky et al. Learning multiple layers of featuresfrom tiny images. 2009. 3

[28] Yann LeCun, Corinna Cortes, and CJ Burges. Mnist hand-written digit database. ATT Labs [Online]. Available:http://yann.lecun.com/exdb/mnist, 2, 1998. 3

[29] Yann LeCun, John S Denker, and Sara A Solla. Optimalbrain damage. In Advances in neural information processingsystems, pages 598–605, 1990. 2

[30] N Lee, T Ajanthan, and P Torr. SNIP: Single-shot networkpruning based on connection sensitivity. In ICLR, 2019. 2, 3

9

[31] Rundong Li, Yan Wang, Feng Liang, Hongwei Qin, JunjieYan, and Rui Fan. Fully quantized network for object detec-tion. In Proceedings of the IEEE Conference on ComputerVision and Pattern Recognition, pages 2810–2819, 2019. 2

[32] Tsung-Yi Lin, Michael Maire, Serge Belongie, James Hays,Pietro Perona, Deva Ramanan, Piotr Dollar, and C LawrenceZitnick. Microsoft coco: Common objects in context. InEuropean conference on computer vision, pages 740–755.Springer, 2014. 2, 6

[33] Zhuang Liu, Jianguo Li, Zhiqiang Shen, Gao Huang,Shoumeng Yan, and Changshui Zhang. Learning efficientconvolutional networks through network slimming. In Pro-ceedings of the IEEE International Conference on ComputerVision, pages 2736–2744, 2017. 2

[34] Christos Louizos, Max Welling, and Diederik P Kingma.Learning sparse neural networks through l 0 regularization.arXiv preprint arXiv:1712.01312, 2017. 2

[35] Liqiang Lu, Jiaming Xie, Ruirui Huang, Jiansong Zhang,Wei Lin, and Yun Liang. An efficient hardware ac-celerator for sparse convolutional neural networks on fp-gas. In 2019 IEEE 27th Annual International Sympo-sium on Field-Programmable Custom Computing Machines(FCCM), pages 17–25. IEEE, 2019. 8

[36] Dhruv Mahajan, Ross Girshick, Vignesh Ramanathan,Kaiming He, Manohar Paluri, Yixuan Li, Ashwin Bharambe,and Laurens van der Maaten. Exploring the limits of weaklysupervised pretraining. In Proceedings of the European Con-ference on Computer Vision (ECCV), pages 181–196, 2018.1, 9

[37] Eran Malach, Gilad Yehudai, Shai Shalev-Shwartz, andOhad Shamir. Proving the lottery ticket hypothesis: Prun-ing is all you need. arXiv preprint arXiv:2002.00585, 2020.3

[38] R Mehta. Sparse transfer learning via winning lottery tickets.In NeurIPS, 2020. 2, 3, 8, 11

[39] Ari Morcos, Haonan Yu, Michela Paganini, and YuandongTian. One ticket to win them all: generalizing lottery ticketinitializations across datasets and optimizers. In Advances inNeural Information Processing Systems, pages 4932–4942,2019. 2, 3, 8

[40] Rajiv Movva and Jason Y. Zhao. Dissecting lottery tickettransformers: Structural and behavioral study of sparse neu-ral machine translation. ArXiv, abs/2009.13270, 2020. 3

[41] Joseph Redmon and Ali Farhadi. Yolov3: An incrementalimprovement. arXiv preprint arXiv:1804.02767, 2018. 3

[42] Shaoqing Ren, Kaiming He, Ross Girshick, and Jian Sun.Faster r-cnn: Towards real-time object detection with regionproposal networks. In Advances in neural information pro-cessing systems, pages 91–99, 2015. 2, 3

[43] Alex Renda, Jonathan Frankle, and Michael Carbin. Com-paring rewinding and fine-tuning in neural network pruning.arXiv preprint arXiv:2003.02389, 2020. 2

[44] Matteo Ruggero Ronchi and Pietro Perona. Benchmarkingand error diagnosis in multi-instance pose estimation. InICCV, 2017. 11

[45] Andrey Salvi and Rodrigo Barros. An experimental analysisof model compression techniques for object detection. In

Anais do VIII Symposium on Knowledge Discovery, Miningand Learning, pages 49–56, Porto Alegre, RS, Brasil, 2020.SBC. 3

[46] Daniel Soudry, Itay Hubara, and Ron Meir. Expectationbackpropagation: Parameter-free training of multilayer neu-ral networks with continuous or discrete weights. In Ad-vances in neural information processing systems, pages 963–971, 2014. 2

[47] Nitish Srivastava, Geoffrey Hinton, Alex Krizhevsky, IlyaSutskever, and Ruslan Salakhutdinov. Dropout: a simple wayto prevent neural networks from overfitting. The journal ofmachine learning research, 15(1):1929–1958, 2014. 2

[48] Chen Sun, Abhinav Shrivastava, Saurabh Singh, and Abhi-nav Gupta. Revisiting Unreasonable Effectiveness of Datain Deep Learning Era. In IEEE International Conference onComputer Vision (ICCV), 2017. 9

[49] Mingxing Tan and Quoc V Le. Efficientnet: Rethinkingmodel scaling for convolutional neural networks. arXivpreprint arXiv:1905.11946, 2019. 1

[50] Robert Tibshirani. Regression shrinkage and selection viathe lasso. Journal of the Royal Statistical Society: Series B(Methodological), 58(1):267–288, 1996. 2

[51] Vincent Vanhoucke, Andrew Senior, and Mark Z Mao. Im-proving the speed of neural networks on cpus. 2011. 2

[52] Chaoqi Wang, Guodong Zhang, and Roger Grosse. Pickingwinning tickets before training by preserving gradient flow.arXiv preprint arXiv:2002.07376, 2020. 3

[53] Wei Wen, Chunpeng Wu, Yandan Wang, Yiran Chen, andHai Li. Learning structured sparsity in deep neural networks.In Advances in neural information processing systems, pages2074–2082, 2016. 2

[54] Jiaxiang Wu, Cong Leng, Yuhang Wang, Qinghao Hu, andJian Cheng. Quantized convolutional neural networks formobile devices. In Proceedings of the IEEE Conferenceon Computer Vision and Pattern Recognition, pages 4820–4828, 2016. 2

[55] H You, C Li, P Xu, Y Fu, Y Wang, X Chen, R G Baraniuk, ZWang, and Y Lin. Drawing early-bird tickets: Toward moreefficient training of deep networks. In ICLR, 2020. 3, 6

[56] H Yu, S Edunov S, Y Tian Y, and A S Morcos. Playing thelottery with rewards and multiple languages: lottery ticketsin RL and NLP. In ICLR, 2020. 3

[57] Sergey Zagoruyko and Nikos Komodakis. Wide residual net-works. arXiv preprint arXiv:1605.07146, 2016. 1

[58] Hattie Zhou, Janice Lan, Rosanne Liu, and Jason Yosinski.Deconstructing lottery tickets: Zeros, signs, and the super-mask. In Advances in Neural Information Processing Sys-tems, pages 3597–3607, 2019. 2

10

AppendixWe provide additional details for some of the experiments pre-sented in the paper. In particular, we provide comparison with asimpler ImageNet ticket transfer alternative in Section A, comparethe different errors made by dense and pruned models in Section B,and finally verify the faster convergence of sparser models in Sec-tion C.

A. Mask Transfer Without RetrainingIn Section 4.2, we analyzed the effects of transferring tickets

only for the ImageNet trained backbones. While this deals withtransferring the ticket mask as well as values, we further analyzewhether transferring only the mask provides winning tickets forthese tasks using the methodology from [38]. We use the defaultImageNet weights in the ResNet-18 and ResNet-50 backbone andkeep the top p% of the weights in convolutional layers while set-ting the rest to zeros and maintaining it throughout the training ofthe entire network. We refer to this method as ‘Mask Transfer’.Since training the backbone on much larger ImageNet data is per-formed only once, ‘Mask Transfer’ is a much cheaper or compu-tationally efficient way of obtaining tickets from parent task. Weobserve that behavior of ‘Mask Transfer‘ is similar to the ‘Trans-fer Ticket‘ obtained by method discussed in Section 4.2 where thesparse subnetwork weights are fully retrained on ImageNet. Eithercases are outperformed by direct pruning on the downstream tasks.The results are summarized in Figure 6 (ResNet-18) and Table 5(ResNet-50).

B. Error analysis on downstream tasksThe mAP score provides us a good way to summarize the per-

formance of an object recognition model with a single number.But it hides a lot of information regarding what kind of mistakesthe model is making. Do the sparse subnetworks obtained by LTHmake same mistakes as the dense models? In order to answer thisquestion, we consider a dense Mask R-CNN model with ResNet-50 backbone and a sparse Mask R-CNN model with 20% of theparameters obtained via LTH. Both the models achieve same per-formance on downstream tasks as also discussed in Section 4.2 ofthe paper.

B.1. Object Detection and Instance SegmentationWe resort to a toolbox from [3] to analyze object detection and

instance segmentation errors. We consider 5 main sources of er-rors in object detection. (i) ‘Cls’ refers to an error corresponding tomiss-classification of a bounding box by a model, (ii) ‘Loc’ refersto the case when bounding box is classified properly but not local-ized properly, (iii) ‘Dupe’ corresponds to the errors when modelmakes multiple predictions at the same location, (iv) ‘Bkgd’ arethe cases when background portion of the image (with no objects)are tagged as an object, and finally (v) ‘Missed’ cases when theobjects are not detected by the model.

Figures 7 and 8 summarize the analysis of detection and seg-mentation errors obtained for dense model as compared to a sparsemodel (with only 20% of the weights). While in the case of objectdetection, the performance of both the models is identical, subtle

24

26

28

30

32Object Detection

24

26

28

mea

n A

vera

ge P

reci

sion

Instance Segmentation

0 5 10 15 20 25 30 35 40Network Sparsity (% pruned)

56

57

58

59

60Keypoint Estimation

Unpruned Transfer Ticket Mask Transfer Direct Pruning via LTH

Figure 6: Transferring ImageNet backbone tickets to object recognitiontasks vs. Direct pruning via LTH on the object recognition tasks. We ex-periment with two variations of transferring ImageNet backbone tickets toobject recognition tasks. ‘Transfer ticket’ refers to the case when we trans-fer the lottery ticket backbone trained on ImageNet data to downstreamtask (also discussed in the Section 4 of the paper). ‘Mask Transfer’ refersto the case when ticket is transferred without retraining on ImageNet, i.e.,only the relevant mask from backbone is transferred keeping ImageNetweights the same. Best viewed in color.

(a) Unpruned model (b) Pruned model

Figure 7: Error analysis of unpruned vs. pruned on object detection. Theerror types of unpruned and pruned models are nearly the same.

differences emerge in case of segmentation where sparse modelmakes fewer localization errors but higher background errors.

B.2. Keypoint EstimationWe use [44] to perform a similar analyses for sparse and dense

models on the task of keypoint estimation. In case of keypoints,

11

Table 5: Performance on the COCO dataset for ImageNet backbones with mask transfer tickets for ResNet-50 at various levels of pruning. The results forVOC are averaged over 5 runs with the standard deviation in parantheses.

Prune %COCO Detection COCO segmentation COCO Keypoint VOC Detection

Networksparsity

mAP AP50Networksparsity

mAP AP50Networksparsity

mAP AP50Networksparsity

mAP

90% 41.99% 35.46 56.51 41.99% 32.40 52.89 31.49% 62.27 84.93 65.37% 61.75(±0.22)

80% 37.33% 36.52 57.28 37.33% 33.53 54.15 28% 63.48 85.72 58.11% 67.30(±0.39)

50% 24.55% 37.99 58.83 24.55% 34.76 55.91 19.71% 64.21 86.32 36.32% 70.32(±0.23)

0% 0% 38.5 59.29 0% 35.13 56.39 0% 64.59 86.48 0% 71.21(±0.32)

(a) Unpruned model (b) Pruned model

Figure 8: Error analysis of unpruned vs. pruned on instance segmentation.The error types of unpruned and pruned models are quite similar.

we compute the Precision Recall Curve of the model while remov-ing the impact of individual errors of following kinds — (i) ‘Miss’- large localization errors, (ii) ‘Swap’ - confusion between samekeypoint of two different persons, (iii) ‘Inversion’ - confusion be-tween two different keypoints of the same person, (iv) ‘Jitter’ -small localization error, and (v) ‘FP’ - background false positives.Fig. 9 summarizes the results. As in the previous case, it appearsthat both the dense and sparse models make similar mistakes.

C. Sparse subnetworks converge fasterThe LTH paper [13] claimed that sparse subnetworks obtained

by pruning, often converge faster than their dense counterparts. Inthis section we verify the claims of the paper on object recogni-tion tasks and found them to hold true. We plot the validation lossduring training for the dense unpruned model, and the sparse sub-network obtained by keeping only 20% of the weights of densemodel. Both the models achieve a similar mAP after convergence.Fig. 10 shows the task loss against the number of epochs dur-ing training. The comparisons confirm that the sparse subnetworkinitialized from the winning ticket weights converge much faster.This observation is consistent for heterogeneous tasks, e.g., objectdetection, instance segmentation, and keypoint estimation.

12

0.0 0.2 0.4 0.6 0.8 1.0recall

0.0

0.2

0.4

0.6

0.8

1.0

prec

isio

n

oksThrs:[0.5], areaRng:[all], maxDets:[20]

w/o FN : 1.00w/o Bkg. FP: .970Opt. Score : .968w/o Jit. : .912w/o Inv. : .912w/o Swap : .900w/o Miss : .893Orig. Dts. : .826

0.0 0.2 0.4 0.6 0.8 1.0recall

0.0

0.2

0.4

0.6

0.8

1.0

prec

isio

n

oksThrs:[0.5], areaRng:[all], maxDets:[20]

w/o FN : 1.00w/o Bkg. FP: .970Opt. Score : .969w/o Jit. : .910w/o Inv. : .910w/o Swap : .901w/o Miss : .897Orig. Dts. : .830

0.0 0.2 0.4 0.6 0.8 1.0recall

0.0

0.2

0.4

0.6

0.8

1.0

prec

isio

n

oksThrs:[0.5], areaRng:[all], maxDets:[20]

w/o FN : 1.00w/o Bkg. FP: .970Opt. Score : .968w/o Jit. : .912w/o Inv. : .912w/o Swap : .900w/o Miss : .893Orig. Dts. : .826

0.0 0.2 0.4 0.6 0.8 1.0recall

0.0

0.2

0.4

0.6

0.8

1.0

prec

isio

n

oksThrs:[0.5], areaRng:[all], maxDets:[20]

w/o FN : 1.00w/o Bkg. FP: .970Opt. Score : .969w/o Jit. : .910w/o Inv. : .910w/o Swap : .901w/o Miss : .897Orig. Dts. : .830

(a)Unpruned model (b) Pruned model

Figure 9: Error analysis of unpruned vs. pruned on kyepoint estimation. The error types of unpruned and pruned models are quite similar while the unprunedone has slightly better performance

Loss

Iterations (k) Iterations (k)

Loss

Iterations (k)

Loss

(a) Object detection (b) Instance segmentation (c) Keypoint estimation

Figure 10: Training curves of dense model vs. sparse model

13

The Lottery Ticket Hypothesis for Object Recognition: Supplementary Material

Sharath Girish**

[email protected]

Shishira R Maiya*

[email protected]

Kamal [email protected]

Hao [email protected]

Larry [email protected]

Abhinav [email protected]

University of Maryland, College Park

We provide additional details for some of the experi-ments presented in the paper. In particular, we provide com-parison with a simpler ImageNet ticket transfer alternativein Section A, compare the different errors made by denseand pruned models in Section B, verify the faster conver-gence of sparser models in Section C and finally analyze thedisk space and number of compute operations in Section D.

A. Mask Transfer Without RetrainingIn Section 4.2, we analyzed the effects of transferring

tickets only for the ImageNet trained backbones. Whilethis deals with transferring the ticket mask as well as val-ues, we further analyze whether transferring only the maskprovides winning tickets for these tasks using the method-ology from [?]. We use the default ImageNet weights in theResNet-18 and ResNet-50 backbone and keep the top p%of the weights in convolutional layers while setting the restto zeros and maintaining it throughout the training of theentire network. We refer to this method as ‘Mask Transfer’.Since training the backbone on much larger ImageNet datais performed only once, ‘Mask Transfer’ is a much cheaperor computationally efficient way of obtaining tickets fromparent task. We observe that behavior of ‘Mask Transfer‘is similar to the ‘Transfer Ticket‘ obtained by method dis-cussed in Section 4.2 where the sparse subnetwork weightsare fully retrained on ImageNet. Either cases are outper-formed by direct pruning on the downstream tasks. Theresults are summarized in Figure 1 (ResNet-18) and Table 1(ResNet-50).

B. Error analysis on downstream tasksThe mAP score provides us a good way to summarize

the performance of an object recognition model with a sin-gle number. But it hides a lot of information regarding whatkind of mistakes the model is making. Do the sparse subnet-works obtained by LTH make same mistakes as the dense

*Equal contribution

24

26

28

30

32Object Detection

24

26

28

mea

n A

vera

ge P

reci

sion

Instance Segmentation

0 5 10 15 20 25 30 35 40Network Sparsity (% pruned)

56

57

58

59

60Keypoint Estimation

Unpruned Transfer Ticket Mask Transfer Direct Pruning via LTH

Figure 1: Transferring ImageNet backbone tickets to object recognitiontasks vs. Direct pruning via LTH on the object recognition tasks. We ex-periment with two variations of transferring ImageNet backbone tickets toobject recognition tasks. ‘Transfer ticket’ refers to the case when we trans-fer the lottery ticket backbone trained on ImageNet data to downstreamtask (also discussed in the Section 4 of the paper). ‘Mask Transfer’ refersto the case when ticket is transferred without retraining on ImageNet, i.e.,only the relevant mask from backbone is transferred keeping ImageNetweights the same. Best viewed in color.

models? In order to answer this question, we consider adense Mask R-CNN model with ResNet-50 backbone and asparse Mask R-CNN model with 20% of the parameters ob-tained via LTH. Both the models achieve same performanceon downstream tasks as also discussed in Section 4.2 of thepaper.

arX

iv:2

012.

0464

3v2

[cs

.CV

] 1

9 A

pr 2

021

Table 1: Performance on the COCO dataset for ImageNet backbones with mask transfer tickets for ResNet-50 at various levels of pruning. The results forVOC are averaged over 5 runs with the standard deviation in parantheses.

Prune %COCO Detection COCO segmentation COCO Keypoint VOC Detection

Networksparsity

mAP AP50Networksparsity

mAP AP50Networksparsity

mAP AP50Networksparsity

mAP

90% 41.99% 35.46 56.51 41.99% 32.40 52.89 31.49% 62.27 84.93 65.37% 61.75(±0.22)

80% 37.33% 36.52 57.28 37.33% 33.53 54.15 28% 63.48 85.72 58.11% 67.30(±0.39)

50% 24.55% 37.99 58.83 24.55% 34.76 55.91 19.71% 64.21 86.32 36.32% 70.32(±0.23)

0% 0% 38.5 59.29 0% 35.13 56.39 0% 64.59 86.48 0% 71.21(±0.32)

(a) Unpruned model (b) Pruned model

Figure 2: Error analysis of unpruned vs. pruned on object detection. Theerror types of unpruned and pruned models are nearly the same.

B.1. Object Detection and Instance Segmentation

We resort to a toolbox from [?] to analyze object de-tection and instance segmentation errors. We consider 5main sources of errors in object detection. (i) ‘Cls’ refers toan error corresponding to miss-classification of a boundingbox by a model, (ii) ‘Loc’ refers to the case when boundingbox is classified properly but not localized properly, (iii)‘Dupe’ corresponds to the errors when model makes mul-tiple predictions at the same location, (iv) ‘Bkgd’ are thecases when background portion of the image (with no ob-jects) are tagged as an object, and finally (v) ‘Missed’ caseswhen the objects are not detected by the model.

Figures 2 and 3 summarize the analysis of detection andsegmentation errors obtained for dense model as comparedto a sparse model (with only 20% of the weights). Whilein the case of object detection, the performance of both themodels is identical, subtle differences emerge in case of seg-mentation where sparse model makes fewer localization er-rors but higher background errors.

B.2. Keypoint Estimation

We use [?] to perform a similar analyses for sparse anddense models on the task of keypoint estimation. In case of

(a) Unpruned model (b) Pruned model

Figure 3: Error analysis of unpruned vs. pruned on instance segmentation.The error types of unpruned and pruned models are quite similar.

0 20 40 60 80Sparsity %

50

100

150

Disk

Spa

ce (M

B)

Disk spaceDet/SegKey

0 20 40 60 80Sparsity %

70

75

80

MAC

(G)

MAC operationsDet/SegKey

Figure 4: Disk space and MAC operations of pruned models withResNet18 backbone for the various tasks on the COCO dataset.

keypoints, we compute the Precision Recall Curve of themodel while removing the impact of individual errors offollowing kinds — (i) ‘Miss’ - large localization errors, (ii)‘Swap’ - confusion between same keypoint of two differentpersons, (iii) ‘Inversion’ - confusion between two differentkeypoints of the same person, (iv) ‘Jitter’ - small localiza-tion error, and (v) ‘FP’ - background false positives. Fig. 5summarizes the results. As in the previous case, it appearsthat both the dense and sparse models make similar mis-takes.

0.0 0.2 0.4 0.6 0.8 1.0recall

0.0

0.2

0.4

0.6

0.8

1.0

prec

isio

n

oksThrs:[0.5], areaRng:[all], maxDets:[20]

w/o FN : 1.00w/o Bkg. FP: .970Opt. Score : .968w/o Jit. : .912w/o Inv. : .912w/o Swap : .900w/o Miss : .893Orig. Dts. : .826

0.0 0.2 0.4 0.6 0.8 1.0recall

0.0

0.2

0.4

0.6

0.8

1.0

prec

isio

n

oksThrs:[0.5], areaRng:[all], maxDets:[20]

w/o FN : 1.00w/o Bkg. FP: .970Opt. Score : .969w/o Jit. : .910w/o Inv. : .910w/o Swap : .901w/o Miss : .897Orig. Dts. : .830

0.0 0.2 0.4 0.6 0.8 1.0recall

0.0

0.2

0.4

0.6

0.8

1.0

prec

isio

n

oksThrs:[0.5], areaRng:[all], maxDets:[20]

w/o FN : 1.00w/o Bkg. FP: .970Opt. Score : .968w/o Jit. : .912w/o Inv. : .912w/o Swap : .900w/o Miss : .893Orig. Dts. : .826

0.0 0.2 0.4 0.6 0.8 1.0recall

0.0

0.2

0.4

0.6

0.8

1.0

prec

isio

n

oksThrs:[0.5], areaRng:[all], maxDets:[20]

w/o FN : 1.00w/o Bkg. FP: .970Opt. Score : .969w/o Jit. : .910w/o Inv. : .910w/o Swap : .901w/o Miss : .897Orig. Dts. : .830

(a)Unpruned model (b) Pruned model

Figure 5: Error analysis of unpruned vs. pruned on kyepoint estimation. The error types of unpruned and pruned models are quite similar while the unprunedone has slightly better performance

Loss

Iterations (k) Iterations (k)

Loss

Iterations (k)

Loss

(a) Object detection (b) Instance segmentation (c) Keypoint estimation

Figure 6: Training curves of dense model vs. sparse model

C. Sparse subnetworks converge faster

The LTH paper [?] claimed that sparse subnetworks ob-tained by pruning, often converge faster than their densecounterparts. In this section we verify the claims of the pa-per on object recognition tasks and found them to hold true.We plot the validation loss during training for the dense un-pruned model, and the sparse subnetwork obtained by keep-ing only 20% of the weights of dense model. Both the mod-els achieve a similar mAP after convergence. Fig. 6 showsthe task loss against the number of epochs during training.The comparisons confirm that the sparse subnetwork initial-

ized from the winning ticket weights converge much faster.This observation is consistent for heterogeneous tasks, e.g.,object detection, instance segmentation, and keypoint esti-mation.

D. Disk space and compute operationsFinally, we analyze the disk space and MAC operations

of pruned models in Figure 4. We store the index and val-ues of only the non zero weights, when above a thresholdsparsity, while we store the full weight values for densersparsity levels. As expected, we observe significant reduc-

tions in disk space for higher levels of sparsity. However,number of operations decreases at a much slower rate. Thiscan possibly be improved further with dedicated hardwarefor sparse operations.