TL;DRMotivationMethodExperimentsTL;DR처음으로 Neural Network compression 제안.pruning9~13Xquantization32bits to 5bitshuffman coding→ accuracy 영향없이 storage saving을 30~40X 늘림. privacy, network bandwidth, storage, energy consumption 이점이 있음.MotivationMethodExperiments