카테고리 없음

Distiller 관련 논문 리스트

SciomageLAB 2024. 10. 18. 00:23
반응형

난 그냥 딥러닝 모델을 Edge에서 빨리 돌리고 싶을 뿐인데 배워야 할게 너무 많다.
Distiller에서 reference하는 논문 자체도 어마어마하게 많으니.. 이 리스트부터 정리하려고 한다.

- 이름 연도 태그
1 Learning both Weights and Connections for Efficient Neural Networks 2015 Pruning
2 Pruning Filters for Efficient ConvNets 2017 Pruning
3 Deep Learning 2017 Regularization
4 DSD: Dense-Sparse-Dense Training for Deep Neural Networks, 2017 Regularization
5 Exploring the Regularity of Sparse Structure in Convolutional Neural Networks 2017 Regularization
6 Structured pruning of deep convolutional neural networks 2015 Regularization
7 High-Performance Hardware for Machine Learning Quantization
8 Deep Compression : Compressing Deep Neural Networks with Pruning, Trained Quantization and Huffman Coding 2016
9 XNOR-Net: ImageNet Classification Using Binary Convolutional Neural Networks 2016 Quantization
10 Training deep neural networks with low precision multiplications Quantization
11 Ristretto: A Framework for Empirical Study of Resource-Efficient Inference in Convolutional Neural Networks. 2018 Quantization
12 8-bit Inference with TensorRT 2017 Quantization
13 DoReFa-Net: Training Low Bitwidth Convolutional Neural Networks with Low Bitwidth Gradients Quantization
14 Incremental Network Quantization: Towards Lossless CNNs with Low-precision Weights Quantization
15 WRPN: Wide Reduced-Precision Networks Quantization
16 PACT: Parameterized Clipping Activation for Quantized Neural Networks Quantization
17 Towards Accurate Binary Convolutional Neural Network Quantization
18 Learning both Weights and Connections for Efficient Neural Network Quantization
19 Ternary Weight Networks Quantization
20 Trained Ternary Quantization Quantization
21 Estimating or Propagating Gradients Through Stochastic Neurons for Conditional Computation Quantization
22 Neural Networks for Machine Learning Quantization
23 Quantization and Training of Neural Networks for Efficient Integer-Arithmetic-Only Inference Quantization
24 Quantizing deep convolutional networks for efficient inference: A whitepaper Quantization
25 ACIQ: Analytical Clipping for Integer Quantization of neural networks Quantization
26 Model Compression Knowledge Distillation
27 Distilling the Knowledge in a Neural Network Knowledge Distillation
28 Hardware-Software Codesign of Accurate, Multiplier-free Deep Neural Networks Knowledge Distillation
29 Apprentice: Using Knowledge Distillation Techniques To Improve Low-Precision Network Accuracy Knowledge Distillation
30 Model compression via distillation and quantization Knowledge Distillation
31 N2N learning: Network to Network Compression via Policy Gradient Reinforcement Learning 32 Knowledge Distillation
33 Faster gaze prediction with dense networks and Fisher pruning Knowledge Distillation

마지막 업데이트 : 2021-05-13

반응형