Related Paper of Efficient Deep Neural Networks
- Efficient Processing of Deep Neural Networks: A Tutorial and Survey
- High-Performance Hardware for Machine Learning
- series of deep learning on iphone blogs
- Learning both Weights and Connections for Efficient Neural Networks [NIPS 2015]
- Dynamic Network Surgery for Efficient DNNs [NIPS2016] Code
- Learning Structured Sparsity in Deep Neural Networks [NIPS 2016] Code
- Sparse Convolutional Neural Networks [CVPR 2015]
- Pruning Filters for Efficient ConvNets [ICLR 2017]
- Incremental Network Quantization: Towards Lossless CNNs with Low-Precision Weights [ICLR 2017]
- [https://arxiv.org/pdf/1706.02393.pdf]
- Deep Compression: Compressing Deep Neural Networks with Pruning, Trained Quantization and Huffman Coding [ICLR 2016]
- XNOR-Net: ImageNet Classification Using Binary Convolutional Neural Networks [ECCV 2016] Code
- Quantized Neural Networks: Training Neural Networks with Low Precision Weights and Activations
- Trained Tenary Quantization [ICLR2017] Code
- DoReFa-Net: Training Low Bitwidth Convolutional Neural Networks with Low Bitwidth Gradients Code
- Binaryconnect: Training deep neural networks with binary weights during propagations [NIPS 2015]
- Binarize neural networks: Training deep neural networks with weights and activations constrained to +1 or -1 [NIPS 2016]
- Google’s Neural Machine Translation System: Bridging the Gap between Human and Machine Translation
- 8-Bit Approximations For Parallelism In Deep Learning [ICLR 2016]
- [Quantized Convolutional Neural Networks for Mobile Devices]
- SqueezeNet: AlexNet-Level Accuracy with 50x Fewer Parameters and < 0.5MB Model Size
- MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications Code
- PVANet: Lightweight Deep Neural Networks for Real-time Object Detection Code
- Distilling the Knowledge in a Neural Network [NIPS 2014]
- Fast Training of Convolutional Networks through FFTs [ICLR 2013] Code
- Fast algorithms for convolutional neural networks [CVPR 2016]
1.Movidius