pith. machine review for the scientific record. sign in

arxiv: 1412.7149 · v4 · submitted 2014-12-22 · 💻 cs.LG · cs.NE· stat.ML

Recognition: unknown

Deep Fried Convnets

Authors on Pith no claims yet
classification 💻 cs.LG cs.NEstat.ML
keywords convolutionaldeepnetworklayersmemorynetworksneuralparameters
0
0 comments X
read the original abstract

The fully connected layers of a deep convolutional neural network typically contain over 90% of the network parameters, and consume the majority of the memory required to store the network parameters. Reducing the number of parameters while preserving essentially the same predictive performance is critically important for operating deep neural networks in memory constrained environments such as GPUs or embedded devices. In this paper we show how kernel methods, in particular a single Fastfood layer, can be used to replace all fully connected layers in a deep convolutional neural network. This novel Fastfood layer is also end-to-end trainable in conjunction with convolutional layers, allowing us to combine them into a new architecture, named deep fried convolutional networks, which substantially reduces the memory footprint of convolutional networks trained on MNIST and ImageNet with no drop in predictive performance.

This paper has not been read by Pith yet.

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Forward citations

Cited by 1 Pith paper

Reviewed papers in the Pith corpus that reference this work. Sorted by Pith novelty score.

  1. Deep Compression: Compressing Deep Neural Networks with Pruning, Trained Quantization and Huffman Coding

    cs.CV 2015-10 conditional novelty 7.0

    A pruning-quantization-Huffman pipeline compresses deep neural networks 35-49x without accuracy loss.