diff --git a/MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications.pdf b/MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications.pdf new file mode 100644 index 0000000000000000000000000000000000000000..4e0f23fd689ca5a3ef468491f54db96647da81e5 Binary files /dev/null and b/MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications.pdf differ diff --git a/README.md b/README.md index 97b3d354a920a3f3940804d11bac9a1eb731a306..22fb1627cca0869ad9b590f702e1a5aa9d986487 100644 --- a/README.md +++ b/README.md @@ -96,4 +96,8 @@ We propose a deep convolutional neural network architecture codenamed "Inception Recent work has shown that convolutional networks can be substantially deeper, more accurate, and efficient to train if they contain shorter connections between layers close to the input and those close to the output. In this paper, we embrace this observation and introduce the Dense Convolutional Network (DenseNet), which connects each layer to every other layer in a feed-forward fashion. Whereas traditional convolutional networks with L layers have L connections - one between each layer and its subsequent layer - our network has L(L+1)/2 direct connections. For each layer, the feature-maps of all preceding layers are used as inputs, and its own feature-maps are used as inputs into all subsequent layers. DenseNets have several compelling advantages: they alleviate the vanishing-gradient problem, strengthen feature propagation, encourage feature reuse, and substantially reduce the number of parameters. We evaluate our proposed architecture on four highly competitive object recognition benchmark tasks (CIFAR-10, CIFAR-100, SVHN, and ImageNet). DenseNets obtain significant improvements over the state-of-the-art on most of them, whilst requiring less computation to achieve high performance. Code and pre-trained models are available at [this https URL](https://github.com/liuzhuang13/DenseNet) . +## MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications + +We present a class of efficient models called MobileNets for mobile and embedded vision applications. MobileNets are based on a streamlined architecture that uses depth-wise separable convolutions to build light weight deep neural networks. We introduce two simple global hyper-parameters that efficiently trade off between latency and accuracy. These hyper-parameters allow the model builder to choose the right sized model for their application based on the constraints of the problem. We present extensive experiments on resource and accuracy tradeoffs and show strong performance compared to other popular models on ImageNet classification. We then demonstrate the effectiveness of MobileNets across a wide range of applications and use cases including object detection, finegrain classification, face attributes and large scale geo-localization. + ![](https://maoxianxin1996.oss-accelerate.aliyuncs.com/codechina/20210615093836.png) \ No newline at end of file