š Paper Info
- Title: MobileNetV2: Inverted Residuals and Linear Bottlenecks
- Authors: Mark Sandler, Andrew Howard, Menglong Zhu, Andrey Zhmoginov, Liang-Chieh Chen
- Link: arXiv 1801.04381
- Published: 2018 (Google Research)
- Code: Available in TensorFlow-Slim
š§ Day 3 Review ā Experiments, Applications, and Conclusions
ā
Step 1: Architecture Expansion
The final MobileNetV2 network consists of:
Initial 3Ć3 Conv
ā Repeating Bottleneck Blocks (19 times)
ā Final 1Ć1 Conv + Global Average Pooling + FC layer
- Each block uses an expansion factor t = 6
- All convolutions are 3Ć3 or 1Ć1
- BatchNorm & Dropout applied where appropriate
- ReLU6 for activation (except final linear projection)
ā
Step 2: Experimental Results Summary
š ImageNet Classification
Model |
Top-1 Acc |
MACs |
Params |
MobileNetV1 |
70.6% |
575M |
4.2M |
MobileNetV2 |
71.8% |
300M |
3.4M |
ā V2 achieves better accuracy with nearly half the computation.
š Object Detection (COCO, with SSDLite)
Model |
mAP |
Latency |
MobileNetV1 + SSD |
19.3 |
27 ms |
MobileNetV2 + SSDLite |
22.1 |
19 ms |
ā V2 provides higher mAP and faster inference.
š Memory Efficiency
In Table 2 (Fig. 2 in the paper), MobileNetV2 shows peak memory usage < 400K during inference.
This is lower than ResNet-50, VGG, Inception, and other baselines.
ā
Step 3: Applications
- Object Detection: Used in real-time detectors like SSDLite
- Semantic Segmentation: Combined with DeepLabv3 for mobile segmentation
- Mobile Transfer Learning: Widely used for fine-tuning on edge devices
ā
Key Insights (3-Line Summary)
- MobileNetV2 achieves strong accuracy with low memory and compute requirements.
- The architecture outperforms V1 and competes with heavier models in speed and accuracy.
- It is ideal for on-device inference tasks such as detection and segmentation.
š New Terms
- MACs (Multiply-Accumulate Ops): A proxy for computation cost
- mAP (mean Average Precision): Detection accuracy averaged across IoU thresholds
- Materialized Memory: Memory required to hold intermediate activations during inference
š GitHub Repository
Visual summary + experimental table:
š github.com/hojjang98/Paper-Review
š Reflections
The experiments confirm that MobileNetV2 is not just lightweight in theory, but in practice.
Its memory efficiency and speed make it one of the most impactful mobile architectures of its time.
Iām especially impressed with how well it balances performance and hardware constraints.