1. 首页
  2. 人工智能
  3. 论文/代码
  4. PareCO:可简化神经网络的帕累托感知通道优化

PareCO:可简化神经网络的帕累托感知通道优化

上传者: 2021-01-22 05:27:38上传 .PDF文件 2.34 MB 热度 25次

可简化的神经网络在预测误差和计算成本(例如浮点运算或FLOP的数量)之间提供了一个灵活的折衷方案,并且存储成本与单个模型相同。最近已针对资源受限的设置(例如移动设备)提出了它们的建议。..

PareCO: Pareto-aware Channel Optimization for Slimmable Neural Networks

Slimmable neural networks provide a flexible trade-off front between prediction error and computational cost (such as the number of floating-point operations or FLOPs) with the same storage cost as a single model. They have been proposed recently for resource-constrained settings such as mobile devices.However, current slimmable neural networks use a single width-multiplier for all the layers to arrive at sub-networks with different performance profiles, which neglects that different layers affect the network's prediction accuracy differently and have different FLOP requirements. Hence, developing a principled approach for deciding width-multipliers across different layers could potentially improve the performance of slimmable networks. To allow for heterogeneous width-multipliers across different layers, we formulate the problem of optimizing slimmable networks from a multi-objective optimization lens, which leads to a novel algorithm for optimizing both the shared weights and the width-multipliers for the sub-networks. We perform extensive empirical analysis with 14 network and dataset combinations and find that less over-parameterized networks benefit more from a joint channel and weight optimization than extremely over-parameterized networks. Quantitatively, improvements up to 1.7% and 1% in top-1 accuracy on the ImageNet dataset can be attained for MobileNetV2 and MobileNetV3, respectively. Our results highlight the potential of optimizing the channel counts for different layers jointly with the weights for slimmable networks.

下载地址
用户评论