1. 首页
  2. 人工智能
  3. 论文/代码
  4. 动态ReLU

动态ReLU

上传者: 2021-01-22 15:13:43上传 .PDF文件 1.24 MB 热度 16次

整流线性单位(ReLU)通常在深度神经网络中使用。到目前为止,ReLU及其概括(非参数或参数)都是静态的,对所有输入样本的性能相同。..

Dynamic ReLU

Rectified linear units (ReLU) are commonly used in deep neural networks. So far ReLU and its generalizations (non-parametric or parametric) are static, performing identically for all input samples.In this paper, we propose dynamic ReLU (DY-ReLU), a dynamic rectifier of which parameters are generated by a hyper function over all in-put elements. The key insight is that DY-ReLU encodes the global context into the hyper function, and adapts the piecewise linear activation function accordingly. Compared to its static counterpart, DY-ReLU has negligible extra computational cost, but significantly more representation capability, especially for light-weight neural networks. By simply using DY-ReLU for MobileNetV2, the top-1 accuracy on ImageNet classification is boosted from 72.0% to 76.2% with only 5% additional FLOPs.

用户评论