AdaBatch: Efficient Gradient Aggregation.pdf
Westudyanewaggregationoperatorforgradientscomingfromamini-batchforstochastic
gradient(SG)methodsthatallowsasignificantspeed-upinthecaseofsparseoptimization
problems.WecallthismethodAdaBatchanditonlyrequiresafewlinesofcodechangecom-
paredtoregularmini-b
下载地址
用户评论