This repository has been archived by the owner on Aug 5, 2022. It is now read-only.
Caffe_v1.0.7
- Features
- [Multi-node] Support weight gradient compression for better scaling efficiency of models with large FC layers, like VGG
- [Multi-node] Integrate LARS (Layer-wise Adaptive Rate Scaling) and apply to Alexnet with BatchNorm layers on 32K global batch size
- Enable pinning internal thread to cores for more stable training performance, e.g., data loader thread
- Merge pull request of supporting Flow LRCN from Github
- Support label smoothing regularization (idea from Inception-V3)
- Bug fixings
- [Multi-node] Fix learning rate message and start first iteration from zero on multi-node to be consistent with single node
- Bug fixes on single node
- Misc
- Upgrade MKLML to 2018.0.1.20171007 and MLSL to V2
- Enhance installation and benchmarking scripts
- Update the optimized models