wistone

Results 4 comments of wistone

The experiments with ResNet101 are trained with 4 GPUs. Actually our training code contains many memory optimization so it needs less memory.

https://github.com/dmlc/mxnet-memonger You can refer to the memory optimization in MXNet

We implement this method in our training platform in caffe.

No it is not public. You can read the paper for the theory, and refer the code in MXNet. It is doable.