unilm
unilm copied to clipboard
Update L. 116 of backbone.py
When fine-tuning the network for our own problem, we encountered the error RuntimeError: one of the variables needed for gradient computation has been modified by an inplace operation: [torch.cuda.FloatTensor [32, 768, 248]].
This seemed to be triggered by the in place operation occuring L. 116. To solve this we replaced x += x_conv by x = x.clone() + x_conv