UM-MAE
UM-MAE copied to clipboard
why can model see mask token when training?
In the MAE paper, kaiming He said, if Vit-mae sees mask token when training, and can't see mask token when testing, the inconsistency exists, which results a bad accuracy. But in your paper, you send mask token to encoder, is it right?