multi-task-learning-example
multi-task-learning-example copied to clipboard
Log var can become negative and explode
The loss function can optimize in a way that keep decreasing the log_var values, which I observe in my experiments. One simple solution is to do torch.abs(log_var). Any thoughts on how this might affect the formulation of the deductions?
I have encountered the same situation during my training, do you have any better solutions?