-
Notifications
You must be signed in to change notification settings - Fork 14
Open
Description
I believe the following chunk that appears throughout src/yoke/torch_training_utils.py:
optimizer.zero_grad(set_to_none=True) # Possible speed-up
loss.mean().backward()
optimizer.step()should instead be:
loss.mean().backward()
optimizer.step()
optimizer.zero_grad(set_to_none=True) # Possible speed-upas is, I think there is gradient leakage across epochs and higher than necessary usage overall:
Reactions are currently unavailable
Metadata
Metadata
Assignees
Labels
No labels

