You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
For example the distributed loss wrapper computes a loss for the global batch, but keeps track of only the local gradients. It would be better to compute the loss relevant to only the local batch by making use of indices_tuple.