bafeed46fb76fa337a771ebb41d65bb95039565a,fairseq/trainer.py,Trainer,train_step,#Trainer#Any#Any#Any#,261
Before Change
// printed out if another exception happens.
// NB(jerry): added a flush to mitigate this
print(msg, file=sys.stderr)
if torch.cuda.is_available() and hasattr(torch.cuda, "memory_summary"):
for device_idx in range(torch.cuda.device_count()):
print(torch.cuda.memory_summary(device=device_idx),
file=sys.stderr)
sys.stderr.flush()
if raise_oom:
raise ValueError(msg)
After Change
if "out of memory" in str(e):
self._log_oom(e)
print("| ERROR: OOM during optimization, irrecoverable")
raise e
if self.args.fp16:
self.meters["loss_scale"].reset()
self.meters["loss_scale"].update(self.optimizer.scaler.loss_scale)
In pattern: SUPERPATTERN
Frequency: 3
Non-data size: 4
Instances Project Name: elbayadm/attn2d
Commit Name: bafeed46fb76fa337a771ebb41d65bb95039565a
Time: 2019-11-06
Author: noreplyspamblackhole@gmail.com
File Name: fairseq/trainer.py
Class Name: Trainer
Method Name: train_step
Project Name: IBM/adversarial-robustness-toolbox
Commit Name: 3bf391c2315bcec722961a2f4a093d1c516dbb10
Time: 2018-08-29
Author: Maria-Irina.Nicolae@ibm.com
File Name: art/classifiers/pytorch.py
Class Name: PyTorchClassifier
Method Name: __init__