6d3ec6c28f32560121d93b8fcab2cf7924f59dc5,thinc/layers/pytorchwrapper.py,,forward,#Any#Any#Any#,27
Before Change
else:
pytorch_model.eval()
x_args, x_kwargs = model.prepare_input(x_data, is_update=False)
with torch.no_grad():
y_var = pytorch_model(*x_args, **x_kwargs)
self._model.train()
return model.prepare_output(y_var)
y = model.prepare_output(y_var)
def backward_pytorch(dy_data):
d_args, d_kwargs = model.prepare_backward_input(dy_data, y_var)
torch.autograd.backward(*d_args, **d_kwargs, retain_graph=True)
After Change
pytorch_model = model.shims[0]
X_torch = xp2torch(X, requires_grad=is_train)
Y_torch, torch_backprop = pytorch_model((X_torch,), {}, is_train)
Y = torch2xp(Y_torch)
def backprop(dY):
In pattern: SUPERPATTERN
Frequency: 3
Non-data size: 6
Instances
Project Name: explosion/thinc
Commit Name: 6d3ec6c28f32560121d93b8fcab2cf7924f59dc5
Time: 2020-01-02
Author: honnibal+gh@gmail.com
File Name: thinc/layers/pytorchwrapper.py
Class Name:
Method Name: forward
Project Name: elbayadm/attn2d
Commit Name: ef17941545c6d742de717d9769b2a412d9924e4e
Time: 2018-06-15
Author: myleott@fb.com
File Name: fairseq/sequence_generator.py
Class Name: SequenceGenerator
Method Name: _decode