e7a02b6293100e21ef15870fd0f9069beaae5290,train.py,,,#,21

Before Change


        batch_size = data.size(0)
        running_batch_sizes += batch_size
        real_label = Variable(torch.ones(batch_size))
        fake_label = Variable(torch.zeros(batch_size))

        ////////////////////////////////////////////////////////
        // (1) Update D network: maximize log(D(x)) + log(1 - D(G(z)))
        //////////////////////////////////////////////////////
        real_img = Variable(target)
        if torch.cuda.is_available():
            real_img = real_img.cuda()
            real_label = real_label.cuda()

        // compute loss of real_img
        real_out = netD(real_img)
        d_loss_real = discriminator_criterion(real_out, real_label)
        real_scores = real_out.data.sum()
        running_real_scores += real_scores

        // compute loss of fake_img
        z = Variable(data)
        if torch.cuda.is_available():
            z = z.cuda()
            fake_label = fake_label.cuda()
        fake_img = netG(z)
        fake_out = netD(fake_img)
        d_loss_fake = discriminator_criterion(fake_out, fake_label)
        fake_scores = fake_out.data.sum()

        d_loss = d_loss_real + d_loss_fake

        // bp and optimize
        optimizerD.zero_grad()
        d_loss.backward(retain_graph=True)
        optimizerD.step()

        ////////////////////////////////////////////////////////
        // (2) Update G network: maximize log(D(G(z)))
        //////////////////////////////////////////////////////
        index = 1
        while ((fabs((real_scores - fake_scores) / batch_size) > G_THRESHOLD) or g_update_first) and (
                    index <= G_STOP_THRESHOLD):
            // compute loss of fake_img
            g_loss = generator_criterion(fake_out, real_label, fake_img, real_img)
            // bp and optimize
            optimizerG.zero_grad()
            g_loss.backward()
            optimizerG.step()
            fake_img = netG(z)
            fake_out = netD(fake_img)
            fake_scores = fake_out.data.sum()
            g_update_first = False
            index += 1

        g_loss = generator_criterion(fake_out, real_label, fake_img, real_img)
        running_g_loss += g_loss.data[0] * batch_size
        d_loss_fake = discriminator_criterion(fake_out, fake_label)
        d_loss = d_loss_real + d_loss_fake
        running_d_loss += d_loss.data[0] * batch_size
        running_fake_scores += fake_scores

        train_bar.set_description(desc="[%d/%d] Loss_D: %.4f Loss_G: %.4f D(x): %.4f D(G(z)): %.4f"

After Change



        g_loss = generator_criterion(fake_out, fake_img, real_img)
        running_g_loss += g_loss.data[0] * batch_size
        d_loss = - torch.mean(torch.log(real_out) + torch.log(1 - fake_out))
        running_d_loss += d_loss.data[0] * batch_size
        running_fake_scores += fake_scores
Italian Trulli
In pattern: SUPERPATTERN

Frequency: 3

Non-data size: 7

Instances


Project Name: leftthomas/SRGAN
Commit Name: e7a02b6293100e21ef15870fd0f9069beaae5290
Time: 2017-12-02
Author: leftthomas@qq.com
File Name: train.py
Class Name:
Method Name:


Project Name: rusty1s/pytorch_geometric
Commit Name: 45a03ac258234ba49c0a43d46ae022493190591e
Time: 2019-03-26
Author: matthias.fey@tu-dortmund.de
File Name: torch_geometric/nn/models/autoencoder.py
Class Name: ARGA
Method Name: discriminator_loss


Project Name: pymc-devs/pymc3
Commit Name: e4bdbe76308dbeff50741415ab8c080ef99a7d70
Time: 2017-03-22
Author: adrian.seyboldt@gmail.com
File Name: pymc3/distributions/multivariate.py
Class Name: LKJCholeskyCov
Method Name: logp