Adaptive threshold possible?

I am trying to threshold Variables using the adaptive value as below.

import torch
from torch.autograd import Variable
import torch.nn as nn

x = Variable(torch.rand(10))
maxx = torch.max(x)
m = nn.Threshold(0.6*maxx,0)
y = m(x)

As you can see, it won’t work because ‘maxx’ is variable not constant.
How can I use the adaptive threshold ?

Thanks in advance for your help.

1 Like


I think you want to use directly the nn functional version of threshold:

import torch
from torch.autograd import Variable
import torch.nn as nn

x = Variable(torch.rand(10))
maxx = torch.max(x)
y = nn.functional.threshold(x, 0.6*[0], 0)

EDIT: access the content of the maxx Variable since we don’t want to propagate through it.

Thanks for your reply.
Unfortunately, your code also won’t work because maxx is Variable.


Yes, my bad.
But given that you will not be able to backpropagate though the threshold value, you don’t want the threshold to be a Variable. You can thus just use the content of the maxx Variable. I updated the code sample above.

Thanks for your reply. It works as I want.

When I use this snippet in my code, I got an error as

RuntimeError: one of the variables needed for gradient computation has been modified by an inplace operation

My test code is as follows.

import torch
import torch.nn as nn
import torch.nn.functional as F
import torch.optim as optim
from torch.autograd import Variable

class AdaReLU(nn.Module):
def __init__(self):
    super(AdaReLU, self).__init__()
    self.relu = nn.ReLU()
def forward(self, x):
    y = x
    for i in range(x.size(0)):
        for ii in range(x.size(1)):
            maxx = torch.max(x[i,ii,0,:])
            y[i,ii,0,:] = F.threshold(x[i,ii,0,:], config.max_th*[0], 0)

    x = self.relu(y)
    return x

class TestNet(nn.Module):
def __init__(self):
    super(TestNet, self).__init__()
    self.conv_1 = nn.Conv2d(len(config.alphabet), config.basic_frames, kernel_size = (1,3), stride = (1,2), padding = (0,0), bias = False)
    self.batchnorm_1 = nn.BatchNorm2d(config.basic_frames)
    self.conv_2 = nn.ConvTranspose2d(config.basic_frames, len(config.alphabet), kernel_size = (1,3), stride = (1,2), padding = (0,0), bias = False)
    self.batchnorm_2 = nn.BatchNorm2d(len(config.alphabet))
    self.adarelu = AdaReLU()

def forward(self, x):
    x = self.adarelu(self.batchnorm_1(self.conv_1(x)))
    x = F.sigmoid(self.batchnorm_2(self.conv_2(x)))
    return x

x = Variable(torch.rand(32,68,1,2283))
model = TestNet()
optimizer = optim.SGD(model.parameters(), lr =0.01, momentum=0.9)
output = model(data)
loss = F.binary_cross_entropy(output, data)

Doing y = x was a good idea to prevent changing the input inplace. Unfortunately, doing it that way leads to x and y begin the same tensors and so you keep modifying the input inplace.
Try to replace it with y = x.clone().

Thanks, it works.
But, due to the for-loops, it is too slow.
Do you have any idea to speed up the process ?


I think you can replace the loop by:

y = x * (1 -

The whole function can be simplified to this:

def AdaReLU(x):                    
    max_val, _ = torch.max(x, 3)   
    mask = x < max_val.expand_as(x)
    x = x.clone()                  
    x[mask] = 0                    
    return F.relu(x)               

The detach_() call shouldn’t be necessary, but I’ve found a bug there :confused:

1 Like

Thanks for your replies.
The process is quite fast now.