I’ve tried:
        optim_parameters = {
            'lr': params.learning_rate, 
        }
        optimizer = optim.Adam([input], optim_parameters) 
Instead of:
        optimizer = optim.Adam([input], lr = learning_rate) 
But I get this error:
    optimizer.step(closure)
  File "/usr/local/lib/python2.7/dist-packages/torch/optim/adam.py", line 104, in step
    step_size = group['lr'] * math.sqrt(bias_correction2) / bias_correction1
TypeError: unsupported operand type(s) for *: 'dict' and 'float'