Pytorch doesen't use GPU

PyTorch does not using GPU on full power.
model and tensors are in GPU and PyTorch use GPU

    print('The train is continue!')
    print(torch.cuda.get_device_name(0))
    print('Memory Usage:')
    print('Allocated:', round(torch.cuda.memory_allocated(0) / 1024 ** 2, 1), 'MB')
    print('Cached:   ', round(torch.cuda.memory_reserved(0) / 1024 ** 2, 1), 'MB')
    double_barrier_net = double_barrier_net.double().to(get_device())

This is the console logs

cuda:0
The train is continue!
NVIDIA GeForce RTX 2070 Super
Memory Usage:
Allocated: 14.5 MB
Cached: 22.0 MB

but training is going too slow and on my time_waste_test we could see that train on GPU even slowly than on CPU.
link to logs
https://drive.google.com/file/d/1RxSd2oLEsGytjNeudIs0hUUSIbV_sX8u/view?usp=sharing

I have no idea how to fix this