Hello,
I have a problem with the model size. The description of my model size is as follows:
| Name | Type | Params
----------------------------------------
0 | model_ | Model | 44.5 M
*1 | loss_function | Loss | 0 *
----------------------------------------
44.5 M Trainable params
0 Non-trainable params
44.5 M Total params
178.099 Total estimated model params size (MB)
So I get this Runtime Error:
RuntimeError: CUDA out of memory. Tried to allocate 2.00 GiB
I am curious why Total estimated model params size is so much bigger than Trainable params. And what is the difference?
How can I reduce the Total params size to avoid this error?
I appreciate any help.