Newbie question here. In my research I found this thread on word embedding dimensions stating:
You can pick a power of 32 (64, 128, 256) to speed up modeling training depending on your software framework and hardware.
Is that also a property in PyTorch and is using a power of 32 increasing speed?