my goal is to understand what is Linear Embedding layer actually is in Swin Transformer
The problem is the Swin Transformer’s paper do not explain what is the linear embedding layer.
arXiv:
-
arxiv org/pdf/2103.14030
A linear embedding layer is applied on this raw-valued feature to project it to an arbitrary dimension (denoted as C).
This is the only explanation given for linear embedding in the paper.
PyTorch documentation:
- Linear pytorch org/docs/stable/generated/torch.nn.Linear.html
- Embedding pytorch org/docs/stable/generated/torch.nn.Embedding.html
YouTube:
- Embedding youtube com/watch?v=aWFllV6WsAs
StackOverflow:
- The difference between Linear and Embedding in PyTorch stackoverflow com/questions/65445174/what-is-the-difference-between-an-embedding-layer-with-a-bias-immediately-afterw
Honestly, I do not know where to start