What is the linear embedding layer in Swin Transformer?

my goal is to understand what is Linear Embedding layer actually is in Swin Transformer

The problem is the Swin Transformer’s paper do not explain what is the linear embedding layer.

arXiv:

  1. arxiv org/pdf/2103.14030

    A linear embedding layer is applied on this raw-valued feature to project it to an arbitrary dimension (denoted as C).

    This is the only explanation given for linear embedding in the paper.

PyTorch documentation:

  1. Linear pytorch org/docs/stable/generated/torch.nn.Linear.html
  2. Embedding pytorch org/docs/stable/generated/torch.nn.Embedding.html

YouTube:

  1. Embedding youtube com/watch?v=aWFllV6WsAs

StackOverflow:

  1. The difference between Linear and Embedding in PyTorch stackoverflow com/questions/65445174/what-is-the-difference-between-an-embedding-layer-with-a-bias-immediately-afterw

Honestly, I do not know where to start

From the PyTorch implementation:

The Conv2D should be responsible for both the patch splitting and then the embedding