I’m going through this tutorial and I’m a bit stuck on the final exercise. I’ve seen lots of solutions that look like this one, where they simply have a single linear layer that maps the embedding from
nn.Embedding to an output layer of
nn.Embedding class get updated by backpropagation? In other words, is it just a lookup table (as the docs imply) or is it actually the first layer of the Neural Net? If it does get updated, how does that work?
If it doesn’t get updated, how does the structure in the solution I listed above generate new embeddings?
EDIT: After some testing, I can see that the values in the
nn.Embedding class do get updated, so I suppose it must be getting updated during backpropagation. How is this happening? Is there a doc anywhere that explains this? I couldn’t figure it out by looking at the