Transformer architecture and theory

Hi! The question is of a theoretical nature. Maybe someone knows or there is a material where a transformer model is described for dummies, so that one pass could be calculated in Excel. With examples, for example, there are 5 sentences, positional coding takes place here, as a result, the matrix of so-and-so. I am especially interested in how Q, K, V and all this magic appear. Is there a matrix of sentences or individual words at the output of the encoder? And how does the decoder turn the matrix into text?

This is the channel you are looking for Pytorch Modules explained And especially starting with Transformer Encoder Overview