Submitted by jayalammar t3_10dfex7 in MachineLearning
blimpyway t1_j4pi713 wrote
Reply to comment by chaosmosis in [D] The Illustrated Stable Diffusion (Video) by jayalammar
The order of the words/tokens is normally encoded via positional embeddings that are added each to their respective token embedding. See e.g. https://machinelearningmastery.com/a-gentle-introduction-to-positional-encoding-in-transformer-models-part-1/
Viewing a single comment thread. View all comments