while I like the drawings, their combination with the purposefully vague and mysterious/mystical terminology like 'embeddings' and 'attention' make the entire document feel like one of those pamphlets that cults or 'science-based' religions produce
- The model is said to be rewarded/penalized based on answers, and the plain meaning would leave someone with the impression of "answer to the question asked/prompt", the visual appears correct (the chalkboard indicates the question is "fill in the next token" and the answer is "the next most likely token")
- You end up with a strong sense there's 3 tables in the model called key, query, and value that carry the weight of the world. Are there only 3?
- "what part of data are we processing and are they relevant?" - what is they? what part of data is it processing? what is data in this context?
- "how well does this data answer my question?" - so the model is picking out answers from the training data and checking if it answers the prompt? This creates a strong sense of copying verbatim from training data
- "how should we improve the contextual representation of the input data?" - "contextual representation of the input data" isn't clear here, in my dummy brain, it's "here's information we can use to decide the next token: in this context, they meant cat cli, not cat the animal"
- "training a model is like doing Q&A with the neutral networks while it attends to the right data": Is it Q&A? If we double down on that (which means also doubling down on the idea being finding the answer in the training data, I don't think it's a good idea), then there's a gap in how the sentence connects within itself that's worth addressing. I assume "like doing Q&A with the neutral networks, and based on whether the answer is correct or not, it'll adjust and learn to attend to the right data"
hey! thanks so much for the feedback. I'd actually love to keep updating / iterating these cartoons so they are more approachable. If you have time, I'd love to hear more on which pages are confusing & how I could have explained it better!
I _tried_ to give a definition to embeddings on page 11, but maybe that's not the most intuitive? Lmk! feel free to DM
The use of the word "relevant" such as in the phrase "are the most relevant" is problematic.
Relevancy is dependent on and relative to a specific task or interest. "Related to" or "associated" might be a better choice, since parts of a text can be statistically associated with each other.
While I feel that you are accurately conveying the terminology in the field I personally feel that terminology overstates things. For example, the notion that words closer in meaning are generally closer in latent space could be more accurately stated as "words closer in latent space are often used in similar ways."
I think this comes down to meaning being largely determined by a person's individual interpretation at a given point in time, similar to the arbitrariness of relevancy.
But you should decrease the amount of the text. Because it's hard to watch and hard to read. Also, text fonts should be changed. Look at some cool design examples.
Before you show info via infographics or animations or anything else you have to watch a lot of cool examples and then find one to follow.