This is a fancy way of saying we’re teaching her to read and understand text, just like how you learned in school (but without all those ***** teachers).
So here’s what happens: first, we give Medusa some data to work with this could be anything from news articles to social media posts. In our case, we’re using the ShareGPT dataset which is a collection of text that people have shared online (hence the name “Share” and “GPT”, which stands for Generative Pre-trained Transformer).
Next, Medusa reads through all this data and tries to figure out what it means. She does this by breaking down each sentence into smaller parts called tokens think of them like building blocks that make up a bigger picture. Once she’s got these tokens, she can start making connections between them and understanding how they relate to one another.
For example, let’s say Medusa comes across the following text: “I went to the store yesterday.” She might break this down into individual tokens like “i”, “went”, “to”, “the”, “store”, “yesterday”. Then she can start making connections between these tokens for instance, she knows that “went” is a verb and it’s usually followed by an object (in this case, the store).
Now let’s say Medusa comes across another text: “The cat sat on the mat.” She might break this down into individual tokens like “the”, “cat”, “sat”, “on”, “the”, “mat”. Then she can start making connections between these tokens for instance, she knows that “sat” is a verb and it’s usually followed by an object (in this case, the mat).
So basically, Medusa learns to read and understand text by breaking it down into smaller parts and figuring out how they relate to one another. And once she’s learned all these connections, she can start using them to generate new text that makes sense just like a human writer would do!