liminal 🦠on Nostr: Think of how you 'hold a memory', no one else can interact with it unless you talk ...
Think of how you 'hold a memory', no one else can interact with it unless you talk about it or draw it, a memory or idea needs to be conveyed somehow to the outside world. An encoding models that produce embeddings are basically half an LLM, it takes in the words/tokens and says "this string is located here", it assigns a coordinate, an address for it. That address is the context, anything with addresses nearby are more related in ideas. The second part of the LLM is the decoder, where it takes the address as a kind of starting point. The decoder uses the context of that coordinate and responds with words that are also in the right context (which is learned by training).
H.T. to TheGuySwann (npub1h8n…rpev) for his fantastic read of "A gentle introduction to Large Large Language Models"
https://fountain.fm/episode/yCpvsos8iUfXsfLeUPon
https://mark-riedl.medium.com/a-very-gentle-introduction-to-large-language-models-without-the-hype-5f67941fa59e
H.T. to TheGuySwann (npub1h8n…rpev) for his fantastic read of "A gentle introduction to Large Large Language Models"
https://fountain.fm/episode/yCpvsos8iUfXsfLeUPon
https://mark-riedl.medium.com/a-very-gentle-introduction-to-large-language-models-without-the-hype-5f67941fa59e