Ever wondered how a transformer (LLM) works? How input text is tokenised, embedded, normalised and passes through multi-head self-attention layers? Well, someone decided to make it into an interactive visualisation. Wow.
Ever wondered how a transformer (LLM) works? How input text is tokenised, embedded, normalised and passes through multi-head self-attention layers? Well, someone decided to make it into an interactive visualisation. Wow.