Attention mechanism

Sparse Activation is the Future of AI (with Adrian Kosowski)

Sparse Activation is the Future of AI (with Adrian Kosowski)

Adrian Kosowski from Pathway explains their groundbreaking research on sparse activation in AI, moving beyond the dense architectures of transformers. Their model, Baby Dragon Hatchling (BDH), mimics the brain's efficiency by activating only a small fraction of its artificial neurons, enabling a new, more scalable, and compositional approach to reasoning that isn't confined by the vector space limitations of current models.

Inside GPT – The Maths Behind the Magic • Alan Smith • GOTO 2024

Inside GPT – The Maths Behind the Magic • Alan Smith • GOTO 2024

A deep dive into the internal workings of Large Language Models like GPT, explaining the journey from a text prompt through tokenization, embeddings, and the attention mechanism to generate a response.