Skip to content

I created a matrix-driven walkthrough of how attention weights are derived. Basically recreated what Transformers do but on Excel..

Notifications You must be signed in to change notification settings

sayeesx/attention

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

3 Commits
 
 
 
 

Repository files navigation

Rather than skimming yet another abstract explanation or relying on an AI tool to summarize it for me, I went hands-on: rebuilt self attention from scratch using just Excel, math, and a stubborn need to actually understand Attention

From tokenizing a simple sentence, to manually crafting the query, key, and value matrices, to computing attention weights and watching how focus shifts across words it’s a deceptively elegant process once you sit with it long enough.

It’s easy to take models like Transformers for granted when they're wrapped up in pre-trained APIs and high-level libraries. But peeling back the abstraction layer reminded me: these systems aren’t magic they’re clever math, stacked purposefully.

About

I created a matrix-driven walkthrough of how attention weights are derived. Basically recreated what Transformers do but on Excel..

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published