Linklog
A curated collection of links and resources I have found over time.
March 2025
- A Visual Guide to LLM Agents (newsletter.maartengrootendorst.com)
It is possibly one of the best summaries out there of how LLMs function, broken down by high-level components (memory, tools), and well illustrated.
- Understanding Attention in LLMs (bartoszmilewski.com)
This is a good example that even if you understand the math behind a concept, there's nothing like good storytelling. I knew how attention worked, but this post brillantly summarized it and clarified some steps for me. A great read!
February 2025
- Deep dive into LLMs like ChatGPT by Andrej Karpathy (TL;DR) (anfalmushtaq.com)
A TL;DR version of Andrej Karpathy's "Deep dive into LLMs like ChatGPT" video. Manages to keep the essentials but presents them in digestible clear chunks.
- Binary vector embeddings are so cool (emschwartz.me)
A description of the effect of binary quantization on embeddings. By restricting the dtype of embedding vectors, you can get a tradeoff between accuracy in latent space and size of the embedding. Using binary dtype seems to conserve a surprisingly high amount of the original information content (about 97%) while yielding a gigantic amount of saving in space (about 97% too here).
- How I program with LLMs (crawshaw.io)
Some interesting reflections on how to use LLMs in daily development work. I personally adhere mostly to the "autocomplete" part with Github Copilot, and I'm getting used to the "search" part where the LLM helps me find information on some language or coding paradigm faster than I can search it. I'm not yet onboard with "Chat-driven programming".
- How to fine-tune open LLMs in 2025 with Hugging Face (www.philschmid.de)
An in-depth example of how to fine-tune an LLM using the Hugging Face ecosystem.
- LoRA (jaketae.github.io)
Explanation of LoRA methods for LLMs.
January 2025
- Building effective agents (www.anthropic.com)
Advice on agentic workflow for practical applications from Anthropic. A good read to better understand what structure you should use when establishing your project.
October 2024
- Transformers From Scratch (blog.matdmiller.com)
Thorough explanation of the Transformers model. If like me you've been confused about what's so special about transformers compared to RNNs or LSTMs, this might help.
September 2024
- dleemiller/WordLlama (github.com)
Natural language processing toolkit optimized for CPU hardware. I haven't tested it yet but it looks really useful for quick clustering, deduplication, similarity search, etc...