LLMs on a Shoestring: The Dynamic Cache Advantage by Arvind Sundararajan
dev.to·21h·
Discuss: DEV
💾Cache Algorithms
StringWa.rs on GPUs: Databases & Bioinformatics 🦠
ashvardanian.com·12h·
🚀Tokenizer Performance
Analog IMC Attention Mechanism For Fast And Energy-Efficient LLMs (FZJ, RWTH Aachen)
semiengineering.com·15h
🗺️Region Inference
Topological Sort: Managing Mutable Structures in Haskell
mmhaskell.com·22h
🪢Rope Data Structures
Rowhammer: TRR on DDR5 DRAM has been broken
comsec.ethz.ch·14h·
🏷️Memory Tagging
Is Recursion in LLMs a Path to Efficiency and Quality?
pub.towardsai.net·7h
🪜Recursive Descent
AQUA: Attention via QUery mAgnitudes for Memory and Compute Efficient Inference in LLMs
arxiv.org·3h
🔍ML Language
Building High-Performance Caching in Go: A Practical Guide
dev.to·1d·
Discuss: DEV
🧠Memory Models
Power Query Secret Tip to Lightning-Fast Approximate Matches
geeky-gadgets.com·18h
📊Query Optimizers
Semantic Dictionary Encoding
falvotech.com·16h·
Discuss: Hacker News
🗂️Type Indexing
Safepoints and Fil-C
fil-c.org·2h·
Discuss: Hacker News
🎯Ring Buffers
Rendezvous Hashing Explained (2020)
randorithms.com·11h·
🔗Hash Algorithms
More hardware won’t fix bad engineering
infoworld.com·22h
🔮Branch Predictors
Model Kombat by HackerRank
producthunt.com·3h
🏁Language Benchmarks
Fastest copy
forums.anandtech.com·15h
Copy Elision
What is Algebraic about Algebraic Effects?
interjectedfuture.com·15h
💫Effect Systems
Conquering the LLM Memory Wall: How to Run 2–4x Longer Contexts with a Single Line of Code
reddit.com·19h·
Discuss: r/LocalLLaMA
🗺️Region Inference
Google releases VaultGemma, its first privacy-preserving LLM
arstechnica.com·10h·
Discuss: Hacker News
🎲Parser Fuzzing
What Facebook's Memcache Taught Me About Systems Thinking
lorbic.com·14h·
Discuss: Hacker News
Cache-Aware Algorithms
Explaining the LMAX Disruptor
lmax-exchange.github.io·6d·
Discuss: DEV
🎯Ring Buffers