Generation at the Speed of Thought: Speculative Decoding
🔵LLM frameworks and AI libraries for TypeScript
Flag this post
TIL: For long-lived LLM sessions, swapping KV Cache to RAM is ~10x faster than recalculating it. Why isn't this a standard feature?
🦙Simple finetuning LLMs
Flag this post
Cycle-accurate 6502 emulator as coroutine in Rust
🔥Svelte
Flag this post
Opportunistically Parallel Lambda Calculus
🔵LLM frameworks and AI libraries for TypeScript
Flag this post
No Cap, This Memory Slaps: Breaking Through the OLTP Memory Wall
🦙Simple finetuning LLMs
Flag this post
How fast can an LLM go?
📊Vector Databases
Flag this post
DGX Spark UMA can trick you
🔥Svelte
Flag this post
Linux/WASM
🧩WASI
Flag this post
From Lossy to Lossless Reasoning
🤖Coding Automation
Flag this post
When Five Dumb AIs Beat One Smart AI: The Case for Multi-Agent Systems
🤖Coding Automation
Flag this post
Learning to program "recycles" preexisting F-P pop codes of logical algorithms
🔄AI Pipeline design and techniques
Flag this post
I built a lightweight HTTP bridge for AnythingLLM to safely run multiple local MCPs inside Docker (Dummy + Time demo included)
🧩WASI
Flag this post
Loading...Loading more...