GoMem is a high-performance memory allocator library for Go
github.com·21h
🧠Memory Allocators
OBCache: Optimal Brain KV Cache Pruning for Efficient Long-Context LLM Inference
arxiv.org·19h
🧠LLM Inference
How Do SSDs Work?
extremetech.com·11h·
Discuss: Hacker News
⚙️Mechanical Sympathy
Looking at my Arduino
boswell.bearblog.dev·7h
🖥️Hardware Architecture
Can AI Co-Design Distributed Systems? Scaling from 1 GPU to 1k
harvard-edge.github.io·1h·
Discuss: Hacker News
🌐Distributed systems
A new method to build more energy-efficient memory devices could lead to a sustainable data future
phys.org·14h
🏭TSMC
Explicit Lossless Vertex Expanders!
gilkalai.wordpress.com·13h
🔬RaBitQ
Benchmarking LLM Inference on RTX 4090 / RTX 5090 / RTX PRO 6000 #2
reddit.com·5h·
Discuss: r/LocalLLaMA
🏗️LLM Infrastructure
Framework for Optimizing Reliability and Thermal Management of 3DICs (National Taiwan Univ., Lamar Univ.)
semiengineering.com·6h
🔬Chip Fabrication
A Manifesto for the Programming Desperado
github.com·7h·
Discuss: Hacker News
🪄Prompt Engineering
Iterated Development and Study of Schemers (IDSS)
lesswrong.com·9h
🆕New AI
QUIC! Jump to User Space!
hackaday.com·7h
QUIC Protocol
How View Caching in Rails Works (2020)
honeybadger.io·9h·
Discuss: Hacker News
💾Prompt Caching
Profiling Your Code: 5 Tips to Significantly Boost Performance
usenix.org·21h
Systems Performance
Multi-Core By Default
rfleury.com·22h·
🧵Concurrency
MECE — The AI Principle You’ll Never Stop Using After Reading This
pub.towardsai.net·12h
🔍AI Interpretability
Intel Bets Recovery on Panther Lake AI Chip as Foundry Bleeds Billions - Decrypt
news.google.com·23h
💻Chips
Supercharge your Enterprise BI: How to approach your migration to AI/BI
databricks.com·2h
🏗️Infrastructure Economics
MultiPar 1.3.3.5 Beta / 1.3.2.9
majorgeeks.com·15h
📄File Formats
Writing regex is pure joy. You can't convince me otherwise.
triangulatedexistence.mataroa.blog·21h·
📑Inverted Indexes