Multi-Core By Default
rfleury.com·22h·
🧵Concurrency
GCC Patches Posted For C++26 SIMD Support
phoronix.com·12h
SIMD
Benchmarking LLM Inference on RTX 4090 / RTX 5090 / RTX PRO 6000 #2
reddit.com·5h·
Discuss: r/LocalLLaMA
🏗️LLM Infrastructure
OBCache: Optimal Brain KV Cache Pruning for Efficient Long-Context LLM Inference
arxiv.org·19h
🧠LLM Inference
Profiling Your Code: 5 Tips to Significantly Boost Performance
usenix.org·21h
Systems Performance
Can AI Co-Design Distributed Systems? Scaling from 1 GPU to 1k
harvard-edge.github.io·1h·
Discuss: Hacker News
🌐Distributed systems
(Forward) automatic implicit differentiation in Rust with num-dual 0.12.0
reddit.com·8h·
Discuss: r/rust
🎭Rust Macros
2025-10-10 # LLMs Are Transpilers
alloc.dev·23h·
Discuss: Hacker News
🏆LLM Benchmarking
GoMem is a high-performance memory allocator library for Go
github.com·21h
🧠Memory Allocators
Building the Reasoning Engine at Axiom
axiommath.ai·3h·
Discuss: Hacker News
💻Programming languages
MECE — The AI Principle You’ll Never Stop Using After Reading This
pub.towardsai.net·12h
🔍AI Interpretability
Explicit Lossless Vertex Expanders!
gilkalai.wordpress.com·13h
🔬RaBitQ
Let's Write a Macro in Rust
hackeryarn.com·7h·
Discuss: Hacker News
🎭Rust Macros
Looking at my Arduino
boswell.bearblog.dev·7h
🖥️Hardware Architecture
Supercharge your Enterprise BI: How to approach your migration to AI/BI
databricks.com·2h
🏗️Infrastructure Economics
Parallelizing Cellular Automata with WebGPU Compute Shaders
vectrx.substack.com·14h·
Discuss: Substack
🏟️Arena Allocators
Evaluating Gemini 2.5 Deep Think's math capabilities
epoch.ai·10h·
Discuss: Hacker News
🏆LLM Benchmarking