Benchmarking LLM Inference on RTX 4090 / RTX 5090 / RTX PRO 6000 #2
reddit.com·4h·
Discuss: r/LocalLLaMA
🏗️LLM Infrastructure
Profiling Your Code: 5 Tips to Significantly Boost Performance
usenix.org·20h
🔬Rust Profiling
Multi-Core By Default
rfleury.com·20h·
🧵Concurrency
Repos with 3,200+ refs: 5s → <0.1s (100x faster)
gitkraken.com·22h·
Discuss: r/programming
📂Git
Supercharge your Enterprise BI: How to approach your migration to AI/BI
databricks.com·1h
🏗️Infrastructure Economics
GoMem is a high-performance memory allocator library for Go
github.com·19h
🧠Memory Allocators
A sprinkle of Rust - bind, don't rewrite, in the age of MCP servers
medium.com·23h·
Discuss: r/rust
🦀Rust
Assuring Agent Safety Evaluations By Analysing Transcripts
lesswrong.com·12h
🏆LLM Benchmarking
OBCache: Optimal Brain KV Cache Pruning for Efficient Long-Context LLM Inference
arxiv.org·18h
🧠LLM Inference
InferenceMAX – open-source Inference Frequent Benchmarking
github.com·2h·
Discuss: Hacker News
🏗️LLM Infrastructure
N8n vs. Windmill vs. Temporal
blog.arcbjorn.com·22h·
Discuss: Hacker News
🚀Async Optimization
Progress being made in porting AMD OpenSIL Turin PoC to Coreboot in a Gigabyte MZ33-AR1
blog.3mdeb.com·2h·
🖥GPUs
Trusted Execution Environments? More Like "Trust Us, Bro" Environments
libroot.org·4h·
Discuss: Hacker News
🔐Hardware Security
Debugging Humidity: Lessons from deploying software in the physical world
physical-ai.ghost.io·2h·
Discuss: Hacker News
🌐Distributed systems
Explicit Lossless Vertex Expanders!
gilkalai.wordpress.com·12h
🔬RaBitQ
Is the End of Detection-Based Security Here?
thenewstack.io·6h
💾Persistence Strategies
How we built a structured Streamlit Application Framework in Snowflake
about.gitlab.com·22h
🔧Developer tools
You don't avoid the chaos. You filter it.
threadreaderapp.com·5h
🧹Spam Filters