Benchmarking LLM Inference on RTX 4090 / RTX 5090 / RTX PRO 6000 #2
reddit.com·10h·
Discuss: r/LocalLLaMA
🏗️LLM Infrastructure
Let's Write a Macro in Rust
hackeryarn.com·12h·
Discuss: Hacker News
🎭Rust Macros
Building and Deploying a RAG Application: From PDF Processing to Production
pub.towardsai.net·4h
🔄LLM RAG Pipelines
PostGIS Performance: Indexing and EXPLAIN
crunchydata.com·14h
🔍Query Optimization
Can AI Co-Design Distributed Systems? Scaling from 1 GPU to 1k
harvard-edge.github.io·6h·
Discuss: Hacker News
🌐Distributed systems
(Forward) automatic implicit differentiation in Rust with num-dual 0.12.0
reddit.com·13h·
Discuss: r/rust
🎭Rust Macros
Explicit Lossless Vertex Expanders!
gilkalai.wordpress.com·18h
🔬RaBitQ
GCC Patches Posted For C++26 SIMD Support
phoronix.com·18h
SIMD
How I Built My Own Tool for Disk Space Cleanup
debamitro.github.io·7h
🔬Rust Profiling
VLLM Predicted Outputs
cascadetech.ai·8h·
Discuss: Hacker News
🏗️LLM Infrastructure
Assuring Agent Safety Evaluations By Analysing Transcripts
lesswrong.com·18h
🏆LLM Benchmarking
Scaling Time-Series Data for AI Models
singlestore.com·13h
🎛️Feed Filtering
🎲 Intel Pentium II introduced May 7, 1997
dfarq.homeip.net·22h
🖥️Hardware Architecture
Parallelizing Cellular Automata with WebGPU Compute Shaders
vectrx.substack.com·19h·
Discuss: Substack
🏟️Arena Allocators
QUIC! Jump to User Space!
hackaday.com·13h
QUIC Protocol
MultiPar 1.3.3.5 Beta / 1.3.2.9
majorgeeks.com·21h
📄File Formats