Run LLMs Locally
⚙️Query Compilers
Flag this post
Inside Pinecone: Slab Architecture
🧮Apache Calcite
Flag this post
The Production Generative AI Stack: Architecture and Components
thenewstack.io·14h
📊Data Lineage
Flag this post
Why Multimodal AI Broke the Data Pipeline — And How Daft Is Beating Ray and Spark to Fix It
hackernoon.com·4d
🔢NumPy
Flag this post
Co-Optimizing GPU Architecture And SW To Enhance Edge Inference Performance (NVIDIA)
semiengineering.com·1d
📈Performance Profiling
Flag this post
Challenging the Fastest OSS Workflow Engine
🔄Concurrency
Flag this post
Understanding multi GPU Parallelism paradigms
🔢NumPy
Flag this post
Accelerating AI inferencing with external KV Cache on Managed Lustre
cloud.google.com·6d
💾Cache Optimization
Flag this post
Cutting LLM Batch Inference Time in Half: Dynamic Prefix Bucketing at Scale
💾Cache Optimization
Flag this post
Physics informed machine learning based predictive control for intelligent operation of edge datacenters
sciencedirect.com·5d
🎮Reinforcement Learning
Flag this post
Optimizing Datalog for the GPU
⚡DataFusion
Flag this post
Which Chip Is Best?
📈Performance Profiling
Flag this post
Hydra: Dual Exponentiated Memory for Multivariate Time Series Analysis
arxiv.org·3d
🧊Iceberg Tables
Flag this post
Supercharging the ML and AI Development Experience at Netflix
netflixtechblog.com·2d
🌊Apache Flink
Flag this post
The state of SIMD in Rust in 2025
⚡SIMD Optimization
Flag this post
Reactive Tree Management in Nuxt 4: How I Modeled Complex Hierarchies with Pinia
🧮Apache Calcite
Flag this post
Why CoreWeave’s Object Storage Launch is Good for AI—and Everyone Building It
backblaze.com·6d
🏛️Lakehouse Architecture
Flag this post
Loading...Loading more...