Opportunistically Parallel Lambda Calculus
💡LSP
Flag this post
AI efficiency advances with spintronic memory chip that combines storage and processing
techxplore.com·3d
⚡Flash Attention
Flag this post
Our newest model: Chandra (OCR)
🏎️TensorRT
Flag this post
How fast can an LLM go?
🏎️TensorRT
Flag this post
Accelerating AI inferencing with external KV Cache on Managed Lustre
cloud.google.com·1d
⚡ONNX Runtime
Flag this post
Machine Learning Fundamentals: Everything I Wish I Knew When I Started
🎓Model Distillation
Flag this post
M5 iPad Pro (Late 2025)
lowendmac.com·14h
🔍Nsight
Flag this post
Custom Intelligence: Building AI that matches your business DNA
aws.amazon.com·1d
🤖AI Coding Tools
Flag this post
VerfCNN, Optimal Complexity zkSNARK for Convolutional Neural Networks
eprint.iacr.org·2d
🧮cuDNN
Flag this post
Polish emerges as top language in multilingual AI benchmark testing
ppc.land·1h
🛠Ml-eng
Flag this post
Resource-Efficient and Robust Inference of Deep and Bayesian Neural Networks on Embedded and Analog Computing Platforms
arxiv.org·3d
🧮cuDNN
Flag this post
Fortytwo's decentralized AI has the answer to life, the universe, and everything
theregister.com·1h
⚡ONNX Runtime
Flag this post
TIL: For long-lived LLM sessions, swapping KV Cache to RAM is ~10x faster than recalculating it. Why isn't this a standard feature?
🔲Loop Tiling
Flag this post
ParallelMind Engine: First AI System with Parallel Logical Reasoning (202+ problems/sec)
🤖AI Coding Tools
Flag this post
Weak-To-Strong Generalization
lesswrong.com·7h
📉Model Quantization
Flag this post
Loading...Loading more...