Serverless: Fast to Market, Faster to the Future • Srushith Repakula • GOTO 2025
youtube.com·4d
☁️Serverless Rust
2025 State of AI Report and Predictions
thezvi.substack.com·2d·
Discuss: Substack
🏗️AI Infrastructure
Introducing OpenZL: An Open Source Format-Aware Compression Framework
engineering.fb.com·6d·
➡️Arrow Format
Control Codegen Spend
oreilly.com·3d
☁️Serverless Rust
(FULLY OPEN SOURCE) open-computer-use: Computer agents working on their own VMs
github.com·15h·
Discuss: Hacker News
☁️Serverless Rust
GNN Predictions: Hidden Bugs and the Verification Nightmare by Arvind Sundararajan
dev.to·1d·
Discuss: DEV
📱Edge AI
The Library Method: Understanding @cache
dev.to·2d·
Discuss: DEV
λFunctional Programming
In-Depth Analysis: "Attention Is All You Need"
dev.to·2d·
Discuss: DEV
🏗️AI Infrastructure
Parameterized Complexity of s-Club Cluster Edge Deletion
arxiv.org·3d
🗳️Raft Algorithm
Krish Naik: Complete RAG Crash Course With Langchain In 2 Hours
dev.to·1d·
Discuss: DEV
🏗️AI Infrastructure
Ollama SDKs in Go: Overview and Code Examples
dev.to·18h·
Discuss: DEV
🦋Tauri
From ML Beginner to Production Engineer: How I’m Leveling Up My AI Projects
dev.to·2h·
Discuss: DEV
🏗️AI Infrastructure
Automated Performance Appraisal System Calibration via Bayesian Hyperparameter Optimization
dev.to·3d·
Discuss: DEV
📈Time-series Analytics
Quantum Autonomy: Algorithms That Design Themselves by Arvind Sundararajan
dev.to·10h·
Discuss: DEV
🧠Neuromorphic Chips
Automated Anomaly Detection in Account Takeover via Multi-Modal Graph Neural Network Fusion
dev.to·1d·
Discuss: DEV
💻Local LLMs
How I Built a $10/mo Headless CMS That Competes with $99/mo Solutions
dev.to·3d·
Discuss: DEV
☁️Serverless Rust
Accelerated Cold Tolerance Breeding via Multi-Modal Phenotyping and Genome-Wide Predictive Modeling
dev.to·21h·
Discuss: DEV
💧Hydroponics Control
Wavelet Predictive Representations for Non-Stationary Reinforcement Learning
arxiv.org·5d
🏗️AI Infrastructure
🚀 Day 38 of My Data Journey
dev.to·4d·
Discuss: DEV
📐Data Modeling
H1B-KV: Hybrid One-Bit Caches for Memory-Efficient Large Language Model Inference
arxiv.org·4d
💻Local LLMs