Skip to main content
Scour
Browse
Getting Started
Login
Sign Up
You are offline. Trying to reconnect...
Copied to clipboard
Unable to share or copy to clipboard
⚙️ MLOps
Specific
model serving, inference, ML pipelines, model monitoring
Filter Results
Timeframe
Fresh
Past Hour
Today
This Week
This Month
Feeds to Scour
Subscribed
All
Scoured
187614
posts in
18.5
ms
Caltech
’s
PrismML
shrinks AI models to fit your phone without losing their mind
🤖
AI Engineering
startupfortune.com
·
3d
Build Strands Agents with
SageMaker
AI models and
MLflow
🤖
AI Engineering
aws.amazon.com
·
4d
Darwinian
Specialization
in AI
🔬
AI Research
tomtunguz.com
·
3d
Best
Practices
for inference on Edge AI
MCUs
🤖
AI Engineering
embedded.com
·
2d
Reinforced
Agent: Inference-Time Feedback for
Tool-Calling
Agents
🔎
AI Interpretability
machinelearning.apple.com
·
1d
AmSach/kvquant
: Drop-in KV cache compressor for local LLM inference - Run 70B models on 8GB RAM
🧠
LLMs
github.com
·
1d
·
DEV
How we built the most performant DeepSeek V3.2, MiniMax-M2.5 and Qwen 3.5
397B
on DigitalOcean NVIDIA
HGX
™ B300 GPU Droplets
📊
Benchmarking
digitalocean.com
·
4d
Belief-Guided
Inference Control for Large Language Model Services via
Verifiable
Observations
🧠
LLMs
arxiv.org
·
1d
MauroCE/m3serve
: Optimised BAAI/bge-m3 serving with dense + sparse + ColBERT embeddings, async dynamic batching and pipeline GPU inference
📊
Benchmarking
github.com
·
4d
·
r/SideProject
Think it, Run it: Autonomous
ML
pipeline generation via
self-healing
multi-agent AI
🤖
AI Engineering
arxiv.org
·
1d
AI
Observability
for Large Language Model Systems: A Multi-Layer Analysis of Monitoring Approaches from Confidence
Calibration
to Infrastructure Tracing
✅
Formal Verification
arxiv.org
·
2d
Introducing
DigitalOcean
AI-Native Cloud for Production AI
Workloads
🤖
AI Engineering
digitalocean.com
·
3d
Adaptive and Fine-grained Module-wise Expert Pruning for Efficient
LoRA-MoE
Fine-Tuning
🧠
LLMs
arxiv.org
·
2d
Compliance versus
Sensibility
: On the Reasoning
Controllability
in Large Language Models
🧠
LLMs
arxiv.org
·
1d
Diagnosing
Capability
Gaps in Fine-Tuning Data
🤖
AI Engineering
arxiv.org
·
1d
Strait:
Perceiving
Priority and
Interference
in ML Inference Serving
🧠
LLMs
arxiv.org
·
1d
Programming with Data: Test-Driven Data Engineering for Self-Improving LLMs from
Raw
Corpora
🧠
LLMs
arxiv.org
·
3d
When Your LLM Reaches End-of-Life: A Framework for
Confident
Model
Migration
in Production Systems
🧠
LLMs
arxiv.org
·
1d
Efficient,
VRAM-Constrained
xLM
Inference on Clients
📊
Benchmarking
arxiv.org
·
2d
Incompressible
Knowledge Probes: Estimating Black-Box LLM Parameter Counts via
Factual
Capacity
🧠
LLMs
arxiv.org
·
3d
·
Hacker News
« Page 1
·
Page 3 »
Log in to enable infinite scrolling
Keyboard Shortcuts
Navigation
Next / previous item
j
/
k
Open post
o
or
Enter
Preview post
v
Post Actions
Love post
a
Like post
l
Dislike post
d
Undo reaction
u
Save / unsave
s
Recommendations
Add interest / feed
Enter
Not interested
x
Go to
Home
g
h
Interests
g
i
Feeds
g
f
Likes
g
l
History
g
y
Changelog
g
c
Settings
g
s
Browse
g
b
Search
/
Pagination
Next page
n
Previous page
p
General
Show this help
?
Submit feedback
!
Close modal / unfocus
Esc
Press
?
anytime to show this help