Skip to main content
Scour
Browse
Getting Started
Login
Sign Up
You are offline. Trying to reconnect...
Close
Copied to clipboard
Close
Unable to share or copy to clipboard
Close
⚙️ MLOps
Specific
mlops, model deployment, ml pipeline, model serving
Filter Results
Timeframe
Fresh
Past Hour
Today
This Week
This Month
Feeds to Scour
Subscribed
All
Scoured
148960
posts in
11.4
ms
Model
Packaging
Tools Every
MLOps
Engineer Should Know
🔥
PyTorch
freecodecamp.org
·
3d
MLOps
in 2026: What Is It and Why Should You Care?
🧠
Machine Learning
flexiana.com
·
18h
Fast
Heterogeneous
Serving: Scalable Mixed-Scale LLM Allocation for
SLO-Constrained
Inference
📝
LLMs
arxiv.org
·
3h
Mastering
MLflow
with Stochastic Gradient
Descent
: A Complete Guide to Diabetes Regression
🧠
Machine Learning
pub.towardsai.net
·
1d
Why Most
ML
Models Die After the
Notebook
(And How to Fix It)
📝
LLMs
photokheecher.medium.com
·
14h
Vulnerabilities
in
Mlflow
software
📝
LLMs
cert.pl
·
2d
Dockerizing
ML Models: A Data Engineer’s Guide to Model
Serving
🧠
Machine Learning
medium.com
·
4d
Benchmarking
LLMs with
Marimo
Pair
📝
LLMs
ericmjl.github.io
·
9h
·
Hacker News
Deep Dive into Google Cloud
Pub/Sub
Single Message
Transforms
and AI Inference
📝
LLMs
medium.com
·
2d
milanm/AutoGrad-Engine
: A complete GPT language model (training and inference) in ~600 lines of pure C#, zero dependencies
📝
LLMs
github.com
·
16h
·
Hacker News
I Built a Production
MLOps
Platform from Scratch :
Kubeflow
, Kafka, Terraform, and Live on GCP
📝
LLMs
medium.com
·
6d
Own your AI.
Optimized
down to the
kernel
📱
Edge AI
runinfra.ai
·
2d
·
Hacker News
How I Built a
Fine-Tuned
Medical AI App and
Deployed
It End-to-End on AWS
📱
Edge AI
medium.com
·
6h
Show HN: Pre-training,
fine-tuning
, and
evals
platform
✨
Generative AI
oumi.ai
·
6d
·
Hacker News
Awesome
Open Source AI
✨
Generative AI
awesomeosai.com
·
5d
·
r/SideProject
The case for Model-as-a-Service over
self-managed
inference
📝
LLMs
news.ycombinator.com
·
3d
·
Hacker News
Blink: CPU-Free LLM Inference by
Delegating
the Serving Stack to GPU and
SmartNIC
📱
Edge AI
arxiv.org
·
3h
PacifAIst/Quansloth
: Based on the implementation of Google's TurboQuant (ICLR 2026) —
Quansloth
brings elite KV cache compression to local LLM inference.
Quansloth
is a fully private, air-gapped AI server that runs massive context models natively on consumer hardware with ease
📝
LLMs
github.com
·
2d
·
Hacker News
benchmarking
inference
of popular models on consumer hardware
🔥
PyTorch
inferena.tech
·
5d
·
Hacker News
Inference
Arena
– new
benchmark
of local inference and training
📱
Edge AI
kvark.github.io
·
4d
·
Hacker News
Loading...
Loading more...
Page 2 »
Keyboard Shortcuts
Navigation
Next / previous item
j
/
k
Open post
o
or
Enter
Preview post
v
Post Actions
Love post
a
Like post
l
Dislike post
d
Undo reaction
u
Save / unsave
s
Recommendations
Add interest / feed
Enter
Not interested
x
Go to
Home
g
h
Interests
g
i
Feeds
g
f
Likes
g
l
History
g
y
Changelog
g
c
Settings
g
s
Browse
g
b
Search
/
Pagination
Next page
n
Previous page
p
General
Show this help
?
Submit feedback
!
Close modal / unfocus
Esc
Press
?
anytime to show this help