Anyone With API Access Can Clone an AI Model and Make It Unsafe
pub.towardsai.net
·1d
Effect Inference
Preview
Report Post

It costs under $100. Security researchers have documented this since 2020. Regulators haven’t caught up.

Source: image by the author

In December 2025, researchers published a paper showing they could clone a safety-aligned medical AI for $12. The clone retained the medical expertise but lost much of the safety alignment, with unsafe output rates rising from 66% to 86% on adversarial prompts.

This sounds like a new discovery. It isn’t. The same attack has been documented since 2020, demonstrated against ChatGPT, GPT-3.5, BERT, and dozens of other models at costs ranging from $0.20 t...

Similar Posts

Loading similar posts...