The Innovations in DeepSeek OCR
x.com·1d·
Discuss: r/LocalLLaMA
Flag this post

DeepSeek just released a pretty shocking new paper. They really buried the lede here by referring to it simply as DeepSeek OCR.

While it’s a very strong OCR model, the purpose of it and the implications of their approach go far beyond what you’d expect of “yet another OCR model.”

Traditionally, vision LLM tokens almost seemed like an afterthought or “bolt on” to the LLM paradigm. And 10k words of English would take up far more space in a multimodal LLM when expressed as intelligible pixels than when expressed as tokens.

So those 10k words may have turned into 15k tokens, or 30k to 60k “visual tokens.” So vision tokens were way less efficient and really only made sense to use for data that couldn’t be effectively conveyed with words. <...

Similar Posts

Loading similar posts...