Small Vs. Large Language Models
semiengineering.com·13w·
Discuss: Hacker News, r/LLM

The proliferation of edge AI will require fundamental changes in language models and chip architectures to make inferencing and learning outside of AI data centers a viable option.

The initial goal for small language models (SLMs) — roughly 10 billion parameters or less, compared to more than a trillion parameters in the biggest LLMs — was to leverage them exclusively for inferencing. Increasingly, however, they also include some learning capability. And because they are purpose-built for narrowly defined tasks, SLMs can generate results in a fraction of the time it takes to send a query, directive, or sensor data to an AI data center and receive a response.

SLMs are not new. EDA companies have been playing around with optimized computational software for years, and scientists hav…

Similar Posts

Loading similar posts...

Keyboard Shortcuts

Navigation
Next / previous item
j/k
Open post
oorEnter
Preview post
v
Post Actions
Love post
a
Like post
l
Dislike post
d
Undo reaction
u
Recommendations
Add interest / feed
Enter
Not interested
x
Go to
Home
gh
Interests
gi
Feeds
gf
Likes
gl
History
gy
Changelog
gc
Settings
gs
Browse
gb
Search
/
General
Show this help
?
Submit feedback
!
Close modal / unfocus
Esc

Press ? anytime to show this help