Skip to main content
Scour
Browse
Getting Started
Login
Sign Up
You are offline. Trying to reconnect...
Close
Copied to clipboard
Close
Unable to share or copy to clipboard
Close
๐ค Transformers
Specific
Attention Mechanism, BERT, GPT, Language Models
Filter Results
Timeframe
Fresh
Past Hour
Today
This Week
This Month
Feeds to Scour
Subscribed
All
Scoured
185738
posts in
24.2
ms
Applications of the Transformer Architecture in
AI-Assisted
English Reading
Comprehension
ย
๐๏ธ
AI Infrastructure
arxiv.org
ยท
2d
Paper page - Large Language Models Explore by
Latent
Distilling
ย
๐
RAG
huggingface.co
ยท
5h
The
Recurrent
Transformer:
Greater
Effective Depth and Efficient Decoding (5 minute read)
ย
๐ฃ๏ธ
Speech Synthesis
alphaxiv.org
ยท
1d
Temporal
Language Models
ย
๐
Parser Combinators
calcifercomputing.com
ยท
2d
ยท
Hacker News
LayerBoost
: Layer-Aware Attention
Reduction
for Efficient LLMs
ย
๐ป
Local LLMs
arxiv.org
ยท
3d
Associative-State
Universal Transformers: Sparse Retrieval Meets Structured
Recurrence
ย
๐
Finite Automata
arxiv.org
ยท
16h
Fine-Grained
Analysis of Shared
Syntactic
Mechanisms in Language Models
ย
๐
Parser Combinators
arxiv.org
ยท
3d
Investigation into In-Context Learning
Capabilities
of
Transformers
ย
๐ค
AI Inference
arxiv.org
ยท
1d
Dissociating
Decodability
and Causal Use in Bracket-Sequence Transformers
ย
๐
Parser Combinators
arxiv.org
ยท
3d
Barriers
to Universal Reasoning With Transformers (And How to
Overcome
Them)
ย
๐ฏ
Hindley-Milner
arxiv.org
ยท
1d
Training
Transformers
as a
Universal
Computer
ย
๐๏ธ
AI Infrastructure
arxiv.org
ยท
1d
Estimating
Tail Risks in Language Model Output
Distributions
ย
๐๏ธ
AI Infrastructure
arxiv.org
ยท
3d
Automating
Categorization
of Scientific Texts with In-Context Learning and
Prompt-Chaining
in Large Language Models
ย
๐
NLP
arxiv.org
ยท
2d
Adaptive ToR: Complexity-Aware Tree-Based Retrieval for
Pareto-Optimal
Multi-Intent
NLU
ย
๐
Parser Combinators
arxiv.org
ยท
2d
The
Recurrent
Transformer:
Greater
Effective Depth and Efficient Decoding
ย
๐ฑ
Edge AI
arxiv.org
ยท
6d
G-Loss:
Graph-Guided
Fine-Tuning
of Language Models
ย
๐ป
Local LLMs
arxiv.org
ยท
1d
The
Structured
Output Benchmark: A Multi-Source Benchmark for
Evaluating
Structured
Output Quality in Large Language Models
ย
๐๏ธ
Whisper
arxiv.org
ยท
1d
An
Information-Geometric
Framework for Stability Analysis of Large Language Models under
Entropic
Stress
ย
๐ก
Information Theory
arxiv.org
ยท
2d
Explainable
AI in Speaker Recognition -- Making Latent Representations
Understandable
ย
๐๏ธ
AI Infrastructure
arxiv.org
ยท
2d
Large Language Models Explore by
Latent
Distilling
ย
๐ป
Local LLMs
arxiv.org
ยท
1d
Log in to enable infinite scrolling
Keyboard Shortcuts
Navigation
Next / previous item
j
/
k
Open post
o
or
Enter
Preview post
v
Post Actions
Love post
a
Like post
l
Dislike post
d
Undo reaction
u
Save / unsave
s
Recommendations
Add interest / feed
Enter
Not interested
x
Go to
Home
g
h
Interests
g
i
Feeds
g
f
Likes
g
l
History
g
y
Changelog
g
c
Settings
g
s
Browse
g
b
Search
/
Pagination
Next page
n
Previous page
p
General
Show this help
?
Submit feedback
!
Close modal / unfocus
Esc
Press
?
anytime to show this help