Skip to main content
Scour
Browse
Getting Started
Login
Sign Up
You are offline. Trying to reconnect...
Close
Copied to clipboard
Close
Unable to share or copy to clipboard
Close
馃 Language Models
transformer architecture, fine-tuning, pretraining, RLHF
Filter Results
Timeframe
Fresh
Past Hour
Today
This Week
This Month
Feeds to Scour
Subscribed
All
Scoured
186578
posts in
21.9
ms
Carbon-Taxed
Transformers: A Green Compression Pipeline for
Overgrown
Language Models
聽
馃
LLMs
arxiv.org
路
1d
Differentially-Private
Text Rewriting reshapes
Linguistic
Style
聽
馃
LLMs
arxiv.org
路
21h
What
Kind
of Language is Easy to Language-Model Under
Curriculum
Learning?
聽
馃
LLMs
arxiv.org
路
21h
Evaluating
Temporal
Consistency
in Multi-Turn Language Models
聽
馃
LLMs
arxiv.org
路
2d
A Multi-Dimensional Audit of
Politically
Aligned
Large Language Models
聽
馃
LLMs
arxiv.org
路
2d
A Survey on Split Learning for LLM
Fine-Tuning
: Models, Systems, and Privacy
Optimizations
聽
馃
LLMs
arxiv.org
路
2d
In-Sync: Adaptation of Speech Aware Large Language Models for
ASR
with Word Level
Timestamp
Predictions
聽
馃
LLMs
arxiv.org
路
2d
Contextual
Linear Activation
Steering
of Language Models
聽
馃
LLMs
arxiv.org
路
2d
Marco-MoE
: Open Multilingual Mixture-of-Expert Language Models with Efficient
Upcycling
聽
馃
LLMs
arxiv.org
路
1d
The
Randomness
Floor: Measuring
Intrinsic
Non-Randomness
in Language Model Token Distributions
聽
馃
LLMs
arxiv.org
路
2d
One
Refiner
to Unlock Them All: Inference-Time Reasoning
Elicitation
via Reinforcement Query Refinement
聽
馃
LLMs
arxiv.org
路
1d
DGLight
:
DQN-Guided
GRPO Fine-Tuning of Large Language Models for Traffic Signal Control
聽
馃
LLMs
arxiv.org
路
1d
All That
Glitters
Is Not Audio: Rethinking Text
Priors
and Audio Reliance in Audio-Language Evaluation
聽
馃
LLMs
arxiv.org
路
2d
Programming with Data: Test-Driven Data Engineering for Self-Improving LLMs from
Raw
Corpora
聽
馃
LLMs
arxiv.org
路
1d
Why Does Reinforcement Learning
Generalize
? A Feature-Level
Mechanistic
Study of Post-Training in Large Language Models
聽
馃
LLMs
arxiv.org
路
1d
Neural Network Optimization
Reimagined
:
Decoupled
Techniques for Scratch and Fine-Tuning
聽
馃敩
AI Research
arxiv.org
路
2d
Parameter
Efficiency Is Not Memory Efficiency:
Rethinking
Fine-Tuning for On-Device LLM Adaptation
聽
馃
LLMs
arxiv.org
路
2d
The
Override
Gap: A Magnitude Account of Knowledge Conflict Failure in
Hypernetwork-Based
Instant LLM Adaptation
聽
馃
LLMs
arxiv.org
路
2d
Compute
Aligned
Training: Optimizing for Test Time Inference
聽
馃
LLMs
arxiv.org
路
1d
Scaling Properties of
Continuous
Diffusion
Spoken
Language Models
聽
馃
LLMs
arxiv.org
路
2d
Sign up or log in to see more results
Sign Up
Login
« Page 2
Log in to enable infinite scrolling
Keyboard Shortcuts
Navigation
Next / previous item
j
/
k
Open post
o
or
Enter
Preview post
v
Post Actions
Love post
a
Like post
l
Dislike post
d
Undo reaction
u
Save / unsave
s
Recommendations
Add interest / feed
Enter
Not interested
x
Go to
Home
g
h
Interests
g
i
Feeds
g
f
Likes
g
l
History
g
y
Changelog
g
c
Settings
g
s
Browse
g
b
Search
/
Pagination
Next page
n
Previous page
p
General
Show this help
?
Submit feedback
!
Close modal / unfocus
Esc
Press
?
anytime to show this help