Skip to main content
Scour
Browse
Getting Started
Login
Sign Up
You are offline. Trying to reconnect...
Close
Copied to clipboard
Close
Unable to share or copy to clipboard
Close
🛡️ AI Safety
Alignment, Interpretability, Adversarial Examples, Ethics
Filter Results
Timeframe
Fresh
Past Hour
Today
This Week
This Month
Feeds to Scour
Subscribed
All
Scoured
8354
posts in
42.4
ms
The
Persistent
Vulnerability of
Aligned
AI Systems
🛡️
AI Security
arxiv.org
·
11h
·
…
The
Ethics
of
Artificial
Intelligence
⚖️
AI Ethics
hackettpublishing.com
·
3d
·
…
The
Ethics
Theater
of AI: Why Switching From ChatGPT to Claude Changes Less Than You Think
⚖️
AI Ethics
hackernoon.com
·
1d
·
…
When AI turns software development inside-out: 170%
throughput
at 80%
headcount
⚡
Code Generation
venturebeat.com
·
4d
·
…
Evaluating Human-AI Safety: A Framework for Measuring Harmful
Capability
Uplift
🤝
Human-AI Collaboration
arxiv.org
·
2d
·
…
Empirical
Validation of the Classification-Verification
Dichotomy
for AI Safety Gates
✍️
Prompt Engineering
arxiv.org
·
11h
·
…
Adversarial
Moral
Stress Testing of Large Language Models
🛡️
AI Security
arxiv.org
·
11h
·
…
Detection of
Adversarial
Attacks in Robotic
Perception
🛡️
AI Security
arxiv.org
·
2d
·
…
Robust Multimodal Safety via
Conditional
Decoding
👁️
Multimodal AI
arxiv.org
·
11h
·
…
A Revealed
Preference
Framework for AI
Alignment
👁️
Multimodal AI
arxiv.org
·
2d
·
…
AI Security in the Foundation Model Era: A Comprehensive Survey from a
Unified
Perspective
🛡️
AI Security
arxiv.org
·
6d
·
…
How Do Language Models Process Ethical
Instructions
?
Deliberation
, Consistency, and Other-Recognition Across Four Models
✍️
Prompt Engineering
arxiv.org
·
11h
·
…
Rethinking AI
Literacy
Education in Higher Education: Bridging Risk Perception and
Responsible
Adoption
⚖️
AI Ethics
arxiv.org
·
1d
·
…
A Unified Memory Perspective for
Probabilistic
Trustworthy
AI
🛡️
AI Security
arxiv.org
·
6d
·
…
A
Provable
Energy-Guided Test-Time Defense Boosting Adversarial
Robustness
of Large Vision-Language Models
🛡️
AI Security
arxiv.org
·
2d
·
…
Lipschitz
verification
of neural networks through training
✓
Formal Verification
arxiv.org
·
2d
·
…
BeSafe-Bench
: Unveiling Behavioral Safety Risks of
Situated
Agents in Functional Environments
🎯
AI Agents
arxiv.org
·
3d
·
…
A
Comparative
Study in
Surgical
AI: Datasets, Foundation Models, and Barriers to Med-AGI
🤖
Agentic AI
arxiv.org
·
2d
·
…
Position: Explainable AI is
Causality
in
Disguise
⚖️
AI Ethics
arxiv.org
·
2d
·
…
FairLLaVA
:
Fairness-Aware
Parameter-Efficient Fine-Tuning for Large Vision-Language Assistants
👁️
Multimodal AI
arxiv.org
·
3d
·
…
Loading...
Loading more...
Page 2 »
Keyboard Shortcuts
Navigation
Next / previous item
j
/
k
Open post
o
or
Enter
Preview post
v
Post Actions
Love post
a
Like post
l
Dislike post
d
Undo reaction
u
Recommendations
Add interest / feed
Enter
Not interested
x
Go to
Home
g
h
Interests
g
i
Feeds
g
f
Likes
g
l
History
g
y
Changelog
g
c
Settings
g
s
Browse
g
b
Search
/
Pagination
Next page
n
Previous page
p
General
Show this help
?
Submit feedback
!
Close modal / unfocus
Esc
Press
?
anytime to show this help