Alignment Research, Model Robustness, Adversarial Examples, Risk Assessment
Safe Pruning LoRA: Robust Distance-Guided Pruning for Safety Alignment in Adaptation of LLMs
arxiv.orgยท20h
I Tested LLM Agents on Simple Safety Rules. They Failed in Surprising and Informative Ways.
lesswrong.comยท2h
Who Would Win: A State-of-the-Art Foundation Model or a Neural Net?
pub.towardsai.netยท8h
A single person with AI and a grudge could create a virus to wipe us out.
threadreaderapp.comยท11h
How AI/LLMs Can Help, Hinder Developers
cacm.acm.orgยท3h
HW Security: Multi-Agent AI Assistant Leveraging LLMs To Automate Key Stages of SoC Security Verification (U. of Florida)
semiengineering.comยท17h
What Inflection AI Learned Porting Its LLM Inference Stack from NVIDIA to Intel Gaudi
thenewstack.ioยท6h
Audit smarter: Introducing Google Cloudโs Recommended AI Controls framework
cloud.google.comยท8h
AI Is Already Crushing the News Industry
theatlantic.comยท3h
From HAL 9000 to ME3AN: what filmโs evil robots tell us about contemporary tech fears
theconversation.comยท4h
It's elementary: Problem-solving AI approach tackles inverse problems used in nuclear physics and beyond
phys.orgยท7h
ReMAR-DS: Recalibrated Feature Learning for Metal Artifact Reduction and CT Domain Transformation
arxiv.orgยท20h
Loading...Loading more...