AI Safety and Alignment Group

AI Safety and Alignment Group

Home
Notes
Archive
About
PostTrainBench: Measuring AI Ability to Perform LLM Post-Training
We expect this to be an important indicator for AI R&D automation as it unfolds over the next few years
Jan 6 • Maksym Andriushchenko, Ben Rank, and Hardik Bhatnagar
Agent Skills Enable a New Class of Realistic and Trivially Simple Prompt Injections
A new high-trust, low-visibility attack surface.
Oct 31, 2025 • David, Sahar Abdelnabi, and Maksym Andriushchenko
Welcome post from a new research group
New group at the ELLIS Institute Tübingen and Max Planck Institute for Intelligent Systems
Oct 27, 2025 • Maksym Andriushchenko
AI Safety and Alignment Group
AI Safety and Alignment Group
AI Safety and Alignment Group at the ELLIS Institute Tübingen and Max Planck Institute for Intelligent Systems

AI Safety and Alignment Group

AboutArchiveSitemap
© 2026 Maksym Andriushchenko · Privacy ∙ Terms ∙ Collection notice
Start your SubstackGet the app
Substack is the home for great culture