AI Security & Safety Research Lab
We secure the systems that think for themselves.
Offensive security research for the AI era. We break AI systems so you can ship them safely.
Schedule a call →Our team previously helped secure
Focus Areas
Projects
View all →Adversarial RL Environment
ActiveReinforcement learning finally works for security. All you need is the right feedback loop. Our RL environments transform pretrained models into offensive and defensive security agents. Robust, packed with realistic attack scenarios, and designed with hundreds of programmatically verifiable security challenges at the edge of frontier capabilities.
AI Safety Research
ResearchResearch into how frontier AI systems fail - from reward hacking and hallucination cascades to alignment drift in autonomous pipelines. We build benchmarks and evaluation frameworks that map the boundaries of safe behavior across model families and real-world deployments.
Agent Security Research
ResearchThreat modeling for autonomous AI agents that make decisions, call tools, and chain actions without human oversight. We study prompt injection, tool poisoning, memory manipulation, and privilege escalation - and publish the attack taxonomies and defense frameworks the industry relies on.
AI for Cybersecurity
ActiveApplying AI to scale offensive security. We use language models and automated analysis to discover vulnerabilities in software - including two CVEs in Apple's software.
Latest
All posts →


Where chaos meets control — the aesthetics of breaking things safely
Connect
Research collaborations, security audits, or just want to talk AI security — hello@arimlabs.ai