AttackIQ

web paid closed source

AttackIQ continuously tests your cybersecurity defenses by simulating real attacks across your entire infrastructure. Instead of waiting for breaches to find gaps, it runs automated adversary emulations 24/7 to validate what's actually working. Security teams who are tired of playing defense with blindfolds finally get to see which expensive tools are earning their keep.

garak

all free open source

garak scans your AI models for security holes before hackers do. It's backed by NVIDIA and constantly updated by the community to stay ahead of new vulnerabilities. Any company building with LLMs needs this to avoid becoming the next AI security disaster story.

Lakera

web freemium closed source

Lakera stops hackers from breaking your AI chatbots and stealing your data. Their secret weapon is learning from over a million actual AI attacks through their Gandalf hacking game. Perfect for any company deploying GenAI who doesn't want to become tomorrow's security nightmare headline.

Orca Security

web paid closed source

Orca AI is your GenAI-powered cloud security analyst that cuts through alert noise to find what actually matters. It speaks plain English so you can ask 'Do I have log4j vulnerabilities exposed?' instead of learning cryptic query languages. Perfect for overwhelmed security teams who need to move fast without hiring more people.

Pentera

web paid closed source

Pentera automatically hacks your own systems to find security holes before bad guys do. It runs full attack simulations using AI to discover complete kill chains and prioritize what actually matters. Security teams at 996+ companies use it to stop playing whack-a-mole with vulnerabilities and focus on real risks.

PromptArmor

web paid closed source

PromptArmor spots AI security risks hiding in your vendor stack before they bite you. Their secret sauce is mapping exactly how third-party AI tools handle your sensitive data across 26 different risk vectors. Enterprise security teams and law firms use this to sleep better knowing their AI vendors aren't leaking client secrets through prompt injection attacks.

Pricing & Filters