Launch HN: MindFort (YC X25) – AI agents for continuous pentesting

Hey HN! We're Brandon, Sam, and Akul from MindFort (https://mindfort.ai). We're building autonomous AI agents that continuously find, validate, and patch security vulnerabilities in web applications—essentially creating an AI red team that runs 24/7.

Here's a demo: https://www.loom.com/share/e56faa07d90b417db09bb4454dce8d5a

Security testing today is increasingly challenging. Traditional scanners generate 30-50% false positives, drowning engineering teams in noise. Manual penetration testing happens quarterly at best, costs tens of thousands per assessment, and takes weeks to complete. Meanwhile, teams are shipping code faster than ever with AI assistance, but security reviews have become an even bigger bottleneck.

All three of us encountered this problem from different angles. Brandon worked at ProjectDiscovery building the Nuclei scanner, then at NetSPI (one of the largest pen testing firms) building AI tools for testers. Sam was a senior engineer at Salesforce leading security for Tableau. He dealt firsthand with juggling security findings and managing remediations. Akul did his master's on AI and security, co-authored papers on using LLMs for ecurity attacks, and participated in red-teams at OpenAI and Anthropic.

We all realized that AI agents were going to fundamentally change security testing, and that the wave of AI-generated code would need an equally powerful solution to keep it secure.

We've built AI agents that perform reconnaissance, exploit vulnerabilities, and suggest patches—similar to how a human penetration tester works. The key difference from traditional scanners is that our agents validate exploits in runtime environments before reporting them, reducing false positives.

We use multiple foundational models orchestrated together. The agents perform recon to understand the attack surface, then use that context to inform testing strategies. When they find potential vulnerabilities, they spin up isolated environments to validate exploitation. If successful, they analyze the codebase to generate contextual patches.

What makes this different from existing tools? Validation through exploitation: We don't just pattern-match—we exploit vulnerabilities to prove they're real; - Codebase integration: The agents understand your code structure to find complex logic bugs and suggest appropriate fixes; - Continuous operation: Instead of point-in-time assessments, we're constantly testing as your code evolves; - Attack chain discovery: The agents can find multi-step vulnerabilities that require chaining different issues together.

We're currently in early access, working with initial partners to refine the platform. Our agents are already finding vulnerabilities that other tools miss and scoring well on penetration testing benchmarks.

Looking forward to your thoughts and comments!


Comments URL: https://news.ycombinator.com/item?id=44117465

Points: 4

# Comments: 0

https://news.ycombinator.com/item?id=44117465

Creată 3d | 28 mai 2025, 16:50:08


Autentifică-te pentru a adăuga comentarii

Alte posturi din acest grup

Show HN: AI Peer Reviewer – Multiagent System for Scientific Manuscript Analysis

After waiting 8 months for a journal response or two months for co-author feedback that consisted of "looks good" and a single comma change, we built an AI-powered peer review system that helps re

31 mai 2025, 16:40:17 | Hacker news
Show HN: Fontofweb – Discover Fonts Used on a Website or Websites Using Font(s)

Hey HN, I've been working on fontofweb.com on and off for the past 4 years, and I'm keen to share it with you. It lets you type in the URL of any website and see exactly how the fonts are used: al

31 mai 2025, 16:40:13 | Hacker news
Ask HN: Anyone making a living from a paid API?

Are there any solo devs or small teams out there genuinely paying their rent from selling API access?

What's your API? How much MRR? What's your pricing model? How did you find your first paying

31 mai 2025, 16:40:11 | Hacker news