Key Moments

Breaking AI to Fix It: Ian Webster's Journey from Discord's Clyde to Promptfoo's $18M Series A

Latent Space PodcastLatent Space Podcast
Science & Technology4 min read44 min video
Oct 24, 2025|1,455 views|27|1
Save to Pod
TL;DR

Promptfoo's Ian Webster discusses AI security testing, shifting from evals to red teaming on a $18M funding round.

Key Insights

1

Promptfoo evolved from an AI evaluation tool to a security red teaming platform due to the critical need for production AI risk management.

2

AI application security focuses on application-specific risks and developer errors, rather than just foundational model toxicity.

3

Promptfoo uses AI-to-AI testing by generating tailored attacks to find vulnerabilities in AI applications, simulating real-world threats.

4

The platform supports various risk areas including technical security, compliance, trust & safety, and brand reputation.

5

Security for AI is shifting earlier in the development lifecycle, moving from runtime solutions to CI/CD integration.

6

The rise of complex agents and RAGs necessitates more sophisticated testing and observability beyond simple input/output checks.

FROM DISCORD'S CLYDE TO PROMPTFU'S ORIGINS

Ian Webster's journey into AI development began at Discord, where he led efforts on AI initiatives like the social chatbot Clyde. Experiencing firsthand the challenges of deploying AI to millions of users, particularly in managing risks and performing evaluations, sparked the idea for Promptfoo. This realization of a critical gap between AI development and safe production deployment became the foundational motivation for creating what is now Promptfoo, an industry-leading AI security testing platform that recently secured $18 million in Series A funding. The initial pain points of shipping AI at scale directly informed the product's direction.

PROMPTFOO: EVOLUTION FROM EVALUATION TO SECURITY RED TEAMING

Initially conceived as an open-source evaluation (eval) tool inspired by side-by-side comparisons used in search algorithm testing, Promptfoo quickly pivoted its go-to-market strategy. Webster argues that evaluations are becoming table stakes and commodities in the AI space. Instead of competing in a crowded eval market, Promptfoo focused on the higher-value, more critical area of AI security and risk management. This strategic shift recognized that while model providers address foundational risks like toxicity, significant vulnerabilities emerge from application-specific implementations and developer errors, an area Promptfoo aims to rigorously test.

UNDERSTANDING AI APPLICATION RISKS

Promptfoo differentiates between foundational model risks, such as toxicity, which are largely addressed by model providers, and application risks. These application risks stem from how developers implement AI, such as insecure access controls in RAG systems or poorly managed agent tool capabilities. Webster emphasizes that entities like OpenAI and Anthropic focus on model safety, but can't 'fix stupid' in application logic. He highlights that the interface between the AI application and the model is where most new vulnerabilities arise, especially when AI is integrated into enterprise workflows where maximum helpfulness isn't always the desired outcome, leading to softer risks like off-label recommendations or competitor praise.

THE PROMPTFOO TESTING METHODOLOGY

Promptfoo operates on an AI-versus-AI principle, generating tailored attacks against target AI applications. Users provide information about the application's purpose, business context, and intended use cases, which informs the testing process. The platform then utilizes 'plugins,' representing various risk areas, to generate attack objectives. These objectives are fed into 'strategies,' which are adapted ML research methods. An attacker LLM interacts with the target application, while a judge LLM evaluates the outputs, iteratively refining attacks to discover vulnerabilities. This dynamic, on-the-fly generation contrasts with static vulnerability databases, making it highly relevant to application security.

EMERGING TRENDS IN AI SECURITY TESTING

The landscape of AI security is rapidly evolving, with a strong push towards integrating security testing earlier into the Software Development Lifecycle (SDLC), ideally within CI/CD pipelines. While guardrails were an initial focus, the realization that they aren't foolproof has led to a demand for pre-deployment risk measurement. Companies are beginning to see AI security not just as securing the AI model itself, but as a broader category of software security, with tools like Promptfoo addressing this gap. The complexity of emerging AI architectures like agents and advanced RAG systems necessitates move sophisticated testing that goes beyond basic conversational checks.

THE ENTERPRISE ADOPTION AND FUTURE OF AI RED TEAMING

Promptfoo is seeing significant traction with large enterprises, including over 10% of the Fortune 500, due to their higher stakes in case of AI-related breaches. While startups might prioritize speed-to-market, established corporations require rigorous security validation. The enterprise version of Promptfoo focuses on extending the vulnerability management lifecycle, offering features for triage, integration with existing systems (like Jira and SIEM), monitoring, and remediation validation. Looking ahead, Webster anticipates the need for even tighter integration with developer workflows, potentially through IDE plugins and code-aware analysis, to address the security challenges posed by increasingly complex AI systems like agents and RAGs.

Promptfoo: AI Application Security Best Practices

Practical takeaways from this episode

Do This

Focus on application-level security risks, not just foundation model toxicity.
Provide detailed business context and user personas for more accurate testing.
Utilize AI vs. AI for generating tailored attacks against your application.
Integrate security testing early in the SDLC, ideally in CI/CD pipelines.
Consider both offensive testing capabilities and defensive measures like proxies for MCP.
If managing MCP, prevent local execution and centralize observability and control.

Avoid This

Assume foundation models alone will solve all security issues.
Rely solely on black-box testing; provide internal context when possible.
Treat guardrails as a complete solution; they often require tuning.
Neglect security testing until after deployment.
Implement MCP in a naive wrapper around APIs without considering security implications.

Common Questions

Promptfoo is a tool designed to find and fix issues in AI applications, primarily focusing on security. It uses an 'AI vs. AI' approach, generating tailored attacks based on your application's context to uncover potential risks that static databases might miss.

Topics

Mentioned in this video

More from Latent Space

View all 84 summaries

Found this useful? Build your knowledge library

Get AI-powered summaries of any YouTube video, podcast, or article in seconds. Save them to your personal pods and access them anytime.

Try Summify free