Building AI Systems You Can Trust

Building AI Systems You Can Trust

May 23, 2025 47 min
🎧 Listen Now

🤖 AI Summary

Overview

This episode explores the critical importance of trust in deploying AI systems, particularly generative AI, in enterprise environments. Scott Clark, CEO of Distributional, and Matt Bornstein from a16z discuss why traditional performance metrics fail to capture the nuanced behaviors of AI systems and how robust testing frameworks can help enterprises confidently scale AI applications. They also delve into the challenges posed by non-deterministic and non-stationary AI behaviors, the rise of centralized AI platforms, and strategies for managing complexity and mitigating risks.

Notable Quotes

- The thing that's holding back people getting value from these AI systems is not performance. It's about being able to confidently trust these systems.Scott Clark, on the shift from optimization to trust.

- Who gets paged in the middle of the night when your AI bot just sold the office building by mistake?Matt Bornstein, highlighting the need for AI-specific operational teams.

- It's not just what you say, but how you say it.Matt Bornstein, on the importance of AI behavior beyond raw outputs.

🧠 The Shift from Performance to Trust

- Scott Clark explains that optimizing AI systems for peak performance often introduces unintended behaviors, undermining trust. Enterprises are now prioritizing reliability and consistency over squeezing out marginal performance gains.

- Generative AI systems, with their expansive output space and agentic capabilities, require new frameworks to define and test behaviors.

- Trust is multifaceted, encompassing reliability, consistency, and alignment with enterprise values. Testing frameworks are essential to verify these attributes.

🔍 Behavioral Testing in AI Systems

- Behavioral testing focuses not just on outputs but on the process leading to those outputs, including tone, toxicity, reasoning steps, and retrieval mechanisms.

- Scott Clark advocates for using a large number of weak estimators to detect subtle shifts in system behavior, enabling root cause analysis and adaptation.

- Enterprises need to move beyond vibe checks and small-scale tests to holistic evaluations that account for non-deterministic and non-stationary behaviors.

🏢 Centralized AI Platforms and Shadow AI

- Centralized AI platforms are emerging as a solution to manage complexity, scale deployments, and mitigate risks like shadow AI, where unauthorized models or data are used.

- These platforms provide value-add services such as logging, testing, and cost optimization, enticing developers to adopt them.

- Scott Clark notes that enterprises often support dozens of models and versions, requiring robust infrastructure to manage and test them effectively.

⚙️ Scaling AI Confidently

- Enterprises face an AI confidence gap, where promising prototypes fail to scale due to fears of unpredictable behavior.

- Behavioral test coverage helps organizations understand trade-offs when tweaking models, prompts, or infrastructure, reducing operational risks.

- Scott Clark likens testing AI systems to traditional software regression tests, ensuring changes improve functionality without breaking the system.

🌍 Enterprise Influence on AI Labs

- AI labs focus on cutting-edge research, but enterprise adoption drives revenue, creating a co-evolution between labs and industry needs.

- Enterprises exert influence by demanding models tailored to specific use cases, while platform owners act as intermediaries, adapting tools for business needs.

- The rise of AI ops teams will be critical to maintaining and troubleshooting AI systems as they become integral to enterprise operations.

AI-generated content may not be accurate or complete and should not be relied upon as a sole source of truth.

📋 Episode Description

In this episode of AI + a16z, Distributional cofounder and CEO Scott Clark, and a16z partner Matt Bornstein, explore why building trust in AI systems matters more than just optimizing performance metrics. From understanding the hidden complexities of generative AI behavior to addressing the challenges of reliability and consistency, they discuss how to confidently deploy AI in production. 

Why is trust becoming a critical factor in enterprise AI adoption? How do traditional performance metrics fail to capture crucial behavioral nuances in generative AI systems? Scott and Matt dive into these questions, examining non-deterministic outcomes, shifting model behaviors, and the growing importance of robust testing frameworks. 

Among other topics, they cover: 

  • The limitations of conventional AI evaluation methods and the need for behavioral testing. 
  • How centralized AI platforms help enterprises manage complexity and ensure responsible AI use. 
  • The rise of "shadow AI" and its implications for security and compliance. 
  • Practical strategies for scaling AI confidently from prototypes to real-world applications.

Follow everyone:

Scott Clark

Distributional

Matt Bornstein

Derrick Harris


Check out everything a16z is doing with artificial intelligence here, including articles, projects, and more podcasts.