SXSW 2025 recap: Building reliable and trustworthy AI

Nick WarnerNick Warner
Nick Warner
|Last updated May 16, 2025
SUMMARIZE WITH
AI tool logoAI tool logoAI tool logoAI tool logo
Create AI videos with 230+ avatars in 140+ languages.
The Summary
As large language models (LLM) integrate into more applications, clear safety standards ensure reliable results and measure performance and effectiveness. At SXSW 2025, Scale AI’s Summer Yue, HeyGen’s Lavanya Poreddy
The Long Version

As large language models (LLM) integrate into more applications, clear safety standards ensure reliable results and measure performance and effectiveness. At SXSW 2025, Scale AI’s Summer Yue, HeyGen’s Lavanya Poreddy, and Fortune’s Sharon Goldman sat down for the Beyond the Hype: Building Reliable and Trustworthy AI panel to explore pioneering AI safety standards, like rigorous model testing and high-quality data, to ensure a more transparent future for AI adoption. They also discussed best practices for mitigating risks and addressing ethical considerations in AI development.

Learn how leading organizations establish AI safety processes to ensure trust and real-world adoption.

Understanding AI and the challenge of evaluation

AI models are often described as “black boxes”—powerful but opaque. Summer noted that even AI researchers rely on informal sources like Twitter and Reddit to understand model capabilities. This trust gap raises concerns about AI’s unpredictability and biases.

Lavanya likened AI to a newborn child—it learns from data but lacks independent reasoning. AI isn’t autonomous intelligence; it’s human-backed intelligence that must be monitored and refined.

One major issue is how to measure AI performance. Traditional AI models recognize patterns, while reasoning models generate outputs based on different thinking times, making evaluations harder. Even with extensive testing, AI models still hallucinate—generating false or misleading information. This is particularly concerning in safety-critical applications like healthcare or law enforcement.

AI bias and content moderation challenges

Bias remains a major challenge. Lavanya explained that if an AI model is trained only on golden retriever images, it will always assume a “dog” looks like a golden retriever. The same problem applies to hiring models or content moderation, where skewed training data can reinforce biases. The solution? More diverse and inclusive training datasets. However, even with improved training, AI still struggles with context and intent, making human oversight essential.

Trust in AI depends on who sets the rules. AI companies have vastly different policies, leading to inconsistencies in content moderation and safety standards. AI automatically filters harmful content like hate speech or nudity. However, gray-area content (e.g., political discussions) is flagged for human review. Lavanya emphasized that AI cannot replace human decision-making, especially in complex ethical situations.

Summer noted that AI-generated content varies significantly across companies. Some models are highly restrictive, while others are more lenient, creating inconsistent user experiences.

AI in real-world applications

The discussion covered AI’s impact across industries, including:

These examples show that AI can enhance efficiency, but human oversight remains crucial. Lavanya and Summer agreed that greater transparency is key to building trust. AI should be seen as a collaborator, not a replacement for human decision-making. Lavanya and Summer believe:

This SXSW session underscored the importance of responsible AI development. While AI’s potential is undeniable, ensuring it is fair, ethical, and transparent remains a critical challenge.

Resources

Continue Reading

Latest blog posts related to SXSW 2025 recap: Building reliable and trustworthy AI.

Content ImageContent Image
Author ImageAuthor Image
NewsA new chapter: Bin Liu joins HeyGen
Content ImageContent Image
Author ImageAuthor Image
NewsBest Synthesia Alternatives & Competitors in 2025
Content ImageContent Image
Author ImageAuthor Image
NewsNavigating Voice Cloning Ethics and AI Video

Use Cases

From Creators to Marketers.

100+ Use Cases for HeyGen.