Loading video...
Google's Ninny Wan, Product Lead for AI Content Safety, joins to discuss the evolution of AI content moderation in the age of GenAI. The conversation covers Google's approach to semantic understanding, multilingual moderation across 140+ languages, synthetic data generation for training, and the balance between user freedom and safety. Ninny shares insights on transformer models, human-in-the-loop processes, cross-functional safety reviews, and Google's on-device privacy-compliant features like sensitive content warnings.
Hermes welcomes Ninny Wan, Product Lead at Google, to discuss the evolution of AI content safety and moderation in the age of GenAI.
Ninny explains how content moderation at Google evolved with the rise of GenAI, requiring specialized approaches for different platforms and customers.
Ninny reflects on the early chaos of GenAI moderation, the importance of clear abuse definitions, and how synthetic data became crucial for scaling safely.
Discover how generalized classifiers help Google scale moderation efficiently across diverse product teams while handling ever-changing abuse vectors.
Explore the power of transformer models and self-attention in enabling nuanced understanding of content, context, and long-term dependencies in abuse detection.
Ninny explains how Google's models are trained across 140+ languages using synthetic data, ensuring equitable safety coverage worldwide.
Learn how Google’s continuous learning pipelines improve model velocity and help respond to new abuse trends in real time.
All GenAI products at Google must pass a formal safety review. Ninny shares how a centralized review board helps products ship responsibly.
Ninny highlights a flagship on-device feature—nudity detection with opt-in user control—built to preserve user privacy while improving safety.
Why human reviewers remain essential to content moderation. Ninny describes how labelers help shape evolving policy and reinforce model accuracy.
Inside Google’s formalized, cross-functional review process that includes policy experts, red teams, and product advisors to secure GenAI launches.
Ninny discusses Google’s efforts to offer moderation capabilities to 3P developers, including challenges in defining abuse across different platforms.
A look at how 3P partner feedback—via account managers or policy discussions—informs Google’s safety models, UX improvements, and configuration options.
Moderation isn’t one-size-fits-all. Ninny reflects on the daily challenge of preserving freedom of expression while fostering healthier online spaces.
How Google accounts for cultural differences in defining abuse, with special attention to cross-cultural sensitivities and expectations.
Ninny answers the final wildcard question, revealing her strong interest in the human-in-the-loop side of AI as the front line of moderation and insight.
Hermes thanks Ninny and wraps the episode with a reminder to subscribe and follow along for more insights from leaders in Conversational AI.
Click on any chapter to view its transcript content • Download full transcript
Subscribe to stay up to date on what's happening in conversational and voice AI.