The Sound of Safety: How Modulate Cleans Up Voice Chat & Beyond with Mike Pappas
Podcasts
•
June 17, 2025





If you've been following my recent posts or caught the latest Player Driven podcast episode (where we chatted with Mike Pappas, CEO & Co-founder of Modulate), you know I'm incredibly passionate about creating safer online spaces. It’s a topic that resonates deeply, both professionally and personally, especially as a parent navigating the digital landscape with my own kids. Today, I want to dive a bit deeper into the fascinating world of voice AI and how companies like Modulate are changing the game in online trust and safety.
Beyond Text: Understanding the Nuances of Voice
For years, moderation efforts in online communities, particularly in gaming, were heavily focused on text-based interactions. It made sense – text is easier to log, analyze, and automate. But anyone who's spent time in an online game knows that a huge part of the social experience, and unfortunately, where a lot of negative behavior can occur, happens through voice chat.
This is where Modulate's flagship product, ToxMod, comes in. They've built "pro-social voice intelligence technology" that goes far beyond simply transcribing words. It's about understanding the emotion, nuance, and behavioral dynamics of a conversation. Think about it: saying "I'm going to shoot you in the head" means something entirely different in a game like Call of Duty versus a casual game like Sudoku. ToxMod is designed to grasp these critical contextual differences.
More Than Just Detecting "Bad" – Promoting "Good"!
While often associated with identifying toxicity and harassment (which it does remarkably well ), Modulate's vision for voice intelligence is much broader. They're also working on detecting and promoting positive behavior. Imagine a system that can identify users who are excellent at "diffusing toxicity" or are "really welcoming to a new player and helps coach them in the game". This proactive approach to fostering healthy communities is truly groundbreaking.
The "Vertical Pivot": From Gaming to Fraud Protection
What's really exciting is how Modulate is now applying this powerful technology beyond gaming. Mike mentioned their "broadening of what the tool is focused on" into new verticals, with a significant push into fraud protection.
You might wonder, how does voice moderation relate to financial fraud? Well, think about the rise of deepfakes and increasingly sophisticated scams. Modulate's technology can detect "falsified audio" and other "signs of syntheticness". But here's the crucial part: they don't just flag any synthetic audio. They look for these signs coupled with other problematic behaviors that indicate an intent to manipulate or defraud. This is incredibly important because, as Mike pointed out, synthetic audio has legitimate uses, such as for individuals who can't speak or those with voice dysphoria.
Their entry into this space was organic. After demonstrating success in detecting nuances in Call of Duty, other industries realized that if Modulate could handle that complexity, they could certainly help with identifying issues in areas like call centers. In fact, one major delivery platform found that Modulate's toxicity detection alone was identifying five times more fraud than their previous tools. That's a massive impact!
Transparency and Community Empowerment
One of the aspects I deeply appreciate about Modulate's approach is their commitment to transparency. They don't dictate what's permissible in a game; instead, they work with studios to understand their code of conduct and then tune their system accordingly. The ultimate decision-making power remains with the studio.
Equally important is the direct engagement with the community. Studios using ToxMod are encouraged to openly communicate why they're moderating voice chat, not just that they are. This proactive messaging helps manage expectations and, crucially, empowers players to be part of the solution by providing feedback and helping to iterate on the system. This collaborative approach builds genuine trust and helps cultivate a truly self-sustaining, positive community.
The Human Element Remains Key
While AI is powerful, it's not magic. Mike emphasized that AI can be "about as perceptive as a very well-trained human being". It helps less experienced agents recognize signs that a trained operator would. This highlights the ongoing importance of the "human in the loop". AI augments human capabilities, allowing teams to focus on the more nuanced cases and appeals, ensuring fairness and continuous improvement.
The journey towards safer, more inclusive online spaces is ongoing, but with innovative technologies like Modulate's ToxMod, we're making significant strides. It's exciting to see how voice AI is not only protecting users but also fostering stronger, healthier communities across various industries.
What are your thoughts on voice AI and online safety? Share them in the comments below!
Beyond Text: Understanding the Nuances of Voice
For years, moderation efforts in online communities, particularly in gaming, were heavily focused on text-based interactions. It made sense – text is easier to log, analyze, and automate. But anyone who's spent time in an online game knows that a huge part of the social experience, and unfortunately, where a lot of negative behavior can occur, happens through voice chat.
This is where Modulate's flagship product, ToxMod, comes in. They've built "pro-social voice intelligence technology" that goes far beyond simply transcribing words. It's about understanding the emotion, nuance, and behavioral dynamics of a conversation. Think about it: saying "I'm going to shoot you in the head" means something entirely different in a game like Call of Duty versus a casual game like Sudoku. ToxMod is designed to grasp these critical contextual differences.
More Than Just Detecting "Bad" – Promoting "Good"!
While often associated with identifying toxicity and harassment (which it does remarkably well ), Modulate's vision for voice intelligence is much broader. They're also working on detecting and promoting positive behavior. Imagine a system that can identify users who are excellent at "diffusing toxicity" or are "really welcoming to a new player and helps coach them in the game". This proactive approach to fostering healthy communities is truly groundbreaking.
The "Vertical Pivot": From Gaming to Fraud Protection
What's really exciting is how Modulate is now applying this powerful technology beyond gaming. Mike mentioned their "broadening of what the tool is focused on" into new verticals, with a significant push into fraud protection.
You might wonder, how does voice moderation relate to financial fraud? Well, think about the rise of deepfakes and increasingly sophisticated scams. Modulate's technology can detect "falsified audio" and other "signs of syntheticness". But here's the crucial part: they don't just flag any synthetic audio. They look for these signs coupled with other problematic behaviors that indicate an intent to manipulate or defraud. This is incredibly important because, as Mike pointed out, synthetic audio has legitimate uses, such as for individuals who can't speak or those with voice dysphoria.
Their entry into this space was organic. After demonstrating success in detecting nuances in Call of Duty, other industries realized that if Modulate could handle that complexity, they could certainly help with identifying issues in areas like call centers. In fact, one major delivery platform found that Modulate's toxicity detection alone was identifying five times more fraud than their previous tools. That's a massive impact!
Transparency and Community Empowerment
One of the aspects I deeply appreciate about Modulate's approach is their commitment to transparency. They don't dictate what's permissible in a game; instead, they work with studios to understand their code of conduct and then tune their system accordingly. The ultimate decision-making power remains with the studio.
Equally important is the direct engagement with the community. Studios using ToxMod are encouraged to openly communicate why they're moderating voice chat, not just that they are. This proactive messaging helps manage expectations and, crucially, empowers players to be part of the solution by providing feedback and helping to iterate on the system. This collaborative approach builds genuine trust and helps cultivate a truly self-sustaining, positive community.
The Human Element Remains Key
While AI is powerful, it's not magic. Mike emphasized that AI can be "about as perceptive as a very well-trained human being". It helps less experienced agents recognize signs that a trained operator would. This highlights the ongoing importance of the "human in the loop". AI augments human capabilities, allowing teams to focus on the more nuanced cases and appeals, ensuring fairness and continuous improvement.
The journey towards safer, more inclusive online spaces is ongoing, but with innovative technologies like Modulate's ToxMod, we're making significant strides. It's exciting to see how voice AI is not only protecting users but also fostering stronger, healthier communities across various industries.
What are your thoughts on voice AI and online safety? Share them in the comments below!
Beyond Text: Understanding the Nuances of Voice
For years, moderation efforts in online communities, particularly in gaming, were heavily focused on text-based interactions. It made sense – text is easier to log, analyze, and automate. But anyone who's spent time in an online game knows that a huge part of the social experience, and unfortunately, where a lot of negative behavior can occur, happens through voice chat.
This is where Modulate's flagship product, ToxMod, comes in. They've built "pro-social voice intelligence technology" that goes far beyond simply transcribing words. It's about understanding the emotion, nuance, and behavioral dynamics of a conversation. Think about it: saying "I'm going to shoot you in the head" means something entirely different in a game like Call of Duty versus a casual game like Sudoku. ToxMod is designed to grasp these critical contextual differences.
More Than Just Detecting "Bad" – Promoting "Good"!
While often associated with identifying toxicity and harassment (which it does remarkably well ), Modulate's vision for voice intelligence is much broader. They're also working on detecting and promoting positive behavior. Imagine a system that can identify users who are excellent at "diffusing toxicity" or are "really welcoming to a new player and helps coach them in the game". This proactive approach to fostering healthy communities is truly groundbreaking.
The "Vertical Pivot": From Gaming to Fraud Protection
What's really exciting is how Modulate is now applying this powerful technology beyond gaming. Mike mentioned their "broadening of what the tool is focused on" into new verticals, with a significant push into fraud protection.
You might wonder, how does voice moderation relate to financial fraud? Well, think about the rise of deepfakes and increasingly sophisticated scams. Modulate's technology can detect "falsified audio" and other "signs of syntheticness". But here's the crucial part: they don't just flag any synthetic audio. They look for these signs coupled with other problematic behaviors that indicate an intent to manipulate or defraud. This is incredibly important because, as Mike pointed out, synthetic audio has legitimate uses, such as for individuals who can't speak or those with voice dysphoria.
Their entry into this space was organic. After demonstrating success in detecting nuances in Call of Duty, other industries realized that if Modulate could handle that complexity, they could certainly help with identifying issues in areas like call centers. In fact, one major delivery platform found that Modulate's toxicity detection alone was identifying five times more fraud than their previous tools. That's a massive impact!
Transparency and Community Empowerment
One of the aspects I deeply appreciate about Modulate's approach is their commitment to transparency. They don't dictate what's permissible in a game; instead, they work with studios to understand their code of conduct and then tune their system accordingly. The ultimate decision-making power remains with the studio.
Equally important is the direct engagement with the community. Studios using ToxMod are encouraged to openly communicate why they're moderating voice chat, not just that they are. This proactive messaging helps manage expectations and, crucially, empowers players to be part of the solution by providing feedback and helping to iterate on the system. This collaborative approach builds genuine trust and helps cultivate a truly self-sustaining, positive community.
The Human Element Remains Key
While AI is powerful, it's not magic. Mike emphasized that AI can be "about as perceptive as a very well-trained human being". It helps less experienced agents recognize signs that a trained operator would. This highlights the ongoing importance of the "human in the loop". AI augments human capabilities, allowing teams to focus on the more nuanced cases and appeals, ensuring fairness and continuous improvement.
The journey towards safer, more inclusive online spaces is ongoing, but with innovative technologies like Modulate's ToxMod, we're making significant strides. It's exciting to see how voice AI is not only protecting users but also fostering stronger, healthier communities across various industries.
What are your thoughts on voice AI and online safety? Share them in the comments below!
Share
Copy link
Share
Copy link
Share
Copy link