Publication: Skin in the Game: Modulate AI and Addressing the Legal and Ethical Challenges of Voice Skin Technology
Open/View Files
Date
2020-01-06
Authors
Published Version
Published Version
Journal Title
Journal ISSN
Volume Title
Publisher
Berkman Klein Center for Internet & Society
The Harvard community has made this article openly available. Please share how this access benefits you.
Citation
Gordon, Rachel, and Ryan Budish. 2020. "Skin in the Game: Modulate AI and Addressing the Legal and Ethical Challenges of Voice Skin Technology." BKC Policy Practice on AI, Berkman Klein Center for Internet & Society.
Research Data
Abstract
This educational toolkit from the BKC Policy Practice on AI includes a case study, a teaching note, and a background primer. Collectively, they comprise a toolkit that can illuminate some of the challenges in moving from AI principles to practice. The case study, Skin in the Game: Modulate AI and Addressing the Legal and Ethical Challenges of Voice Skin Technology, follows MIT alums and friends Carter Huffman and Mike Pappas, who co-founded Modulate in 2017 to commercialize their technology for creating synthetic voice skins. By innovatively applying concepts from artificial intelligence systems called Generative Adversarial Networks (GANs), the two men had created a novel approach to make one voice sound like another in real time. Although Modulate had extremely limited human and financial resources, Huffman and Pappas wanted to ensure that this technology, with its ability to match the timbre of almost any individual on the planet, would not be misused. How could they simultaneously push Modulate forward, maintain its technological and competitive edges, and make their investors happy, while also upholding a code of ethics? For a tiny company like Modulate, what did this look like?
Written by Rachel Gordon, Research Associate, Teaching Learning and Curriculum Solutions, Harvard Law School Library, and Ryan Budish, Assistant Research Director, Berkman Klein Center for Internet & Society at Harvard University, this case was developed as a basis for discussion in educational and training environments. It is intended to spark discussions among students as they put themselves in the place of the Modulate co-founders, and explore the conflicts and tensions that they faced. The case is not an endorsement of any one approach or business, but instead highlights the complex and dynamic challenges that AI ethics can present in the real world.
Description
Other Available Sources
Keywords
Terms of Use
This article is made available under the terms and conditions applicable to Other Posted Material (LAA), as set forth at Terms of Service