/voices-and-voids

A net art project that explores how we might reclaim, transcode, and embody voice assistant interactions and data. Developed in collaboration with Audrey Desjardins and many other amazing researchers, artists, and musicians.

Primary LanguageJavaScriptMIT LicenseMIT

Voices and Voids

echo speaking to each other

visit site


Explorations around how we might reclaim, transcode, and embody voice assistant interactions and data

Responding to current concerns about the ubiquity of voice assistants, this artistic research focuses on building a series of performative artifacts that aim to challenge AI and ML technologies, and to examine automation through the prism of “ghost work” that constantly support these systems. By allowing AI agents to listen to our most private conversations, we become receptive to this mediated care, while forgetting or ignoring how much these automated interactions have been pre-scripted. While these interactions cultivate a sense of familiarization with the non-human, they also corroborate the impact of Late Capitalism and the Anthropocene. Within these contradictions, we see an opportunity to reclaim, examine, and ultimately transcode this data through an interdisciplinary performance project, by developing embodied experiments using a combination of design, data-driven art, cyber crafts, found-object, and traditional percussion instruments, spoken word, and movement.

For more about our process, take a look at our blog at blog.voicesandvoids.net

Credits

Project leads:
Audrey Desjardins, Afroditi Psarra, Bonnie Whiting

Artists in residence

Danny Clay, Laura Devendorf, Kate Sicchio, Jordan Wirfs Brock, Yvonne Wu

Web development:
Amanda Yeh

Collaborators:
Esteban Yosef Agosin, Gabrielle Benabdallah, Darcy Copeland, Jonathan Rodriguez, James Wenlock

Funded by:
Mellon Foundation, College of Arts and Sciences - University of Washington