Turk Interpreter Update

The Turk Interpreter has been updated!

Here’s a list of new features

  1. The backend has been updated with a more comprehensive playback and FX chain
  2. The GUI has been modified to include envelope and FX control
  3. It now incorporates x-OSC, using it’s built in accelerometer to trigger sample playback
  4. Two additional turk worker data sets have been added
  5. The person option now contains an all mode, randomly selecting from all of the available data-sets

A demo is available here and code is available here

The First of a Few Miniatures

As previously reported, we brainstormed some great ideas and some possible ideas for what to do with the voice assistants and data archives.

My approach is to make a series of relatively short movements or “miniatures” using said data sets. The data sets were used as the generative material to create the fixed media track which you can listen to below. All sounds you hear in the track come directly from the voice clips and are then further processed for desired sounds and effects. There are no samples or sounds that came from other sources. The intention is to continue generating fixed media under that parameter.

To accompany the fixed media track will be a graphic notation score for performer(s). If you’ve never seen one, here are some examples: https://www.classicfm.com/discover-music/latest/graphic-scores-art-music-pictures/. While I am still in the process of sketching the score, it will be loosely based off a spectrum analysis of the frequencies found in the various voice clips you hear in the fixed media. The idea for instrumentation for this miniature is leaning towards resonant sounds of some type. I hope to post a page from the score soon so please check back.


Turk Interpreter

The Turk Interpreter is a system that plays back voice commands issued by Mechanical Turk workers. This blog post will briefly overview its features and present a short demo.

Turk Worker GUI

  1. Output -> Specifies the output of the turk interpreter. There are two options, human and Computer. Human will play back recordings of voice commands issued by turk workers, computer will read the commands with Mac’s built in text to speech software.
  2. Person -> Specifies the specific turk worker for playback. 
  3. Search -> Allows user to search turk worker data for specific word/phrase. Playback is then refined to results of this search.
  4. Result -> Results of search. Text can be edited to alter READ functionality.
  5. Play -> Plays back random voice command based on search results with varied output depending on output settings.
  6. Read -> Reads whatever is in results window with computer
  7. Stop -> Kills all audio

Turk Worker Demo

The person’s name has been blacked out to preserve their autonomy

Experiments with Alexa’s voice

Working in the Developer Console for Alexa, Audrey worked on ‘hacking’ (or modifying) Alexa’s voice. Through simple commands, it is possible to make the voice whisper, slow down, speed up, or change pitch. It is also possible to change the voice itself, with the opportunity to hear from Ivy, Joanna, Joey, Justin, Kendra, Kimberly, Matthew, Salli. The commands are explained here.

Broadening ideas through sketching

Audrey used sketching as a way to visually map the various ideas we have had and to open new proposals for what a future performance could hold. She also started to collect ideas for points of departure for our design and artistic process. These points include:

  • A lot of AI responses are hard coded
  • People ask many things to voices assistants that are not planned for (not hard coded)
  • Voice assistants are still not very good at understanding human language, generating many bloopers
  • Voice assistants lack context and memory
  • Data is part of an ecosystem of algorithms, data collection, voice detection, voice to text technology, AWS….
  • Corporations benefit from clean and ‘true’ data
  • Voice assistants can’t pretend to be human — instead, how can they honor the reality that they are software.

New Friends and a Little Brainstorming

Hi, all!

It has been a little while since we’ve posted and we have a lot to share! Recently, a few students with interest in this project joined the research group on voice assistants and experimental performance. The following students have backgrounds in music composition, experimental percussion and digital arts and experimental media:

Darcy C.

Gabrielle B.

Esteban A.

James W.

Jonathan R.

While Bonnie and Afroditi were attending and presenting at the Transmediale Festival (a festival and year-round project that draws out new connections between art, culture, and technology), Audrey and the team got together for a brainstorming session in an attempt to identify the trajectory and next steps for this project. We felt like we came up with some great initial ideas and concepts, here are a few: voice assistants and language manipulation, sound manipulation using digital audio workstations, voice assistants within the context of defined spaces.

The ideas are flowing and we’ve got momentum! Can’t wait to share more soon!