/crowd-eval-audio

Primary LanguageJupyter NotebookMIT LicenseMIT

Crowd-Eval-Audio Framework

License

A frontend focused framework to setup and conduct perceptual listening tests on crowdsourced platforms such as Amazon's Mechanical Turk. This project resulted from a need for a lightweight plug and play framework for conducting listening tests w/ minimal setup requirements (no DB installations etc for basic setup).

This framework used in conjunction with the External API on Mechanical Turk provides an configurable, multi-page, intuitive listening test interface for participants.

This project has the following structure. Please see individual directories for instructions to setup for development and deployment.

  • ui/ : Source for the configurable web interface iframed inside Mechanical Turk
  • serverless/ : Source for AWS Lambda. Currently (optionally) used by the web interface to log unique participant ids for Between-Groups experiments.
  • notebooks/ : Set of notebooks used to setup experiments on Mechanical Turk. Also to download responses and programatically pay Mechanical Turk participants etc.

Please cite this work as -

@inproceedings{10.1145/3581641.3584083,
author = {Kamath, Purnima and Li, Zhuoyao and Gupta, Chitralekha and Jaidka, Kokil and Nanayakkara, Suranga and Wyse, Lonce},
title = {Evaluating Descriptive Quality of AI-Generated Audio Using Image-Schemas},
year = {2023},
isbn = {9798400701061},
publisher = {Association for Computing Machinery},
address = {New York, NY, USA},
url = {https://doi.org/10.1145/3581641.3584083},
doi = {10.1145/3581641.3584083},
booktitle = {Proceedings of the 28th International Conference on Intelligent User Interfaces},
pages = {621–632},
numpages = {12},
keywords = {listening tests, audio evaluations, web-based interactions, AI-generated audio},
location = {Sydney, NSW, Australia},
series = {IUI '23}
}