This research project started with a literature review on relevant articles pertaining to large language models (LLM), sociphonetics (the study of how voices affect behaviour), and the general development of trust in the human-computer interaction (HCI) space. It was found that literature was lacking especially in the voice user interface space, where guidelines on how best to design a voice assistant were scarce.

I then started to get a better understanding of voice assistant users through quantitative research. Creating a survey that aligned with an existing scale, the “Trust in Automation” scale then allowed me to conduct a series of data analyses to find any potential correlations to discuss/focus on in the interviews.

For the interviews, I developed a Generative AI-powered voice assistant prototype for participants to directly interact with. I developed an Amazon Alexa Skill, connected to Open AI’s Completions API - imagine Chat GPT but voice based. The conversation design was made on Voiceflow, allowing me to quickly prototype the Skill, using JavaScript to call, receive, and manipulate data from the API.

The interviews were then done in a semi-structured way to allow for an open conversation yet have consistency between participants. Through the interview, it was concluded that people did not really value references and prefer quick concise answers. A lot of the participants blindly trusted the answers coming from the Alexa, even when programmed to provide mistake-filled answers.
