ChristiÂna Y. Tzeng, San JosĂ© State University
Full TranÂscript:
ChristiÂna Y. Tzeng:
All right. Thank you, Rachel, for the introÂducÂtion at the beginÂning of the sesÂsion and to both you and Joshua for gathÂerÂing us in this space. I also want to say thank you upfront for everyÂone who is still here at the last talk of the day. I am excitÂed to share about my expeÂriÂence conÂductÂing speech perÂcepÂtion experÂiÂments online, highÂlightÂing the powÂer for these online experÂiÂments to democÂraÂtize sciÂence and teachÂing. I’ll aim to achieve two objecÂtives in my talk today. The first is to share some findÂings that add to what we now know is a growÂing piece of eviÂdence that online speech perÂcepÂtion experÂiÂments are highÂly effiÂcient and do yield robust data. The secÂond objecÂtive is to share some thoughts on how online experÂiÂments, more broadÂly, can make sciÂence more accesÂsiÂble for both researchers and participants.
ChristiÂna Y. Tzeng:
In my work, I study how we, as lisÂtenÂers, overÂcome the enorÂmous amount of variÂaÂtion that we encounter when we lisÂten to difÂferÂent voicÂes and utterÂances. Our experÂiÂments typÂiÂcalÂly require parÂticÂiÂpants to lisÂten to audiÂtoÂry stimÂuli and make subÂseÂquent responsÂes on a comÂputÂer to each one. For in-perÂson or in-lab experÂiÂments, this would typÂiÂcalÂly require what’s picÂtured on the left: a sound-attenÂuÂatÂed, disÂtracÂtion-free booth, high-qualÂiÂty headÂphones, and speÂcialÂized softÂware, as well as hardware.
ChristiÂna Y. Tzeng:
As a disÂclaimer, I have to state that my first real dive into the world of online experÂiÂments was in late 2019, which makes me a relÂaÂtiveÂly novÂel user of these online experÂiÂmenÂtal methÂods, but this is when I startÂed to wonÂder, “Is such a highÂly conÂtrolled lisÂtenÂing enviÂronÂment realÂly necessary?”
ChristiÂna Y. Tzeng:
In the interÂest of achievÂing this first objecÂtive, I’d like to share what are now pubÂlished findÂings from my first forÂay into the online experÂiÂment world. This is work done in colÂlabÂoÂraÂtion with my colÂleagues, Dr. Lynne Nygaard and Dr. Rachel Theodore, where we examÂined the time course of a pheÂnomÂeÂnon called lexÂiÂcalÂly guidÂed perÂcepÂtuÂal learning.
ChristiÂna Y. Tzeng:
We know that lisÂtenÂers use a whole host of cues to map the acoustics of the speech sigÂnal onto linÂguisÂtic units. One of these cues is lexÂiÂcal knowlÂedge. ImagÂine hearÂing a fricaÂtive sound that’s between an S and an SH sound. If that ambiguÂous sound is embedÂded into this word on the left, the lisÂtenÂer hears that sound as an S as in dinosaur. But if that same ambiguÂous sound is instead embedÂded in the word on the right, the lisÂtenÂer hears that sound instead as an SH as in effiÂcient. But if lisÂtenÂers are exposed to these ambiguÂous sounds in staÂble lexÂiÂcal conÂtexts, that bias them to hear either S or SH sound.
ChristiÂna Y. Tzeng:
What we then see are changes in the lisÂtenÂer’s repÂreÂsenÂtaÂtions of their S and SH catÂeÂgoÂry. These changes in sound catÂeÂgoÂry repÂreÂsenÂtaÂtion are what we call lexÂiÂcalÂly guidÂed perÂcepÂtuÂal learnÂing. In both the online and in-perÂson verÂsions of this task, the lexÂiÂcalÂly guidÂed perÂcepÂtuÂal learnÂing parÂaÂdigm takes about 20 minÂutes to comÂplete. So here, lisÂtenÂers comÂplete an expoÂsure phase folÂlowed by a test phase. And in the expoÂsure phase, they comÂplete a lexÂiÂcal deciÂsion task where they hear an ambiguÂous sound such as a fricaÂtive between S and SH. One group hears this ambiguÂous sound that’s embedÂded in words, biasÂing them to hear it as an S, whereÂas anothÂer group is biased to hear that same sound as an SH. So after expoÂsure, the lisÂtenÂers comÂplete a phoÂnetÂic catÂeÂgoÂrizaÂtion task where they idenÂtiÂfy ambiguÂous sounds on a non-word conÂtinÂuÂum here, either as asi or ashi.
ChristiÂna Y. Tzeng:
We drew our samÂples from ProÂlifÂic and exeÂcutÂed the experÂiÂments in GorilÂla. We comÂpletÂed a total of six experÂiÂments in this pubÂliÂcaÂtion, but in the interÂest of time, I’ll share the findÂings from one. What will appear here are the results of the phoÂnetÂic catÂeÂgoÂrizaÂtion task at test, whereÂupon hearÂing ambiguÂous sound on the asi/ashi conÂtinÂuÂum, we meaÂsured the likeÂliÂhood that parÂticÂiÂpants heard those sounds as asi. Here, we see robust eviÂdence for lexÂiÂcalÂly guidÂed perÂcepÂtuÂal learnÂing. As lisÂtenÂers, we’re more likeÂly to hear the ambiguÂous sounds as asi when they were biased to hear S durÂing expoÂsure indiÂcatÂed by the red line, then when they were biased to hear the sounds as SH durÂing expoÂsure shown here by the green line.
ChristiÂna Y. Tzeng:
To showÂcase the high levÂel of data qualÂiÂty that we see at the indiÂvidÂual levÂel, here are sepÂaÂrate plots for each of the 70 parÂticÂiÂpants at test where we can see the expectÂed psyÂchoÂmeÂtÂric curves for every sinÂgle parÂticÂiÂpant. We only excludÂed 5% of our parÂticÂiÂpants across the six experÂiÂments due to failÂure to perÂform the task. We did have to exclude 16% of the total numÂber of parÂticÂiÂpants due to failÂure to pass the woods at all, headÂphone check that Dr. Theodore described at the beginÂning of the sesÂsion. But this was a small price to pay, givÂen the speed of data colÂlecÂtion. So for examÂple, we colÂlectÂed data from the 70 parÂticÂiÂpants preÂsentÂed in ExperÂiÂment 1 in under a sinÂgle hour.
ChristiÂna Y. Tzeng:
I hope what I’ve shared has supÂportÂed the idea that online speech perÂcepÂtion experÂiÂments are highÂly effiÂcient and yields robust findÂings even with audiÂtoÂry tasks that require fine-grained phoÂnetÂic disÂcrimÂiÂnaÂtions like the one I presented.
ChristiÂna Y. Tzeng:
I now want to turn to the idea that online experÂiÂments can proÂvide us with two things in parÂticÂuÂlar: access to a largÂer and more diverse pool of parÂticÂiÂpants and also more user-friendÂly experÂiÂment buildÂing interÂfaces for our stuÂdents and research mentees.
ChristiÂna Y. Tzeng:
This is the figÂure I showed earÂliÂer. We repliÂcatÂed the findÂing with anothÂer end of 70 parÂticÂiÂpants using a secÂond stimÂuÂlus set shown here on the right, meanÂing we ran a total of 150 parÂticÂiÂpants withÂin the span of about an hour and a half, which using in-perÂson methÂods would have takÂen us weeks or even months.
ChristiÂna Y. Tzeng:
For his masÂter’s theÂsis, one of my stuÂdent colÂlabÂoÂraÂtors, UlisÂes QuinÂtero, is interÂestÂed in recruitÂing parÂticÂiÂpants who speak EngÂlish and a secÂond lanÂguage. So in ProÂlifÂic, if we use our stanÂdard incluÂsion criÂteÂria, includÂing this criÂteÂriÂon of speakÂing EngÂlish plus anothÂer lanÂguage, we autoÂmatÂiÂcalÂly have access to over 3,000 parÂticÂiÂpants, which is magÂniÂtudes greater than what we would have access to using in-perÂson methÂods. For his underÂgradÂuÂate honÂors theÂsis, Justin Au built a talkÂer ID task in GorilÂla on his own using priÂmarÂiÂly the tutoÂrÂiÂal supÂport that is on GorilÂla’s webÂsite as a guide.
ChristiÂna Y. Tzeng:
And by addressÂing the two quesÂtions about audiÂtoÂry research more broadÂly that Rachel shared at the beginÂning of the sesÂsion, the first is, “What do you think is the biggest chalÂlenge for audiÂtoÂry research online, and how do you overÂcome it?” As Jason menÂtioned, due to the panÂdemÂic, we have all been forced to some extent to embrace online methÂods more readÂiÂly, but I think we are still very much in the process of estabÂlishÂing both the validÂiÂty and the reliÂaÂbilÂiÂty of these methÂods. And one way for us to do this is to run online and in-perÂson experÂiÂments in parÂalÂlel so that we, not just as indiÂvidÂual researchers but as a field, can be reasÂsured that our tasks can be sucÂcessÂfulÂly transÂferred across these difÂferÂent platforms.
ChristiÂna Y. Tzeng:
And the secÂond quesÂtion, “What can audiÂtoÂry research gain most from online methÂods?” My take on this is that, with how quickÂly, we can colÂlect data from a whole numÂber of difÂferÂent popÂuÂlaÂtions. We’ve essenÂtialÂly elimÂiÂnatÂed the data colÂlecÂtion botÂtleÂneck. AdaptÂing in-perÂson experÂiÂments to the online world takes a lot of triÂal and error, and I’m still very much in that learnÂing phase, but I think that the reducÂtion of this botÂtleÂneck drasÂtiÂcalÂly changes the pace of audiÂtoÂry research and sciÂence more broadly.
ChristiÂna Y. Tzeng:
With that, I’d like to extend my gratÂiÂtude to my recent colÂlabÂoÂraÂtors as well as to all of you for your attenÂtion. I look forÂward to your quesÂtions and comments.
Rachel Theodore:
ExcelÂlent, ChristiÂna. Thank you so much for those realÂly careÂful thoughts. QuesÂtions. Yeah, here’s one for you, ChristiÂna. “I was wonÂderÂing if, in your work, you’ve observed the use of difÂferÂent expoÂsure phase methÂods besides lexÂiÂcal deciÂsion in an online world. How’s the stoÂry lisÂtenÂing closed senÂtences and if you’ve noticed any difÂferÂences at test as a funcÂtion of those expoÂsure phase methods?”
ChristiÂna Y. Tzeng:
Thanks for that quesÂtion. So again, comÂing back to this disÂclaimer that I’m a relÂaÂtiveÂly novÂel user of online methodÂolÂoÂgy in genÂerÂal for audiÂtoÂry research, we’ve only done some pilot work using othÂer kinds of expoÂsure methodÂolÂoÂgy. We’re in the process of pilotÂing a talkÂer idenÂtiÂfiÂcaÂtion task, where durÂing the expoÂsure phase, lisÂtenÂers well hear utterÂances spoÂken by speÂcifÂic talkÂers and have to indiÂcate which talkÂer they think they’re hearÂing with the ultiÂmate goal being able to idenÂtiÂfy the difÂferÂent voicÂes in the task. And so far, we haven’t seen any kind of noticeÂable difÂferÂence in perÂforÂmance for in-perÂsonÂ/in-lab verÂsions of that and online verÂsions. What we do notice is that, someÂtimes, parÂticÂiÂpants will take self-inflictÂed breaks. And so one lesÂson we’ve learned is that in addiÂtion to keepÂing the task relÂaÂtiveÂly short, we will build in some breaks so that they’re not leavÂing the comÂputÂer for an extendÂed periÂod of time. But the short response to that quesÂtion is at least with talkÂer idenÂtiÂfiÂcaÂtion tasks and this lexÂiÂcalÂly guidÂed perÂcepÂtuÂal learnÂing task, we haven’t seen reaÂsons to not transÂfer these into the online world.


