{"_id":"4gqwBNRhqMCiN9eho","bibbaseid":"arakawa-kashino-takamichi-verhulst-inami-digitalspeechmakeupvoiceconversionbasedalteredauditoryfeedbackfortransformingselfrepresentation-2021","author_short":["Arakawa, R.","Kashino, Z.","Takamichi, S.","Verhulst, A.","Inami, M."],"bibdata":{"bibtype":"inproceedings","type":"inproceedings","address":"New York, NY, USA","title":"Digital Speech Makeup: Voice Conversion Based Altered Auditory Feedback for Transforming Self-Representation","isbn":"978-1-4503-8481-0","shorttitle":"Digital Speech Makeup","url":"https://doi.org/10.1145/3462244.3479934","abstract":"Makeup (i.e., cosmetics) has long been used to transform not only one’s appearance but also their self-representation. Previous studies have demonstrated that visual transformations can induce a variety of effects on self-representation. Herein, we introduce Digital Speech Makeup (DSM), the novel concept of using voice conversion (VC) based auditory feedback to transform human self-representation. We implemented a proof-of-concept system that leverages a state-of-the-art algorithm for near real-time VC and bone-conduction headphones for resolving speech disruptions caused by delayed auditory feedback. Our user study confirmed that conversing for a few dozen minutes using the system influenced participants’ speech ownership and implicit bias. Furthermore, we reviewed the participants’ comments about the experience of DSM and gained additional qualitative insight into possible future directions for the concept. Our work represents the first step towards utilizing VC to design various interpersonal interactions, centered on influencing the users’ psychological state.","urldate":"2021-12-08","booktitle":"Proceedings of the 2021 International Conference on Multimodal Interaction","publisher":"Association for Computing Machinery","author":[{"propositions":[],"lastnames":["Arakawa"],"firstnames":["Riku"],"suffixes":[]},{"propositions":[],"lastnames":["Kashino"],"firstnames":["Zendai"],"suffixes":[]},{"propositions":[],"lastnames":["Takamichi"],"firstnames":["Shinnosuke"],"suffixes":[]},{"propositions":[],"lastnames":["Verhulst"],"firstnames":["Adrien"],"suffixes":[]},{"propositions":[],"lastnames":["Inami"],"firstnames":["Masahiko"],"suffixes":[]}],"month":"October","year":"2021","keywords":"auditory feedback, self-representation, speech transformation, voice conversion","pages":"159–167","bibtex":"@inproceedings{arakawa_digital_2021,\n\taddress = {New York, NY, USA},\n\ttitle = {Digital {Speech} {Makeup}: {Voice} {Conversion} {Based} {Altered} {Auditory} {Feedback} for {Transforming} {Self}-{Representation}},\n\tisbn = {978-1-4503-8481-0},\n\tshorttitle = {Digital {Speech} {Makeup}},\n\turl = {https://doi.org/10.1145/3462244.3479934},\n\tabstract = {Makeup (i.e., cosmetics) has long been used to transform not only one’s appearance but also their self-representation. Previous studies have demonstrated that visual transformations can induce a variety of effects on self-representation. Herein, we introduce Digital Speech Makeup (DSM), the novel concept of using voice conversion (VC) based auditory feedback to transform human self-representation. We implemented a proof-of-concept system that leverages a state-of-the-art algorithm for near real-time VC and bone-conduction headphones for resolving speech disruptions caused by delayed auditory feedback. Our user study confirmed that conversing for a few dozen minutes using the system influenced participants’ speech ownership and implicit bias. Furthermore, we reviewed the participants’ comments about the experience of DSM and gained additional qualitative insight into possible future directions for the concept. Our work represents the first step towards utilizing VC to design various interpersonal interactions, centered on influencing the users’ psychological state.},\n\turldate = {2021-12-08},\n\tbooktitle = {Proceedings of the 2021 {International} {Conference} on {Multimodal} {Interaction}},\n\tpublisher = {Association for Computing Machinery},\n\tauthor = {Arakawa, Riku and Kashino, Zendai and Takamichi, Shinnosuke and Verhulst, Adrien and Inami, Masahiko},\n\tmonth = oct,\n\tyear = {2021},\n\tkeywords = {auditory feedback, self-representation, speech transformation, voice conversion},\n\tpages = {159--167},\n}\n\n","author_short":["Arakawa, R.","Kashino, Z.","Takamichi, S.","Verhulst, A.","Inami, M."],"key":"arakawa_digital_2021","id":"arakawa_digital_2021","bibbaseid":"arakawa-kashino-takamichi-verhulst-inami-digitalspeechmakeupvoiceconversionbasedalteredauditoryfeedbackfortransformingselfrepresentation-2021","role":"author","urls":{"Paper":"https://doi.org/10.1145/3462244.3479934"},"keyword":["auditory feedback","self-representation","speech transformation","voice conversion"],"metadata":{"authorlinks":{}}},"bibtype":"inproceedings","biburl":"https://api.zotero.org/users/2805047/collections/ICBS3JG5/items?key=vu5JAgEl9brxvxihKOTCpenn&format=bibtex&limit=100","dataSources":["ypSzwLqyQJEZpq9yK"],"keywords":["auditory feedback","self-representation","speech transformation","voice conversion"],"search_terms":["digital","speech","makeup","voice","conversion","based","altered","auditory","feedback","transforming","self","representation","arakawa","kashino","takamichi","verhulst","inami"],"title":"Digital Speech Makeup: Voice Conversion Based Altered Auditory Feedback for Transforming Self-Representation","year":2021}