File Download

There are no files associated with this item.

Supplementary

Conference Paper: Bottom-up learning of a phonetic system using an autoencoder

TitleBottom-up learning of a phonetic system using an autoencoder
Authors
Issue Date26-May-2023
Citation
Hanyang International Symposium on Phonetics & Cognitive Sciences of Language 2023 (HISPhonCog 2023), 2023 How to Cite?
Abstract

This study investigates an autoencoder model’s effectiveness in learning phonemes and distinctive features from unsegmented, non-transcribed wave data, mirroring infants’ early language acquisition stages. The results of the experiment on Mandarin and English show that features can be learned through repeated projection of sounds to a hidden space and reconstruction back to their original form without prior knowledge of segmentation. The acquired knowledge is represented by different distributions in the hidden space. The model was capable of clustering segments of the same phoneme and projecting different phonemes to separate regions in the hidden space. However, the model struggled to cluster allophones closely together, which may indicate the boundary between the contributions of bottom-up and top-down information. Overall, the study suggests that sound knowledge can be learned to a certain degree through unsupervised learning techniques that do not require labeled data or prior knowledge of segmentation, and that this approach may provide insights into the early stages of human language acquisition.


Persistent Identifierhttp://hdl.handle.net/10722/329225

 

DC FieldValueLanguage
dc.contributor.authorTan, Frank Lihui-
dc.contributor.authorDo, Youngah-
dc.date.accessioned2023-08-09T02:19:19Z-
dc.date.available2023-08-09T02:19:19Z-
dc.date.issued2023-05-26-
dc.identifier.citationHanyang International Symposium on Phonetics & Cognitive Sciences of Language 2023 (HISPhonCog 2023), 2023-
dc.identifier.urihttp://hdl.handle.net/10722/329225-
dc.description.abstract<p>This study investigates an autoencoder model’s effectiveness in learning phonemes and distinctive features from unsegmented, non-transcribed wave data, mirroring infants’ early language acquisition stages. The results of the experiment on Mandarin and English show that features can be learned through repeated projection of sounds to a hidden space and reconstruction back to their original form without prior knowledge of segmentation. The acquired knowledge is represented by different distributions in the hidden space. The model was capable of clustering segments of the same phoneme and projecting different phonemes to separate regions in the hidden space. However, the model struggled to cluster allophones closely together, which may indicate the boundary between the contributions of bottom-up and top-down information. Overall, the study suggests that sound knowledge can be learned to a certain degree through unsupervised learning techniques that do not require labeled data or prior knowledge of segmentation, and that this approach may provide insights into the early stages of human language acquisition.</p>-
dc.languageeng-
dc.relation.ispartofHanyang International Symposium on Phonetics & Cognitive Sciences of Language 2023 (HISPhonCog 2023) (26/05/2023-27/05/2023, Hanyang University, Seoul, Korea)-
dc.rightsThis work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.-
dc.titleBottom-up learning of a phonetic system using an autoencoder-
dc.typeConference_Paper-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats