2024-03-28T09:27:19Z
https://jcaa.caa-aca.ca/index.php/jcaa/oai
oai:jcaa.caa-aca.ca:article/2598
2013-04-19T04:53:52Z
jcaa:ART-SPC
nmb a2200000Iu 4500
2291-1391
0711-6659
dc
An Index for Quantifying Tongue Curvature
Stolar, Stade
Gick, Bryan
University of British Columbia, Dept. of Linguistics, Totem Field Studios
2613 West Mall, Vancouver, British Columbia, Canada, V6T 1Z4
This study develops a method of quantifying tongue curvature by modeling the shape of the tongue surface in any anatomical plane using a polynomial approximation. In a validation experiment, the curvature indices of English vowel and consonant sounds were calculated across ten native speakers’ productions based on midsagittal ultrasound images of the tongue. Indices confirm substantially higher curvature values for liquids /r/ and /l/ than for all other sounds in the inventory. This method is both more generalized and less dependent upon fixed locations than previous methods, and provides a simple, powerful metric for evaluating shape complexity with applications in areas such as motor development and aeroacoustics.
Canadian Acoustical Association / Association canadienne d'acoustique
2013-03-01 00:00:00
application/pdf
application/pdf
https://jcaa.caa-aca.ca/index.php/jcaa/article/view/2598
Canadian Acoustics; Vol. 41 No. 1 (2013)
eng
Copyright (c)
oai:jcaa.caa-aca.ca:article/2602
2014-03-08T18:37:23Z
jcaa:ART-SPC
nmb a2200000Iu 4500
"131204 2013 eng "
2291-1391
0711-6659
dc
Perceptual Integration of Visual Evidence of the Airstream from Aspirated Stops
Mayer, Connor
University of British Columbia
Gick, Bryan
University of British Columbia
Haskins Laboratories http://www.linguistics.ubc.ca/people/gick
Weigel, Tamra
University of British Columbia
Whalen, Douglas H
Haskins Laboratories
City University of New York http://www.haskins.yale.edu/staff/whalen.html
This study investigates whether indirect visual evidence of aspiration can influence speech perception as previously found for tactile information. Participants were shown video of a speaker producing the sequence "pom" and "bomb" in a noisy setting. In some tokens, a candle was visibly perturbed by aspiration. All participants were more likely to correctly identify “pom” and incorrectly identify “bomb” in the presence of visible perturbation, indicating that perceptual integration was taking place. This effect was stronger for participants who reported being consciously aware of the candle as a predictor. This indicates that ambient information can be incorporated in speech perception even when presented via an indirect modality, and that active attention can amplify this effect.
Canadian Acoustical Association / Association canadienne d'acoustique
2014-01-30 00:00:00
application/pdf
https://jcaa.caa-aca.ca/index.php/jcaa/article/view/2602
Canadian Acoustics; Vol. 41 No. 3 (2013)
eng
Copyright (c)
oai:jcaa.caa-aca.ca:article/2603
2017-04-03T19:06:39Z
jcaa:ART-SPC
nmb a2200000Iu 4500
"170403 2017 eng "
2291-1391
0711-6659
dc
Ultrasound Study of Emphatics, Uvulars, Pharyngeals and Laryngeals in Three Arabic Dialects
Al-solami, Majed Abdullah
PhD student, University of Toronto
ifferent studies show different and to some extent inconsistent mechanisms for the articulation of Arabic uvular, pharyngeal and laryngeal sounds, i.e. gutturals, and emphatic sounds. The current study uses ultrasound imaging of the tongue to examine the articulations involved in guttural and emphatic sounds in three Arabic dialects, Egyptian, Saudi and Palestinian. This investigation attempts to answer the question: what are the tongue shapes during the production of both inherently retracted /?, ?, ?, ?, q/ and secondarily retracted /ð?, s?, t?, d?/ sounds. Articulatory results indicate that these sounds are produced with different tongue root and tongue dorsum retraction mechanisms. Pharyngeals are articulated with tongue root retraction and statistically do not involve significant tongue dorsum retraction. Uvulars and emphatics show tongue root and tongue dorsum retraction with inconsistent tongue root retraction. Laryngeals do not show any significant tongue retraction.
Keywords: ultrasound, Arabic dialects, emphatics, uvulars, pharyngeals, laryngeals
Canadian Acoustical Association / Association canadienne d'acoustique
2017-04-03 12:06:39
application/pdf
application/pdf
https://jcaa.caa-aca.ca/index.php/jcaa/article/view/2603
Canadian Acoustics; Vol. 45 No. 1 (2017)
eng
Copyright (c) 2017 Majed Abdullah Alsolami
oai:jcaa.caa-aca.ca:article/2654
2015-04-01T23:01:10Z
jcaa:ART-SPC
nmb a2200000Iu 4500
"150401 2015 eng "
2291-1391
0711-6659
dc
Validation of a Finite Element Code for a Continuum Model of Vocal Fold Vibration Under the Influence of a Sessile Polyp
Greiss, Raymond
Carleton University
Rocha, Joana
Carleton University
Matida, Edgar
Carleton University
Vocal fold vibration has been extensively investigated using numerical simulation through the use of lumped element models, and more recently, through the use of finite element continuum models. Finite element models offer the ability to analyze the effects of detailed and complex geometric models, allowing for the study of the influence of pathologies and phonosurgery on the process of phonation. The present study details the development of a finite element code of vocal fold vibration and a continuum model of a vocal fold with a sessile polyp. The capability of the code to capture major structural vibration trends are illustrated through a validation process, wherein previously explored models are replicated, and computed results are sub- sequently compared to gauge the code’s efficacy. An overview of literature pertinent to the modeling of vocal fold polyps is presented, followed by the discussion of the creation of a continuum model of a vocal fold affected by a sessile polyp. This pathological model is parameterized according to the size and position of the polyp, and trends are explored based on varying these parameters. Polyp size is concluded to have a more profound influence on the fundamental frequency of vibration than position. An inversely proportional relationship is found between polyp size and fundamental frequency, as well as proximity to the anterior-posterior surface and fundamental frequency.
Canadian Acoustical Association / Association canadienne d'acoustique
2015-03-01 00:00:00
application/pdf
application/pdf
https://jcaa.caa-aca.ca/index.php/jcaa/article/view/2654
Canadian Acoustics; Vol. 43 No. 1 (2015)
eng
Copyright (c) 2015
oai:jcaa.caa-aca.ca:article/2655
2016-03-16T17:11:05Z
jcaa:ART-SPC
driver
nmb a2200000Iu 4500
"160303 2016 eng "
2291-1391
0711-6659
dc
Effects of mouthing and interlocutor presence on movements of visible vs. non-visible articulators
Bicevskis, Katie
University of British Columbia
de Vries, Jonathan
University of British Columbia
Green, Laurie
University of British Columbia
Heim, Johannes
University of British Columbia
Božič, Jurij
University of British Columbia
D'Aquisto, Joe
University of British Columbia
Fry, Michael
University of British Columbia
Sadlier-Brown, Emily
University of British Columbia
Tkachman, Oksana
Yamane, Noriko
University of British Columbia
Gick, Bryan
University of British Columbia
Speakers take into account what information a conversation partner requires in a given context in order to best understand an utterance. Despite growing evidence showing that movements of visible articulators such as the lips are augmented in mouthed speech relative to vocalized speech, little to date has been done comparing this effect in visible vs. non-visible articulators. In addition, no studies have examined whether interlocutor engagement differentially impacts these. Building on a basic present/not-present design, we investigated whether presence of audible speech information and/or an interlocutor affect the movements of the lips and the tongue. Participants were asked to a) speak or b) mouth three target syllables in interlocutor-present and interlocutor-not-present conditions, while lip and tongue movements were recorded using video and ultrasound imaging. Results show that lip protrusion was greater in mouthed conditions compared to vocalized ones and tongue movements were either attenuated (/wa/) or unaffected (/ri/, /ra/) by these same conditions, indicating differential effects for the visible and non-visible articulators in the absence of an auditory signal. A significant interaction between the social engagement and vocalizing conditions in reference to lip aperture showed that participants produced smaller lip apertures when vocalizing alone, as compared to when in the presence of an interlocutor. However, measures of lip protrusion failed to find an effect of social engagement. We conclude that speakers make use of both auditory and visual modalities in the presence of an interlocutor, and that when acoustic information is unavailable, compensatory increases are made in the visual domain. Our findings shed new light on the multimodal nature of speech, and pose new questions about differential adaptations made by visible and non-visible articulators in different speech conditions.
Canadian Acoustical Association / Association canadienne d'acoustique
2016-03-03 00:00:00
application/pdf
application/pdf
https://jcaa.caa-aca.ca/index.php/jcaa/article/view/2655
Canadian Acoustics; Vol. 44 No. 1 (2016)
eng
Copyright (c) 2016 Katie Bicevskis, Jonathan de Vries, Laurie Green, Johannes Heim, Jurij Boži?, Joe D'Aquisto, Michael Fry, Emily Sadlier-Brown, Oksana Tkachman, Noriko Yamane, Bryan Gick
oai:jcaa.caa-aca.ca:article/2670
2015-12-21T19:26:33Z
jcaa:ART-SPC
nmb a2200000Iu 4500
"151215 2015 eng "
2291-1391
0711-6659
dc
Strategies to Enhance Whispered Speech Speaker Verification: A Comparative Analysis
Sarria-Paja, Milton O.
Institut National de la Recherche Scientifique, Centre EMT, University of Quebec http://miltonsp.com/
Falk, Tiago H.
Institut National de la Recherche Scientifique, Centre EMT, University of Quebec http://musaelab.ca/team-view/prof-tiago-h-falk/
Today, automated speech-enabled tools are increasingly being used in everyday environments. This mobility has created new challenges for developers, who are now faced with input speech of varying styles (e.g. whispered) and corrupted by different noise sources. In this paper, special emphasis is placed on whispered speech, an underexplored yet burgeoning area due to the rapid proliferation of smartphones around the world. More specifically, this paper explores the performance boundaries achievable with whispered speech for a speaker verification task, both in matched and mismatched train/test conditions. Several strategies are investigated to improve the performance in the mismatched scenario, as well as in situations involving ambient noise. Our results agree with previously reported studies in adjacent areas, that significant gains could be obtained by training speaker models with both naturally voiced and whispered speech data. Moreover, additional gains could be achieved with speaking style and gender dependent systems. Overall, speaker verification performance inline with that obtained with naturally-voiced speech could be attained for whispered speech once specific strategies were put in place. Particularly, feature fusion showed to be an important strategy for practical applications in both clean and noisy conditions.
Canadian Acoustical Association / Association canadienne d'acoustique
2015-12-15 00:00:00
application/pdf
application/pdf
application/pdf
https://jcaa.caa-aca.ca/index.php/jcaa/article/view/2670
Canadian Acoustics; Vol. 43 No. 4 (2015)
eng
Copyright (c) 2015 Milton O. Sarria-Paja, Tiago H. Falk
oai:jcaa.caa-aca.ca:article/2740
2017-04-03T19:06:39Z
jcaa:ART-SPC
nmb a2200000Iu 4500
"170403 2017 eng "
2291-1391
0711-6659
dc
Using Optical Flow Analysis on Ultrasound of the Tongue to Examine Phonological Relationships
Hall, Kathleen Currie
University of British Columbia
Smith, Hanna
University of British Columbia
McMullin, Kevin
University of Ottawa
Allen, Blake
University of British Columbia
Yamane, Noriko
University of British Columbia
This paper examines whether there are articulatory correlates of differing degrees of phonological contrast. English tense vowels are found to be produced with greater average magnitudes of movement when they occur in closed syllables, where they are generally contrastive with their lax vowel counterparts, than when they occur in open syllables, where they are less contrastive. Magnitude of tongue movement was determined by optical flow analysis of ultrasound videos of tongue movements; optical flow analysis allows for direct comparison of results across speakers and for the extraction of data from the entire timecourse of productions.
Canadian Acoustical Association / Association canadienne d'acoustique
2017-04-03 12:06:39
application/pdf
application/pdf
application/pdf
https://jcaa.caa-aca.ca/index.php/jcaa/article/view/2740
Canadian Acoustics; Vol. 45 No. 1 (2017)
eng
Copyright (c) 2017 Kathleen Currie Hall, Hanna Smith, Kevin McMullin, Blake Allen, Noriko Yamane
oai:jcaa.caa-aca.ca:article/3263
2020-11-14T00:28:57Z
jcaa:ART-SPC
nmb a2200000Iu 4500
"201113 2020 eng "
2291-1391
0711-6659
dc
Effects of Altered Intensity Feedback on Speech in Healthy Speakers
Senthinathan, Anita
SUNY Buffalo State https://orcid.org/0000-0002-7368-2538
Adams, Scott G
Western University
This study examined the role of auditory feedback in speech intensity regulation in healthy adults. Ten participants completed five speech production tasks (vowel, sentence reading with/without instructions to ignore feedback, and conversation at one/six meter interlocutor distances) under various altered intensity feedback (AIF) conditions. AIF conditions ranged from plus/minus 2.5, 5, 7.5, 10, 12.5, and 15dB SPL relative to the equivalent (0dB) condition. Speech intensity values obtained for each of the thirteen AIF conditions were submitted to a linear regression analysis and the resulting slope and intercept values were compared. The average slope for the altered intensity feedback function was found to be significantly steeper for the conversation tasks (-.16) than for the reading tasks (-.07). No difference in slope was found for the comparison of the reading tasks with and without instructions to ignore altered feedback or for the comparison of conversations at one- and six-meter interlocutor distances. It appears that speech tasks with greater communicative demands (i.e. conversation) show larger compensatory responses to altered intensity feedback than tasks with lower communicative demands (reading or prolonged vowels). Results suggest that demands of the speech task can interact with auditory feedback to influence the regulation of speech intensity.
Canadian Acoustical Association / Association canadienne d'acoustique
2020-10-05 12:42:34
application/pdf
https://jcaa.caa-aca.ca/index.php/jcaa/article/view/3263
Canadian Acoustics; Vol. 48 No. 3 (2020)
eng
Copyright (c) 2020 Anita Abeyesekera, Scott G Adams
oai:jcaa.caa-aca.ca:article/3305
2021-01-12T06:07:34Z
jcaa:ART-SPC
nmb a2200000Iu 4500
"200317 2020 eng "
2291-1391
0711-6659
dc
Real-time Ultrasound-enhanced Multimodal Imaging of Tongue using 3D Printable Stabilizer System: A Deep Learning Approach
Mozaffari, M. Hamed
University of Ottawa http://web5.uottawa.ca/www5/mmaaref/index.html http://orcid.org/0000-0002-2297-6114
Lee, Won-Sook
Professor, School of Electrical Engineering and Computer Science (EECS), University of Ottawa, Ottawa, Ontario, Canada, K1N 6N5
Despite renewed awareness of the importance of articulation, it remains a challenge for instructors to handle the pronunciation needs of language learners. There are relatively scarce pedagogical tools for pronunciation teaching and learning. Unlike inefficient, traditional pronunciation instructions like listening and repeating, electronic visual feedback (EVF) systems such as ultrasound technology have been employed in new approaches.
Recently, an ultrasound-enhanced multimodal method has been developed for visualizing tongue movements of a language learner overlaid on face-side of the speaker's head. That system was evaluated for several language courses via a blended learning paradigm at the university level.
The result was asserted that visualizing articulator’s system as biofeedback to language learners will significantly improve articulation learning efficiency.
In spite of that successful usage of multimodal technique for pronunciation training, it still requires manual works and human manipulation. In this article, we aim to contribute to this growing body of research by addressing difficulties of the previous approaches by proposing a new comprehensive, automatic, real-time multimodal pronunciation training system, benefits from powerful artificial intelligence techniques.
The main objective of this research was to combine the advantages of ultrasound technology, three-dimensional printing, and deep learning algorithms to enhance the performance of previous systems. Our preliminary pedagogical evaluation of the proposed system revealed a significant improvement in flexibility, control, robustness, and autonomy.
Canadian Acoustical Association / Association canadienne d'acoustique
2020-04-07 00:00:00
application/pdf
https://jcaa.caa-aca.ca/index.php/jcaa/article/view/3305
Canadian Acoustics; Vol. 48 No. 1 (2020)
eng
Copyright (c) 2020 M. Hamed Mozaffari, Won-Sook Lee
oai:jcaa.caa-aca.ca:article/3383
2020-12-28T01:21:54Z
jcaa:ART-SPC
nmb a2200000Iu 4500
"201213 2020 eng "
2291-1391
0711-6659
dc
Vowel Discrimination Abilities in Quebec French School-Aged Children
Ménard, Lucie
UQAM
Trudeau-Fisette, Pamela
Laboratoire de phonétique, UQAM
Although auditory perception develops in infancy, it continues to mature until mid-adolescence. Many studies have documented early auditory perceptual development, yet very little is known about changes that occur in childhood. This study aimed to investigate the development of vowel discrimination in school-aged French-speaking children. Synthesized vowels contrasting along height and rounding were used as stimuli in a discrimination test given to 49 children aged 6 to 10 years old and twelve adults. Peak vowel discrimination scores and category boundaries shifted between 7 and 9 years of age, which suggests non-linear changes in speech processing.
Canadian Acoustical Association / Association canadienne d'acoustique
2020-12-13 00:00:00
application/pdf
https://jcaa.caa-aca.ca/index.php/jcaa/article/view/3383
Canadian Acoustics; Vol. 48 No. 4 (2020)
eng
Copyright (c) 2020 Lucie Ménard, Madame
oai:ojs.pkp.sfu.ca:article/3383
2020-12-14T00:38:06Z
jcaa:ART-SPC
oai:jcaa.caa-aca.ca:article/3437
2022-07-25T20:50:51Z
jcaa:ART-SPC
nmb a2200000Iu 4500
"220725 2022 eng "
2291-1391
0711-6659
dc
An An Acoustic Analysis of Oromo and Amharic Ejective Stops
Negesse, Feda
Addis Ababa University https://orcid.org/0000-0002-0515-4215
Previous acoustic studies of ejective sounds seem to have concentrated on the comparison of ejective stops with their pulmonic voiceless stops. The current study investigates ejective stops of Amharic and Oromo in order to examine if there are significant variations within the sounds and between the two languages. The audio data for the study were collected from 36 undergraduate students who are native speakers of their respective languages. Nine acoustic measures, which include temporal and spectral parameters, were extracted from the data. Analysis of the acoustic measures reveals a significant variation within ejective stops with respect to most acoustic variables under investigation. However, there is no significant variation between languages in all acoustic measures. Most tokens of the sounds are correctly classified with the spectral moments of their burst noise. Typologically, ejective stops of the languages cannot be categorically classified as slack or stiff and diffuse or compact based on their acoustic properties. Overall, it is concluded that more differences are observed within the sounds than between the languages.
Canadian Acoustical Association / Association canadienne d'acoustique
2021-12-27 00:00:00
application/pdf
https://jcaa.caa-aca.ca/index.php/jcaa/article/view/3437
Canadian Acoustics; Vol. 49 No. 4 (2021)
eng
Copyright (c) 2021 Feda Negesse