Results for 'Speech Acoustics'

970 found
Order:
  1.  25
    Speech and Music Acoustics, Rhythms of the Brain and their Impact on the Ability to Accept Information.I. V. Pavlov & V. M. Tsaplev - 2020 - Дискурс 6 (1):96-105.
    Introduction. A radical tendency in modern approaches to understanding the mechanisms of the brain is the tendency of some scientists to believe that the brain is a receptor capable of capturing thoughts; the nature of the occurrence of the thoughts themselves, however, is not to be clarified. However, speech expressing thoughts is undoubtedly the result of the work of the brain, so studies of the frequency structure of speech can be the basis for considering the material structure of (...)
    No categories
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  2.  17
    Vowel acoustics of Nungon child-directed speech, adult dyadic conversation, and foreigner-directed monologues.Hannah S. Sarvasy, Weicong Li, Jaydene Elvin & Paola Escudero - 2022 - Frontiers in Psychology 13.
    In many communities around the world, speech to infants and small children has increased mean pitch, increased pitch range, increased vowel duration, and vowel hyper-articulation when compared to speech directed to adults. Some of these IDS and CDS features are also attested in foreigner-directed speech, which has been studied for a smaller range of languages, generally major national languages, spoken by millions of people. We examined vowel acoustics in CDS, conversational ADS, and monologues directed to a (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  3.  19
    Acoustic Correlates and Adult Perceptions of Distress in Infant Speech-Like Vocalizations and Cries.Hyunjoo Yoo, Eugene H. Buder, Dale D. Bowman, Gavin M. Bidelman & D. Kimbrough Oller - 2019 - Frontiers in Psychology 10.
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark   1 citation  
  4.  62
    Articulatory-to-Acoustic Conversion of Mandarin Emotional Speech Based on PSO-LSSVM.Guofeng Ren, Jianmei Fu, Guicheng Shao & Yanqin Xun - 2021 - Complexity 2021:1-10.
    The production of emotional speech is determined by the movement of the speaker’s tongue, lips, and jaw. In order to combine articulatory data and acoustic data of speakers, articulatory-to-acoustic conversion of emotional speech has been studied. In this paper, parameters of LSSVM model have been optimized using the PSO method, and the optimized PSO-LSSVM model was applied to the articulatory-to-acoustic conversion. The root mean square error and mean Mel-cepstral distortion have been used to evaluate the results of conversion; (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  5.  8
    Interactions between acoustic challenges and processing depth in speech perception as measured by task-evoked pupil response.Jing Shen, Laura P. Fitzgerald & Erin R. Kulick - 2022 - Frontiers in Psychology 13.
    Speech perception under adverse conditions is a multistage process involving a dynamic interplay among acoustic, cognitive, and linguistic factors. Nevertheless, prior research has primarily focused on factors within this complex system in isolation. The primary goal of the present study was to examine the interaction between processing depth and the acoustic challenge of noise and its effect on processing effort during speech perception in noise. Two tasks were used to represent different depths of processing. The speech recognition (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  6.  21
    Beyond Correlation: Acoustic Transformation Methods for the Experimental Study of Emotional Voice and Speech.Pablo Arias, Laura Rachman, Marco Liuni & Jean-Julien Aucouturier - 2020 - Emotion Review 13 (1):12-24.
    While acoustic analysis methods have become a commodity in voice emotion research, experiments that attempt not only to describe but to computationally manipulate expressive cues in emotional voice...
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark   2 citations  
  7.  40
    Disordered speech disrupts conversational entrainment: a study of acoustic-prosodic entrainment and communicative success in populations with communication challenges.Stephanie A. Borrie, Nichola Lubold & Heather Pon-Barry - 2015 - Frontiers in Psychology 6.
    Direct download (8 more)  
     
    Export citation  
     
    Bookmark   1 citation  
  8.  39
    Acoustic correlates and perceptual cues in speech.James R. Sawusch - 1998 - Behavioral and Brain Sciences 21 (2):283-284.
    Locus equations are supposed to capture a perceptual invariant of place of articulation in consonants. Synthetic speech data show that human classification deviates systematically from the predictions of locus equations. The few studies that have contrasted predictions from competing theories yield mixed results, indicating that no current theory adequately characterizes the perceptual mapping from sound to phonetic symbol.
    Direct download (6 more)  
     
    Export citation  
     
    Bookmark  
  9.  51
    The mapping from acoustic structure to the phonetic categories of speech: The invariance problem.Sheila E. Blumstein - 1998 - Behavioral and Brain Sciences 21 (2):260-260.
    This commentary focuses on the nature of combinatorial properties for speech and the locus equation. The presence of some overlap in locus equation space suggests that this higher order property may not be strictly invariant and may require other cues or properties for the perception of place of articulation. Moreover, combinatorial analysis in two-dimensional space and the resultant linearity appear to have a “special” status in the development of this theoretical framework. However, place of articulation is only one of (...)
    Direct download (6 more)  
     
    Export citation  
     
    Bookmark  
  10. Perception of Nigerian Dùndún Talking Drum Performances as Speech-Like vs. Music-Like: The Role of Familiarity and Acoustic Cues.Cecilia Durojaye, Lauren Fink, Tina Roeske, Melanie Wald-Fuhrmann & Pauline Larrouy-Maestri - 2021 - Frontiers in Psychology 12.
    It seems trivial to identify sound sequences as music or speech, particularly when the sequences come from different sound sources, such as an orchestra and a human voice. Can we also easily distinguish these categories when the sequence comes from the same sound source? On the basis of which acoustic features? We investigated these questions by examining listeners’ classification of sound sequences performed by an instrument intertwining both speech and music: the dùndún talking drum. The dùndún is commonly (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  11.  28
    Linear correlates in the speech signal: Consequences of the specific use of an acoustic tube?René Carré - 1998 - Behavioral and Brain Sciences 21 (2):261-262.
    The debate on the origin of the locus equation is circular. In this commentary the locus equation is obtained by way of a theoretical model based on acoustics without recourse to articulatory knowledge or perceptual constraints. The proposed model is driven by criteria of minimum energy and maximum simplicity.
    Direct download (6 more)  
     
    Export citation  
     
    Bookmark  
  12.  15
    Acoustic Detail But Not Predictability of Task-Irrelevant Speech Disrupts Working Memory.Malte Wöstmann & Jonas Obleser - 2016 - Frontiers in Human Neuroscience 10.
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  13.  20
    Two Servants, One Master: The Common Acoustic Origins of the Divergent Communicative Media of Music and Speech.Nicholas Bannan - 2022 - Evolutionary Studies in Imaginative Culture 6 (2):21-42.
    This article explores and examines research in the field of human vocalization, proposing an evolutionary sequence for human acoustic perception and productive response. This involves updating and extending Charles Darwin’s 1871 proposal that musical communi­cation predated language, while providing the anatomical and behavioral foundations for the articulacy on which it depends. In presenting evidence on which a new consensus regarding the emergence of human vocal ability may be based, we present and review contributions from a wide range of disciplines, illustrating (...)
    No categories
    Direct download (3 more)  
     
    Export citation  
     
    Bookmark  
  14.  26
    Moving to the Speed of Sound: Context Modulation of the Effect of Acoustic Properties of Speech.Hadas Shintel & Howard C. Nusbaum - 2008 - Cognitive Science 32 (6):1063-1074.
    Suprasegmental acoustic patterns in speech can convey meaningful information and affect listeners' interpretation in various ways, including through systematic analog mapping of message‐relevant information onto prosody. We examined whether the effect of analog acoustic variation is governed by the acoustic properties themselves. For example, fast speech may always prime the concept of speed or a faster response. Alternatively, the effect may be modulated by the context‐dependent interpretation of those properties; the effect of rate may depend on how listeners (...)
    No categories
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark   3 citations  
  15.  12
    Discriminatively trained continuous Hindi speech recognition using integrated acoustic features and recurrent neural network language modeling.R. K. Aggarwal & A. Kumar - 2020 - Journal of Intelligent Systems 30 (1):165-179.
    This paper implements the continuous Hindi Automatic Speech Recognition (ASR) system using the proposed integrated features vector with Recurrent Neural Network (RNN) based Language Modeling (LM). The proposed system also implements the speaker adaptation using Maximum-Likelihood Linear Regression (MLLR) and Constrained Maximum likelihood Linear Regression (C-MLLR). This system is discriminatively trained by Maximum Mutual Information (MMI) and Minimum Phone Error (MPE) techniques with 256 Gaussian mixture per Hidden Markov Model(HMM) state. The training of the baseline system has been done (...)
    No categories
    Direct download  
     
    Export citation  
     
    Bookmark  
  16.  9
    N1 Repetition-Attenuation for Acoustically Variable Speech and Spectrally Rotated Speech.Ellen Marklund, Lisa Gustavsson, Petter Kallioinen & Iris-Corinna Schwarz - 2020 - Frontiers in Human Neuroscience 14.
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  17.  28
    Motor theory of speech perception or acoustic theory of speech production?Lyn Frazier - 1983 - Behavioral and Brain Sciences 6 (2):213-214.
  18.  55
    The Semantics of Prosody: Acoustic and Perceptual Evidence of Prosodic Correlates to Word Meaning.Lynne C. Nygaard, Debora S. Herold & Laura L. Namy - 2009 - Cognitive Science 33 (1):127-146.
    This investigation examined whether speakers produce reliable prosodic correlates to meaning across semantic domains and whether listeners use these cues to derive word meaning from novel words. Speakers were asked to produce phrases in infant‐directed speech in which novel words were used to convey one of two meanings from a set of antonym pairs (e.g., big/small). Acoustic analyses revealed that some acoustic features were correlated with overall valence of the meaning. However, each word meaning also displayed a unique acoustic (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark   16 citations  
  19.  15
    Representation of speech sounds in precategorical acoustic storage.Robert G. Crowder - 1973 - Journal of Experimental Psychology 98 (1):14.
  20.  10
    Formant transitions as acoustic cues for place of articulation in speech perception.Zora Jachova, Lidija Ristovska & Ljudmil Spasov - 2021 - Годишен зборник на Филозофскиот факултет/The Annual of the Faculty of Philosophy in Skopje 74:567-580.
    No categories
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  21.  44
    Phonetic features and acoustic invariance in speech.Sheila E. Blumstein & Kenneth N. Stevens - 1981 - Cognition 10 (1-3):25-32.
    Direct download (3 more)  
     
    Export citation  
     
    Bookmark   4 citations  
  22. Cue integration with categories: Weighting acoustic cues in speech using unsupervised learning and distributional statistics.Joseph C. Toscano & Bob McMurray - 2010 - Cognitive Science 34 (3):434.
    Direct download (8 more)  
     
    Export citation  
     
    Bookmark   29 citations  
  23.  34
    Acoustic correlates of emotional dimensions in laughter: Arousal, dominance, and valence.Diana P. Szameitat, Chris J. Darwin, Dirk Wildgruber, Kai Alter & André J. Szameitat - 2011 - Cognition and Emotion 25 (4):599-611.
    Although laughter plays an essential part in emotional vocal communication, little is known about the acoustical correlates that encode different emotional dimensions. In this study we examined the acoustical structure of laughter sounds differing along four emotional dimensions: arousal, dominance, sender's valence, and receiver-directed valence. Correlation of 43 acoustic parameters with individual emotional dimensions revealed that each emotional dimension was associated with a number of vocal cues. Common patterns of cues were found with emotional expression in speech, supporting the (...)
    Direct download (3 more)  
     
    Export citation  
     
    Bookmark   3 citations  
  24.  9
    Acoustic Correlates of English Lexical Stress Produced by Chinese Dialect Speakers Compared to Native English Speakers.Xingrong Guo - 2022 - Frontiers in Psychology 13:796252.
    English second language learners often experience difficulties in producing native-like English lexical stress. It is unknown which acoustic correlates, such as fundamental frequency (F0), duration, and intensity, are the most problematic for Chinese dialect speakers. The present study investigated the prosodic transfer effects of first language (L1) regional dialects on the production of English stress contrasts. Native English speakers (N = 20) and Chinese learners (N = 60) with different dialect backgrounds (Beijing, Changsha, and Guangzhou dialects) produced the same stimulus (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  25.  64
    Linear correlates in the speech signal: The orderly output constraint.Harvey M. Sussman, David Fruchter, Jon Hilbert & Joseph Sirosh - 1998 - Behavioral and Brain Sciences 21 (2):241-259.
    Neuroethological investigations of mammalian and avian auditory systems have documented species-specific specializations for processing complex acoustic signals that could, if viewed in abstract terms, have an intriguing and striking relevance for human speech sound categorization and representation. Each species forms biologically relevant categories based on combinatorial analysis of information-bearing parameters within the complex input signal. This target article uses known neural models from the mustached bat and barn owl to develop, by analogy, a conceptualization of human processing of consonant (...)
    Direct download (7 more)  
     
    Export citation  
     
    Bookmark   6 citations  
  26. Eros, Beauty, and Phon-Aesthetic Judgements of Language Sound. We Like It Flat and Fast, but Not Melodious. Comparing Phonetic and Acoustic Features of 16 European Languages.Vita V. Kogan & Susanne M. Reiterer - 2021 - Frontiers in Human Neuroscience 15:578594.
    This article concerns sound aesthetic preferences for European foreign languages. We investigated the phonetic-acoustic dimension of the linguistic aesthetic pleasure to describe the “music” found in European languages. The Romance languages, French, Italian, and Spanish, take a lead when people talk about melodious language – the music-like effects in the language (a.k.a., phonetic chill). On the other end of the melodiousness spectrum are German and Arabic that are often considered sounding harsh and un-attractive. Despite the public interest, limited research has (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  27.  4
    Acoustic Processing and the Origin of Human Vocal Communication.Nicholas Bannan, Robin I. M. Dunbar, Alan R. Harvey & Piotr Podlipniak - forthcoming - Evolutionary Studies in Imaginative Culture:1006-1039.
    Humans have inherited from their remotest mammalian ancestors an integration of the sensory and motor systems that permits the exchange of signals and information, led by an instinctive response to harmonicity. The transition, from capacity for animal communication involving calls, facial expression and gestures, to modern human culture that embraces language, music, and dance, has resulted from anatomical adaptations such as upright posture, a distinct oro-facial and respiratory tract arrangement, and important changes in neural architecture, connectivity and plasticity. A key (...)
    No categories
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  28.  29
    The sound of motion in spoken language: Visual information conveyed by acoustic properties of speech.Hadas Shintel & Howard C. Nusbaum - 2007 - Cognition 105 (3):681-690.
  29.  69
    Are you a good mimic? Neuro-acoustic signatures for speech imitation ability.Susanne M. Reiterer, Xiaochen Hu, T. A. Sumathi & Nandini C. Singh - 2013 - Frontiers in Psychology 4.
    Direct download (5 more)  
     
    Export citation  
     
    Bookmark   4 citations  
  30.  12
    A procedure for adaptive control of the interaction between acoustic classification and linguistic decoding in automatic recognition of continuous speech.C. C. Tappert & N. R. Dixon - 1974 - Artificial Intelligence 5 (2):95-113.
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  31. Speech Perception: A Philosophical Analysis.Irene Appelbaum - 1995 - Dissertation, The University of Chicago
    The overall goal of speech perception research is to explain how spoken language is recognized and understood. In the current research framework it is assumed that the key to achieving this overall goal is to solve the lack of invariance problem. But nearly half a century of sustained effort in a variety of theoretical perspectives has failed to solve this problem. Indeed, not only has the problem not been solved, virtually no empirical candidates for solving the problem have been (...)
     
    Export citation  
     
    Bookmark  
  32.  48
    Brain mechanisms of acoustic communication in humans and nonhuman primates: An evolutionary perspective.Hermann Ackermann, Steffen R. Hage & Wolfram Ziegler - 2014 - Behavioral and Brain Sciences 37 (6):529-546.
    Any account of “what is special about the human brain” (Passingham 2008) must specify the neural basis of our unique ability to produce speech and delineate how these remarkable motor capabilities could have emerged in our hominin ancestors. Clinical data suggest that the basal ganglia provide a platform for the integration of primate-general mechanisms of acoustic communication with the faculty of articulate speech in humans. Furthermore, neurobiological and paleoanthropological data point at a two-stage model of the phylogenetic evolution (...)
    Direct download (3 more)  
     
    Export citation  
     
    Bookmark   18 citations  
  33.  41
    A computational model of word segmentation from continuous speech using transitional probabilities of atomic acoustic events.Okko Räsänen - 2011 - Cognition 120 (2):149-176.
    Direct download (5 more)  
     
    Export citation  
     
    Bookmark   3 citations  
  34.  31
    Emotional Connotations of Musical Instrument Timbre in Comparison With Emotional Speech Prosody: Evidence From Acoustics and Event-Related Potentials.Xiaoluan Liu, Yi Xu, Kai Alter & Jyrki Tuomainen - 2018 - Frontiers in Psychology 9.
    Direct download (3 more)  
     
    Export citation  
     
    Bookmark   3 citations  
  35.  14
    Word and Mystery: The Acoustics of Cultural Transmission During the Protestant Reformation.Braxton Boren - 2021 - Frontiers in Psychology 12.
    To a first-order approximation we can place most worship services on a continuum between clarity and mystery, depending on the setting and content of the service. This liturgical space can be thought of as a combination of the physical acoustics of the worship space and the qualities of the sound created during the worship service. A very clear acoustic channel emphasizes semantic content, especially speech intelligibility. An immersive, reverberant acoustic emphasizes mystery and music. One of the chief challenges (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  36.  36
    A Neuromotor to Acoustical Jaw-Tongue Projection Model With Application in Parkinson’s Disease Hypokinetic Dysarthria.Andrés Gómez, Pedro Gómez, Daniel Palacios, Victoria Rodellar, Víctor Nieto, Agustín Álvarez & Athanasios Tsanas - 2021 - Frontiers in Human Neuroscience 15.
    Aim: The present work proposes the study of the neuromotor activity of the masseter-jaw-tongue articulation during diadochokinetic exercising to establish functional statistical relationships between surface Electromyography, 3D Accelerometry, and acoustic features extracted from the speech signal, with the aim of characterizing Hypokinetic Dysarthria. A database of multi-trait signals of recordings from an age-matched control and PD participants are used in the experimental study. Hypothesis: The main assumption is that information between sEMG and 3D acceleration, and acoustic features may be (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  37. Neutrosophic speech recognition Algorithm for speech under stress by Machine learning.Florentin Smarandache, D. Nagarajan & Said Broumi - 2023 - Neutrosophic Sets and Systems 53.
    It is well known that the unpredictable speech production brought on by stress from the task at hand has a significant negative impact on the performance of speech processing algorithms. Speech therapy benefits from being able to detect stress in speech. Speech processing performance suffers noticeably when perceptually produced stress causes variations in speech production. Using the acoustic speech signal to objectively characterize speaker stress is one method for assessing production variances brought on (...)
    Direct download  
     
    Export citation  
     
    Bookmark  
  38.  17
    Charting speech with bats without requiring maps.Jagmeet S. Kanwal - 1998 - Behavioral and Brain Sciences 21 (2):272-273.
    The effort to understand speech perception on the basis of relationships between acoustic parameters of speech sounds is to be recommended. Neural specializations (combination-sensitivity) for echolocation, communication, and sound localization probably constitute the common mechanisms of vertebrate auditory processing and may be essential for speech production as well as perception. There is, however, no need for meaningful maps.
    Direct download (6 more)  
     
    Export citation  
     
    Bookmark  
  39.  17
    Standing-out and Fitting-in: The Acoustic-Space of Extemporised Speech.Tim Flanagan - 2022 - Journal of Intercultural Studies 6 (43):758-772.
    An explicit feature of the World Health Organisation’s response to the COVID-19 pandemic has been to ensure that naming conventions, both for the disease itself and for the variants of its underlying virus, should not have a stigmatising effect on any one population or region. An implicit feature of this undertaking is the recognition that the relation between ‘what is said’ and ‘what is heard’ involves an ongoing and even generative tension that cannot be mapped following a defined set of (...)
    Direct download  
     
    Export citation  
     
    Bookmark  
  40.  57
    The ConDialInt Model: Condensation, Dialogality, and Intentionality Dimensions of Inner Speech Within a Hierarchical Predictive Control Framework.Romain Grandchamp, Lucile Rapin, Marcela Perrone-Bertolotti, Cédric Pichat, Célise Haldin, Emilie Cousin, Jean-Philippe Lachaux, Marion Dohen, Pascal Perrier, Maëva Garnier, Monica Baciu & Hélène Lœvenbruck - 2019 - Frontiers in Psychology 10:454766.
    Inner speech has been shown to vary in form along several dimensions. Along condensation, condensed inner speech forms have been described, that are supposed to be deprived of acoustic, phonological and even syntactic qualities. Expanded forms, on the other extreme, display articulatory and auditory properties. Along dialogality, inner speech can be monologal, when we engage in internal soliloquy, or dialogal, when we recall past conversations or imagine future dialogs involving our own voice as well as that of (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark   11 citations  
  41.  24
    Modeling Co‐evolution of Speech and Biology.Bart de Boer - 2016 - Topics in Cognitive Science 8 (2):459-468.
    Two computer simulations are investigated that model interaction of cultural evolution of language and biological evolution of adaptations to language. Both are agent‐based models in which a population of agents imitates each other using realistic vowels. The agents evolve under selective pressure for good imitation. In one model, the evolution of the vocal tract is modeled; in the other, a cognitive mechanism for perceiving speech accurately is modeled. In both cases, biological adaptations to using and learning speech evolve, (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark   5 citations  
  42. Speech perception deficits and the effect of envelope-enhanced story listening combined with phonics intervention in pre-readers at risk for dyslexia.Femke Vanden Bempt, Shauni Van Herck, Maria Economou, Jolijn Vanderauwera, Maaike Vandermosten, Jan Wouters & Pol Ghesquière - 2022 - Frontiers in Psychology 13.
    Developmental dyslexia is considered to be most effectively addressed with preventive phonics-based interventions, including grapheme-phoneme coupling and blending exercises. These intervention types require intact speech perception abilities, given their large focus on exercises with auditorily presented phonemes. Yet some children with dyslexia experience problems in this domain due to a poorer sensitivity to rise times, i.e., rhythmic acoustic cues present in the speech envelope. As a result, the often subtle speech perception problems could potentially constrain an optimal (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  43.  57
    Modeling Co‐evolution of Speech and Biology.Bart Boer - 2016 - Topics in Cognitive Science 8 (2):459-468.
    Two computer simulations are investigated that model interaction of cultural evolution of language and biological evolution of adaptations to language. Both are agent-based models in which a population of agents imitates each other using realistic vowels. The agents evolve under selective pressure for good imitation. In one model, the evolution of the vocal tract is modeled; in the other, a cognitive mechanism for perceiving speech accurately is modeled. In both cases, biological adaptations to using and learning speech evolve, (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark   5 citations  
  44.  50
    DLD: An Optimized Chinese Speech Recognition Model Based on Deep Learning.Hong Lei, Yue Xiao, Yanchun Liang, Dalin Li & Heow Pueh Lee - 2022 - Complexity 2022:1-8.
    Speech recognition technology has played an indispensable role in realizing human-computer intelligent interaction. However, most of the current Chinese speech recognition systems are provided online or offline models with low accuracy and poor performance. To improve the performance of offline Chinese speech recognition, we propose a hybrid acoustic model of deep convolutional neural network, long short-term memory, and deep neural network. This model utilizes DCNN to reduce frequency variation and adds a batch normalization layer after its convolutional (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  45.  13
    Multi-Talker Speech Promotes Greater Knowledge-Based Spoken Mandarin Word Recognition in First and Second Language Listeners.Seth Wiener & Chao-Yang Lee - 2020 - Frontiers in Psychology 11.
    Spoken word recognition involves a perceptual tradeoff between the reliance on the incoming acoustic signal and knowledge about likely sound categories and their co-occurrences as words. This study examined how adult second language (L2) learners navigate between acoustic-based and knowledge-based spoken word recognition when listening to highly variable, multi-talker truncated speech, and whether this perceptual tradeoff changes as L2 listeners gradually become more proficient in their L2 after multiple months of structured classroom learning. First language (L1) Mandarin Chinese listeners (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark   2 citations  
  46.  69
    Electrocorticographic representations of segmental features in continuous speech.Fabien Lotte, Jonathan S. Brumberg, Peter Brunner, Aysegul Gunduz, Anthony L. Ritaccio, Cuntai Guan & Gerwin Schalk - 2015 - Frontiers in Human Neuroscience 9:119171.
    Acoustic speech output results from coordinated articulation of dozens of muscles, bones and cartilages of the vocal mechanism. While we commonly take the fluency and speed of our speech productions for granted, the neural mechanisms facilitating the requisite muscular control are not completely understood. Previous neuroimaging and electrophysiology studies of speech sensorimotor control has typically concentrated on speech sounds (i.e., phonemes, syllables and words) in isolation; sentence-length investigations have largely been used to inform coincident linguistic processing. (...)
    Direct download (7 more)  
     
    Export citation  
     
    Bookmark   2 citations  
  47.  34
    Locus equation and hidden parameters of speech.Li Deng - 1998 - Behavioral and Brain Sciences 21 (2):263-264.
    Locus equations contain an economical set of hidden (i.e., not directly observable in the data) parameters of speech that provide an elegant way of characterizing the ubiquitous context-dependent behaviors exhibited in speech acoustics. These hidden parameters can be effectively exploited to constrain the huge set of context-dependent speech model parameters currently in use in modern, mainstream speech recognition technology.
    Direct download (6 more)  
     
    Export citation  
     
    Bookmark  
  48.  22
    Visual Speech Perception Cues Constrain Patterns of Articulatory Variation and Sound Change.Jonathan Havenhill & Youngah Do - 2018 - Frontiers in Psychology 9:337534.
    What are the factors that contribute to (or inhibit) diachronic sound change? While acoustically motivated sound changes are well documented, research on the articulatory and audiovisual-perceptual aspects of sound change is limited. This paper investigates the interaction of articulatory variation and audiovisual speech perception in the Northern Cities Vowel Shift (NCVS), a pattern of sound change observed in the Great Lakes region of the United States. We focus specifically on the maintenance of the contrast between the vowels /ɑ/ and (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  49.  13
    The Principle of Inverse Effectiveness in Audiovisual Speech Perception.Luuk P. H. van de Rijt, Anja Roye, Emmanuel A. M. Mylanus, A. John van Opstal & Marc M. van Wanrooij - 2019 - Frontiers in Human Neuroscience 13:468577.
    We assessed how synchronous speech listening and lipreading affects speech recognition in acoustic noise. In simple audiovisual perceptual tasks, inverse effectiveness is often observed, which holds that the weaker the unimodal stimuli, or the poorer their signal-to-noise ratio, the stronger the audiovisual benefit. So far, however, inverse effectiveness has not been demonstrated for complex audiovisual speech stimuli. Here we assess whether this multisensory integration effect can also be observed for the recognizability of spoken words. To that end, (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark   3 citations  
  50.  63
    Learning Foreign Sounds in an Alien World: Videogame Training Improves Non-Native Speech Categorization.Sung-joo Lim & Lori L. Holt - 2011 - Cognitive Science 35 (7):1390-1405.
    Although speech categories are defined by multiple acoustic dimensions, some are perceptually weighted more than others and there are residual effects of native-language weightings in non-native speech perception. Recent research on nonlinguistic sound category learning suggests that the distribution characteristics of experienced sounds influence perceptual cue weights: Increasing variability across a dimension leads listeners to rely upon it less in subsequent category learning (Holt & Lotto, 2006). The present experiment investigated the implications of this among native Japanese learning (...)
    No categories
    Direct download (3 more)  
     
    Export citation  
     
    Bookmark   8 citations  
1 — 50 / 970