Results for 'Human speech recognition'

988 found
Order:
  1.  17
    EARSHOT: A Minimal Neural Network Model of Incremental Human Speech Recognition.James S. Magnuson, Heejo You, Sahil Luthra, Monica Li, Hosung Nam, Monty Escabí, Kevin Brown, Paul D. Allopenna, Rachel M. Theodore, Nicholas Monto & Jay G. Rueckl - 2020 - Cognitive Science 44 (4):e12823.
    Despite the lack of invariance problem (the many‐to‐many mapping between acoustics and percepts), human listeners experience phonetic constancy and typically perceive what a speaker intends. Most models of human speech recognition (HSR) have side‐stepped this problem, working with abstract, idealized inputs and deferring the challenge of working with real speech. In contrast, carefully engineered deep learning networks allow robust, real‐world automatic speech recognition (ASR). However, the complexities of deep learning architectures and training regimens (...)
    No categories
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark   4 citations  
  2.  80
    Neutrosophic speech recognition Algorithm for speech under stress by Machine learning.Florentin Smarandache, D. Nagarajan & Said Broumi - 2023 - Neutrosophic Sets and Systems 53.
    It is well known that the unpredictable speech production brought on by stress from the task at hand has a significant negative impact on the performance of speech processing algorithms. Speech therapy benefits from being able to detect stress in speech. Speech processing performance suffers noticeably when perceptually produced stress causes variations in speech production. Using the acoustic speech signal to objectively characterize speaker stress is one method for assessing production variances brought on (...)
    Direct download  
     
    Export citation  
     
    Bookmark  
  3.  37
    DLD: An Optimized Chinese Speech Recognition Model Based on Deep Learning.Hong Lei, Yue Xiao, Yanchun Liang, Dalin Li & Heow Pueh Lee - 2022 - Complexity 2022:1-8.
    Speech recognition technology has played an indispensable role in realizing human-computer intelligent interaction. However, most of the current Chinese speech recognition systems are provided online or offline models with low accuracy and poor performance. To improve the performance of offline Chinese speech recognition, we propose a hybrid acoustic model of deep convolutional neural network, long short-term memory, and deep neural network. This model utilizes DCNN to reduce frequency variation and adds a batch normalization (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  4.  26
    A salience-driven approach to speech recognition for human-robot interaction.Pierre Lison - 2010 - In T. Icard & R. Muskens (eds.), Interfaces: Explorations in Logic, Language and Computation. Springer Berlin. pp. 102--113.
  5.  11
    Recognition of English speech – using a deep learning algorithm.Shuyan Wang - 2023 - Journal of Intelligent Systems 32 (1).
    The accurate recognition of speech is beneficial to the fields of machine translation and intelligent human–computer interaction. After briefly introducing speech recognition algorithms, this study proposed to recognize speech with a recurrent neural network (RNN) and adopted the connectionist temporal classification (CTC) algorithm to align input speech sequences and output text sequences forcibly. Simulation experiments compared the RNN-CTC algorithm with the Gaussian mixture model–hidden Markov model and convolutional neural network-CTC algorithms. The results demonstrated (...)
    Direct download  
     
    Export citation  
     
    Bookmark  
  6.  11
    Sexist Hate Speech and the International Human Rights Law: Towards Legal Recognition of the Phenomenon by the United Nations and the Council of Europe.Katarzyna Sękowska-Kozłowska, Grażyna Baranowska & Aleksandra Gliszczyńska-Grabias - 2022 - International Journal for the Semiotics of Law - Revue Internationale de Sémiotique Juridique 35 (6):2323-2345.
    For many women and girls sexist and misogynistic language is an everyday experience. Some instances of this speech can be categorized as ‘sexist hate speech’, as not only having an insulting or degrading character towards the individuals to whom the speech is addressed, but also resonating with the entire group, contributing to its silencing, marginalization and exclusion. The aim of this article is to examine how sexist hate speech is handled in international human rights law. (...)
    Direct download (3 more)  
     
    Export citation  
     
    Bookmark  
  7.  34
    How Should a Speech Recognizer Work?Odette Scharenborg, Dennis Norris, Louis Bosch & James M. McQueen - 2005 - Cognitive Science 29 (6):867-918.
    Although researchers studying human speech recognition (HSR) and automatic speech recognition (ASR) share a common interest in how information processing systems (human or machine) recognize spoken language, there is little communication between the two disciplines. We suggest that this lack of communication follows largely from the fact that research in these related fields has focused on the mechanics of how speech can be recognized. In Marr's (1982) terms, emphasis has been on the algorithmic (...)
    No categories
    Direct download (4 more)  
     
    Export citation  
     
    Bookmark   7 citations  
  8.  14
    How Should a Speech Recognizer Work?Odette Scharenborg, Dennis Norris, Louis ten Bosch & James M. McQueen - 2005 - Cognitive Science 29 (6):867-918.
    Although researchers studying human speech recognition (HSR) and automatic speech recognition (ASR) share a common interest in how information processing systems (human or machine) recognize spoken language, there is little communication between the two disciplines. We suggest that this lack of communication follows largely from the fact that research in these related fields has focused on the mechanics of how speech can be recognized. In Marr's (1982) terms, emphasis has been on the algorithmic (...)
    No categories
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark   5 citations  
  9. Intention Recognition as the Mechanism of Human Communication.Daniel W. Harris - 2019 - In Arthur Sullivan (ed.), Sensations, Thoughts, and Language: Essays in Honor of Brian Loar. New York, NY: Routledge.
    Intentionalism is a research program that seeks to explain facts about meaning and communication in psychological terms, with our capacity for intention recognition playing a starring role. My aim here is to recommend a methodological reorientation in this program. Instead of a focus on intuitive counterexamples to proposals about necessary-and-sufficient conditions, we should aim to investigate the psychological mechanisms whose activities and interactions explain our capacity to communicate. Taking this methodologi- cal reorientation to heart, I sketch a theory of (...)
    Direct download  
     
    Export citation  
     
    Bookmark   3 citations  
  10.  10
    Understanding Miscommunication: Speech Act Recognition in Digital Contexts.Thomas Holtgraves - 2021 - Cognitive Science 45 (10):e13023.
    Successful language use requires accurate intention recognition. However, sometimes this can be undermined because communication occurs within an interpersonal context. In this research, I used a relatively large set of speech acts (n = 32) and explored how variability in their inherent face‐threat influences the extent to which they are successfully recognized by a recipient, as well as the confidence of senders and receivers in their communicative success. Participants in two experiments either created text messages (senders) designed to (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark   1 citation  
  11. From monkey-like action recognition to human language: An evolutionary framework for neurolinguistics.Michael A. Arbib - 2005 - Behavioral and Brain Sciences 28 (2):105-124.
    The article analyzes the neural and functional grounding of language skills as well as their emergence in hominid evolution, hypothesizing stages leading from abilities known to exist in monkeys and apes and presumed to exist in our hominid ancestors right through to modern spoken and signed languages. The starting point is the observation that both premotor area F5 in monkeys and Broca's area in humans contain a “mirror system” active for both execution and observation of manual actions, and that F5 (...)
    Direct download (8 more)  
     
    Export citation  
     
    Bookmark   126 citations  
  12.  4
    A hidden Markov optimization model for processing and recognition of English speech feature signals.Yinchun Chen - 2022 - Journal of Intelligent Systems 31 (1):716-725.
    Speech recognition plays an important role in human–computer interaction. The higher the accuracy and efficiency of speech recognition are, the larger the improvement of human–computer interaction performance. This article briefly introduced the hidden Markov model -based English speech recognition algorithm and combined it with a back-propagation neural network to further improve the recognition accuracy and reduce the recognition time of English speech. Then, the BPNN-combined HMM algorithm was simulated and (...)
    No categories
    Direct download  
     
    Export citation  
     
    Bookmark  
  13.  20
    Audiovisual cues benefit recognition of accented speech in noise but not perceptual adaptation.Briony Banks, Emma Gowen, Kevin J. Munro & Patti Adank - 2015 - Frontiers in Human Neuroscience 9.
  14.  11
    Realization of Self-Adaptive Higher Teaching Management Based Upon Expression and Speech Multimodal Emotion Recognition.Huihui Zhou & Zheng Liu - 2022 - Frontiers in Psychology 13.
    In the process of communication between people, everyone will have emotions, and different emotions will have different effects on communication. With the help of external performance information accompanied by emotional expression, such as emotional speech signals or facial expressions, people can easily communicate with each other and understand each other. Emotion recognition is an important network of affective computers and research centers for signal processing, pattern detection, artificial intelligence, and human-computer interaction. Emotions convey important information in (...) communication and communication. Since the end of the last century, people have started the research on emotion recognition, especially how to correctly judge the emotion type has invested a lot of time and energy. In this paper, multi-modal emotion recognition is introduced to recognize facial expressions and speech, and conduct research on adaptive higher education management. Language and expression are the most direct ways for people to express their emotions. After obtaining the framework of the dual-modal emotion recognition system, the BOW model is used to identify the characteristic movement of local areas or key points. The recognition rates of emotion recognition for 1,000 audios of anger, disgust, fear, happiness, sadness and surprise are: 97.3, 83.75, 64.87, 89.87, 84.12, and 86.68%, respectively. (shrink)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  15.  22
    Speech trasformations solutions.Dimitri Kanevsky, Sara Basson, Alexander Faisman, Leonid Rachevsky, Alex Zlatsin & Sarah Conrod - 2006 - Pragmatics and Cognition 14 (2):411-442.
    This paper outlines the background development of “intelligent“ technologies such as speech recognition. Despite significant progress in the development of these technologies, they still fall short in many areas, and rapid advances in areas such as dictation are actually stalled. In this paper we have proposed semi-automatic solutions — smart integration of human and intelligent efforts. One such technique involves improvement to the speech recognition editing interface, thereby reducing the perception of errors to the viewer. (...)
    Direct download (4 more)  
     
    Export citation  
     
    Bookmark  
  16.  71
    Computational Validation of the Motor Contribution to Speech Perception.Leonardo Badino, Alessandro D'Ausilio, Luciano Fadiga & Giorgio Metta - 2014 - Topics in Cognitive Science 6 (3):461-475.
    Action perception and recognition are core abilities fundamental for human social interaction. A parieto-frontal network (the mirror neuron system) matches visually presented biological motion information onto observers' motor representations. This process of matching the actions of others onto our own sensorimotor repertoire is thought to be important for action recognition, providing a non-mediated “motor perception” based on a bidirectional flow of information along the mirror parieto-frontal circuits. State-of-the-art machine learning strategies for hand action identification have shown better (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  17.  6
    The Principle of Inverse Effectiveness in Audiovisual Speech Perception.Luuk P. H. van de Rijt, Anja Roye, Emmanuel A. M. Mylanus, A. John van Opstal & Marc M. van Wanrooij - 2019 - Frontiers in Human Neuroscience 13:468577.
    We assessed how synchronous speech listening and lipreading affects speech recognition in acoustic noise. In simple audiovisual perceptual tasks, inverse effectiveness is often observed, which holds that the weaker the unimodal stimuli, or the poorer their signal-to-noise ratio, the stronger the audiovisual benefit. So far, however, inverse effectiveness has not been demonstrated for complex audiovisual speech stimuli. Here we assess whether this multisensory integration effect can also be observed for the recognizability of spoken words. To that (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark   3 citations  
  18.  18
    Revisiting Human-Agent Communication: The Importance of Joint Co-construction and Understanding Mental States.Stefan Kopp & Nicole Krämer - 2021 - Frontiers in Psychology 12.
    The study of human-human communication and the development of computational models for human-agent communication have diverged significantly throughout the last decade. Yet, despite frequently made claims of “super-human performance” in, e.g., speech recognition or image processing, so far, no system is able to lead a half-decent coherent conversation with a human. In this paper, we argue that we must start to re-consider the hallmarks of cooperative communication and the core capabilities that we have (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark   2 citations  
  19.  64
    Automatic phonetic segmentation of Hindi speech using hidden Markov model.Archana Balyan, S. S. Agrawal & Amita Dev - 2012 - AI and Society 27 (4):543-549.
    In this paper, we study the performance of baseline hidden Markov model (HMM) for segmentation of speech signals. It is applied on single-speaker segmentation task, using Hindi speech database. The automatic phoneme segmentation framework evolved imitates the human phoneme segmentation process. A set of 44 Hindi phonemes were chosen for the segmentation experiment, wherein we used continuous density hidden Markov model (CDHMM) with a mixture of Gaussian distribution. The left-to-right topology with no skip states has been selected (...)
    Direct download (4 more)  
     
    Export citation  
     
    Bookmark  
  20.  21
    The Deuteros Plous, Simmias' Speech, and Socrates' Answer to Cebes in Plato's 'Phaedo'.Donald Ross - 1982 - Hermes 110 (1):19-25.
    There is growing recognition in Phaedo scholarship of a parallel between the deuteros plous passage and the introduction to Simmias' speech: both speak of attempting to discover or to learn the truth about things, and then, if that proves impossible, to resort to divine or human logoi, the former being the "safer" of the two. It is contended that that model governs Socrates reply to Cebes: he first tried to discover the truth about causes by himself; then (...)
    Direct download  
     
    Export citation  
     
    Bookmark  
  21.  38
    Ghost-in-the-Machine reveals human social signals for human–robot interaction.Sebastian Loth, Katharina Jettka, Manuel Giuliani & Jan P. de Ruiter - 2015 - Frontiers in Psychology 6.
    We used a new method called “Ghost-in-the-Machine” (GiM) to investigate social interactions with a robotic bartender taking orders for drinks and serving them. Using the GiM paradigm allowed us to identify how human participants recognize the intentions of customers on the basis of the output of the robotic recognizers. Specifically, we measured which recognizer modalities (e.g., speech, the distance to the bar) were relevant at different stages of the interaction. This provided insights into human social behavior necessary (...)
    Direct download (9 more)  
     
    Export citation  
     
    Bookmark   3 citations  
  22.  26
    Effects of Attention on the Strength of Lexical Influences on Speech Perception: Behavioral Experiments and Computational Mechanisms.Daniel Mirman, James L. McClelland, Lori L. Holt & James S. Magnuson - 2008 - Cognitive Science 32 (2):398-417.
    The effects of lexical context on phonological processing are pervasive and there have been indications that such effects may be modulated by attention. However, attentional modulation in speech processing is neither well documented nor well understood. Experiment 1 demonstrated attentional modulation of lexical facilitation of speech sound recognition when task and critical stimuli were identical across attention conditions. We propose modulation of lexical activation as a neurophysiologically plausible computational mechanism that can account for this type of modulation. (...)
    Direct download (4 more)  
     
    Export citation  
     
    Bookmark   4 citations  
  23.  29
    Intersections between Paul Ricoeur’s Conception of Narrative Identity and Mikhail Bakhtin’s Notion of the Polyphony of Speech.Małgorzata Hołda - 2016 - Forum Philosophicum: International Journal for Philosophy 21 (2):225-247.
    Proposing his conception of narrative identity in Oneself as Another, Paul Ricoeur holds that human life is comprehensible, once the story of a man’s life has actually been told, and it is the narrative of one’s life which constructs one’s identity. Developing his theory of heteroglossia and the polyphony of human speech, explicated chiefly in Speech Genres and The Dialogic Imagination, Mikhail Bakhtin recognizes the intrinsically intertwining character of utterance and response. According to him, utterance is (...)
    Direct download (3 more)  
     
    Export citation  
     
    Bookmark  
  24.  9
    Intersections between Paul Ricœur’s Conception of Narrative Identity and Mikhail Bakhtin’s Notion of the Polyphony of Speech.Małgorzata Hołda - 2017 - Forum Philosophicum: International Journal for Philosophy 21 (2):227-249.
    Proposing his conception of narrative identity in Oneself as Another, Paul Ricœur holds that human life is comprehensible, once the story of a man’s life has actually been told, and it is the narrative of one’s life which constructs one’s identity. Developing his theory of heteroglossia and the polyphony of human speech, explicated chiefly in Speech Genres and The Dialogic Imagination, Mikhail Bakhtin recognizes the intrinsically intertwining character of utterance and response. According to him, utterance is (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  25.  12
    Intersections between Paul Ricœur’s Conception of Narrative Identity and Mikhail Bakhtin’s Notion of the Polyphony of Speech.Małgorzata Hołda - 2017 - Forum Philosophicum: International Journal for Philosophy 21 (2):227-249.
    Proposing his conception of narrative identity in Oneself as Another, Paul Ricœur holds that human life is comprehensible, once the story of a man’s life has actually been told, and it is the narrative of one’s life which constructs one’s identity. Developing his theory of heteroglossia and the polyphony of human speech, explicated chiefly in Speech Genres and The Dialogic Imagination, Mikhail Bakhtin recognizes the intrinsically intertwining character of utterance and response. According to him, utterance is (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  26.  29
    Lexical and Sublexical Units in Speech Perception.Ibrahima Giroux & Arnaud Rey - 2009 - Cognitive Science 33 (2):260-272.
    Saffran, Newport, and Aslin (1996a) found that human infants are sensitive to statistical regularities corresponding to lexical units when hearing an artificial spoken language. Two sorts of segmentation strategies have been proposed to account for this early word‐segmentation ability: bracketing strategies, in which infants are assumed to insert boundaries into continuous speech, and clustering strategies, in which infants are assumed to group certain speech sequences together into units (Swingley, 2005). In the present study, we test the predictions (...)
    No categories
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark   18 citations  
  27.  72
    Maasai Concepts of Personhood: The Roles of Recognition, Community, and Individuality.Gail M. Presbey - 2002 - International Studies in Philosophy 34 (2):57-82.
    There has been a debate, popularized by Ifenyi Menkiti and Kwame Gyekye, regarding philosophical understandings of the human person in Africa. The debate revolves around the saying "So and so is not a person." Gyekye convincingly argues that the saying is a manner of speech, intended to be a moral evaluation of a person's actions. Menkiti, however, goes further and suggests that many of the African conceptions of a person are based on a dynamic understanding of the self. (...)
    Direct download (3 more)  
     
    Export citation  
     
    Bookmark   5 citations  
  28.  12
    A State-of-the-Art Review of EEG-Based Imagined Speech Decoding.Diego Lopez-Bernal, David Balderas, Pedro Ponce & Arturo Molina - 2022 - Frontiers in Human Neuroscience 16:867281.
    Currently, the most used method to measure brain activity under a non-invasive procedure is the electroencephalogram (EEG). This is because of its high temporal resolution, ease of use, and safety. These signals can be used under a Brain Computer Interface (BCI) framework, which can be implemented to provide a new communication channel to people that are unable to speak due to motor disabilities or other neurological diseases. Nevertheless, EEG-based BCI systems have presented challenges to be implemented in real life situations (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  29. Human Rights, Women's Rights, Gender Mainstreaming, and Diversity: The Language Question.Yvanka B. Raynova - 2015 - In Community, Praxis, and Values in a Postmetaphysical Age: Studies on Exclusion and Social Integration in Feminist Theory and Contemporary Philosophy. Axia Academic Publishers. pp. 38-89.
    In the following study the author goes back to the beginnings of the Women's Rights movements in order to pose the question on gender equality by approaching it through the prism of language as a powerful tool in human rights battles. This permits her to show the deep interrelation between women's struggle for recognition and some particular women rights, like the "feminization" of professional titles and the implementation of a gender sensitive language. Hence she argues the thesis that (...)
     
    Export citation  
     
    Bookmark  
  30.  77
    Influences of lexical tone and pitch on word recognition in bilingual infants.Leher Singh & Joanne Foong - 2012 - Cognition 124 (2):128-142.
  31.  25
    Intelligent service robots for elderly or disabled people and human dignity: legal point of view.Katarzyna Pfeifer-Chomiczewska - 2023 - AI and Society 38 (2):789-800.
    This article aims to present the problem of the impact of artificial intelligence on respect for human dignity in the sphere of care for people who, for various reasons, are described as particularly vulnerable, especially seniors and people with various disabilities. In recent years, various initiatives and works have been undertaken on the European scene to define the directions in which the development and use of artificial intelligence should go. According to the human-centric approach, artificial intelligence should be (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark   1 citation  
  32.  92
    Merging information in speech recognition: Feedback is never necessary.Dennis Norris, James M. McQueen & Anne Cutler - 2000 - Behavioral and Brain Sciences 23 (3):299-325.
    Top-down feedback does not benefit speech recognition; on the contrary, it can hinder it. No experimental data imply that feedback loops are required for speech recognition. Feedback is accordingly unnecessary and spoken word recognition is modular. To defend this thesis, we analyse lexical involvement in phonemic decision making. TRACE (McClelland & Elman 1986), a model with feedback from the lexicon to prelexical processes, is unable to account for all the available data on phonemic decision making. (...)
    Direct download (8 more)  
     
    Export citation  
     
    Bookmark   46 citations  
  33.  8
    Restricted Speech Recognition in Noise and Quality of Life of Hearing-Impaired Children and Adolescents With Cochlear Implants – Need for Studies Addressing This Topic With Valid Pediatric Quality of Life Instruments.Maria Huber & Clara Havas - 2019 - Frontiers in Psychology 10.
    Cochlear implants (CI) support the development of oral language in hearing-impaired children. However, even with CI, speech recognition in noise (SRiN) is limited. This raised the question, whether these restrictions are related to the quality of life (QoL) of children and adolescents with CI and how SRiN and QoL are related to each other. As a result of a systematic literature research only three studies were found, indicating positive moderating effects between SRiN and QoL of young CI users. (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  34.  24
    Automatic Speech Recognition: A Comprehensive Survey.Arbana Kadriu & Amarildo Rista - 2020 - Seeu Review 15 (2):86-112.
    Speech recognition is an interdisciplinary subfield of natural language processing (NLP) that facilitates the recognition and translation of spoken language into text by machine. Speech recognition plays an important role in digital transformation. It is widely used in different areas such as education, industry, and healthcare and has recently been used in many Internet of Things and Machine Learning applications. The process of speech recognition is one of the most difficult processes in computer (...)
    No categories
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  35.  49
    Intelligent Aging Home Control Method and System for Internet of Things Emotion Recognition.Xu Wu & Qian Zhang - 2022 - Frontiers in Psychology 13.
    To solve a series of pension problems caused by aging, based on the emotional recognition of the Internet of Things, the control method and system research of smart homes are proposed. This article makes a detailed analysis and research on the necessity, feasibility, and how to realize speech emotion recognition technology in smart families, introduces the definition and classification of emotion, and puts forward five main emotions to be recognized in speech emotion recognition based on (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  36.  6
    Human rights and ethics: proceedings of the 22nd IVR World Congress, Granada 2005, volume III = Derechos humanos y ética.Andrés Ollero (ed.) - 2007 - Stuttgart: Franz Steiner Verlag.
    This volume reflects on questions of human rights in the context of globalization. The essays responding to this subject are rich and varied: they focus on legal acceptance as well as consequences of human rights with regard to social rights and the necessary protection of the environment connected or close to those rights. Another approach to the subject featured in the volume is the legal recognition and the consideration of human rights as moral rights. With concepts (...)
    Direct download  
     
    Export citation  
     
    Bookmark  
  37. Speech recognition: Statistical methods.L. R. Rabiner & B. H. Juang - 2006 - In Keith Brown (ed.), Encyclopedia of Language and Linguistics. Elsevier. pp. 1--18.
     
    Export citation  
     
    Bookmark  
  38.  15
    A Review on Five Recent and Near-Future Developments in Computational Processing of Emotion in the Human Voice.Dagmar M. Schuller & Björn W. Schuller - 2020 - Emotion Review 13 (1):44-50.
    We provide a short review on the recent and near-future developments of computational processing of emotion in the voice, highlighting self-learning of representations moving continuously away from traditional expert-crafted or brute-forced feature representations to end-to-end learning, a movement towards the coupling of analysis and synthesis of emotional voices to foster better mutual understanding, weakly supervised learning at a large scale, transfer learning from related domains such as speech recognition or cross-modal transfer learning, and reinforced learning through interactive applications (...)
    Direct download  
     
    Export citation  
     
    Bookmark   2 citations  
  39. Speech recognition technology.F. Beaufays, H. Bourlard, Horacio Franco & Nelson Morgan - 2002 - In M. Arbib (ed.), The Handbook of Brain Theory and Neural Networks. MIT Press.
     
    Export citation  
     
    Bookmark  
  40.  3
    Longitudinal Speech Recognition in Noise in Children: Effects of Hearing Status and Vocabulary.Elizabeth A. Walker, Caitlin Sapp, Jacob J. Oleson & Ryan W. McCreery - 2019 - Frontiers in Psychology 10.
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  41.  27
    Modelling asynchrony in automatic speech recognition using loosely coupled hidden Markov models.H. J. Nock & S. J. Young - 2002 - Cognitive Science 26 (3):283-301.
    Hidden Markov models (HMMs) have been successful for modelling the dynamics of carefully dictated speech, but their performance degrades severely when used to model conversational speech. Since speech is produced by a system of loosely coupled articulators, stochastic models explicitly representing this parallelism may have advantages for automatic speech recognition (ASR), particularly when trying to model the phonological effects inherent in casual spontaneous speech. This paper presents a preliminary feasibility study of one such model (...)
    Direct download (4 more)  
     
    Export citation  
     
    Bookmark  
  42.  16
    Masked Speech Recognition in School-Age Children.Lori J. Leibold & Emily Buss - 2019 - Frontiers in Psychology 10.
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  43.  24
    Deep learning approach to text analysis for human emotion detection from big data.Jia Guo - 2022 - Journal of Intelligent Systems 31 (1):113-126.
    Emotional recognition has arisen as an essential field of study that can expose a variety of valuable inputs. Emotion can be articulated in several means that can be seen, like speech and facial expressions, written text, and gestures. Emotion recognition in a text document is fundamentally a content-based classification issue, including notions from natural language processing (NLP) and deep learning fields. Hence, in this study, deep learning assisted semantic text analysis (DLSTA) has been proposed for human (...)
    Direct download  
     
    Export citation  
     
    Bookmark  
  44.  9
    Discriminatively trained continuous Hindi speech recognition using integrated acoustic features and recurrent neural network language modeling.R. K. Aggarwal & A. Kumar - 2020 - Journal of Intelligent Systems 30 (1):165-179.
    This paper implements the continuous Hindi Automatic Speech Recognition (ASR) system using the proposed integrated features vector with Recurrent Neural Network (RNN) based Language Modeling (LM). The proposed system also implements the speaker adaptation using Maximum-Likelihood Linear Regression (MLLR) and Constrained Maximum likelihood Linear Regression (C-MLLR). This system is discriminatively trained by Maximum Mutual Information (MMI) and Minimum Phone Error (MPE) techniques with 256 Gaussian mixture per Hidden Markov Model(HMM) state. The training of the baseline system has been (...)
    No categories
    Direct download  
     
    Export citation  
     
    Bookmark  
  45. On Human Communication. [REVIEW]M. E. - 1967 - Review of Metaphysics 20 (4):714-714.
    An excellent introduction to communication theory, this book is a comprehensive study of its subject; fields such as linguistics, logic, mathematics, and psychology are considered in terms of their relevance for communication theory. No material that appeared in the first edition has been deleted from this second edition, but some comments have been added, some figures updated, and the bibliography extended to include the new publications in the field. Cherry begins with an examination of the concept of "communication"; he also (...)
    No categories
     
    Export citation  
     
    Bookmark  
  46.  18
    Human speech: A tinkerer's delight.Harvey M. Sussman, David Fruchter, Jon Hilbert & Joseph Sirosh - 1998 - Behavioral and Brain Sciences 21 (2):287-295.
    The most frequent criticism of the target article is the lack of clear separability of human speech data relative to neuroethological data. A rationalization for this difference was sought in the tinkered nature of such new adaptations as human speech. Basic theoretical premises were defended, and new data were presented to support a claim that speakers maintain a low-noise relationship between F2 transition onset and offset frequencies for stops in pre-vocalic positions through articulatory choices. It remains (...)
    Direct download (6 more)  
     
    Export citation  
     
    Bookmark  
  47.  6
    Human Posture Recognition and Estimation Method Based on 3D Multiview Basketball Sports Dataset.Xuhui Song & Linyuan Fan - 2021 - Complexity 2021:1-10.
    In traditional 3D reconstruction methods, using a single view to predict the 3D structure of an object is a very difficult task. This research mainly discusses human pose recognition and estimation based on 3D multiview basketball sports dataset. The convolutional neural network framework used in this research is VGG11, and the basketball dataset Image Net is used for pretraining. This research uses some modules of the VGG11 network. For different feature fusion methods, different modules of the VGG11 network (...)
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark  
  48.  35
    Merging information versus speech recognition.Irene Appelbaum - 2000 - Behavioral and Brain Sciences 23 (3):325-326.
    Norris, McQueen & Cutler claim that all known speech recognition data can be accounted for with their autonomous model, “Merge.” But this claim is doubly misleading. (1) Although speech recognition is autonomous in their view, the Merge model is not. (2) The body of data which the Merge model accounts for, is not, in their view, speech recognition data. Footnotes1 Author is also affiliated with the Center for the Study of Language and Information, Stanford (...)
    Direct download (4 more)  
     
    Export citation  
     
    Bookmark  
  49.  10
    Human Face Recognition in Horses: Data in Favor of a Holistic Process.Léa Lansade, Violaine Colson, Céline Parias, Fabrice Reigner, Aline Bertin & Ludovic Calandreau - 2020 - Frontiers in Psychology 11.
    Direct download (2 more)  
     
    Export citation  
     
    Bookmark   1 citation  
  50.  34
    Perceptual units in speech recognition.Dominic W. Massaro - 1974 - Journal of Experimental Psychology 102 (2):199.
1 — 50 / 988