Articles

Sort by
Previous Page Page 1 of 12 Next Page
  1. Research

    A computational study of auditory models in music recognition tasks for normal-hearing and hearing-impaired listeners

    The benefit of auditory models for solving three music recognition tasks—onset detection, pitch estimation, and instrument recognition—is analyzed. Appropriate features are introduced which enable the use of s...

    Klaus Friedrichs, Nadja Bauer, Rainer Martin and Claus Weihs

    EURASIP Journal on Audio, Speech, and Music Processing 2017 2017:7

    Published on: 2 March 2017

  2. Research

    Context-dependent factored language models

    The incorporation of grammatical information into speech recognition systems is often used to increase performance in morphologically rich languages. However, this introduces demands for sufficiently large tra...

    Gregor Donaj and Zdravko Kačič

    EURASIP Journal on Audio, Speech, and Music Processing 2017 2017:6

    Published on: 28 February 2017

  3. Research

    Efficiency of chosen speech descriptors in relation to emotion recognition

    This research paper presents parametrization of emotional speech using a pool of common features utilized in emotion recognition such as fundamental frequency, formants, energy, MFCC, PLP, and LPC coefficients. T...

    Dorota Kamińska, Tomasz Sapiński and Gholamreza Anbarjafari

    EURASIP Journal on Audio, Speech, and Music Processing 2017 2017:3

    Published on: 20 February 2017

  4. Research

    Cantor Digitalis: chironomic parametric synthesis of singing

    Cantor Digitalis is a performative singing synthesizer that is composed of two main parts: a chironomic control interface and a parametric voice synthesizer. The control interface is based on a pen/touch graph...

    Lionel Feugère, Christophe d’Alessandro, Boris Doval and Olivier Perrotin

    EURASIP Journal on Audio, Speech, and Music Processing 2017 2017:2

    Published on: 23 January 2017

  5. Research

    New approach for determining the QoS of MP3-coded voice signals in IP networks

    Present-day IP transport platforms being what they are, it will never be possible to rule out conflicts between the available services. The logical consequence of this assertion is the inevitable conclusion th...

    Tadeus Uhl, Stefan Paulsen and Krzysztof Nowicki

    EURASIP Journal on Audio, Speech, and Music Processing 2017 2017:1

    Published on: 19 January 2017

  6. Research

    Structure of pauses in speech in the context of speaker verification and classification of speech type

    Statistics of pauses appearing in Polish as a potential source of biometry information for automatic speaker recognition were described. The usage of three main types of acoustic pauses (silent, filled and bre...

    Magdalena Igras-Cybulska, Bartosz Ziółko, Piotr Żelasko and Marcin Witkowski

    EURASIP Journal on Audio, Speech, and Music Processing 2016 2016:18

    Published on: 9 November 2016

  7. Research

    Fast fundamental frequency determination via adaptive autocorrelation

    We present an algorithm for the estimation of fundamental frequencies in voiced audio signals. The method is based on an autocorrelation of a signal with a segment of the same signal. During operation, frequen...

    Michael Staudacher, Viktor Steixner, Andreas Griessner and Clemens Zierhofer

    EURASIP Journal on Audio, Speech, and Music Processing 2016 2016:17

    Published on: 24 October 2016

  8. Research

    Single microphone speech separation by diffusion-based HMM estimation

    We present a novel non-iterative and rigorously motivated approach for estimating hidden Markov models (HMMs) and factorial hidden Markov models (FHMMs) of high-dimensional signals. Our approach utilizes the a...

    Yochay R. Yeminy, Yosi Keller and Sharon Gannot

    EURASIP Journal on Audio, Speech, and Music Processing 2016 2016:16

    Published on: 18 October 2016

  9. Research

    A hybrid input-type recurrent neural network for LVCSR language modeling

    Substantial amounts of resources are usually required to robustly develop a language model for an open vocabulary speech recognition system as out-of-vocabulary (OOV) words can hurt recognition accuracy. In th...

    Vataya Chunwijitra, Ananlada Chotimongkol and Chai Wutiwiwatchai

    EURASIP Journal on Audio, Speech, and Music Processing 2016 2016:15

    Published on: 8 August 2016

  10. Research

    Voice activity detection algorithm based on long-term pitch information

    A new voice activity detection algorithm based on long-term pitch divergence is presented. The long-term pitch divergence not only decomposes speech signals with a bionic decomposition but also makes full use ...

    Xu-Kui Yang, Liang He, Dan Qu and Wei-Qiang Zhang

    EURASIP Journal on Audio, Speech, and Music Processing 2016 2016:14

    Published on: 7 July 2016

  11. Research

    JND-based spatial parameter quantization of multichannel audio signals

    In multichannel spatial audio coding (SAC), the accurate representations of virtual sounds and the efficient compressions of spatial parameters are the key to perfect reproduction of spatial sound effects in 3...

    Li Gao, Ruimin Hu, Xiaochen Wang, Gang Li, Yuhong Yang and Weiping Tu

    EURASIP Journal on Audio, Speech, and Music Processing 2016 2016:13

    Published on: 21 May 2016

  12. Research

    Audio bandwidth extension using ensemble of recurrent neural networks

    In audio communication systems, the perceptual audio quality of the reproduced audio signals such as the naturalness of the sound is limited by the available audio bandwidth. In this paper, a wideband to super...

    Xin Liu and Chang-Chun Bao

    EURASIP Journal on Audio, Speech, and Music Processing 2016 2016:12

    Published on: 12 May 2016

  13. Research

    Wise teachers train better DNN acoustic models

    Automatic speech recognition is becoming more ubiquitous as recognition performance improves, capable devices increase in number, and areas of new application open up. Neural network acoustic models that can u...

    Ryan Price, Ken-ichi Iso and Koichi Shinoda

    EURASIP Journal on Audio, Speech, and Music Processing 2016 2016:10

    Published on: 12 April 2016

  14. Research

    Semi-supervised feature selection for audio classification based on constraint compensated Laplacian score

    Audio classification, classifying audio segments into broad categories such as speech, non-speech, and silence, is an important front-end problem in speech signal processing. Dozens of features have been propo...

    Xu-Kui Yang, Liang He, Dan Qu, Wei-Qiang Zhang and Michael T. Johnson

    EURASIP Journal on Audio, Speech, and Music Processing 2016 2016:9

    Published on: 15 March 2016

  15. Research

    Developing a unit selection voice given audio without corresponding text

    Today, a large amount of audio data is available on the web in the form of audiobooks, podcasts, video lectures, video blogs, news bulletins, etc. In addition, we can effortlessly record and store audio data s...

    Tejas Godambe, Sai Krishna Rallabandi, Suryakanth V. Gangashetty, Ashraf Alkhairy and Afshan Jafri

    EURASIP Journal on Audio, Speech, and Music Processing 2016 2016:6

    Published on: 1 March 2016

  16. Research

    iSargam: music notation representation for Indian Carnatic music

    Indian classical music, including its two varieties, Carnatic and Hindustani music, has a rich music tradition and enjoys a wide audience from various parts of the world. The Carnatic music which is more popul...

    Stanly Mammen, Ilango Krishnamurthi, A. Jalaja Varma and G. Sujatha

    EURASIP Journal on Audio, Speech, and Music Processing 2016 2016:5

    Published on: 16 February 2016

  17. Research

    Hybrid statistical/unit-selection Turkish speech synthesis using suffix units

    Unit selection based text-to-speech synthesis (TTS) has been the dominant TTS approach of the last decade. Despite its success, unit selection approach has its disadvantages. One of the most significant disadv...

    Cenk Demiroğlu and Ekrem Güner

    EURASIP Journal on Audio, Speech, and Music Processing 2016 2016:4

    Published on: 2 February 2016

  18. Research

    Grid-based approximation for voice conversion in low resource environments

    The goal of voice conversion is to modify a source speaker’s speech to sound as if spoken by a target speaker. Common conversion methods are based on Gaussian mixture modeling (GMM). They aim to statistically ...

    Hadas Benisty, David Malah and Koby Crammer

    EURASIP Journal on Audio, Speech, and Music Processing 2016 2016:3

    Published on: 21 January 2016

  19. Research

    Detecting fingering of overblown flute sound using sparse feature learning

    In woodwind instruments such as a flute, producing a higher-pitched tone than a standard tone by increasing the blowing pressure is called overblowing, and this allows several distinct fingerings for the same ...

    Yoonchang Han and Kyogu Lee

    EURASIP Journal on Audio, Speech, and Music Processing 2016 2016:2

    Published on: 21 January 2016

Previous Page Page 1 of 12 Next Page