Articles

Sort by
Previous Page Page 1 of 12 Next Page
  1. Research

    Efficiency of chosen speech descriptors in relation to emotion recognition

    This research paper presents parametrization of emotional speech using a pool of common features utilized in emotion recognition such as fundamental frequency, formants, energy, MFCC, PLP, and LPC coefficients. T...

    Dorota Kamińska, Tomasz Sapiński and Gholamreza Anbarjafari

    EURASIP Journal on Audio, Speech, and Music Processing 2017 2017:3

    Published on: 20 February 2017

  2. Research

    Cantor Digitalis: chironomic parametric synthesis of singing

    Cantor Digitalis is a performative singing synthesizer that is composed of two main parts: a chironomic control interface and a parametric voice synthesizer. The control interface is based on a pen/touch graph...

    Lionel Feugère, Christophe d’Alessandro, Boris Doval and Olivier Perrotin

    EURASIP Journal on Audio, Speech, and Music Processing 2017 2017:2

    Published on: 23 January 2017

  3. Research

    New approach for determining the QoS of MP3-coded voice signals in IP networks

    Present-day IP transport platforms being what they are, it will never be possible to rule out conflicts between the available services. The logical consequence of this assertion is the inevitable conclusion th...

    Tadeus Uhl, Stefan Paulsen and Krzysztof Nowicki

    EURASIP Journal on Audio, Speech, and Music Processing 2017 2017:1

    Published on: 19 January 2017

  4. Research

    Structure of pauses in speech in the context of speaker verification and classification of speech type

    Statistics of pauses appearing in Polish as a potential source of biometry information for automatic speaker recognition were described. The usage of three main types of acoustic pauses (silent, filled and bre...

    Magdalena Igras-Cybulska, Bartosz Ziółko, Piotr Żelasko and Marcin Witkowski

    EURASIP Journal on Audio, Speech, and Music Processing 2016 2016:18

    Published on: 9 November 2016

  5. Research

    Fast fundamental frequency determination via adaptive autocorrelation

    We present an algorithm for the estimation of fundamental frequencies in voiced audio signals. The method is based on an autocorrelation of a signal with a segment of the same signal. During operation, frequen...

    Michael Staudacher, Viktor Steixner, Andreas Griessner and Clemens Zierhofer

    EURASIP Journal on Audio, Speech, and Music Processing 2016 2016:17

    Published on: 24 October 2016

  6. Research

    Single microphone speech separation by diffusion-based HMM estimation

    We present a novel non-iterative and rigorously motivated approach for estimating hidden Markov models (HMMs) and factorial hidden Markov models (FHMMs) of high-dimensional signals. Our approach utilizes the a...

    Yochay R. Yeminy, Yosi Keller and Sharon Gannot

    EURASIP Journal on Audio, Speech, and Music Processing 2016 2016:16

    Published on: 18 October 2016

  7. Research

    A hybrid input-type recurrent neural network for LVCSR language modeling

    Substantial amounts of resources are usually required to robustly develop a language model for an open vocabulary speech recognition system as out-of-vocabulary (OOV) words can hurt recognition accuracy. In th...

    Vataya Chunwijitra, Ananlada Chotimongkol and Chai Wutiwiwatchai

    EURASIP Journal on Audio, Speech, and Music Processing 2016 2016:15

    Published on: 8 August 2016

  8. Research

    Voice activity detection algorithm based on long-term pitch information

    A new voice activity detection algorithm based on long-term pitch divergence is presented. The long-term pitch divergence not only decomposes speech signals with a bionic decomposition but also makes full use ...

    Xu-Kui Yang, Liang He, Dan Qu and Wei-Qiang Zhang

    EURASIP Journal on Audio, Speech, and Music Processing 2016 2016:14

    Published on: 7 July 2016

  9. Research

    JND-based spatial parameter quantization of multichannel audio signals

    In multichannel spatial audio coding (SAC), the accurate representations of virtual sounds and the efficient compressions of spatial parameters are the key to perfect reproduction of spatial sound effects in 3...

    Li Gao, Ruimin Hu, Xiaochen Wang, Gang Li, Yuhong Yang and Weiping Tu

    EURASIP Journal on Audio, Speech, and Music Processing 2016 2016:13

    Published on: 21 May 2016

  10. Research

    Audio bandwidth extension using ensemble of recurrent neural networks

    In audio communication systems, the perceptual audio quality of the reproduced audio signals such as the naturalness of the sound is limited by the available audio bandwidth. In this paper, a wideband to super...

    Xin Liu and Chang-Chun Bao

    EURASIP Journal on Audio, Speech, and Music Processing 2016 2016:12

    Published on: 12 May 2016

  11. Research

    Wise teachers train better DNN acoustic models

    Automatic speech recognition is becoming more ubiquitous as recognition performance improves, capable devices increase in number, and areas of new application open up. Neural network acoustic models that can u...

    Ryan Price, Ken-ichi Iso and Koichi Shinoda

    EURASIP Journal on Audio, Speech, and Music Processing 2016 2016:10

    Published on: 12 April 2016

  12. Research

    Semi-supervised feature selection for audio classification based on constraint compensated Laplacian score

    Audio classification, classifying audio segments into broad categories such as speech, non-speech, and silence, is an important front-end problem in speech signal processing. Dozens of features have been propo...

    Xu-Kui Yang, Liang He, Dan Qu, Wei-Qiang Zhang and Michael T. Johnson

    EURASIP Journal on Audio, Speech, and Music Processing 2016 2016:9

    Published on: 15 March 2016

  13. Research

    Developing a unit selection voice given audio without corresponding text

    Today, a large amount of audio data is available on the web in the form of audiobooks, podcasts, video lectures, video blogs, news bulletins, etc. In addition, we can effortlessly record and store audio data s...

    Tejas Godambe, Sai Krishna Rallabandi, Suryakanth V. Gangashetty, Ashraf Alkhairy and Afshan Jafri

    EURASIP Journal on Audio, Speech, and Music Processing 2016 2016:6

    Published on: 1 March 2016

  14. Research

    iSargam: music notation representation for Indian Carnatic music

    Indian classical music, including its two varieties, Carnatic and Hindustani music, has a rich music tradition and enjoys a wide audience from various parts of the world. The Carnatic music which is more popul...

    Stanly Mammen, Ilango Krishnamurthi, A. Jalaja Varma and G. Sujatha

    EURASIP Journal on Audio, Speech, and Music Processing 2016 2016:5

    Published on: 16 February 2016

  15. Research

    Hybrid statistical/unit-selection Turkish speech synthesis using suffix units

    Unit selection based text-to-speech synthesis (TTS) has been the dominant TTS approach of the last decade. Despite its success, unit selection approach has its disadvantages. One of the most significant disadv...

    Cenk Demiroğlu and Ekrem Güner

    EURASIP Journal on Audio, Speech, and Music Processing 2016 2016:4

    Published on: 2 February 2016

  16. Research

    Grid-based approximation for voice conversion in low resource environments

    The goal of voice conversion is to modify a source speaker’s speech to sound as if spoken by a target speaker. Common conversion methods are based on Gaussian mixture modeling (GMM). They aim to statistically ...

    Hadas Benisty, David Malah and Koby Crammer

    EURASIP Journal on Audio, Speech, and Music Processing 2016 2016:3

    Published on: 21 January 2016

  17. Research

    Detecting fingering of overblown flute sound using sparse feature learning

    In woodwind instruments such as a flute, producing a higher-pitched tone than a standard tone by increasing the blowing pressure is called overblowing, and this allows several distinct fingerings for the same ...

    Yoonchang Han and Kyogu Lee

    EURASIP Journal on Audio, Speech, and Music Processing 2016 2016:2

    Published on: 21 January 2016

  18. Research

    Comparison of ALBAYZIN query-by-example spoken term detection 2012 and 2014 evaluations

    Query-by-example spoken term detection (QbE STD) aims at retrieving data from a speech repository given an acoustic query containing the term of interest as input. Nowadays, it is receiving much interest due t...

    Javier Tejedor, Doroteo T. Toledano, Paula Lopez-Otero, Laura Docio-Fernandez and Carmen Garcia-Mateo

    EURASIP Journal on Audio, Speech, and Music Processing 2016 2016:1

    Published on: 13 January 2016

  19. Research

    Speech signal modeling using multivariate distributions

    Using a proper distribution function for speech signal or for its representations is of crucial importance in statistical-based speech processing algorithms. Although the most commonly used probability density...

    Ali Aroudi, Hadi Veisi, Hossein Sameti and Zahra Mafakheri

    EURASIP Journal on Audio, Speech, and Music Processing 2015 2015:35

    Published on: 30 December 2015

  20. Research

    Albayzín-2014 evaluation: audio segmentation and classification in broadcast news domains

    Audio segmentation is important as a pre-processing task to improve the performance of many speech technology tasks and, therefore, it has an undoubted research interest. This paper describes the database, the...

    Diego Castán, David Tavarez, Paula Lopez-Otero, Javier Franco-Pedroso, Héctor Delgado, Eva Navas, Laura Docio-Fernández, Daniel Ramos, Javier Serrano, Alfonso Ortega and Eduardo Lleida

    EURASIP Journal on Audio, Speech, and Music Processing 2015 2015:33

    Published on: 1 December 2015

Previous Page Page 1 of 12 Next Page