%Aigaion2 BibTeX export from Idiap Publications
%Friday 03 May 2024 07:13:23 AM

@INPROCEEDINGS{Kabil_INTERSPEECH_2018,
         author = {Kabil, Selen Hande and Muckenhirn, Hannah and Magimai.-Doss, Mathew},
       keywords = {automatic gender recognition, Convolutional Neural Networks, end-to-end training, multilayer perceptron},
       projects = {UNITS, Idiap},
          month = sep,
          title = {On Learning to Identify Genders from Raw Speech Signal Using CNNs},
      booktitle = {Proceedings of Interspeech},
           year = {2018},
          pages = {287-291},
       location = {Hyderabad, INDIA},
           issn = {2308-457X},
           isbn = {978-1-5108-7221-9},
            doi = {10.21437/Interspeech.2018-1240},
       abstract = {Automatic Gender Recognition (AGR) is the task of identifying the gender of a speaker given a speech signal. Standard approaches extract features like fundamental frequency and cepstral features from the speech signal and train a binary classifier. Inspired from recent works in the area of automatic speech recognition (ASR), speaker recognition and presentation attack detection, we present a novel approach where relevant features and classifier are jointly learned from the raw speech signal in end-to-end manner. We propose a convolutional neural networks (CNN) based gender classifier that consists of: (1) convolution layers, which can be interpreted as a feature learning stage and (2) a multilayer perceptron (MLP), which can be interpreted as a classification stage. The system takes raw speech signal as input, and outputs gender posterior probabilities. Experimental studies conducted on two datasets, namely AVspoof and ASVspoof 2015, with different architectures show that with simple architectures the proposed approach yields better system than standard acoustic features based approach. Further analysis of the CNNs show that the CNNs learn formant and fundamental frequency information for gender identification.},
            pdf = {https://publications.idiap.ch/attachments/papers/2018/Kabil_INTERSPEECH_2018.pdf}
}