5
978-1-4244-6742-6/10/$26.00 ©2010 IEEE Classification of voice aging using ANN and glottal signal parameters Leonardo Alfredo Forero Mendoza 1 , Edson Cataldo 2 , Marley Vellasco 1 , Marco Aurélio Silva 1 1 Eletrical Engineering Department Pontificia Universidade Catolica do Rio de Janeiro Rio de Janeiro, Brazil 2 Telecommunications Engineering Deparment Universidade Federal Fluminense Niterói, RJ Brasil [email protected] Alvaro David Orjuela Cañón 1,2 , Jose Manoel de Seixas 1 1 Eletrical Engineering Department Universidade Federal do Rio de Janeiro, UFRJ Rio de Janeiro, Brazil 2 Electronics and Biomedical Engineering Department Universidad Antonio Nariño Bogotá. Colombia [email protected] Abstract— Classification of voice aging has many applications in health care and geriatrics. This work focuses on finding the most relevant parameters to classify voice aging. The most significant parameters extracted from the glottal signal are chosen to identify the voice aging process of men and women. After analyzing their statistics, the chosen parameters are used as entries to a neural network to classify male and female Brazilian speakers in three different age groups: young (from 15 to 30 years old), adult (from 31 to 60 years old), and senior (from 61 to 90 years old). The corpus used for this work was composed by one hundred and twenty Brazilian speakers (both males and females) of different ages. As compared to similar works, we employ a larger corpus and obtain a superior classification rate. Keywords:artificial neural networks ANN, inverse filtering, glottal signal parameters. I. INTRODUCTION The human aging is of great interest in various areas of research in voice, because the aging makes human body vulnerable, which directly reflects in voice production [1]. Some physiological process are continuos and irreversible, the main changes occur in the elderly. As causes of aging voice, we can cite the changes that occur in the body, for example, loss of lung capacity, changes in the epithelium of the larynx, calcification of cartilage and changes in the vibrations of the vocal folds. From 1995 to 2005, the life expectancy of Brazilian people has increased more than 8% [1]. This scenario is a global tendency, resulting in the need for developing new treatments to provide better life quality to senior people, including care for their voice. The most common method for extracting voice features is directly from the voice signal. However, many researchers have looked for some characteristics extracted from the so called glottal signal, which is the signal obtained just after the vocal folds and before the vocal tract. Since one of the consequences of aging is the change of the vocal fold structures, this signal is, therefore, important because it preserves characteristics from the movement of the vocal folds without the influence of the vocal tract. Moreover, it is known that some parameters extracted from the glottal signal can help the diagnosis of pathologies of the vocal folds better than those ones extracted from the voice signal [2]. Nowadays, obtaining this signal is easier due to the development of algorithms that can perform an inverse filtering from the voice signal, eliminating the influence of the vocal tract [2] [3]. Before, it used to be necessary to use equipment coupled with micro cameras to record sounds just after air passed from the vocal folds. This was an invasive technique and very difficult to be performed. In previous work, it has been used methods to classification of voice as Bayes, Hidden, Markov Models (HMM), Gaussian Mixture Models (GMM) and neural networks, using as input Mel-frequency cepstral Coefficients (MFCC) and the parameters jitter and shimmer, adding 37 parameters entry, using bases of 12 recordings for men and women not getting the results that rise from 80% [4]. The MFCC have proven effectiveness for speaker recognition [6], but their performance is not equal to the classification of voice aging. Comparing these methods for classification of voice aging, the best results are obtained with the techniques of neural networks and support vector machine (SVM). Differently, when the objective is the recognition of speaker, the best results are obtained with GMM and HMM methods [5] [6]. This is because for classification of voice aging are not as important temporal features of the voice and to change the voice signal with age [5]. An advantage of this work is the number of examples of voice speakers used in the classification, using only the most relevant parameters as input in a neural network. II. FEATURES FROM THE GLOTTAL SIGNAL A. The glottal signal The voice signal, particularly the one related to voiced sounds, e.g. vowels, starts with the contraction-expansion of the lungs, generating a pressure difference between the air in the lungs and the air near the mouth. The airflow created passes

[IEEE 2010 IEEE ANDESCON - Bogota, Colombia (2010.09.15-2010.09.17)] 2010 IEEE ANDESCON - Classification of voice aging using ANN and glottal signal parameters

Embed Size (px)

Citation preview

Page 1: [IEEE 2010 IEEE ANDESCON - Bogota, Colombia (2010.09.15-2010.09.17)] 2010 IEEE ANDESCON - Classification of voice aging using ANN and glottal signal parameters

978-1-4244-6742-6/10/$26.00 ©2010 IEEE

Classification of voice aging using ANN and glottal signal parameters

Leonardo Alfredo Forero Mendoza1, Edson Cataldo2, Marley Vellasco1, Marco Aurélio Silva1

1 Eletrical Engineering Department Pontificia Universidade Catolica do Rio de Janeiro

Rio de Janeiro, Brazil 2 Telecommunications Engineering Deparment

Universidade Federal Fluminense Niterói, RJ Brasil

[email protected]

Alvaro David Orjuela Cañón1,2, Jose Manoel de Seixas1

1Eletrical Engineering Department Universidade Federal do Rio de Janeiro, UFRJ

Rio de Janeiro, Brazil 2Electronics and Biomedical Engineering Department

Universidad Antonio Nariño Bogotá. Colombia [email protected]

Abstract— Classification of voice aging has many applications in health care and geriatrics. This work focuses on finding the most relevant parameters to classify voice aging. The most significant parameters extracted from the glottal signal are chosen to identify the voice aging process of men and women. After analyzing their statistics, the chosen parameters are used as entries to a neural network to classify male and female Brazilian speakers in three different age groups: young (from 15 to 30 years old), adult (from 31 to 60 years old), and senior (from 61 to 90 years old). The corpus used for this work was composed by one hundred and twenty Brazilian speakers (both males and females) of different ages. As compared to similar works, we employ a larger corpus and obtain a superior classification rate.

Keywords:artificial neural networks ANN, inverse filtering,

glottal signal parameters.

I. INTRODUCTION

The human aging is of great interest in various areas of research in voice, because the aging makes human body vulnerable, which directly reflects in voice production [1]. Some physiological process are continuos and irreversible, the main changes occur in the elderly. As causes of aging voice, we can cite the changes that occur in the body, for example, loss of lung capacity, changes in the epithelium of the larynx, calcification of cartilage and changes in the vibrations of the vocal folds. From 1995 to 2005, the life expectancy of Brazilian people has increased more than 8% [1]. This scenario is a global tendency, resulting in the need for developing new treatments to provide better life quality to senior people, including care for their voice.

The most common method for extracting voice features is directly from the voice signal. However, many researchers have looked for some characteristics extracted from the so called glottal signal, which is the signal obtained just after the vocal folds and before the vocal tract. Since one of the consequences of aging is the change of the vocal fold structures, this signal is, therefore, important because it preserves characteristics from

the movement of the vocal folds without the influence of the vocal tract.

Moreover, it is known that some parameters extracted from the glottal signal can help the diagnosis of pathologies of the vocal folds better than those ones extracted from the voice signal [2]. Nowadays, obtaining this signal is easier due to the development of algorithms that can perform an inverse filtering from the voice signal, eliminating the influence of the vocal tract [2] [3]. Before, it used to be necessary to use equipment coupled with micro cameras to record sounds just after air passed from the vocal folds. This was an invasive technique and very difficult to be performed.

In previous work, it has been used methods to classification of voice as Bayes, Hidden, Markov Models (HMM), Gaussian Mixture Models (GMM) and neural networks, using as input Mel-frequency cepstral Coefficients (MFCC) and the parameters jitter and shimmer, adding 37 parameters entry, using bases of 12 recordings for men and women not getting the results that rise from 80% [4]. The MFCC have proven effectiveness for speaker recognition [6], but their performance is not equal to the classification of voice aging. Comparing these methods for classification of voice aging, the best results are obtained with the techniques of neural networks and support vector machine (SVM). Differently, when the objective is the recognition of speaker, the best results are obtained with GMM and HMM methods [5] [6]. This is because for classification of voice aging are not as important temporal features of the voice and to change the voice signal with age [5]. An advantage of this work is the number of examples of voice speakers used in the classification, using only the most relevant parameters as input in a neural network.

II. FEATURES FROM THE GLOTTAL SIGNAL

A. The glottal signal

The voice signal, particularly the one related to voiced sounds, e.g. vowels, starts with the contraction-expansion of the lungs, generating a pressure difference between the air in the lungs and the air near the mouth. The airflow created passes

Page 2: [IEEE 2010 IEEE ANDESCON - Bogota, Colombia (2010.09.15-2010.09.17)] 2010 IEEE ANDESCON - Classification of voice aging using ANN and glottal signal parameters

through the vocal folds, which oscillate in a frequency called the fundamental frequency of the voice.

This oscillation modifies the airflow coming from the lungs, changing it into air pulses. The pressure signal formed by the air pulses is quasi-periodic and it is called the glottal signal. Fig. 1 presents time steps of the vocal fold vibration.

Figure 1. Formation of the glottal signal.

B. Formation glottal signal

As mentioned before, to obtain the glottal signal it used to

be necessary to perform an invasive process, by installing micro cameras to record sounds just after air passed from the vocal folds. Nowadays, it is possible to obtain the glottal signal using non invasive methods, by performing an inverse filtering on the voice signal, which consists in eliminating the influence of the vocal tract and the voice radiation caused by the mouth, preserving the glottal signal characteristics [7].

Algorithms that permit to obtain the glottal signal can be classified into two categories: semi-automatic and manual. In this paper, the inverse filtering algorithm used is of the semi-automatic category, called PSIAIF (Pitch Synchronous Iterative Adaptive Inverse Filtering) [8] [9].

It was chosen due to its high performance and ease development. There is a toolbox implementation in Matlab, called Aparat [10], which was constructed especially based on PSIAIF method to obtain the glottal signal as well as the extraction of its main features or parameters.

The parameters which can be extracted from the glottal signal can be divided into three groups: time domain, frequency domain and the ones that represent the variations of the fundamental frequency [8].

1) The time domain parameters

The time domain parameters that can be extracted from the

glottal signal are described below [8], [9] .

Closing phase (Ko): it describes the interval between the instant of the maximum opening of the vocal folds and the instant where they close [8] (see Fig. 2).

Opening phase (Ka): it describes the interval between the instant where the vocal folds start the oscillation up to their maximum opening [8] (see Fig. 2).

Opening quotient (OQ): The ratio between the total time of the vocal folds opening and the total time of a cycle (or period) of the glottal signal (T). It is inversely proportional to the intensity of the voice. The smaller it is, the higher the voice intensity [1][8]. It is calculated by Eq. (1) (see Fig. 3):

(1)

Figure 2. Parameters Ko e Ka obtained from the glottal signal.

Closing quotient (CIQ): The ratio between the closing phase parameter (Ko) and the total length of a glottal pulse (T) [8]. It is inversely proportional to voice intensity. The smaller it is, the higher the voice intensity. It is given by Eq. (2) (see Fig. 3):

(2)

Amplitude quotient (AQ): The ratio between the glottal signal amplitude (Av) and the minimum value of the glottal signal derivative (dmin) [11]. It is related to the speaker phonation [9] and it is given by Eq. (3) (see Fig. 3):

(3)

Normalized amplitude quotient (NAQ): It is calculated by the ratio between the amplitude quotient (AQ) and the total time length of the glottal pulse (T) [12]. It is given by Eq. (4) (see Fig. 3):

(4)

Opening quotient defined by the Liljencrants-Fant model (OQa): This is another opening quotient but calculated by the Liljencrants-Fant model for inverse filtering. Details about this model can be found in [13].

Quasi opening quotient (QoQ): It is a relationship between the glottal signal opening at the exact instant of the oscillation and the closing time [9]. It is used in some research to classify emotions [14].

Page 3: [IEEE 2010 IEEE ANDESCON - Bogota, Colombia (2010.09.15-2010.09.17)] 2010 IEEE ANDESCON - Classification of voice aging using ANN and glottal signal parameters

Speed quotient (SQ) is defined as the ratio of the opening phase length to the closing phase length [8].

(5)

Figure 3. Parameters of the glottal signal in time domain.

2) The frecuency domain parameters

Difference between harmonics (DH12): Also known as H1-

H2 and it is the difference between the values of the first and second harmonics of the glottal signal [15][16] (see Fig. 4 (a)). This parameter has been used to measure vocal quality.

Harmonics relation factor (HRF): It relates the first harmonic (Hi) with the sum of the energy of the other harmonics (Hk) (see Fig. 4 (b)) [17]. It has also been used to measure the vocal quality and it is given by Eq. (6).

(6)

3) Parameters that represent variations and perturbations

in the fundamental frequency

Jitter: variations in fundamental frequency between successive vibratory cycles, as shown in Fig. 5 [18] [19]. Changes in jitter may be indicative of neurological or psychological difficulties [1].

Shimmer: variations in amplitude of the glottal flow between successive vibratory cycles, as also illustrated in Fig. 5 (b) [18] [19]. Changing the shimmer is found mainly in the presence of mass lesions in the vocal folds, such as polyps, edema or carcinomas [1].

Harmonics-to-Noise-Ratio (HNR): is an acoustic measure that determines how much of harmonics (periodic) and of noise (aperiodic) exists in a voice. This measurement has proved to be a sensitive and strong indicator of vocal productions in adults [1].

Figure 4. Parameters in the frequency domain signal glottal,

Figure 5. Jitter and Shimmer

All the parameters described above can be obtained from the Aparat Software [10].

III. METHODOLOGY

A. Database

The database used for this work is composed of 60 male voice registers and 60 female voice registers for the sustained vowel /e/ for the Portuguese spoken in the city of Rio de Janeiro city, Brazil. This database was divided in three different age groups: from 15 to 30 years old, from 31 to 60 years old and from 61 to 90 years old. There were 20 voice registers for each age group.

B. Inverse Filtering

For each vocal register the corresponding glottal signal was

obtained by inverse filtering PSIAIF and the parameters were extracted using the Aparat [10] and Praat [20] software. The following parameters were obtained: fundamental frequency (fo), jitter, shimmer, HNR, Ko, Ka, NAQ, AQ, CIQ, OQ1, OQ2,

Page 4: [IEEE 2010 IEEE ANDESCON - Bogota, Colombia (2010.09.15-2010.09.17)] 2010 IEEE ANDESCON - Classification of voice aging using ANN and glottal signal parameters

Oqa, Qoq, SQ1, SQ2, DH12 and HRF. The parameters were separated according to the groups to which they belonged. OQ was divided into OQ1, OQ2, the open quotients calculated from the so-called primary and secondary openings of the glottal flow; SQ was divided into speed quotients calculated from the primary and the secondary openings of glottal signal.

C. Choice of parameters

With the parameters of the glottal signal obtained from the database, it was calculated the median, the mean value and the mean dispersion value to increase their behavior in each age group. The parameters that showed the greatest differences between the individual age groups are for women Shimmer, AQ, OQ1 and OQ2 the results are shown in Table I. For men, the parameters are Ko OQ1 Oqaand SQ2 results are shown in Table II.

TABLE I. PARAMETERS FOR FEMALES GROUP

Parameters Measures Groups

15-30 31-60 61-90

Median 5.6 7.8 8.9 Shimmer

Mean value 7.370 10.53 10.40 Dispersion 36.76 83.13 37.44

Median 0.32 0.56 0.65 Mean value 0.6 0.7 0.8

AQ

Dispersion 0.059 0 0.018 Median 0.426 0.626 0.822

Mean value 0.587 0.614 0.799 OQ1

Dispersion 0.082 0.062 0.0339 Median 0.28 0.448 0.598 OQ2

Mean value 0.409 0.477 0.617 Dispersion 0.039 0.051 0.0258

Shimmer and AQ has mean value and median very similar to the range of 31-60 and 61-90, but the variance is lower in the range of 61-90, that explains why women after menopause 45-55 years old or so has a sudden change in hormones produce changes in voice.

TABLE II. PARAMETERS FOR MALE GROUP

Parameters Mesures Gropus

15-30 31-60 61-90

Median 0.61 0.81 1.7 Ko

Mean value 0,0011 0,001 0,002 Dispersion 7,2*10-6 3*10-6 6,6*10-6

Median 0.4 0.41 0.83 Mean value 0,5 0,4 0,8

OQ1

Dispersion 0,135 0,015 0,798 Median 0.165 0.18 0.35

Mean value 0,161 0,1705 0,3045 Oqa

Dispersion 0,0025 0,0023 0,0144 Median 1.95 1.7 1.3 SQ2

Mean value 1,92 1,73 1,32 Dispersion 0,4198 0,4190 0,3338

The OQ1 and OQ2 are good discriminators already that its value is greater for the age group 61-90 it colud be that the older the time between the opening chords of vowels is the biggest cause of lost elasticity of the strings vowels.

The mean value for Ko is high in the 61-90 group when compared to other groups. It may occur due to the vocal folds relax, during the aging process according to the aging, because the closing of the vocal folds takes more time. The mean value and median for Ko and OQ1 for male group is high in the 61-90 group when compared to other groups. It may occur due to the vocal folds relax, during the aging process according to the aging, because the closing of the vocal folds takes more time.

The speed ratio is lower for the group 61-90 years old, this is why the speed of opening of the glottis decreases with age.

In the fig 6 and fig 7 shows the box plots of the parameters for female and male groups. In these graphs shows that the median parameters for each age group and the behavior of the data.

Figure 6. Box plot parameters for female group

IV. RESULTS

With the parameters of the glottal signal, the classification

is developed with a neural network (MLP), obtaining the speaker kind from the three defined groups. These groups are the outputs of the neural network in which 65% from the database were used for training, 15% as validation set (to avoid the network overtraining and to choose the number of processors and the hidden layers) and 20% for testing.

For the female database, the best result was achieved using 3 neurons with only one hidden layer classifying correctly 10 speakers and incorrectly 2 speakers; that is, 91.66% for guess rate, as shown in Table III.

Page 5: [IEEE 2010 IEEE ANDESCON - Bogota, Colombia (2010.09.15-2010.09.17)] 2010 IEEE ANDESCON - Classification of voice aging using ANN and glottal signal parameters

Figure 7. Box plot parameters for male group

For the male database, the best result was achieved using 3 neurons with only one hidden layer classifying correctly 11 speakers and incorrectly 1 speaker; that is, 91,66% for guess rate, as shown in Table IV.

TABLE III. CONFUSION MATRIX OF CLASSIFICATION FOR FEMALE GROUP.

Age 15-30 Age 31-60 Age 61-90

Age 15-30 4 0 0 Age 31-60 0 1 1 Age 61-90 0 0 5

TABLE IV. CONFUSION MATRIX OF CLASSIFICATION FOR MALE GROUP.

Age 15-30 Age 31-60 Age 61-90

Age 15-30 4 1 0 Age 31-60 0 3 0 Age 61-90 0 0 4

V. CONCLUSIONS

The results obtained herein were satisfactory and better than those achieved in previous similar work. The main contribution of this paper is the use of relevant parameters of the glottal signal in classifying aging voices. Both automatic classifiers used in this work were efficient in combining the parameters to provide an improved overall performance.

VI. ACKNOWLEDGMENTS

The authors thank CAPES, CNPq, and FAPERJ by the partial funding of this work.

REFERENCES

[1] Rosa I. S. Analise acústica da voz de indivíduos na terceira idade. Tese de mestrado Universidade de São Carlos (2005).

[2] Gómez P., Fernández R., Glottal Source biometrical signature for voice pathology detection Speech Communication 51 759–781 (2009).

[3] Verdonck I., Mahieu H. Vocal aging and the impact on daily life: a longitudinal study. doi:10.1016 jornal voice (2003).

[4] Sadeghi N. A., Homayounpour M.M.: ‘Speaker age interval and sex identification based on jitters, shimmers and mean MFCC using supervised and unsupervised discriminative classification methods’. Proc. ICSP, Guilin, China, (2006)

[5] Sedaaghi M. H. A Comparative Study of Gender and Age Classification in

[6] Speech Signals Iranian Journal of Electrical & Electronic Engineering, Vol. 5, No. 1, March 2009

[7] Campbell, Jr. J. P., “Speaker Recognition: A tutorial”, Proceedings of the IEEE, vol. 85, no. 9, pp. 1437-1462, 1997.

[8] Alku P., Glottal wave analysis with Pitch Synchronous Adaptive Inverse Filtering", Speech Communication, vol. 11, pp. 109-118 (1992).

[9] Pulakka H. Analysis of Human Voice Production Using Inverse Filtering, High-Speed Imaging, and Electroglottography. University of Technology Helsinki ( 2005).

[10] Juliano S. M., Um estudo comparativo entre o sinal electroglotográfico e o sinal de voz”, Dissertação de mestrado em Engenharia de Telecomunicações, UFF (2008)

[11] Software Aparat http://aparat.sourceforge.net/index.php/Main_Page Helsinki University of Technology Laboratory of Acoustics and Audio Signal Processing.

[12] Alku P. and Vilkman. E. Amplitude Domain Quotient for Characterization of the Glottal Volume Velocity Waveform Estimated by Inverse Filtering.Speech Communication, 18(2):131-138, (1996)

[13] Alku P., Bäckström T., and Vilkman E. Normalized Amplitude Quotient for Parameterization of the Glottal Flow. Journal of the Acoustical Society of America, 112(2):701-710, August (2002)

[14] Gobl, C., Chasaide. A., Amplitude-based source parameters for measuring voice quality , In VOQUAL 151-156 (2003)

[15] A.-M. Laukkanen, E. Vilkman, P. Alku, Related to Stress and Emotional State: a Preliminary Study. Journal of Phonetics, 24(3):313-335, (1996).

[16] I. Titze and J. Sundberg. Vocal intensity in speakers and singers. Journal of the Acoustical Society of America, 2936-2946, May (1992)

[17] Airas M., Methods and studies of laringeal voice quality analysis in speech production. Dissertation for the degree of Doctor Helsinki University of Technology (2008)

[18] Childers D. G. and Lee. C. K Vocal quality factors: Analysis, synthesis, and perception. Journal of the Acoustical Society of America, 2394-2410 May (1990)

[19] Cataldo, E. ; Rodrigues, Fabiana; BRANDÃO, A. ; LUCERO, Jorge . Usando Redes Neurais para Classificação de padrões de voz. In: XXVIII CNMAC - Congresso Nacional de Matemática Aplicada e Computacional, 2005, São Paulo. Anais do XXVIII CNMAC, (2005)

[20] Vieira, M. N. Automated Measures of Dysphonias and the Phonatory Effects of Asymmetries in the Posterior Larynx, Ph.D Thesis, University of Edinburgh, UK (1997).

[21] Software Praat http://www.fon.hum.uva.nl/praat/ University of Amsterdam.