intelligibility of speech - troels...

43
Technical University of Denmark Intelligibility of Speech 2 nd semester project DTU Electrical Engineering Acoustic Technology Spring semester 2008 Group 5 Troels Schmidt Lindgreen – s073081 David Pelegrin Garcia – s070610 Eleftheria Georganti – s073203 Instructor Thomas Ulrich Christiansen

Upload: duongkhue

Post on 28-Aug-2018

235 views

Category:

Documents


0 download

TRANSCRIPT

Technical University of Denmark

Intelligibility of Speech

2nd semester project

DTU Electrical Engineering

Acoustic Technology

Spring semester 2008

Group 5

Troels Schmidt Lindgreen – s073081David Pelegrin Garcia – s070610

Eleftheria Georganti – s073203

Instructor

Thomas Ulrich Christiansen

DTU Electrical Engineering

Ørsteds Plads, Building 348

2800 Lyngby

Denmark

Telephone +45 4525 3800

http://www.elektro.dtu.dk

Technical University of Denmark

Title:

Intelligibility of Speech

Course:

31236 Auditory Signal Processingand PerceptionSpring semester 2008

Project group:

5

Participants:

Troels Schmidt LindgreenDavid Pelegrin GarciaEleftheria Georganti

Supervisor:

Torsten Dau

Instructor:

Thomas Ulrich Christiansen

Date of exercise: April 17th

Pages: 27

Copies: 4

Synopsis:

This report deals with the intelligibility

of speech and objective methods that can

be used to predict it. Three different

methods are used to estimate speech

intelligibility. Namely, Articulation Index

(AI), Speech Intelligibility Index (SII) and

Speech Transmission Index (STI).

It has been concluded that it is pos-

sible to determine the gender of the

speaker, but not the spoken language from

a long-term magnitude spectrum. The

main peak in the spectrum corresponds

to the fundamental frequency, which is

around 125 Hz for male speakers and

250 Hz for females. AI and SII are useful

estimators of speech intelligibility for

band-limited systems whereas STI can

only be calculated if the impulse response

is available and it is used mainly in room

acoustics measurements.

No part of this report may be published in any form without the consent of the writers.

Introduction

This exercise investigates the intelligibility of speech and three different methods to es-timate it are examined. Namely, the Articulation Index (AI), the Speech IntelligibilityIndex (SII) and the Speech Transmission Index (STI).

The magnitude spectrum of speech and speech shaped noise of four existing speech sam-ples is examined. It is investigated whether regional and gender characteristics can beobtained through these spectra.

A Matlab function is composed in order to calculate the Articulation Index (AI) andthe Speech Intelligibility Index (SII). Low-pass and high-pass transmission channels withdifferent cut-off frequencies are used. The predictions are plotted as a function of thecut-off frequency. Further investigation of the influence of the speech level on AI and SIIcalculation is carried out.

In the last part, the presence of non-linear distortions in the transmission channel isinvestigated and the concept of predicting the speech intelligibility with the Speech Trans-mission Index (STI) is used. Simulated impulse responses are used and the correspondingSTI is calculated.

Technical University of Denmark, April 25, 2008

Troels Schmidt Lindgreen David Pelegrin Garcia Eleftheria Georganti

Contents

1 Theory 1

1.1 Definition of speech intelligibility . . . . . . . . . . . . . . . . . . . . . . . . 21.2 Speech reception threshold . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21.3 Speech shaped noise . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31.4 Articulation index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41.5 Speech intelligibility index . . . . . . . . . . . . . . . . . . . . . . . . . . . . 51.6 Speech transmission index . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5

2 Results 9

2.1 Spectrum of speech . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 92.2 Speech shaped noise . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 102.3 High-pass and low-pass filtered speech . . . . . . . . . . . . . . . . . . . . . 112.4 Speech intelligibility in simulated rooms . . . . . . . . . . . . . . . . . . . . 15

3 Discussion 17

3.1 Spectrum of speech . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 173.2 Speech shaped noise . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 183.3 High-pass and low-pass filtered speech . . . . . . . . . . . . . . . . . . . . . 213.4 Comparison of AI, SII and STI . . . . . . . . . . . . . . . . . . . . . . . . . 21

4 Conclusion 25

Bibliography 27

A Matlab code A1

B Long-term average spectrum of speech A5

April 2008 Chapter 1. Theory

Chapter 1Theory

To get a better understanding in speech generation and perception, methods have beendeveloped in order to determine the speech intelligibility threshold. Speech intelligibilityindicates how many correct words the listener understands from speech in a given situa-tion. Speech is generated from a talker and is received by a listener, as illustrated in figure1.1.

Figure 1.1: Speech from one person to another. Inspired from [Poulsen, 2007, p. 3]

Page 1

Chapter 1. Theory Technical University of Denmark

1.1 Definition of speech intelligibility

Speech intelligibility is a number that indicates how much speech is understood correctlyin a certain situation. Speech intelligibility can be formulated as:

Speech intelligibility =100T· R (1.1)

Where:T is the number of speech units in the test.R is the number of correct speech units.

Speech intelligibility is defined as the percentage of speech units that are understoodcorrectly from a speech intelligibility test [Poulsen, 2005, p. 62]. These units can bewords (and equation (1.1) is referred to as word score), syllables, or other units. Speechintelligibility is typically drawn as a function of signal-to-noise ratio, as shown in figure1.2.

1.2 Speech reception threshold

The Speech Reception Threshold (SRT) is defined as the signal-to-noise ratio at which thepsychometric function that describes the speech intelligibility is equal to 0.5. Thus, theSRT depends on the test subject, the material and the score units used to determine theintelligibility.

40%50%60%70%80%90%

100%

ect a

nsw

ers

[%]

0%10%20%30%

-25 -20 -15 -10 -5 0

Corr

e

Speech relative to noise [dB]

Figure 1.2: Word score from Dantale as a function of signal-to-noise ratio. Redrawn from [Poulsen, 2005, p.67].

A typical speech intelligibility test consists of a number of words being read out aloud bya speaker. The listener writes what he/she hears, and the lists are compared afterwards

Page 2 Intelligibility of Speech

April 2008 Chapter 1. Theory

to find the speech intelligibility, see figure 1.3.

Text-in

Speaker Transmisson system Listener

Text-out

Comparing texts

Figure 1.3: Elements involved in speech intelligibility test. Inspired by [Poulsen, 2005].

There are a lot of factors to consider in speech communication systems. The transmissionsystem is an important part of speech intelligibility and it can be anything from the roomwhich is spoken in, a telephone line or the ears of the listener. It depends on the situation,where the transmission system can be influenced by reverberation, distortion or noise etc.

Figure 1.3 shows that the transmission system is only a small part of the system. Otherimportant factors are:Word material Could be sentences, words or numbers.Presentation method Single words presented or a small text/sentence.Open/closed response test Whether or not the words are known to the listener be-

forehand.Speaker Speed of talking, pronunciation and dialect. Normal

speech, raised voice or shouting.Listener Hearing ability, training, vocabulary, dialect, mother

tongue.Scoring method Scoring by exact word or phoneme. Oral or written an-

swers.

1.3 Speech shaped noise

Speech Shaped Noise (SSN) is defined as a random noise that has the same long-termspectrum as a given speech signal. There are many techniques to produce this kind ofnoise, e.g.:

• Gaussian noise can be equalized to have the same magnitude spectrum as a givenspeech signal.

Page 3

Chapter 1. Theory Technical University of Denmark

• The phase of a speech signal can be randomized. In this way, the signal will berandom noise but it will keep the magnitude spectrum of the original speech.

• Multiple replicas of the signal (reversed or not) can be delayed and added together.

However, these techniques do not preserve the intrinsic fluctuations of speech and thus themodulations are reduced. Nevertheless, it is possible to produce speech shaped noise thatkeeps the envelope spectrum of the speech. According to [Dau and Christiansen, 2008, p.4], one example of this kind of shaped noise is found in the so-called ICRA-noises, whichare created by the following stages, illustrated in figure 1.4:

1. Filtering of the speech signal in three adjacent bands covering the whole frequencyrange of the signal.

2. Randomly reversing the sign of the samples in each band. This process modifies thespectrum but keeps the envelope.

3. Filtering of each band with the same filters as in stage 1.

4. Add and normalise the contributions from the three bands.

Figure 1.4: Block diagram of the ICRA-noise generation for a given speech signal s(t). r1(t), r2(t) and r3(t)are uncorrelated random signals that take values −1 or 1 and have the same length as the original signal.

1.4 Articulation index

The Articulation Index (AI) was developed in order to evaluate how noise over a telephoneline affects speech intelligibility. AI estimates the physical signal-to-noise ratio as a cor-related internal representation in the listener by calculating an ”effective signal-to-noiseratio” for 20 frequency bands. Masking effects are taken into account in the calculationsuch that the noise may mask adjacent frequency bands. The effective contributions ofeach frequency band are weighted using a band importance function to yield an index valuefrom 0 to 1, where 1 marks maximum intelligibility and 0 marks no intelligibility. [Dau

Page 4 Intelligibility of Speech

April 2008 Chapter 1. Theory

and Christiansen, 2008, p. 4]

It must be notes that AI quantifies the degradations of the signal introduced by thetransmission channel only. In order to predict speech intelligibility, knowledge about thespecific task is required. It is then possible to read off the speech intelligibility from AIgraphs which are based on speech intelligibility measurements representing average pop-ulation responses of normal hearing talkers and listeners in stationary background noise.Such a graph can be seen on figure 1.5

Figure 1.5: Relation between speech intelligibility (in %) and AI. From [Poulsen, 2005, p. 76]

.

AI is standardized in ANSI S3.5-1969.

1.5 Speech intelligibility index

The speech intelligibility index (SII) is based on the AI principle. SII uses other weightingfunction and a number of modifications are implemented. One of the modifications is thecorrection for change in speech spectrum according to vocal effort.

SII in standardized in ANSI S3.5-1997.

1.6 Speech transmission index

An approach to quantify the speech intelligibility from the objective sound field datais based on the Modulation Transfer Function (MTF). MTF is a function m(Ω) (where

Page 5

Chapter 1. Theory Technical University of Denmark

Ω = i · ω) which can be calculated from the impulse response h of the transmission systemand equals to:

m(Ω) =

∫ ∞0 h2(t)e−iωtdt∫ ∞

0 h2(t)dt(1.2)

In other words equation 1.2 is the complex Fourier Transform of the squared impulse re-sponse divided by its total energy. The MTF quantifies the leveling effect of reverberationon the envelope of speech signals. [Houtgast and Steeneken, 1973] have developed a pro-cedure to convert MTF data measured in seven octave bands and at several modulationfrequencies into one single quantity, which they called “Speech Transmission Index (STI)”.This conversion involves averaging over a certain range of modulation frequencies andtakes into account the contribution of the various frequency bands to speech quality andthe masking effect between adjacent frequency bands that occurs in the hearing system.

The exact calculations needed to determine the STI requires firstly, the calculation ofthe complex MTF for the octave bands centered at 125 Hz, 250 Hz, 500 Hz, 1 kHz, 2 kHz,4 kHz, and 8 kHz, according to equation (1.2). Then a reduction of modulation mF,k forthe different 14 modulation frequencies (0.63 Hz, 0.8 Hz, 1 Hz, 1.25 Hz, 1.6 Hz, 2 Hz, 2.5Hz, 3.15 Hz, 4 Hz, 5 Hz, 6.3 Hz, 8 Hz, 10 Hz, and 12.5 Hz) is required, see figure 1.6. Thisequals to a transformation to a signal-to-noise ratio:

SNRF,k = 10 log10

(mF,k

1 − mF,k

)(1.3)

where S NRF,k is the signal-to-noise ratio in a given combination of an octave band F anda modulation frequency band k and mF,k is the modulation frequency of the correspondingcombination of bands.

The 14 values for each octave band are then averaged across modulation frequency yielding7 values that correspond to an octave band. These values are then multiplied by a factoraF :

SNRF = aF

14∑k=1

SNRF,k

14(1.4)

Afterwards, the 7 octave band values S NRF are limited to ±15 dB, averaged and thennormalized. This leads to the expression of the STI:

STI =SNRav + 15

30(1.5)

where SNRav corresponds the average value of the SNR over the 7 octave bands. STIscores are rated as seen in table 1.1.

Page 6 Intelligibility of Speech

April 2008 Chapter 1. Theory

Acoustic Communication T. Poulsen

80

modulation depth will be reduced at the receiving point if noise or reverberation has been present in the course from sender to receiver. The reduction of the modulation is described by the MTF, m(F), which is a function of the modulation frequency, F. MTF is the basis for the calculation of STI. A noise signal with the same long term spectrum as speech (speech shaped noise) is used in the measurement of STI. The noise signal is divided into 7 octave bands from 125 Hz to 8 kHz, see figure 11.2. Each octave band is separately modulated with 14 modulation frequencies (one at a time) from 0.63 Hz to 12.5 Hz. The modulation is sinusoidal and 100% and the signal is emitted from the sender (the speaker).

Figure 11-2. Overview of frequency bands and modulation frequencies in the STI calculation.

For each of the 98 combinations of octave band and modulation frequency the reduction in modulation is determined. This is the MTF as a function of octave band centre frequency, k, and modulation frequency, F. The reduction in modulation, m(F,k), is transformed to a signal-to-noise ratio by means of SNRF,k = 10 log (m / (1-m)) dB As in the AI method the SNR is truncated to a dynamic range of 30 dB but here the limits are ±15 dB (in AI the limits are +12 dB and –18 dB). If SNR is >15 dB then SNR is set to 15 dB and if SNR is < –15 dB then SNR is set to –15 dB.

Figure 1.6: Overview of frequency bands and modulation frequencies in the STI calculation. The graymarks are the 9 calculations used in RASTI (RApid STI method). From [Steeneken and Houtgast, 1985, p.14].

STI value < 0.30 0.30 – 0.45 0.45 – 0.60 0.60 – 0.75 > 0.75

STI rating Bad Poor Fair Good Excellent

Table 1.1: STI ratings. From [Houtgast and Steeneken, 1985].

Like in AI, the relation between STI and speech intelligibility score is not linear. It ispossible to read off the speech intelligibility from STI graphs like figure 1.7.

STI is standardized in IEC 286-16-1988, and has been revised in 2003.

Page 7

Chapter 1. Theory Technical University of Denmark

Acoustic Communication T. Poulsen

82

and in the STI procedure. It is seen that especially around 2 kHz the weights are different. The STI in itself is not enough to predict the speech intelligibility. In order to use the STI it is necessary to know the relation between speech intelligibility and STI. This is illustrated in figure 11.4 for some typical word materials.

Figure 11-4. Relation between intelligibility and STI for different word materials. Compare with Figure 10-6.

11.2 RASTI As seen in the STI section it is a somewhat complicated matter to calculate STI. This is not satisfying for an objective method that should replace the cumbersome subjective measuring methods. Therefore a simpler method, RASTI (= RApid STI) has been developed based on the same principles as in STI. In RASTI the number of combinations of octave bands and modulation frequencies are reduced from 98 to 9. In Figure 11-2 the combinations of frequency bands and modulation frequencies used in RASTI are shown in grey. Only the octave bands 500 Hz and 2 kHz are used and the modulation frequencies are selected so that they cover the most important range. Figure 11.5 show the envelopes curves for the 500 Hz and the 2 kHz noise bands. Figure 11.6 shows the calculation procedure. There are no weighting factors in the RASTI calculation but because of the 5 modulation frequencies used in the 2 kHz band and only 4 modulation frequencies in the 500 Hz band, there will be a slightly higher weight for the 2 kHz band in the average calculation.

Figure 1.7: Relation between intelligibility and STI for different word materials. From [Houtgast andSteeneken, 1985, p. 11].

Page 8 Intelligibility of Speech

April 2008 Chapter 2. Results

Chapter 2Results

2.1 Spectrum of speech

Four speech recordings are listened and their full bandwidth long-term magnitude spec-trum is calculated with the script provided in appendix A.

The long-term magnitude spectrum from 100 Hz to 10 kHz of the 4 speech signals and thecorresponding smoothed spectrum (using the supplied script Oct3Smooth.m) can be seenin figures 2.1, 2.2, 2.3 and 2.4.

Frequency [Hz]

Mag

nit

ude

[dB

]

Long term spectrum

1/3-octave smoothed spectrum

100 250 500 1000 2000 4000 8000-40

-20

0

20

40

60

Figure 2.1: Long-term magnitude spectrum from100 Hz to 10 kHz and the corresponding smoothedspectrum for a danish female speaker.

Frequency [Hz]

Mag

nit

ude

[dB

]

Long term spectrum

1/3-octave smoothed spectrum

100 250 500 1000 2000 4000 8000-40

-20

0

20

40

60

Figure 2.2: Long-term magnitude spectrum from100 Hz to 10 kHz and the corresponding smoothedspectrum for a danish male speaker.

Page 9

Chapter 2. Results Technical University of Denmark

Frequency [Hz]

Mag

nit

ude

[dB

]

Long term spectrum

1/3-octave smoothed spectrum

100 250 500 1000 2000 4000 8000-40

-20

0

20

40

60

Figure 2.3: Long-term magnitude spectrum from100 Hz to 10 kHz and the corresponding smoothedspectrum for an english female speaker.

Frequency [Hz]

Mag

nit

ude

[dB

]

Long term spectrum

1/3-octave smoothed spectrum

100 250 500 1000 2000 4000 8000-40

-20

0

20

40

60

Figure 2.4: Long-term magnitude spectrum from100 Hz to 10 kHz and the corresponding smoothedspectrum for an english male speaker.

The sampling frequency ( fs), the number of samples (N), the bandwidth (B) and thefrequency resolution (Fres) of the above speech signals are related when performing anFFT. More precisely the frequency resolution (Fres) and the bandwidth of the transformedsignal (B) equal to:

Fres =1N

(2.1)

B =fs

2(2.2)

2.2 Speech shaped noise

The script randomizePhase is used to generate speech shaped noise from the 4 speechsamples. This script assigns a random value to the phase of each frequency componentof the signal. As explained in section 1.3, this kind of procedure does not preserve themodulation of the signal. As an example, in figure 2.5 a speech signal and the speechshaped noise calculated with this method are shown, both in the time domain and themagnitude spectrum.

Page 10 Intelligibility of Speech

April 2008 Chapter 2. Results

Time [s]

0 10 20 30 40 50-0.4

-0.2

0

0.2

0.4

(a)

Time [s]

0 10 20 30 40 50-0.4

-0.2

0

0.2

0.4

(b)

Frequency [Hz]

Magnitude

[dB

]

102 103 104

20

30

40

50

60

(c)

Frequency [Hz]

Magnitude

[dB

]

102 103 104

20

30

40

50

60

(d)

Figure 2.5: Speech signal (a) and its magnitude spectrum (c). The phase of the signal has been randomizedand results in a signal (b) with magnitude spectrum (d).

2.3 High-pass and low-pass filtered speech

In order to determine the importance of each frequency region for the intelligibility ofspeech, low-pass and high-pass transmission channels with different cut-off frequenciescan be used.

High-pass and low-pass filters with 20 different cut-off frequencies in the frequency rangebetween 100 Hz and 10 kHz were calculated using Matlab (see appendix A). The transferfunctions of the high-pass filters with different cut-off frequencies have been plotted andcan be seen in figure 2.6. Additionally, the transfer functions of the low-pass filters withdifferent cut-off frequencies have been plotted and can be seen in figure 2.7.

For the low-pass and high-pass filters the AI and SII are calculated using the Matlab codethat can be found in appendix A. The speech spectrum level is assumed to be “normal”,“raised”, “loud”or“shout”, corresponding to the SPL values shown in table 2.1. The resultsfor the high-pass and low-pass filters can be seen in figures 2.8 and 2.9 respectively.

Page 11

Chapter 2. Results Technical University of Denmark

Mag

nit

ude

(lin

ear)

Frequency [Hz]

50 100 200 500 1000 2000 5000 100000

0.2

0.4

0.6

0.8

1

Figure 2.6: Transfer functions of high-pass filters with different cut-off frequencies in the frequency rangebetween 100 Hz and 10 kHz.

Voice level dB SPL

normal 63raised 68loud 75shout 82

Table 2.1: Overall speech sound pressure levels for a female speaker [Poulsen, 2005].

Page 12 Intelligibility of Speech

April 2008 Chapter 2. Results

Mag

nit

ude

(lin

ear)

Frequency [Hz]

50 100 200 500 1000 2000 5000 100000

0.2

0.4

0.6

0.8

1

Figure 2.7: Transfer functions of low-pass filters with different cut-off frequencies in the frequency rangebetween 100 Hz and 10 kHz.

AI

or

SII

Filter cut-off frequency [Hz]

AI

SII

125 250 500 1000 2000 4000 80000

0.5

1

AI

or

SII

Filter cut-off frequency [Hz]

AI

SII

125 250 500 1000 2000 4000 80000

0.5

1

AI

or

SII

Filter cut-off frequency [Hz]

AI

SII

125 250 500 1000 2000 4000 80000

0.5

1

AI

or

SII

Filter cut-off frequency [Hz]

AI

SII

125 250 500 1000 2000 4000 80000

0.5

1

Figure 2.8: AI and SII for high-pass filters as a function of the cutoff frequency when the speech spectrumlevel is assumed to be (a) “normal”, (b) “raised”, (c) “loud” and (d) “shout”.

Page 13

Chapter 2. Results Technical University of DenmarkA

Ior

SII

Filter cut-off frequency [Hz]

AI

SII

125 250 500 1000 2000 4000 80000

0.5

1A

Ior

SII

Filter cut-off frequency [Hz]

AI

SII

125 250 500 1000 2000 4000 80000

0.5

1

AI

or

SII

Filter cut-off frequency [Hz]

AI

SII

125 250 500 1000 2000 4000 80000

0.5

1

AI

or

SII

Filter cut-off frequency [Hz]

AI

SII

125 250 500 1000 2000 4000 80000

0.5

1

Figure 2.9: AI and SII for low-pass filters as a function of the cutoff frequency when the speech spectrumlevel is assumed to be (a) “normal”, (b) “raised”, (c) “loud” and (d) “shout”.

Page 14 Intelligibility of Speech

April 2008 Chapter 2. Results

2.4 Speech intelligibility in simulated rooms

The STI for four different impulse responses from different rooms was calculated withMatlab and the results can be seen in table 2.2. The code can be found in the appendixA.

Simulated impulse response STI

auditorium11J02.wav 0.62BostonSHJ01.wav 0.44GrundvigsJ03.wav 0.33Listen04J01.wav 0.78

Table 2.2: STI values for different rooms.

Page 15

Chapter 2. Results Technical University of Denmark

Page 16 Intelligibility of Speech

April 2008 Chapter 3. Discussion

Chapter 3Discussion

3.1 Spectrum of speech

It is possible to determine the sex of from a long-term magnitude spectrum by looking atthe low frequencies of the long-term magnitude spectrum of the speech signals that appearin figures 2.1, 2.2, 2.3 and 2.4. Male have lower fundamental frequencies than females.Fundamental frequencies for males are around 125 Hz whereas females have fundamentalfrequencies around 250 Hz. [Poulsen, 2005, p. 47]. Figures 2.1 and 2.3 imply female speak-ers as a peak appears close to 200 Hz, that corresponds to their fundamental frequency.On the other hand, figures 2.2 and 2.4 imply male speakers as the fundamental frequencyappears to be around 100 Hz.

Is it very difficult though to determine the language used from a given long-term mag-nitude spectrum. This is because languages have similar long-term magnitude spectraeven though the languages are fundamentally different. This can be seen on the figuresappearing in appendix B, which show the long-term magnitude spectra for speakers ofboth genders and different languages.

There is a lot of research going on in the topic of speaker identification. Several tech-niques and strategies are used and most of them are based on the scheme shown in figure3.1.

Most of the systems use the dynamic properties of the speech signal rather than ana-lyzing the long-term spectrum. The latter requires a long signal acquisition time to betext-independent. According to figure 3.1, the different blocks can be implemented withdifferent strategies. The features extracted from the speech signal can be the cepstrumand different cepstral coefficients or the coefficients from a linear predictor (LP) based

Page 17

Chapter 3. Discussion Technical University of Denmark

Figure 3.1: Basic speaker identification system.

on different filter models. The length of the different speech segments is also dependingon the strategy. The output of the block after the A/D conversion is a set of vectors(describing different features of speech) that should be compared to the ones stored in adatabase, corresponding to different individuals, using the concept of distance measure-ment. Different strategies are used to determine this distance, such as pattern matchingbased on vector Euclidean distances or the use of Hidden Markov Models (HMM). Finally,a decision device has to determine the individual that provides the maximum likelihoodfor a given input signal [Campbell, 1997].

3.2 Speech shaped noise

A speech shaped noise generated from a speech signal is shown in figure 2.5. The timefluctuations in the original signal (a), i.e. the envelope, disappear when the phase is ran-domized (b), while the magnitude spectrum remains unaffected by the process, (c) and (d).

In order to incorporate the fluctuations of the speech signal into the speech shaped noise,two methods are suggested, apart from the ICRA-noises generation described in section1.3. The first method, illustrated in figure 3.2, uses two inputs: random noise nI(t) anda speech signal s(t). The spectrum of the noise is equalized with the one of the speechsignal. Using the Hilbert transform, the envelope of the speech is extracted and appliedto the equalized noise. The last step is the normalization of the energy, extracted fromthe spectra of the signal and the modulated speech shaped noise.

Another method is designed to obtain speech shaped noise with modulation features usingonly a speech signal as shown in figure 3.3. Thus, the noise is obtained by making thephase of the speech random, while the magnitude spectrum is preserved. Then the enve-lope is applied and the energy is normalized in the same way as in the previous method.

The method shown in figure 3.2 has been implemented in Matlab (see Analysis_syn-

Page 18 Intelligibility of Speech

April 2008 Chapter 3. Discussion

Figure 3.2: Block diagram of the used method to generate speech shaped noise with modulation features,based on the adaption of random noise to a given speech signal.

Figure 3.3: Block diagram of the used method to generate speech shaped noise with modulation features,based on the phase randomization of a given speech signal.

thesis.m function in appendix A) and it has been used to produce speech shaped noisewith speech envelope from a recorded sound file. The results are shown in figure 3.4. Itcan be seen that the amplitude of the noise (b) is higher than the amplitude of the speech(a). This is due to the fact that the fine structure of the noise is random noise and thefine structure of the speech is a periodic signal. This last signal carries the same energyas the noise with less amplitude, i.e. the speech has a lower crest factor. It can also benoted that the spectrum of the speech shaped noise is slightly different from the one ofthe speech signal (c). This is due to the convolution of the original spectrum with theenvelope spectrum, which has a reduced bandwidth as can be seen in figure 3.5.

The highest values of envelope magnitude appear at frequencies below 10 Hz, as shownin figure 3.5. It can also be seen that a secondary peak is present around 200 Hz. Thispeak represents the fundamental frequency of the speaker, in this case a female, whichhas also been observed in the long-term magnitude spectrum of the same signal in figure2.1. Therefore, it might be possible to use the envelope spectrum as a feature for speakerrecognition.

Page 19

Chapter 3. Discussion Technical University of Denmark

Time [s]

0 10 20 30 40 50-0.4

-0.2

0

0.2

0.4

(a)

Time [s]

0 10 20 30 40 50-0.4

-0.2

0

0.2

0.4

(b)

Frequency [Hz]

Magnitude

[dB

]

Input noise

Speech

Output noise

102 103 104

20

25

30

35

40

45

50

55

(c)

Figure 3.4: Temporal waveforms of the speech signal (a) and the modulated speech shaped noise (b). Themagnitude spectra of both signals and the noise are used as input are shown in (c).

Envel

ope

magnit

ude

[dB

]

Modulation frequency [Hz]

1 10 100 100020

30

40

50

60

70

80

Figure 3.5: Envelope spectrum of the speech signal.

Page 20 Intelligibility of Speech

April 2008 Chapter 3. Discussion

3.3 High-pass and low-pass filtered speech

In order to derive which frequency band is the most important for the intelligibility ofspeech one can convert the values of AI for the low-pass and high-pass filters, shown infigures 2.8 and 2.9, to the corresponding speech intelligibility values according to figure1.5 on page 5. The gradient of the resulting curves will be highest at the most impor-tant frequency bands for speech intelligibility. This is because, at these frequency regions,small changes in the passing bands of the filters give the highest improvements in speechintelligibility.

In figures 3.6 the dependence of AI and SII on the speech spectrum level can be seen.It can be seen that the AI increases with an increase of the speech level (figures 3.6 (a)and (b)), because the SNR increases. The slopes of the curves are rather similar amongstthem and at very low frequencies (125 Hz - 400 Hz) and very high frequencies (5 kHz -8 kHz) the values of AI, when the speech levels are “loud” and “shout” are identical. Onthe other hand, the SII presents a quite different behavior. For the low cut-off frequenciesof the high-pass filtering it can be seen that the higher the speech level the lower the SII,for cut-off frequencies up to 1.5 kHz. From that cut-off frequency and above the SII isimproving as the speech level is increasing. This behavior, which is different to the oneobserved for the AI concept, is due to the more sophisticated way that the SII procedure isimplemented. SII takes into account the different distribution of the energy in the speechspectrum at different levels and the effect of masking.

Thus, taking into consideration figure 3.6 it can be concluded that speech level affects ina different way the two indices.

3.4 Comparison of AI, SII and STI

The idea behind AI (and SII) is that background noise can influence intelligibility due tomasking and not all frequency components are equally important. STI is based on theassumption that the envelope of a speech signal must be perceived correctly in order thespeech to be understood correctly. This means that AI and SII are calculated directly fromthe magnitude response of the signals. On the other hand, the results given by STI arecorrelated with the preservation of the envelope. AI and SII are suited to estimate speechintelligibility of transmission channels in communication systems with limited bandwidth,whereas STI is calculated from the impulse response and is more suited for speech intelli-gibility in rooms.

Page 21

Chapter 3. Discussion Technical University of Denmark

Filter cut-off frequency [Hz]

AI

NormalRaisedLoudShout

125 250 500 1000 2000 4000 80000

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

(a) AI - highpass

Filter cut-off frequency [Hz]

AI

NormalRaisedLoudShout

125 250 500 1000 2000 4000 80000

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

(b) AI - lowpass

Filter cut-off frequency [Hz]

SII

NormalRaisedLoudShout

125 250 500 1000 2000 4000 80000

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

(c) SII - highpass

Filter cut-off frequency [Hz]

SII

NormalRaisedLoudShout

125 250 500 1000 2000 4000 80000

0.1

0.2

0.3

0.4

0.5

0.6

0.7

0.8

0.9

1

(d) SII - lowpass

Figure 3.6: Effect of different speech levels in the AI (a),(b) and SII (c), (d) for high-pass and low-passfiltering.

Page 22 Intelligibility of Speech

April 2008 Chapter 3. Discussion

STI describes the degradation of intelligibility due to masking in the modulation domain.This correlates with the concept discussed in previous reports about the existence of amodulation filterbank that seems to play an important role in human perception. Theidea of a modulation filterbank supports the assumption that STI is based on.

Page 23

Chapter 3. Discussion Technical University of Denmark

Page 24 Intelligibility of Speech

April 2008 Chapter 4. Conclusion

Chapter 4Conclusion

In this report, the intelligibility of speech has been discussed and the main conclusions arethe following:

• From the long-term magnitude spectrum, it is possible to determine the gender of thespeaker, but not the spoken language. The main peak in the spectrum correspondsto the fundamental frequency, which is around 125 Hz for male speakers and 250 Hzfor females.

• Speech-shaped noise is used as a masker for speech intelligibility tests. However, thespeech envelope is not always preserved, and some methods have been proposed togenerate speech shaped noise with the same speech envelope.

• AI and SII are affected in a different way by the change in the voice level. This isdue to the corrections on speech spectrum for different voice levels that SII applies.

• The STI is a useful parameter to assess speech intelligibility in different rooms andits calculation is based in the preservation of modulation in the speech signal. Thisseems to be in agreement with the way the auditory system perceives sounds andreinforces the assumptions of the existence of a modulation filterbank.

Page 25

Chapter 4. Conclusion Technical University of Denmark

Page 26 Intelligibility of Speech

April 2008 Bibliography

Bibliography

[Byrne et al., 1994] Byrne, D., Dillon, H., Tran, K., Arlinger, S., Wilbraham, K., Cox, R.,Hayerman, B., Hetu, R., Kei, J., Lui, C., Kiessling, J., Notby, M. N., Nasser, N. H. A.,Kholy, W. A. H. E., Nakanish, Y., Oyer, H., Powell, R., Stephens, D., Meredith, R.,Sirimanna, T., Tavartkiladze, G., Frolenkov, G. I., Westerman, S., and Ludvigsen, C.(1994). An international comparison of long-term average speech spectra. J. Acoust.Soc. Am, 96(4):2108–2120.

[Campbell, 1997] Campbell, J. (1997). Speaker recognition: A tutorial. Proceedings of theIEEE, 85(9):1437–1462.

[Dau and Christiansen, 2008] Dau, T. and Christiansen, T. U. (2008). Intelligibility ofspeech. 1.4.0 edition.

[Houtgast and Steeneken, 1973] Houtgast, T. and Steeneken, H. (1973). The modulationtransfer function in room acoustics as a predictor of speech intelligibility. Acustica,28(1):66–73.

[Houtgast and Steeneken, 1985] Houtgast, T. and Steeneken, H. (1985). The modulationtransfer function in room acoustics. B&K Technical Review, 3.

[Poulsen, 2005] Poulsen, T. (2005). Acoustic Communication, Hearing and Speech. 2.0edition. Only available from DTU campusnet.

[Poulsen, 2007] Poulsen, T. (2007). Speech. Only available from DTU campusnet.

[Steeneken and Houtgast, 1985] Steeneken, H. and Houtgast, T. (1985). Rasti: A tool forevaluating auditoria. B&K Technical Review, 3.

Page 27

Bibliography Technical University of Denmark

Page 28 Intelligibility of Speech

Project group 5 April 2008

Appendix AMatlab code

1 clc

2 clear a l l

3 close a l l

4

5 % Read wa v f i l e s

6 [y1 ,fs,Nbits]=wavread(’Speech sample 01. wav’); % Female , Eng l i sh

7 [y2 ,fs,Nbits]=wavread(’Speech sample 02. wav’); % Male , Eng l i sh

8 [y3 ,fs,Nbits]=wavread(’Speech sample 03. wav’); % Female , Danish

9 [y4 ,fs,Nbits]=wavread(’Speech sample 04. wav’); % Male , Danish

10

11 % Do something

12 N = length(y1);

13 fres=N ;%f s /N; % Reso lu t ion

14 f=fs/fres .*(0: fres /2-1); % Frequency ax i s

15

16 y1longterm =20* log10(abs( f f t (y1 ,fres)));

17 y2longterm =20* log10(abs( f f t (y2 ,fres)));

18 y3longterm =20* log10(abs( f f t (y3 ,fres)));

19 y4longterm =20* log10(abs( f f t (y4 ,fres)));

20

21 vLout1 = Oct3Smooth(f,y1longterm (1: fres /2),f_AI);

22 vLout2 = Oct3Smooth(f,y2longterm (1: fres /2),f_AI);

23 vLout3 = Oct3Smooth(f,y3longterm (1: fres /2),f_AI);

24 vLout4 = Oct3Smooth(f,y4longterm (1: fres /2),f_AI);

25

26 figure ;

27 semilogx(f,y1longterm (1: fres /2),’Color’ ,[.6 .6 .6])

28 hold on;

29 semilogx(f_AI ,vLout1 ,’k’,’LineWidth ’ ,2);

30 xlim ([100 10000])

31 ylim ([-40 70]);

32 set(gca,’XTick’ ,[100 250 500 1000 2000 4000 8000]);

33 nicefigure;

34 legend(’Long term spectrum ’,’1/3- octave smoothed spectrum ’,’Location ’,’SouthWest ’);

35 xlabel(’Frequency [Hz]’);

36 ylabel(’Magnitude [dB]’);

Page A1

APPENDIX

37 % pr i n t l a t e x ( ’ EnglishFemale ’ ,11 ,10 , ’ nof igcopy ’ ) ;

38

39 figure ;

40 semilogx(f,y2longterm (1: fres /2),’Color’ ,[.6 .6 .6])

41 hold on;

42 semilogx(f_AI ,vLout2 ,’k’,’LineWidth ’ ,2);

43 xlim ([100 10000])

44 ylim ([-40 70]);

45 set(gca,’XTick’ ,[100 250 500 1000 2000 4000 8000]);

46 nicefigure;

47 legend(’Long term spectrum ’,’1/3- octave smoothed spectrum ’,’Location ’,’SouthWest ’);

48 xlabel(’Frequency [Hz]’);

49 ylabel(’Magnitude [dB]’);

50 % pr i n t l a t e x ( ’ EnglishMale ’ ,11 ,10 , ’ nof igcopy ’ ) ;

51

52 figure ;

53 semilogx(f,y3longterm (1: fres /2),’Color’ ,[.6 .6 .6])

54 hold on;

55 semilogx(f_AI ,vLout3 ,’k’,’LineWidth ’ ,2);

56 xlim ([100 10000])

57 ylim ([-40 70]);

58 set(gca,’XTick’ ,[100 250 500 1000 2000 4000 8000]);

59 nicefigure;

60 legend(’Long term spectrum ’,’1/3- octave smoothed spectrum ’,’Location ’,’SouthWest ’);

61 xlabel(’Frequency [Hz]’);

62 ylabel(’Magnitude [dB]’);

63 % pr i n t l a t e x ( ’ DanishFemale ’ ,11 ,10 , ’ nof igcopy ’ ) ;

64

65 figure ;

66 semilogx(f,y4longterm (1: fres /2),’Color’ ,[.6 .6 .6])

67 hold on;

68 semilogx(f_AI ,vLout4 ,’k’,’LineWidth ’ ,2);

69 xlim ([100 10000])

70 ylim ([-40 70]);

71 set(gca,’XTick’ ,[100 250 500 1000 2000 4000 8000]);

72 nicefigure;

73 legend(’Long term spectrum ’,’1/3- octave smoothed spectrum ’,’Location ’,’SouthWest ’);

74 xlabel(’Frequency [Hz]’);

75 ylabel(’Magnitude [dB]’);

76 % pr i n t l a t e x ( ’ DanishMale ’ ,11 ,10 , ’ nof igcopy ’ ) ;

77

78 figure ;

79 semilogx(f_AI ,vLout1 ,’Color ’ ,[.5 .5 .5]);

80 hold on

81 semilogx(f_AI ,vLout2 ,’--’,’Color ’ ,[.5 .5 .5]);

82 semilogx(f_AI ,vLout3 ,’k’);

83 semilogx(f_AI ,vLout4 ,’--k’);

84 xlim ([100 10000])

85 ylim ([20 60]);

86 set(gca,’XTick’ ,[100 250 500 1000 2000 4000 8000]);

87 nicefigure;

88 legend(’English female ’,’English male’,’Danish female ’,’Danish male’,’Location ’,’

SouthWest ’);

89 xlabel(’Frequency [Hz]’);

Page A2 Intelligibility of Speech

Project group 5 April 2008

90 ylabel(’Magnitude [dB]’);

91 % pr i n t l a t e x ( ’ SpecComparison ’ ,11 ,10 , ’ nof igcopy ’ ) ;

92

93 % r y1 = randomizePhase ( y1 ) ;

94 % r y2 = randomizePhase ( y2 ) ;

95 % r y3 = randomizePhase ( y3 ) ;

96 % r y4 = randomizePhase ( y4 ) ; 1 clc

2 close a l l ;

3 clear variables;

4

5 [y,fs,Nbits]=wavread(’Speech sample 01. wav’); % Female , Eng l i sh

6 N = length(y);

7 t=1/fs*(1:N);

8 f=fs/N.*(0:N/2-1); % Frequency ax i s

9 yfft= f f t (y);

10 ylongterm =20* log10(abs(yfft));

11 vLout = Oct3Smooth(f,ylongterm (1:N/2),f_AI);

12 y_energy = sum(10.^( vLout /10));

13

14 y_envelope = abs(hilbert(y));

15

16 % f i g u r e ;

17 % p l o t ( t , y , ’ Color ’ , [ . 5 .5 . 5 ] ) ;

18 % hold on

19 % p l o t ( t , y enve lope , ’ k ’ , ’ LineWidth ’ , 2 ) ;

20

21 x=rand(N,1);

22 % f i g u r e ; p l o t ( t , x , ’ Color ’ , [ . 5 .5 . 5 ] ) ;

23 % x=x .∗ y enve lope ;

24 % hold on ;

25 % p l o t ( t , x , ’ k ’ ) ;

26

27 xfft= f f t (x);

28 xlongterm =20* log10(abs(xfft));

29 xvLout = Oct3Smooth(f,xlongterm (1:N/2),f_AI);

30

31 figure (111);

32 semilogx(f_AI ,xvLout ,’--k’);

33 hold on

34

35 xfft=xfft.*abs(yfft)./abs(xfft);

36 x2longterm =20* log10(abs(xfft));

37 x2vLout = Oct3Smooth(f,x2longterm (1:N/2),f_AI);

38 semilogx(f_AI ,x2vLout ,’-k’);

39 x2= i f f t (xfft);

40

41 x2=x2.* y_envelope;

42 x2fft= f f t (x2);

43 x2longterm =20* log10(abs(x2fft));

44 x2vLout = Oct3Smooth(f,x2longterm (1:N/2),f_AI);

45 x2_energy = sum(10.^( x2vLout /10));

46 x2 = x2 * sqrt(y_energy/x2_energy);

Page A3

APPENDIX

47 figure ;

48 plot(t,x2,’k’);

49 xlabel(’Time [s]’);

50 xlim ([0 50]);

51 ylim ([-.4 .4]);

52 nicefigure;

53 printlatex(’noise2_time ’,7,8,’nofigcopy ’);

54

55

56 figure (111);

57 semilogx(f_AI ,x2vLout +10* log10(y_energy/x2_energy),’-’,’Color’ ,[.5 .5 .5]);

58 xlim ([100 10000]);

59 xlabel(’Frequency [Hz]’);

60 ylim ([20 55]);

61 ylabel(’Magnitude [dB]’);

62 nicefigure;

63 legend(’Input noise’,’Speech ’,’Output noise’,’Location ’,’SouthWest ’);

64 printlatex(’specs_mod_shape ’,9,8,’nofigcopy ’);

Page A4 Intelligibility of Speech

Project group 5 April 2008

Appendix BLong-term average spectrum of speech

• s• ..................... I:" • •' •'"'"-" '•• '.'".. ': • • ' ' ':' ß : ' ' ' : : : : t' I'1 ...'.' •.•o.. ,, .. ..

--'i'" i--'i---::', 1 ',::::: :: , ,',','!11, ;!i ,;7!',',', ,

"'•1-'"'"'"'""'"'''"''" :,., ,,... •,•• .... .... ....,...,.._.: ,.

-"- ............... • :', O:

•"[•'1•i--•"'•---:".•'"! " "ß .... i: -,- _ .,.. _, - _ .,.

•,o [:• --•.- - ? - - -! ....... ::.-- i- ?-•,- • •'•1::'•,':- - ' :' ' ' -: ....... :"-':-".•"? -:•,•i'"!"-'i' ' ' :'":"":'".•'": ' "' • ....... ©,: I• 3• I , , I • I I • , I , • ! • • I • • I , • I • • I

• 12S 2S0 •00 lk 2k •k •k l•k

,.-, 65 . ,•-T•r1•'•r-• •. .• Frequency (Hz)

"øIi"'!, .... i"-!-"i-'-il ,-, LEGEND

'•"-o - ':---;-- -! •"•:-•'•";"':"'i o- "ør ;,o-/•---'r---i---!---i'"!- -': - •z_.-] , •,,•, • o. o, • f:v,• .., ,, . , ,, •"•"'"'""'"". ß . ß "'.... ,, ":• '•' •..' - •---i • . •. • o. oo,

63 !25 250 S00 ] ß 2k t•k 8k ]6k

Frequency

FIG. 1. Male and female long-term average speech sp:ctrum (LTASS) values for five samples of English. Solid line shows LTASS average across 17 speech samples (all samples except Arabic), males and femal :s separately for frequencies below 160 Hz, combined for higher frequencies.

no statistical examination of whether variability interacted with sample. For all samples combined, Fig. 8 shows the standard deviation of individual variations from the mean

value at each frequency for males and for females. The

analysis is based on data which had been normalized to a 70 dB overall level for each talker. The deviation values are

similar for both sexes and at all frequencies from 630 to 4000 Hz. Variability shows a small but consistent increase,

2113 J. Acoust. Soc. Am., Vol. 96, No. 4, October 1994 Byrne et aL: Long-term average speech spectra 2113

Figure B.1: Long-term average spectrum of speech. From [Byrne et al., 1994, p. 2113].

Page A5

APPENDIX

• ' •c'•) SUEDISH , •, l•li•) SUEDISH (• A• • , , STpCKHqLM , i • LINKOPINe

,,0 .... ,...,..., ...... .... ....................

•,, • ............................... •...•...•...•...?..•...•...• •30 I, , • , , • , , • , , • , , , , , • , , • , , I r• . . , , , • , , • . , • , F,. , , , . • , . •

.... .............. • ,o .•--i-•'• -'•/••' '•"":'"•'

• •o .•. •... •-..• -•• .•..-•- .y• ...................... • •'• ....... '•. J•...•..2_•.•.

•65•• Fr equenc• (Hz) • LEGEND

'•--•: 7-:---:-•L•':•-- :•--'•-•

::.. :.oo•... ;..-:• • AVERAGE

• 0 •LE • < O. O i

• •0 ...................... • FEMALE p ( O. O]

• •5 • •LE p < O. OOl

• .... •'' '• '' '%•• • FEMME p ( 0.00] 0 •0 6• ] 2S 250 •00 lk 2k 4k 8k 16k

Fr equenc• (Hz)

FIG. 2. Male and female LTASS values for Swedish (two samples), Danish, German, and Russian. Solid line shows LTASS average across 17 speech samples.

for both sexes, for bands above 4000 Hz. The only substan- tial increases in variability are in the bands 80 and 100 Hz for males and 125 and 160 Hz for females.

E. LTASS for Arabic

The LTASS values (dB) for Arabic, normalized to 70 dB overall level (linear), are shown in Fig. 9.

III. DISCUSSION

The overall finding of this study is that the LTASS is very similar over the wide range of languages that were ana- lyzed. Indeed, there is no single language or group of lan- guages which could be regarded as being markedly different from the others. Therefore, it is feasible to propose a univer-

2114 J. Acoust. Soc. Am., Vol. 96, No. 4, October 1994 Byrne et aL: Long-term average speech spectra 2114

Figure B.2: Long-term average spectrum of speech. From [Byrne et al., 1994, p. 2114].

Page A6 Intelligibility of Speech

Project group 5 April 2008

•'!'"•'T-•"•;'": .... :'";---:'• ,---•-- ,---,---•---•---,-

•,• .:.!.il..:: ...... ::...•..•...:• ; '1 •x •'"•"'•-" ;r '":' ' ' • - ": •t

030'1• • I, I Ill III I• I • • I •11i i I• • •1 I • • ' • • • • • I • • I • • I • • I • i I /

•o1: • :-••---•---:---•---• •' i ....

..... ',...;...:

• :, , , ', , ', ,, ',

•,•1-:•-•---::- ...... :• ..... •-• .... • , : : : : • , •,o1•__•___• ..... • +••

..,,,,,,,,.,,,,,,

• soE', -• ;-•-:•-;---:--- ;---;---:•

: • : : : :• ' ' ' b'" .... ',' ' ' • ' ' ':-:

......... •...•..•.

•,•[•./. J•..•...•... •...•.••..• : •o[•../...•...•... •...?...•...,:..• • •': • • : : : : : •'•F•'•---• .... , ..................... ••,, [,.[,.,,,,,,,.,,.,,

63 125 250 500 lk 2k •k 8k 16k

Frequenc•

FIG. 3. Male and female LTASS values for Canton.•se, Mandarin, Vietnamese, Japanese, and French. Solid line shows LTASS average across 17 speech samples.

sal LTASS that would be applicable to most (possibly all) languages and would be sufficiently prec se for many pur- poses. Nonetheless, there are small but (statistically) signifi- cant differences among languages and more substantial dif- ferences, at the low frequencies, between male and female talkers.

A. Male/female differences

Considering first the comparison between males and fe- males, the most notable feature is that their spectra are vir- tually identical over the frequency range from 250 to 5000 Hz. Within this range, the normalized male and female lev-

2115 J. Acoust. Sec. Am., Vol. 96, No. 4, Octo3er 1994 Byrne et al.: Long-term average speech spectra 2115

Figure B.3: Long-term average spectrum of speech. From [Byrne et al., 1994, p. 2115].

---•65

L

,,0 L

o 3O 63 !2S 2S0 SO0 !k 2k •k $k lSk

Fr equencw (Hz)

LEOEND

MALE /• FEMALE

MALE p < 0.0! •, FEMALE p < 0.0l

MALE p < 0.001 ß FEMALE p < 0.00l

AVERROE

FIG. 4. Male and female LTASS values for Singhalese and Welsh. Solid line shows LTASS average across 17 speech samples.

TABLE II. Male, Female, and combined speech spectra, normalized for 70 dB SPL overall level and averaged across samples (excluding Arabic). Com- bined is equal to male for frequencies up to 160 Hz; it is average of male and females for other frequencies. The spectrum (combined male and fe- male) recommended by Cox and Moore (1988) is also shown.

Frequency Cox and Moore (Hz) Male Female Combined (1988)

63 38.6 37.0 38.6 ...

80 43.5 36.0 43.5 --.

100 54.4 37.5 54.4 ...

125 57.7 40.1 57.7 ...

160 56.8 53.4 56.8 ...

200 58.2 62.2 60.2 '" 250 59.7 60.9 60.3 60.0

315 60.0 58.1 59.0 57.0 400 62.4 61.7 62.1 61.0 500 62.6 61.7 62.1 62.0

630 60.6 60.4 60.5 59.0 800 55.7 58.0 56.8 56.5

1000 53.1 54.3 53.7 55.0

1250 53.7 52.3 53.0 54.5 1600 52.3 51.7 52.0 52.0 2000 48.7 48.8 48.7 49.0

2500 48.9 47.3 48.1 48.0 3150 47.0 46.7 46.8 46.5 4000 46.0 45.3 45.6 46.0

5000 44.4 44.6 44.5 44.0 6300 43.3 45.2 44.3 45.5 8000 42.4 44.9 43.7 "'

10 000 41.9 45.0 43.4 ."

12 500 39.8 42.8 41.3 ... 16 000 40.4 41.1 40.7 ...

els, averaged over all languages, agree within 2 dB at all third-octave frequencies except 800 Hz, where the difference is 2.3 dB. (The non-normalized values agree almost as closely as there was little difference between the average overall male and female speech levels.) For frequencies of 160 Hz and below, male levels greatly exceeded female lev- els undoubtably because of the difference in the fundamental frequency ranges. These findings are consistent across lan- guages and consistent with previous research (Benson and Hirsh, 1953; Tarnoczy and Fant, 1964; Tarnoczy, 1971; Ni- emoller et al., 1974; Byrne, 1977; Pearsons et al., 1977; Cox and Moore, 1988).

140

120

100 80 6O

40

2O

0

54 58 62 66 70 74 78 82 86

Long term rms level (rib $PL)

FIG. 5. Distribution of overall rms levels (measured at 20 cm from mouth). Curve shows normal distribution fitted to data.

2116 J. Acoust. Soc. Am., Vol. 96, No. 4, October 1994 Byrne et al.: Long-term average speech spectra 2116

Figure B.4: Long-term average spectrum of speech. From [Byrne et al., 1994, p. 2116].

Page A7