Académique Documents
Professionnel Documents
Culture Documents
Introduction
245
and an aperiodic part has been proposed in [6][7]. The method aims at decomposing the excitation signal (linear prediction residual) rather than the original
speech signal.
In this paper we describe and compare three versions of a Harmonic plus
Noise Model, HNM, for speech decomposition. The periodic (or quasi-periodic)
part is supposed to be harmonic. For the rst version of HNM, denoted by
HNM1 , the harmonic part designates sums of harmonically related sinusoidal
components with constant amplitudes within each analysis frame. The phase
is modeled by a rst-order polynomial (i.e., is assumed to be linear). For the
second version, denoted by HNM2 , the periodic part is also a sum of harmonically
related sinusoidal components, however, with piece-wise linearly varying complex
amplitudes. The third version, denoted by HNM3 , makes use of a p-th order
polynomial with real coecients for the harmonic amplitudes, and the phase is
assumed to be linear. Given the harmonic part, the non-periodic part is obtained
by subtracting the harmonic part from the original speech signal. The nonperiodic part (or residual signal ) thus accounts for everything in the signal that
is not described by harmonic components. It includes the friction noise, the
period-to-period uctuations produced by the turbulences of the glottal airow,
etc [8].
The paper is organized as follows. The development and description of the
three versions of HNM is given rst. The second part is devoted to estimating
parameters for the harmonic part of the three models, followed by a comparison
of the three harmonic models. The comparison will be done on the variance of
the residual signal obtained from the three models, the spectral content of the
residual signal and the error modeling. Although the paper focuses on the harmonic part modeling, some models for the noise part are also proposed. Finally,
the paper concludes with a discussion of the usefulness of the proposed models
for prosodic modications of speech.
The basic approaches to speech analysis proposed in this paper are to approximate a discrete-time sequence s[n] using three versions of HNM[9][10]. The
sequence s[n] represents the sampling version of the continuous speech signal
s(t) sampled at a rate of Fs samples per second. HNM assumes the speech signal
to be composed of a periodic component h[n] and a non-periodic component
r[n]. The periodic component are designated as sums of harmonically related
sinusoids
HNM1 : Sum of exponential functions without slope
L(nia )
h1 [n] =
k=L(nia )
(1)
246
Yannis Stylianou
L(n
a )
i
i
Ak (n) expj2kf0 (na )(nna )
h2 [n] =
(2)
k=1
where
(3)
with ak (nia ), bk (nia ) to be complex numbers (amplitude and slope respectively). denotes taking the real part.
HNM3 : Sum of sinusoids with time-varying real amplitudes
L(nia )
h3 [n] =
ak (n)cos(k (n))
(4)
k=0
where
(5)
where p(n) is the order of the amplitude polynomial, which is, in general, a
time-varying parameter.
The parameters of the three models are updated at specic time-instants
denoted by nia . f0 (n) and L(n)1 represent the fundamental frequency and the
number of pitch-harmonics included in the harmonic part, respectively. These
two parameters are held constant within each analysis frame, and they equal the
values at the center of the analysis window, nia .
The non-periodic part is just the residual signal obtained by subtracting the
periodic-part (harmonic part) from the original speech signal in the time-domain
r[n] = s[n] h[n]
(6)
3.1
Assuming voiced speech, the rst step of the analysis process consists of estimating the fundamental frequency and the number of harmonics to be included in
the harmonic part. In a rst step, an initial pitch estimation is obtained by using
a standard time-domain pitch detector based on a normalized cross-correlation
function [11]. Then, this initial pitch estimate is used to separate voiced from
unvoiced frames based on an analysis-by-synthesis method.
The voiced/unvoiced estimation has also to be based on a criterion which
takes into account how close the harmonic model will be to the original speech
1
For convenience, the index nia will be omitted from the number of harmonics, L
247
where f0 is the initial fundamental frequency estimate. If the error E is below the
threshold, this frame is marked as voiced. Otherwise, it is marked as unvoiced.
The next step of the process is the estimation of the maximum voiced frequency, Fm . The largest sine-wave amplitude (peak) in the frequency range
[f0 /2, 3f0/2] is found. For this peak, two amplitudes and a frequency location
are determined. The location is dened simply as the frequency at which the
actual peak occurs and it is denoted here by fc . The rst amplitude is dened
as the magnitude of the sample at the peak, while the second amplitude is dened as the non-normalized sum of the amplitudes of all of the samples from
the previous valley to the following valley. (see Fig.1). This second amplitude is
called cumulative amplitude and it is denoted by Amc in contrast to the simple amplitude : Am. Using the cumulative amplitudes also resulted in a better
Am
*
*
*
*
fc
Amc
separation between true peaks and spurious peaks compared to using just the
simple amplitude at the peak [12]. The peaks which are in the frequency range
[fc f0 /2, fc + f0 /2] are also considered and the two types of the amplitudes are
calculated for each peak. Denote the frequencies of these neighboring peaks and
the mean value of their cumulative amplitudes, by fi and Amc(fi ) respectively,
the following harmonic tests are applied to the largest peak:
248
Yannis Stylianou
if
Amc(fc )
>2
Amc(fi )
(8)
(9)
|fc lf0 |
< 10%
lf0
(10)
or
then, if
Using the stream of the estimated pitch values, the position of the analysis timeinstants, nia , are set at a pitch-synchronous rate 2 .
The second step of the analysis process for voiced frames of speech consists of
estimating the harmonic amplitudes and phases. This is done using a weighted
least-squares method aiming at minimizing the following criterion with respect
to ak (nia ) for HNM1 , to ak (nia ) and bk (nia ) for HNM2 and to unknown coecients
of the amplitude and phase polynomial for HNM3 :
2
One can use a non pitch-synchronous analysis. However, as our goal is to use these
models for prosodic modications, we prefer to work in a pitch-synchronous context
249
nia +N
=
w2 [n](s[n] h[n])
(12)
n=nia N
in which
h[n] is dened as in (1), (2) or (4) for HNM1 , HNM2 or HNM3 respectively. w[n] is a weighting window and N is the integer closest to the local
pitch period T (nia ). The above criterion has a quadratic form for the parameters in HNM1 and HNM2 , and is solved by inverting an over-determined system
of linear equations [13]. For HNM3 , a non-linear system of equations has to be
solved.
At this point in the analysis, we nd it convenient to switch to matrix notation. In particular we may write the harmonic parts of the three models as
for HNM1 : P1 x1 = h1
for HNM2 : P2 x2 = h2
for HNM3 : P3 x3 = h3
(13)
..
..
..
..
P1 = bL . bL+1 . bL+2 . . bL
(14)
where bk is a (2N + 1)-by-1 vector corresponding to k-th harmonic and it is
dened by :
T
i
i
i
i
(15)
bk = ej2kf0 (ta N ) ej2kf0 (ta N +1) ej2kf0 (ta N +2) ej2kf0 (ta +N )
where the superscript T denotes the transpose operator.
The matrix P2 is formed by concatenating the following matrices
with
P2 = [B1 |B2 ]
(16)
(17)
.
.
.
.
P3 = B1 .. diag(n)B1 .. diag(n)2 B1 .. ..diag(n)p B1
T
(18)
.. ..
.. ..
(19)
B1 = b1 . b2 . . bL . 1
250
Yannis Stylianou
for k = 1 L
(20)
T
x1 = aL aL+1 aL+2 aL
(21)
(22)
..
.
.
. c11 c21 cL1 c01 .. .. c1p c2p cLp c0p
T
(23)
HNM2 :
1
P1 h WT Ws
x1 = P1 h WT WP1
(24)
1
x2 = P2 h WT WP2
P2 h WT Ws
(25)
where s denotes the (2N + 1)-by-1 vector of the original speech samples of the
current frame
T
(26)
s = [s[N ] s[N + 1] s[N ]]
and W denotes the (2N + 1)-by-(2N + 1) diagonal matrix with diagonal entries
the weight vector
T
w = [w[N ] w[N + 1] w[N ]]
(27)
which is typically a Hamming window. Note that the choice of the weighting
window w[n] is important in order to control the variance of the residual signal. An extended discussion of the importance of the choice of the window can
be found in [10] [8]. The superscript h denotes the transpose and conjugate
operation whereas the superscript T denotes just the transpose operator.
It can be shown[10] that for HNM1 the matrix to invert in (24) is Toeplitz
which means that fast algorithms can be used to solve the respective linear set
of equations.
251
The solution for the third model can not be obtained so easily. Specically,
as the coecients k and are unknown and the amplitude polynomial has an
order greater than zero, the above error criterion leads to nonlinear equations
for HNM3 and then, the solution must be calculated using iterative methods.
We have found that using only one iteration was sucient and so the following
analysis scheme was adapted. As the instanteneous frequency, f0 (n), has been
already calculated during the rst step of the analysis and given that within the
analysis frame |n nia | N we suppose that f0 (n) = f0 (nia ), it is straightforward to show that the parameter of the phase polynomial is equal to f0 (nia ).
Denoting by ik the phase value of the k-th harmonic at the time-instant nia , and
evaluating the phase polynomial at nia , using (5), the coecients k are given by
k = ik for k = 1, , L
(28)
Note that ik can be eciently estimated using the HNM1 analysis process,
presented above. In fact, HNM1 makes use of a zero order polynomial for the
harmonic amplitudes and a rst order polynomial for the phase. Hence, using the
HNM1 analysis step we estimate the phase ik and then by (28) the coecients
k . Having determined all the coecients of the phase polynomial, the next step
consists of estimating the coecients of the amplitude polynomial. As (t) is
known for each n [nia N nia + N ], it turns out that the solution to the
least-squares error is given by
1
P3 T WT Ws
x3 = P3 T WT WP3
(29)
Unfortunately, the matrix to invert in HNM3 does not have such an attractive
structure (like in HNM1 and HNM2 ). However, it is symmetric and positive
denite. One method that could be used is the Cholesky decomposition[14] for
solving this linear set of equations. Note that this method is about a factor of
two faster than alternative methods for solving linear equations[14].
Lastly, the harmonic part is readily obtained by
i = W1 Pi xi
h
(30)
(31)
Note that the subscript i in (31) denotes the model that we use.
Given that the analysis frame is always two local pitch periods long, 2T (nia )
(see Eq.12), to avoid conditioning problems during the analysis, the number of
parameters that must be estimated should be less than 2T (nia ). In order to take
into account harmonics up to Fs /2 (the maximum number of harmonics), where
252
Yannis Stylianou
Fs denotes the sampling frequency, one would choose L(nia ) = T (nia )/2, meaning
that the maximum number of harmonics should be less (in theory, equal) to half
of the local pitch period, T (nia )/2. For HNM1 , 2L(nia ) complex parameters must
be estimated so the analysis frame should be 4L(nia ) long, or 2T (nia ). For HNM2 ,
2L(nia ) complex parameters must be estimated, meaning that the analysis frame
should be as long as for HNM1 . Finally, for HNM3 , (p + 2)L(nia) real parameters
must be estimated, where p is the order of the amplitude polynomial. The length
of the analysis frame constrains the maximum order, pmax that can be used. For
example, if we choose the analysis frame to be two local pitch periods long, it
turns out that the order p has an upper bound pmax = 2.
We have found that the maximum voiced frequency seldom exceeds the
threshold of 6500Hz (a sampling frequency of 16kHz was used). Thus, taking
as upper bound for the maximum voiced frequency of 6500Hz, and respecting
p 2, the analysis frame could be as short as 1.6 T (nia). This means that as the
analysis frame is always equal to twice the local pitch period, we avoid in practice the ill-conditioning of matrix Pi T WT WPi , where the subscript i denotes
the model that we use.
Before addressing modeling of the noise part, some properties of the residual
signals obtained from the proposed HNM versions using (31) should be discussed.
In the following, we will assume that the order p of the amplitude polynomial of
HNM3 is xed: p = 2.
5.1
It can be shown [10] [8] that if the input signal, s[n], is a white noise with unit
variance then the covariance matrix is given by3 :
E(rrh ) = I WP(Ph Wh WP)1 Ph Wh
(32)
and the variance of the residual signal, r[n], is the diagonal of the above matrix. Using the same weighting window w(t), typically a Hamming window, the
variance of the residual signal from each of the harmonic models has been computed. The three variances are depicted in the Fig.2. The variance of the HNM1
residual signal is represented by a solid line, the variance of the HNM2 residual
by a dashed line, and nally the variance of the HNM3 residual signal by a dashdotted line. As is made clear by Fig.2 the variance of the HNM1 residual signal
is not evenly distributed across the analysis frame (as it ideally should be) in
contrast to the variance of the residual signal obtained from the two other models. The variance of the HNM2 residual error is comparable with the variance
3
To simplify the notation the the subscript i which represents the model version has
been omitted and the upscript h denotes conjugate and/or transpose depending
on the case
253
Variance
0.6
0.5
0.4
0.3
0.2
0.1
0
10
20
30
40
50
Time in samples
60
70
80
90
Fig. 2. Variance of the least-squares residual signals from: HNM1 (solid line), HNM2
(dashed line) and HNM3 (dash dotted line). The weighting window was a typical hamming window
of the residual from HNM3 . Furthermore, it can be seen that the variance from
HNM3 is closest to the ideal least-squares variance for most of the analysis time
than the variance of the HNM2 residual signal. However, at the boundaries of
the analysis frame the variance of the HNM3 residual is irregular.
5.2
In this section we compare the residual signals in the time-domain. From the
discussion presented in previous sections, we recall that HNM1 makes use of a
stationary model for the estimation of the harmonic amplitudes and phases. The
amplitudes and the phases estimated by this model correspond at the center of
each analysis frame and there is, therefore, no information about the parameters
variation within the analysis frame. This causes low frequencies to appear in the
HNM1 residual signal. To demonstrate this characteristic of the HNM1 residual
signal, a voiced fricative frame from a speech signal sampled at 16kHz was
selected. Data of this frame is plotted in Fig.3(a); the length of the analysis frame
is twice the local pitch period. Harmonic peaks have been found up to 4000Hz.
To synthesize this frame, the harmonic parameters from the frame vicinity (one
frame before and one next) have been used and then the overlap-add approach
has been applied. This means that we postulate a linear model for the harmonic
amplitudes (approximately the case of HNM3 when p = 1).
Fig.3(b) shows the residual signal from HNM1 . This gure clearly shows the
behavior of the HNM1 residual signal indicated above. For the same frame the
residual signals for HNM2 and HNM3 are presented in Fig.3(c) and (d), respectively. At rst glance in Fig.3, it seems that information about the periodicity
of the original speech signal is not detectable in the residual signals of HNM2
254
Yannis Stylianou
4
1.5
x 10
1000
0.5
0
0
1000
0.5
1
0
2000
100
200
(a) Number of samples
3000
0
3000
3000
2000
2000
1000
1000
1000
1000
2000
2000
3000
0
100
200
(c) Number of samples
100
200
(a) Number of samples
3000
0
100
200
(d) Number of samples
Fig. 3. (a) a voiced fricative speech signal and the residual error signals from (b)
HNM1 , (c) HNM2 and (d) HNM3
and HNM3 (they appear to be noisy). Also, based on the same gure we are
tempted to say that the residual errors from HNM2 and HNM3 are very similar.
As we will see later, when the magnitude spectrum of the residual signals will
be compared, this is not really true.
To measure the similarity between the harmonic part and the original speech
signal the following expression for the error has been used:
E = 10log10
2
r(t)
2
s(t)
(33)
2
2
denotes the variance of the residual signal r(t) and s(t)
denotes the
where r(t)
variance of the original speech signal s(t). For example, the error produced by
the three models for the original speech frame in Fig.3 was : 15.8dB for HNM1 ,
25.56dB for HNM2 and 25.58dB for HNM3 . Fig.4(a) shows a segment of a
speech signal (wazi waza) and in (b) the modeling error in dB produced from
the three models is plotted. The modeling error of HNM1 is presented by a
solid line, the error of HNM2 by a dashed line, and the error of HNM3 by a
dash-dotted line. It is clear by this gure that the modeling error using HNM1
is greater than those produced by the two other models. As the full frequency
band (from 0Hz up to half of the sampling frequency) of the residual signal is
used in the denition of the modeling error, the error is large on voiced fricative
regions (for example, between the 0.25 and 0.35 sec.). To give an acoustic notion
to the modeling error, note that if the modeling error is less than 25dB, the
synthetic speech signal produced from the model is indistinguishable from the
original speech signal.
255
x 10
2
1
0
1
2
0
0.1
0.2
0.3
0.4
0.5
0.6
Time in sec
0.7
0.8
0.9
0.1
0.2
0.3
0.4
0.5
0.6
Time in sec
0.7
0.8
0.9
Error in dB
10
20
30
40
50
60
0
5.3
As was mentioned above, although the HNM2 and the HNM3 residual signals
seem to be similar in the time domain, in the frequency domain they do not
have the same properties. This can be seen in the next gure.
Fig.5(a) shows, for the same voiced frame as before (Fig.3), the magnitude (in
dB) of the HNM2 residual signal (solid line) and of the original speech (dashed
Residual from HNM2
Magnitude (db)
80
60
40
20
0
20
0
1000
2000
3000
4000
5000
6000
7000
8000
6000
7000
8000
Magnitude (db)
80
60
40
20
0
20
0
1000
2000
3000
4000
5000
Frequency in Hz
Fig. 5. Magnitude of the Fourier transform of the residual signal (solid line) of (a)
HNM2 and (b) HNM3 . The magnitude of the Fourier transform of the original speech
signal has also been included (dashed line)
256
Yannis Stylianou
line). The same representation is shown in Fig.5(b) for the HNM3 residual signal.
The comparison of part (a) and (b) of the Fig.5 reveals that HNM2 has cancelled
the noise part in the harmonic band (in this example from 0 to 4000Hz) in contrast to HNM3 which leaves the noise part untouched. This means that this part
of the noise signal was passed on into the harmonic part in the case of HNM2 .
Fortunately, this is not the case for HNM3 . To make this property of HNM3
clearer, the following test was done: white noise was ltered by a low pass lter
with cuto frequency equal to the maximum voiced frequency (4000Hz). The resulting noise signal was added to the original speech signal. Then the HNM2 and
the HNM3 parameters were estimated using the noise-corrupted speech signal.
The frequency content of the two residual signals (HNM2 and HNM3 ) are shown
in Fig.6. This gure clearly shows that the additive noise (about 25dB) below the
speech level has been absorbed in the harmonic part in the case of HNM2 , as it is
not present in the low (harmonic) band of the HNM2 residual signal. In contrast,
for HNM3 , the noise level into the harmonic band has clearly been augmented.
(a) Residual from HNM2
80
60
dB
40
20
0
20
0
1000
2000
3000
4000
5000
(a) Frequency in Hz
6000
7000
8000
6000
7000
8000
dB
40
20
0
20
0
1000
2000
3000
4000
5000
(a) Frequency in Hz
Once the harmonic part has been obtained using any of the above mentioned
harmonic models, the noise part can be obtained from (6). For high-quality
re-synthesis of the residual signal, a hybrid representation is necessary: a joint
time-domain and frequency-domain representation. In previously reported work,
the stochastic part was modeled in the frequency-domain exclusively [1, 3, 15].
However, residual signals obtained from real speech recordings (voiced fricatives, plosives etc...) exhibit a specic time-domain structure in terms of energy
localization[9]. The model proposed here follows this observation: the frequency
contents of the stochastic part is described by a time-varying AR envelope; its
time-domain structure is represented by a time-domain energy-envelope function. In each time-frame, the residual signals spectral density function is mod-
257
eled by tting an lth-order all-pole lter by use of a standard autocorrelationbased LPC method. Finally, the temporal energy distribution-function d(t) is
calculated by squaring the residual signal, low-pass ltering it with a positive
impulse-response lter and decimating it by a factor R. In practice, for a signal
sampled at 16kHz, the AR model order is set to 15, and the decimating factor
to 10.
Note that the above noise model can be applied only to HNM2 and to HNM3
residual signals; the HNM1 residual signal contains low frequencies which would
bias the AR model. For HNM1 , it is better to t an lth-order AR model to
the original signals spectral density function. Then, during the synthesis step,
a high-pass lter with cuto frequency equal to the local maximum voiced frequency can be used to eliminate the low frequencies. Obviously, in that case the
temporal energy distribution function can not be estimated as before. Some deterministic envelopes can be used instead. Fig.7 shows the envelopes that have
been tested. It was found that they are speaker-dependent. However, the envelope in Fig.7(a) was found to be more consistent for the most part of our
database speakers and it was the envelope nally used.
(a)
(b)
1
T1
l1
l2
T2
T1
T2
T1
l1
l2
(c)
T2
T1
T2
(d)
Fig. 7. The time-domain envelope used for the time-behaviour of the noise part from
HNM1 . The T1 and T2 represent two successive analysis time-instants. Typical values
for l1 and l2 are : l1 = 0.15 (T2-T1) and l2 = 0.85 (T2-T1)
Synthesis
The synthesis of the speech signal can be performed in two dierent ways using
either a pitch-synchronous overlap-add technique, or a time-domain sinusoidal
synthesis. In this paper we will only present the rst alternative while for the
258
Yannis Stylianou
second one the interested reader may check in [10]. The harmonic and the noise
components are synthesized separately and then added together.
The deterministic part is simply obtained by overlap-adding a stream of
short-time signals si (t) at the specic time-instants ti in pitch synchronous manner. Here, the short-time signals are obtained from the harmonic parameters
using a synthesis equation depending on the model that is used (e.g., 1), (2), or
(4)and applying a Hamming window centered at ti . The length of the synthesis window is twice the local pitch-period, and the overlap between successive
short-term signals is one pitch-period.
The noise component is obtained by ltering a unit-variance white Gaussian
noise through a time-varying normalized lattice lter whose coecients ki (t) are
derived from the stream of A(ti , z) using a linear interpolation of the model coecients on a sample-by-sample basis. Note that, since the lattice is normalized,
the energy of the output signal is constant in time (and equal to unity). The
time-domain energy envelope-function is then applied directly on this synthetic
signal.
This paper concentrated on the harmonic modeling of speech, based on a Harmonic plus Noise Model. The analysis has been concentrated on the harmonic
part development; the noise part is assumed to be obtained from the substraction of the harmonic part from the original speech signal. Some directions have
been given for the modeling of the noise part. Three versions have been proposed for modeling periodic speech aimed at text-to-speech synthesis based on
acoustical units concatenation. From the three models presented in this paper,
HNM1 is the simplest one. HNM1 has been found to have the largest modeling error among the proposed models. This was expected, as HNM1 does not
include any information about the variation of the harmonic amplitudes; only
during the synthesis step these parameters are changed linearly. HNM1 seems
to be an acceptable model for strongly voiced speech frames where the period
does not change quickly from one frame to another. The analysis step in HNM1
is faster than in the other two models. The noise model proposed for HNM1 is
very simple and it can be easily manipulated during prosodic modications. The
output speech signal produced by HNM1 is almost indistinguishable from the
original one; some slight distortions are observed in transitional voiced speech
frames where pitch changes quickly, or in voiced/unvoiced to unvoiced/voiced
passages.
HNM2 has low modeling error but it cancels completely the noise part in the
harmonic band (from 0Hz up to the maximum voiced frequency), absorbing it
in the harmonic part. Thus, during prosodic modications the noise part of the
harmonic band will be repeated (if, for example, a time scaling by factor 2 is
applied) producing an articial long-term autocorrelation in the output signal
that is perceived as some sort of periodicity [10]. This is also a problem of the
PSOLA approach for prosodic modications[16]. The signal produced by HNM2
259
is indistinguishable from the original one, when no prosodic changes are made.
We use a large number of parameters to model speech, and taking into account
that this model it is not useful for speech modications, it tend to be excluded
from further analysis.
HNM3 also has a small modeling error and, in contrast to HNM2 , does not
cancel the noise part of the harmonic band. Original speech signals are indistinguishable from their synthetic version produced by HNM3 . Using a second
order amplitude polynomial quasi-periodicity and fast transitions are eciently
modeled. A nice property of the amplitude polynomial coecients, ck1 and ck2 ,
is that they exhibit a strong peak during transitions from one phoneme to another inside a voiced island. During steady regions of voiced speech they have
almost constant values. This can be useful in two ways. First, one could combine
HNM1 and HNM3 using a time-dependent order p, thus reducing the number
of parameters of HNM3 , and second, HNM3 could be used for automatic speech
segmentation or detection of steady speech regions.
Comparing the noise models proposed in this paper, HNM2 and HNM3 are
not as exible as HNM1 , but are more accurate; the temporal energy distribution
function that they use is estimated in the analysis step and it is not imposed as
for the HNM1 noise model. However, in HNM1 , a parametric energy distribution
function is simple to stretch in the case of pitch and time-scale modications. A
disadvantage of this approach is that the same envelope is used for all speakers.
Further developments will aim at combining HNM1 and HNM3 for highquality speech synthesis and prosodic modications. Also, a speaker-dependent
parametric temporal energy envelope needs to be explored in further research,
as this is an important factor for high-quality speech reproduction.
Acknowledgment
The author would like to thank J. Laroche, and E. Moulines for their valuable
remarks during this work.
References
1. Grin, D., Lim, J.: Multiband-excitation vocoder. IEEE Trans. Acoust., Speech,
Signal Processing ASSP-36 (1988) 236243
2. Dutoit, T., Leich, H.: Text-To-Speech synthesis based on a MBE re-synthesis of
the segments database. Speech Communication 13 (1993) 435440
3. Serra, X., Smith, J.: Spectral modeling synthesis: A sound analysis/synthesis system based on a deterministic plus stochastic decomposition. Computer Music J.
14 (1990) 1224
4. Serra, X.: A System for Sound Analysis/Transformation/Synthesis Based on a
Deterministic Plus Stochastic Decomposition. PhD thesis, Stanford University,
Stanford, CA (1989) STAN-M-58.
5. Rodet, X., Depalle, P., Poirot, G.: Speech Analysis and Synthesis Methods Based
on Spectral Envelopes nad Voiced/Unvoiced Functions. In: Proc. EUROSPEECH,
Edinburgh, U.K. (1987)
260
Yannis Stylianou