• No results found

Spoken language variety recognition in Dutch

N/A
N/A
Protected

Academic year: 2021

Share "Spoken language variety recognition in Dutch"

Copied!
8
0
0

Loading.... (view fulltext now)

Full text

(1)

PDF hosted at the Radboud Repository of the Radboud University

Nijmegen

The following full text is a publisher's version.

For additional information about this publication click this link.

http://hdl.handle.net/2066/135159

Please be advised that this information was generated on 2017-12-05 and may be subject to

change.

(2)

Spoken language variety recognition in Dutch

David A. van Leeuwen d.vanleeuwen@let.ru.nl

Centre for Language and Speech Technoloy, CLS, Radboud University Nijmegen, The Netherlands

Daan Henselmans d.r.henselmans@gmail.com

Department of Electrical and Electronic Engineering, Stellenbosch University, South Africa

Thomas Niesler trn@sun.ac.za

Department of Electrical and Electronic Engineering, Stellenbosch University, South Africa Keywords: speech, accent recognition, spoken language recognition, Dutch, Flemish

Abstract

We used the Duch Spoken Corpus CGN Cor-

pus Gesproken Nederlands to train a spo-

ken language variety recognition system, that is capable of discriminating between Dutch spoken in The Netherlands and Belgium. We use a Support Vector Machine classifier to recognise utterances, with a linear ker- nel in a high-dimensional (50176) supervec- tor space. The supervectors are scaled shifts of the concatenated means of a Gaussian Mixture Model (GMM). The shifted means for an utterance are found by Maximum A Posteriori adaptation of a so-called Univer- sal Background Model, a 1024-component GMM that is trained with all available train- ing speech in both language varieties. The acoustic features used are Shifted Delta Cep- stra, based on Mel Frequency Cepstral Coef- ficients obtained from short duration frames of the speech signal. The discrimination per- formance varies with test segment duration, from an Equal Error Rate of 16.1 % for 3 sec- ond segments to 3.4 % for 30 s segments.

1. Introduction

Spoken language recognition is an area in speech tech- nology covering tasks where information about the identity of the language spoken in a speech utterance is automatically extracted. Spoken language is one of the ‘meta data’ items in speech, for which we usually con- sider the spoken content, i.e., the words, as the main

Appearing in Proceedings of BENELEARN 2014. Copy- right 2014 by the author(s)/owner(s).

information carrying type of data. Other meta data types include speaker identity, age, gender, and phys- ical, mental, and emotional state. Some of these have discrete, well defined, reference values, such as speaker identity. For other types the best reference value may be a self-reported value on a continuous scale, e.g., arousal as an emotional state parameter. Spoken lan- guage variety (Despres et al., 2009) is a politically neutral description for what popularly is known as

accent —but the latter term has many connotations

that may be perceived judgemental (e.g., countryside, lower-class, foreign, etc.) Sometimes the difference in language variety can be originating from political be- liefs, such as the difference between Hindi and Urdu, whereas other language varieties are the results of his- toric geographical separation and divergent develop- ment of the languages, such as Dutch and Afrikaans. Spoken language variety recognition can therefore be seen as a task similar to language recognition, but less well defined w.r.t. the reference truth, and in general harder.

In this paper we study the task of discriminating be- tween the language variety of Dutch spoken in The Netherlands and Belgium. The latter, spoken in the provinces in Flanders, is often referred to as Flem- ish. The difference in language varieties is perceived by most speakers of both countries as quite easy to recognize, although accent difference in regions near the border can be less trivial to detect. The difference is apparent from different lexical choices, subtle differ- ences in word order, and different acoustic realisations of the phonemes, i.e., allophones. Because the infor- mation density of the acoustics is much higher than in lexical choices and word order, it is beter suited for shorter duration utterances. Computational ap- proaches at the acoustic feature level further have the

(3)

Spoken language variety recognition in Dutch

92

advantage that no detailed knowledge of the language is needed, and that the implicit task of speech recog- nition, which is necessary for lexical features, does not need to be carried out. In this paper we therefore limit ourselves to these ‘acoustic approaches’ of lan- guage variety recognition. The main challenge of lan- guage (variety) recognition is that the variabilities of speaker, lexical content and acoustic channel that are inherent in the speech signal need to be dissociated from the language identity encoded implicitly in the speech signal.

This research was carried out in the context of the project ‘Elftal’ (Eleven Likelihoods For The (South) African Languages), sponsored by the Dutch Language Union and the South African Ministry of Arts and Cul- ture. In this project we carried out language recogni- tion in the eleven official languages of South Africa us- ing a Parallel Phone Recogniser Language Modelling (PPRLM) approach (Zissman, 1996; Matˇejka et al., 2005), and Dutch language variety recognition using an acoustic approach. Two different approaches were taken for the different tasks, in order to gain more ex- perience with the approaches. The assignment of the approach to task was—from a scientific point of view— arbitrary, and because of the confounding of task and approach we will not make any attempt to compare either in this paper.

The purpose of the paper is to make the general ma- chine learning community more familiar with speech as an application thereof, and the classification of Dutch and Flemish as two language variants of Dutch— covering the geographical realm of Benelearn —is just an example of the various speech technologies. The focus in speaker, language and accent recognition has shifted from signal processing to machine learning over the years, and we believe the speech and machine learning communities can benefit from more interac- tion.

The paper is organised as follows. In Section 2 we re- port quite elaborately on the technical details of acous- tic spoken language variety recognition. In Section 3 we describe the experiments carried out, and finally in Section 4 the results are presented and put in context.

2. Acoustic language variety

recognition

The approach we present in this paper is based on work by Campbell (Campbell et al., 2006) where Sup- port Vector Machines (SVMs) were applied in the con- text of automatic speaker recognition. This approach was later applied in automatic language recognition in the NIST Language Recognition Evaluation by sev- eral researchers (Castaldo et al., 2007; van Leeuwen &

Bru¨mmer, 2008). The system architecture is based on acoustic feature extraction and normalisation, followed by feature space modelling resulting in a utterance- dependent model. The parameters of the model are then used as coordinates in an SVM to discriminate between the labels of the utterances, which in our case are the language varieties. The classifier can produce a probabilistic output, which we can directly evaluate for its classification and probabilistic properties, or fur- ther process through a re-calibration step to improve these properties.

The features represent sort-time dynamics of the spec- tral properties of the signal, and include correlations of these over the duration of a short syllable (approx- imately 200 ms). Thus they encode the dynamics of the sounds rather than the sounds themselves, and the idea is that this is less speaker and content specific but more language specific. The acoustic models form a fixed-length encoding of the variable-length utterance, which makes it more easy to implement the classifier. Finally, the SVM carries out the actual classification, treating speaker and lexical variabilities as noise. 2.1. Acoustic features

Contrary to other application fields of pattern recogni- tion, almost all areas of speech processing are based on a limited set of standard feature sets, that appear to encode all important aspects of speech. More specif- ically, Mel Frequency Cepstral Coefficients (MFCC) and Perceptual Linear Prediction (Hermansky, 1990) are amongst the most popular choices. We have opted for MFCCs, which are computed by analysing the signal in short duration (25 ms) frames, by comput- ing the power spectrum and binning the energy in 40 bands chosen linearly on the Mel scale, which is inspired by human perception. The power is com- pressed logarithmically, and the binned power spec- trum is converted to a ‘cepstrum’ by applying a dis- crete cosine transform, which decorrelates the features to some extent. For language recognition purposes, we take the first n = 7 cepstral coefficients, and com- pute time-wise derivatives by fitting 1st order poly- nomials over D = 3 consecutive frames, which are spaced 10 ms, for each of the features. These so-called ‘delta’ cepstral features are stacked k = 7 times by interleaving each with p = 3 frames, such that the total the final feature vector consists of 49 parame- ters, spanning a period of 200 ms which is typical for a short syllable in speech. These acoustic features have been coined “shifted delta cepstra” (SDC) (Torres- Carrasquillo et al., 2002), and the specific settings for (n, D, k, p) were reported in (Singer et al., 2003) as op- timal settings for language recognition. Speech Activ- ity Detection retains only frames with enough energy

(4)

93

N

(within 30 dB from the maximum found in the utter- ance), and after SDC extraction each feature stream within an utterance is short-time Gaussianized using

where Fi are the 1st order statistics aligning the speech

with the UBM

) a window of 400 frames, which in speaker recognition

is known as feature warping (Pelecanos & Sridharan,

Fi = fk p(i | fk , M ). (5) k

2001). The short-time Gaussianization has been found to give slightly better performance than other fea- ture normalization techniques such as mean/variance-

The components mij of this vector, where i is the

Gaussian component index and j is the feature dimen- sion index, can further be scaled by √w /σij , where

normalization (shifting and scaling such that for each

utterance the mean is 0 and the standard deviation wi are the GMM component weights

i

and σij are the

is 1).

2.2. Feature space modelling

The feature space of speech is modelled by a sin- gle, large, Gaussian Mixture Model (GMM) known under the name “Universal Background Model” (UBM) (Reynolds et al., 2000), a term from speaker recognition referring to the modelling of a background model of ‘all other speakers’ in a likelihood ratio test. The likelihood function of the UBM M given a frame

f is

L(f | M ) =

)

wi N (f | µi , Σi ) (1) i

where wi are the weights of the Gaussians N with

mean µi and covariance Σi . We used a GMM of 1024

diagonal covariance components, trained by starting with a single multivariate Gaussian and successively splitting the Gaussians in two and separating the means. After each split five Expectation Maximisation iterations are run adapting wi , µi and Σi to maximize

the likelihood of the training data. 2.3. Supervectors

Given the UBM and a speech utterance, a new GMM can be derived from the UBM through Maximum A Posteriori adaptation of the means of the UBM (Gau- vain et al., 1994; Reynolds et al., 2000). For each Gaussian, a coefficient ζi governs the shift of that

Gaussian according to

Ni

corresponding diagonal covariances. The scaled shifts

xij = √wi mij /σij form supervectors characterising

the speech utterance that show good distance prop- erties. The inner product between two supervectors a and b can be seen as an upper bound to the Kullback- Leibler divergence between the GMMs represented by a and b (Campbell et al., 2006).

2.4. Support Vector Machines

The inner product distance properties of the supervec- tors make them suitable for the application in Support Vector Machines using the linear kernel

K (a, b) = a · b (6) since it satisfies the Mercer condition (Campbell et al., 2006). The SVM model is formed by a sparse set of weights αi and offset d for training vectors xi with

labels yi = ±1. The vectors xi support a hyperplane

that separates the training data with a maximum mar- gin criterion. Classification of a test vector x is carried out by computing the classifier function

f (x) = ) αi yi xi · x + d (7) i = () αi yi xi l · x + d ≡ wt x + d, (8) i

where w is the expression in parenthesis. This is a simple inner product which makes the classification of multiple test samples equivalent to a matrix-vector product. ζi = i , (2) + r

3. Experiment

where r is a ‘relevance factor’, usually chosen 16. Ni

are the 0th order statistics aligning the speech with the UBM

Ni =

)

p(i | fk , M ), (3) k

where the sum is over all frames fk of the utterance of the posterior probabilities of the Gaussian i given frame fk . These are also known as the responsibilities of Gaussian i. The shift in the mean is computed as

Fi

The speech data was obtained from the Corpus Gesproken Nederlands (CGN (Oostdijk & Broeder, 2003)), components c and d, which are conversational telephone recordings between acquaintances recorded with a telephony platform and a local mini-disc, re- spectively. The database consists of speakers recruited from The Netherlands and Belgium, making up the two language varieties under study. We split the data in two parts, a training set and a test set, making sure there was no overlap in speaker identity. The test mi = ζi

(

Ni l

− µi , (4)

(5)

Spoken language variety recognition in Dutch

94

llr speakers that were used in the development of Dutch

speech recognition systems in the STEVIN N-Best project (van Leeuwen et al., 2009; van Leeuwen, 2013). Instead of considering the speaker identity only, the split in test and training was extended to produce completely disjoint conversations. This means that there are no conversations in which one side is in the training data and the other side is in the test set. The conversation sides of the 48 N-Best development test speakers were augmented with the other sides of the conversations to form the test set of this research. All conversations in which any of the 29 conversation partners occurred, were further removed from the training set. The result was a test set of 128 conversations, 256 conversation sides, and a remaining

the classifier can make, and expressed as probabilities these are P (NL | VL) and P (VL | NL). These error probabilities trade-off if the threshold is varied, and this leads to a so-called ‘Receiver Operating Curve’ (ROC). The post-hoc threshold θ= that results in the same value for these two error probabilities is know as the ‘Equal Error Rate’ E= . We will use E= as a measure for discrimination of the log likelihood ratio scores. We use the ‘ROC convex hull’ definition of

E= (de Villiers & Bru¨mmer, 2010).

To evaluate the probabilistic interpretation, we will use the empirical cross entropy (Ramos, 2007) at a prior of 0.5, which is known as Cllr , the Cost of the Log Likelihood Ratio (Bru¨mmer & du Preez, 2006) part of 1102 conversations (2204 conversation sides)

for the training set, covering 850 speakers.

The training data (77.6 hours after speech activity

de-1 Cllr = 2 log 2 ( ) i∈NL ) log(1 + e− λi )

tection) is used to train a 1024-component UBM. Next, all training utterances are used individually to obtain

+

i∈VL

log(1 + eλi )

l

(9) supervectors xi . These 2204 points are used together

with the supervised class labels yi ∈ {NL, VL}, encod-

ing ‘Nederlands’ and ‘Vlaams,’ to train an SVM with a linear kernel. We used the Julia interface to the LIBSVM implementation (Chang & Lin, 2011). Clas- sification experiments were carried out on the test set of the data (256 conversation sides). For each test ut- terance a supervector was computed using the UBM, and classified using the SVM.

No parameter tuning was carried out on the test set, and each trial from the test set is treated indepen- dently of all others, i.e., no normalization over the test set is carried out.

3.1. Duration variation

One of the most important experimental conditions is the duration of the test segment. Traditionally, in NIST Speaker Recognition Evaluations (Martin & Le, 2008), the test duration has been quantised to 3, 10, and 30 second segments. We generated multiple shorter test segments from the original 256 test con- versation sides by cutting non-overlapping fragments of the desired duration out of the feature streams. The cut durations were sampled from a Gaussian distri- bution with the standard deviation a quarter of the desired duration.

3.2. Performance metrics

LIBSVM can produce posterior probabilities (Chang & Lin, 2011) for the two classes, which, assuming a flat prior, can be converted to a log likelihood ratio score λ = log P (x | NL)/P (x | VL). A hard decision for ‘NL’ vs ‘VL’ can be obtained by applying a thresh- old θ to the score λ. There are two types of errors

Cllr is a so-called proper scoring rule (DeGroot & Fien- berg, 1983) penalising likelihood ratios that are either under- or overconfident (Bru¨mmer & du Preez, 2006; van Leeuwen & Bru¨mmer, 2007). It is zero for a per- fect classifier giving scores λ = ±∞ for the correct class, it is normalized to 1 for a classifier that gives no information λ = 0 for all trials, but may be larger than 1 for classifiers that are too confident about ei- ther class. The latter situation is known as ‘bad cal- ibration,’ and can be the result of a single trial for which |λ| is large, but of the wrong sign, i.e., strongly supporting the wrong hypothesis. Well-calibrated like- lihood ratios mean that, for a given discrimination per- formance (i.e., ROC curve), the minimum Bayes risk decision based on λ and a prior π is optimal for any π. In other words, for any cost function the actual costs are the same (and represented by the same point on the ROC) as the minimum costs that can be obtained by post-hoc optimising the threshold.

Finally, the metric ‘minimum Cllr ’, C min , represents the value of Cllr that can be obtained under optimal warping of the score axis, keeping the order of scores the same. This shows what Cllr could have been if the probabilistic scores would have been perfectly cal- ibrated, and the metric is therefore a measure of dis- crimination performance, similar to E= but sensitive to the behaviour along the entire ROC curve.

4. Results and Discussion

In Fig. 1 the so-called ‘Detection Error Trade-off ’ plot is shown (Martin et al., 1997). In the graph, the trade-off for the two types of errors P (NL | VL) and

(6)

95 dur (s) NNL NVL E= (%) Cllr C llr min 3 10 30 2151 614 178 4706 1354 390 16.1 7.93 3.35 1.11 0.680 0.251 0.518 0.251 0.083

dur (s) Cllr C min C cal a b

3 10 30 1.11 0.680 0.251 0.518 0.251 0.082 0.528 0.270 0.101 4.1 2.9 2.3 −6.4 −4.4 −3.4 pooled 0.971 0.444 0.452 3.8 −5.9 llr llr P( VL | N L) in % 0. 1 0 .5 2 5 10 20 40

Language variety detection

Table 1. Results of the language variation recognition ex-

periment. llr llr 3 s 10 s 30 s 0.1 0.5 2 5 10 20 40 P(NL | VL) in %

Figure 1. The DET plot for the language variety detection

experiment, for three different duration conditions. The horizontal axis is the fraction of VL trials classified as NL, for a given threshold θ, and at the vertical axis is the frac- tion of NL trials classified as VL, for the same threshold. The lines are the result of varying θ.

threshold θ is varied. It is similar to a Receiver Op- erating Characteristic (ROC) plot, with errors instead of hits vs misses, but the axes are warped according to the inverse of the cumulative normal distribution. If the underlying score distribution for NL and VL classes are normal, then this gives a straight line in the DET plot. The reverse is not true, since any mono- tonically increasing warping function applied to the scores does not change the position of the curve in the DET plot. For low error rates, and longer duration recordings, the curve becomes ragged due to lack of data. The most dramatic effect of this can be seen for the 30 s curve, which was obtained with only 390 trials. For a more accurate determination of the error rates, more data from more speakers is needed. The relatively smooth 3 s curve is the result of almost 7000 trials, but we must warn that these cannot be treated truly independently since many come from the same utterance.

Table 2. The calibration parameters and their effect on Cllr .

warn the reader here for a potential methodological cause of such good results. The CGN database was not designed for language variety recognition. The collec- tion of the data in The Netherlands and Flanders was carried out quite independently, and as a result there may be “channel differences” between the NL and VL parts. These channel differences are confounded with language variety differences, so it is possible that chan- nel effects help the discrimination between the vari- eties. This is an undesirable effect, and for a proper evaluation we need data from an independent collec- tion that covers both language varieties.

Yet, with all these reservations in mind, we want to conclude that language variety discrimination between NL and VL using acoustic features is possible. In a next step we need to check with newly collected data if any possible channel effect has given a too optimistic performance.

The discrimination capabilities of the SVM work fine, but for a probabilistic interpretation of the result, an additional calibration step is necessary. In Table 2 we show the results of the calibration analysis. The differ- ence between Cllr and C min is quite large, and for 3 s we have Cllr > 1, meaning that on average the recog- niser does worse in terms of expected Bayes risk than a trivial recogniser basing its decision on the prior only. In order to find out what the cause of this calibration error is, we re-calibrated the scores by transforming them as

In Table 1 the main results of the language variety re- cognition system are tabulated. The Equal Error Rate varies from about 3.5–16 % for the duration range 3– 30 sec, which is not unlike values reported in literature for language discrimination tasks (Matˇejka et al., 2005; Martin & Le, 2008), in fact, a little lower than typical language variety detection error rates. Again, we must

λcal = aλ + b, (10)

where parameters a and b are found through logistic regression. This is a form of ‘self calibration’ which is similar to determining C min , but is much more con- strained by the low number of parameters, 2.

The calibration parameters (cf. Table 2) show that the log-likelihood estimate from the SVM has a large off-

(7)

Spoken language variety recognition in Dutch

96

set and quite a substantial scaling. But this simple re- calibration does solve the mis-calibration almost com- pletely: there is not very much more to be gained in

Cllr . Interestingly, the scaling factor a > 1, whereas in speaker recognition we usually find that models compute a log-likelihood ratio that needs to be scaled down—which might be an effect of the frame indepen- dence assumptions made in the score function.

5. Conclusions

In this paper we have quite elaborately described our spoken language variety recognition system. We have shown the typical approach to such a problem taken in speech processing: starting from typical speech fea- tures, using a generative model for the feature space and a discriminative model for classifying the utter- ances. We touched upon the issue of calibration of the probabilistic output of the SVM, but for proper evalu- ation we would need a separate evaluation set. It will be the next step in our research in this direction to evaluate using an independently collected database.

6. Acknowledgements

This research has received support form the Dutch Language Union / Ministry of Arts and Culture project ‘Elftal.’

References

Bru¨mmer, N., & du Preez, J. (2006). Application- independent evaluation of speaker detection. Com-

puter Speech and Language, 20, 230–275.

Campbell, W., Sturim, D., & Reynolds, D. (2006). Support vector machines using GMM supervectors for speaker verification. IEEE Signal Processing Let-

ters, 13, 308–311.

Castaldo, F., Colibro, D., Dalmasso, E., Laface, P., & Vair, C. (2007). Acoustic language identification using fast discriminative training. Proc. Interspeech (pp. 346–349). Antwerp.

Chang, C.-C., & Lin, C.-J. (2011). LIBSVM: A library for support vector machines. ACM Transactions on

Intelligent Systems and Technology, 2, 27:1–27:27.

Software available at http://www.csie.ntu.edu. tw/~cjlin/libsvm.

de Villiers, E., & Bru¨mmer, N. (2010). The bosaris

toolkit. BOSARIS. Software available at https://

sites.google.com/site/bosaristoolkit/. DeGroot, M., & Fienberg, S. (1983). The comparison

and evaluation of forecasters. The Statistician, 12– 22.

Despres, J., Fousek, P., Gauvain, J.-L., Gay, S., Josse, Y., Lamel, L., & Messaoudi, A. (2009). Modeling

Northern and Southern varieties of Dutch for STT.

Proc. Interspeech (pp. 96–99). Brighton.

Gauvain, J. L., Lamel, L. F., Adda, G., & Adda- Decker, M. (1994). Speaker-independent continuous speech dictation. Speech Comm., 15, 21–37. Hermansky, H. (1990). Perceptual linear predictive

(PLP) analysis of speech. JASA, 87, 1738–1752. Martin, A., Doddington, G., Kamm, T., Ordowski,

M., & Przybocki, M. (1997). The DET curve in as- sessment of detection task performance. Proc. Eu-

rospeech 1997 (pp. 1895–1898). Rhodes, Greece.

Martin, A. F., & Le, A. N. (2008). NIST 2007 language recognition evaluation. Proc. Speaker and Language

Odyssey. Stellenbosch, South Afrika: IEEE.

Matˇejka, P., Schwarz, P., Cˇ ernocky´, J., & Chytil, P. (2005). Phonotactic language identification using high quality phoneme recognition. Proc. Interspeech (pp. 2237–2240). ISCA.

Oostdijk, N. H. J., & Broeder, D. (2003). The Spo- ken Dutch Corpus and its exploitation environment.

Proceedings of the 4th International Workshop on Linguistically Interpreted Corpora (LINC-03).. Bu-

dapest, Hungary.

Pelecanos, J., & Sridharan, S. (2001). Feature warp- ing for robust speaker verification. Proc. Speaker

Odyssey (pp. 213–218). Crete, Greece.

Ramos, D. (2007). Forensic evaluation of the evi-

dence using automatic speaker recognition systems.

Doctoral dissertation, Universidad Autonoma de Madrid.

Reynolds, D. A., Quatieri, T. F., & Dunn, R. B. (2000). Speaker verification using adapted gaussian mixture models. Digital Signal Processing, 10, 19– 41.

Singer, E., Torres-Carrasquillo, P. A., Gleason, T. P., Campbell, W. M., & Reynolds, R. A. (2003). Acous- tic, phonetic, and discriminative approaches to au- tomatic language identification. Proc. Eurospeech (pp. 1345–1349).

Torres-Carrasquillo, P. A., Singer, E., Kohler, M. A., Greene, R. J., Reynolds, D. A., & Deller Jr, J. R. (2002). Approaches to language identification using gaussian mixture models and shifted delta cepstral features. ICSLP.

van Leeuwen, D. A. (2013). Essential speech and lan-

guage technology for Dutch, chapter N-Best 2008: a

benchmark evaluation for large vocabulary speech recognition in Dutch, 271–288. Springer.

van Leeuwen, D. A., & Bru¨mmer, N. (2007). An in- troduction to application-independent evaluation of speaker recognition systems. In C. Mu¨ller (Ed.),

(8)

97

Speaker classification, vol. 4343 of Lecture Notes in Computer Science / Artificial Intelligence. Springer.

van Leeuwen, D. A., & Bru¨mmer, N. (2008). Building language detectors using small amounts of training data. Proc. Speaker and Language Odyssey. Stellen- bosch, South Afrika: IEEE.

van Leeuwen, D. A., Kessens, J., Sanders, E., &

van den Heuvel, H. (2009). Results of the N-Best 2008 Dutch speech recognition evaluation. Proc. In-

terspeech (pp. 2571–2574). Brighton: ISCA.

Zissman, M. A. (1996). Comparison of four approaches to automatic language identification of telephone speech. IEEE Trans. Speech and Audio Proc., SAP-

References

Related documents

gies that have a positive effect (positive beta value) on firm’s performance include: long-range supply chain strategies where the firm continuously plans. its supply

Arguably, a systems perspective is appropriate for studying the ways in which rural electrification (RE) processes change how energy supply and use are embedded

Methods: A hospital-based cross-sectional study was conducted in Hiwot Fana Specialized University Hospital on assessment of the prescribing, dispensing, and patient use pattern of

In this method, the ratings of the alternatives on attributes in individual decision matrices given by each expert are in the form of interval-valued intuitionistic fuzzy sets

When Concept Mapping is compared with other sorts of activities, such as outlining or defining concepts, that also can induce the learner to take a thoughtful, systematic approach

Argumentet for ikke å ta mer hensyn til realøkonomien var at Norges Bank ikke ønsket å være ledende når de satte opp rentene i frykt for at dette skulle føre til en styrking

The size of the informal sector has two opposing effects on the welfare change induced by the tax reform: (i) the consumption tax distortion increases (yielding a negative effect);

If you choose this section, read the following passages carefully, and answer Questions 1–4.. (Note: there are two options in