# Evaluating an ASR in a Spoken Dialogue System

Swaraj Dalmia Jan 21, 2022 · 9 mins read

An ASR (automatic speech recognition) is an integral component of any voice bot. The most popular metric that is used to evaluate the accuracy of an ASR model is WER or the word error rate. In this blog, we discuss metrics that can be used to evaluate an ASR, their flaws and suggestions for improvement in the context of a conversational agent.

The ASR module takes as input a spoken utterance and outputs the most likely transcription. Most ASR’s output multiple alternatives with certain confidence scores. Some ASR’s including Kaldi’s implementations output N-Best alternatives in an order not necessarily reflective of the confidence. Some ASR systems output likelihood information of word sequences in the form of word-lattices or confusion networks [1] with probability information.

## What is WER ?

Word Error Rate measures the transcription errors, treating words as the smallest unit. It takes 2 inputs, the actual transcript and a hypothesis transcript.

There are two types of WER:

• Isolated word recognition (IWR-WER)
• Connected Speech Recognition (CSR-WER)

Isolated Word Recognition WER

This considers the words in isolation and is not based on any alignment. It simply measures the number of non-hits.

$IWR-WER = 1-\frac{H}{N}$

where,
H = number of hits
N = total matched I/O words

Connected Speech Recognition WER

This is calculated on the basis of alignment and uses the Levenshtein distance for words which measures the minimum edit distance. This is efficiently calculated using dynamic programming. It calculates the best alignment which minimises the number of substitutions, insertions and deletions necessary to map the actual transcript to the hypothesis transcript, giving equal weight to all the operations. WER is not input/output symmetric, as N is the total number of words in the actual transcripts.

$CSR-WER = (I+S+D)/(N)$

where, I = insertions
S = substitutions
D = deletions
N = total number of words in the actual transcript

An example calculation of the best alignment to calculate the CSR WER is shown below.

The CSR WER for the above hypothesis is = (6+3+1)/13 = 0.77

The upper bound of CSR WER is not 1, but $$\frac{max(N1, N2)}{N1}$$ where N1 is length of true transcripts and N2 is length of hypothesis transcript.

## Statistical Significance of WER ?

Improvements in WER’s over a test set is one of the standard ways of evaluating upgrades in an ASR. Often, what is missed out is in this evaluation is whether the gain is statistically significant or not. One of the standard statistical tests is the Matched-Pair Sentence Segment Word Error (MAPSSWE) test, introduced in Gillick and Cox (1989) [2]. For an example on how to calculate the statistic, consult the book by Jurafsky.

## Issues with WER

• It gives the same importance to words like “a”, “the” compared to verbs and nouns that carry more semantic value.
• WER is a purely 1 : 1 transcript based metric, and doesn’t take into account the rich output of ASR systems like alternatives or word lattices.
• The concept of edit-distance that WER is based on is appropriate for a dictation machine where the additional cost is that of correcting the transcripts rather than applications where communicating the meaning is of primary importance
• It doesn’t take into account the performance of an ASR in the context of the dialogue pipeline. For example if a higher WER makes no difference in the information retrieval for the downstream SLU then is the improvement worth it ?
• An example of a context dependent metric is discussed in [4].
• It is not a true % based metric, because it has no upper bound therefore it doesn’t tell you how good a system is, but only that one is better than another. Even for the later, it provides only a heuristic for ranking of performance.
• Consider two ASR systems, ASR-1 that replaces 2 wrong words for every word it listens too, and another ASR system that replaces 1 wrong word for every word it listens to. Both communicate zero information, but the WER for ASR-1 is 2 and the WER for ASR-1 is 1. This 50% difference in WER is not reflective of performance, since both the systems communicate no correct information whatsoever.

## Debunking Conventional Wisdom

One might assume that better ASR’s improve the performance of all downstream SLU systems. A paper in 2003 [7] found this was not always the case. Their model had a 17% better slot accuracy despite a 46% worse WER performance. They have this gain by using a SLU model as the language model for speech recognition. Therefore it is not necessary that an oracular ASR will solve downstream inaccuracies. It is important to be aware that there might be non-linear correlations in metrics for the ASR vs the downstream task.

## WER Variants:

A few WER variants are discussed below.

Metrics that look at a different granularity than that of words : **

• Sentence Error Rate (SER) : The percentage of sentences with at least one word error.
• Character Error Rate (CER) : Similar as WER with the smallest units as characters and not words

The next two error rates are variations of WER, are are bounded between [0,1].

• Match Error Rate (MER) [3] : Measures the probability of a given match being incorrect.
$MER = \frac{I+S+D}{I+S+D+H} = 1 - \frac{H}{I+S+D+H}$

where,
H = number of hits N = total number of words in the actual transcript

MER is always <= WER.

• Word Information Lost (WIL) [3] : Information theoretic measure based on entropy. For more details look at the paper.

Example values, comparing WER, MER and WIL are shown below:

Note : For implementations of WER, WIL, and MER, have a look at Jiwer.

Apart from looking at just single metrics for evaluating an ASR there are few additional metrics that one should use to evaluate a goal oriented ASR deployed for a given application:

• Which speaker demographic is most often misrecognised ?
• What (context-dependent) phones are least well recognised ?
• Which words are most confused ? Generate a confusion matrix of confused words.

These often help prioritise improvements in terms of what might benefit the goal the most.

# Semantics based Metrics

WER doesn’t look at the meaning of what has been transcribed despite the fact that is the semantics that are most relevant in an ASR that is a part of a spoken dialogue system.

## Concept accuracy

It is a simple metric that looks at the accuracy of the concepts that are of relevance in the transcripts.

Example:
Reference - I want to go from Boston to Baltimore on September 29 Hypothesis - Go from Boston to Baltimore on December 29

The WER is $$45\%$$. However if one looks at concept accuracy, it is 2/3. Out of the 3 concepts : “Boston”, “Baltimore” and “September 29” it gets 2 of them right.

## WER with embeddings

In this paper [9] they look at augmenting the WER metric for an ASR that is used for a Spoken Language Translation (SLT) task.

They argue that some morphological operations, like adding a plural doesn’t impact the translation task and that such substitution errors should be penalised differently. To decide which ones to penalise they use word embeddings. They call their new metric, WER-E i.e. WER with embeddings. The only change in this metric is that the substitution cost in WER is replaced by the cosine distance between the two words, so near identical words get assigned a very low cost.

## Other Metrics

There are lots of different papers that augment the WER metric or introduce a new metric specific to a downstream task. Mentioning a few of them below :

• In [10] a new measure called Automatic Transcription Evaluation for Named Entity (ATENE) is introduced for the NER downstream task
• 3 new evaluation metrics are introduced in [12], where the downstream application is Information Retrieval
• SemDist metric is introduced in [13] for downstream SLU

Downstream applications aside, it is quite pertinent to evaluate and benchmarks ASR across different social and demographic groups to evaluate bias and fairness. Systems don’t exist in isolation from the society in which they are deployed in and it is important that ML Engineers pay head to such metrics while deploying ASRs.

## References :

[5] : http://www.cs.columbia.edu/~julia/courses/CS4706/asreval.pdf

[12] : Evaluating ASR Output for Information Retrieval (2007)