Speech recognition in Linux is generally considered tough, but it can be done with very good results. Julian is a special version of Julius which performs grammar based speech recognition.  The video shows the the speech recognition in action.

.

As we can see, there occurs no error in the recognition of sentences.

All Speech Recognition Engines (“SRE“s) are made up of the following components:

  • Language Model or Grammar – Language Models contain a very large list of words and their probability of occurrence in a given sequence.  They are used in dictation applications.  Grammars are a much smaller file containing sets of predefined combinations of words.  Grammars are used in IVR or desktop Command and Control applications.   Each word in a Language Model or Grammar has an associated list of phonemes (which correspond to the distinct sounds that make up a word).
  • Acoustic Model – Contains a statistical representation of the distinct sounds that make up each word in the Language Model or Grammar.  Each distinct sound corresponds to a phoneme.
  • Decoder – Software program that takes the sounds spoken by a user and searches the Acoustic Model for the equivalent sounds.  When a match is made, the Decoder determines the phoneme corresponding to the sound.  It keeps track of the matching phonemes until it reaches a pause in the users speech.  It then searches the Language Model or Grammar file for the equivalent series of phonemes.  If a match is made it returns the text of the corresponding word or phrase to the calling program.

Although Julian (a special version of Julius which performs grammar-based speech recognition) uses Acoustic Models created with theHTK toolkit, Julian uses its own Grammar definition format.

Grammar

A recognition Grammar essentially defines constraints on what the SRE can expect as input.  It is a list of words and/or phrases that the SRE listens for.  When one of these predefined words or phrases is heard, the SRE returns the word or phrase to the calling program – usually a Dialog Manager (but could also be a script written in Perl, Python, etc.).  The Dialog Manager then does some processing based on this word or phrase.

The example video shown above  that of a voice-operated interface to for a robot control.  If the SRE hears the sequence of words: ‘Chippu Move Forward’, it returns the textual representation of this phrase to the Dialog Manager, which then produces the control signals to turn the motors.

It is very important to understand that the words that you can use in your Grammar are limited to the words that you have ‘trained’ in your Acoustic Model.  The two are tied very closely together.

Acoustic Model

An Acoutic Model is a file that contains a statistical representation of each distinct sound that makes up a spoken word.  It must contain the sounds for each word used in your grammar.  The words in your grammar give the SRE the sequence of sounds it must listen for.  The SRE then listens for the sequence of sounds that make up a particular word, and when it finds a particular sequence, returns the textual representation of the word to the calling program (usually a Dialog Manager).  Thus, when an SRE is listening for words, it is actually listening for the sequence of sounds that make up one of the words you defined in your Grammar.  The Grammar and the Acoustic Model work together.

Therefore, when you train your Acoustic Model to recognize the phrase ‘CHIPPU MOVE FORWARD’, the SRE is actually listening for the phoneme sequence “ch” “iy” “p” “ax” “m” “uw” “v”  “f” “ao” “r” “w” “er” and “d”.  If you say each of these phonemes aloud in sequence, it will give you an idea of what the SRE is looking for.

Commercial SREs use large databases of speech audio to create their Acoustic Models.  Because of this, most common words that might be used in a Grammar are already included in their Acoustic Model.

When creating your own Acoustic Models and Grammars, you need to make sure that all the phonemes that make up the words in your Grammar are included in your Acoustic Model.

Background – Julian Grammars

In Julian, a recognition grammar is separated into two files:

  • the “.grammar” file which defines a set of rules governing the words the SRE is expected to recognize;  rather than listing out each word in the .grammar file, a Julian grammar file uses “Word Categories” – which is the name for a list of words to be recognized (which are defined in a separate “.voca” file);
  • the “.voca” file which defines the actual “Word Candidates” in each Word Category and their pronunciation information (Note: the phonemes that make up this pronunciation information must be the same as will be used to train your Acoustic Model).

.grammar file

The rules governing the allowed words are defined in the .grammar file using a modified BNF format.  A .grammar specification in Julian uses a set of derivation rules, written as:

Symbol: [expression with Symbols]

where:

  • Symbol is a nonterminal; and
  • [expression with Symbols] is an expression which consists of sequences of Symbols, which can be terminals and/ornonterminals.

terminal is BNF jargon for a symbol that represents a constant value.  It never appears to the left of the colon.  In Julian terminals represent Word Categories – lists of words that are further defined in a separate “.voca” file.

nonterminal is BNF jargon for a symbol that can be expressed in terms of other symbols.  It can be replaced as a result of substitution rules.

For example, look at the the following derivation rules:

S : NS_B MOVE NS_E
MOVE: NAME ACTION DIRECTION

In this example, “S” is the initial sentence symbol.   NS_B and NS_E correspond to the silence that occurs just before the utterance you want to recognize and after.   “S”, “NS_B” and “NS_E” are required in all Julian grammars.

“NS_B”, “NS_E”, “NAME”,”ACTION” and “DIRECTION” are terminals, and represent Word Categories that must be defined in the “.voca” file.  In the “.voca” file,”ACTION” corresponds to two words: “MOVE” and “LOOK” and their pronunciations.  “NAME” corresponds to the word “CHIPPU” . “DIRECTION” corresponds to four words: “LEFT”, “RIGHT”, “FORWARD”, “BACKWARDS” and their pronunciations.

“MOVE” is a nonterminal, and does not have any definition in the .voca file.  It does have a further definition in the .grammar file, where it is replaced by the expression “NAME ACTION DIRECTION”.  All nonterminals must be further defined in the .grammar file until they are finally represented by terminals (which are then defined in the .voca file as Word Categories).

With Julian, only one Substitution Rule per line is permitted, with the colon “:” as the separator.   Alphanumeric ASCII characters and the underscore are permitted for Symbol names, and these are case sensitive

.voca file

The “.voca” file contains Word Definitions for each Word Category defined in the .grammar file.

Each Word Category must be defined with “%” preceding it.  Word Definitions in each Word Category are then defined one per line. The first column is the string which will be output when recognized, and the rest is the pronunciation.  Spaces and/or tabs can act field separators.

Format:

%[Word Category]
[Word Definition]   [pronunciation ...]
...

For example the Word Categories “NS_B”, “NS_E”,  “NAME” ,”ACTION” and “DIRECTION” were referenced in the “.grammar” file above and are defined in a “.voca” as follows:

% NS_B
<s>                           sil

% NS_E
</s>                         sil

% NAME

CHIPPU                    ch iy p ax

% ACTION

MOVE                          m uv w

LOOK                           l uw k

% DIRECTION

FORWARD                 f ao r w er d

BACKWARDS            b ae k w er d z

LEFT                             l eh f t

RIGHT                          r ay t

In the above example, the NS_B and NS_E Word Categories each have one Word Definition with a silence model (‘sil’ is a special silence model defined in your Acoustic Model).  These correspond to the head and tail silence in speech input.

“ACTION” is broken out into two words “MOVE” and “LOOK” with pronunciation information, which are the phonemes that make up the words to be recognized (and which correspond to phonemes that will be included in your Acoustic Model).   “DIRECTION” is broken out into four words: “FORWARD” and “BACKWARDS”, “LEFT”, “RIGHT” and their phonemes

If you have words with different pronunciations, simply create the additional entries on separate lines for the same word but with the different pronunciation.

The .grammar and .voca files working together

Julian needs a predefined word lattice file where each word and each word-to-word transition is listed explicitly.  We get this by compiling the “.grammar” and “.voca” files together to generate the word lattice file (actually it is two files, but more on that later) with a script.  The mkdfa.pl script does this by looking for the Initial Sentence Symbol “S” in the .grammar file and replacing the Word Categories with all the possible Word Candidates from the .voca file, and making a predefined list of all the possible combinations of words and phrases Julian must recognize.  In this case, the list of all possible sentences would be:

<s> CHIPPU MOVE FORWARD </s>

<s> CHIPPU MOVE BACKWARDS</s>

<s> CHIPPU MOVE RIGHT</s>

<s> CHIPPU MOVE LEFT</s>

<s> CHIPPU LOOK FORWARD</s>

<s> CHIPPU LOOK BACKWARDS</s>

<s>CHIPPU LOOK LEFT</s>

<s>CHIPPU LOOK RIGHT</s>

Compiling your Grammar

The .grammar and .voca files now need to be compiled into “.dfa”  and “.dict” files so that Julian can use them.  This is done using the Julian “mkdfa.pl” grammar compiler. The .grammar and .voca files need to have the same file prefix, and this prefix is then specified to the mkdfa.pl script.   Compile your files (sample.grammar and sample.voca) as follows:

mkdfa.pl sample
The generated sample.dfa and  sample.term   files contain finite automaton information, and the sample.dict file contains word dictionary information.  All are in Julian format.
Now run the Julius to see the speech recognition in action
julius -quiet -input mic -C julian.jconf 2>/dev/null