NAME
Speech::Recognizer::SPX - Perl extension for the PocketSphinx speech recognizer
SYNOPSIS
use Speech::Recognizer::SPX qw(:fbs :uttproc)
fbs_init([arg1 => $val, arg2 => $val, ...]);
uttproc_begin_utt();
uttproc_end_utt();
fbs_end();
DESCRIPTION
This module provides a Perl interface to the PocketSphinx speech recognizer library.
USING THIS MODULE
use Speech::Recognizer::SPX qw(:fbs :uttproc :lm);
Because most parts of the PocketSphinx library contain a lot of global internal state, it makes no sense to use an object-oriented interface at this time. However I don't want to clobber your namespace with a billion functions you may or may not use. To make things easier on your typing hands, the available functions have been grouped in to tags representing modules inside the library itself. These tags and the functions they import are listed below.
:fbs
-
This is somewhat of a misnomer - FBS stands for Fast Beam Search, but in actual fact this module (the fbs_main.c file in PocketSphinx) just wraps around the other modules in sphinx (one of which actually does fast beam search :-) and initializes the recognizer for you. Functions imported by this tag are:
fbs_init fbs_end
:uttproc
-
This is the utterance processing module. You feed it data (either raw audio data or feature data - which currently means vectors of mel-frequency cepstral coefficients), and it feeds back search hypotheses based on a language model. Functions imported by this tag are:
uttfile_open uttproc_begin_utt uttproc_rawdata uttproc_cepdata uttproc_end_utt uttproc_abort_utt uttproc_stop_utt uttproc_restart_utt uttproc_result uttproc_result_seg uttproc_partial_result uttproc_partial_result_seg uttproc_get_uttid uttproc_set_auto_uttid_prefix uttproc_set_lm uttproc_lmupdate uttproc_set_context uttproc_set_rawlogdir uttproc_set_mfclogdir uttproc_set_logfile search_get_alt
:lm
-
This is the language model module. It loads and unloads language models.
lm_read lm_delete
STARTUP
fbs_init(\@args);
The fbs_init
function is the main entry point to the Sphinx library. If given no arguments, it will snarf options from the global @ARGV
array (because that's what its C equivalent does). To make life easier, and to entice people to write Sphinx programs in Perl instead of C, we also give you a way around this by allowing you to also pass a reference to an array whose contents are arranged in the same way @ARGV
might be, i.e. a list of option/value pairs.
To make things pretty, you can use the magical => operator, like this:
fbs_init([samp => 16000,
datadir => '/foo/bar/baz']);
Note that you can omit the leading dash from argument names (if you like).
Calling this function will block your process for a long time and print unbelievable amounts of debugging gunk to STDOUT and STDERR. This will get better eventually.
This function has a large number of options. Someday they will be documented. Until then, either look in the example code, or go straight to the source, namely the file include/cmdln_macro.h
.
FEED ME
uttproc_begin_utt() or die;
uttproc_rawdata($buf [, $block]) or die;
uttproc_cepdata(\@cepvecs [, $block]) or die;
uttproc_end_utt() or die;
To actually recognize some speech data, you use the functions exported by the :uttproc
tag. Before calling any of them, you must successfully call uttproc_begin_utt
, or Bad Things are certain to happen (I can't speculate on exactly what things, but I'm sure they're bad).
You should call uttproc_begin_utt
before each distinct utterance (to the extent that you can predict when individual utterances begin or end, of course...), and uttproc_end_utt
at the end of each.
After calling uttproc_begin_utt
, you can pass either raw audio data or cepstral feature vectors (see Audio::MFCC), using uttproc_rawdata
or uttproc_cepdata
, respectively. Due to the way feature extraction works, you cannot mix the two types of data within the same utterance.
If live mode is in effect (i.e. -livemode =
TRUE> was passed to fbs_init
), the optional $block
parameter controls whether these functions will return immediately after processing a single frame of data, or whether they will process all pending frames of data. If you need partial results, you probably want to pass a non-zero value (FIXME: should be a true value but I don't know how to test for truth in XS code) for $block
, though this may increase latency elsewhere in the system.
Unfortunately, it appears that there is no specific function to flush all unprocessed frames before getting a partial result. Calling uttproc_rawdata with an empty $buf
and $block
non-zero seems to have the desired effect.
GETTING RESULTS
my ($frames, $hypothesis) = uttproc_result($block);
my ($frames, $hypothesis) = uttproc_partial_result();
my ($frames, $hypseg) = uttproc_result_seg($block);
my ($frames, $hypseg) = uttproc_partial_result_seg();
my $hypothesis = $hypseg->sent;
my $segs = $hypseg->segs;
my @nbest = search_get_alt($n); # Must call uttproc_result first!
foreach my $nhyp (@nbest) {
my $nsent = $nhyp->sent;
my $nsegs = $nhyp->segs;
print "Hypothesis: $nsent\n";
foreach my $seg (@$nsegs) {
printf " Start frame %d end frame %d word %s\n",
$seg->sf, $seg->ef, $seg->word;
}
}
At any point during utterance processing, you may call uttproc_partial_result
to obtain the current "best guess". Note that this function does not flush unprocessed frames, so you might want to use the trick mentioned above to do so before calling it if you are operating in non-blocking mode.
By contrast, you may not call uttproc_result
until after you have called uttproc_end_utt
(or uttproc_abort_utt
or also possibly uttproc_stop_utt
). The $block
flag is also optional here, but I strongly suggest you use it.
The functions uttproc_result_seg
and uttproc_partial_result_seg
functions work similarly except that instead of returning a string, they return a Speech::Recognizer::SPX::Hypothesis
object which contains probability and word segmentation information. You can access its fields with the following accessor functions:
sent
senscale
ascr
lscr
segs
The sent
field contains the string representation of the hypothesis, and is equivalent to the string returned by uttproc_result
. The senscale
, ascr
, and lscr
fields are currently unimplemented.
The segs
field contains a reference to an array of Speech::Recognizer::SPX::Segment
objects. Each of these objects contains fields which can be accessed with the following accessors:
word
sf
ef
ascr
lscr
conf
latden
phone_perp
fsg_state_to
fsg_state_from
The word
field contains the string representation of the word. The sf
and ef
fields contain the start and end frames for this word. The ascr
and lscr
fields contain the acoustic and language model scores for the word. The fsg_state_from
and fsg_state_to
fields indicate the finite-state grammar states in which this entry starts and terminates, if a finite-state grammar is used. The latden
field contains the average lattice density for this word, while the phone_perp
contains the average phoneme perplexity. The conf
field contains a confidence score which is an estimated probability that this word was recognized correctly.
You can also obtain an N-best list of hypotheses using the search_get_alt
function. This function returns a list of the number of hypotheses requested, or as many as can be found. Each element in this list is a Speech::Recognizer::SPX::Hypothesis
object like the above, except that the acoustic/language model score is not filled in.
FIDDLY BITS
Changing language models, etc, etc... This documentation is under construction.
EXAMPLES
For now there are just some example programs in the distribution.
AUTHOR
David Huggins-Daines <dhuggins@cs.cmu.edu>. Support for N-best hypotheses was funded by SingleTouch Interactive, Inc (http://www.singletouch.net/).
SEE ALSO
perl(1), Speech::Recognizer::SPX::Server, Audio::SPX, Audio::MFCC