AdityaYadavalli1 / Kaldi-on-ADA

My documentation of which script does what in Kaldi

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

[this is what I have understood over the years independently. So there might be some factual errors here. If found please report]

Kaldi on ADA

[SAT 20th Jun '20]

Understanding the recipe

I will be going through run.sh. The top level script that calls other scripts.

  • Before Stage 0: The script local/download_and_untar.sh downloads the data you want and unzips it. First argument is the should tell the program where the user wants to download the data, second is url and the third is corpus part. This will any way be shown if you provide incorrect arguments.

  • Stage 0: This downloads 3 language models (why three? I don't know). First argument to the script called (local/download_lm.sh) is the base url and the second argument is the where those models are stored (folder).

    [THURS 25th Jun '20]

  • Stage 1: In this stage, 5 scripts are called:

    • local/data_prep.sh: Creates all the important files (wav.scp, transcript, utt2spk, spk2gender, utt2dur) given that the data has SPEAKERS.txt file. This takes source directory as the first argument and destination directory as the second argument.
    • local/prepare_dict.sh: Creates all the necessary steps for g2p. If lexicon_nosil.txt is not available, using CMUdict g2p (or/and Sequitur G2P*) lexicon_nosil.txt is created. With this, all g2p files are created (lexicon.txt, silence.txt, nonsilence.txt, optional.txt, extraquestions.txt) . Arguments need:
    • 1st argument: LM directory (in our case data/local/lm)
    • 2nd argument: G2P model directory (it was not used in our case because we started from stage 3)
    • 3rd argument: destination directory (where all the files made are stored)
    • utils/prepare_lang.sh:
      • 1st argument: Source directory (the destination directory for local/prepare_dict.sh)
      • 2nd argument: OOV word representation (eg: "UNK")
      • 3rd argument: Temporary directory (this script needs a space to store some temporary files)
      • 4th argument: Destination directory

    With these, essentially, it will make L.fst (lexicon in fst format. The format is used by kaldi to calculate scores) [all of this is usually stored in data/lang]

    • local/format_lms.sh:

      • optional argument: utils/prepare_lang.sh destination directory should be source directory for this file (by default it is data/lang)
      • 2nd argument: LM directory (data/local/lm in this case)

      This script converts arpa formatted language model to G.fst (grammer in fst format). Separate fst models (L and G) are avaiable in ${src_dir}_test_${lm_suffix} where lm suffix could something of the form of tgsmall or tgmed or tglarge.

    • utils/build_const_arpa_lm.sh: Converts arpa formatted language model to const arpa format. Arpa is the format a language model is in and constant arpa is the format that is needed for decoding (in run.sh this is run only on tglarge. We can run the previous script on tglarge too but that'll be too time consuming)

    [FRI 26th Jun '20]

  • Stage 2: In stage 2 we extract mfcc features from wav files. For this, 3 scripts are called:

    • steps/make_mfcc.sh:

      • 1st argument: Directory where the data is stored. (training)
      • 2nd argument: Where you want want the log files to be written. (optional argument)
      • 3rd argument: Where you want the extracted mfcc feats to be written. (optional argument)

      This extracts mfcc features from the wav files.

    • steps/compute_cmvn_stats.sh:

      • 1st argument: Training data directory
      • 2nd argument: Log directory
      • 3rd argument: Result files' directory Compute cepstral mean and variance statistics per speaker. (probably for speaker normalisation). MFCC features are stored in *.ark files and the absolute filepaths are stored in *.scp inside the 3rd argument (directory). For a more in depth explanation (I do not need it and for lack better documentation skills) you can refer this
    • utils/subset_data_dir.sh: A nice utility script. Given a data file with a specified number of utterances (a subset of utterances), it outputs a subset of utterances from the data file. You can shortest, last or first n utterances from the file by specifying it with the corresponding argument.

      • 1st argument: Option ([shortest, last, first n] utterances)
      • 2nd argument: Data
      • 3rd argument: Resultant data
  • Stage 3: In stage 3, we train a monophone system. For this, 3 scripts are called: [SUN 5th Jul'20]

    • steps/train_mono.sh:

      • 1st argument: Training data directory

      • 2nd argument: lang directory (eg: data/lang_nosp or data/lang)

      • 3rd argument: A place where the model can be stored (exp/mono). Apparently this can take a few gigabytes. Scripts it calls:

        • gmm-init-mono with training features extracted in the previous stage. This creates a phonetic decision tree with only the root because all GMMs are initialised to the same value. One can see the topology of HMMs in data/lang/topo.
        • compile-train-graphs which generates a training graph - one FST per training utterance. These FSTs encode HMM structure for that training utterance. The FSTs' input-symbols are transitions-ids which includes pdf-ids (which represents GMM acoustic state) [transition ids essentially encode audio frame] and output-symbols are words. This FST also includes cost (mostly including the cost coming from lexicon i.e pronunciation probability) but the transition probability of the HMM model will only be added later during training (word-word or phone-phone cost is added later)
        • align-equal-compiled this performs the most naïve yet the best the guess when we don't know anything - equally spaced alignments. That is this assumes that all HMM states are equally spaced.
        • gmm-est force aligns using Viterbi Training. Using this, GMMs are re-estimated. Inside the training loop the following scripts are called: [basically EM algorithm is applied. Baum-Welch in a sense]
        • gmm-align-compiled aligns phone states according to the GMM models.
        • gmm-acc-stats-ali accumulate stats for GMM training.
        • gmm-est performs Maximum Likelihood to re-estimate the GMM-based acoustic models. (This time with different options)

        NOTE: In this script we can change the beam search size. It is a hyper-parameter, change this from data to data to get optimal results.
        NOTE1: For any graph, the output symbol is words (specifically word.txt) and inputs are transition-ids (arcs in CD HMMs) (as opposed to pdf-ids which represent GMM states)

    • utils/mkgraph.sh: This script creates a fully expanded decoding graph (HoCoLoG) that represents all the language-model, pronunciation dictionary (lexicon), context-dependency, and HMM structure in our model. The output is a Finite State Transducer that has word-ids on the output, and pdf-ids on the input (these are indexes that resolve to Gaussian Mixture Models).

      • 1st argument: lang directory
      • 2nd argument: directory where the model is stored (exp/mono)
      • 3rd argument: directory where the decoded graph can be stored (exp/mono/graph_nosp) [graph generation]
    • steps/decode.sh: This script finally decodes the graph compiled using mkgraph.sh i.e we generate lattices (a graph-based record of the most likely utterances) using scores made by local/score.sh

    Note: “decoding” refers to the computation where we find the best sentence given the model.

    • steps/align_si.sh: It combines all the alignments learnt in different passes/epochs by gmm-est and gmm-align-compiled. (audio frames with monophones) [check this once]
  • Stage 4

    • steps/train_deltas.sh: This takes number of GMMs for the model to use, maximum number of Gaussians each GMM should use, monophone alignments, training data directory and the language directory (one that contains phones, questions non-silent phones etc). This script calls:

      • acc-tree-stats.cc: This program accumulates the phonetic-context tree based on previously made alignments and returns it. (The context width and centre phone can be given as arguments) [at this stage it is made sure that delta features extracted are taken into consideration]
      • sum-tree-stats.cc: This program summarises (or is it just sum?) all the stats accumulated (for example if you run acc-tree-stats.cc on many processors simultaneously those many files are created. All these files are then combined)
      • cluster-phones.cc: This program takes the statistics found in sum-tree-stats.cc and the phones as input and finds out the questions to be asked to make a split in the phonetic-decision tree (or) cluster phones. Here, we cluster based on acoustic similarity of the phones.
      • compile-questions.cc: (Not able to understand what this program this?). This program, taking the output of cluster-phones.cc and HMM topology maps these questions to HMM states.
      • build-tree.cc: This finally builds the tree phonetic tree using the statistics and questions found/formed previously. We can configure the tree to have multiple roots (3 for triphones typically)
      • gmm-init-model: Initialize a GMM acoustic model (e.g. 1.mdl) from a decision tree (e.g. tree), accumulated tree stats (e.g. treeacc), and an HMM model topology
      • gmm-mixup.cc: This program splits existing GMMs to have more gaussian components to capture finer detail. It does this by taking in transition-id statistics (occupation counts 1.occ) and the current GMMs.
      • convert-ali.cc: Given we have new information (triphone phonetic decision tree, new GMM model made by gmm-mixup.cc) and the old information (monophone alignments), this program uses both to come up with new set of alignments (triphone alignments)
      • compile-train-graphs.cc: Explained in detail in stage 3 (monophone modelling). Here it is used in triphone context.
        Then inside the training loop again an EM algorithm is run (similar to monphone modelling). The following scripts are called: gmm-align-compiled, gmm-acc-stats-ali, gmm-est. All three scripts are explained in detail in monophone modelling.
    • utils/mkgraph.sh: Explained in detail in stage 3

    • steps/decode.sh: Explained in detail in stage 3

    • steps/lmrescore.sh: At this stage the lattice (which essentially contains a graph that encodes n-best hypothesis) is rescored. Here, we decouple acoustic and language modelling scores on lattice, rescore acoustic based on the new model. We also use a more complex language model (since the search space is decreased now) -possibly trained on more data or higher n-gram model. Still have a few doubts here:

      1. I am still wondering what phi matcher is in mode 4? Ans: adds backoff arcs. Essentially the new lm that it takes is the backoff model.
      2. Why is lattice scale for newlm also -1 (from mode 2 to mode 4)? Ans: I think when removing the old lm we scale the acoustic and language model to -1. Now when we pipe this lattice to a later stage this scaling for new lm should be done relative to old lm. Thats why new lm scaling is -1.
    • steps/lmrescore_const_arpa.sh: This also does something similar to lmrescore.sh but on a larger scale due to the way the language model is formatted (arpa)

      Note: After lmrescore.sh and lmrescore_const_arpa.sh score.sh is called. This script essentially find the WER (or CER) whichever is specified for that the model currently created.

    • steps/align_si.sh: Explained in stage 3. Here we align audio frames with triphones (context dependent phones)

[SAT 25th Jul '20]

  • Stage 5

    • steps/train_lda_mllt.sh: This script splices across frames, runs LDA to reduce the dimension of the features to 40. MLLT estimates a linear transform that when applied to GMM's means adapts to a specific speaker. Here, multiple speaker classes can use the same MLL transform. We create these classes using either a decision tree (mostly) or clustering approach.
    • Rest of the scripts run at this stage are explained in detail in stage 4 or 3
  • Stage 6

    • steps/train_sat: Here, the model is retrained on fMLLR transformed features. fMLLR just means instead of transforming at the acoustic model level like MLLR you transform at the feature level to adapt the model to a specific speaker (or a set of speakers). This script can be run after just computing delta+delta delta features as well.
    • Rest of the scripts run at this stage are explained in detail in stage 4 or 3

[SUN 26th Jul '20]

  • Stage 7
    Since, generating lexicon directly from words doesn't silence in the lexicon we try to rescore the lexicon using silence here.

    • steps/get_prons.sh: The script outputs pronunciation i.e phones for each utterance using the latest acoustic and language model (from the latest alignments). There are a few files generated here for further use. They are:
      • prons.*.gz: The script takes in the latest alignments (ali*.gz) turns it into a nbest lattice, adds timestamps to that lattice i.e begin_frame, num_frame (duration), and then convert that nbest lattice to a format in which every word in every utterance has a line for itself. So, the final format in which prons.*.gz stores data is
      <utt_id> <begin_frame> <num_frame> <word> <phone1> <phone2> ... <phoneN>
      
      Note: From my understanding, we use nbest lattice but in reality it only stores one word graph with only one path but for the lack of specific 1best command we use nbest command to generate that lattice. Also, these phones are represented using unique numbers at this stage.
      • pron_counts.int: This stores the number of times a particular sequence of phones occur. Format:
          <number_of_times_seq_of_phones_occur> <word> <phone1> <phone2> ... <phoneN>
        
      • pron_counts.txt: This is same as pron_counts.int but phone itself is replaced by the number that represents it.
      • pron_counts_nowb.txt: Same as the previous file but phones don't have B,I,S,E tags.
      • pron_perutt_nowb.txt: In this file, each utterance has one line instead of each word having one line. Starting of the sentence is represented using <s> and ending using </s>. Apart from that each word (including silence using ) is present along with its phone representation after the word. Format:
        <utt_id> <s> <word1> <phone_rep> <word2> <phone_rep> ... <wordN> <phone_rep> </s>
      
      • pron_bigram_counts_nowb.txt: Bigram counts of the words is calculated here. Its format is:
        <bigram_count> <word> <word's_phone_representation>
      
      Note: Here, one should note that the silence is not considered when counting the bigrams.
      • sil_counts_nowb.txt: At this stage, we collect silence/non-silence counts before and after a word and store those counts [silence bigram essentially]. Each word has one line (including <s> and </s>). Format:
        <sil-before-count> <nonsil-before-count> <sil-after-count> <nonsil-after-count> <word> <phone1> <phone2> ... <phoneN>
      
    • utils/dict_dir_add_pronprobs.sh: Smoothing is done on these bigram counts (both silence and non-silence). This probability is added to L.fst directly. (I have to read this paper and code in depth). 2 new files lexiconp.txt and lexicon_silprob.txt are created for this.
    • utils/prepare_lang.sh: Explained in stage 1
    • local/format_lms.sh: Explained in stage 1
    • utils/build_const_arpa_lm.sh: Explained in stage 1
    • steps/align_fmllr.sh: Use model tri3b (including SAT) to create better alignments i.e frames to phones.
  • Stage 8:
    At this we only make a new graph using the previous alignments, decode code it to make a lattice, rescore using new LM and AM. Final HMM-GMM model WERs are based on this lattice/model.

  • Stage 9
    Run a chain TDNN model using these alignments. I am planning to cover that in another document as of now.

Some Useful Sources

About

My documentation of which script does what in Kaldi