Menu

Where is the latest an4 file to download?

Help
rezaee
2016-10-09
2016-10-14
  • rezaee

    rezaee - 2016-10-09

    I have downloaded an4 file from the link inside the tutorial :http://www.speech.cs.cmu.edu/databases/an4/an4_sphere.tar.gz but I saw there are some differences between the content of the files and what it says about in tutorial. also when I ran traning I got many of these Errors!
    where is the latest version of the an4 file to download?

    e/mm/n/an4/feat/an4_clstk/mtos/cen3-mtos-b.mfc' does not exist, or is empty
    WARNING: Error in '/home/mm/n/an4/etc/an4_train.fileids', the feature file '/home/mm/n/an4/feat/an4_clstk/mtos/cen4-mtos-b.mfc' does not exist, or is empty
    WARNING: Error in '/home/mm/n/an4/etc/an4_train.fileids', the feature file '/home/mm/n/an4/feat/an4_clstk/mtos/cen5-mtos-b.mfc' does not exist, or is empty
    WARNING: Error in '/home/mm/n/an4/etc/an4_train.fileids', the feature file '/home/mm/n/an4/feat/an4_clstk/mtos/cen6-mtos-b.mfc' does not exist, or is empty
    WARNING: Error in '/home/mm/n/an4/etc/an4_train.fileids', the feature file '/home/mm/n/an4/feat/an4_clstk/mtos/cen7-mtos-b.mfc' does not exist, or is empty
    WARNING: Error in '/home/mm/n/an4/etc/an4_train.fileids', the feature file '/home/mm/n/an4/feat/an4_clstk/mtos/cen8-mtos-b.mfc' does not exist, or is empty
    WARNING: Error in '/home/mm/n/an4/etc/an4_train.fileids', the feature file '/home/mm/n/an4/feat/an4_clstk/mtxj/an376-mtxj-b.mfc' does not exist, or is empty
    WARNING: Error in '/home/mm/n/an4/etc/an4_train.fileids', the feature file '/home/mm/n/an4/feat/an4_clstk/mtxj/an377-mtxj-b.mfc' does not exist, or is empty
    WARNING: Error in '/home/mm/n/an4/etc/an4_train.fileids', the feature file '/home/mm/n/an4/feat/an4_clstk/mtxj/an378-mtxj-b.mfc' does not exist, or is empty
    WARNING: Error in '/home/mm/n/an4/etc/an4_train.fileids', the feature file '/home/mm/n/an4/feat/an4_clstk/mtxj/an379-mtxj-b.mfc' does not exist, or is empty
    WARNING: Error in '/home/mm/n/an4/etc/an4_train.fileids', the feature file '/home/mm/n/an4/feat/an4_clstk/mtxj/an380-mtxj-b.mfc' does not exist, or is empty
    WARNING: Error in '/home/mm/n/an4/etc/an4_train.fileids', the feature file '/home/mm/n/an4/feat/an4_clstk/mtxj/cen1-mtxj-b.mfc' does not exist, or is empty
    WARNING: Error in '/home/mm/n/an4/etc/an4_train.fileids', the feature file '/home/mm/n/an4/feat/an4_clstk/mtxj/cen2-mtxj-b.mfc' does not exist, or is empty
    WARNING: Error in '/home/mm/n/an4/etc/an4_train.fileids', the feature file '/home/mm/n/an4/feat/an4_clstk/mtxj/cen3-mtxj-b.mfc' does not exist, or is empty
    WARNING: Error in '/home/mm/n/an4/etc/an4_train.fileids', the feature file '/home/mm/n/an4/feat/an4_clstk/mtxj/cen4-mtxj-b.mfc' does not exist, or is empty
    WARNING: Error in '/home/mm/n/an4/etc/an4_train.fileids', the feature file '/home/mm/n/an4/feat/an4_clstk/mtxj/cen5-mtxj-b.mfc' does not exist, or is empty
    WARNING: Error in '/home/mm/n/an4/etc/an4_train.fileids', the feature file '/home/mm/n/an4/feat/an4_clstk/mtxj/cen6-mtxj-b.mfc' does not exist, or is empty
    WARNING: Error in '/home/mm/n/an4/etc/an4_train.fileids', the feature file '/home/mm/n/an4/feat/an4_clstk/mtxj/cen7-mtxj-b.mfc' does not exist, or is empty
    WARNING: Error in '/home/mm/n/an4/etc/an4_train.fileids', the feature file '/home/mm/n/an4/feat/an4_clstk/mtxj/cen8-mtxj-b.mfc' does not exist, or is empty
    WARNING: Error in '/home/mm/n/an4/etc/an4_train.fileids', the feature file '/home/mm/n/an4/feat/an4_clstk/mwhw/an151-mwhw-b.mfc' does not exist, or is empty
    WARNING: Error in '/home/mm/n/an4/etc/an4_train.fileids', the feature file '/home/mm/n/an4/feat/an4_clstk/mwhw/an152-mwhw-b.mfc' does not exist, or is empty
    WARNING: Error in '/home/mm/n/an4/etc/an4_train.fileids', the feature file '/home/mm/n/an4/feat/an4_clstk/mwhw/an153-mwhw-b.mfc' does not exist, or is empty
    WARNING: Error in '/home/mm/n/an4/etc/an4_train.fileids', the feature file '/home/mm/n/an4/feat/an4_clstk/mwhw/an154-mwhw-b.mfc' does not exist, or is empty
    WARNING: Error in '/home/mm/n/an4/etc/an4_train.fileids', the feature file '/home/mm/n/an4/feat/an4_clstk/mwhw/an155-mwhw-b.mfc' does not exist, or is empty
    WARNING: Error in '/home/mm/n/an4/etc/an4_train.fileids', the feature file '/home/mm/n/an4/feat/an4_clstk/mwhw/cen1-mwhw-b.mfc' does not exist, or is empty
    WARNING: Error in '/home/mm/n/an4/etc/an4_train.fileids', the feature file '/home/mm/n/an4/feat/an4_clstk/mwhw/cen2-mwhw-b.mfc' does not exist, or is empty
    WARNING: Error in '/home/mm/n/an4/etc/an4_train.fileids', the feature file '/home/mm/n/an4/feat/an4_clstk/mwhw/cen3-mwhw-b.mfc' does not exist, or is empty
    WARNING: Error in '/home/mm/n/an4/etc/an4_train.fileids', the feature file '/home/mm/n/an4/feat/an4_clstk/mwhw/cen4-mwhw-b.mfc' does not exist, or is empty
    WARNING: Error in '/home/mm/n/an4/etc/an4_train.fileids', the feature file '/home/mm/n/an4/feat/an4_clstk/mwhw/cen5-mwhw-b.mfc' does not exist, or is empty
    WARNING: Error in '/home/mm/n/an4/etc/an4_train.fileids', the feature file '/home/mm/n/an4/feat/an4_clstk/mwhw/cen6-mwhw-b.mfc' does not exist, or is empty
    WARNING: Error in '/home/mm/n/an4/etc/an4_train.fileids', the feature file '/home/mm/n/an4/feat/an4_clstk/mwhw/cen7-mwhw-b.mfc' does not exist, or is empty
    WARNING: Error in '/home/mm/n/an4/etc/an4_train.fileids', the feature file '/home/mm/n/an4/feat/an4_clstk/mwhw/cen8-mwhw-b.mfc' does not exist, or is empty
    
     
  • rezaee

    rezaee - 2016-10-09

    Also downloaded this: http://www.speech.cs.cmu.edu/databases/an4/an4_raw.littleendian.tar.gz
    And did all of the configuretion as it said in tutorial, but got these Errors again:

    WARNING: Error in '/home/mm/an4/etc/an4_train.fileids', the feature file '/home/mm/an4/feat/an4_clstk/meht/cen2-meht-b.mfc' does not exist, or is empty
    WARNING: Error in '/home/mm/an4/etc/an4_train.fileids', the feature file '/home/mm/an4/feat/an4_clstk/meht/cen3-meht-b.mfc' does not exist, or is empty
    WARNING: Error in '/home/mm/an4/etc/an4_train.fileids', the feature file '/home/mm/an4/feat/an4_clstk/meht/cen4-meht-b.mfc' does not exist, or is empty
    WARNING: Error in '/home/mm/an4/etc/an4_train.fileids', the feature file '/home/mm/an4/feat/an4_clstk/meht/cen5-meht-b.mfc' does not exist, or is empty
    WARNING: Error in '/home/mm/an4/etc/an4_train.fileids', the feature file '/home/mm/an4/feat/an4_clstk/meht/cen6-meht-b.mfc' does not exist, or is empty
    WARNING: Error in '/home/mm/an4/etc/an4_train.fileids', the feature file '/home/mm/an4/feat/an4_clstk/meht/cen7-meht-b.mfc' does not exist, or is empty
    WARNING: Error in '/home/mm/an4/etc/an4_train.fileids', the feature file '/home/mm/an4/feat/an4_clstk/meht/cen8-meht-b.mfc' does not exist, or is empty
    WARNING: Error in '/home/mm/an4/etc/an4_train.fileids', the feature file '/home/mm/an4/feat/an4_clstk/mema/an286-mema-b.mfc' does not exist, or is empty
    WARNING: Error in '/home/mm/an4/etc/an4_train.fileids', the feature file '/home/mm/an4/feat/an4_clstk/mema/an287-mema-b.mfc' does not exist, or is empty
    
     
  • rezaee

    rezaee - 2016-10-09

    I have changed these lines in my code:

    $CFG_WAVFILE_EXTENSION = 'sph';
    $CFG_WAVFILE_TYPE = 'nist'; # one of nist, mswav, raw
    

    and got this new Error:

    Sphinxtrain path: /usr/local/lib/sphinxtrain
    Sphinxtrain binaries path: /usr/local/libexec/sphinxtrain
    Running the training
    MODULE: 000 Computing feature from audio files
    Extracting features from  segments starting at  (part 1 of 1) 
    Extracting features from  segments starting at  (part 1 of 1) 
    Feature extraction is done
    MODULE: 00 verify training files
        Phase 1: Checking to see if the dict and filler dict agrees with the phonelist file.
            Found 133 words using 34 phones
        Phase 2: Checking to make sure there are not duplicate entries in the dictionary
        Phase 3: Check general format for the fileids file; utterance length (must be positive); files exist
        Phase 4: Checking number of lines in the transcript file should match lines in fileids file
        Phase 5: Determine amount of training data, see if n_tied_states seems reasonable.
            Estimated Total Hours Training: 0.702463888888889
            This is a small amount of data, no comment at this time
        Phase 6: Checking that all the words in the transcript are in the dictionary
            Words in dictionary: 130
            Words in filler dictionary: 3
        Phase 7: Checking that all the phones in the transcript are in the phonelist, and all phones in the phonelist appear at least once
    MODULE: 0000 train grapheme-to-phoneme model
    Skipped (set $CFG_G2P_MODEL = 'yes' to enable)
    MODULE: 01 Train LDA transformation
    Skipped (set $CFG_LDA_MLLT = 'yes' to enable)
    MODULE: 02 Train MLLT transformation
    Skipped (set $CFG_LDA_MLLT = 'yes' to enable)
    MODULE: 05 Vector Quantization
    Skipped for continuous models
    MODULE: 10 Training Context Independent models for forced alignment and VTLN
    Skipped:  $ST::CFG_FORCEDALIGN set to 'no' in sphinx_train.cfg
    Skipped:  $ST::CFG_VTLN set to 'no' in sphinx_train.cfg
    MODULE: 11 Force-aligning transcripts
    Skipped:  $ST::CFG_FORCEDALIGN set to 'no' in sphinx_train.cfg
    MODULE: 12 Force-aligning data for VTLN
    Skipped:  $ST::CFG_VTLN set to 'no' in sphinx_train.cfg
    MODULE: 20 Training Context Independent models
        Phase 1: Cleaning up directories:
        accumulator...logs...qmanager...models...
        Phase 2: Flat initialize
    ERROR: FATAL: "cmn.c", line 126: Unknown CMN type 'batch'
    
     
    • Nickolay V. Shmyrev

      You need to update sphinxbase.

       
  • rezaee

    rezaee - 2016-10-13

    How can I do it?

     
  • rezaee

    rezaee - 2016-10-13

    I have downloaded and reinstalled the sphinxbase and solved the problem.
    thanks!

     
  • rezaee

    rezaee - 2016-10-13

    But this is new Errors :D

    Can't open /result/an4-1-2.match
    Can't open /result/an4-2-2.match
    Can't open /etc/_test.fileids for reading
    

    I have no results file in my downloaded an4 file.

     
  • rezaee

    rezaee - 2016-10-13

    I have lost my OS because of a mistake and reinstalled it, and everythin, and this is my new Error when I want to run sphinxtrain run

    Sphinxtrain path: /usr/local/lib/sphinxtrain
    Sphinxtrain binaries path: /usr/local/libexec/sphinxtrain
    Running the training
    MODULE: 000 Computing feature from audio files
    Extracting features from  segments starting at  (part 1 of 1) 
    Extracting features from  segments starting at  (part 1 of 1) 
    Feature extraction is done
    MODULE: 00 verify training files
        Phase 1: Checking to see if the dict and filler dict agrees with the phonelist file.
            Found 133 words using 34 phones
        Phase 2: Checking to make sure there are not duplicate entries in the dictionary
        Phase 3: Check general format for the fileids file; utterance length (must be positive); files exist
        Phase 4: Checking number of lines in the transcript file should match lines in fileids file
        Phase 5: Determine amount of training data, see if n_tied_states seems reasonable.
            Estimated Total Hours Training: 0.702463888888889
            This is a small amount of data, no comment at this time
        Phase 6: Checking that all the words in the transcript are in the dictionary
            Words in dictionary: 130
            Words in filler dictionary: 3
        Phase 7: Checking that all the phones in the transcript are in the phonelist, and all phones in the phonelist appear at least once
    MODULE: 0000 train grapheme-to-phoneme model
    Skipped (set $CFG_G2P_MODEL = 'yes' to enable)
    MODULE: 01 Train LDA transformation
    Skipped (set $CFG_LDA_MLLT = 'yes' to enable)
    MODULE: 02 Train MLLT transformation
    Skipped (set $CFG_LDA_MLLT = 'yes' to enable)
    MODULE: 05 Vector Quantization
    Skipped for continuous models
    MODULE: 10 Training Context Independent models for forced alignment and VTLN
    Skipped:  $ST::CFG_FORCEDALIGN set to 'no' in sphinx_train.cfg
    Skipped:  $ST::CFG_VTLN set to 'no' in sphinx_train.cfg
    MODULE: 11 Force-aligning transcripts
    Skipped:  $ST::CFG_FORCEDALIGN set to 'no' in sphinx_train.cfg
    MODULE: 12 Force-aligning data for VTLN
    Skipped:  $ST::CFG_VTLN set to 'no' in sphinx_train.cfg
    MODULE: 20 Training Context Independent models
        Phase 1: Cleaning up directories:
        accumulator...logs...qmanager...models...
        Phase 2: Flat initialize
        Phase 3: Forward-Backward
            Baum welch starting for 1 Gaussian(s), iteration: 1 (1 of 1)
            0% 
    ERROR: Only 0 parts of 1 of Baum Welch were successfully completed
    ERROR: Parts 1 failed to run!
    ERROR: Training failed in iteration 1
    
     
    • Nickolay V. Shmyrev

      You can check logs for details, most likely you didn't set LD_LIBRARY_PATH, so binaries fail to find sphinxbase.

       
  • rezaee

    rezaee - 2016-10-13

    This is the Error:

    MODULE: DECODE Decoding using models previously trained (2016-10-13 23:08)
    
    Decoding 0 segments starting at 0 (part 2 of 2)
    
    pocketsphinx_batch Log File
    
    ERROR: FATAL: "batch.c", line 815: Failed to open control file '/etc/_test.fileids': No such file or directory
    FAILED
    
    ERROR: Failed to start pocketsphinx_batch
    
    Decoding 0 segments starting at 0 (part 1 of 2)
    
    pocketsphinx_batch Log File
    
    ERROR: FATAL: "batch.c", line 815: Failed to open control file '/etc/_test.fileids': No such file or directory
    FAILED
    
    ERROR: Failed to start pocketsphinx_batch
    
     
    • Nickolay V. Shmyrev

      This error suggests you didn't run setup command properly

       
      • rezaee

        rezaee - 2016-10-13

        I did it in this way:

        root@dhcppc3:/home/m/cmusphinx/an4# sphinxtrain -t an4 setup
        Sphinxtrain path: /usr/local/lib/sphinxtrain
        Sphinxtrain binaries path: /usr/local/libexec/sphinxtrain
        Setting up the database an4
        

        and also did all the configurations in sphinx_train.cfg

         
        • Nickolay V. Shmyrev

          share etc/sphinx_train.cfg you got.

           
          • rezaee

            rezaee - 2016-10-14

            This is the file after configuration:

            # Configuration script for sphinx trainer                  -*-mode:Perl-*-
            
            $CFG_VERBOSE = 1;       # Determines how much goes to the screen.
            
            # These are filled in at configuration time
            $CFG_DB_NAME = "an4";
            # Experiment name, will be used to name model files and log files
            $CFG_EXPTNAME = "$CFG_DB_NAME";
            
            # Directory containing SphinxTrain binaries
            $CFG_BASE_DIR = "/home/m/cmusphinx/an4";
            $CFG_SPHINXTRAIN_DIR = "/usr/local/lib/sphinxtrain";
            $CFG_BIN_DIR = "/usr/local/libexec/sphinxtrain";
            $CFG_SCRIPT_DIR = "/usr/local/lib/sphinxtrain/scripts";
            
            
            # Audio waveform and feature file information
            $CFG_WAVFILES_DIR = "$CFG_BASE_DIR/wav";
            $CFG_WAVFILE_EXTENSION = 'wav';
            $CFG_WAVFILE_TYPE = 'mswav'; # one of nist, mswav, raw
            $CFG_FEATFILES_DIR = "$CFG_BASE_DIR/feat";
            $CFG_FEATFILE_EXTENSION = 'mfc';
            
            # Feature extraction parameters
            $CFG_WAVFILE_SRATE = 16000.0;
            $CFG_NUM_FILT = 25; # For wideband speech it's 25, for telephone 8khz reasonable value is 15
            $CFG_LO_FILT = 130; # For telephone 8kHz speech value is 200
            $CFG_HI_FILT = 6800; # For telephone 8kHz speech value is 3500
            $CFG_TRANSFORM = "dct"; # Previously legacy transform is used, but dct is more accurate
            $CFG_LIFTER = "22"; # Cepstrum lifter is smoothing to improve recognition
            $CFG_VECTOR_LENGTH = 13; # 13 is usually enough
            
            $CFG_MIN_ITERATIONS = 1;  # BW Iterate at least this many times
            $CFG_MAX_ITERATIONS = 10; # BW Don't iterate more than this, somethings likely wrong.
            
            # (none/max) Type of AGC to apply to input files
            $CFG_AGC = 'none';
            # (current/none) Type of cepstral mean subtraction/normalization
            # to apply to input files
            $CFG_CMN = 'batch';
            # (yes/no) Normalize variance of input files to 1.0
            $CFG_VARNORM = 'no';
            # (yes/no) Train full covariance matrices
            $CFG_FULLVAR = 'no';
            # (yes/no) Use diagonals only of full covariance matrices for
            # Forward-Backward evaluation (recommended if CFG_FULLVAR is yes)
            $CFG_DIAGFULL = 'no';
            
            # (yes/no) Perform vocal tract length normalization in training.  This
            # will result in a "normalized" model which requires VTLN to be done
            # during decoding as well.
            $CFG_VTLN = 'no';
            # Starting warp factor for VTLN
            $CFG_VTLN_START = 0.80;
            # Ending warp factor for VTLN
            $CFG_VTLN_END = 1.40;
            # Step size of warping factors
            $CFG_VTLN_STEP = 0.05;
            
            # Directory to write queue manager logs to
            $CFG_QMGR_DIR = "$CFG_BASE_DIR/qmanager";
            # Directory to write training logs to
            $CFG_LOG_DIR = "$CFG_BASE_DIR/logdir";
            # Directory for re-estimation counts
            $CFG_BWACCUM_DIR = "$CFG_BASE_DIR/bwaccumdir";
            # Directory to write model parameter files to
            $CFG_MODEL_DIR = "$CFG_BASE_DIR/model_parameters";
            
            # Directory containing transcripts and control files for
            # speaker-adaptive training
            $CFG_LIST_DIR = "$CFG_BASE_DIR/etc";
            
            # Decoding variables for MMIE training
            $CFG_LANGUAGEWEIGHT = "11.5";
            $CFG_BEAMWIDTH      = "1e-100";
            $CFG_WORDBEAM       = "1e-80";
            $CFG_LANGUAGEMODEL  = "$CFG_LIST_DIR/$CFG_DB_NAME.lm.DMP";
            $CFG_WORDPENALTY    = "0.2";
            
            # Lattice pruning variables
            $CFG_ABEAM              = "1e-50";
            $CFG_NBEAM              = "1e-10";
            $CFG_PRUNED_DENLAT_DIR  = "$CFG_BASE_DIR/pruned_denlat";
            
            # MMIE training related variables
            $CFG_MMIE = "no";
            $CFG_MMIE_MAX_ITERATIONS = 5;
            $CFG_LATTICE_DIR = "$CFG_BASE_DIR/lattice";
            $CFG_MMIE_TYPE   = "rand"; # Valid values are "rand", "best" or "ci"
            $CFG_MMIE_CONSTE = "3.0";
            $CFG_NUMLAT_DIR  = "$CFG_BASE_DIR/numlat";
            $CFG_DENLAT_DIR  = "$CFG_BASE_DIR/denlat";
            
            # Variables used in main training of models
            $CFG_DICTIONARY     = "$CFG_LIST_DIR/$CFG_DB_NAME.dic";
            $CFG_RAWPHONEFILE   = "$CFG_LIST_DIR/$CFG_DB_NAME.phone";
            $CFG_FILLERDICT     = "$CFG_LIST_DIR/$CFG_DB_NAME.filler";
            $CFG_LISTOFFILES    = "$CFG_LIST_DIR/${CFG_DB_NAME}_train.fileids";
            $CFG_TRANSCRIPTFILE = "$CFG_LIST_DIR/${CFG_DB_NAME}_train.transcription";
            $CFG_FEATPARAMS     = "$CFG_LIST_DIR/feat.params";
            
            # Variables used in characterizing models
            
            $CFG_HMM_TYPE = '.cont.'; # Sphinx 4, PocketSphinx
            #$CFG_HMM_TYPE  = '.semi.'; # PocketSphinx
            #$CFG_HMM_TYPE  = '.ptm.'; # PocketSphinx (larger data sets)
            
            if (($CFG_HMM_TYPE ne ".semi.")
                and ($CFG_HMM_TYPE ne ".ptm.")
                and ($CFG_HMM_TYPE ne ".cont.")) {
              die "Please choose one CFG_HMM_TYPE out of '.cont.', '.ptm.', or '.semi.', " .
                "currently $CFG_HMM_TYPE\n";
            }
            
            # This configuration is fastest and best for most acoustic models in
            # PocketSphinx and Sphinx-III.  See below for Sphinx-II.
            $CFG_STATESPERHMM = 3;
            $CFG_SKIPSTATE = 'no';
            
            if ($CFG_HMM_TYPE eq '.semi.') {
              $CFG_DIRLABEL = 'semi';
            # Four stream features for PocketSphinx
              $CFG_FEATURE = "s2_4x";
              $CFG_NUM_STREAMS = 4;
              $CFG_INITIAL_NUM_DENSITIES = 256;
              $CFG_FINAL_NUM_DENSITIES = 256;
              die "For semi continuous models, the initial and final models have the same density" 
                if ($CFG_INITIAL_NUM_DENSITIES != $CFG_FINAL_NUM_DENSITIES);
            } elsif ($CFG_HMM_TYPE eq '.ptm.') {
              $CFG_DIRLABEL = 'ptm';
            # Four stream features for PocketSphinx
              $CFG_FEATURE = "s2_4x";
              $CFG_NUM_STREAMS = 4;
              $CFG_INITIAL_NUM_DENSITIES = 64;
              $CFG_FINAL_NUM_DENSITIES = 64;
              die "For phonetically tied models, the initial and final models have the same density" 
                if ($CFG_INITIAL_NUM_DENSITIES != $CFG_FINAL_NUM_DENSITIES);
            } elsif ($CFG_HMM_TYPE eq '.cont.') {
              $CFG_DIRLABEL = 'cont';
            # Single stream features - Sphinx 3
              $CFG_FEATURE = "1s_c_d_dd";
              $CFG_NUM_STREAMS = 1;
              $CFG_INITIAL_NUM_DENSITIES = 1;
              $CFG_FINAL_NUM_DENSITIES = 8;
              die "The initial has to be less than the final number of densities" 
                if ($CFG_INITIAL_NUM_DENSITIES > $CFG_FINAL_NUM_DENSITIES);
            }
            
            # Number of top gaussians to score a frame. A little bit less accurate computations
            # make training significantly faster. Uncomment to apply this during the training
            # For good accuracy make sure you are using the same setting in decoder
            # In theory this can be different for various training stages. For example 4 for
            # CI stage and 16 for CD stage
            # $CFG_CI_TOPN = 4;
            # $CFG_CD_TOPN = 16;
            
            # (yes/no) Train multiple-gaussian context-independent models (useful
            # for alignment, use 'no' otherwise) in the models created
            # specifically for forced alignment
            $CFG_FALIGN_CI_MGAU = 'no';
            # (yes/no) Train multiple-gaussian context-independent models (useful
            # for alignment, use 'no' otherwise)
            $CFG_CI_MGAU = 'no';
            # (yes/no) Train context-dependent models
            $CFG_CD_TRAIN = 'yes';
            # Number of tied states (senones) to create in decision-tree clustering
            $CFG_N_TIED_STATES = 1000;
            # How many parts to run Forward-Backward estimatinon in
            $CFG_NPART = 2;
            
            # (yes/no) Train a single decision tree for all phones (actually one
            # per state) (useful for grapheme-based models, use 'no' otherwise)
            $CFG_CROSS_PHONE_TREES = 'no';
            
            # Use force-aligned transcripts (if available) as input to training
            $CFG_FORCEDALIGN = 'no';
            
            # Use a specific set of models for force alignment.  If not defined,
            # context-independent models for the current experiment will be used.
            $CFG_FORCE_ALIGN_MODELDIR = "$CFG_MODEL_DIR/$CFG_EXPTNAME.falign_ci_$CFG_DIRLABEL";
            
            # Use a specific dictionary and filler dictionary for force alignment.
            # If these are not defined, a dictionary and filler dictionary will be
            # created from $CFG_DICTIONARY and $CFG_FILLERDICT, with noise words
            # removed from the filler dictionary and added to the dictionary (this
            # is because the force alignment is not very good at inserting them)
            
            # $CFG_FORCE_ALIGN_DICTIONARY = "$ST::CFG_BASE_DIR/falignout$ST::CFG_EXPTNAME.falign.dict";;
            # $CFG_FORCE_ALIGN_FILLERDICT = "$ST::CFG_BASE_DIR/falignout/$ST::CFG_EXPTNAME.falign.fdict";;
            
            # Use a particular beam width for force alignment.  The wider
            # (i.e. smaller numerically) the beam, the fewer sentences will be
            # rejected for bad alignment.
            $CFG_FORCE_ALIGN_BEAM = 1e-60;
            
            # Calculate an LDA/MLLT transform?
            $CFG_LDA_MLLT = 'no';
            # Dimensionality of LDA/MLLT output
            $CFG_LDA_DIMENSION = 29;
            
            # This is actually just a difference in log space (it doesn't make
            # sense otherwise, because different feature parameters have very
            # different likelihoods)
            $CFG_CONVERGENCE_RATIO = 0.1;
            
            # Queue::POSIX for multiple CPUs on a local machine
            # Queue::PBS to use a PBS/TORQUE queue
            $CFG_QUEUE_TYPE = "Queue::POSIX";
            
            # Name of queue to use for PBS/TORQUE
            $CFG_QUEUE_NAME = "workq";
            
            # (yes/no) Build questions for decision tree clustering automatically
            $CFG_MAKE_QUESTS = "yes";
            # If CFG_MAKE_QUESTS is yes, questions are written to this file.
            # If CFG_MAKE_QUESTS is no, questions are read from this file.
            $CFG_QUESTION_SET = "${CFG_BASE_DIR}/model_architecture/${CFG_EXPTNAME}.tree_questions";
            #$CFG_QUESTION_SET = "${CFG_BASE_DIR}/linguistic_questions";
            
            $CFG_CP_OPERATION = "${CFG_BASE_DIR}/model_architecture/${CFG_EXPTNAME}.cpmeanvar";
            
            # Configuration for grapheme-to-phoneme model
            $CFG_G2P_MODEL= 'no';
            
            # Configuration script for sphinx decoder 
            
            # Variables starting with $DEC_CFG_ refer to decoder specific
            # arguments, those starting with $CFG_ refer to trainer arguments,
            # some of them also used by the decoder.
            
            $DEC_CFG_VERBOSE = 1;       # Determines how much goes to the screen.
            
            # These are filled in at configuration time
            
            # Name of the decoding script to use (psdecode.pl or s3decode.pl, probably)
            $DEC_CFG_SCRIPT = 'psdecode.pl';
            
            $DEC_CFG_EXPTNAME = "$CFG_EXPTNAME";
            $DEC_CFG_JOBNAME  = "$CFG_EXPTNAME"."_job";
            
            # Models to use.
            $DEC_CFG_MODEL_NAME = "$CFG_EXPTNAME.cd_${CFG_DIRLABEL}_${CFG_N_TIED_STATES}";
            
            $DEC_CFG_FEATFILES_DIR = "$CFG_BASE_DIR/feat";
            $DEC_CFG_FEATFILE_EXTENSION = '.mfc';
            $DEC_CFG_AGC = $CFG_AGC;
            $DEC_CFG_CMN = $CFG_CMN;
            $DEC_CFG_VARNORM = $CFG_VARNORM;
            
            $DEC_CFG_QMGR_DIR = "$CFG_BASE_DIR/qmanager";
            $DEC_CFG_LOG_DIR = "$CFG_BASE_DIR/logdir";
            $DEC_CFG_MODEL_DIR = "$CFG_MODEL_DIR";
            
            #$DEC_CFG_DICTIONARY     = "$CFG_BASE_DIR/etc/$CFG_DB_NAME.dic";
            #$DEC_CFG_FILLERDICT     = "$CFG_BASE_DIR/etc/$CFG_DB_NAME.filler";
            #$DEC_CFG_LISTOFFILES    = "$CFG_BASE_DIR/etc/${CFG_DB_NAME}_test.fileids";
            #$DEC_CFG_TRANSCRIPTFILE = "$CFG_BASE_DIR/etc/${CFG_DB_NAME}_test.transcription";
            #$DEC_CFG_RESULT_DIR     = "$CFG_BASE_DIR/result";
            $DEC_CFG_PRESULT_DIR     = "$DEC_CFG_BASE_DIR/presult";
            
            # This variables, used by the decoder, have to be user defined, and
            # may affect the decoder output
            
            
            
            $DEC_CFG_DICTIONARY     = "$DEC_CFG_BASE_DIR/etc/$DEC_CFG_DB_NAME.dic";
            $DEC_CFG_FILLERDICT     = "$DEC_CFG_BASE_DIR/etc/$DEC_CFG_DB_NAME.filler";
            $DEC_CFG_LISTOFFILES    = "$DEC_CFG_BASE_DIR/etc/${DEC_CFG_DB_NAME}_test.fileids";
            $DEC_CFG_TRANSCRIPTFILE = "$DEC_CFG_BASE_DIR/etc/${DEC_CFG_DB_NAME}_test.transcription";
            $DEC_CFG_RESULT_DIR     = "$DEC_CFG_BASE_DIR/result";
            
            # These variables, used by the decoder, have to be user defined, and
            # may affect the decoder output
            
            $DEC_CFG_LANGUAGEMODEL_DIR = "$DEC_CFG_BASE_DIR/etc";
            $DEC_CFG_LANGUAGEMODEL  = "$DEC_CFG_LANGUAGEMODEL_DIR/an4.ug.lm.DMP";
            
            #$DEC_CFG_LANGUAGEMODEL  = "$CFG_BASE_DIR/etc/${CFG_DB_NAME}.lm.DMP";
            # Or can be JSGF or FSG too, used if uncommented
            # $DEC_CFG_GRAMMAR  = "$CFG_BASE_DIR/etc/${CFG_DB_NAME}.jsgf";
            # $DEC_CFG_FSG  = "$CFG_BASE_DIR/etc/${CFG_DB_NAME}.fsg";
            
            $DEC_CFG_LANGUAGEWEIGHT = "10";
            $DEC_CFG_BEAMWIDTH = "1e-80";
            $DEC_CFG_WORDBEAM = "1e-40";
            $DEC_CFG_WORDPENALTY = "0.2";
            
            $DEC_CFG_ALIGN = "builtin";
            
            $DEC_CFG_NPART = 2;     #  Define how many pieces to split decode in
            
            # This variable has to be defined, otherwise utils.pl will not load.
            $CFG_DONE = 1;
            
            return 1;
            
             
            • rezaee

              rezaee - 2016-10-14

              I only replaced this part with what it said in tutorial:

              #$DEC_CFG_DICTIONARY     = "$CFG_BASE_DIR/etc/$CFG_DB_NAME.dic";
              #$DEC_CFG_FILLERDICT     = "$CFG_BASE_DIR/etc/$CFG_DB_NAME.filler";
              #$DEC_CFG_LISTOFFILES    = "$CFG_BASE_DIR/etc/${CFG_DB_NAME}_test.fileids";
              #$DEC_CFG_TRANSCRIPTFILE = "$CFG_BASE_DIR/etc/${CFG_DB_NAME}_test.transcription";
              #$DEC_CFG_RESULT_DIR     = "$CFG_BASE_DIR/result";
              

              and added this line from tutorial (there wasn't in my file):

              `$DEC_CFG_LANGUAGEMODEL_DIR = "$DEC_CFG_BASE_DIR/etc";
              

              finally changed the number of processors to 2.
              And also changed the an4.lm.DMP to an4.ug.lm.DMP

              $DEC_CFG_LANGUAGEMODEL  = "$DEC_CFG_LANGUAGEMODEL_DIR/an4.ug.lm.DMP";
              

              All the other parts were as same as the tutorial.

               

              Last edit: rezaee 2016-10-14
              • Nickolay V. Shmyrev

                Actually you should not do this change, this was an error in tutorial, you only need to change DEC_CFG_LANGAUGEMODEL line. I fixed a tutorial and suggest you to start from setup again.

                 
  • rezaee

    rezaee - 2016-10-14

    Ok thanks. solved the problem. this is the result:

            SENTENCE ERROR: 61.5% (80/130)   WORD ERROR RATE: 27.4% (212/773)
    

    What a good results! 61.5% sentence Error!!

    How many hours of sounds, and the sounds of how many people this database has?

     
  • rezaee

    rezaee - 2016-10-14

    In the tutorial says you should at least run all these files:

    perl scripts_pl/000.comp_feat/slave_feat.pl
    perl scripts_pl/00.verify/verify_all.pl
    perl scripts_pl/10.vector_quantize/slave.VQ.pl
    perl scripts_pl/20.ci_hmm/slave_convg.pl
    perl scripts_pl/30.cd_hmm_untied/slave_convg.pl
    perl scripts_pl/40.buildtrees/slave.treebuilder.pl
    perl scripts_pl/45.prunetree/slave-state-tying.pl
    perl scripts_pl/50.cd_hmm_tied/slave_convg.pl
    perl scripts_pl/90.deleted_interpolation/deleted_interpolation.pl
    

    But in the configuration section, doesn't say anything about enabling this commands:

    $ST::CFG_FORCEDALIGN
    $ST::CFG_VTLN
    

    So I have changed the conf file to :

    $CFG_VTLN = 'yes';
    $CFG_FORCEDALIGN = 'yes';
    

    But the running command "sphinxtrain run" stop with this command:

    root@dhcppc3:/home/m/cmusphinx/an4# sphinxtrain runSphinxtrain path: /usr/local/lib/sphinxtrain
    Sphinxtrain binaries path: /usr/local/libexec/sphinxtrain
    Running the training
    MODULE: 000 Computing feature from audio files
    Extracting features from  segments starting at  (part 1 of 2) 
    Extracting features from  segments starting at  (part 2 of 2) 
    Extracting features from  segments starting at  (part 1 of 2) 
    Extracting features from  segments starting at  (part 2 of 2) 
    Feature extraction is done
    MODULE: 00 verify training files
        Phase 1: Checking to see if the dict and filler dict agrees with the phonelist file.
            Found 133 words using 34 phones
        Phase 2: Checking to make sure there are not duplicate entries in the dictionary
        Phase 3: Check general format for the fileids file; utterance length (must be positive); files exist
        Phase 4: Checking number of lines in the transcript file should match lines in fileids file
        Phase 5: Determine amount of training data, see if n_tied_states seems reasonable.
            Estimated Total Hours Training: 0.702463888888889
            This is a small amount of data, no comment at this time
        Phase 6: Checking that all the words in the transcript are in the dictionary
            Words in dictionary: 130
            Words in filler dictionary: 3
        Phase 7: Checking that all the phones in the transcript are in the phonelist, and all phones in the phonelist appear at least once
    MODULE: 0000 train grapheme-to-phoneme model
    Skipped (set $CFG_G2P_MODEL = 'yes' to enable)
    MODULE: 01 Train LDA transformation
    Skipped (set $CFG_LDA_MLLT = 'yes' to enable)
    MODULE: 02 Train MLLT transformation
    Skipped (set $CFG_LDA_MLLT = 'yes' to enable)
    MODULE: 05 Vector Quantization
    Skipped for continuous models
    MODULE: 10 Training Context Independent models for forced alignment and VTLN
        Phase 1: Cleaning up directories:
        accumulator...logs...qmanager...models...
        Phase 2: Flat initialize
        Phase 3: Forward-Backward
            Baum welch starting for 1 Gaussian(s), iteration: 1 (1 of 2)
            0% 10% 20% 30% 40% 50% 60% 70% 80% 90% 
            Baum welch starting for 1 Gaussian(s), iteration: 1 (2 of 2)
            0% 10% 20% 30% 40% 50% 60% 70% 80% 90% 100% 
            Normalization for iteration: 1
            Current Overall Likelihood Per Frame = -156.043331606607
            Baum welch starting for 1 Gaussian(s), iteration: 2 (1 of 2)
            0% 10% 20% 30% 40% 50% 60% 70% 80% 90% 
            Baum welch starting for 1 Gaussian(s), iteration: 2 (2 of 2)
            0% 10% 20% 30% 40% 50% 60% 70% 80% 90% 100% 
            Normalization for iteration: 2
            Current Overall Likelihood Per Frame = -154.047736736171
            Convergence Ratio = 1.99559487043632
            Baum welch starting for 1 Gaussian(s), iteration: 3 (1 of 2)
            0% 10% 20% 30% 40% 50% 60% 70% 80% 90% 
            Baum welch starting for 1 Gaussian(s), iteration: 3 (2 of 2)
            0% 10% 20% 30% 40% 50% 60% 70% 80% 90% 100% 
            Normalization for iteration: 3
            Current Overall Likelihood Per Frame = -151.482163970469
            Convergence Ratio = 2.56557276570197
            Baum welch starting for 1 Gaussian(s), iteration: 4 (1 of 2)
            0% 10% 20% 30% 40% 50% 60% 70% 80% 90% 
            Baum welch starting for 1 Gaussian(s), iteration: 4 (2 of 2)
            0% 10% 20% 30% 40% 50% 60% 70% 80% 90% 100% 
            Normalization for iteration: 4
            Current Overall Likelihood Per Frame = -149.963659658266
            Convergence Ratio = 1.51850431220265
            Baum welch starting for 1 Gaussian(s), iteration: 5 (1 of 2)
            0% 10% 20% 30% 40% 50% 60% 70% 80% 90% 
            Baum welch starting for 1 Gaussian(s), iteration: 5 (2 of 2)
            0% 10% 20% 30% 40% 50% 60% 70% 80% 90% 100% 
            Normalization for iteration: 5
            Current Overall Likelihood Per Frame = -149.573999454302
            Convergence Ratio = 0.389660203964297
            Baum welch starting for 1 Gaussian(s), iteration: 6 (1 of 2)
            0% 10% 20% 30% 40% 50% 60% 70% 80% 90% 
            Baum welch starting for 1 Gaussian(s), iteration: 6 (2 of 2)
            0% 10% 20% 30% 40% 50% 60% 70% 80% 90% 100% 
            Normalization for iteration: 6
            Current Overall Likelihood Per Frame = -149.436428127978
            Convergence Ratio = 0.137571326323894
            Baum welch starting for 1 Gaussian(s), iteration: 7 (1 of 2)
            0% 10% 20% 30% 40% 50% 60% 70% 80% 90% 
            Baum welch starting for 1 Gaussian(s), iteration: 7 (2 of 2)
            0% 10% 20% 30% 40% 50% 60% 70% 80% 90% 100% 
            Normalization for iteration: 7
            Current Overall Likelihood Per Frame = -149.319182085279
            Convergence Ratio = 0.117246042698781
            Baum welch starting for 1 Gaussian(s), iteration: 8 (1 of 2)
            0% 10% 20% 30% 40% 50% 60% 70% 80% 90% 
            Baum welch starting for 1 Gaussian(s), iteration: 8 (2 of 2)
            0% 10% 20% 30% 40% 50% 60% 70% 80% 90% 100% 
            Normalization for iteration: 8
            Current Overall Likelihood Per Frame = -149.274023575747
            Training completed after 8 iterations
    MODULE: 11 Force-aligning transcripts
    Skipped: No sphinx3_align(.exe) found in /usr/local/libexec/sphinxtrain
    If you wish to do force-alignment, please copy or link the
    sphinx3_align binary from Sphinx 3 to /usr/local/libexec/sphinxtrain
    and either define $CFG_MODEL_DIR in sphinx_train.cfg or
    run context-independent training first.
    
     
  • rezaee

    rezaee - 2016-10-14

    I think the module 10 works only for sphinx3 and shouldn't be run for pocketsphinx?

     

Log in to post a comment.

Want the latest updates on software, tech news, and AI?
Get latest updates about software, tech news, and AI from SourceForge directly in your inbox once a month.