https://github.com/freewym/espresso

sort by:
Revision Author Date Message Commit Date
660facf allows dictionary files w/o the counts column; rename task's --max-num-expansions-per-step to --transducer-max-num-expansions-per-step (same as generation's) and its default is 20; prints out word counts after WER evaluation; fixes decoding log write out 24 July 2023, 00:01:56 UTC
e0e61e2 add options to be able to share learnable positional embeddings across layers or attention heads 14 July 2023, 01:36:27 UTC
929ad76 Bugfix in transformer model creation. Making the signatures of methods compatible: `SpeechTransformerModelBase.build_encoder(...)` `SpeechTransformerModel.build_encoder(...)` (otherwise, the model build crashing) 08 December 2022, 09:00:45 UTC
09c7a91 makes BatchNorm configurable in preencoder; adds the bias term to joiner's final fc_out and uses its default initializer 08 December 2022, 09:00:45 UTC
f823f17 adds an option to include EOS in transducer model training; fixes a bug regarding attn_mask; rewrites masked_copy_cached_state() to make it clearer and more general; removes clone_cached_state(); code adaptation/changes according to the commits on Nov 2, 2022 08 December 2022, 09:00:45 UTC
f45c936 transducer decoding fix; cosmatic changes 08 December 2022, 09:00:44 UTC
3b60635 code adaptation/changes according to the commits on Oct 17, 2022 08 December 2022, 09:00:44 UTC
54df33a adds chunk streaming mask support for Transformer 08 December 2022, 09:00:44 UTC
3c21b82 handle installation of packaging 08 December 2022, 09:00:44 UTC
7fa5d35 code adaptation/changes according to the commits on Sep 8, 2022 08 December 2022, 09:00:44 UTC
4834687 CTC model is supported; adjusts num_workers and required_batch_size_m… (#76) * CTC model is supported; adjusts num_workers and required_batch_size_multiple; adjusts num updates accordingly; cosmetic changes 08 December 2022, 09:00:43 UTC
4bf0874 cosmetic improvement for specaugment config and transducer loss; add extra_symbols_to_ignore when printing out sample predictions 08 December 2022, 09:00:43 UTC
3c5be1b Revert "code adaptation/changes according to the commits on June 15, 2022" This reverts commit 63d3458de0e6846e95e992e738f2a823d8a9ffd8. 08 December 2022, 09:00:43 UTC
45b30ae code adaptation/changes according to the commits on June 15, 2022 08 December 2022, 09:00:43 UTC
844c858 add the "learnable" property for relative positional embedding and enquiry that in MultiheadAttention 08 December 2022, 09:00:43 UTC
5a2b90e code adaptation/changes according to the commits on May 4, 2022 08 December 2022, 09:00:43 UTC
77621cc code adaptation/changes according to the commits on Apr 20, 2022 08 December 2022, 09:00:43 UTC
99b272e code adaptation/changes according to the commits on Mar 24, 2022 08 December 2022, 09:00:43 UTC
16900d9 adaptive specaugment 08 December 2022, 09:00:43 UTC
312ac68 use feature_transforms' API for applying SpecAugment; get rid of hard-coding the train_subset name in tasks 08 December 2022, 09:00:42 UTC
4204aac conformer (#73) 08 December 2022, 09:00:42 UTC
bd2961f code adaptation/changes according to the commits on Jan 20, 2022 08 December 2022, 09:00:41 UTC
13c2817 update relative positional embedding implementations; tune a bit of transducer models 08 December 2022, 09:00:41 UTC
79d5a2b batch data based on both src and tgt lengths for transducer models 08 December 2022, 09:00:40 UTC
c941705 transformer transducer code check-in (#72) Includes: * a transformer-transducer model * transducer loss * Noam lr scheduler * greedy decoding and beam search decoding with LM shallow fusion * a LibriSpeech recipe 08 December 2022, 09:00:40 UTC
6f4558a in-place masked_copy_incremental_state; fix import module issue by dummy importing espresso in sequence_generator.py 08 December 2022, 09:00:40 UTC
f9220e9 remove an unnecessary check for LM in sequence_generator.py; other cosmetic changes 08 December 2022, 09:00:39 UTC
5c4c288 suppress deprecation warning for division with rounding 08 December 2022, 09:00:39 UTC
2ead96c reformat for the previous commit 08 December 2022, 09:00:39 UTC
c3f66c2 added relaxed attention to encoder-decoder attention (arxiv.org/abs/2107.01275) (#71) * added relaxed attention to encoder-decoder attention (arxiv.org/abs/2107.01275) * Adressed change requests for PR * override build_encoder_attn for relaxed attention PR Co-authored-by: Timo Lohrenz <lohrenz@ifn.ing.tu-bs.de> 08 December 2022, 09:00:39 UTC
2cded21 code adaptation/changes according to the commits on Nov 29, 2021 (Add linting with black) 08 December 2022, 09:00:38 UTC
c5518d3 code adaptation/changes according to the commits on Oct 13, 2021 08 December 2022, 09:00:38 UTC
de30590 code adaptation/changes according to the commits on Sep 9, 2021 08 December 2022, 09:00:38 UTC
9c3c820 refactor transformer 2 (according to the commits on July 16-28, 2021) 08 December 2022, 09:00:38 UTC
7c5037f refactor transformer 1 (according to the commits on July 16, 2021) 08 December 2022, 09:00:38 UTC
9fb004a code adaptation/changes according to the commits on June 24-July 1, 2021 08 December 2022, 09:00:38 UTC
6e73519 Relative Positional Embedding (fixed or learnable) (#68) * Relative Positional Embedding (fixed or learnable) * support relative positional embeddings in transformer models * bug fixes; fixes command line args passing for SpeechTransformerModel * f * ff 08 December 2022, 09:00:38 UTC
c14194a code adaptation/changes according to the commits on June 11, 2021 08 December 2022, 09:00:38 UTC
e5eff0e code adaptation/changes according to the commits on June 3, 2021; empty GPU memory cache of the WSJ recipe more frequently to avoid occasional CUDNN error; fixes the command line args parsing problem for SpeechTransformerModel 08 December 2022, 09:00:37 UTC
b3de88b fix a comment about the shape of encoder_padding_mask in speech_lstm; fix a bug in the task speech_recognition_hybrid about converting to an iterator; fix an import bug in feat_text_dataset 08 December 2022, 09:00:37 UTC
81d86a1 code adaptation/changes according to the commits on May 21, 2021 08 December 2022, 09:00:37 UTC
c5fcfa1 decapitalize bool values in yaml configs 08 December 2022, 09:00:36 UTC
026faf1 add an option to allow multi-layer LSTM defined as a single module (faster and may fix a possible cuDNN error when defined as a ModuleList); add missing global cmvn to the decoding stage 08 December 2022, 09:00:36 UTC
e4213c7 hydra migration for librispeech torchaudio recipes (#65) 08 December 2022, 09:00:36 UTC
76106fa code adaptation/changes according to the commits on Apr 23, 2021 08 December 2022, 09:00:36 UTC
0184238 more informative error msgs when corpus dirs are not specified 08 December 2022, 09:00:36 UTC
dcae0c8 move some task-specifc code in speech_train.py to task functions; move two remaining ones to fairseq's train.py and remove speech_train.py 08 December 2022, 09:00:36 UTC
096cddc on-the-fly feature extraction with torchaudio + LibriSpeech recipe without dependency on Kaldi (#61) 08 December 2022, 09:00:36 UTC
6ff8231 fix misleading error msgs on non-CLSP machines; normalize LM training by update_freq; other style changes 08 December 2022, 09:00:36 UTC
236c0d0 several bug fixes for hybrid systems (thanks to liuhongji2000) 08 December 2022, 09:00:36 UTC
0939b1d support bos in class AsrDictionary for CTC loss, etc 08 December 2022, 09:00:36 UTC
1edb65c code adaptation/changes according to the commits on Mar 29-30, 2021 08 December 2022, 09:00:35 UTC
1b603c6 on-the-fly tokenization 08 December 2022, 09:00:35 UTC
6102cde code adaptation/changes according to the commits on Mar 22-23, 2021 08 December 2022, 09:00:35 UTC
0916351 move convolutions from lstm to a separate file 08 December 2022, 09:00:35 UTC
9dca701 fixes a bug in FeatScpCachedDataset; code adaptation/changes according to the commits on Mar 20, 2021 08 December 2022, 09:00:35 UTC
86cf6de code adaptation/changes according to the commits on Mar 4-12, 2021 08 December 2022, 09:00:35 UTC
a291aeb tokenization within datasets (need to re-run json prep stage right before training stage to be compatible with this change) (#58) 08 December 2022, 09:00:35 UTC
3c8712a code adaptation/changes according to the commits on Feb 8-Mar 2, 2021 08 December 2022, 09:00:35 UTC
537b3b2 fix labeled smoothed cross entropy computation according to the corresponding changes in fairseq on Feb 3rd, 2021; affects recipes using uniform smoothing 08 December 2022, 09:00:34 UTC
a3ae124 code adaptation/changes according to the commits on Jan 20-28, 2021 08 December 2022, 09:00:34 UTC
a9558b3 fix typos in argument names; code adaptation/changes according to the commits on Dec 28, 2020 - Jan 5, 2021 08 December 2022, 09:00:34 UTC
a8b1c99 code adaptation/changes according to the commits on Dec 22, 2020 08 December 2022, 09:00:34 UTC
907c9bc fix length tensor device issue in lf_mmi loss; code adaptation/changes according to the commits on Dec 3-12, 2020 08 December 2022, 09:00:34 UTC
f351215 code adaptation/changes according to the commits on Nov 16-20, 2020; fix a bug in Multi-level LM when getting cached states 08 December 2022, 09:00:33 UTC
83390bf fix an error when more than one external LMs are used for shallow fusion 08 December 2022, 09:00:33 UTC
976dfa8 code adaptation/changes according to the commits on Nov 11, 2020; obtain feat_dim in setup_task() instead 08 December 2022, 09:00:33 UTC
4734e37 code adaptation/changes according to the commits on Nov 4-9, 2020 08 December 2022, 09:00:33 UTC
8028172 code adaptation/changes according to the commits on Oct 18-Nov 3, 2020 (lots of changes, mostly for adapting to hydra configs and code formatting) 08 December 2022, 09:00:33 UTC
91973ad code adaptation/changes according to the commits on Oct 2-15, 2020 08 December 2022, 09:00:32 UTC
63bdcdf code adaptation/changes according to the commits on Oct 1, 2020 08 December 2022, 09:00:32 UTC
3987af6 code adaptation/changes according to the commits on Sep 17-26, 2020 08 December 2022, 09:00:32 UTC
f5bd61f code adaptation/changes according to the commits on Sep 9-11, 2020 08 December 2022, 09:00:31 UTC
094ecc7 code adaptation/changes according to the commits on Aug 31, 2020 08 December 2022, 09:00:31 UTC
38b199c code adaptation/changes according to the commits on Aug 20-24, 2020 08 December 2022, 09:00:31 UTC
eb6507e code adaptation/changes according to the commits on Aug 10-18, 2020 08 December 2022, 09:00:31 UTC
ef05c79 updates for new PyChain (#37) * add support for output l2 regularization and xent regularization; add a bichar WSJ recipe; add missing soft links to kaldi files * move ChainLossFunction here from PyChain 08 December 2022, 09:00:31 UTC
509be47 reorder the elements of the returned tuple of TdnnModel.forward(); export KALDI_ROOT to adapt to the recent changes in kaldi_io; code adaptation/changes according to the commits on Aug 3-4, 2020 08 December 2022, 09:00:31 UTC
e916ef5 fix reorder_encoder_out in SpeechChunkTransformerEncoder; code adaptation/changes according to the commits on Jul 28, 2020 08 December 2022, 09:00:31 UTC
271bd2a code adaptation/changes according to the commits on Jul 20-25, 2020 08 December 2022, 09:00:31 UTC
33e5bd9 code adaptation/changes according to the commits on Jul 16, 2020 08 December 2022, 09:00:31 UTC
7b491aa code adaptation/changes according to the commits on Jul 14, 2020 08 December 2022, 09:00:31 UTC
fdc42c1 ignore flake8's FileNotFoundError for soft links to kaldi files; code adaptation/changes according to the commits on Jul 8, 2020 08 December 2022, 09:00:31 UTC
591e19d Update Transformer models (#31) * update transformer * initial recipe * fix transformer * add encoder positional embeddings * add more recipes 08 December 2022, 09:00:30 UTC
35b0e71 code adaptation/changes according to the commits on Jun 24-25, 2020; fix validation loss in LSTM models 08 December 2022, 09:00:30 UTC
c09a878 code adaptation/changes according to the commits on Jun 18-23, 2020 08 December 2022, 09:00:30 UTC
08ddb46 remove useless max_{source,target}_positions arguments 08 December 2022, 09:00:30 UTC
f7359ad fix lf-mmi loss; code adaptation/changes according to the commits on May 27 08 December 2022, 09:00:30 UTC
8af6d18 code adaptation/changes according to the commits on May 18 08 December 2022, 09:00:29 UTC
95a109d code adaptation/changes according to the commits on May 10 08 December 2022, 09:00:29 UTC
a7c391e Hybrid ASR code (E2E LF-MMI and cross-entropy) and WSJ examples (#29) 08 December 2022, 09:00:29 UTC
c34a7c2 use EncoderOut for SpeechLSTMEncoder's output; code adaptation/changes according to the commits on Apr 21 08 December 2022, 09:00:29 UTC
c748fde update the qsub script for gpu jobs; code adaptation/changes according to the commits on Apr 16 08 December 2022, 09:00:28 UTC
46dcea9 code adaptation/changes according to the commits on Apr 7 08 December 2022, 09:00:28 UTC
6d0eb55 code adaptation/changes according to the commits on Mar 24-Apr 3, 2020; use data.encoders.{bpe,tokenizer} for wordpiece decode 08 December 2022, 09:00:28 UTC
444510f fix specaug indexing 08 December 2022, 09:00:28 UTC
6366fff code adaptation/changes according to the commits on Mar 11, 2020; change logs/->log/; rename SpeechDataset->AsrDataset, Scp*Dataset->FeatScp*Dataset, score*.sh->score*_e2e.sh; remove validation on train subset from wsj recipe 08 December 2022, 09:00:28 UTC
2084287 SpecAugment (#21) 08 December 2022, 09:00:28 UTC
86346e7 code adaptation/changes according to the commits on Mar 3-10, 2020 08 December 2022, 09:00:28 UTC
01c5fbe code adaptation/changes according to the commits on Feb 27-29, 2020 08 December 2022, 09:00:27 UTC
back to top