* add ctcloss training scripts. * update compute_det_ctc * fix typo. * add fsmn model, can use pretrained kws model from modelscope. * Add streaming detection of CTC model. Add CTC model onnx export. Add CTC model's result in README; For now CTC model runtime is not supported yet. * QA run.sh, maxpooling training scripts is compatible. Ready to PR. * Add a streaming kws demo, support fsmn online forward * fix typo. * Align Stream FSMN and Non-Stream FSMN, both in feature extraction and model forward. * fix repeat activation, add a interval restrict. * fix timestamp when subsampling!=1. * fix flake8, update training script and README, give pretrained ckpt. * fix quickcheck and flake8 * Add realtime CTC-KWS demo in README. --------- Co-authored-by: dujing <dujing@xmov.ai>
176 lines
5.5 KiB
Bash
176 lines
5.5 KiB
Bash
#!/bin/bash
|
|
# Copyright 2021 Binbin Zhang(binbzha@qq.com)
|
|
# 2023 Jing Du(thuduj12@163.com)
|
|
|
|
. ./path.sh
|
|
|
|
stage=$1
|
|
stop_stage=$2
|
|
num_keywords=2599
|
|
|
|
config=conf/fsmn_ctc.yaml
|
|
norm_mean=true
|
|
norm_var=true
|
|
gpus="0"
|
|
|
|
checkpoint=
|
|
dir=exp/fsmn_ctc
|
|
average_model=true
|
|
num_average=30
|
|
if $average_model ;then
|
|
score_checkpoint=$dir/avg_${num_average}.pt
|
|
else
|
|
score_checkpoint=$dir/final.pt
|
|
fi
|
|
|
|
download_dir=/mnt/52_disk/back/DuJing/data/nihaowenwen # your data dir
|
|
|
|
. tools/parse_options.sh || exit 1;
|
|
window_shift=50
|
|
|
|
if [ ${stage} -le -2 ] && [ ${stop_stage} -ge -2 ]; then
|
|
echo "Download and extracte all datasets"
|
|
local/mobvoi_data_download.sh --dl_dir $download_dir
|
|
fi
|
|
|
|
|
|
if [ ${stage} -le -1 ] && [ ${stop_stage} -ge -1 ]; then
|
|
echo "Preparing datasets..."
|
|
mkdir -p dict
|
|
echo "<filler> -1" > dict/words.txt
|
|
echo "Hi_Xiaowen 0" >> dict/words.txt
|
|
echo "Nihao_Wenwen 1" >> dict/words.txt
|
|
|
|
for folder in train dev test; do
|
|
mkdir -p data/$folder
|
|
for prefix in p n; do
|
|
mkdir -p data/${prefix}_$folder
|
|
json_path=$download_dir/mobvoi_hotword_dataset_resources/${prefix}_$folder.json
|
|
local/prepare_data.py $download_dir/mobvoi_hotword_dataset $json_path \
|
|
data/${prefix}_$folder
|
|
done
|
|
cat data/p_$folder/wav.scp data/n_$folder/wav.scp > data/$folder/wav.scp
|
|
cat data/p_$folder/text data/n_$folder/text > data/$folder/text
|
|
rm -rf data/p_$folder data/n_$folder
|
|
done
|
|
fi
|
|
|
|
if [ ${stage} -le -0 ] && [ ${stop_stage} -ge -0 ]; then
|
|
# Here we Use Paraformer Large(https://www.modelscope.cn/models/damo/speech_paraformer-large_asr_nat-zh-cn-16k-common-vocab8404-pytorch/summary)
|
|
# to transcribe the negative wavs, and upload the transcription to modelscope.
|
|
git clone https://www.modelscope.cn/datasets/thuduj12/mobvoi_kws_transcription.git
|
|
for folder in train dev test; do
|
|
if [ -f data/$folder/text ];then
|
|
mv data/$folder/text data/$folder/text.label
|
|
fi
|
|
cp mobvoi_kws_transcription/$folder.text data/$folder/text
|
|
done
|
|
|
|
# and we also copy the tokens and lexicon that used in
|
|
# https://modelscope.cn/models/damo/speech_charctc_kws_phone-xiaoyun/summary
|
|
cp mobvoi_kws_transcription/tokens.txt data/tokens.txt
|
|
cp mobvoi_kws_transcription/lexicon.txt data/lexicon.txt
|
|
|
|
fi
|
|
|
|
if [ ${stage} -le 1 ] && [ ${stop_stage} -ge 1 ]; then
|
|
echo "Compute CMVN and Format datasets"
|
|
tools/compute_cmvn_stats.py --num_workers 16 --train_config $config \
|
|
--in_scp data/train/wav.scp \
|
|
--out_cmvn data/train/global_cmvn
|
|
|
|
for x in train dev test; do
|
|
tools/wav_to_duration.sh --nj 8 data/$x/wav.scp data/$x/wav.dur
|
|
|
|
# Here we use tokens.txt and lexicon.txt to convert txt into index
|
|
tools/make_list.py data/$x/wav.scp data/$x/text \
|
|
data/$x/wav.dur data/$x/data.list \
|
|
--token_file data/tokens.txt \
|
|
--lexicon_file data/lexicon.txt
|
|
done
|
|
fi
|
|
|
|
if [ ${stage} -le 2 ] && [ ${stop_stage} -ge 2 ]; then
|
|
|
|
echo "Use the base model from modelscope"
|
|
if [ ! -d speech_charctc_kws_phone-xiaoyun ] ;then
|
|
git lfs install
|
|
git clone https://www.modelscope.cn/damo/speech_charctc_kws_phone-xiaoyun.git
|
|
fi
|
|
checkpoint=speech_charctc_kws_phone-xiaoyun/train/base.pt
|
|
cp speech_charctc_kws_phone-xiaoyun/train/feature_transform.txt.80dim-l2r2 data/global_cmvn.kaldi
|
|
|
|
echo "Start training ..."
|
|
mkdir -p $dir
|
|
cmvn_opts=
|
|
$norm_mean && cmvn_opts="--cmvn_file data/global_cmvn.kaldi"
|
|
$norm_var && cmvn_opts="$cmvn_opts --norm_var"
|
|
num_gpus=$(echo $gpus | awk -F ',' '{print NF}')
|
|
|
|
torchrun --standalone --nnodes=1 --nproc_per_node=$num_gpus \
|
|
wekws/bin/train.py --gpus $gpus \
|
|
--config $config \
|
|
--train_data data/train/data.list \
|
|
--cv_data data/dev/data.list \
|
|
--model_dir $dir \
|
|
--num_workers 8 \
|
|
--num_keywords $num_keywords \
|
|
--min_duration 50 \
|
|
--seed 666 \
|
|
$cmvn_opts \
|
|
${checkpoint:+--checkpoint $checkpoint}
|
|
fi
|
|
|
|
if [ ${stage} -le 3 ] && [ ${stop_stage} -ge 3 ]; then
|
|
echo "Do model average, Compute FRR/FAR ..."
|
|
if $average_model; then
|
|
python wekws/bin/average_model.py \
|
|
--dst_model $score_checkpoint \
|
|
--src_path $dir \
|
|
--num ${num_average} \
|
|
--val_best
|
|
fi
|
|
result_dir=$dir/test_$(basename $score_checkpoint)
|
|
mkdir -p $result_dir
|
|
stream=true # we detect keyword online with ctc_prefix_beam_search
|
|
score_prefix=""
|
|
if $stream ; then
|
|
score_prefix=stream_
|
|
fi
|
|
python wekws/bin/${score_prefix}score_ctc.py \
|
|
--config $dir/config.yaml \
|
|
--test_data data/test/data.list \
|
|
--gpu 0 \
|
|
--batch_size 256 \
|
|
--checkpoint $score_checkpoint \
|
|
--score_file $result_dir/score.txt \
|
|
--num_workers 8 \
|
|
--keywords "\u55e8\u5c0f\u95ee,\u4f60\u597d\u95ee\u95ee" \
|
|
--token_file data/tokens.txt \
|
|
--lexicon_file data/lexicon.txt
|
|
|
|
python wekws/bin/compute_det_ctc.py \
|
|
--keywords "\u55e8\u5c0f\u95ee,\u4f60\u597d\u95ee\u95ee" \
|
|
--test_data data/test/data.list \
|
|
--window_shift $window_shift \
|
|
--step 0.001 \
|
|
--score_file $result_dir/score.txt \
|
|
--token_file data/tokens.txt \
|
|
--lexicon_file data/lexicon.txt
|
|
fi
|
|
|
|
|
|
if [ ${stage} -le 4 ] && [ ${stop_stage} -ge 4 ]; then
|
|
jit_model=$(basename $score_checkpoint | sed -e 's:.pt$:.zip:g')
|
|
onnx_model=$(basename $score_checkpoint | sed -e 's:.pt$:.onnx:g')
|
|
# For now, FSMN can not export to JITScript
|
|
# python wekws/bin/export_jit.py \
|
|
# --config $dir/config.yaml \
|
|
# --checkpoint $score_checkpoint \
|
|
# --jit_model $dir/$jit_model
|
|
python wekws/bin/export_onnx.py \
|
|
--config $dir/config.yaml \
|
|
--checkpoint $score_checkpoint \
|
|
--onnx_model $dir/$onnx_model
|
|
fi
|