- Flask==0.12.2
- gunicorn==19.7.1
- librosa==0.5.0
- numpy==1.14.2
- pylint==1.8.4
- scikit-learn==0.19.1
- scipy==1.0.1
- six==1.11.0
- sugartensor==1.0.0.2
- tensorflow==1.0.0
- Werkzeug==0.14.1
- Scrapy==1.5.0
- Data sources:
- vais
- vivos
- KTSpeech (Chị Đào Chị Lý, Kẻ làm người chịu)
- audiobook (Trúng số độc đắc)
- King-ASR-M-008-Promotion
- Processed data:
- Train set: 45182 samples ( > 41hrs)
- Test set: 5006 samples ( > 4hrs)
- FINAL_DATA
- Preprocessed_data
- Scripts:
gather_data.py: Gathering all data from above sets then rearranging and preprocessing them. After gathering, data will be stored in/data processing/asset/data/FINAL_DATAfolder.preprocess.py: Preprocessing gathered data to get mfcc features. After preprocessing, preprocessed data will be stored in/data processing/asset/data/preprocessfolder.
- Preprocessing:
- Run this to gather data and preprocess them (You might need to modify codes):
/data processing$ python gather_data.py - Or run this if you've already have the
FINAL_DATAto preprocess them:/data processing$ python preprocess.py
- Run this to gather data and preprocess them (You might need to modify codes):
- Architecture:

- Training:
- Put
preprocess(preprocessed data) folder in/training/asset/data. - You can modify codes in
train.pyandmodel.pyto run in your case. - Before training, remember to adjust the
max_ep(max epoch) parameter intrain.py(default learning rate is 0.001, you can modify learning rate by thelrparameter intrain.py). - Run this to begin training (trained models will be stored in
/training/asset/train):/training$ python train.py - Run this to test your latest model with the preprocessed test set (this will create a
loss.csvfile to store testing loss information and store the best model in thebest_modelfolder):/training$ python test.py
- Put
- Data source: VnExpress
- Categories: Thời sự, Góc nhìn, Thế giới, Kinh doanh, Giải trí, Thể thao, Pháp luật, Giáo dục, Sức khỏe, Gia đình, Du lịch, Khoa học, Số hóa, Xe, Cộng đồng, Tâm sự.
- Number of articles: 198,139
- Number of different words: 147,378
- Total number of words: 77,907,100
- Top 10 common words:

- Running crawler:
- You might need to modify codes in
vnexpress_spider.pyto run in your case. - Run this to crawl from listed categories above (500 pages each category, 100 pages for Góc nhìn), this will return a txt file and a csv file in
/language model/crawlers/vnexpress_crawler:/language model/crawlers/vnexpress_crawler$ scrapy crawl --nolog vnexpress_spider -o vnexpress.csv -t csv
- You might need to modify codes in
- Correcting spell:
- Put this VnExpress data into
/language modelfolder to test the precrawled data withcorrect_spell.py. - You might need to modify codes in
correct_spell.pyto test the language model.
- Put this VnExpress data into
- Trained Model:
- Epoch: 20
- Step: 59283
- Training loss: 11.84
- Testing loss: 19.09
- Running Web App:
- Put this VnExpress data into
/web appfolder to run the Web App. - Run this to open the Web App on your localhost, you can test our model there:
/web app$ FLASK_APP=app.py flask run
- Put this VnExpress data into
- Heroku Web App:
- You can find our Web App here: https://thesis-vnmese-s2t-api-heroku.herokuapp.com/
- But due to the lack of memory on our Heroku server, size of the language model using on the site is much smaller than the one running on your localhost here.
- Language Model: Upgrading from single word to compound word.
- Model libraries: Upgrading to fit with newer version of libraries.
- Model Architecture: Modifying the model architecture to get better result.
- Speech-to-Text-WaveNet : End-to-end sentence level English speech recognition using DeepMind's WaveNet
- WaveNet: A Generative Model for Raw Audio
Kim and Park. Speech-to-Text-WaveNet. 2016. GitHub repository. https://github.com/buriburisuri/.