[1] The History of Speech Recognition to the Year 2030, Awni Hannun https://arxiv.org/pdf/2108.00084v1.pdf
[2] Vassil Panayotov, Guoguo Chen, Daniel Povey, and Sanjeev Khudanpur. Librispeech:
an asr corpus based on public domain audio books. ICASSP, pages 5206-5210. IEEE, 2015.
[3] Daniel Povey, Arnab Ghoshal, Gilles Boulianne, Lukas Burget, Ondrej Glembek, Nagendra Goel, Mirko Hannemann, Petr Motlicek, Yanmin Qian, Petr Schwarz, et al.
The kaldi speech recognition toolkit. In IEEE 2011 workshop on automatic speech recognition and understanding, number CONF. IEEE Signal Processing Society, 2011.
[4] Geoffrey Hinton, Li Deng, Dong Yu, George E Dahl, Abdel-rahman Mohamed, Navdeep Jaitly, Andrew Senior, Vincent Vanhoucke, Patrick Nguyen, Tara N Sainath, et al. Deep neural networks for acoustic modeling in speech recognition: The shared views of four research groups. IEEE Signal processing magazine, 29(6):82-97, 2012.
[5] Awni Hannun, Carl Case, Jared Casper, Bryan Catanzaro, Greg Diamos, Erich Elsen, Ryan Prenger, Sanjeev Satheesh, Shubho Sengupta, Adam Coates, et al. Deep speech: Scaling up end-to-end speech recognition. https://arxiv.org/pdf/1412.5567.pdf
[6] Jan Chorowski, Dzmitry Bahdanau, Dmitriy Serdyuk, Kyunghyun Cho, and Yoshua Bengio. Attention-based models for speech recognition. NIPS Volume 1, pages 577-585,
2015.
[7] Yanzhang He, Tara N Sainath, Rohit Prabhavalkar, Ian McGraw, Raziel Alvarez, Ding Zhao, David Rybach, Anjuli Kannan, Yonghui Wu, Ruoming Pang, et al. Streaming end-to-end speech recognition for mobile devices. ICASSP, pages 6381-6385. IEEE, 2019.
[8] https://paperswithcode.com/sota/speech-recognition-on-librispeech-test-clean
[9] https://arxiv.org/pdf/1912.07875.pdf
[10] https://catalog.ldc.upenn.edu/LDC97S62
[11] https://catalog.ldc.upenn.edu/LDC93S6A
[12] https://catalog.ldc.upenn.edu/LDC93S1
[13] https://aihub.or.kr/aihub-data/natural-language/about
[14] https://arxiv.org/abs/1904.05862
[15] http://aihub-competition.or.kr/hangeul