[2]
Hinton G, Deng L, Yu D, et al. Deep neural networks for acoustic modeling in speech recognition. IEEE Signal Process Mag 2012; •••: 29.
[3]
Khanna T, Nand P, Bali V. Permissioned Blockchain Model for End-to-End Trackability in Supply Chain Management 2020.
[4]
Aggarwal D, Bali V, Mittal S. An Insight into Machine Learning Techniques for Predictive Analysis and Feature Selection
[5]
Gangwar S, Bali V, Kumar A. Comparative Analysis of Wind Speed Forecasting Using LSTM and SVM. EAI Endorsed Transactions on Scalable Information Systems 2020; 7(25)159407
[9]
Hernandez F, Nguyen N, Ghannay S, Tomashenko N, Estève Y. TED-LIUM 3: Twice as much data and corpus repartition for experiments on speaker adaptation. International Conference on Speech and Computer. 198-08.
[10]
Barker J, Watanabe S, Vincent E, Trmal J. 2018.
[11]
Panayotov V, Chen G, Povey D, Khudanpur S. Librispeech: An ASR corpus based on public domain audio books. 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). 5206-10.
[12]
Hain T, Woodland PC, Evermann G, Povey D. The cu-htk Proc Speech Transcription Workshop.
[13]
Chu SMD, Povey D. IEEE International Conference on Acoustics, Speech and Signal Processing.
[22]
Dash D, Kim M, Teplansky K, Wang J. Automatic speech recognition with articulatory information and a unified dictionary for Hindi, Marathi, Bengali, and Oriya. Proceedings of the Annual Conference of the International Speech Communication Association.
[23]
Chellapriyadharshini M, Toffy A, Ramasubramanian V. 2018.
[24]
Grézl F, Karafiat M, Janda M. Study of probabilistic and bottle-neck features in multilingual environment. IEEE Workshop on Automatic Speech Recognition & Understanding.
[25]
Sahraeian R, Van Compernolle D. Cross-Entropy Training of DNN Ensemble Acoustic Models for Low-Resource ASR. IEEE/ACM Trans Audio Speech Lang Process 2018; 26(11): 1991-2001. [TASLP]
[26]
Lazaridis A, Himawan I, Motlicek P, Mporas I, Garner PN. Investigating cross-lingual multi-level adaptive networks: The importance of the correlation of source and target languages, No. CONF 2016.
[27]
Vu NT, Schultz T. Multilingual multilayer perceptron for rapid language adaptation between and across language families. Interspeech 2013.
[28]
Tüske Z, Pinto J, Willett D, Schlüter R. Investigation on cross-and multilingual MLP features under matched and mismatched acoustical conditions IEEE International Conference on Acoustics, Speech and Signal Processing.
[29]
Huang JT, Li J, Yu D, Deng L, Gong Y. Cross-language knowledge transfer using multilingual deep neural network with shared hidden layers IEEE International Conference on Acoustics, Speech and Signal Processing.
[30]
Gales MJ, Knill KM, Ragni A, Rath SP. Speech recognition and keyword spotting for low-resource languages: BABEL project research at CUED. Spoken Language Technologies for Under-Resourced Languages 2014.
[31]
Li X, Dalmia S, Black AW, Metze F. 1908.
[32]
Kawakami K, Wang L, Dyer C, Blunsom P, Oord AV. 2001.
[33]
Biswas A, Menon R, van der Westhuizen E, Niesler T. 1907.
[35]
Veselý K, Karafiát M, Grézl F, Janda M, Egorova E. The language-independent bottleneck features. IEEE Spoken Language Technology Workshop (SLT).
[36]
Dua M, Aggarwal RK, Biswas M. Discriminatively trained continuous Hindi speech recognition system using interpolated recurrent neural network language modeling. Neural Comput Appl 2018; •••: 1-9.
[38]
Passricha V, Aggarwal RK. Convolutional support vector machines for speech recognition. Int J Speech Technol 2018; •••: 1-9.
[40]
Vydana HK, Gurugubelli K, Vegesna VV, et al. An Exploration towards joint acoustic modeling for indian languages: IIIT-H submission for low resource speech recognition challenge for indian languages. INTERSPEECH 2018; pp. 3192-6.
[41]
Shetty VM, Sharon RA, et al. Articulatory and stacked bottleneck features for low resource speech recognition. 2018.
[42]
Sailor HB, Krishna MV, et al. 2018.
[43]
Fathima N, Patel T, Mahima C, et al. 2018.
[44]
Ghoshal A, Swietojanski P, Renals S. Multilingual training of deep neural networks IEEE International Conference on Acoustics, Speech and Signal Processing.
[45]
Ni C, Leung CC, Wang L, et al. Efficient methods to train multilingual bottleneck feature extractors for low resource keyword search. 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). 5650-4.
[46]
Zhou S, Zhao Y, Xu S, et al. Multilingual recurrent neural networks with residual learning for low-resource speech recognition. 2017.
[48]
Tong S, Garner P N, Bourlard H. An investigation of deep neural networks for multilingual speech recognition training and adaptation 2017.
[50]
Biswas A, de Wet F, van der Westhuizen E, et al. Multilingual neural network acoustic modelling for ASR of under-resourced English-isiZulu code-switched speech. 2018.
[51]
Cui J, Kingsbury B, Ramabhadran B, et al. Multilingual representations for low resource speech recognition and keyword search. 2015 IEEE Workshop on Automatic Speech Recognition and Understanding (ASRU). 259-66.
[52]
Lin H, Deng L, Yu D, et al. A study on multilingual acoustic modeling for large vocabulary ASR. 2009 IEEE International Conference on Acoustics, Speech and Signal Processing. 4333-36.
[53]
Amodei D, Ananthanarayanan S, Anubhai R, et al. Deep speech 2: End-to-end speech recognition in english and mandarin. International Conference on Machine Learning. 173-82.
[54]
Seide F, Li G, Yu D. Conversational speech transcription using context-dependent deep neural networks.
[55]
Abdel-Hamid O, Mohamed AR, Jiang H, et al. Applying convolutional neural networks concepts to hybrid NN-HMM model for speech recognition. 2012 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). 4277-80.
[56]
Yu D, Jinyu L. Recent progresses in deep learning based acoustic models 2017.
[57]
Peddinti V, Povey D, Khudanpur S. A time delay neural network architecture for efficient modeling of long temporal contexts. Sixteenth Annual Conference of the International Speech Communication Association.
[59]
Peddinti V, Chen G, Povey D, et al. Reverberation robust acoustic modeling using i-vectors with time delay neural networks. Sixteenth Annual Conference of the International Speech Communication Association.
[60]
Kreyssig FL, Zhang C, Woodland PC. Improved TDNNs using deep kernels and frequency dependent Grid-RNNs. IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[61]
Peddinti V, Chen G, Manohar V, et al. Jhu aspire system: Robust lvcsr with tdnns, ivector adaptation and rnn-lms. 2015 IEEE Workshop on Automatic Speech Recognition and Understanding (ASRU). 539-46.
[62]
Sak H, Senior A, Beaufays F, et al. 2015.
[64]
Sak H, Senior A, Beaufays F. Long short-term memory recurrent neural network architectures for large scale acoustic modeling. 2014.
[65]
Feng S, Lee T. Improving cross-lingual knowledge transferability using multilingual TDNN-BLSTM with language-dependent pre-final layer. 2018.
[66]
Smit P, Gangireddy SR, Enarvi S, et al. Aalto. 2017; pp. 2017 IEEE Automatic Speech Recognition and Understanding Workshop (ASRU). 338-45.
[67]
Ali A, Vogel S, Renals S. IEEE Automatic Speech Recognition and Understanding Workshop (ASRU).
[69]
Eastwood M, Jayne C. Restricted Boltzmann machines for pre-training deep Gaussian networks. International Joint Conference on Neural Networks (IJCNN).
[70]
Veselý K, et al. Sequence-discriminative training of deep neural networks. Interspeech 2013.
[71]
Ko T, Peddinti V, Povey D, Khudanpur S. Audio augmentation for speech recognition. Sixteenth annual conference of the international speech communication association.
[73]
Xu H, Li K, Wang Y, et al. Neural network language modeling with letter-based features and importance sampling. 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). 6109-13.
[74]
Samudravijaya K, Rao PV, Agrawal SS. Hindi speech database Sixth international conference on spoken language processing.