Automatic speech recognition (ASR) is vital for very low-resource languages for mitigating the extinction trouble. Chaha is one of the low-resource languages, which suffers from the problem of resource insufficiency a...Automatic speech recognition (ASR) is vital for very low-resource languages for mitigating the extinction trouble. Chaha is one of the low-resource languages, which suffers from the problem of resource insufficiency and some of its phonological, morphological, and orthographic features challenge the development and initiatives in the area of ASR. By considering these challenges, this study is the first endeavor, which analyzed the characteristics of the language, prepared speech corpus, and developed different ASR systems. A small 3-hour read speech corpus was prepared and transcribed. Different basic and rounded phone unit-based speech recognizers were explored using multilingual deep neural network (DNN) modeling methods. The experimental results demonstrated that all the basic phone and rounded phone unit-based multilingual models outperformed the corresponding unilingual models with the relative performance improvements of 5.47% to 19.87% and 5.74% to 16.77%, respectively. The rounded phone unit-based multilingual models outperformed the equivalent basic phone unit-based models with relative performance improvements of 0.95% to 4.98%. Overall, we discovered that multilingual DNN modeling methods are profoundly effective to develop Chaha speech recognizers. Both the basic and rounded phone acoustic units are convenient to build Chaha ASR system. However, the rounded phone unit-based models are superior in performance and faster in recognition speed over the corresponding basic phone unit-based models. Hence, the rounded phone units are the most suitable acoustic units to develop Chaha ASR systems.展开更多
Recent research has shown a burgeoning interest in exploring sparse models for massively Multilingual Neural Machine Translation(MNMT).In this paper,we present a comprehensive survey of this emerging topic.Massively M...Recent research has shown a burgeoning interest in exploring sparse models for massively Multilingual Neural Machine Translation(MNMT).In this paper,we present a comprehensive survey of this emerging topic.Massively MNMT,when based on sparse models,offers significant improvements in parameter efficiency and reduces interference compared to its dense model counterparts.Various methods have been proposed to leverage sparse models for enhancing translation quality.However,the lack of a thorough survey has hindered the identification and further investigation of the most promising approaches.To address this gap,we provide an exhaustive examination of the current research landscape in massively MNMT,with a special emphasis on sparse models.Initially,we categorize the various sparse model-based approaches into distinct classifications.We then delve into each category in detail,elucidating their fundamental modeling principles,core issues,and the challenges they face.Wherever possible,we conduct comparative analyses to assess the strengths and weaknesses of different methodologies.Moreover,we explore potential future research avenues for MNMT based on sparse models.This survey serves as a valuable resource for both newcomers and established experts in the field of MNMT,particularly those interested in sparse model applications.展开更多
The rise of social networking enables the development of multilingual Internet-accessible digital documents in several languages.The digital document needs to be evaluated physically through the Cross-Language Text Su...The rise of social networking enables the development of multilingual Internet-accessible digital documents in several languages.The digital document needs to be evaluated physically through the Cross-Language Text Summarization(CLTS)involved in the disparate and generation of the source documents.Cross-language document processing is involved in the generation of documents from disparate language sources toward targeted documents.The digital documents need to be processed with the contextual semantic data with the decoding scheme.This paper presented a multilingual crosslanguage processing of the documents with the abstractive and summarising of the documents.The proposed model is represented as the Hidden Markov Model LSTM Reinforcement Learning(HMMlstmRL).First,the developed model uses the Hidden Markov model for the computation of keywords in the cross-language words for the clustering.In the second stage,bi-directional long-short-term memory networks are used for key word extraction in the cross-language process.Finally,the proposed HMMlstmRL uses the voting concept in reinforcement learning for the identification and extraction of the keywords.The performance of the proposed HMMlstmRL is 2%better than that of the conventional bi-direction LSTM model.展开更多
文摘Automatic speech recognition (ASR) is vital for very low-resource languages for mitigating the extinction trouble. Chaha is one of the low-resource languages, which suffers from the problem of resource insufficiency and some of its phonological, morphological, and orthographic features challenge the development and initiatives in the area of ASR. By considering these challenges, this study is the first endeavor, which analyzed the characteristics of the language, prepared speech corpus, and developed different ASR systems. A small 3-hour read speech corpus was prepared and transcribed. Different basic and rounded phone unit-based speech recognizers were explored using multilingual deep neural network (DNN) modeling methods. The experimental results demonstrated that all the basic phone and rounded phone unit-based multilingual models outperformed the corresponding unilingual models with the relative performance improvements of 5.47% to 19.87% and 5.74% to 16.77%, respectively. The rounded phone unit-based multilingual models outperformed the equivalent basic phone unit-based models with relative performance improvements of 0.95% to 4.98%. Overall, we discovered that multilingual DNN modeling methods are profoundly effective to develop Chaha speech recognizers. Both the basic and rounded phone acoustic units are convenient to build Chaha ASR system. However, the rounded phone unit-based models are superior in performance and faster in recognition speed over the corresponding basic phone unit-based models. Hence, the rounded phone units are the most suitable acoustic units to develop Chaha ASR systems.
基金supported by the Key Research and Development Program of Yunnan Province(No.202203AA080004)the National Natural Science Foundation of China Youth Found(No.62306210).
文摘Recent research has shown a burgeoning interest in exploring sparse models for massively Multilingual Neural Machine Translation(MNMT).In this paper,we present a comprehensive survey of this emerging topic.Massively MNMT,when based on sparse models,offers significant improvements in parameter efficiency and reduces interference compared to its dense model counterparts.Various methods have been proposed to leverage sparse models for enhancing translation quality.However,the lack of a thorough survey has hindered the identification and further investigation of the most promising approaches.To address this gap,we provide an exhaustive examination of the current research landscape in massively MNMT,with a special emphasis on sparse models.Initially,we categorize the various sparse model-based approaches into distinct classifications.We then delve into each category in detail,elucidating their fundamental modeling principles,core issues,and the challenges they face.Wherever possible,we conduct comparative analyses to assess the strengths and weaknesses of different methodologies.Moreover,we explore potential future research avenues for MNMT based on sparse models.This survey serves as a valuable resource for both newcomers and established experts in the field of MNMT,particularly those interested in sparse model applications.
文摘The rise of social networking enables the development of multilingual Internet-accessible digital documents in several languages.The digital document needs to be evaluated physically through the Cross-Language Text Summarization(CLTS)involved in the disparate and generation of the source documents.Cross-language document processing is involved in the generation of documents from disparate language sources toward targeted documents.The digital documents need to be processed with the contextual semantic data with the decoding scheme.This paper presented a multilingual crosslanguage processing of the documents with the abstractive and summarising of the documents.The proposed model is represented as the Hidden Markov Model LSTM Reinforcement Learning(HMMlstmRL).First,the developed model uses the Hidden Markov model for the computation of keywords in the cross-language words for the clustering.In the second stage,bi-directional long-short-term memory networks are used for key word extraction in the cross-language process.Finally,the proposed HMMlstmRL uses the voting concept in reinforcement learning for the identification and extraction of the keywords.The performance of the proposed HMMlstmRL is 2%better than that of the conventional bi-direction LSTM model.