Automatic recognition of human emotions in a continuous dialog model remains challenging where a speaker’s utterance includes several sentences that may not always carry a single emotion.Limited work with standalone ...Automatic recognition of human emotions in a continuous dialog model remains challenging where a speaker’s utterance includes several sentences that may not always carry a single emotion.Limited work with standalone speech emotion recognition(SER)systems proposed for continuous speech only has been reported.In the recent decade,various effective SER systems have been proposed for discrete speech,i.e.,short speech phrases.It would be more helpful if these systems could also recognize emotions from continuous speech.However,if these systems are applied directly to test emotions from continuous speech,emotion recognition performance would not be similar to that achieved for discrete speech due to the mismatch between training data(from training speech)and testing data(from continuous speech).The problem may possibly be resolved if an existing SER system for discrete speech is enhanced.Thus,in this work the author’s existing effective SER system for multilingual and mixed-lingual discrete speech is enhanced by enriching the cepstral speech feature set with bi-spectral speech features and a unique functional set of Mel frequency cepstral coefficient features derived from a sine filter bank.Data augmentation is applied to combat skewness of the SER system toward certain emotions.Classification using random forest is performed.This enhanced SER system is used to predict emotions from continuous speech with a uniform segmentation method.Due to data scarcity,several audio samples of discrete speech from the SAVEE database that has recordings in a universal language,i.e.,English,are concatenated resulting in multi-emotional speech samples.Anger,fear,sad,and neutral emotions,which are vital during the initial investigation of mentally disordered individuals,are selected to build six categories of multi-emotional samples.Experimental results demonstrate the suitability of the proposed method for recognizing emotions from continuous speech as well as from discrete speech.展开更多
Automated grading of colon biopsy images across all magnifications is challenging because of tailored segmentation and dependent features on each magnification.This work presents a novel approach of robust magnificati...Automated grading of colon biopsy images across all magnifications is challenging because of tailored segmentation and dependent features on each magnification.This work presents a novel approach of robust magnification-independent colon cancer grading framework to distinguish colon biopsy images into four classes:normal,well,moderate,and poor.The contribution of this research is to develop a magnification invariant hybrid feature set comprising cartoon feature,Gabor wavelet,wavelet moments,HSV histogram,color auto-correlogram,color moments,and morphological features that can be used to characterize different grades.Besides,the classifier is modeled as a multiclass structure with six binary class Bayesian optimized random forest(BO-RF)classifiers.This study uses four datasets(two collected from Indian hospitals—Ishita Pathology Center(IPC)of 4X,10X,and 40X and Aster Medcity(AMC)of 10X,20X,and 40X—two benchmark datasets—gland segmentation(GlaS)of 20X and IMEDIATREAT of 10X)comprising multiple microscope magnifications.Experimental results demonstrate that the proposed method outperforms the other methods used for colon cancer grading in terms of accuracy(97.25%-IPC,94.40%-AMC,97.58%-GlaS,99.16%-Imediatreat),sensitivity(0.9725-IPC,0.9440-AMC,0.9807-GlaS,0.9923-Imediatreat),specificity(0.9908-IPC,0.9813-AMC,0.9907-GlaS,0.9971-Imediatreat)and F-score(0.9725-IPC,0.9441-AMC,0.9780-GlaS,0.9923-Imediatreat).The generalizability of the model to any magnified input image is validated by training in one dataset and testing in another dataset,highlighting strong concordance in multiclass classification and evidencing its effective use in the first level of automatic biopsy grading and second opinion.展开更多
基金This work was partially supported by the Research Groups Program(Research Group Number RG-1439-033),under the Deanship of Scientific Research,King Saud University,Riyadh,Saudi Arabia.
文摘Automatic recognition of human emotions in a continuous dialog model remains challenging where a speaker’s utterance includes several sentences that may not always carry a single emotion.Limited work with standalone speech emotion recognition(SER)systems proposed for continuous speech only has been reported.In the recent decade,various effective SER systems have been proposed for discrete speech,i.e.,short speech phrases.It would be more helpful if these systems could also recognize emotions from continuous speech.However,if these systems are applied directly to test emotions from continuous speech,emotion recognition performance would not be similar to that achieved for discrete speech due to the mismatch between training data(from training speech)and testing data(from continuous speech).The problem may possibly be resolved if an existing SER system for discrete speech is enhanced.Thus,in this work the author’s existing effective SER system for multilingual and mixed-lingual discrete speech is enhanced by enriching the cepstral speech feature set with bi-spectral speech features and a unique functional set of Mel frequency cepstral coefficient features derived from a sine filter bank.Data augmentation is applied to combat skewness of the SER system toward certain emotions.Classification using random forest is performed.This enhanced SER system is used to predict emotions from continuous speech with a uniform segmentation method.Due to data scarcity,several audio samples of discrete speech from the SAVEE database that has recordings in a universal language,i.e.,English,are concatenated resulting in multi-emotional speech samples.Anger,fear,sad,and neutral emotions,which are vital during the initial investigation of mentally disordered individuals,are selected to build six categories of multi-emotional samples.Experimental results demonstrate the suitability of the proposed method for recognizing emotions from continuous speech as well as from discrete speech.
基金This work was partially supported by the Research Groups Program(Research Group Number RG-1439-033),under the Deanship of Scientific Research,King Saud University,Riyadh,Saudi Arabia.
文摘Automated grading of colon biopsy images across all magnifications is challenging because of tailored segmentation and dependent features on each magnification.This work presents a novel approach of robust magnification-independent colon cancer grading framework to distinguish colon biopsy images into four classes:normal,well,moderate,and poor.The contribution of this research is to develop a magnification invariant hybrid feature set comprising cartoon feature,Gabor wavelet,wavelet moments,HSV histogram,color auto-correlogram,color moments,and morphological features that can be used to characterize different grades.Besides,the classifier is modeled as a multiclass structure with six binary class Bayesian optimized random forest(BO-RF)classifiers.This study uses four datasets(two collected from Indian hospitals—Ishita Pathology Center(IPC)of 4X,10X,and 40X and Aster Medcity(AMC)of 10X,20X,and 40X—two benchmark datasets—gland segmentation(GlaS)of 20X and IMEDIATREAT of 10X)comprising multiple microscope magnifications.Experimental results demonstrate that the proposed method outperforms the other methods used for colon cancer grading in terms of accuracy(97.25%-IPC,94.40%-AMC,97.58%-GlaS,99.16%-Imediatreat),sensitivity(0.9725-IPC,0.9440-AMC,0.9807-GlaS,0.9923-Imediatreat),specificity(0.9908-IPC,0.9813-AMC,0.9907-GlaS,0.9971-Imediatreat)and F-score(0.9725-IPC,0.9441-AMC,0.9780-GlaS,0.9923-Imediatreat).The generalizability of the model to any magnified input image is validated by training in one dataset and testing in another dataset,highlighting strong concordance in multiclass classification and evidencing its effective use in the first level of automatic biopsy grading and second opinion.