期刊文献+
共找到3篇文章
< 1 >
每页显示 20 50 100
Explainable Diabetic Retinopathy Detection Using a Distributed CNN and LightGBM Framework
1
作者 Pooja Bidwai Shilpa Gite +1 位作者 Biswajeet Pradhan Abdullah Almari 《Computers, Materials & Continua》 2025年第8期2645-2676,共32页
Diabetic Retinopathy(DR)is a critical disorder that affects the retina due to the constant rise in diabetics and remains the major cause of blindness across the world.Early detection and timely treatment are essential... Diabetic Retinopathy(DR)is a critical disorder that affects the retina due to the constant rise in diabetics and remains the major cause of blindness across the world.Early detection and timely treatment are essential to mitigate the effects of DR,such as retinal damage and vision impairment.Several conventional approaches have been proposed to detect DR early and accurately,but they are limited by data imbalance,interpretability,overfitting,convergence time,and other issues.To address these drawbacks and improve DR detection accurately,a distributed Explainable Convolutional Neural network-enabled Light Gradient Boosting Machine(DE-ExLNN)is proposed in this research.The model combines an explainable Convolutional Neural Network(CNN)and Light Gradient Boosting Machine(LightGBM),achieving highly accurate outcomes in DR detection.LightGBM serves as the detection model,and the inclusion of an explainable CNN addresses issues that conventional CNN classifiers could not resolve.A custom dataset was created for this research,containing both fundus and OCTA images collected from a realtime environment,providing more accurate results compared to standard conventional DR datasets.The custom dataset demonstrates notable accuracy,sensitivity,specificity,and Matthews Correlation Coefficient(MCC)scores,underscoring the effectiveness of this approach.Evaluations against other standard datasets achieved an accuracy of 93.94%,sensitivity of 93.90%,specificity of 93.99%,and MCC of 93.88%for fundus images.For OCTA images,the results obtained an accuracy of 95.30%,sensitivity of 95.50%,specificity of 95.09%,andMCC of 95%.Results prove that the combination of explainable CNN and LightGBMoutperforms othermethods.The inclusion of distributed learning enhances the model’s efficiency by reducing time consumption and complexity while facilitating feature extraction. 展开更多
关键词 Diabetic retinopathy explainable convolutional neural network light gradient boosting machine fundus image custom dataset
在线阅读 下载PDF
Audiovisual speech recognition based on a deep convolutional neural network 被引量:2
2
作者 Shashidhar Rudregowda Sudarshan Patilkulkarni +2 位作者 Vinayakumar Ravi Gururaj H.L. Moez Krichen 《Data Science and Management》 2024年第1期25-34,共10页
Audiovisual speech recognition is an emerging research topic.Lipreading is the recognition of what someone is saying using visual information,primarily lip movements.In this study,we created a custom dataset for India... Audiovisual speech recognition is an emerging research topic.Lipreading is the recognition of what someone is saying using visual information,primarily lip movements.In this study,we created a custom dataset for Indian English linguistics and categorized it into three main categories:(1)audio recognition,(2)visual feature extraction,and(3)combined audio and visual recognition.Audio features were extracted using the mel-frequency cepstral coefficient,and classification was performed using a one-dimension convolutional neural network.Visual feature extraction uses Dlib and then classifies visual speech using a long short-term memory type of recurrent neural networks.Finally,integration was performed using a deep convolutional network.The audio speech of Indian English was successfully recognized with accuracies of 93.67%and 91.53%,respectively,using testing data from 200 epochs.The training accuracy for visual speech recognition using the Indian English dataset was 77.48%and the test accuracy was 76.19%using 60 epochs.After integration,the accuracies of audiovisual speech recognition using the Indian English dataset for training and testing were 94.67%and 91.75%,respectively. 展开更多
关键词 Audiovisual speech recognition Custom dataset 1D Convolution neural network(CNN) Deep CNN(DCNN) Long short-term memory(LSTM) LIPREADING Dlib Mel-frequency cepstral coefficient(MFCC)
在线阅读 下载PDF
Indian traffic sign detection and recognition using deep learning 被引量:1
3
作者 Rajesh Kannan Megalingam Kondareddy Thanigundala +2 位作者 Sreevatsava Reddy Musani Hemanth Nidamanuru Lokesh Gadde 《International Journal of Transportation Science and Technology》 2023年第3期683-699,共17页
Traffic signs play a crucial role in managing traffic on the road,disciplining the drivers,thereby preventing injury,property damage,and fatalities.Traffic sign management with automatic detection and recognition is v... Traffic signs play a crucial role in managing traffic on the road,disciplining the drivers,thereby preventing injury,property damage,and fatalities.Traffic sign management with automatic detection and recognition is very much part of any Intelligent Transportation System(ITS).In this era of self-driving vehicles,calls for automatic detection and recognition of traffic signs cannot be overstated.This paper presents a deep-learning-based autonomous scheme for cognizance of traffic signs in India.The automatic traffic sign detection and recognition was conceived on a Convolutional Neural Network(CNN)-Refined Mask R-CNN(RM R-CNN)-based end-to-end learning.The proffered concept was appraised via an innovative dataset comprised of 6480 images that constituted 7056 instances of Indian traffic signs grouped into 87 categories.We present several refinements to the Mask R-CNN model both in architecture and data augmentation.We have considered highly challenging Indian traffic sign categories which are not yet reported in previous works.The dataset for training and testing of the proposed model is obtained by capturing images in real-time on Indian roads.The evaluation results indicate lower than 3%error.Furthermore,RM R-CNN’s performance was compared with the conventional deep neural network architectures such as Fast R-CNN and Mask R-CNN.Our proposed model achieved precision of 97.08%which is higher than precision obtained by Mask R-CNN and Faster RCNN models. 展开更多
关键词 Refined Mask R-CNN Fast R-CNN Data augmentation PRE-PROCESSING Custom dataset Indian Traffic Sign
在线阅读 下载PDF
上一页 1 下一页 到第
使用帮助 返回顶部