In first aid, traditional information interchange has numerous shortcomings. For example, delayed information and disorganized departmental communication cause patients to miss out on critical rescue time. Information...In first aid, traditional information interchange has numerous shortcomings. For example, delayed information and disorganized departmental communication cause patients to miss out on critical rescue time. Information technology is becoming more and more mature, and as a result, its use across numerous industries is now standard. China is still in the early stages of developing its integration of emergency medical services with modern information technology;despite our progress, there are still numerous obstacles and constraints to overcome. Our goal is to integrate information technology into every aspect of emergency patient care, offering robust assistance for both patient rescue and the efforts of medical personnel. Information may be communicated in a fast, multiple, and effective manner by utilizing modern information technology. This study aims to examine the current state of this field’s development, current issues, and the field’s future course of development.展开更多
Lower back pain is one of the most common medical problems in the world and it is experienced by a huge percentage of people everywhere.Due to its ability to produce a detailed view of the soft tissues,including the s...Lower back pain is one of the most common medical problems in the world and it is experienced by a huge percentage of people everywhere.Due to its ability to produce a detailed view of the soft tissues,including the spinal cord,nerves,intervertebral discs,and vertebrae,Magnetic Resonance Imaging is thought to be the most effective method for imaging the spine.The semantic segmentation of vertebrae plays a major role in the diagnostic process of lumbar diseases.It is difficult to semantically partition the vertebrae in Magnetic Resonance Images from the surrounding variety of tissues,including muscles,ligaments,and intervertebral discs.U-Net is a powerful deep-learning architecture to handle the challenges of medical image analysis tasks and achieves high segmentation accuracy.This work proposes a modified U-Net architecture namely MU-Net,consisting of the Meijering convolutional layer that incorporates the Meijering filter to perform the semantic segmentation of lumbar vertebrae L1 to L5 and sacral vertebra S1.Pseudo-colour mask images were generated and used as ground truth for training the model.The work has been carried out on 1312 images expanded from T1-weighted mid-sagittal MRI images of 515 patients in the Lumbar Spine MRI Dataset publicly available from Mendeley Data.The proposed MU-Net model for the semantic segmentation of the lumbar vertebrae gives better performance with 98.79%of pixel accuracy(PA),98.66%of dice similarity coefficient(DSC),97.36%of Jaccard coefficient,and 92.55%mean Intersection over Union(mean IoU)metrics using the mentioned dataset.展开更多
Terms of intelligence in 20th and 21th century mean the methods of automatic extraction, analysis, interpretation and use of information. Thus, the intelligence services in the future created an electronic database in...Terms of intelligence in 20th and 21th century mean the methods of automatic extraction, analysis, interpretation and use of information. Thus, the intelligence services in the future created an electronic database in which to their being classified intelligence products, users could choose between the latter themselves relevant information. The EU (European Union) that activities are carried out from at least in year 1996, terrorist attacks in year 200l is only accelerating. Proposals to increase surveillance and international cooperation in this field have been drawn up before September 11 2011. On the Web you can fmd a list of networks (Cryptome, 2011), which could be connected, or are under the control of the security service--NSA (National Security Agency). United States of America in year 1994 enacted a law for telephone communication--Digital Telephony Act, which would require manufacturers of telecommunications equipment, leaving some security holes for control. In addition, we monitor the Internet and large corporations. The example of the United States of America in this action reveals the organization for electronic freedoms against a telecom company that the NSA illegally gains access to data on information technology users and Internet telephony.展开更多
In the rapidly evolving landscape of intelligent transportation systems,the security and authenticity of vehicular communication have emerged as critical challenges.As vehicles become increasingly interconnected,the n...In the rapidly evolving landscape of intelligent transportation systems,the security and authenticity of vehicular communication have emerged as critical challenges.As vehicles become increasingly interconnected,the need for robust authentication mechanisms to safeguard against cyber threats and ensure trust in an autonomous ecosystem becomes essential.On the other hand,using intelligence in the authentication system is a significant attraction.While existing surveys broadly address vehicular security,a critical gap remains in the systematic exploration of Deep Learning(DL)-based authentication methods tailored to these communication paradigms.This survey fills that gap by offering a comprehensive analysis of DL techniques—including supervised,unsupervised,reinforcement,and hybrid learning—for vehicular authentication.This survey highlights novel contributions,such as a taxonomy of DL-driven authentication protocols,real-world case studies,and a critical evaluation of scalability and privacy-preserving techniques.Additionally,this paper identifies unresolved challenges,such as adversarial resilience and real-time processing constraints,and proposes actionable future directions,including lightweight model optimization and blockchain integration.By grounding the discussion in concrete applications,such as biometric authentication for driver safety and adaptive key management for infrastructure security,this survey bridges theoretical advancements with practical deployment needs,offering a roadmap for next-generation secure intelligent vehicular ecosystems for the modern world.展开更多
Purpose:Generally,the scientific comparison has been done with the help of the overall impact of scholars.Although it is very easy to compare scholars,but how can we assess the scientific impact of scholars who have d...Purpose:Generally,the scientific comparison has been done with the help of the overall impact of scholars.Although it is very easy to compare scholars,but how can we assess the scientific impact of scholars who have different research careers?It is very obvious,the scholars may gain a high impact if they have more research experience or have spent more time(in terms of research career in a year).Then we cannot compare two scholars who have different research careers.Many bibliometrics indicators address the time-span of scholars.In this series,the h-index sequence and EM/EM’-index sequence have been introduced for assessment and comparison of the scientific impact of scholars.The h-index sequence,EM-index sequence,and EM’-index sequence consider the yearly impact of scholars,and comparison is done by the index value along with their component value.The time-series indicators fail to give a comparative analysis between senior and junior scholars if there is a huge difference in both scholars’research careers.Design/methodology/approach:We have proposed the cumulative index calculation method to appraise the scientific impact of scholars till that age and tested it with 89 scholars data.Findings:The proposed mechanism is implemented and tested on 89 scholars’publication data,providing a clear difference between the scientific impact of two scholars.This also helps in predicting future prominent scholars based on their research impact.Research limitations:This study adopts a simplistic approach by assigning equal credit to all authors,regardless of their individual contributions.Further,the potential impact of career breaks on research productivity is not taken into account.These assumptions may limit the generalizability of our findings Practical implications:The proposed method can be used by respected institutions to compare their scholars impact.Funding agencies can also use it for similar purposes.Originality/value:This research adds to the existing literature by introducing a novel methodology for comparing the scientific impact of scholars.The outcomes of this research have notable implications for the development of more precise and unbiased research assessment frameworks,enabling a more equitable evaluation of scholarly contributions.展开更多
Distributed denial of service(DDoS)attacks are common network attacks that primarily target Internet of Things(IoT)devices.They are critical for emerging wireless services,especially for applications with limited late...Distributed denial of service(DDoS)attacks are common network attacks that primarily target Internet of Things(IoT)devices.They are critical for emerging wireless services,especially for applications with limited latency.DDoS attacks pose significant risks to entrepreneurial businesses,preventing legitimate customers from accessing their websites.These attacks require intelligent analytics before processing service requests.Distributed denial of service(DDoS)attacks exploit vulnerabilities in IoT devices by launchingmulti-point distributed attacks.These attacks generate massive traffic that overwhelms the victim’s network,disrupting normal operations.The consequences of distributed denial of service(DDoS)attacks are typically more severe in software-defined networks(SDNs)than in traditional networks.The centralised architecture of these networks can exacerbate existing vulnerabilities,as these weaknesses may not be effectively addressed in this model.The preliminary objective for detecting and mitigating distributed denial of service(DDoS)attacks in software-defined networks(SDN)is to monitor traffic patterns and identify anomalies that indicate distributed denial of service(DDoS)attacks.It implements measures to counter the effects ofDDoS attacks,and ensure network reliability and availability by leveraging the flexibility and programmability of SDN to adaptively respond to threats.The authors present a mechanism that leverages the OpenFlow and sFlow protocols to counter the threats posed by DDoS attacks.The results indicate that the proposed model effectively mitigates the negative effects of DDoS attacks in an SDN environment.展开更多
Multiple Sclerosis(MS)poses significant health risks.Patients may face neurodegeneration,mobility issues,cognitive decline,and a reduced quality of life.Manual diagnosis by neurologists is prone to limitations,making ...Multiple Sclerosis(MS)poses significant health risks.Patients may face neurodegeneration,mobility issues,cognitive decline,and a reduced quality of life.Manual diagnosis by neurologists is prone to limitations,making AI-based classification crucial for early detection.Therefore,automated classification using Artificial Intelligence(AI)techniques has a crucial role in addressing the limitations of manual classification and preventing the development of MS to advanced stages.This study developed hybrid systems integrating XGBoost(eXtreme Gradient Boosting)with multi-CNN(Convolutional Neural Networks)features based on Ant Colony Optimization(ACO)and Maximum Entropy Score-based Selection(MESbS)algorithms for early classification of MRI(Magnetic Resonance Imaging)images in a multi-class and binary-class MS dataset.All hybrid systems started by enhancing MRI images using the fusion processes of a Gaussian filter and Contrast-Limited Adaptive Histogram Equalization(CLAHE).Then,the Gradient Vector Flow(GVF)algorithm was applied to select white matter(regions of interest)within the brain and segment them from the surrounding brain structures.These regions of interest were processed by CNN models(ResNet101,DenseNet201,and MobileNet)to extract deep feature maps,which were then combined into fused feature vectors of multi-CNN model combinations(ResNet101-DenseNet201,DenseNet201-MobileNet,ResNet101-MobileNet,and ResNet101-DenseNet201-MobileNet).The multi-CNN features underwent dimensionality reduction using ACO and MESbS algorithms to remove unimportant features and retain important features.The XGBoost classifier employed the resultant feature vectors for classification.All developed hybrid systems displayed promising outcomes.For multiclass classification,the XGBoost model using ResNet101-DenseNet201-MobileNet features selected by ACO attained 99.4%accuracy,99.45%precision,and 99.75%specificity,surpassing prior studies(93.76%accuracy).It reached 99.6%accuracy,99.65%precision,and 99.55%specificity in binary-class classification.These results demonstrate the effectiveness of multi-CNN fusion with feature selection in improving MS classification accuracy.展开更多
Efficient resource management within Internet of Things(IoT)environments remains a pressing challenge due to the increasing number of devices and their diverse functionalities.This study introduces a neural network-ba...Efficient resource management within Internet of Things(IoT)environments remains a pressing challenge due to the increasing number of devices and their diverse functionalities.This study introduces a neural network-based model that uses Long-Short-Term Memory(LSTM)to optimize resource allocation under dynam-ically changing conditions.Designed to monitor the workload on individual IoT nodes,the model incorporates long-term data dependencies,enabling adaptive resource distribution in real time.The training process utilizes Min-Max normalization and grid search for hyperparameter tuning,ensuring high resource utilization and consistent performance.The simulation results demonstrate the effectiveness of the proposed method,outperforming the state-of-the-art approaches,including Dynamic and Efficient Enhanced Load-Balancing(DEELB),Optimized Scheduling and Collaborative Active Resource-management(OSCAR),Convolutional Neural Network with Monarch Butterfly Optimization(CNN-MBO),and Autonomic Workload Prediction and Resource Allocation for Fog(AWPR-FOG).For example,in scenarios with low system utilization,the model achieved a resource utilization efficiency of 95%while maintaining a latency of just 15 ms,significantly exceeding the performance of comparative methods.展开更多
Livestock transportation is a key factor that contributes to the spatial spread of brucellosis.To analyze the impact of sheep transportation on brucellosis transmission,we develop a human–sheep coupled brucellosis mo...Livestock transportation is a key factor that contributes to the spatial spread of brucellosis.To analyze the impact of sheep transportation on brucellosis transmission,we develop a human–sheep coupled brucellosis model within a metapopulation network framework.Theoretically,we examine the positively invariant set,the basic reproduction number,the existence,uniqueness,and stability of disease-free equilibrium and the existence of the endemic equilibrium of the model.For practical application,using Heilongjiang province as a case study,we simulate brucellosis transmission across 12 cities based on data using three network types:the BA network,the ER network,and homogeneous mixing network.The simulation results indicate that the network's average degree plays a role in the spread of brucellosis.For BA and ER networks,the basic reproduction number and cumulative incidence of brucellosis stabilize when the network's average degree reaches 4 or 5.In contrast,sheep transport in a homogeneous mixing network accelerates the cross-regional spread of brucellosis,whereas transportation in a BA network helps to control it effectively.Furthermore,the findings suggest that the movement of sheep is not always detrimental to controlling the spread of brucellosis.For cities with smaller sheep populations,such as Shuangyashan and Qitaihe,increasing the transport of sheep outward amplifies the spatial spread of the disease.In contrast,in cities with larger sheep populations,such as Qiqihar,Daqing,and Suihua,moderate sheep outflow can help reduce the spread.In addition,cities with large livestock populations play a dominant role in the overall transmission dynamics,underscoring the need for stricter supervision in these areas.展开更多
The rapid expansion of Internet of Things(IoT)networks has introduced challenges in network management,primarily in maintaining energy efficiency and robust connectivity across an increasing array of devices.This pape...The rapid expansion of Internet of Things(IoT)networks has introduced challenges in network management,primarily in maintaining energy efficiency and robust connectivity across an increasing array of devices.This paper introduces the Adaptive Blended Marine Predators Algorithm(AB-MPA),a novel optimization technique designed to enhance Quality of Service(QoS)in IoT systems by dynamically optimizing network configurations for improved energy efficiency and stability.Our results represent significant improvements in network performance metrics such as energy consumption,throughput,and operational stability,indicating that AB-MPA effectively addresses the pressing needs ofmodern IoT environments.Nodes are initiated with 100 J of stored energy,and energy is consumed at 0.01 J per square meter in each node to emphasize energy-efficient networks.The algorithm also provides sufficient network lifetime extension to a resourceful 7000 cycles for up to 200 nodes with a maximum Packet Delivery Ratio(PDR)of 99% and a robust network throughput of up to 1800 kbps in more compact node configurations.This study proposes a viable solution to a critical problem and opens avenues for further research into scalable network management for diverse applications.展开更多
IoT has emerged as a game-changing technology that connects numerous gadgets to networks for communication,processing,and real-time monitoring across diverse applications.Due to their heterogeneous nature and constrai...IoT has emerged as a game-changing technology that connects numerous gadgets to networks for communication,processing,and real-time monitoring across diverse applications.Due to their heterogeneous nature and constrained resources,as well as the growing trend of using smart gadgets,there are privacy and security issues that are not adequately managed by conventional securitymeasures.This review offers a thorough analysis of contemporary AI solutions designed to enhance security within IoT ecosystems.The intersection of AI technologies,including ML,and blockchain,with IoT privacy and security is systematically examined,focusing on their efficacy in addressing core security issues.The methodology involves a detailed exploration of existing literature and research on AI-driven privacy-preserving security mechanisms in IoT.The reviewed solutions are categorized based on their ability to tackle specific security challenges.The review highlights key advancements,evaluates their practical applications,and identifies prevailing research gaps and challenges.The findings indicate that AI solutions,particularly those leveraging ML and blockchain,offerpromising enhancements to IoT privacy and security by improving threat detection capabilities and ensuring data integrity.This paper highlights how AI technologies might strengthen IoT privacy and security and offer suggestions for upcoming studies intended to address enduring problems and improve the robustness of IoT networks.展开更多
Machine learning models can predict material properties quickly and accurately at a low computational cost.This study generated novel hybridized nanocomposites with unsaturated polyester resin as the matrix and Areca ...Machine learning models can predict material properties quickly and accurately at a low computational cost.This study generated novel hybridized nanocomposites with unsaturated polyester resin as the matrix and Areca fruit husk fiber(AFHF),tamarind fruit fiber(TFF),and nano-sized coconut shell powder(NCSP).It is challenging to determine the optimal proportion of raw materials in this composite to achieve maximum mechanical properties.This task was accomplished with the help of ML techniques in this study.The tensile strength of the hybridized nanocomposite was increased by 134.06% compared to the neat unsaturated polyester resin at a 10:5:2 wt.% ratio,AFHF:TFF:NCSP.The stiffness and impact behavior of hybridized nanocomposites were similar.The scanning electron microscope showed homogeneous reinforcement and nanofiller distribution in the matrix.However,the hybridized nanocomposite with a 20:5:0 wt.% combination ratio had the highest strain at break of 5.98%,AFHF:TFF:NCSP.The effectiveness of recurrent neural networks and recurrent neural networks with Levenberg’s algorithm was assessed using R2,mean absolute errors,and minimum squared errors.Tensile and impact strength of hybridized nanocomposites were well predicted by the recurrent neural network with Levenberg’s model with 2 and 3 hidden layers,80 neurons and 80 neurons,respectively.A recurrent neural network model with 4 hidden layers,60 neurons,and 2 hidden layers,100 neurons predicted hybridized nanocomposites’Young’s modulus and elongation at break with maximum R2 values.The mean absolute errors and minimum squared errors were evaluated to ensure the reliability of the machine learning algorithms.The models optimize hybridized nanocomposites’mechanical properties,saving time and money during experimental characterization.展开更多
Early detection of Forest and Land Fires(FLF)is essential to prevent the rapid spread of fire as well as minimize environmental damage.However,accurate detection under real-world conditions,such as low light,haze,and ...Early detection of Forest and Land Fires(FLF)is essential to prevent the rapid spread of fire as well as minimize environmental damage.However,accurate detection under real-world conditions,such as low light,haze,and complex backgrounds,remains a challenge for computer vision systems.This study evaluates the impact of three image enhancement techniques—Histogram Equalization(HE),Contrast Limited Adaptive Histogram Equalization(CLAHE),and a hybrid method called DBST-LCM CLAHE—on the performance of the YOLOv11 object detection model in identifying fires and smoke.The D-Fire dataset,consisting of 21,527 annotated images captured under diverse environmental scenarios and illumination levels,was used to train and evaluate the model.Each enhancement method was applied to the dataset before training.Model performance was assessed using multiple metrics,including Precision,Recall,mean Average Precision at 50%IoU(mAP50),F1-score,and visual inspection through bounding box results.Experimental results show that all three enhancement techniques improved detection performance.HE yielded the highest mAP50 score of 0.771,along with a balanced precision of 0.784 and recall of 0.703,demonstrating strong generalization across different conditions.DBST-LCM CLAHE achieved the highest Precision score of 79%,effectively reducing false positives,particularly in scenes with dispersed smoke or complex textures.CLAHE,with slightly lower overall metrics,contributed to improved local feature detection.Each technique showed distinct advantages:HE enhanced global contrast;CLAHE improved local structure visibility;and DBST-LCM CLAHE provided an optimal balance through dynamic block sizing and local contrast preservation.These results underline the importance of selecting preprocessing methods according to detection priorities,such as minimizing false alarms or maximizing completeness.This research does not propose a new model architecture but rather benchmarks a recent lightweight detector,YOLOv11,combined with image enhancement strategies for practical deployment in FLF monitoring.The findings support the integration of preprocessing techniques to improve detection accuracy,offering a foundation for real-time FLF detection systems on edge devices or drones,particularly in regions like Indonesia.展开更多
Electric Vehicle Charging Systems(EVCS)are increasingly vulnerable to cybersecurity threats as they integrate deeply into smart grids and Internet ofThings(IoT)environments,raising significant security challenges.Most...Electric Vehicle Charging Systems(EVCS)are increasingly vulnerable to cybersecurity threats as they integrate deeply into smart grids and Internet ofThings(IoT)environments,raising significant security challenges.Most existing research primarily emphasizes network-level anomaly detection,leaving critical vulnerabilities at the host level underexplored.This study introduces a novel forensic analysis framework leveraging host-level data,including system logs,kernel events,and Hardware Performance Counters(HPC),to detect and analyze sophisticated cyberattacks such as cryptojacking,Denial-of-Service(DoS),and reconnaissance activities targeting EVCS.Using comprehensive forensic analysis and machine learning models,the proposed framework significantly outperforms existing methods,achieving an accuracy of 98.81%.The findings offer insights into distinct behavioral signatures associated with specific cyber threats,enabling improved cybersecurity strategies and actionable recommendations for robust EVCS infrastructure protection.展开更多
Aquila Optimizer(AO)is a recently proposed population-based optimization technique inspired by Aquila’s behavior in catching prey.AO is applied in various applications and its numerous variants were proposed in the l...Aquila Optimizer(AO)is a recently proposed population-based optimization technique inspired by Aquila’s behavior in catching prey.AO is applied in various applications and its numerous variants were proposed in the literature.However,chaos theory has not been extensively investigated in AO.Moreover,it is still not applied in the parameter estimation of electro-hydraulic systems.In this work,ten well-defined chaotic maps were integrated into a narrowed exploitation of AO for the development of a robust chaotic optimization technique.An extensive investigation of twenty-three mathematical benchmarks and ten IEEE Congress on Evolutionary Computation(CEC)functions shows that chaotic Aquila optimization techniques perform better than the baseline technique.The investigation is further conducted on parameter estimation of an electro-hydraulic control system,which is performed on various noise levels and shows that the proposed chaotic AO with Piecewise map(CAO6)achieves the best fitness values of and at noise levels and respectively.Friedman test 2.873E-05,1.014E-04,8.728E-031.300E-03,1.300E-02,1.300E-01,for repeated measures,computational analysis,and Taguchi test reflect the superiority of CAO6 against the state of the arts,demonstrating its potential for addressing various engineering optimization problems.However,the sensitivity to parameter tuning may limit its direct application to complex optimization scenarios.展开更多
Cold atmospheric plasma(CAP)has emerged as a promising technology for the degradation of organic dyes,but the underlying mechanisms at the molecular level remain poorly understood.Using density-functional tight-bindin...Cold atmospheric plasma(CAP)has emerged as a promising technology for the degradation of organic dyes,but the underlying mechanisms at the molecular level remain poorly understood.Using density-functional tight-binding(DFTB)-based quantum chemical molecular dynamics at 300 K,we have performed numerical simulations to investigate the degradation mechanism of Disperse Red 1(DR)interacting with CAP-generated oxygen radicals.One hundred directdynamics trajectories were calculated for up to 100 ps simulation time,after which hydrogenabstraction,benzene ring-opening/expanding,formaldehyde formation and modification in the chromophoric azo group which can lead to color-losing were observed.The latter was obtained with yields of around 6%at the given temperature.These findings not only enhance our understanding of CAP treatment processes but also have implications for the development of optimized purification systems for sustainable wastewater treatment.This study underscores the utility of DFTB simulations in unraveling complex chemical processes and guiding the design of advanced treatment strategies in the context of CAP technology.展开更多
Background:Pneumoconioses,a group of occupational lung diseases caused by inhalation of mineral dust,pose significant health risks to affected individuals.Accurate assessment of profusion(extent of lung involvement)in...Background:Pneumoconioses,a group of occupational lung diseases caused by inhalation of mineral dust,pose significant health risks to affected individuals.Accurate assessment of profusion(extent of lung involvement)in chest radiographs is essential for screening,diagnosis and monitoring of the diseases along with epidemiological classification.This study explores an automated classification system combining U-Net-based segmentation for lung field delineation and DenseNet121 with ImageNet-based transfer learning for profusion classification.Methods:Lung field segmentation using U-Net achieved precise delineation,ensuring accurate region-of-interest definition.Transfer learning with DenseNet121 leveraged pre-trained knowledge from ImageNet,minimizing the need for extensive training.The model was fine-tuned with International Labour Organization(ILO)-2022 version standard chest radiographs and evaluated on a diverse dataset of ILO-2000 version standardized radiographs.Results:The U-Net-based segmentation demonstrated robust performance(Accuracy 94%and Dice Coefficient 90%),facilitating subsequent profusion classification.The DenseNet121-based transfer learning model exhibited high accuracy(95%),precision(92%),and recall(94%)for classifying four profusion levels on test ILO 2000/2011D dataset.The final Evaluation on ILO-2000 radiographs highlighted its generalization capability.Conclusion:The proposed system offers clinical promise,aiding radiologists,pulmonologists,general physicians,and occupational health specialists in pneumoconioses screening,diagnosis,monitoring and epidemiological classification.Best of our knowledge,this is the first work in the field of automated Classification of Profusion in Chest Radiographs of Pneumoconioses based on recently published latest ILO-2022 standard.Future research should focus on further refinement and real-world validation.This approach exemplifies the potential of deep learning for enhancing the accuracy and efficiency of pneumoconioses assessment,benefiting industrial workers,patients,and healthcare providers.展开更多
Deep Learning(DL)offers promising solutions for analyzing wearable signals and gaining valuable insights into cognitive disorders.While previous review studies have explored various aspects of DL in cognitive healthca...Deep Learning(DL)offers promising solutions for analyzing wearable signals and gaining valuable insights into cognitive disorders.While previous review studies have explored various aspects of DL in cognitive healthcare,there remains a lack of comprehensive analysis that integrates wearable signals,data processing techniques,and the broader applications,benefits,and challenges of DL methods.Addressing this limitation,our study provides an extensive review of DL’s role in cognitive healthcare,with a particular emphasis on wearables,data processing,and the inherent challenges in this field.This review also highlights the considerable promise of DL approaches in addressing a broad spectrum of cognitive issues.By enhancing the understanding and analysis of wearable signal modalities,DL models can achieve remarkable accuracy in cognitive healthcare.Convolutional Neural Network(CNN),Recurrent Neural Network(RNN),and Long Short-term Memory(LSTM)networks have demonstrated improved performance and effectiveness in the early diagnosis and progression monitoring of neurological disorders.Beyond cognitive impairment detection,DL has been applied to emotion recognition,sleep analysis,stress monitoring,and neurofeedback.These applications lead to advanced diagnosis,personalized treatment,early intervention,assistive technologies,remote monitoring,and reduced healthcare costs.Nevertheless,the integration of DL and wearable technologies presents several challenges,such as data quality,privacy,interpretability,model generalizability,ethical concerns,and clinical adoption.These challenges emphasize the importance of conducting future research in areas such as multimodal signal analysis and explainable AI.The findings of this review aim to benefit clinicians,healthcare professionals,and society by facilitating better patient outcomes in cognitive healthcare.展开更多
Automated classification of retinal fundus images is essential for identifying eye diseases,though there is earlier research on applying deep learning models designed especially for detecting tessellation in retinal f...Automated classification of retinal fundus images is essential for identifying eye diseases,though there is earlier research on applying deep learning models designed especially for detecting tessellation in retinal fundus images.This study classifies 4 classes of retinal fundus images with 3 diseased fundus images and 1 normal fundus image,by creating a refined VGG16 model to categorize fundus pictures into tessellated,normal,myopia,and choroidal neovascularization groups.The approach utilizes a VGG16 architecture that has been altered with unique fully connected layers and regularization using dropouts,along with data augmentation techniques(rotation,flip,and rescale)on a dataset of 302 photos.Training involves class weighting and critical callbacks(early halting,learning rate reduction,checkpointing)to maximize performance.Gains in accuracy(93.42%training,77.5%validation)and improved class-specific F1 scores are attained.Grad-CAM’s Explainable AI(XAI)highlights areas of the images that are important for each categorization,making it interpretable for better understanding of medical experts.These results highlight the model’s potential as a helpful diagnostic tool in ophthalmology,providing a clear and practical method for the early identification and categorization of retinal disorders,especially in cases such as tessellated fundus images.展开更多
In this paper,the problem of increasing information transfer authenticity is formulated.And to reach a decision,the control methods and algorithms based on the use of statistical and structural information redundancy ...In this paper,the problem of increasing information transfer authenticity is formulated.And to reach a decision,the control methods and algorithms based on the use of statistical and structural information redundancy are presented.It is assumed that the controllable information is submitted as the text element images and it contains redundancy,caused by statistical relations and non-uniformity probability distribution of the transmitted data.The use of statistical redundancy allows to develop the adaptive rules of the authenticity control which take into account non-stationarity properties of image data while transferring the information.The structural redundancy peculiar to the container of image in a data transfer package is used for developing new rules to control the information authenticity on the basis of pattern recognition mechanisms.The techniques offered in this work are used to estimate the authenticity in structure of data transfer packages.The results of comparative analysis for developed methods and algorithms show that their parameters of efficiency are increased by criterion of probability of undetected mistakes,labour input and cost of realization.展开更多
文摘In first aid, traditional information interchange has numerous shortcomings. For example, delayed information and disorganized departmental communication cause patients to miss out on critical rescue time. Information technology is becoming more and more mature, and as a result, its use across numerous industries is now standard. China is still in the early stages of developing its integration of emergency medical services with modern information technology;despite our progress, there are still numerous obstacles and constraints to overcome. Our goal is to integrate information technology into every aspect of emergency patient care, offering robust assistance for both patient rescue and the efforts of medical personnel. Information may be communicated in a fast, multiple, and effective manner by utilizing modern information technology. This study aims to examine the current state of this field’s development, current issues, and the field’s future course of development.
文摘Lower back pain is one of the most common medical problems in the world and it is experienced by a huge percentage of people everywhere.Due to its ability to produce a detailed view of the soft tissues,including the spinal cord,nerves,intervertebral discs,and vertebrae,Magnetic Resonance Imaging is thought to be the most effective method for imaging the spine.The semantic segmentation of vertebrae plays a major role in the diagnostic process of lumbar diseases.It is difficult to semantically partition the vertebrae in Magnetic Resonance Images from the surrounding variety of tissues,including muscles,ligaments,and intervertebral discs.U-Net is a powerful deep-learning architecture to handle the challenges of medical image analysis tasks and achieves high segmentation accuracy.This work proposes a modified U-Net architecture namely MU-Net,consisting of the Meijering convolutional layer that incorporates the Meijering filter to perform the semantic segmentation of lumbar vertebrae L1 to L5 and sacral vertebra S1.Pseudo-colour mask images were generated and used as ground truth for training the model.The work has been carried out on 1312 images expanded from T1-weighted mid-sagittal MRI images of 515 patients in the Lumbar Spine MRI Dataset publicly available from Mendeley Data.The proposed MU-Net model for the semantic segmentation of the lumbar vertebrae gives better performance with 98.79%of pixel accuracy(PA),98.66%of dice similarity coefficient(DSC),97.36%of Jaccard coefficient,and 92.55%mean Intersection over Union(mean IoU)metrics using the mentioned dataset.
文摘Terms of intelligence in 20th and 21th century mean the methods of automatic extraction, analysis, interpretation and use of information. Thus, the intelligence services in the future created an electronic database in which to their being classified intelligence products, users could choose between the latter themselves relevant information. The EU (European Union) that activities are carried out from at least in year 1996, terrorist attacks in year 200l is only accelerating. Proposals to increase surveillance and international cooperation in this field have been drawn up before September 11 2011. On the Web you can fmd a list of networks (Cryptome, 2011), which could be connected, or are under the control of the security service--NSA (National Security Agency). United States of America in year 1994 enacted a law for telephone communication--Digital Telephony Act, which would require manufacturers of telecommunications equipment, leaving some security holes for control. In addition, we monitor the Internet and large corporations. The example of the United States of America in this action reveals the organization for electronic freedoms against a telecom company that the NSA illegally gains access to data on information technology users and Internet telephony.
基金funded and supported by the UCSI University Research Excellence&Innovation Grant(REIG),REIG-ICSDI-2024/044.
文摘In the rapidly evolving landscape of intelligent transportation systems,the security and authenticity of vehicular communication have emerged as critical challenges.As vehicles become increasingly interconnected,the need for robust authentication mechanisms to safeguard against cyber threats and ensure trust in an autonomous ecosystem becomes essential.On the other hand,using intelligence in the authentication system is a significant attraction.While existing surveys broadly address vehicular security,a critical gap remains in the systematic exploration of Deep Learning(DL)-based authentication methods tailored to these communication paradigms.This survey fills that gap by offering a comprehensive analysis of DL techniques—including supervised,unsupervised,reinforcement,and hybrid learning—for vehicular authentication.This survey highlights novel contributions,such as a taxonomy of DL-driven authentication protocols,real-world case studies,and a critical evaluation of scalability and privacy-preserving techniques.Additionally,this paper identifies unresolved challenges,such as adversarial resilience and real-time processing constraints,and proposes actionable future directions,including lightweight model optimization and blockchain integration.By grounding the discussion in concrete applications,such as biometric authentication for driver safety and adaptive key management for infrastructure security,this survey bridges theoretical advancements with practical deployment needs,offering a roadmap for next-generation secure intelligent vehicular ecosystems for the modern world.
文摘Purpose:Generally,the scientific comparison has been done with the help of the overall impact of scholars.Although it is very easy to compare scholars,but how can we assess the scientific impact of scholars who have different research careers?It is very obvious,the scholars may gain a high impact if they have more research experience or have spent more time(in terms of research career in a year).Then we cannot compare two scholars who have different research careers.Many bibliometrics indicators address the time-span of scholars.In this series,the h-index sequence and EM/EM’-index sequence have been introduced for assessment and comparison of the scientific impact of scholars.The h-index sequence,EM-index sequence,and EM’-index sequence consider the yearly impact of scholars,and comparison is done by the index value along with their component value.The time-series indicators fail to give a comparative analysis between senior and junior scholars if there is a huge difference in both scholars’research careers.Design/methodology/approach:We have proposed the cumulative index calculation method to appraise the scientific impact of scholars till that age and tested it with 89 scholars data.Findings:The proposed mechanism is implemented and tested on 89 scholars’publication data,providing a clear difference between the scientific impact of two scholars.This also helps in predicting future prominent scholars based on their research impact.Research limitations:This study adopts a simplistic approach by assigning equal credit to all authors,regardless of their individual contributions.Further,the potential impact of career breaks on research productivity is not taken into account.These assumptions may limit the generalizability of our findings Practical implications:The proposed method can be used by respected institutions to compare their scholars impact.Funding agencies can also use it for similar purposes.Originality/value:This research adds to the existing literature by introducing a novel methodology for comparing the scientific impact of scholars.The outcomes of this research have notable implications for the development of more precise and unbiased research assessment frameworks,enabling a more equitable evaluation of scholarly contributions.
基金supported by the Deanship of Graduate Studies and Scientific Research at Qassim University for financial support(QU-APC-2025).
文摘Distributed denial of service(DDoS)attacks are common network attacks that primarily target Internet of Things(IoT)devices.They are critical for emerging wireless services,especially for applications with limited latency.DDoS attacks pose significant risks to entrepreneurial businesses,preventing legitimate customers from accessing their websites.These attacks require intelligent analytics before processing service requests.Distributed denial of service(DDoS)attacks exploit vulnerabilities in IoT devices by launchingmulti-point distributed attacks.These attacks generate massive traffic that overwhelms the victim’s network,disrupting normal operations.The consequences of distributed denial of service(DDoS)attacks are typically more severe in software-defined networks(SDNs)than in traditional networks.The centralised architecture of these networks can exacerbate existing vulnerabilities,as these weaknesses may not be effectively addressed in this model.The preliminary objective for detecting and mitigating distributed denial of service(DDoS)attacks in software-defined networks(SDN)is to monitor traffic patterns and identify anomalies that indicate distributed denial of service(DDoS)attacks.It implements measures to counter the effects ofDDoS attacks,and ensure network reliability and availability by leveraging the flexibility and programmability of SDN to adaptively respond to threats.The authors present a mechanism that leverages the OpenFlow and sFlow protocols to counter the threats posed by DDoS attacks.The results indicate that the proposed model effectively mitigates the negative effects of DDoS attacks in an SDN environment.
文摘Multiple Sclerosis(MS)poses significant health risks.Patients may face neurodegeneration,mobility issues,cognitive decline,and a reduced quality of life.Manual diagnosis by neurologists is prone to limitations,making AI-based classification crucial for early detection.Therefore,automated classification using Artificial Intelligence(AI)techniques has a crucial role in addressing the limitations of manual classification and preventing the development of MS to advanced stages.This study developed hybrid systems integrating XGBoost(eXtreme Gradient Boosting)with multi-CNN(Convolutional Neural Networks)features based on Ant Colony Optimization(ACO)and Maximum Entropy Score-based Selection(MESbS)algorithms for early classification of MRI(Magnetic Resonance Imaging)images in a multi-class and binary-class MS dataset.All hybrid systems started by enhancing MRI images using the fusion processes of a Gaussian filter and Contrast-Limited Adaptive Histogram Equalization(CLAHE).Then,the Gradient Vector Flow(GVF)algorithm was applied to select white matter(regions of interest)within the brain and segment them from the surrounding brain structures.These regions of interest were processed by CNN models(ResNet101,DenseNet201,and MobileNet)to extract deep feature maps,which were then combined into fused feature vectors of multi-CNN model combinations(ResNet101-DenseNet201,DenseNet201-MobileNet,ResNet101-MobileNet,and ResNet101-DenseNet201-MobileNet).The multi-CNN features underwent dimensionality reduction using ACO and MESbS algorithms to remove unimportant features and retain important features.The XGBoost classifier employed the resultant feature vectors for classification.All developed hybrid systems displayed promising outcomes.For multiclass classification,the XGBoost model using ResNet101-DenseNet201-MobileNet features selected by ACO attained 99.4%accuracy,99.45%precision,and 99.75%specificity,surpassing prior studies(93.76%accuracy).It reached 99.6%accuracy,99.65%precision,and 99.55%specificity in binary-class classification.These results demonstrate the effectiveness of multi-CNN fusion with feature selection in improving MS classification accuracy.
基金funding of the Deanship of Graduate Studies and Scientific Research,Jazan University,Saudi Arabia,through Project Number:ISP-2024.
文摘Efficient resource management within Internet of Things(IoT)environments remains a pressing challenge due to the increasing number of devices and their diverse functionalities.This study introduces a neural network-based model that uses Long-Short-Term Memory(LSTM)to optimize resource allocation under dynam-ically changing conditions.Designed to monitor the workload on individual IoT nodes,the model incorporates long-term data dependencies,enabling adaptive resource distribution in real time.The training process utilizes Min-Max normalization and grid search for hyperparameter tuning,ensuring high resource utilization and consistent performance.The simulation results demonstrate the effectiveness of the proposed method,outperforming the state-of-the-art approaches,including Dynamic and Efficient Enhanced Load-Balancing(DEELB),Optimized Scheduling and Collaborative Active Resource-management(OSCAR),Convolutional Neural Network with Monarch Butterfly Optimization(CNN-MBO),and Autonomic Workload Prediction and Resource Allocation for Fog(AWPR-FOG).For example,in scenarios with low system utilization,the model achieved a resource utilization efficiency of 95%while maintaining a latency of just 15 ms,significantly exceeding the performance of comparative methods.
基金Project supported by the National Natural Science Foundation of China(Grant No.12101443,12371493)the Natural Science Foundation of Shanxi Province(Grant Nos.20210302124260 and 202303021221024)。
文摘Livestock transportation is a key factor that contributes to the spatial spread of brucellosis.To analyze the impact of sheep transportation on brucellosis transmission,we develop a human–sheep coupled brucellosis model within a metapopulation network framework.Theoretically,we examine the positively invariant set,the basic reproduction number,the existence,uniqueness,and stability of disease-free equilibrium and the existence of the endemic equilibrium of the model.For practical application,using Heilongjiang province as a case study,we simulate brucellosis transmission across 12 cities based on data using three network types:the BA network,the ER network,and homogeneous mixing network.The simulation results indicate that the network's average degree plays a role in the spread of brucellosis.For BA and ER networks,the basic reproduction number and cumulative incidence of brucellosis stabilize when the network's average degree reaches 4 or 5.In contrast,sheep transport in a homogeneous mixing network accelerates the cross-regional spread of brucellosis,whereas transportation in a BA network helps to control it effectively.Furthermore,the findings suggest that the movement of sheep is not always detrimental to controlling the spread of brucellosis.For cities with smaller sheep populations,such as Shuangyashan and Qitaihe,increasing the transport of sheep outward amplifies the spatial spread of the disease.In contrast,in cities with larger sheep populations,such as Qiqihar,Daqing,and Suihua,moderate sheep outflow can help reduce the spread.In addition,cities with large livestock populations play a dominant role in the overall transmission dynamics,underscoring the need for stricter supervision in these areas.
文摘The rapid expansion of Internet of Things(IoT)networks has introduced challenges in network management,primarily in maintaining energy efficiency and robust connectivity across an increasing array of devices.This paper introduces the Adaptive Blended Marine Predators Algorithm(AB-MPA),a novel optimization technique designed to enhance Quality of Service(QoS)in IoT systems by dynamically optimizing network configurations for improved energy efficiency and stability.Our results represent significant improvements in network performance metrics such as energy consumption,throughput,and operational stability,indicating that AB-MPA effectively addresses the pressing needs ofmodern IoT environments.Nodes are initiated with 100 J of stored energy,and energy is consumed at 0.01 J per square meter in each node to emphasize energy-efficient networks.The algorithm also provides sufficient network lifetime extension to a resourceful 7000 cycles for up to 200 nodes with a maximum Packet Delivery Ratio(PDR)of 99% and a robust network throughput of up to 1800 kbps in more compact node configurations.This study proposes a viable solution to a critical problem and opens avenues for further research into scalable network management for diverse applications.
基金The author Dr.Arshiya Sajid Ansari extends the appreciation to the Deanship of Postgraduate Studies and Scientific Research at Majmaah University for funding this research work through the project number(R-2025-1706).
文摘IoT has emerged as a game-changing technology that connects numerous gadgets to networks for communication,processing,and real-time monitoring across diverse applications.Due to their heterogeneous nature and constrained resources,as well as the growing trend of using smart gadgets,there are privacy and security issues that are not adequately managed by conventional securitymeasures.This review offers a thorough analysis of contemporary AI solutions designed to enhance security within IoT ecosystems.The intersection of AI technologies,including ML,and blockchain,with IoT privacy and security is systematically examined,focusing on their efficacy in addressing core security issues.The methodology involves a detailed exploration of existing literature and research on AI-driven privacy-preserving security mechanisms in IoT.The reviewed solutions are categorized based on their ability to tackle specific security challenges.The review highlights key advancements,evaluates their practical applications,and identifies prevailing research gaps and challenges.The findings indicate that AI solutions,particularly those leveraging ML and blockchain,offerpromising enhancements to IoT privacy and security by improving threat detection capabilities and ensuring data integrity.This paper highlights how AI technologies might strengthen IoT privacy and security and offer suggestions for upcoming studies intended to address enduring problems and improve the robustness of IoT networks.
文摘Machine learning models can predict material properties quickly and accurately at a low computational cost.This study generated novel hybridized nanocomposites with unsaturated polyester resin as the matrix and Areca fruit husk fiber(AFHF),tamarind fruit fiber(TFF),and nano-sized coconut shell powder(NCSP).It is challenging to determine the optimal proportion of raw materials in this composite to achieve maximum mechanical properties.This task was accomplished with the help of ML techniques in this study.The tensile strength of the hybridized nanocomposite was increased by 134.06% compared to the neat unsaturated polyester resin at a 10:5:2 wt.% ratio,AFHF:TFF:NCSP.The stiffness and impact behavior of hybridized nanocomposites were similar.The scanning electron microscope showed homogeneous reinforcement and nanofiller distribution in the matrix.However,the hybridized nanocomposite with a 20:5:0 wt.% combination ratio had the highest strain at break of 5.98%,AFHF:TFF:NCSP.The effectiveness of recurrent neural networks and recurrent neural networks with Levenberg’s algorithm was assessed using R2,mean absolute errors,and minimum squared errors.Tensile and impact strength of hybridized nanocomposites were well predicted by the recurrent neural network with Levenberg’s model with 2 and 3 hidden layers,80 neurons and 80 neurons,respectively.A recurrent neural network model with 4 hidden layers,60 neurons,and 2 hidden layers,100 neurons predicted hybridized nanocomposites’Young’s modulus and elongation at break with maximum R2 values.The mean absolute errors and minimum squared errors were evaluated to ensure the reliability of the machine learning algorithms.The models optimize hybridized nanocomposites’mechanical properties,saving time and money during experimental characterization.
基金funded by the Directorate of Research,Technology,and Community Service,Ministry of Higher Education,Science,and Technology of the Republic of Indonesia the Regular Fundamental Research scheme,with grant numbers 001/LL6/PL/AL.04/2025,011/SPK-PFR/RIK/05/2025.
文摘Early detection of Forest and Land Fires(FLF)is essential to prevent the rapid spread of fire as well as minimize environmental damage.However,accurate detection under real-world conditions,such as low light,haze,and complex backgrounds,remains a challenge for computer vision systems.This study evaluates the impact of three image enhancement techniques—Histogram Equalization(HE),Contrast Limited Adaptive Histogram Equalization(CLAHE),and a hybrid method called DBST-LCM CLAHE—on the performance of the YOLOv11 object detection model in identifying fires and smoke.The D-Fire dataset,consisting of 21,527 annotated images captured under diverse environmental scenarios and illumination levels,was used to train and evaluate the model.Each enhancement method was applied to the dataset before training.Model performance was assessed using multiple metrics,including Precision,Recall,mean Average Precision at 50%IoU(mAP50),F1-score,and visual inspection through bounding box results.Experimental results show that all three enhancement techniques improved detection performance.HE yielded the highest mAP50 score of 0.771,along with a balanced precision of 0.784 and recall of 0.703,demonstrating strong generalization across different conditions.DBST-LCM CLAHE achieved the highest Precision score of 79%,effectively reducing false positives,particularly in scenes with dispersed smoke or complex textures.CLAHE,with slightly lower overall metrics,contributed to improved local feature detection.Each technique showed distinct advantages:HE enhanced global contrast;CLAHE improved local structure visibility;and DBST-LCM CLAHE provided an optimal balance through dynamic block sizing and local contrast preservation.These results underline the importance of selecting preprocessing methods according to detection priorities,such as minimizing false alarms or maximizing completeness.This research does not propose a new model architecture but rather benchmarks a recent lightweight detector,YOLOv11,combined with image enhancement strategies for practical deployment in FLF monitoring.The findings support the integration of preprocessing techniques to improve detection accuracy,offering a foundation for real-time FLF detection systems on edge devices or drones,particularly in regions like Indonesia.
文摘Electric Vehicle Charging Systems(EVCS)are increasingly vulnerable to cybersecurity threats as they integrate deeply into smart grids and Internet ofThings(IoT)environments,raising significant security challenges.Most existing research primarily emphasizes network-level anomaly detection,leaving critical vulnerabilities at the host level underexplored.This study introduces a novel forensic analysis framework leveraging host-level data,including system logs,kernel events,and Hardware Performance Counters(HPC),to detect and analyze sophisticated cyberattacks such as cryptojacking,Denial-of-Service(DoS),and reconnaissance activities targeting EVCS.Using comprehensive forensic analysis and machine learning models,the proposed framework significantly outperforms existing methods,achieving an accuracy of 98.81%.The findings offer insights into distinct behavioral signatures associated with specific cyber threats,enabling improved cybersecurity strategies and actionable recommendations for robust EVCS infrastructure protection.
基金funded by Taif University,Saudi Arabia,Project No.(TU-DSPP-2024-52).
文摘Aquila Optimizer(AO)is a recently proposed population-based optimization technique inspired by Aquila’s behavior in catching prey.AO is applied in various applications and its numerous variants were proposed in the literature.However,chaos theory has not been extensively investigated in AO.Moreover,it is still not applied in the parameter estimation of electro-hydraulic systems.In this work,ten well-defined chaotic maps were integrated into a narrowed exploitation of AO for the development of a robust chaotic optimization technique.An extensive investigation of twenty-three mathematical benchmarks and ten IEEE Congress on Evolutionary Computation(CEC)functions shows that chaotic Aquila optimization techniques perform better than the baseline technique.The investigation is further conducted on parameter estimation of an electro-hydraulic control system,which is performed on various noise levels and shows that the proposed chaotic AO with Piecewise map(CAO6)achieves the best fitness values of and at noise levels and respectively.Friedman test 2.873E-05,1.014E-04,8.728E-031.300E-03,1.300E-02,1.300E-01,for repeated measures,computational analysis,and Taguchi test reflect the superiority of CAO6 against the state of the arts,demonstrating its potential for addressing various engineering optimization problems.However,the sensitivity to parameter tuning may limit its direct application to complex optimization scenarios.
基金the financial support from the Ministry of Higher Education,Science,and Innovations of the Republic of Uzbekistan (Nos.AL-4821012320 and AL-5921122141)。
文摘Cold atmospheric plasma(CAP)has emerged as a promising technology for the degradation of organic dyes,but the underlying mechanisms at the molecular level remain poorly understood.Using density-functional tight-binding(DFTB)-based quantum chemical molecular dynamics at 300 K,we have performed numerical simulations to investigate the degradation mechanism of Disperse Red 1(DR)interacting with CAP-generated oxygen radicals.One hundred directdynamics trajectories were calculated for up to 100 ps simulation time,after which hydrogenabstraction,benzene ring-opening/expanding,formaldehyde formation and modification in the chromophoric azo group which can lead to color-losing were observed.The latter was obtained with yields of around 6%at the given temperature.These findings not only enhance our understanding of CAP treatment processes but also have implications for the development of optimized purification systems for sustainable wastewater treatment.This study underscores the utility of DFTB simulations in unraveling complex chemical processes and guiding the design of advanced treatment strategies in the context of CAP technology.
文摘Background:Pneumoconioses,a group of occupational lung diseases caused by inhalation of mineral dust,pose significant health risks to affected individuals.Accurate assessment of profusion(extent of lung involvement)in chest radiographs is essential for screening,diagnosis and monitoring of the diseases along with epidemiological classification.This study explores an automated classification system combining U-Net-based segmentation for lung field delineation and DenseNet121 with ImageNet-based transfer learning for profusion classification.Methods:Lung field segmentation using U-Net achieved precise delineation,ensuring accurate region-of-interest definition.Transfer learning with DenseNet121 leveraged pre-trained knowledge from ImageNet,minimizing the need for extensive training.The model was fine-tuned with International Labour Organization(ILO)-2022 version standard chest radiographs and evaluated on a diverse dataset of ILO-2000 version standardized radiographs.Results:The U-Net-based segmentation demonstrated robust performance(Accuracy 94%and Dice Coefficient 90%),facilitating subsequent profusion classification.The DenseNet121-based transfer learning model exhibited high accuracy(95%),precision(92%),and recall(94%)for classifying four profusion levels on test ILO 2000/2011D dataset.The final Evaluation on ILO-2000 radiographs highlighted its generalization capability.Conclusion:The proposed system offers clinical promise,aiding radiologists,pulmonologists,general physicians,and occupational health specialists in pneumoconioses screening,diagnosis,monitoring and epidemiological classification.Best of our knowledge,this is the first work in the field of automated Classification of Profusion in Chest Radiographs of Pneumoconioses based on recently published latest ILO-2022 standard.Future research should focus on further refinement and real-world validation.This approach exemplifies the potential of deep learning for enhancing the accuracy and efficiency of pneumoconioses assessment,benefiting industrial workers,patients,and healthcare providers.
基金the Asian Institute of Technology,Khlong Nueng,Thailand for their support in carrying out this study。
文摘Deep Learning(DL)offers promising solutions for analyzing wearable signals and gaining valuable insights into cognitive disorders.While previous review studies have explored various aspects of DL in cognitive healthcare,there remains a lack of comprehensive analysis that integrates wearable signals,data processing techniques,and the broader applications,benefits,and challenges of DL methods.Addressing this limitation,our study provides an extensive review of DL’s role in cognitive healthcare,with a particular emphasis on wearables,data processing,and the inherent challenges in this field.This review also highlights the considerable promise of DL approaches in addressing a broad spectrum of cognitive issues.By enhancing the understanding and analysis of wearable signal modalities,DL models can achieve remarkable accuracy in cognitive healthcare.Convolutional Neural Network(CNN),Recurrent Neural Network(RNN),and Long Short-term Memory(LSTM)networks have demonstrated improved performance and effectiveness in the early diagnosis and progression monitoring of neurological disorders.Beyond cognitive impairment detection,DL has been applied to emotion recognition,sleep analysis,stress monitoring,and neurofeedback.These applications lead to advanced diagnosis,personalized treatment,early intervention,assistive technologies,remote monitoring,and reduced healthcare costs.Nevertheless,the integration of DL and wearable technologies presents several challenges,such as data quality,privacy,interpretability,model generalizability,ethical concerns,and clinical adoption.These challenges emphasize the importance of conducting future research in areas such as multimodal signal analysis and explainable AI.The findings of this review aim to benefit clinicians,healthcare professionals,and society by facilitating better patient outcomes in cognitive healthcare.
基金support from the"Intelligent Recognition Industry Service Center"as part of the Featured Areas Research Center Program under the Higher Education Sprout Project by the Ministry of Education(MOE)in Taiwan,and the National Science and Technology Council,Taiwan,under grants[113-2622-E-224-002]and[113-2221-E-224-041]support was provided by Isuzu Optics Corporation.
文摘Automated classification of retinal fundus images is essential for identifying eye diseases,though there is earlier research on applying deep learning models designed especially for detecting tessellation in retinal fundus images.This study classifies 4 classes of retinal fundus images with 3 diseased fundus images and 1 normal fundus image,by creating a refined VGG16 model to categorize fundus pictures into tessellated,normal,myopia,and choroidal neovascularization groups.The approach utilizes a VGG16 architecture that has been altered with unique fully connected layers and regularization using dropouts,along with data augmentation techniques(rotation,flip,and rescale)on a dataset of 302 photos.Training involves class weighting and critical callbacks(early halting,learning rate reduction,checkpointing)to maximize performance.Gains in accuracy(93.42%training,77.5%validation)and improved class-specific F1 scores are attained.Grad-CAM’s Explainable AI(XAI)highlights areas of the images that are important for each categorization,making it interpretable for better understanding of medical experts.These results highlight the model’s potential as a helpful diagnostic tool in ophthalmology,providing a clear and practical method for the early identification and categorization of retinal disorders,especially in cases such as tessellated fundus images.
文摘In this paper,the problem of increasing information transfer authenticity is formulated.And to reach a decision,the control methods and algorithms based on the use of statistical and structural information redundancy are presented.It is assumed that the controllable information is submitted as the text element images and it contains redundancy,caused by statistical relations and non-uniformity probability distribution of the transmitted data.The use of statistical redundancy allows to develop the adaptive rules of the authenticity control which take into account non-stationarity properties of image data while transferring the information.The structural redundancy peculiar to the container of image in a data transfer package is used for developing new rules to control the information authenticity on the basis of pattern recognition mechanisms.The techniques offered in this work are used to estimate the authenticity in structure of data transfer packages.The results of comparative analysis for developed methods and algorithms show that their parameters of efficiency are increased by criterion of probability of undetected mistakes,labour input and cost of realization.