The analysis of Android malware shows that this threat is constantly increasing and is a real threat to mobile devices since traditional approaches,such as signature-based detection,are no longer effective due to the ...The analysis of Android malware shows that this threat is constantly increasing and is a real threat to mobile devices since traditional approaches,such as signature-based detection,are no longer effective due to the continuously advancing level of sophistication.To resolve this problem,efficient and flexible malware detection tools are needed.This work examines the possibility of employing deep CNNs to detect Android malware by transforming network traffic into image data representations.Moreover,the dataset used in this study is the CIC-AndMal2017,which contains 20,000 instances of network traffic across five distinct malware categories:a.Trojan,b.Adware,c.Ransomware,d.Spyware,e.Worm.These network traffic features are then converted to image formats for deep learning,which is applied in a CNN framework,including the VGG16 pre-trained model.In addition,our approach yielded high performance,yielding an accuracy of 0.92,accuracy of 99.1%,precision of 98.2%,recall of 99.5%,and F1 score of 98.7%.Subsequent improvements to the classification model through changes within the VGG19 framework improved the classification rate to 99.25%.Through the results obtained,it is clear that CNNs are a very effective way to classify Android malware,providing greater accuracy than conventional techniques.The success of this approach also shows the applicability of deep learning in mobile security along with the direction for the future advancement of the real-time detection system and other deeper learning techniques to counter the increasing number of threats emerging in the future.展开更多
Palmprint identification has been conducted over the last two decades in many biometric systems.High-dimensional data with many uncorrelated and duplicated features remains difficult due to several computational compl...Palmprint identification has been conducted over the last two decades in many biometric systems.High-dimensional data with many uncorrelated and duplicated features remains difficult due to several computational complexity issues.This paper presents an interactive authentication approach based on deep learning and feature selection that supports Palmprint authentication.The proposed model has two stages of learning;the first stage is to transfer pre-trained VGG-16 of ImageNet to specific features based on the extraction model.The second stage involves the VGG-16 Palmprint feature extraction in the Siamese network to learn Palmprint similarity.The proposed model achieves robust and reliable end-to-end Palmprint authentication by extracting the convolutional features using VGG-16 Palmprint and the similarity of two input Palmprint using the Siamese network.The second stage uses the CASIA dataset to train and test the Siamese network.The suggested model outperforms comparable studies based on the deep learning approach achieving accuracy and EER of 91.8%and 0.082%,respectively,on the CASIA left-hand images and accuracy and EER of 91.7%and 0.084,respectively,on the CASIA right-hand images.展开更多
The nutritional value of perishable food items,such as fruits and vegetables,depends on their freshness levels.The existing approaches solve a binary class problem by classifying a known fruit\vegetable class into fre...The nutritional value of perishable food items,such as fruits and vegetables,depends on their freshness levels.The existing approaches solve a binary class problem by classifying a known fruit\vegetable class into fresh or rotten only.We propose an automated fruits and vegetables categorization approach that first recognizes the class of object in an image and then categorizes that fruit or vegetable into one of the three categories:purefresh,medium-fresh,and rotten.We gathered a dataset comprising of 60K images of 11 fruits and vegetables,each is further divided into three categories of freshness,using hand-held cameras.The recognition and categorization of fruits and vegetables are performed through two deep learning models:Visual Geometry Group(VGG-16)and You Only Look Once(YOLO),and their results are compared.VGG-16 classifies fruits and vegetables and categorizes their freshness,while YOLO also localizes them within the image.Furthermore,we have developed an android based application that takes the image of the fruit or vegetable as input and returns its class label and its freshness degree.A comprehensive experimental evaluation of proposed approach demonstrates that the proposed approach can achieve a high accuracy and F1score on gathered FruitVeg Freshness dataset.The dataset is publicly available for further evaluation by the research community.展开更多
Nowadays,the COVID-19 virus disease is spreading rampantly.There are some testing tools and kits available for diagnosing the virus,but it is in a lim-ited count.To diagnose the presence of disease from radiological i...Nowadays,the COVID-19 virus disease is spreading rampantly.There are some testing tools and kits available for diagnosing the virus,but it is in a lim-ited count.To diagnose the presence of disease from radiological images,auto-mated COVID-19 diagnosis techniques are needed.The enhancement of AI(Artificial Intelligence)has been focused in previous research,which uses X-ray images for detecting COVID-19.The most common symptoms of COVID-19 are fever,dry cough and sore throat.These symptoms may lead to an increase in the rigorous type of pneumonia with a severe barrier.Since medical imaging is not suggested recently in Canada for critical COVID-19 diagnosis,computer-aided systems are implemented for the early identification of COVID-19,which aids in noticing the disease progression and thus decreases the death rate.Here,a deep learning-based automated method for the extraction of features and classi-fication is enhanced for the detection of COVID-19 from the images of computer tomography(CT).The suggested method functions on the basis of three main pro-cesses:data preprocessing,the extraction of features and classification.This approach integrates the union of deep features with the help of Inception 14 and VGG-16 models.At last,a classifier of Multi-scale Improved ResNet(MSI-ResNet)is developed to detect and classify the CT images into unique labels of class.With the support of available open-source COVID-CT datasets that consists of 760 CT pictures,the investigational validation of the suggested method is estimated.The experimental results reveal that the proposed approach offers greater performance with high specificity,accuracy and sensitivity.展开更多
The use of the internet is increasing all over the world on a daily basis in the last two decades.The increase in the internet causes many sexual crimes,such as sexual misuse,domestic violence,and child pornography.Va...The use of the internet is increasing all over the world on a daily basis in the last two decades.The increase in the internet causes many sexual crimes,such as sexual misuse,domestic violence,and child pornography.Various research has been done for pornographic image detection and classification.Most of the used models used machine learning techniques and deep learning models which show less accuracy,while the deep learning model ware used for classification and detection performed better as compared to machine learning.Therefore,this research evaluates the performance analysis of intelligent neural-based deep learning models which are based on Convolution neural network(CNN),Visual geometry group(VGG-16),VGG-14,and Residual Network(ResNet-50)with the expanded dataset,trained using transfer learning approaches applied in the fully connected layer for datasets to classify rank(Pornographic vs.Nonpornographic)classification in images.The simulation result shows that VGG-16 performed better than the used model in this study without augmented data.The VGG-16 model with augmented data reached a training and validation accuracy of 0.97,0.94 with a loss of 0.070,0.16.The precision,recall,and f-measure values for explicit and non-explicit images are(0.94,0.94,0.94)and(0.94,0.94,0.94).Similarly,The VGG-14 model with augmented data reached a training and validation accuracy of 0.98,0.96 with a loss of 0.059,0.11.The f-measure,recall,and precision values for explicit and non-explicit images are(0.98,0.98,0.98)and(0.98,0.98,0.98).The CNN model with augmented data reached a training and validation accuracy of 0.776&0.78 with losses of 0.48&0.46.The f-measure,recall,and precision values for explicit and non-explicit images are(0.80,0.80,0.80)and(0.78,0.79,0.78).The ResNet-50 model with expanded data reached with training accuracy of 0.89 with a loss of 0.389 and 0.86 of validation accuracy and a loss of 0.47.The f-measure,recall,and precision values for explicit and non-explicit images are(0.86,0.97,0.91)and(0.86,0.93,0.89).Where else without augmented data the VGG-16 model reached a training and validation accuracy of 0.997,0.986 with a loss of 0.008,0.056.The f-measure,recall,and precision values for explicit and non-explicit images are(0.94,0.99,0.97)and(0.99,0.93,0.96)which outperforms the used models with the augmented dataset in this study.展开更多
Currently,distracted driving is among the most important causes of traffic accidents.Consequently,intelligent vehicle driving systems have become increasingly important.Recently,interest in driver-assistance systems t...Currently,distracted driving is among the most important causes of traffic accidents.Consequently,intelligent vehicle driving systems have become increasingly important.Recently,interest in driver-assistance systems that detect driver actions and help them drive safely has increased.In these studies,although some distinct data types,such as the physical conditions of the driver,audio and visual features,and vehicle information,are used,the primary data source is images of the driver that include the face,arms,and hands taken with a camera inside the car.In this study,an architecture based on a convolution neural network(CNN)is proposed to classify and detect driver distraction.An efficient CNN with high accuracy is implemented,and to implement intense convolutional networks for large-scale image recognition,a new architecture was proposed based on the available Visual Geometry Group(VGG-16)architecture.The proposed architecture was evaluated using the StateFarm dataset for driver-distraction detection.This dataset is publicly available on Kaggle and is frequently used for this type of research.The proposed architecture achieved 96.95%accuracy.展开更多
Aiming at the problems such as diverse target scales and large-scale changes in crowds in dense crowd scenarios,a crowd density estimation method based on multi-scale feature fusion and information en-hancement is pro...Aiming at the problems such as diverse target scales and large-scale changes in crowds in dense crowd scenarios,a crowd density estimation method based on multi-scale feature fusion and information en-hancement is proposed.Firstly,considering that small-scale targets account for a relatively large proportion in the image,based on the VGG-16 network,the dilated convolution module is introduced to mine the detailed information of the image.Secondly,in order to make full use of the multi-scale information of the target,a new context-aware module is constructed to extract the contrast features between different scales.Finally,con-sidering the characteristic of continuous changes in the target scale,a multi-scale feature aggregation module is designed to enhance the sampling range of dense scales and multi-scale information interaction,thereby improving the network performance.Experiments on public datasets show that the proposed method in this paper can effectively estimate the population density compared with other advanced methods.展开更多
基金funded by the Deanship of Scientific Research at Princess Nourah bint Abdulrahman University,through the Research Funding Program,Grant No.(FRP-1443-15).
文摘The analysis of Android malware shows that this threat is constantly increasing and is a real threat to mobile devices since traditional approaches,such as signature-based detection,are no longer effective due to the continuously advancing level of sophistication.To resolve this problem,efficient and flexible malware detection tools are needed.This work examines the possibility of employing deep CNNs to detect Android malware by transforming network traffic into image data representations.Moreover,the dataset used in this study is the CIC-AndMal2017,which contains 20,000 instances of network traffic across five distinct malware categories:a.Trojan,b.Adware,c.Ransomware,d.Spyware,e.Worm.These network traffic features are then converted to image formats for deep learning,which is applied in a CNN framework,including the VGG16 pre-trained model.In addition,our approach yielded high performance,yielding an accuracy of 0.92,accuracy of 99.1%,precision of 98.2%,recall of 99.5%,and F1 score of 98.7%.Subsequent improvements to the classification model through changes within the VGG19 framework improved the classification rate to 99.25%.Through the results obtained,it is clear that CNNs are a very effective way to classify Android malware,providing greater accuracy than conventional techniques.The success of this approach also shows the applicability of deep learning in mobile security along with the direction for the future advancement of the real-time detection system and other deeper learning techniques to counter the increasing number of threats emerging in the future.
基金This work was funded by the Deanship of Scientific Research at Jouf University under Grant No.(DSR-2022-RG-0104).
文摘Palmprint identification has been conducted over the last two decades in many biometric systems.High-dimensional data with many uncorrelated and duplicated features remains difficult due to several computational complexity issues.This paper presents an interactive authentication approach based on deep learning and feature selection that supports Palmprint authentication.The proposed model has two stages of learning;the first stage is to transfer pre-trained VGG-16 of ImageNet to specific features based on the extraction model.The second stage involves the VGG-16 Palmprint feature extraction in the Siamese network to learn Palmprint similarity.The proposed model achieves robust and reliable end-to-end Palmprint authentication by extracting the convolutional features using VGG-16 Palmprint and the similarity of two input Palmprint using the Siamese network.The second stage uses the CASIA dataset to train and test the Siamese network.The suggested model outperforms comparable studies based on the deep learning approach achieving accuracy and EER of 91.8%and 0.082%,respectively,on the CASIA left-hand images and accuracy and EER of 91.7%and 0.084,respectively,on the CASIA right-hand images.
文摘The nutritional value of perishable food items,such as fruits and vegetables,depends on their freshness levels.The existing approaches solve a binary class problem by classifying a known fruit\vegetable class into fresh or rotten only.We propose an automated fruits and vegetables categorization approach that first recognizes the class of object in an image and then categorizes that fruit or vegetable into one of the three categories:purefresh,medium-fresh,and rotten.We gathered a dataset comprising of 60K images of 11 fruits and vegetables,each is further divided into three categories of freshness,using hand-held cameras.The recognition and categorization of fruits and vegetables are performed through two deep learning models:Visual Geometry Group(VGG-16)and You Only Look Once(YOLO),and their results are compared.VGG-16 classifies fruits and vegetables and categorizes their freshness,while YOLO also localizes them within the image.Furthermore,we have developed an android based application that takes the image of the fruit or vegetable as input and returns its class label and its freshness degree.A comprehensive experimental evaluation of proposed approach demonstrates that the proposed approach can achieve a high accuracy and F1score on gathered FruitVeg Freshness dataset.The dataset is publicly available for further evaluation by the research community.
基金Supporting this research through Taif University Researchers Supporting Project number(TURSP-2020/231),Taif University,Taif,Saudi Arabia.
文摘Nowadays,the COVID-19 virus disease is spreading rampantly.There are some testing tools and kits available for diagnosing the virus,but it is in a lim-ited count.To diagnose the presence of disease from radiological images,auto-mated COVID-19 diagnosis techniques are needed.The enhancement of AI(Artificial Intelligence)has been focused in previous research,which uses X-ray images for detecting COVID-19.The most common symptoms of COVID-19 are fever,dry cough and sore throat.These symptoms may lead to an increase in the rigorous type of pneumonia with a severe barrier.Since medical imaging is not suggested recently in Canada for critical COVID-19 diagnosis,computer-aided systems are implemented for the early identification of COVID-19,which aids in noticing the disease progression and thus decreases the death rate.Here,a deep learning-based automated method for the extraction of features and classi-fication is enhanced for the detection of COVID-19 from the images of computer tomography(CT).The suggested method functions on the basis of three main pro-cesses:data preprocessing,the extraction of features and classification.This approach integrates the union of deep features with the help of Inception 14 and VGG-16 models.At last,a classifier of Multi-scale Improved ResNet(MSI-ResNet)is developed to detect and classify the CT images into unique labels of class.With the support of available open-source COVID-CT datasets that consists of 760 CT pictures,the investigational validation of the suggested method is estimated.The experimental results reveal that the proposed approach offers greater performance with high specificity,accuracy and sensitivity.
基金funded by the Deanship of Scientific Research at Jouf University under Gran Number DSR–2022–RG–0101.
文摘The use of the internet is increasing all over the world on a daily basis in the last two decades.The increase in the internet causes many sexual crimes,such as sexual misuse,domestic violence,and child pornography.Various research has been done for pornographic image detection and classification.Most of the used models used machine learning techniques and deep learning models which show less accuracy,while the deep learning model ware used for classification and detection performed better as compared to machine learning.Therefore,this research evaluates the performance analysis of intelligent neural-based deep learning models which are based on Convolution neural network(CNN),Visual geometry group(VGG-16),VGG-14,and Residual Network(ResNet-50)with the expanded dataset,trained using transfer learning approaches applied in the fully connected layer for datasets to classify rank(Pornographic vs.Nonpornographic)classification in images.The simulation result shows that VGG-16 performed better than the used model in this study without augmented data.The VGG-16 model with augmented data reached a training and validation accuracy of 0.97,0.94 with a loss of 0.070,0.16.The precision,recall,and f-measure values for explicit and non-explicit images are(0.94,0.94,0.94)and(0.94,0.94,0.94).Similarly,The VGG-14 model with augmented data reached a training and validation accuracy of 0.98,0.96 with a loss of 0.059,0.11.The f-measure,recall,and precision values for explicit and non-explicit images are(0.98,0.98,0.98)and(0.98,0.98,0.98).The CNN model with augmented data reached a training and validation accuracy of 0.776&0.78 with losses of 0.48&0.46.The f-measure,recall,and precision values for explicit and non-explicit images are(0.80,0.80,0.80)and(0.78,0.79,0.78).The ResNet-50 model with expanded data reached with training accuracy of 0.89 with a loss of 0.389 and 0.86 of validation accuracy and a loss of 0.47.The f-measure,recall,and precision values for explicit and non-explicit images are(0.86,0.97,0.91)and(0.86,0.93,0.89).Where else without augmented data the VGG-16 model reached a training and validation accuracy of 0.997,0.986 with a loss of 0.008,0.056.The f-measure,recall,and precision values for explicit and non-explicit images are(0.94,0.99,0.97)and(0.99,0.93,0.96)which outperforms the used models with the augmented dataset in this study.
基金supported by Grant Number 13-INF2456-08-R from King Abdul Aziz City for Sciences and Technology(KACST),Riyadh,Kingdom of Saudi Arabia.
文摘Currently,distracted driving is among the most important causes of traffic accidents.Consequently,intelligent vehicle driving systems have become increasingly important.Recently,interest in driver-assistance systems that detect driver actions and help them drive safely has increased.In these studies,although some distinct data types,such as the physical conditions of the driver,audio and visual features,and vehicle information,are used,the primary data source is images of the driver that include the face,arms,and hands taken with a camera inside the car.In this study,an architecture based on a convolution neural network(CNN)is proposed to classify and detect driver distraction.An efficient CNN with high accuracy is implemented,and to implement intense convolutional networks for large-scale image recognition,a new architecture was proposed based on the available Visual Geometry Group(VGG-16)architecture.The proposed architecture was evaluated using the StateFarm dataset for driver-distraction detection.This dataset is publicly available on Kaggle and is frequently used for this type of research.The proposed architecture achieved 96.95%accuracy.
文摘Aiming at the problems such as diverse target scales and large-scale changes in crowds in dense crowd scenarios,a crowd density estimation method based on multi-scale feature fusion and information en-hancement is proposed.Firstly,considering that small-scale targets account for a relatively large proportion in the image,based on the VGG-16 network,the dilated convolution module is introduced to mine the detailed information of the image.Secondly,in order to make full use of the multi-scale information of the target,a new context-aware module is constructed to extract the contrast features between different scales.Finally,con-sidering the characteristic of continuous changes in the target scale,a multi-scale feature aggregation module is designed to enhance the sampling range of dense scales and multi-scale information interaction,thereby improving the network performance.Experiments on public datasets show that the proposed method in this paper can effectively estimate the population density compared with other advanced methods.