In order to prevent possible casualties and economic loss, it is critical to accurate prediction of the Remaining Useful Life (RUL) in rail prognostics health management. However, the traditional neural networks is di...In order to prevent possible casualties and economic loss, it is critical to accurate prediction of the Remaining Useful Life (RUL) in rail prognostics health management. However, the traditional neural networks is difficult to capture the long-term dependency relationship of the time series in the modeling of the long time series of rail damage, due to the coupling relationship of multi-channel data from multiple sensors. Here, in this paper, a novel RUL prediction model with an enhanced pulse separable convolution is used to solve this issue. Firstly, a coding module based on the improved pulse separable convolutional network is established to effectively model the relationship between the data. To enhance the network, an alternate gradient back propagation method is implemented. And an efficient channel attention (ECA) mechanism is developed for better emphasizing the useful pulse characteristics. Secondly, an optimized Transformer encoder was designed to serve as the backbone of the model. It has the ability to efficiently understand relationship between the data itself and each other at each time step of long time series with a full life cycle. More importantly, the Transformer encoder is improved by integrating pulse maximum pooling to retain more pulse timing characteristics. Finally, based on the characteristics of the front layer, the final predicted RUL value was provided and served as the end-to-end solution. The empirical findings validate the efficacy of the suggested approach in forecasting the rail RUL, surpassing various existing data-driven prognostication techniques. Meanwhile, the proposed method also shows good generalization performance on PHM2012 bearing data set.展开更多
Recently,video-based fire detection technology has become an important research topic in the field of machine vision.This paper proposes a method of combining the classification model and target detection model in dee...Recently,video-based fire detection technology has become an important research topic in the field of machine vision.This paper proposes a method of combining the classification model and target detection model in deep learning for fire detection.Firstly,the depthwise separable convolution is used to classify fire images,which saves a lot of detection time under the premise of ensuring detection accuracy.Secondly,You Only Look Once version 3(YOLOv3)target regression function is used to output the fire position information for the images whose classification result is fire,which avoids the problem that the accuracy of detection cannot be guaranteed by using YOLOv3 for target classification and position regression.At the same time,the detection time of target regression for images without fire is greatly reduced saved.The experiments were tested using a network public database.The detection accuracy reached 98%and the detection rate reached 38fps.This method not only saves the workload of manually extracting flame characteristics,reduces the calculation cost,and reduces the amount of parameters,but also improves the detection accuracy and detection rate.展开更多
One of the most obvious clinical reasons of dementia or The Behavioral and Psychological Symptoms of Dementia(BPSD)are the lack of emotional expression,the increased frequency of negative emotions,and the impermanence...One of the most obvious clinical reasons of dementia or The Behavioral and Psychological Symptoms of Dementia(BPSD)are the lack of emotional expression,the increased frequency of negative emotions,and the impermanence of emotions.Observing the reduction of BPSD in dementia through emotions can be considered effective and widely used in the field of non-pharmacological therapy.At present,this article will verify whether the image recognition artificial intelligence(AI)system can correctly reflect the emotional performance of the elderly with dementia through a questionnaire survey of three professional elderly nursing staff.The ANOVA(sig.=0.50)is used to determine that the judgment given by the nursing staff has no obvious deviation,and then Kendall's test(0.722**)and spearman's test(0.863**)are used to verify the judgment severity of the emotion recognition system and the nursing staff unanimously.This implies the usability of the tool.Additionally,it can be expected to be further applied in the research related to BPSD elderly emotion detection.展开更多
The distinctive fault characteristics of battery energy storage stations(BESSs)significantly affect the reliability of conventional protection methods for transmission lines.In this paper,the three-dimensional(3D)data...The distinctive fault characteristics of battery energy storage stations(BESSs)significantly affect the reliability of conventional protection methods for transmission lines.In this paper,the three-dimensional(3D)data scattergrams are constructed using current data from both sides of the transmission line and their sum.Following a comprehensive analysis of the varying characteristics of 3D data scattergrams under different conditions,a 3D data scattergram image classification based protection method is developed.The depth-wise separable convolution is used to ensure a lightweight convolutional neural network(CNN)structure without compromising performance.In addition,a Bayesian hyperparameter optimization algorithm is used to achieve a hyperparametric search to simplify the training process.Compared with artificial neural networks and CNNs,the depth-wise separable convolution based CNN(DPCNN)achieves a higher recognition accuracy.The 3D data scattergram image classification based protection method using DPCNN can accurately separate internal faults from other disturbances and identify fault phases under different operating states and fault conditions.The proposed protection method also shows first-class tolerability against current transformer(CT)saturation and CT measurement errors.展开更多
Traditional data-driven fault diagnosis methods depend on expert experience to manually extract effective fault features of signals,which has certain limitations.Conversely,deep learning techniques have gained promine...Traditional data-driven fault diagnosis methods depend on expert experience to manually extract effective fault features of signals,which has certain limitations.Conversely,deep learning techniques have gained prominence as a central focus of research in the field of fault diagnosis by strong fault feature extraction ability and end-to-end fault diagnosis efficiency.Recently,utilizing the respective advantages of convolution neural network(CNN)and Transformer in local and global feature extraction,research on cooperating the two have demonstrated promise in the field of fault diagnosis.However,the cross-channel convolution mechanism in CNN and the self-attention calculations in Transformer contribute to excessive complexity in the cooperative model.This complexity results in high computational costs and limited industrial applicability.To tackle the above challenges,this paper proposes a lightweight CNN-Transformer named as SEFormer for rotating machinery fault diagnosis.First,a separable multiscale depthwise convolution block is designed to extract and integrate multiscale feature information from different channel dimensions of vibration signals.Then,an efficient self-attention block is developed to capture critical fine-grained features of the signal from a global perspective.Finally,experimental results on the planetary gearbox dataset and themotor roller bearing dataset prove that the proposed framework can balance the advantages of robustness,generalization and lightweight compared to recent state-of-the-art fault diagnosis models based on CNN and Transformer.This study presents a feasible strategy for developing a lightweight rotating machinery fault diagnosis framework aimed at economical deployment.展开更多
In this paper,a new bolt fault diagnosis method is developed to solve the fault diagnosis problem of wind turbine flange bolts using one-dimensional depthwise separable convolutions.The main idea is to use a one-dimen...In this paper,a new bolt fault diagnosis method is developed to solve the fault diagnosis problem of wind turbine flange bolts using one-dimensional depthwise separable convolutions.The main idea is to use a one-dimensional convolutional neural network model to classify and identify the acoustic vibration signals of bolts,which represent different bolt damage states.Through the methods of knock test and modal simulation,it is concluded that the damage state of wind turbine flange bolt is related to the natural frequency distribution of acoustic vibration signal.It is found that the bolt damage state affects the modal shape of the structure,and then affects the natural frequency distribution of the bolt vibration signal.Therefore,the damage state can be obtained by identifying the natural frequency distribution of the bolt acoustic vibration signal.In the present one-dimensional depth-detachable convolutional neural network model,the one-dimensional vector is first convolved into multiple channels,and then each channel is separately learned by depth-detachable convolution,which can effectively improve the feature quality and the effect of data classification.From the perspective of the realization mechanism of convolution operation,the depthwise separable convolution operation has fewer parameters and faster computing speed,making it easier to build lightweight models and deploy them to mobile devices.展开更多
Channel pruning can reduce memory consumption and running time with least performance damage,and is one of the most important techniques in network compression.However,existing channel pruning methods mainly focus on ...Channel pruning can reduce memory consumption and running time with least performance damage,and is one of the most important techniques in network compression.However,existing channel pruning methods mainly focus on the pruning of standard convolutional networks,and they rely intensively on time-consuming fine-tuning to achieve the performance improvement.To this end,we present a novel efficient probability-based channel pruning method for depthwise separable convolutional networks.Our method leverages a new simple yet effective probability-based channel pruning criterion by taking the scaling and shifting factors of batch normalization layers into consideration.A novel shifting factor fusion technique is further developed to improve the performance of the pruned networks without requiring extra time-consuming fine-tuning.We apply the proposed method to five representative deep learning networks,namely MobileNetV1,MobileNetV2,ShuffleNetV1,ShuffleNetV2,and GhostNet,to demonstrate the efficiency of our pruning method.Extensive experimental results and comparisons on publicly available CIFAR10,CIFAR100,and ImageNet datasets validate the feasibility of the proposed method.展开更多
Image classification using Convolutional Neural Network(CNN)achieves optimal perfor-mance with a particular strategy.MobileNet reduces the parameter number for learning features by switching from the standard convolut...Image classification using Convolutional Neural Network(CNN)achieves optimal perfor-mance with a particular strategy.MobileNet reduces the parameter number for learning features by switching from the standard convolution paradigm to the depthwise separable convolution(DSC)paradigm.However,there are not enough features to learn for identify-ing the freshness of fish eyes.Furthermore,minor variances in features should not require complicated CNN architecture.In this paper,our first contribution proposed DSC Bottle-neck with Expansion for learning features of the freshness of fish eyes with a Bottleneck Multiplier.The second contribution proposed Residual Transition to bridge current feature maps and skip connection feature maps to the next convolution block.The third contribu-tion proposed MobileNetV1 Bottleneck with Expansion(MB-BE)for classifying the freshness of fish eyes.The result obtained from the Freshness of the Fish Eyes dataset shows that MB-BE outperformed other models such as original MobileNet,VGG16,Densenet,Nasnet Mobile with 63.21%accuracy.展开更多
Convolutional neural networks (CNNs) are widely used in image classification tasks, but their increasing model size and computation make them challenging to implement on embedded systems with constrained hardware reso...Convolutional neural networks (CNNs) are widely used in image classification tasks, but their increasing model size and computation make them challenging to implement on embedded systems with constrained hardware resources. To address this issue, the MobileNetV1 network was developed, which employs depthwise convolution to reduce network complexity. MobileNetV1 employs a stride of 2 in several convolutional layers to decrease the spatial resolution of feature maps, thereby lowering computational costs. However, this stride setting can lead to a loss of spatial information, particularly affecting the detection and representation of smaller objects or finer details in images. To maintain the trade-off between complexity and model performance, a lightweight convolutional neural network with hierarchical multi-scale feature fusion based on the MobileNetV1 network is proposed. The network consists of two main subnetworks. The first subnetwork uses a depthwise dilated separable convolution (DDSC) layer to learn imaging features with fewer parameters, which results in a lightweight and computationally inexpensive network. Furthermore, depthwise dilated convolution in DDSC layer effectively expands the field of view of filters, allowing them to incorporate a larger context. The second subnetwork is a hierarchical multi-scale feature fusion (HMFF) module that uses parallel multi-resolution branches architecture to process the input feature map in order to extract the multi-scale feature information of the input image. Experimental results on the CIFAR-10, Malaria, and KvasirV1 datasets demonstrate that the proposed method is efficient, reducing the network parameters and computational cost by 65.02% and 39.78%, respectively, while maintaining the network performance compared to the MobileNetV1 baseline.展开更多
Background The use of remote photoplethysmography(rPPG)to estimate blood volume pulse in a noncontact manner has been an active research topic in recent years.Existing methods are primarily based on a singlescale regi...Background The use of remote photoplethysmography(rPPG)to estimate blood volume pulse in a noncontact manner has been an active research topic in recent years.Existing methods are primarily based on a singlescale region of interest(ROI).However,some noise signals that are not easily separated in a single-scale space can be easily separated in a multi-scale space.Also,existing spatiotemporal networks mainly focus on local spatiotemporal information and do not emphasize temporal information,which is crucial in pulse extraction problems,resulting in insufficient spatiotemporal feature modelling.Methods Here,we propose a multi-scale facial video pulse extraction network based on separable spatiotemporal convolution(SSTC)and dimension separable attention(DSAT).First,to solve the problem of a single-scale ROI,we constructed a multi-scale feature space for initial signal separation.Second,SSTC and DSAT were designed for efficient spatiotemporal correlation modeling,which increased the information interaction between the long-span time and space dimensions;this placed more emphasis on temporal features.Results The signal-to-noise ratio(SNR)of the proposed network reached 9.58dB on the PURE dataset and 6.77dB on the UBFC-rPPG dataset,outperforming state-of-the-art algorithms.Conclusions The results showed that fusing multi-scale signals yielded better results than methods based on only single-scale signals.The proposed SSTC and dimension-separable attention mechanism will contribute to more accurate pulse signal extraction.展开更多
Multiple suppression is an important element of marine seismic data processing.Intelligent suppression of multiples us-ing artificial intelligence reduces labor costs,minimizes dependence on unknown prior information,...Multiple suppression is an important element of marine seismic data processing.Intelligent suppression of multiples us-ing artificial intelligence reduces labor costs,minimizes dependence on unknown prior information,and improves data processing ef-ficiency.In this study,we propose an intelligent method for suppressing marine seismic multiples using deep learning approaches.The proposed method enables the intelligent suppression of free-surface-related multiples from seismic records.Initially,we construct a multi-category marine seismic multiple dataset through finite difference forward modeling under different boundary conditions.We use various models and data augmentation methods,including sample rotation,noise addition,and random channel omission.Then,we apply depthwise separable convolution to develop our deep learning Mobilenet-Unet model.The Mobilenet-Unet framework sig-nificantly reduces the number of operations required for multiple elimination without sacrificing model performance,ultimately reali-zing the optimal multiple suppression model.The trained Mobilenet-Unet is applied to the test set for verification.Moreover,to deter-mine its generalization ability,it is implemented to seismic records containing multiples generated by two marine geophysical models that were not included in the training process.The performance of Mobilenet-Unet is also compared with that of different network structures.The results indicate that,despite its small size,our proposed Mobilenet-Unet deep learning model can rapidly and effective-ly separate multiples in marine seismic data,possessing reasonable generalization ability.展开更多
To address the limitations of existing abnormal traffic detection methods,such as insufficient temporal and spatial feature extraction,high false positive rate(FPR),poor generalization,and class imbalance,this study p...To address the limitations of existing abnormal traffic detection methods,such as insufficient temporal and spatial feature extraction,high false positive rate(FPR),poor generalization,and class imbalance,this study proposed an intelligent detection method that combines a Stacked Convolutional Network(SCN),Bidirectional Long Short-Term Memory(BiLSTM)network,and Equalization Loss v2(EQL v2).This method was divided into two components:a feature extraction model and a classification and detection model.First,SCN was constructed by combining a Convolutional Neural Network(CNN)with a Depthwise Separable Convolution(DSC)network to capture the abstract spatial features of traffic data.These features were then input into the BiLSTM to capture temporal dependencies.An attention mechanism was incorporated after SCN and BiLSTM to enhance the extraction of key spatiotemporal features.To address class imbalance,the classification detection model applied EQL v2 to adjust the weights of the minority classes,ensuring that they received equal focus during training.The experimental results indicated that the proposed method outperformed the existing methods in terms of accuracy,FPR,and F1-score and significantly improved the identification rate of minority classes.展开更多
The accurate and automatic segmentation of retinal vessels fromfundus images is critical for the early diagnosis and prevention ofmany eye diseases,such as diabetic retinopathy(DR).Existing retinal vessel segmentation...The accurate and automatic segmentation of retinal vessels fromfundus images is critical for the early diagnosis and prevention ofmany eye diseases,such as diabetic retinopathy(DR).Existing retinal vessel segmentation approaches based on convolutional neural networks(CNNs)have achieved remarkable effectiveness.Here,we extend a retinal vessel segmentation model with low complexity and high performance based on U-Net,which is one of the most popular architectures.In view of the excellent work of depth-wise separable convolution,we introduce it to replace the standard convolutional layer.The complexity of the proposed model is reduced by decreasing the number of parameters and calculations required for themodel.To ensure performance while lowering redundant parameters,we integrate the pre-trained MobileNet V2 into the encoder.Then,a feature fusion residual module(FFRM)is designed to facilitate complementary strengths by enhancing the effective fusion between adjacent levels,which alleviates extraneous clutter introduced by direct fusion.Finally,we provide detailed comparisons between the proposed SepFE and U-Net in three retinal image mainstream datasets(DRIVE,STARE,and CHASEDB1).The results show that the number of SepFE parameters is only 3%of U-Net,the Flops are only 8%of U-Net,and better segmentation performance is obtained.The superiority of SepFE is further demonstrated through comparisons with other advanced methods.展开更多
Deep kernel mapping support vector machines have achieved good results in numerous tasks by mapping features from a low-dimensional space to a high-dimensional space and then using support vector machines for classifi...Deep kernel mapping support vector machines have achieved good results in numerous tasks by mapping features from a low-dimensional space to a high-dimensional space and then using support vector machines for classification.However,the depth kernel mapping support vector machine does not take into account the connection of different dimensional spaces and increases the model parameters.To further improve the recognition capability of deep kernel mapping support vector machines while reducing the number of model parameters,this paper proposes a framework of Lightweight Deep Convolutional Cross-Connected Kernel Mapping Support Vector Machines(LC-CKMSVM).The framework consists of a feature extraction module and a classification module.The feature extraction module first maps the data from low-dimensional to high-dimensional space by fusing the representations of different dimensional spaces through cross-connections;then,it uses depthwise separable convolution to replace part of the original convolution to reduce the number of parameters in the module;The classification module uses a soft margin support vector machine for classification.The results on 6 different visual datasets show that LC-CKMSVM obtains better classification accuracies on most cases than the other five models.展开更多
With the widespread use of Internet of Things(IoT)technology in daily life and the considerable safety risks of falls for elderly individuals,research on IoT-based fall detection systems has gainedmuch attention.This ...With the widespread use of Internet of Things(IoT)technology in daily life and the considerable safety risks of falls for elderly individuals,research on IoT-based fall detection systems has gainedmuch attention.This paper proposes an IoT-based spatiotemporal data processing framework based on a depthwise separable convolution generative adversarial network using skip-connection(Skip-DSCGAN)for fall detection.The method uses spatiotemporal data from accelerometers and gyroscopes in inertial sensors as input data.A semisupervised learning approach is adopted to train the model using only activities of daily living(ADL)data,which can avoid data imbalance problems.Furthermore,a quantile-based approach is employed to determine the fall threshold,which makes the fall detection frameworkmore robust.This proposed fall detection framework is evaluated against four other generative adversarial network(GAN)models with superior anomaly detection performance using two fall public datasets(SisFall&MobiAct).The test results show that the proposed method achieves better results,reaching 96.93% and 92.75% accuracy on the above two test datasets,respectively.At the same time,the proposed method also achieves satisfactory results in terms ofmodel size and inference delay time,making it suitable for deployment on wearable devices with limited resources.In addition,this paper also compares GAN-based semisupervised learning methods with supervised learning methods commonly used in fall detection.It clarifies the advantages of GAN-based semisupervised learning methods in fall detection.展开更多
With the growth of the Internet,more and more business is being done online,for example,online offices,online education and so on.While this makes people’s lives more convenient,it also increases the risk of the netw...With the growth of the Internet,more and more business is being done online,for example,online offices,online education and so on.While this makes people’s lives more convenient,it also increases the risk of the network being attacked by malicious code.Therefore,it is important to identify malicious codes on computer systems efficiently.However,most of the existing malicious code detection methods have two problems:(1)The ability of the model to extract features is weak,resulting in poor model performance.(2)The large scale of model data leads to difficulties deploying on devices with limited resources.Therefore,this paper proposes a lightweight malicious code identification model Lightweight Malicious Code Classification Method Based on Improved SqueezeNet(LCMISNet).In this paper,the MFire lightweight feature extraction module is constructed by proposing a feature slicing module and a multi-size depthwise separable convolution module.The feature slicing module reduces the number of parameters by grouping features.The multi-size depthwise separable convolution module reduces the number of parameters and enhances the feature extraction capability by replacing the standard convolution with depthwise separable convolution with different convolution kernel sizes.In addition,this paper also proposes a feature splicing module to connect the MFire lightweight feature extraction module based on the feature reuse and constructs the lightweight model LCMISNet.The malicious code recognition accuracy of LCMISNet on the BIG 2015 dataset and the Malimg dataset reaches 98.90% and 99.58%,respectively.It proves that LCMISNet has a powerful malicious code recognition performance.In addition,compared with other network models,LCMISNet has better performance,and a lower number of parameters and computations.展开更多
In the model of the vehicle recognition algorithm implemented by the convolutional neural network,the model needs to compute and store a lot of parameters.Too many parameters occupy a lot of computational resources ma...In the model of the vehicle recognition algorithm implemented by the convolutional neural network,the model needs to compute and store a lot of parameters.Too many parameters occupy a lot of computational resources making it difficult to run on computers with poor performance.Therefore,obtaining more efficient feature information of target image or video with better accuracy on computers with limited arithmetic power becomes the main goal of this research.In this paper,a lightweight densely connected,and deeply separable convolutional network(DCDSNet)algorithmis proposed to achieve this goal.Visual Geometry Group(VGG)model is improved by utilizing the convolution instead of the fully connected module,the deeply separable convolution module,and the densely connected network module,with the first two modules reducing the parameters and the third module allowing the algorithm to have more features in a limited number of parameters.The algorithm achieves better results in the mine vehicle recognition dataset.Experiments show that the recognition accuracy is improved by 4.41% compared to VGG19 and the amount of parameters is reduced by 71% compared to VGG19.展开更多
Multispectral pedestrian detection technology leverages infrared images to provide reliable information for visible light images, demonstrating significant advantages in low-light conditions and background occlusion s...Multispectral pedestrian detection technology leverages infrared images to provide reliable information for visible light images, demonstrating significant advantages in low-light conditions and background occlusion scenarios. However, while continuously improving cross-modal feature extraction and fusion, ensuring the model’s detection speed is also a challenging issue. We have devised a deep learning network model for cross-modal pedestrian detection based on Resnet50, aiming to focus on more reliable features and enhance the model’s detection efficiency. This model employs a spatial attention mechanism to reweight the input visible light and infrared image data, enhancing the model’s focus on different spatial positions and sharing the weighted feature data across different modalities, thereby reducing the interference of multi-modal features. Subsequently, lightweight modules with depthwise separable convolution are incorporated to reduce the model’s parameter count and computational load through channel-wise and point-wise convolutions. The network model algorithm proposed in this paper was experimentally validated on the publicly available KAIST dataset and compared with other existing methods. The experimental results demonstrate that our approach achieves favorable performance in various complex environments, affirming the effectiveness of the multispectral pedestrian detection technology proposed in this paper.展开更多
Birds play a crucial role in maintaining ecological balance,making bird recognition technology a hot research topic.Traditional recognition methods have not achieved high accuracy in bird identification.This paper pro...Birds play a crucial role in maintaining ecological balance,making bird recognition technology a hot research topic.Traditional recognition methods have not achieved high accuracy in bird identification.This paper proposes an improved ResNet18 model to enhance the recognition rate of local bird species in Yunnan.First,a dataset containing five species of local birds in Yunnan was established:C.amherstiae,T.caboti,Syrmaticus humiae,Polyplectron bicalcaratum,and Pucrasia macrolopha.The improved ResNet18 model was then used to identify these species.This method replaces traditional convolution with depth wise separable convolution and introduces an SE(Squeeze and Excitation)module to improve the model’s efficiency and accuracy.Compared to the traditional ResNet18 model,this improved model excels in implementing a wild bird classification solution,significantly reducing computational overhead and accelerating model training using low-power,lightweight hardware.Experimental analysis shows that the improved ResNet18 model achieved an accuracy of 98.57%,compared to 98.26%for the traditional Residual Network 18 layers(ResNet18)model.展开更多
The basic theory of YOLO series object detection algorithms is discussed, the dangerous driving behavior dataset is collected and produced, and then the YOLOv7 network is introduced in detail, the deep separable convo...The basic theory of YOLO series object detection algorithms is discussed, the dangerous driving behavior dataset is collected and produced, and then the YOLOv7 network is introduced in detail, the deep separable convolution and CA attention mechanism are introduced, the YOLOv7 bounding box loss function and clustering algorithm are optimized, and the DB-YOLOv7 network structure is constructed. In the first stage of the experiment, the PASCAL VOC public dataset was utilized for pre-training. A comparative analysis was conducted to assess the recognition accuracy and inference time before and after the proposed improvements. The experimental results demonstrated an increase of 1.4% in the average recognition accuracy, alongside a reduction in the inference time by 4 ms. Subsequently, a model for the recognition of dangerous driving behaviors was trained using a specialized dangerous driving behavior dataset. A series of experiments were performed to evaluate the efficacy of the DB-YOLOv7 algorithm in this context. The findings indicate a significant enhancement in detection performance, with a 4% improvement in accuracy compared to the baseline network. Furthermore, the model’s inference time was reduced by 20%, from 25 ms to 20 ms. These results substantiate the effectiveness of the DB-YOLOv7 recognition algorithm for detecting dangerous driving behaviors, providing comprehensive validation of its practical applicability.展开更多
文摘In order to prevent possible casualties and economic loss, it is critical to accurate prediction of the Remaining Useful Life (RUL) in rail prognostics health management. However, the traditional neural networks is difficult to capture the long-term dependency relationship of the time series in the modeling of the long time series of rail damage, due to the coupling relationship of multi-channel data from multiple sensors. Here, in this paper, a novel RUL prediction model with an enhanced pulse separable convolution is used to solve this issue. Firstly, a coding module based on the improved pulse separable convolutional network is established to effectively model the relationship between the data. To enhance the network, an alternate gradient back propagation method is implemented. And an efficient channel attention (ECA) mechanism is developed for better emphasizing the useful pulse characteristics. Secondly, an optimized Transformer encoder was designed to serve as the backbone of the model. It has the ability to efficiently understand relationship between the data itself and each other at each time step of long time series with a full life cycle. More importantly, the Transformer encoder is improved by integrating pulse maximum pooling to retain more pulse timing characteristics. Finally, based on the characteristics of the front layer, the final predicted RUL value was provided and served as the end-to-end solution. The empirical findings validate the efficacy of the suggested approach in forecasting the rail RUL, surpassing various existing data-driven prognostication techniques. Meanwhile, the proposed method also shows good generalization performance on PHM2012 bearing data set.
基金This work was supported by Liaoning Provincial Science Public Welfare Research Fund Project(No.2016002006)Liaoning Provincial Department of Education Scientific Research Service Local Project(No.L201708).
文摘Recently,video-based fire detection technology has become an important research topic in the field of machine vision.This paper proposes a method of combining the classification model and target detection model in deep learning for fire detection.Firstly,the depthwise separable convolution is used to classify fire images,which saves a lot of detection time under the premise of ensuring detection accuracy.Secondly,You Only Look Once version 3(YOLOv3)target regression function is used to output the fire position information for the images whose classification result is fire,which avoids the problem that the accuracy of detection cannot be guaranteed by using YOLOv3 for target classification and position regression.At the same time,the detection time of target regression for images without fire is greatly reduced saved.The experiments were tested using a network public database.The detection accuracy reached 98%and the detection rate reached 38fps.This method not only saves the workload of manually extracting flame characteristics,reduces the calculation cost,and reduces the amount of parameters,but also improves the detection accuracy and detection rate.
文摘One of the most obvious clinical reasons of dementia or The Behavioral and Psychological Symptoms of Dementia(BPSD)are the lack of emotional expression,the increased frequency of negative emotions,and the impermanence of emotions.Observing the reduction of BPSD in dementia through emotions can be considered effective and widely used in the field of non-pharmacological therapy.At present,this article will verify whether the image recognition artificial intelligence(AI)system can correctly reflect the emotional performance of the elderly with dementia through a questionnaire survey of three professional elderly nursing staff.The ANOVA(sig.=0.50)is used to determine that the judgment given by the nursing staff has no obvious deviation,and then Kendall's test(0.722**)and spearman's test(0.863**)are used to verify the judgment severity of the emotion recognition system and the nursing staff unanimously.This implies the usability of the tool.Additionally,it can be expected to be further applied in the research related to BPSD elderly emotion detection.
基金supported by the Fundamental Research Funds for Central Universities(No.2024JCCXJD01).
文摘The distinctive fault characteristics of battery energy storage stations(BESSs)significantly affect the reliability of conventional protection methods for transmission lines.In this paper,the three-dimensional(3D)data scattergrams are constructed using current data from both sides of the transmission line and their sum.Following a comprehensive analysis of the varying characteristics of 3D data scattergrams under different conditions,a 3D data scattergram image classification based protection method is developed.The depth-wise separable convolution is used to ensure a lightweight convolutional neural network(CNN)structure without compromising performance.In addition,a Bayesian hyperparameter optimization algorithm is used to achieve a hyperparametric search to simplify the training process.Compared with artificial neural networks and CNNs,the depth-wise separable convolution based CNN(DPCNN)achieves a higher recognition accuracy.The 3D data scattergram image classification based protection method using DPCNN can accurately separate internal faults from other disturbances and identify fault phases under different operating states and fault conditions.The proposed protection method also shows first-class tolerability against current transformer(CT)saturation and CT measurement errors.
基金supported by the National Natural Science Foundation of China(No.52277055).
文摘Traditional data-driven fault diagnosis methods depend on expert experience to manually extract effective fault features of signals,which has certain limitations.Conversely,deep learning techniques have gained prominence as a central focus of research in the field of fault diagnosis by strong fault feature extraction ability and end-to-end fault diagnosis efficiency.Recently,utilizing the respective advantages of convolution neural network(CNN)and Transformer in local and global feature extraction,research on cooperating the two have demonstrated promise in the field of fault diagnosis.However,the cross-channel convolution mechanism in CNN and the self-attention calculations in Transformer contribute to excessive complexity in the cooperative model.This complexity results in high computational costs and limited industrial applicability.To tackle the above challenges,this paper proposes a lightweight CNN-Transformer named as SEFormer for rotating machinery fault diagnosis.First,a separable multiscale depthwise convolution block is designed to extract and integrate multiscale feature information from different channel dimensions of vibration signals.Then,an efficient self-attention block is developed to capture critical fine-grained features of the signal from a global perspective.Finally,experimental results on the planetary gearbox dataset and themotor roller bearing dataset prove that the proposed framework can balance the advantages of robustness,generalization and lightweight compared to recent state-of-the-art fault diagnosis models based on CNN and Transformer.This study presents a feasible strategy for developing a lightweight rotating machinery fault diagnosis framework aimed at economical deployment.
基金supported in part by the National Key R&D Program of China(Nos.2021YFE0206100 and 2018YFB1702300)the National Natural Science Foundation of China(No.62073321)+1 种基金the National Defense Basic Scientific Research Program(No.JCKY2019203C029)the Science and Technology Development Fund,Macao SAR(No.0015/2020/AMJ).
文摘In this paper,a new bolt fault diagnosis method is developed to solve the fault diagnosis problem of wind turbine flange bolts using one-dimensional depthwise separable convolutions.The main idea is to use a one-dimensional convolutional neural network model to classify and identify the acoustic vibration signals of bolts,which represent different bolt damage states.Through the methods of knock test and modal simulation,it is concluded that the damage state of wind turbine flange bolt is related to the natural frequency distribution of acoustic vibration signal.It is found that the bolt damage state affects the modal shape of the structure,and then affects the natural frequency distribution of the bolt vibration signal.Therefore,the damage state can be obtained by identifying the natural frequency distribution of the bolt acoustic vibration signal.In the present one-dimensional depth-detachable convolutional neural network model,the one-dimensional vector is first convolved into multiple channels,and then each channel is separately learned by depth-detachable convolution,which can effectively improve the feature quality and the effect of data classification.From the perspective of the realization mechanism of convolution operation,the depthwise separable convolution operation has fewer parameters and faster computing speed,making it easier to build lightweight models and deploy them to mobile devices.
基金the National Natural Science Foundation of China under Grant Nos.62036010 and 62072340the Zhejiang Provincial Natural Science Foundation of China under Grant Nos.LZ21F020001 and LSZ19F020001the Open Project Program of the State Key Laboratory of CAD&CG,Zhejiang University under Grant No.A2220.
文摘Channel pruning can reduce memory consumption and running time with least performance damage,and is one of the most important techniques in network compression.However,existing channel pruning methods mainly focus on the pruning of standard convolutional networks,and they rely intensively on time-consuming fine-tuning to achieve the performance improvement.To this end,we present a novel efficient probability-based channel pruning method for depthwise separable convolutional networks.Our method leverages a new simple yet effective probability-based channel pruning criterion by taking the scaling and shifting factors of batch normalization layers into consideration.A novel shifting factor fusion technique is further developed to improve the performance of the pruned networks without requiring extra time-consuming fine-tuning.We apply the proposed method to five representative deep learning networks,namely MobileNetV1,MobileNetV2,ShuffleNetV1,ShuffleNetV2,and GhostNet,to demonstrate the efficiency of our pruning method.Extensive experimental results and comparisons on publicly available CIFAR10,CIFAR100,and ImageNet datasets validate the feasibility of the proposed method.
文摘Image classification using Convolutional Neural Network(CNN)achieves optimal perfor-mance with a particular strategy.MobileNet reduces the parameter number for learning features by switching from the standard convolution paradigm to the depthwise separable convolution(DSC)paradigm.However,there are not enough features to learn for identify-ing the freshness of fish eyes.Furthermore,minor variances in features should not require complicated CNN architecture.In this paper,our first contribution proposed DSC Bottle-neck with Expansion for learning features of the freshness of fish eyes with a Bottleneck Multiplier.The second contribution proposed Residual Transition to bridge current feature maps and skip connection feature maps to the next convolution block.The third contribu-tion proposed MobileNetV1 Bottleneck with Expansion(MB-BE)for classifying the freshness of fish eyes.The result obtained from the Freshness of the Fish Eyes dataset shows that MB-BE outperformed other models such as original MobileNet,VGG16,Densenet,Nasnet Mobile with 63.21%accuracy.
文摘Convolutional neural networks (CNNs) are widely used in image classification tasks, but their increasing model size and computation make them challenging to implement on embedded systems with constrained hardware resources. To address this issue, the MobileNetV1 network was developed, which employs depthwise convolution to reduce network complexity. MobileNetV1 employs a stride of 2 in several convolutional layers to decrease the spatial resolution of feature maps, thereby lowering computational costs. However, this stride setting can lead to a loss of spatial information, particularly affecting the detection and representation of smaller objects or finer details in images. To maintain the trade-off between complexity and model performance, a lightweight convolutional neural network with hierarchical multi-scale feature fusion based on the MobileNetV1 network is proposed. The network consists of two main subnetworks. The first subnetwork uses a depthwise dilated separable convolution (DDSC) layer to learn imaging features with fewer parameters, which results in a lightweight and computationally inexpensive network. Furthermore, depthwise dilated convolution in DDSC layer effectively expands the field of view of filters, allowing them to incorporate a larger context. The second subnetwork is a hierarchical multi-scale feature fusion (HMFF) module that uses parallel multi-resolution branches architecture to process the input feature map in order to extract the multi-scale feature information of the input image. Experimental results on the CIFAR-10, Malaria, and KvasirV1 datasets demonstrate that the proposed method is efficient, reducing the network parameters and computational cost by 65.02% and 39.78%, respectively, while maintaining the network performance compared to the MobileNetV1 baseline.
基金Supported by the National Natural Science Foundation of China(61903336,61976190)the Natural Science Foundation of Zhejiang Province(LY21F030015)。
文摘Background The use of remote photoplethysmography(rPPG)to estimate blood volume pulse in a noncontact manner has been an active research topic in recent years.Existing methods are primarily based on a singlescale region of interest(ROI).However,some noise signals that are not easily separated in a single-scale space can be easily separated in a multi-scale space.Also,existing spatiotemporal networks mainly focus on local spatiotemporal information and do not emphasize temporal information,which is crucial in pulse extraction problems,resulting in insufficient spatiotemporal feature modelling.Methods Here,we propose a multi-scale facial video pulse extraction network based on separable spatiotemporal convolution(SSTC)and dimension separable attention(DSAT).First,to solve the problem of a single-scale ROI,we constructed a multi-scale feature space for initial signal separation.Second,SSTC and DSAT were designed for efficient spatiotemporal correlation modeling,which increased the information interaction between the long-span time and space dimensions;this placed more emphasis on temporal features.Results The signal-to-noise ratio(SNR)of the proposed network reached 9.58dB on the PURE dataset and 6.77dB on the UBFC-rPPG dataset,outperforming state-of-the-art algorithms.Conclusions The results showed that fusing multi-scale signals yielded better results than methods based on only single-scale signals.The proposed SSTC and dimension-separable attention mechanism will contribute to more accurate pulse signal extraction.
基金supported by the Key Laboratory of Ma-rine Mineral Resources,Ministry of Natural Resources,Guangzhou(No.KLMMR-2022-G09)the Guangzhou Ba-sic Research Program-Basic and Basic Applied Research Project(No.2023A04J0917)the PI Project of South-ern Marine Science and Engineering Guangdong Labora-tory(Guangzhou)(No.GML2020GD0802).
文摘Multiple suppression is an important element of marine seismic data processing.Intelligent suppression of multiples us-ing artificial intelligence reduces labor costs,minimizes dependence on unknown prior information,and improves data processing ef-ficiency.In this study,we propose an intelligent method for suppressing marine seismic multiples using deep learning approaches.The proposed method enables the intelligent suppression of free-surface-related multiples from seismic records.Initially,we construct a multi-category marine seismic multiple dataset through finite difference forward modeling under different boundary conditions.We use various models and data augmentation methods,including sample rotation,noise addition,and random channel omission.Then,we apply depthwise separable convolution to develop our deep learning Mobilenet-Unet model.The Mobilenet-Unet framework sig-nificantly reduces the number of operations required for multiple elimination without sacrificing model performance,ultimately reali-zing the optimal multiple suppression model.The trained Mobilenet-Unet is applied to the test set for verification.Moreover,to deter-mine its generalization ability,it is implemented to seismic records containing multiples generated by two marine geophysical models that were not included in the training process.The performance of Mobilenet-Unet is also compared with that of different network structures.The results indicate that,despite its small size,our proposed Mobilenet-Unet deep learning model can rapidly and effective-ly separate multiples in marine seismic data,possessing reasonable generalization ability.
基金supported by the National Natural Science Foundation of China(Grant No.62102449).
文摘To address the limitations of existing abnormal traffic detection methods,such as insufficient temporal and spatial feature extraction,high false positive rate(FPR),poor generalization,and class imbalance,this study proposed an intelligent detection method that combines a Stacked Convolutional Network(SCN),Bidirectional Long Short-Term Memory(BiLSTM)network,and Equalization Loss v2(EQL v2).This method was divided into two components:a feature extraction model and a classification and detection model.First,SCN was constructed by combining a Convolutional Neural Network(CNN)with a Depthwise Separable Convolution(DSC)network to capture the abstract spatial features of traffic data.These features were then input into the BiLSTM to capture temporal dependencies.An attention mechanism was incorporated after SCN and BiLSTM to enhance the extraction of key spatiotemporal features.To address class imbalance,the classification detection model applied EQL v2 to adjust the weights of the minority classes,ensuring that they received equal focus during training.The experimental results indicated that the proposed method outperformed the existing methods in terms of accuracy,FPR,and F1-score and significantly improved the identification rate of minority classes.
基金supported by the Hunan Provincial Natural Science Foundation of China(2021JJ50074)the Scientific Research Fund of Hunan Provincial Education Department(19B082)+6 种基金the Science and Technology Development Center of the Ministry of Education-New Generation Information Technology Innovation Project(2018A02020)the Science Foundation of Hengyang Normal University(19QD12)the Science and Technology Plan Project of Hunan Province(2016TP1020)the Subject Group Construction Project of Hengyang Normal University(18XKQ02)theApplication Oriented SpecialDisciplines,Double First ClassUniversity Project of Hunan Province(Xiangjiaotong[2018]469)the Hunan Province Special Funds of Central Government for Guiding Local Science and Technology Development(2018CT5001)the First Class Undergraduate Major in Hunan Province Internet of Things Major(Xiangjiaotong[2020]248,No.288).
文摘The accurate and automatic segmentation of retinal vessels fromfundus images is critical for the early diagnosis and prevention ofmany eye diseases,such as diabetic retinopathy(DR).Existing retinal vessel segmentation approaches based on convolutional neural networks(CNNs)have achieved remarkable effectiveness.Here,we extend a retinal vessel segmentation model with low complexity and high performance based on U-Net,which is one of the most popular architectures.In view of the excellent work of depth-wise separable convolution,we introduce it to replace the standard convolutional layer.The complexity of the proposed model is reduced by decreasing the number of parameters and calculations required for themodel.To ensure performance while lowering redundant parameters,we integrate the pre-trained MobileNet V2 into the encoder.Then,a feature fusion residual module(FFRM)is designed to facilitate complementary strengths by enhancing the effective fusion between adjacent levels,which alleviates extraneous clutter introduced by direct fusion.Finally,we provide detailed comparisons between the proposed SepFE and U-Net in three retinal image mainstream datasets(DRIVE,STARE,and CHASEDB1).The results show that the number of SepFE parameters is only 3%of U-Net,the Flops are only 8%of U-Net,and better segmentation performance is obtained.The superiority of SepFE is further demonstrated through comparisons with other advanced methods.
基金This work is supported by the National Natural Science Foundation of China(61806013,61876010,61906005,62166002)General project of Science and Technology Plan of Beijing Municipal Education Commission(KM202110005028)+1 种基金Project of Interdisciplinary Research Institute of Beijing University of Technology(2021020101)International Research Cooperation Seed Fund of Beijing University of Technology(2021A01).
文摘Deep kernel mapping support vector machines have achieved good results in numerous tasks by mapping features from a low-dimensional space to a high-dimensional space and then using support vector machines for classification.However,the depth kernel mapping support vector machine does not take into account the connection of different dimensional spaces and increases the model parameters.To further improve the recognition capability of deep kernel mapping support vector machines while reducing the number of model parameters,this paper proposes a framework of Lightweight Deep Convolutional Cross-Connected Kernel Mapping Support Vector Machines(LC-CKMSVM).The framework consists of a feature extraction module and a classification module.The feature extraction module first maps the data from low-dimensional to high-dimensional space by fusing the representations of different dimensional spaces through cross-connections;then,it uses depthwise separable convolution to replace part of the original convolution to reduce the number of parameters in the module;The classification module uses a soft margin support vector machine for classification.The results on 6 different visual datasets show that LC-CKMSVM obtains better classification accuracies on most cases than the other five models.
基金supported partly by the Natural Science Foundation of Zhejiang Province,China(LGF21F020017).
文摘With the widespread use of Internet of Things(IoT)technology in daily life and the considerable safety risks of falls for elderly individuals,research on IoT-based fall detection systems has gainedmuch attention.This paper proposes an IoT-based spatiotemporal data processing framework based on a depthwise separable convolution generative adversarial network using skip-connection(Skip-DSCGAN)for fall detection.The method uses spatiotemporal data from accelerometers and gyroscopes in inertial sensors as input data.A semisupervised learning approach is adopted to train the model using only activities of daily living(ADL)data,which can avoid data imbalance problems.Furthermore,a quantile-based approach is employed to determine the fall threshold,which makes the fall detection frameworkmore robust.This proposed fall detection framework is evaluated against four other generative adversarial network(GAN)models with superior anomaly detection performance using two fall public datasets(SisFall&MobiAct).The test results show that the proposed method achieves better results,reaching 96.93% and 92.75% accuracy on the above two test datasets,respectively.At the same time,the proposed method also achieves satisfactory results in terms ofmodel size and inference delay time,making it suitable for deployment on wearable devices with limited resources.In addition,this paper also compares GAN-based semisupervised learning methods with supervised learning methods commonly used in fall detection.It clarifies the advantages of GAN-based semisupervised learning methods in fall detection.
文摘With the growth of the Internet,more and more business is being done online,for example,online offices,online education and so on.While this makes people’s lives more convenient,it also increases the risk of the network being attacked by malicious code.Therefore,it is important to identify malicious codes on computer systems efficiently.However,most of the existing malicious code detection methods have two problems:(1)The ability of the model to extract features is weak,resulting in poor model performance.(2)The large scale of model data leads to difficulties deploying on devices with limited resources.Therefore,this paper proposes a lightweight malicious code identification model Lightweight Malicious Code Classification Method Based on Improved SqueezeNet(LCMISNet).In this paper,the MFire lightweight feature extraction module is constructed by proposing a feature slicing module and a multi-size depthwise separable convolution module.The feature slicing module reduces the number of parameters by grouping features.The multi-size depthwise separable convolution module reduces the number of parameters and enhances the feature extraction capability by replacing the standard convolution with depthwise separable convolution with different convolution kernel sizes.In addition,this paper also proposes a feature splicing module to connect the MFire lightweight feature extraction module based on the feature reuse and constructs the lightweight model LCMISNet.The malicious code recognition accuracy of LCMISNet on the BIG 2015 dataset and the Malimg dataset reaches 98.90% and 99.58%,respectively.It proves that LCMISNet has a powerful malicious code recognition performance.In addition,compared with other network models,LCMISNet has better performance,and a lower number of parameters and computations.
基金supported by the open project of National Local Joint Engineering Research Center for Agro-Ecological Big Data Analysis and Application Technology,“Adaptive Agricultural Machinery Motion Detection and Recognition in Natural Scenes”,AE202210By the school-level key discipline of Suzhou University in China with No.2019xjzdxk12022 Anhui Province College Research Program Project of the Suzhou Vocational College of Civil Aviation,No.2022AH053155.
文摘In the model of the vehicle recognition algorithm implemented by the convolutional neural network,the model needs to compute and store a lot of parameters.Too many parameters occupy a lot of computational resources making it difficult to run on computers with poor performance.Therefore,obtaining more efficient feature information of target image or video with better accuracy on computers with limited arithmetic power becomes the main goal of this research.In this paper,a lightweight densely connected,and deeply separable convolutional network(DCDSNet)algorithmis proposed to achieve this goal.Visual Geometry Group(VGG)model is improved by utilizing the convolution instead of the fully connected module,the deeply separable convolution module,and the densely connected network module,with the first two modules reducing the parameters and the third module allowing the algorithm to have more features in a limited number of parameters.The algorithm achieves better results in the mine vehicle recognition dataset.Experiments show that the recognition accuracy is improved by 4.41% compared to VGG19 and the amount of parameters is reduced by 71% compared to VGG19.
基金supported by the Henan Provincial Science and Technology Research Project under Grants 232102211006,232102210044,232102211017,232102210055 and 222102210214the Science and Technology Innovation Project of Zhengzhou University of Light Industry under Grant 23XNKJTD0205+1 种基金the Undergraduate Universities Smart Teaching Special Research Project of Henan Province under Grant Jiao Gao[2021]No.489-29the Doctor Natural Science Foundation of Zhengzhou University of Light Industry under Grants 2021BSJJ025 and 2022BSJJZK13.
文摘Multispectral pedestrian detection technology leverages infrared images to provide reliable information for visible light images, demonstrating significant advantages in low-light conditions and background occlusion scenarios. However, while continuously improving cross-modal feature extraction and fusion, ensuring the model’s detection speed is also a challenging issue. We have devised a deep learning network model for cross-modal pedestrian detection based on Resnet50, aiming to focus on more reliable features and enhance the model’s detection efficiency. This model employs a spatial attention mechanism to reweight the input visible light and infrared image data, enhancing the model’s focus on different spatial positions and sharing the weighted feature data across different modalities, thereby reducing the interference of multi-modal features. Subsequently, lightweight modules with depthwise separable convolution are incorporated to reduce the model’s parameter count and computational load through channel-wise and point-wise convolutions. The network model algorithm proposed in this paper was experimentally validated on the publicly available KAIST dataset and compared with other existing methods. The experimental results demonstrate that our approach achieves favorable performance in various complex environments, affirming the effectiveness of the multispectral pedestrian detection technology proposed in this paper.
文摘Birds play a crucial role in maintaining ecological balance,making bird recognition technology a hot research topic.Traditional recognition methods have not achieved high accuracy in bird identification.This paper proposes an improved ResNet18 model to enhance the recognition rate of local bird species in Yunnan.First,a dataset containing five species of local birds in Yunnan was established:C.amherstiae,T.caboti,Syrmaticus humiae,Polyplectron bicalcaratum,and Pucrasia macrolopha.The improved ResNet18 model was then used to identify these species.This method replaces traditional convolution with depth wise separable convolution and introduces an SE(Squeeze and Excitation)module to improve the model’s efficiency and accuracy.Compared to the traditional ResNet18 model,this improved model excels in implementing a wild bird classification solution,significantly reducing computational overhead and accelerating model training using low-power,lightweight hardware.Experimental analysis shows that the improved ResNet18 model achieved an accuracy of 98.57%,compared to 98.26%for the traditional Residual Network 18 layers(ResNet18)model.
文摘The basic theory of YOLO series object detection algorithms is discussed, the dangerous driving behavior dataset is collected and produced, and then the YOLOv7 network is introduced in detail, the deep separable convolution and CA attention mechanism are introduced, the YOLOv7 bounding box loss function and clustering algorithm are optimized, and the DB-YOLOv7 network structure is constructed. In the first stage of the experiment, the PASCAL VOC public dataset was utilized for pre-training. A comparative analysis was conducted to assess the recognition accuracy and inference time before and after the proposed improvements. The experimental results demonstrated an increase of 1.4% in the average recognition accuracy, alongside a reduction in the inference time by 4 ms. Subsequently, a model for the recognition of dangerous driving behaviors was trained using a specialized dangerous driving behavior dataset. A series of experiments were performed to evaluate the efficacy of the DB-YOLOv7 algorithm in this context. The findings indicate a significant enhancement in detection performance, with a 4% improvement in accuracy compared to the baseline network. Furthermore, the model’s inference time was reduced by 20%, from 25 ms to 20 ms. These results substantiate the effectiveness of the DB-YOLOv7 recognition algorithm for detecting dangerous driving behaviors, providing comprehensive validation of its practical applicability.