Dear Editor,The letter proposes a tensor low-rank orthogonal compression(TLOC)model for a convolutional neural network(CNN),which facilitates its efficient and highly-accurate low-rank representation.Model compression...Dear Editor,The letter proposes a tensor low-rank orthogonal compression(TLOC)model for a convolutional neural network(CNN),which facilitates its efficient and highly-accurate low-rank representation.Model compression is crucial for deploying deep neural network(DNN)models on resource-constrained embedded devices.展开更多
Convolutional neural networks(CNNs)have shown remarkable success across numerous tasks such as image classification,yet the theoretical understanding of their convergence remains underdeveloped compared to their empir...Convolutional neural networks(CNNs)have shown remarkable success across numerous tasks such as image classification,yet the theoretical understanding of their convergence remains underdeveloped compared to their empirical achievements.In this paper,the first filter learning framework with convergence-guaranteed learning laws for end-to-end learning of deep CNNs is proposed.Novel update laws with convergence analysis are formulated based on the mathematical representation of each layer in convolutional neural networks.The proposed learning laws enable concurrent updates of weights across all layers of the deep convolutional neural network and the analysis shows that the training errors converge to certain bounds which are dependent on the approximation errors.Case studies are conducted on benchmark datasets and the results show that the proposed concurrent filter learning framework guarantees the convergence and offers more consistent and reliable results during training with a trade-off in performance compared to stochastic gradient descent methods.This framework represents a significant step towards enhancing the reliability and effectiveness of deep convolutional neural network by developing a theoretical analysis which allows practical implementation of the learning laws with automatic tuning of the learning rate to guarantee the convergence during training.展开更多
Landslide susceptibility mapping(LSM)plays a crucial role in assessing geological risks.The current LSM techniques face a significant challenge in achieving accurate results due to uncertainties associated with region...Landslide susceptibility mapping(LSM)plays a crucial role in assessing geological risks.The current LSM techniques face a significant challenge in achieving accurate results due to uncertainties associated with regional-scale geotechnical parameters.To explore rainfall-induced LSM,this study proposes a hybrid model that combines the physically-based probabilistic model(PPM)with convolutional neural network(CNN).The PPM is capable of effectively capturing the spatial distribution of landslides by incorporating the probability of failure(POF)considering the slope stability mechanism under rainfall conditions.This significantly characterizes the variation of POF caused by parameter uncertainties.CNN was used as a binary classifier to capture the spatial and channel correlation between landslide conditioning factors and the probability of landslide occurrence.OpenCV image enhancement technique was utilized to extract non-landslide points based on the POF of landslides.The proposed model comprehensively considers physical mechanics when selecting non-landslide samples,effectively filtering out samples that do not adhere to physical principles and reduce the risk of overfitting.The results indicate that the proposed PPM-CNN hybrid model presents a higher prediction accuracy,with an area under the curve(AUC)value of 0.85 based on the landslide case of the Niangniangba area of Gansu Province,China compared with the individual CNN model(AUC=0.61)and the PPM(AUC=0.74).This model can also consider the statistical correlation and non-normal probability distributions of model parameters.These results offer practical guidance for future research on rainfall-induced LSM at the regional scale.展开更多
The integration of image analysis through deep learning(DL)into rock classification represents a significant leap forward in geological research.While traditional methods remain invaluable for their expertise and hist...The integration of image analysis through deep learning(DL)into rock classification represents a significant leap forward in geological research.While traditional methods remain invaluable for their expertise and historical context,DL offers a powerful complement by enhancing the speed,objectivity,and precision of the classification process.This research explores the significance of image data augmentation techniques in optimizing the performance of convolutional neural networks(CNNs)for geological image analysis,particularly in the classification of igneous,metamorphic,and sedimentary rock types from rock thin section(RTS)images.This study primarily focuses on classic image augmentation techniques and evaluates their impact on model accuracy and precision.Results demonstrate that augmentation techniques like Equalize significantly enhance the model's classification capabilities,achieving an F1-Score of 0.9869 for igneous rocks,0.9884 for metamorphic rocks,and 0.9929 for sedimentary rocks,representing improvements compared to the baseline original results.Moreover,the weighted average F1-Score across all classes and techniques is 0.9886,indicating an enhancement.Conversely,methods like Distort lead to decreased accuracy and F1-Score,with an F1-Score of 0.949 for igneous rocks,0.954 for metamorphic rocks,and 0.9416 for sedimentary rocks,exacerbating the performance compared to the baseline.The study underscores the practicality of image data augmentation in geological image classification and advocates for the adoption of DL methods in this domain for automation and improved results.The findings of this study can benefit various fields,including remote sensing,mineral exploration,and environmental monitoring,by enhancing the accuracy of geological image analysis both for scientific research and industrial applications.展开更多
The migration of tasks aided by machine learning(ML)predictions IN(DPM)is a system-level design technique that is used to reduce energy by enhancing the overall performance of the processor.In this paper,we address th...The migration of tasks aided by machine learning(ML)predictions IN(DPM)is a system-level design technique that is used to reduce energy by enhancing the overall performance of the processor.In this paper,we address the issue of system-level higher task dissipation during the execution of parallel workloads with common deadlines by introducing a machine learning-based framework that includes task migration using energy-efficient earliest deadline first scheduling(EA-EDF).ML-based EA-EDF enhances the overall throughput and optimizes the energy to avoid delay and performance degradation in a multiprocessor system.The proposed system model allocates processors to the ready task set in such a way that their deadlines are guaranteed.A full task migration policy is also integrated to ensure proper task mapping that ensures inter-process linkage among the arrived tasks with the same deadlines.The execution of a task can halt on one CPU and reschedule the execution on a different processor to avoid delay and ensure meeting the deadline.Our approach shows promising potential for machine-learning-based schedulability analysis enables a comparison between different ML models and shows a promising reduction in energy as compared with other ML-aware task migration techniques for SoC like Multi-Layer Feed-Forward Neural Networks(MLFNN)based on convolutional neural network(CNN),Random Forest(RF)and Deep learning(DL)algorithm.The Simulations are conducted using super pipelined microarchitecture of advanced micro devices(AMD)XScale PXA270 using instruction and data cache per core 32 Kbyte I-cache and 32 Kbyte D-cache on various utilization factors(u_(i))12%,31%and 50%.The proposed approach consumes 5.3%less energy when almost half of the CPU is running and on a lower workload consumes 1.04%less energy.The proposed design accumulatively gives significant improvements by reducing the energy dissipation on three clock rates by 4.41%,on 624 MHz by 5.4%and 5.9%on applications operating on 416 and 312 MHz standard operating frequencies.展开更多
Hemoglobin is a vital protein in red blood cells responsible for transporting oxygen throughout the body.Its accurate measurement is crucial for diagnosing and managing conditions such as anemia and diabetes,where abn...Hemoglobin is a vital protein in red blood cells responsible for transporting oxygen throughout the body.Its accurate measurement is crucial for diagnosing and managing conditions such as anemia and diabetes,where abnormal hemoglobin levels can indicate significant health issues.Traditional methods for hemoglobin measurement are invasive,causing pain,risk of infection,and are less convenient for frequent monitoring.PPG is a transformative technology in wearable healthcare for noninvasive monitoring and widely explored for blood pressure,sleep,blood glucose,and stress analysis.In this work,we propose a hemoglobin estimation method using an adaptive lightweight convolutional neural network(HMALCNN)from PPG.The HMALCNN is designed to capture both fine-grained local waveform characteristics and global contextual patterns,ensuring robust performance across acquisition settings.We validated our approach on two multi-regional datasets containing 152 and 68 subjects,respectively,employing a subjectindependent 5-fold cross-validation strategy.The proposed method achieved root mean square errors(RMSE)of 0.90 and 1.20 g/dL for the two datasets,with strong Pearson correlations of 0.82 and 0.72.We conducted extensive posthoc analyses to assess clinical utility and interpretability.A±1 g/dL clinical error tolerance evaluation revealed that 91.3%and 86.7%of predictions for the two datasets fell within the acceptable clinical range.Hemoglobin range-wise analysis demonstrated consistently high accuracy in the normal and low hemoglobin categories.Statistical significance testing using the Wilcoxon signed-rank test confirmed the stability of performance across validation folds(p>0.05 for both RMSE and correlation).Furthermore,model interpretability was enhanced using Gradient-weighted Class Activation Mapping(Grad-CAM),supporting the model’s clinical trustworthiness.The proposed HMALCNN offers a computationally efficient,clinically interpretable,and generalizable framework for noninvasive hemoglobin monitoring,with strong potential for integration into wearable healthcare systems as a practical alternative to invasive measurement techniques.展开更多
To enhance the inference efficiency of convolutional neural network(CNN),tensor parallelism is employed to improve the parallelism within operators.However,existing methods are customized to specific networks and hard...To enhance the inference efficiency of convolutional neural network(CNN),tensor parallelism is employed to improve the parallelism within operators.However,existing methods are customized to specific networks and hardware,limiting their generalizability.This paper proposes an approach called resource-adaptive tensor decomposition(RATD)for CNN operators,which aims to achieve an optimal match between computational resources and parallel computing tasks.Firstly,CNN is represented with fine-grained tensors at the lower graph level,thereby decoupling tensors that can be computed in parallel within operators.Secondly,the convolution and pooling operators are fused,and the decoupled tensor blocks are scheduled in parallel.Finally,a cost model is constructed,based on runtime and resource utilization,to iteratively refine the process of tensor block decomposition and automatically determine the optimal tensor decomposition.Experimental results demonstrate that the proposed RATD improves the accuracy of the model by 11%.Compared with CUDA(compute unified device architecture)deep neural network library(cuDNN),RATD achieves an average speedup ratio of 1.21 times in inference time across various convolution kernels,along with a 12%increase in computational resource utilization.展开更多
Bearing pitting,one of the common faults in mechanical systems,is a research hotspot in both academia and industry.Traditional fault diagnosis methods for bearings are based on manual experience with low diagnostic ef...Bearing pitting,one of the common faults in mechanical systems,is a research hotspot in both academia and industry.Traditional fault diagnosis methods for bearings are based on manual experience with low diagnostic efficiency.This study proposes a novel bearing fault diagnosis method based on deep separable convolution and spatial dropout regularization.Deep separable convolution extracts features from the raw bearing vibration signals,during which a 3×1 convolutional kernel with a one-step size selects effective features by adjusting its weights.The similarity pruning process of the channel convolution and point convolution can reduce the number of parameters and calculation quantities by evaluating the size of the weights and removing the feature maps of smaller weights.The spatial dropout regularization method focuses on bearing signal fault features,improving the independence between the bearing signal features and enhancing the robustness of the model.A batch normalization algorithm is added to the convolutional layer for gradient explosion control and network stability improvement.To validate the effectiveness of the proposed method,we collect raw vibration signals from bearings in eight different health states.The experimental results show that the proposed method can effectively distinguish different pitting faults in the bearings with a better accuracy than that of other typical deep learning methods.展开更多
The recent surge of mobile subscribers and user data traffic has accelerated the telecommunication sector towards the adoption of the fifth-generation (5G) mobile networks. Cloud radio access network (CRAN) is a promi...The recent surge of mobile subscribers and user data traffic has accelerated the telecommunication sector towards the adoption of the fifth-generation (5G) mobile networks. Cloud radio access network (CRAN) is a prominent framework in the 5G mobile network to meet the above requirements by deploying low-cost and intelligent multiple distributed antennas known as remote radio heads (RRHs). However, achieving the optimal resource allocation (RA) in CRAN using the traditional approach is still challenging due to the complex structure. In this paper, we introduce the convolutional neural network-based deep Q-network (CNN-DQN) to balance the energy consumption and guarantee the user quality of service (QoS) demand in downlink CRAN. We first formulate the Markov decision process (MDP) for energy efficiency (EE) and build up a 3-layer CNN to capture the environment feature as an input state space. We then use DQN to turn on/off the RRHs dynamically based on the user QoS demand and energy consumption in the CRAN. Finally, we solve the RA problem based on the user constraint and transmit power to guarantee the user QoS demand and maximize the EE with a minimum number of active RRHs. In the end, we conduct the simulation to compare our proposed scheme with nature DQN and the traditional approach.展开更多
This study introduces an innovative“Big Model”strategy to enhance Bridge Structural Health Monitoring(SHM)using a Convolutional Neural Network(CNN),time-frequency analysis,and fine element analysis.Leveraging ensemb...This study introduces an innovative“Big Model”strategy to enhance Bridge Structural Health Monitoring(SHM)using a Convolutional Neural Network(CNN),time-frequency analysis,and fine element analysis.Leveraging ensemble methods,collaborative learning,and distributed computing,the approach effectively manages the complexity and scale of large-scale bridge data.The CNN employs transfer learning,fine-tuning,and continuous monitoring to optimize models for adaptive and accurate structural health assessments,focusing on extracting meaningful features through time-frequency analysis.By integrating Finite Element Analysis,time-frequency analysis,and CNNs,the strategy provides a comprehensive understanding of bridge health.Utilizing diverse sensor data,sophisticated feature extraction,and advanced CNN architecture,the model is optimized through rigorous preprocessing and hyperparameter tuning.This approach significantly enhances the ability to make accurate predictions,monitor structural health,and support proactive maintenance practices,thereby ensuring the safety and longevity of critical infrastructure.展开更多
Recent advances in deep neural networks have shed new light on physics,engineering,and scientific computing.Reconciling the data-centered viewpoint with physical simulation is one of the research hotspots.The physicsi...Recent advances in deep neural networks have shed new light on physics,engineering,and scientific computing.Reconciling the data-centered viewpoint with physical simulation is one of the research hotspots.The physicsinformedneural network(PINN)is currently the most general framework,which is more popular due to theconvenience of constructing NNs and excellent generalization ability.The automatic differentiation(AD)-basedPINN model is suitable for the homogeneous scientific problem;however,it is unclear how AD can enforce fluxcontinuity across boundaries between cells of different properties where spatial heterogeneity is represented bygrid cells with different physical properties.In this work,we propose a criss-cross physics-informed convolutionalneural network(CC-PINN)learning architecture,aiming to learn the solution of parametric PDEs with spatialheterogeneity of physical properties.To achieve the seamless enforcement of flux continuity and integration ofphysicalmeaning into CNN,a predefined 2D convolutional layer is proposed to accurately express transmissibilitybetween adjacent cells.The efficacy of the proposedmethodwas evaluated through predictions of several petroleumreservoir problems with spatial heterogeneity and compared against state-of-the-art(PINN)through numericalanalysis as a benchmark,which demonstrated the superiority of the proposed method over the PINN.展开更多
In order to improve the spatial resolution of hyperspectral(HS)image and minimize the spectral distortion,an HS and multispectral(MS)image fusion approach based on convolutional neural network(CNN)is proposed.The prop...In order to improve the spatial resolution of hyperspectral(HS)image and minimize the spectral distortion,an HS and multispectral(MS)image fusion approach based on convolutional neural network(CNN)is proposed.The proposed approach incorporates the linear spectral mixture model and spatial-spectral spread transform model into the learning phase of network,aiming to fully exploit the spatial-spectral information of HS and MS images,and improve the spectral fidelity of fusion images.Experiments on two real remote sensing data under different resolutions demonstrate that compared with some state-of-the-art HS and MS image fusion methods,the proposed approach achieves superior spectral fidelities and lower fusion errors.展开更多
With the high-speed development of the Internet,a growing number of Internet users like giving their subjective comments in the BBS,blog and shopping website.These comments contains critics’attitudes,emotions,views a...With the high-speed development of the Internet,a growing number of Internet users like giving their subjective comments in the BBS,blog and shopping website.These comments contains critics’attitudes,emotions,views and other information.Using these information reasonablely can help understand the social public opinion and make a timely response and help dealer to improve quality and service of products and make consumers know merchandise.This paper mainly discusses using convolutional neural network(CNN)for the operation of the text feature extraction.The concrete realization are discussed.Then combining with other text classifier make class operation.The experiment result shows the effectiveness of the method which is proposed in this paper.展开更多
A demodulator based on convolutional neural networks( CNNs) is proposed to demodulate bipolar extended binary phase shifting keying( EBPSK) signals transmitted at a faster-thanNyquist( FTN) rate, solving the pro...A demodulator based on convolutional neural networks( CNNs) is proposed to demodulate bipolar extended binary phase shifting keying( EBPSK) signals transmitted at a faster-thanNyquist( FTN) rate, solving the problem of severe inter symbol interference( ISI) caused by FTN rate signals. With the characteristics of local connectivity, pooling and weight sharing,a six-layer CNNs structure is used to demodulate and eliminate ISI. The results showthat with the symbol rate of 1. 07 k Bd, the bandwidth of the band-pass filter( BPF) in a transmitter of 1 k Hz and the changing number of carrier cycles in a symbol K = 5,10,15,28, the overall bit error ratio( BER) performance of CNNs with single-symbol decision is superior to that with a doublesymbol united-decision. In addition, the BER performance of single-symbol decision is approximately 0. 5 d B better than that of the coherent demodulator while K equals the total number of carrier circles in a symbol, i. e., K = N = 28. With the symbol rate of 1. 07 k Bd, the bandwidth of BPF in a transmitter of 500 Hz and K = 5,10,15,28, the overall BER performance of CNNs with double-symbol united-decision is superior to those with single-symbol decision. Moreover, the double-symbol uniteddecision method is approximately 0. 5 to 1. 5 d B better than that of the coherent demodulator while K = N = 28. The demodulators based on CNNs successfully solve the serious ISI problems generated during the transmission of FTN rate bipolar EBPSK signals, which is beneficial for the improvement of spectrum efficiency.展开更多
Flood probability maps are essential for a range of applications,including land use planning and developing mitigation strategies and early warning systems.This study describes the potential application of two archite...Flood probability maps are essential for a range of applications,including land use planning and developing mitigation strategies and early warning systems.This study describes the potential application of two architectures of deep learning neural networks,namely convolutional neural networks(CNN)and recurrent neural networks(RNN),for spatially explicit prediction and mapping of flash flood probability.To develop and validate the predictive models,a geospatial database that contained records for the historical flood events and geo-environmental characteristics of the Golestan Province in northern Iran was constructed.The step-wise weight assessment ratio analysis(SWARA)was employed to investigate the spatial interplay between floods and different influencing factors.The CNN and RNN models were trained using the SWARA weights and validated using the receiver operating characteristics technique.The results showed that the CNN model(AUC=0.832,RMSE=0.144)performed slightly better than the RNN model(AUC=0.814,RMSE=0.181)in predicting future floods.Further,these models demonstrated an improved prediction of floods compared to previous studies that used different models in the same study area.This study showed that the spatially explicit deep learning neural network models are successful in capturing the heterogeneity of spatial patterns of flood probability in the Golestan Province,and the resulting probability maps can be used for the development of mitigation plans in response to the future floods.The general policy implication of our study suggests that design,implementation,and verification of flood early warning systems should be directed to approximately 40%of the land area characterized by high and very susceptibility to flooding.展开更多
Nowadays,the amount of wed data is increasing at a rapid speed,which presents a serious challenge to the web monitoring.Text sentiment analysis,an important research topic in the area of natural language processing,is...Nowadays,the amount of wed data is increasing at a rapid speed,which presents a serious challenge to the web monitoring.Text sentiment analysis,an important research topic in the area of natural language processing,is a crucial task in the web monitoring area.The accuracy of traditional text sentiment analysis methods might be degraded in dealing with mass data.Deep learning is a hot research topic of the artificial intelligence in the recent years.By now,several research groups have studied the sentiment analysis of English texts using deep learning methods.In contrary,relatively few works have so far considered the Chinese text sentiment analysis toward this direction.In this paper,a method for analyzing the Chinese text sentiment is proposed based on the convolutional neural network(CNN)in deep learning in order to improve the analysis accuracy.The feature values of the CNN after the training process are nonuniformly distributed.In order to overcome this problem,a method for normalizing the feature values is proposed.Moreover,the dimensions of the text features are optimized through simulations.Finally,a method for updating the learning rate in the training process of the CNN is presented in order to achieve better performances.Experiment results on the typical datasets indicate that the accuracy of the proposed method can be improved compared with that of the traditional supervised machine learning methods,e.g.,the support vector machine method.展开更多
基金supported by the Science and Technology Innovation Key R&D Program of Chongqing(CSTB2025TIAD-STX0032)National Key Research and Development Program of China(2024YFF0908200)+1 种基金the Chongqing Technology Innovation and Application Development Special Key Project(CSTB2024TIAD-KPX0018)the Southwest University Graduate Student Research Innovation(SWUB24051)。
文摘Dear Editor,The letter proposes a tensor low-rank orthogonal compression(TLOC)model for a convolutional neural network(CNN),which facilitates its efficient and highly-accurate low-rank representation.Model compression is crucial for deploying deep neural network(DNN)models on resource-constrained embedded devices.
基金supported by the Ministry of Education(MOE)Singapore,Academic Research Fund(AcRF)Tier 1(RG65/22)。
文摘Convolutional neural networks(CNNs)have shown remarkable success across numerous tasks such as image classification,yet the theoretical understanding of their convergence remains underdeveloped compared to their empirical achievements.In this paper,the first filter learning framework with convergence-guaranteed learning laws for end-to-end learning of deep CNNs is proposed.Novel update laws with convergence analysis are formulated based on the mathematical representation of each layer in convolutional neural networks.The proposed learning laws enable concurrent updates of weights across all layers of the deep convolutional neural network and the analysis shows that the training errors converge to certain bounds which are dependent on the approximation errors.Case studies are conducted on benchmark datasets and the results show that the proposed concurrent filter learning framework guarantees the convergence and offers more consistent and reliable results during training with a trade-off in performance compared to stochastic gradient descent methods.This framework represents a significant step towards enhancing the reliability and effectiveness of deep convolutional neural network by developing a theoretical analysis which allows practical implementation of the learning laws with automatic tuning of the learning rate to guarantee the convergence during training.
基金funding support from the National Natural Science Foundation of China(Grant Nos.U22A20594,52079045)Hong-Zhi Cui acknowledges the financial support of the China Scholarship Council(Grant No.CSC:202206710014)for his research at Universitat Politecnica de Catalunya,Barcelona.
文摘Landslide susceptibility mapping(LSM)plays a crucial role in assessing geological risks.The current LSM techniques face a significant challenge in achieving accurate results due to uncertainties associated with regional-scale geotechnical parameters.To explore rainfall-induced LSM,this study proposes a hybrid model that combines the physically-based probabilistic model(PPM)with convolutional neural network(CNN).The PPM is capable of effectively capturing the spatial distribution of landslides by incorporating the probability of failure(POF)considering the slope stability mechanism under rainfall conditions.This significantly characterizes the variation of POF caused by parameter uncertainties.CNN was used as a binary classifier to capture the spatial and channel correlation between landslide conditioning factors and the probability of landslide occurrence.OpenCV image enhancement technique was utilized to extract non-landslide points based on the POF of landslides.The proposed model comprehensively considers physical mechanics when selecting non-landslide samples,effectively filtering out samples that do not adhere to physical principles and reduce the risk of overfitting.The results indicate that the proposed PPM-CNN hybrid model presents a higher prediction accuracy,with an area under the curve(AUC)value of 0.85 based on the landslide case of the Niangniangba area of Gansu Province,China compared with the individual CNN model(AUC=0.61)and the PPM(AUC=0.74).This model can also consider the statistical correlation and non-normal probability distributions of model parameters.These results offer practical guidance for future research on rainfall-induced LSM at the regional scale.
文摘The integration of image analysis through deep learning(DL)into rock classification represents a significant leap forward in geological research.While traditional methods remain invaluable for their expertise and historical context,DL offers a powerful complement by enhancing the speed,objectivity,and precision of the classification process.This research explores the significance of image data augmentation techniques in optimizing the performance of convolutional neural networks(CNNs)for geological image analysis,particularly in the classification of igneous,metamorphic,and sedimentary rock types from rock thin section(RTS)images.This study primarily focuses on classic image augmentation techniques and evaluates their impact on model accuracy and precision.Results demonstrate that augmentation techniques like Equalize significantly enhance the model's classification capabilities,achieving an F1-Score of 0.9869 for igneous rocks,0.9884 for metamorphic rocks,and 0.9929 for sedimentary rocks,representing improvements compared to the baseline original results.Moreover,the weighted average F1-Score across all classes and techniques is 0.9886,indicating an enhancement.Conversely,methods like Distort lead to decreased accuracy and F1-Score,with an F1-Score of 0.949 for igneous rocks,0.954 for metamorphic rocks,and 0.9416 for sedimentary rocks,exacerbating the performance compared to the baseline.The study underscores the practicality of image data augmentation in geological image classification and advocates for the adoption of DL methods in this domain for automation and improved results.The findings of this study can benefit various fields,including remote sensing,mineral exploration,and environmental monitoring,by enhancing the accuracy of geological image analysis both for scientific research and industrial applications.
文摘The migration of tasks aided by machine learning(ML)predictions IN(DPM)is a system-level design technique that is used to reduce energy by enhancing the overall performance of the processor.In this paper,we address the issue of system-level higher task dissipation during the execution of parallel workloads with common deadlines by introducing a machine learning-based framework that includes task migration using energy-efficient earliest deadline first scheduling(EA-EDF).ML-based EA-EDF enhances the overall throughput and optimizes the energy to avoid delay and performance degradation in a multiprocessor system.The proposed system model allocates processors to the ready task set in such a way that their deadlines are guaranteed.A full task migration policy is also integrated to ensure proper task mapping that ensures inter-process linkage among the arrived tasks with the same deadlines.The execution of a task can halt on one CPU and reschedule the execution on a different processor to avoid delay and ensure meeting the deadline.Our approach shows promising potential for machine-learning-based schedulability analysis enables a comparison between different ML models and shows a promising reduction in energy as compared with other ML-aware task migration techniques for SoC like Multi-Layer Feed-Forward Neural Networks(MLFNN)based on convolutional neural network(CNN),Random Forest(RF)and Deep learning(DL)algorithm.The Simulations are conducted using super pipelined microarchitecture of advanced micro devices(AMD)XScale PXA270 using instruction and data cache per core 32 Kbyte I-cache and 32 Kbyte D-cache on various utilization factors(u_(i))12%,31%and 50%.The proposed approach consumes 5.3%less energy when almost half of the CPU is running and on a lower workload consumes 1.04%less energy.The proposed design accumulatively gives significant improvements by reducing the energy dissipation on three clock rates by 4.41%,on 624 MHz by 5.4%and 5.9%on applications operating on 416 and 312 MHz standard operating frequencies.
基金funded by the Deanship of Graduate Studies and Scientific Research at Qassim University for financial support(QU-APC-2025).
文摘Hemoglobin is a vital protein in red blood cells responsible for transporting oxygen throughout the body.Its accurate measurement is crucial for diagnosing and managing conditions such as anemia and diabetes,where abnormal hemoglobin levels can indicate significant health issues.Traditional methods for hemoglobin measurement are invasive,causing pain,risk of infection,and are less convenient for frequent monitoring.PPG is a transformative technology in wearable healthcare for noninvasive monitoring and widely explored for blood pressure,sleep,blood glucose,and stress analysis.In this work,we propose a hemoglobin estimation method using an adaptive lightweight convolutional neural network(HMALCNN)from PPG.The HMALCNN is designed to capture both fine-grained local waveform characteristics and global contextual patterns,ensuring robust performance across acquisition settings.We validated our approach on two multi-regional datasets containing 152 and 68 subjects,respectively,employing a subjectindependent 5-fold cross-validation strategy.The proposed method achieved root mean square errors(RMSE)of 0.90 and 1.20 g/dL for the two datasets,with strong Pearson correlations of 0.82 and 0.72.We conducted extensive posthoc analyses to assess clinical utility and interpretability.A±1 g/dL clinical error tolerance evaluation revealed that 91.3%and 86.7%of predictions for the two datasets fell within the acceptable clinical range.Hemoglobin range-wise analysis demonstrated consistently high accuracy in the normal and low hemoglobin categories.Statistical significance testing using the Wilcoxon signed-rank test confirmed the stability of performance across validation folds(p>0.05 for both RMSE and correlation).Furthermore,model interpretability was enhanced using Gradient-weighted Class Activation Mapping(Grad-CAM),supporting the model’s clinical trustworthiness.The proposed HMALCNN offers a computationally efficient,clinically interpretable,and generalizable framework for noninvasive hemoglobin monitoring,with strong potential for integration into wearable healthcare systems as a practical alternative to invasive measurement techniques.
基金Supported by the National Science and Technology Major Project of China(No.2022ZD0119003)the National Natural Science Foundation of China(No.61834005).
文摘To enhance the inference efficiency of convolutional neural network(CNN),tensor parallelism is employed to improve the parallelism within operators.However,existing methods are customized to specific networks and hardware,limiting their generalizability.This paper proposes an approach called resource-adaptive tensor decomposition(RATD)for CNN operators,which aims to achieve an optimal match between computational resources and parallel computing tasks.Firstly,CNN is represented with fine-grained tensors at the lower graph level,thereby decoupling tensors that can be computed in parallel within operators.Secondly,the convolution and pooling operators are fused,and the decoupled tensor blocks are scheduled in parallel.Finally,a cost model is constructed,based on runtime and resource utilization,to iteratively refine the process of tensor block decomposition and automatically determine the optimal tensor decomposition.Experimental results demonstrate that the proposed RATD improves the accuracy of the model by 11%.Compared with CUDA(compute unified device architecture)deep neural network library(cuDNN),RATD achieves an average speedup ratio of 1.21 times in inference time across various convolution kernels,along with a 12%increase in computational resource utilization.
基金the National Key Research and Development Program of China (No. 2019YFB1704500)the State Ministry of Science and Technology Innovation Fund of China (No. 2018IM030200)+1 种基金the National Natural Foundation of China (No. U1708255)the China Scholarship Council (No. 201906080059)
文摘Bearing pitting,one of the common faults in mechanical systems,is a research hotspot in both academia and industry.Traditional fault diagnosis methods for bearings are based on manual experience with low diagnostic efficiency.This study proposes a novel bearing fault diagnosis method based on deep separable convolution and spatial dropout regularization.Deep separable convolution extracts features from the raw bearing vibration signals,during which a 3×1 convolutional kernel with a one-step size selects effective features by adjusting its weights.The similarity pruning process of the channel convolution and point convolution can reduce the number of parameters and calculation quantities by evaluating the size of the weights and removing the feature maps of smaller weights.The spatial dropout regularization method focuses on bearing signal fault features,improving the independence between the bearing signal features and enhancing the robustness of the model.A batch normalization algorithm is added to the convolutional layer for gradient explosion control and network stability improvement.To validate the effectiveness of the proposed method,we collect raw vibration signals from bearings in eight different health states.The experimental results show that the proposed method can effectively distinguish different pitting faults in the bearings with a better accuracy than that of other typical deep learning methods.
基金supported by the Universiti Tunku Abdul Rahman (UTAR) Malaysia under UTARRF (IPSR/RMC/UTARRF/2021-C1/T05)
文摘The recent surge of mobile subscribers and user data traffic has accelerated the telecommunication sector towards the adoption of the fifth-generation (5G) mobile networks. Cloud radio access network (CRAN) is a prominent framework in the 5G mobile network to meet the above requirements by deploying low-cost and intelligent multiple distributed antennas known as remote radio heads (RRHs). However, achieving the optimal resource allocation (RA) in CRAN using the traditional approach is still challenging due to the complex structure. In this paper, we introduce the convolutional neural network-based deep Q-network (CNN-DQN) to balance the energy consumption and guarantee the user quality of service (QoS) demand in downlink CRAN. We first formulate the Markov decision process (MDP) for energy efficiency (EE) and build up a 3-layer CNN to capture the environment feature as an input state space. We then use DQN to turn on/off the RRHs dynamically based on the user QoS demand and energy consumption in the CRAN. Finally, we solve the RA problem based on the user constraint and transmit power to guarantee the user QoS demand and maximize the EE with a minimum number of active RRHs. In the end, we conduct the simulation to compare our proposed scheme with nature DQN and the traditional approach.
文摘This study introduces an innovative“Big Model”strategy to enhance Bridge Structural Health Monitoring(SHM)using a Convolutional Neural Network(CNN),time-frequency analysis,and fine element analysis.Leveraging ensemble methods,collaborative learning,and distributed computing,the approach effectively manages the complexity and scale of large-scale bridge data.The CNN employs transfer learning,fine-tuning,and continuous monitoring to optimize models for adaptive and accurate structural health assessments,focusing on extracting meaningful features through time-frequency analysis.By integrating Finite Element Analysis,time-frequency analysis,and CNNs,the strategy provides a comprehensive understanding of bridge health.Utilizing diverse sensor data,sophisticated feature extraction,and advanced CNN architecture,the model is optimized through rigorous preprocessing and hyperparameter tuning.This approach significantly enhances the ability to make accurate predictions,monitor structural health,and support proactive maintenance practices,thereby ensuring the safety and longevity of critical infrastructure.
基金the National Natural Science Foundation of China(No.52274048)Beijing Natural Science Foundation(No.3222037)+1 种基金the CNPC 14th Five-Year Perspective Fundamental Research Project(No.2021DJ2104)the Science Foundation of China University of Petroleum,Beijing(No.2462021YXZZ010).
文摘Recent advances in deep neural networks have shed new light on physics,engineering,and scientific computing.Reconciling the data-centered viewpoint with physical simulation is one of the research hotspots.The physicsinformedneural network(PINN)is currently the most general framework,which is more popular due to theconvenience of constructing NNs and excellent generalization ability.The automatic differentiation(AD)-basedPINN model is suitable for the homogeneous scientific problem;however,it is unclear how AD can enforce fluxcontinuity across boundaries between cells of different properties where spatial heterogeneity is represented bygrid cells with different physical properties.In this work,we propose a criss-cross physics-informed convolutionalneural network(CC-PINN)learning architecture,aiming to learn the solution of parametric PDEs with spatialheterogeneity of physical properties.To achieve the seamless enforcement of flux continuity and integration ofphysicalmeaning into CNN,a predefined 2D convolutional layer is proposed to accurately express transmissibilitybetween adjacent cells.The efficacy of the proposedmethodwas evaluated through predictions of several petroleumreservoir problems with spatial heterogeneity and compared against state-of-the-art(PINN)through numericalanalysis as a benchmark,which demonstrated the superiority of the proposed method over the PINN.
基金National Natural Science Foundation of China(No.61902060)Natural Science Foundation of Shanghai,China(No.19ZR1453800)Fundamental Research Funds for the Central Universities,China(No.2232021D-33)。
文摘In order to improve the spatial resolution of hyperspectral(HS)image and minimize the spectral distortion,an HS and multispectral(MS)image fusion approach based on convolutional neural network(CNN)is proposed.The proposed approach incorporates the linear spectral mixture model and spatial-spectral spread transform model into the learning phase of network,aiming to fully exploit the spatial-spectral information of HS and MS images,and improve the spectral fidelity of fusion images.Experiments on two real remote sensing data under different resolutions demonstrate that compared with some state-of-the-art HS and MS image fusion methods,the proposed approach achieves superior spectral fidelities and lower fusion errors.
文摘With the high-speed development of the Internet,a growing number of Internet users like giving their subjective comments in the BBS,blog and shopping website.These comments contains critics’attitudes,emotions,views and other information.Using these information reasonablely can help understand the social public opinion and make a timely response and help dealer to improve quality and service of products and make consumers know merchandise.This paper mainly discusses using convolutional neural network(CNN)for the operation of the text feature extraction.The concrete realization are discussed.Then combining with other text classifier make class operation.The experiment result shows the effectiveness of the method which is proposed in this paper.
文摘针对不同磁密幅值、频率、谐波组合等复杂激励工况下磁致伸缩建模面临的精准性问题,该文利用空间注意力机制(spatial attention mechanism,SAM)对传统的卷积神经网络(convolutional neural network,CNN)进行改进,将SAM嵌套入CNN网络中,建立SAMCNN改进型网络。再结合双向长短期记忆(bidirectional long short-term memory,BiLSTM)网络,提出电工钢片SAMCNN-BiLSTM磁致伸缩模型。首先,利用灰狼优化算法(grey wolf optimization,GWO)寻优神经网络结构的参数,实现复杂工况下磁致伸缩效应的准确表征;然后,建立中低频范围单频与叠加谐波激励等复杂工况下的磁致伸缩应变数据库,开展数据预处理与特征分析;最后,对SAMCNN-BiLSTM模型开展对比验证。对比叠加3次谐波激励下的磁致伸缩应变频谱主要分量,SAMCNN-BiLSTM模型计算值最大相对误差为3.70%,其比Jiles-Atherton-Sablik(J-A-S)、二次畴转等模型能更精确地表征电工钢片的磁致伸缩效应。
基金The National Natural Science Foundation of China(No.6504000089)
文摘A demodulator based on convolutional neural networks( CNNs) is proposed to demodulate bipolar extended binary phase shifting keying( EBPSK) signals transmitted at a faster-thanNyquist( FTN) rate, solving the problem of severe inter symbol interference( ISI) caused by FTN rate signals. With the characteristics of local connectivity, pooling and weight sharing,a six-layer CNNs structure is used to demodulate and eliminate ISI. The results showthat with the symbol rate of 1. 07 k Bd, the bandwidth of the band-pass filter( BPF) in a transmitter of 1 k Hz and the changing number of carrier cycles in a symbol K = 5,10,15,28, the overall bit error ratio( BER) performance of CNNs with single-symbol decision is superior to that with a doublesymbol united-decision. In addition, the BER performance of single-symbol decision is approximately 0. 5 d B better than that of the coherent demodulator while K equals the total number of carrier circles in a symbol, i. e., K = N = 28. With the symbol rate of 1. 07 k Bd, the bandwidth of BPF in a transmitter of 500 Hz and K = 5,10,15,28, the overall BER performance of CNNs with double-symbol united-decision is superior to those with single-symbol decision. Moreover, the double-symbol uniteddecision method is approximately 0. 5 to 1. 5 d B better than that of the coherent demodulator while K = N = 28. The demodulators based on CNNs successfully solve the serious ISI problems generated during the transmission of FTN rate bipolar EBPSK signals, which is beneficial for the improvement of spectrum efficiency.
基金conducted by the Basic Research Project of the Korea Institute of Geoscience and Mineral Resources(KIGAM)funded by the Ministry of Science and ICT。
文摘Flood probability maps are essential for a range of applications,including land use planning and developing mitigation strategies and early warning systems.This study describes the potential application of two architectures of deep learning neural networks,namely convolutional neural networks(CNN)and recurrent neural networks(RNN),for spatially explicit prediction and mapping of flash flood probability.To develop and validate the predictive models,a geospatial database that contained records for the historical flood events and geo-environmental characteristics of the Golestan Province in northern Iran was constructed.The step-wise weight assessment ratio analysis(SWARA)was employed to investigate the spatial interplay between floods and different influencing factors.The CNN and RNN models were trained using the SWARA weights and validated using the receiver operating characteristics technique.The results showed that the CNN model(AUC=0.832,RMSE=0.144)performed slightly better than the RNN model(AUC=0.814,RMSE=0.181)in predicting future floods.Further,these models demonstrated an improved prediction of floods compared to previous studies that used different models in the same study area.This study showed that the spatially explicit deep learning neural network models are successful in capturing the heterogeneity of spatial patterns of flood probability in the Golestan Province,and the resulting probability maps can be used for the development of mitigation plans in response to the future floods.The general policy implication of our study suggests that design,implementation,and verification of flood early warning systems should be directed to approximately 40%of the land area characterized by high and very susceptibility to flooding.
文摘Nowadays,the amount of wed data is increasing at a rapid speed,which presents a serious challenge to the web monitoring.Text sentiment analysis,an important research topic in the area of natural language processing,is a crucial task in the web monitoring area.The accuracy of traditional text sentiment analysis methods might be degraded in dealing with mass data.Deep learning is a hot research topic of the artificial intelligence in the recent years.By now,several research groups have studied the sentiment analysis of English texts using deep learning methods.In contrary,relatively few works have so far considered the Chinese text sentiment analysis toward this direction.In this paper,a method for analyzing the Chinese text sentiment is proposed based on the convolutional neural network(CNN)in deep learning in order to improve the analysis accuracy.The feature values of the CNN after the training process are nonuniformly distributed.In order to overcome this problem,a method for normalizing the feature values is proposed.Moreover,the dimensions of the text features are optimized through simulations.Finally,a method for updating the learning rate in the training process of the CNN is presented in order to achieve better performances.Experiment results on the typical datasets indicate that the accuracy of the proposed method can be improved compared with that of the traditional supervised machine learning methods,e.g.,the support vector machine method.