The integration of image analysis through deep learning(DL)into rock classification represents a significant leap forward in geological research.While traditional methods remain invaluable for their expertise and hist...The integration of image analysis through deep learning(DL)into rock classification represents a significant leap forward in geological research.While traditional methods remain invaluable for their expertise and historical context,DL offers a powerful complement by enhancing the speed,objectivity,and precision of the classification process.This research explores the significance of image data augmentation techniques in optimizing the performance of convolutional neural networks(CNNs)for geological image analysis,particularly in the classification of igneous,metamorphic,and sedimentary rock types from rock thin section(RTS)images.This study primarily focuses on classic image augmentation techniques and evaluates their impact on model accuracy and precision.Results demonstrate that augmentation techniques like Equalize significantly enhance the model's classification capabilities,achieving an F1-Score of 0.9869 for igneous rocks,0.9884 for metamorphic rocks,and 0.9929 for sedimentary rocks,representing improvements compared to the baseline original results.Moreover,the weighted average F1-Score across all classes and techniques is 0.9886,indicating an enhancement.Conversely,methods like Distort lead to decreased accuracy and F1-Score,with an F1-Score of 0.949 for igneous rocks,0.954 for metamorphic rocks,and 0.9416 for sedimentary rocks,exacerbating the performance compared to the baseline.The study underscores the practicality of image data augmentation in geological image classification and advocates for the adoption of DL methods in this domain for automation and improved results.The findings of this study can benefit various fields,including remote sensing,mineral exploration,and environmental monitoring,by enhancing the accuracy of geological image analysis both for scientific research and industrial applications.展开更多
Landslide susceptibility mapping(LSM)plays a crucial role in assessing geological risks.The current LSM techniques face a significant challenge in achieving accurate results due to uncertainties associated with region...Landslide susceptibility mapping(LSM)plays a crucial role in assessing geological risks.The current LSM techniques face a significant challenge in achieving accurate results due to uncertainties associated with regional-scale geotechnical parameters.To explore rainfall-induced LSM,this study proposes a hybrid model that combines the physically-based probabilistic model(PPM)with convolutional neural network(CNN).The PPM is capable of effectively capturing the spatial distribution of landslides by incorporating the probability of failure(POF)considering the slope stability mechanism under rainfall conditions.This significantly characterizes the variation of POF caused by parameter uncertainties.CNN was used as a binary classifier to capture the spatial and channel correlation between landslide conditioning factors and the probability of landslide occurrence.OpenCV image enhancement technique was utilized to extract non-landslide points based on the POF of landslides.The proposed model comprehensively considers physical mechanics when selecting non-landslide samples,effectively filtering out samples that do not adhere to physical principles and reduce the risk of overfitting.The results indicate that the proposed PPM-CNN hybrid model presents a higher prediction accuracy,with an area under the curve(AUC)value of 0.85 based on the landslide case of the Niangniangba area of Gansu Province,China compared with the individual CNN model(AUC=0.61)and the PPM(AUC=0.74).This model can also consider the statistical correlation and non-normal probability distributions of model parameters.These results offer practical guidance for future research on rainfall-induced LSM at the regional scale.展开更多
The recent surge of mobile subscribers and user data traffic has accelerated the telecommunication sector towards the adoption of the fifth-generation (5G) mobile networks. Cloud radio access network (CRAN) is a promi...The recent surge of mobile subscribers and user data traffic has accelerated the telecommunication sector towards the adoption of the fifth-generation (5G) mobile networks. Cloud radio access network (CRAN) is a prominent framework in the 5G mobile network to meet the above requirements by deploying low-cost and intelligent multiple distributed antennas known as remote radio heads (RRHs). However, achieving the optimal resource allocation (RA) in CRAN using the traditional approach is still challenging due to the complex structure. In this paper, we introduce the convolutional neural network-based deep Q-network (CNN-DQN) to balance the energy consumption and guarantee the user quality of service (QoS) demand in downlink CRAN. We first formulate the Markov decision process (MDP) for energy efficiency (EE) and build up a 3-layer CNN to capture the environment feature as an input state space. We then use DQN to turn on/off the RRHs dynamically based on the user QoS demand and energy consumption in the CRAN. Finally, we solve the RA problem based on the user constraint and transmit power to guarantee the user QoS demand and maximize the EE with a minimum number of active RRHs. In the end, we conduct the simulation to compare our proposed scheme with nature DQN and the traditional approach.展开更多
This study introduces an innovative“Big Model”strategy to enhance Bridge Structural Health Monitoring(SHM)using a Convolutional Neural Network(CNN),time-frequency analysis,and fine element analysis.Leveraging ensemb...This study introduces an innovative“Big Model”strategy to enhance Bridge Structural Health Monitoring(SHM)using a Convolutional Neural Network(CNN),time-frequency analysis,and fine element analysis.Leveraging ensemble methods,collaborative learning,and distributed computing,the approach effectively manages the complexity and scale of large-scale bridge data.The CNN employs transfer learning,fine-tuning,and continuous monitoring to optimize models for adaptive and accurate structural health assessments,focusing on extracting meaningful features through time-frequency analysis.By integrating Finite Element Analysis,time-frequency analysis,and CNNs,the strategy provides a comprehensive understanding of bridge health.Utilizing diverse sensor data,sophisticated feature extraction,and advanced CNN architecture,the model is optimized through rigorous preprocessing and hyperparameter tuning.This approach significantly enhances the ability to make accurate predictions,monitor structural health,and support proactive maintenance practices,thereby ensuring the safety and longevity of critical infrastructure.展开更多
绵羊的姿态与其健康及福利密切相关。随着智能化畜牧业需求的增长,自动、准确地检测绵羊姿态尤为尤为重要。本研究提出基于Mask R-CNN基准网络的新型RDS-Mask R-CNN绵羊姿态检测算法,以Res2Net101作为特征提取网络,同时引入可变形卷积(D...绵羊的姿态与其健康及福利密切相关。随着智能化畜牧业需求的增长,自动、准确地检测绵羊姿态尤为尤为重要。本研究提出基于Mask R-CNN基准网络的新型RDS-Mask R-CNN绵羊姿态检测算法,以Res2Net101作为特征提取网络,同时引入可变形卷积(Deformable convolution network,DCN),以更精准捕捉绵羊在不同位置的姿态特征,并运用软非极大值抑制(Soft non-maximum suppression,Soft NMS)算法实现重叠实例目标的准确分割。结果表明:1)目标检测框架算法对比:与该领域最经典的YOLOv3和Faster R-CNN相比,改进的算法在平均精度均值(Mean average precision,mAP)上分别提升了16.68%和8.64%;2)不同改进策略的算法对比:改进算法相较于基准网络,边界框平均精度均值(Bounding box mean average precision,Bbox mAP)提高6.21%,分割平均精度均值(Segmentation mean average precision,Segm mAP)提高6.61%,分别达到87.34%和81.50%;3)相较于Mask R-CNN,改进模型在识别绵羊站立与躺卧姿态时边界框平均精度(Bounding box average precision,Bbox AP)分别提高了6.84%和5.58%,分割平均精度(Segmentation average precision,Segm AP)分别提高了7.25%和5.17%;4)模型可解释性可视化结果表明RDS-Mask R-CNN能精准捕获绵羊站立和躺卧姿态关键部位深度特征,表明模型自动检测可行且具有可解释性。综上,本研究提出的RDS-Mask R-CNN算法,有效提升了绵羊姿态检测的精准度,为智慧养殖提供了技术支撑。展开更多
Hemoglobin is a vital protein in red blood cells responsible for transporting oxygen throughout the body.Its accurate measurement is crucial for diagnosing and managing conditions such as anemia and diabetes,where abn...Hemoglobin is a vital protein in red blood cells responsible for transporting oxygen throughout the body.Its accurate measurement is crucial for diagnosing and managing conditions such as anemia and diabetes,where abnormal hemoglobin levels can indicate significant health issues.Traditional methods for hemoglobin measurement are invasive,causing pain,risk of infection,and are less convenient for frequent monitoring.PPG is a transformative technology in wearable healthcare for noninvasive monitoring and widely explored for blood pressure,sleep,blood glucose,and stress analysis.In this work,we propose a hemoglobin estimation method using an adaptive lightweight convolutional neural network(HMALCNN)from PPG.The HMALCNN is designed to capture both fine-grained local waveform characteristics and global contextual patterns,ensuring robust performance across acquisition settings.We validated our approach on two multi-regional datasets containing 152 and 68 subjects,respectively,employing a subjectindependent 5-fold cross-validation strategy.The proposed method achieved root mean square errors(RMSE)of 0.90 and 1.20 g/dL for the two datasets,with strong Pearson correlations of 0.82 and 0.72.We conducted extensive posthoc analyses to assess clinical utility and interpretability.A±1 g/dL clinical error tolerance evaluation revealed that 91.3%and 86.7%of predictions for the two datasets fell within the acceptable clinical range.Hemoglobin range-wise analysis demonstrated consistently high accuracy in the normal and low hemoglobin categories.Statistical significance testing using the Wilcoxon signed-rank test confirmed the stability of performance across validation folds(p>0.05 for both RMSE and correlation).Furthermore,model interpretability was enhanced using Gradient-weighted Class Activation Mapping(Grad-CAM),supporting the model’s clinical trustworthiness.The proposed HMALCNN offers a computationally efficient,clinically interpretable,and generalizable framework for noninvasive hemoglobin monitoring,with strong potential for integration into wearable healthcare systems as a practical alternative to invasive measurement techniques.展开更多
To enhance the inference efficiency of convolutional neural network(CNN),tensor parallelism is employed to improve the parallelism within operators.However,existing methods are customized to specific networks and hard...To enhance the inference efficiency of convolutional neural network(CNN),tensor parallelism is employed to improve the parallelism within operators.However,existing methods are customized to specific networks and hardware,limiting their generalizability.This paper proposes an approach called resource-adaptive tensor decomposition(RATD)for CNN operators,which aims to achieve an optimal match between computational resources and parallel computing tasks.Firstly,CNN is represented with fine-grained tensors at the lower graph level,thereby decoupling tensors that can be computed in parallel within operators.Secondly,the convolution and pooling operators are fused,and the decoupled tensor blocks are scheduled in parallel.Finally,a cost model is constructed,based on runtime and resource utilization,to iteratively refine the process of tensor block decomposition and automatically determine the optimal tensor decomposition.Experimental results demonstrate that the proposed RATD improves the accuracy of the model by 11%.Compared with CUDA(compute unified device architecture)deep neural network library(cuDNN),RATD achieves an average speedup ratio of 1.21 times in inference time across various convolution kernels,along with a 12%increase in computational resource utilization.展开更多
文摘针对不同磁密幅值、频率、谐波组合等复杂激励工况下磁致伸缩建模面临的精准性问题,该文利用空间注意力机制(spatial attention mechanism,SAM)对传统的卷积神经网络(convolutional neural network,CNN)进行改进,将SAM嵌套入CNN网络中,建立SAMCNN改进型网络。再结合双向长短期记忆(bidirectional long short-term memory,BiLSTM)网络,提出电工钢片SAMCNN-BiLSTM磁致伸缩模型。首先,利用灰狼优化算法(grey wolf optimization,GWO)寻优神经网络结构的参数,实现复杂工况下磁致伸缩效应的准确表征;然后,建立中低频范围单频与叠加谐波激励等复杂工况下的磁致伸缩应变数据库,开展数据预处理与特征分析;最后,对SAMCNN-BiLSTM模型开展对比验证。对比叠加3次谐波激励下的磁致伸缩应变频谱主要分量,SAMCNN-BiLSTM模型计算值最大相对误差为3.70%,其比Jiles-Atherton-Sablik(J-A-S)、二次畴转等模型能更精确地表征电工钢片的磁致伸缩效应。
文摘The integration of image analysis through deep learning(DL)into rock classification represents a significant leap forward in geological research.While traditional methods remain invaluable for their expertise and historical context,DL offers a powerful complement by enhancing the speed,objectivity,and precision of the classification process.This research explores the significance of image data augmentation techniques in optimizing the performance of convolutional neural networks(CNNs)for geological image analysis,particularly in the classification of igneous,metamorphic,and sedimentary rock types from rock thin section(RTS)images.This study primarily focuses on classic image augmentation techniques and evaluates their impact on model accuracy and precision.Results demonstrate that augmentation techniques like Equalize significantly enhance the model's classification capabilities,achieving an F1-Score of 0.9869 for igneous rocks,0.9884 for metamorphic rocks,and 0.9929 for sedimentary rocks,representing improvements compared to the baseline original results.Moreover,the weighted average F1-Score across all classes and techniques is 0.9886,indicating an enhancement.Conversely,methods like Distort lead to decreased accuracy and F1-Score,with an F1-Score of 0.949 for igneous rocks,0.954 for metamorphic rocks,and 0.9416 for sedimentary rocks,exacerbating the performance compared to the baseline.The study underscores the practicality of image data augmentation in geological image classification and advocates for the adoption of DL methods in this domain for automation and improved results.The findings of this study can benefit various fields,including remote sensing,mineral exploration,and environmental monitoring,by enhancing the accuracy of geological image analysis both for scientific research and industrial applications.
基金funding support from the National Natural Science Foundation of China(Grant Nos.U22A20594,52079045)Hong-Zhi Cui acknowledges the financial support of the China Scholarship Council(Grant No.CSC:202206710014)for his research at Universitat Politecnica de Catalunya,Barcelona.
文摘Landslide susceptibility mapping(LSM)plays a crucial role in assessing geological risks.The current LSM techniques face a significant challenge in achieving accurate results due to uncertainties associated with regional-scale geotechnical parameters.To explore rainfall-induced LSM,this study proposes a hybrid model that combines the physically-based probabilistic model(PPM)with convolutional neural network(CNN).The PPM is capable of effectively capturing the spatial distribution of landslides by incorporating the probability of failure(POF)considering the slope stability mechanism under rainfall conditions.This significantly characterizes the variation of POF caused by parameter uncertainties.CNN was used as a binary classifier to capture the spatial and channel correlation between landslide conditioning factors and the probability of landslide occurrence.OpenCV image enhancement technique was utilized to extract non-landslide points based on the POF of landslides.The proposed model comprehensively considers physical mechanics when selecting non-landslide samples,effectively filtering out samples that do not adhere to physical principles and reduce the risk of overfitting.The results indicate that the proposed PPM-CNN hybrid model presents a higher prediction accuracy,with an area under the curve(AUC)value of 0.85 based on the landslide case of the Niangniangba area of Gansu Province,China compared with the individual CNN model(AUC=0.61)and the PPM(AUC=0.74).This model can also consider the statistical correlation and non-normal probability distributions of model parameters.These results offer practical guidance for future research on rainfall-induced LSM at the regional scale.
基金supported by the Universiti Tunku Abdul Rahman (UTAR) Malaysia under UTARRF (IPSR/RMC/UTARRF/2021-C1/T05)
文摘The recent surge of mobile subscribers and user data traffic has accelerated the telecommunication sector towards the adoption of the fifth-generation (5G) mobile networks. Cloud radio access network (CRAN) is a prominent framework in the 5G mobile network to meet the above requirements by deploying low-cost and intelligent multiple distributed antennas known as remote radio heads (RRHs). However, achieving the optimal resource allocation (RA) in CRAN using the traditional approach is still challenging due to the complex structure. In this paper, we introduce the convolutional neural network-based deep Q-network (CNN-DQN) to balance the energy consumption and guarantee the user quality of service (QoS) demand in downlink CRAN. We first formulate the Markov decision process (MDP) for energy efficiency (EE) and build up a 3-layer CNN to capture the environment feature as an input state space. We then use DQN to turn on/off the RRHs dynamically based on the user QoS demand and energy consumption in the CRAN. Finally, we solve the RA problem based on the user constraint and transmit power to guarantee the user QoS demand and maximize the EE with a minimum number of active RRHs. In the end, we conduct the simulation to compare our proposed scheme with nature DQN and the traditional approach.
文摘This study introduces an innovative“Big Model”strategy to enhance Bridge Structural Health Monitoring(SHM)using a Convolutional Neural Network(CNN),time-frequency analysis,and fine element analysis.Leveraging ensemble methods,collaborative learning,and distributed computing,the approach effectively manages the complexity and scale of large-scale bridge data.The CNN employs transfer learning,fine-tuning,and continuous monitoring to optimize models for adaptive and accurate structural health assessments,focusing on extracting meaningful features through time-frequency analysis.By integrating Finite Element Analysis,time-frequency analysis,and CNNs,the strategy provides a comprehensive understanding of bridge health.Utilizing diverse sensor data,sophisticated feature extraction,and advanced CNN architecture,the model is optimized through rigorous preprocessing and hyperparameter tuning.This approach significantly enhances the ability to make accurate predictions,monitor structural health,and support proactive maintenance practices,thereby ensuring the safety and longevity of critical infrastructure.
文摘绵羊的姿态与其健康及福利密切相关。随着智能化畜牧业需求的增长,自动、准确地检测绵羊姿态尤为尤为重要。本研究提出基于Mask R-CNN基准网络的新型RDS-Mask R-CNN绵羊姿态检测算法,以Res2Net101作为特征提取网络,同时引入可变形卷积(Deformable convolution network,DCN),以更精准捕捉绵羊在不同位置的姿态特征,并运用软非极大值抑制(Soft non-maximum suppression,Soft NMS)算法实现重叠实例目标的准确分割。结果表明:1)目标检测框架算法对比:与该领域最经典的YOLOv3和Faster R-CNN相比,改进的算法在平均精度均值(Mean average precision,mAP)上分别提升了16.68%和8.64%;2)不同改进策略的算法对比:改进算法相较于基准网络,边界框平均精度均值(Bounding box mean average precision,Bbox mAP)提高6.21%,分割平均精度均值(Segmentation mean average precision,Segm mAP)提高6.61%,分别达到87.34%和81.50%;3)相较于Mask R-CNN,改进模型在识别绵羊站立与躺卧姿态时边界框平均精度(Bounding box average precision,Bbox AP)分别提高了6.84%和5.58%,分割平均精度(Segmentation average precision,Segm AP)分别提高了7.25%和5.17%;4)模型可解释性可视化结果表明RDS-Mask R-CNN能精准捕获绵羊站立和躺卧姿态关键部位深度特征,表明模型自动检测可行且具有可解释性。综上,本研究提出的RDS-Mask R-CNN算法,有效提升了绵羊姿态检测的精准度,为智慧养殖提供了技术支撑。
基金funded by the Deanship of Graduate Studies and Scientific Research at Qassim University for financial support(QU-APC-2025).
文摘Hemoglobin is a vital protein in red blood cells responsible for transporting oxygen throughout the body.Its accurate measurement is crucial for diagnosing and managing conditions such as anemia and diabetes,where abnormal hemoglobin levels can indicate significant health issues.Traditional methods for hemoglobin measurement are invasive,causing pain,risk of infection,and are less convenient for frequent monitoring.PPG is a transformative technology in wearable healthcare for noninvasive monitoring and widely explored for blood pressure,sleep,blood glucose,and stress analysis.In this work,we propose a hemoglobin estimation method using an adaptive lightweight convolutional neural network(HMALCNN)from PPG.The HMALCNN is designed to capture both fine-grained local waveform characteristics and global contextual patterns,ensuring robust performance across acquisition settings.We validated our approach on two multi-regional datasets containing 152 and 68 subjects,respectively,employing a subjectindependent 5-fold cross-validation strategy.The proposed method achieved root mean square errors(RMSE)of 0.90 and 1.20 g/dL for the two datasets,with strong Pearson correlations of 0.82 and 0.72.We conducted extensive posthoc analyses to assess clinical utility and interpretability.A±1 g/dL clinical error tolerance evaluation revealed that 91.3%and 86.7%of predictions for the two datasets fell within the acceptable clinical range.Hemoglobin range-wise analysis demonstrated consistently high accuracy in the normal and low hemoglobin categories.Statistical significance testing using the Wilcoxon signed-rank test confirmed the stability of performance across validation folds(p>0.05 for both RMSE and correlation).Furthermore,model interpretability was enhanced using Gradient-weighted Class Activation Mapping(Grad-CAM),supporting the model’s clinical trustworthiness.The proposed HMALCNN offers a computationally efficient,clinically interpretable,and generalizable framework for noninvasive hemoglobin monitoring,with strong potential for integration into wearable healthcare systems as a practical alternative to invasive measurement techniques.
基金Supported by the National Science and Technology Major Project of China(No.2022ZD0119003)the National Natural Science Foundation of China(No.61834005).
文摘To enhance the inference efficiency of convolutional neural network(CNN),tensor parallelism is employed to improve the parallelism within operators.However,existing methods are customized to specific networks and hardware,limiting their generalizability.This paper proposes an approach called resource-adaptive tensor decomposition(RATD)for CNN operators,which aims to achieve an optimal match between computational resources and parallel computing tasks.Firstly,CNN is represented with fine-grained tensors at the lower graph level,thereby decoupling tensors that can be computed in parallel within operators.Secondly,the convolution and pooling operators are fused,and the decoupled tensor blocks are scheduled in parallel.Finally,a cost model is constructed,based on runtime and resource utilization,to iteratively refine the process of tensor block decomposition and automatically determine the optimal tensor decomposition.Experimental results demonstrate that the proposed RATD improves the accuracy of the model by 11%.Compared with CUDA(compute unified device architecture)deep neural network library(cuDNN),RATD achieves an average speedup ratio of 1.21 times in inference time across various convolution kernels,along with a 12%increase in computational resource utilization.