A method based on multiple images captured under different light sources at different incident angles was developed to recognize the coal density range in this study.The innovation is that two new images were construc...A method based on multiple images captured under different light sources at different incident angles was developed to recognize the coal density range in this study.The innovation is that two new images were constructed based on images captured under four single light sources.Reconstruction image 1 was constructed by fusing greyscale versions of the original images into one image,and Reconstruction image2 was constructed based on the differences between the images captured under the different light sources.Subsequently,the four original images and two reconstructed images were input into the convolutional neural network AlexNet to recognize the density range in three cases:-1.5(clean coal) and+1.5 g/cm^(3)(non-clean coal);-1.8(non-gangue) and+1.8 g/cm^(3)(gangue);-1.5(clean coal),1.5-1.8(middlings),and+1.8 g/cm^(3)(gangue).The results show the following:(1) The reconstructed images,especially Reconstruction image 2,can effectively improve the recognition accuracy for the coal density range compared with images captured under single light source.(2) The recognition accuracies for gangue and non-gangue,clean coal and non-clean coal,and clean coal,middlings,and gangue reached88.44%,86.72% and 77.08%,respectively.(3) The recognition accuracy increases as the density moves further away from the boundary density.展开更多
Convolutional neural networks(CNNs)-based medical image segmentation technologies have been widely used in medical image segmentation because of their strong representation and generalization abilities.However,due to ...Convolutional neural networks(CNNs)-based medical image segmentation technologies have been widely used in medical image segmentation because of their strong representation and generalization abilities.However,due to the inability to effectively capture global information from images,CNNs can easily lead to loss of contours and textures in segmentation results.Notice that the transformer model can effectively capture the properties of long-range dependencies in the image,and furthermore,combining the CNN and the transformer can effectively extract local details and global contextual features of the image.Motivated by this,we propose a multi-branch and multi-scale attention network(M2ANet)for medical image segmentation,whose architecture consists of three components.Specifically,in the first component,we construct an adaptive multi-branch patch module for parallel extraction of image features to reduce information loss caused by downsampling.In the second component,we apply residual block to the well-known convolutional block attention module to enhance the network’s ability to recognize important features of images and alleviate the phenomenon of gradient vanishing.In the third component,we design a multi-scale feature fusion module,in which we adopt adaptive average pooling and position encoding to enhance contextual features,and then multi-head attention is introduced to further enrich feature representation.Finally,we validate the effectiveness and feasibility of the proposed M2ANet method through comparative experiments on four benchmark medical image segmentation datasets,particularly in the context of preserving contours and textures.展开更多
Diabetic Retinopathy(DR)is a critical disorder that affects the retina due to the constant rise in diabetics and remains the major cause of blindness across the world.Early detection and timely treatment are essential...Diabetic Retinopathy(DR)is a critical disorder that affects the retina due to the constant rise in diabetics and remains the major cause of blindness across the world.Early detection and timely treatment are essential to mitigate the effects of DR,such as retinal damage and vision impairment.Several conventional approaches have been proposed to detect DR early and accurately,but they are limited by data imbalance,interpretability,overfitting,convergence time,and other issues.To address these drawbacks and improve DR detection accurately,a distributed Explainable Convolutional Neural network-enabled Light Gradient Boosting Machine(DE-ExLNN)is proposed in this research.The model combines an explainable Convolutional Neural Network(CNN)and Light Gradient Boosting Machine(LightGBM),achieving highly accurate outcomes in DR detection.LightGBM serves as the detection model,and the inclusion of an explainable CNN addresses issues that conventional CNN classifiers could not resolve.A custom dataset was created for this research,containing both fundus and OCTA images collected from a realtime environment,providing more accurate results compared to standard conventional DR datasets.The custom dataset demonstrates notable accuracy,sensitivity,specificity,and Matthews Correlation Coefficient(MCC)scores,underscoring the effectiveness of this approach.Evaluations against other standard datasets achieved an accuracy of 93.94%,sensitivity of 93.90%,specificity of 93.99%,and MCC of 93.88%for fundus images.For OCTA images,the results obtained an accuracy of 95.30%,sensitivity of 95.50%,specificity of 95.09%,andMCC of 95%.Results prove that the combination of explainable CNN and LightGBMoutperforms othermethods.The inclusion of distributed learning enhances the model’s efficiency by reducing time consumption and complexity while facilitating feature extraction.展开更多
Multi-source information can be obtained through the fusion of infrared images and visible light images,which have the characteristics of complementary information.However,the existing acquisition methods of fusion im...Multi-source information can be obtained through the fusion of infrared images and visible light images,which have the characteristics of complementary information.However,the existing acquisition methods of fusion images have disadvantages such as blurred edges,low contrast,and loss of details.Based on convolution sparse representation and improved pulse-coupled neural network this paper proposes an image fusion algorithm that decompose the source images into high-frequency and low-frequency subbands by non-subsampled Shearlet Transform(NSST).Furthermore,the low-frequency subbands were fused by convolutional sparse representation(CSR),and the high-frequency subbands were fused by an improved pulse coupled neural network(IPCNN)algorithm,which can effectively solve the problem of difficulty in setting parameters of the traditional PCNN algorithm,improving the performance of sparse representation with details injection.The result reveals that the proposed method in this paper has more advantages than the existing mainstream fusion algorithms in terms of visual effects and objective indicators.展开更多
Real-time,contact-free temperature monitoring of low to medium range(30℃-150℃)has been extensively used in industry and agriculture,which is usually realized by costly infrared temperature detection methods.This pap...Real-time,contact-free temperature monitoring of low to medium range(30℃-150℃)has been extensively used in industry and agriculture,which is usually realized by costly infrared temperature detection methods.This paper proposes an alternative approach of extracting temperature information in real time from the visible light images of the monitoring target using a convolutional neural network(CNN).A mean-square error of<1.119℃was reached in the temperature measurements of low to medium range using the CNN and the visible light images.Imaging angle and imaging distance do not affect the temperature detection using visible optical images by the CNN.Moreover,the CNN has a certain illuminance generalization ability capable of detection temperature information from the images which were collected under different illuminance and were not used for training.Compared to the conventional machine learning algorithms mentioned in the recent literatures,this real-time,contact-free temperature measurement approach that does not require any further image processing operations facilitates temperature monitoring applications in the industrial and civil fields.展开更多
In this study,we proposed a multi-source approach for mapping local-scale population density of England.Specifically,we mapped both the working and daytime population densities by integrating the multi-source data suc...In this study,we proposed a multi-source approach for mapping local-scale population density of England.Specifically,we mapped both the working and daytime population densities by integrating the multi-source data such as residential population density,point-of-interest density,point-of-interest category mix,and nighttime light intensity.It is demonstrated that combining remote sensing and social sensing data provides a plausible way to map annual working or daytime population densities.In this paper,we trained models with England-wide data and subsequently tested these models with Wales-wide data.In addition,we further tested the models with England-wide data at a higher level of spatial granularity.Particularly,the random forest and convolutional neural network models were adopted to map population density.The estimated results and validation suggest that the three built models have high prediction accuracies at the local authority district level.It is shown that the convolutional neural network models have the greatest prediction accuracies at the local authority district level though they are most time-consuming.The models trained with the data at the local authority district level are less appropriately applicable to test data at a higher level of spatial granularity.The proposed multi-source approach performs well in mapping local-scale population density.It indicates that combining remote sensing and social sensing data is advantageous to mapping socioeconomic variables.展开更多
文摘A method based on multiple images captured under different light sources at different incident angles was developed to recognize the coal density range in this study.The innovation is that two new images were constructed based on images captured under four single light sources.Reconstruction image 1 was constructed by fusing greyscale versions of the original images into one image,and Reconstruction image2 was constructed based on the differences between the images captured under the different light sources.Subsequently,the four original images and two reconstructed images were input into the convolutional neural network AlexNet to recognize the density range in three cases:-1.5(clean coal) and+1.5 g/cm^(3)(non-clean coal);-1.8(non-gangue) and+1.8 g/cm^(3)(gangue);-1.5(clean coal),1.5-1.8(middlings),and+1.8 g/cm^(3)(gangue).The results show the following:(1) The reconstructed images,especially Reconstruction image 2,can effectively improve the recognition accuracy for the coal density range compared with images captured under single light source.(2) The recognition accuracies for gangue and non-gangue,clean coal and non-clean coal,and clean coal,middlings,and gangue reached88.44%,86.72% and 77.08%,respectively.(3) The recognition accuracy increases as the density moves further away from the boundary density.
基金supported by the Natural Science Foundation of the Anhui Higher Education Institutions of China(Grant Nos.2023AH040149 and 2024AH051915)the Anhui Provincial Natural Science Foundation(Grant No.2208085MF168)+1 种基金the Science and Technology Innovation Tackle Plan Project of Maanshan(Grant No.2024RGZN001)the Scientific Research Fund Project of Anhui Medical University(Grant No.2023xkj122).
文摘Convolutional neural networks(CNNs)-based medical image segmentation technologies have been widely used in medical image segmentation because of their strong representation and generalization abilities.However,due to the inability to effectively capture global information from images,CNNs can easily lead to loss of contours and textures in segmentation results.Notice that the transformer model can effectively capture the properties of long-range dependencies in the image,and furthermore,combining the CNN and the transformer can effectively extract local details and global contextual features of the image.Motivated by this,we propose a multi-branch and multi-scale attention network(M2ANet)for medical image segmentation,whose architecture consists of three components.Specifically,in the first component,we construct an adaptive multi-branch patch module for parallel extraction of image features to reduce information loss caused by downsampling.In the second component,we apply residual block to the well-known convolutional block attention module to enhance the network’s ability to recognize important features of images and alleviate the phenomenon of gradient vanishing.In the third component,we design a multi-scale feature fusion module,in which we adopt adaptive average pooling and position encoding to enhance contextual features,and then multi-head attention is introduced to further enrich feature representation.Finally,we validate the effectiveness and feasibility of the proposed M2ANet method through comparative experiments on four benchmark medical image segmentation datasets,particularly in the context of preserving contours and textures.
基金funded by the Centre for Advanced Modelling and Geospatial Information Systems(CAMGIS),Faculty of Engineering and IT,University of Technology Sydneysupported by the Research Funding Program,King Saud University,Riyadh,Saudi Arabia,under Project Ongoing Research Funding program(ORF-2025-14).
文摘Diabetic Retinopathy(DR)is a critical disorder that affects the retina due to the constant rise in diabetics and remains the major cause of blindness across the world.Early detection and timely treatment are essential to mitigate the effects of DR,such as retinal damage and vision impairment.Several conventional approaches have been proposed to detect DR early and accurately,but they are limited by data imbalance,interpretability,overfitting,convergence time,and other issues.To address these drawbacks and improve DR detection accurately,a distributed Explainable Convolutional Neural network-enabled Light Gradient Boosting Machine(DE-ExLNN)is proposed in this research.The model combines an explainable Convolutional Neural Network(CNN)and Light Gradient Boosting Machine(LightGBM),achieving highly accurate outcomes in DR detection.LightGBM serves as the detection model,and the inclusion of an explainable CNN addresses issues that conventional CNN classifiers could not resolve.A custom dataset was created for this research,containing both fundus and OCTA images collected from a realtime environment,providing more accurate results compared to standard conventional DR datasets.The custom dataset demonstrates notable accuracy,sensitivity,specificity,and Matthews Correlation Coefficient(MCC)scores,underscoring the effectiveness of this approach.Evaluations against other standard datasets achieved an accuracy of 93.94%,sensitivity of 93.90%,specificity of 93.99%,and MCC of 93.88%for fundus images.For OCTA images,the results obtained an accuracy of 95.30%,sensitivity of 95.50%,specificity of 95.09%,andMCC of 95%.Results prove that the combination of explainable CNN and LightGBMoutperforms othermethods.The inclusion of distributed learning enhances the model’s efficiency by reducing time consumption and complexity while facilitating feature extraction.
基金supported in part by the National Natural Science Foundation of China under Grant 41505017.
文摘Multi-source information can be obtained through the fusion of infrared images and visible light images,which have the characteristics of complementary information.However,the existing acquisition methods of fusion images have disadvantages such as blurred edges,low contrast,and loss of details.Based on convolution sparse representation and improved pulse-coupled neural network this paper proposes an image fusion algorithm that decompose the source images into high-frequency and low-frequency subbands by non-subsampled Shearlet Transform(NSST).Furthermore,the low-frequency subbands were fused by convolutional sparse representation(CSR),and the high-frequency subbands were fused by an improved pulse coupled neural network(IPCNN)algorithm,which can effectively solve the problem of difficulty in setting parameters of the traditional PCNN algorithm,improving the performance of sparse representation with details injection.The result reveals that the proposed method in this paper has more advantages than the existing mainstream fusion algorithms in terms of visual effects and objective indicators.
基金Project supported by the National Natural Science Foundation of China (Grant Nos.61975072 and 12174173)the Natural Science Foundation of Fujian Province,China (Grant Nos.2022H0023,2022J02047,ZZ2023J20,and 2022G02006)。
文摘Real-time,contact-free temperature monitoring of low to medium range(30℃-150℃)has been extensively used in industry and agriculture,which is usually realized by costly infrared temperature detection methods.This paper proposes an alternative approach of extracting temperature information in real time from the visible light images of the monitoring target using a convolutional neural network(CNN).A mean-square error of<1.119℃was reached in the temperature measurements of low to medium range using the CNN and the visible light images.Imaging angle and imaging distance do not affect the temperature detection using visible optical images by the CNN.Moreover,the CNN has a certain illuminance generalization ability capable of detection temperature information from the images which were collected under different illuminance and were not used for training.Compared to the conventional machine learning algorithms mentioned in the recent literatures,this real-time,contact-free temperature measurement approach that does not require any further image processing operations facilitates temperature monitoring applications in the industrial and civil fields.
文摘In this study,we proposed a multi-source approach for mapping local-scale population density of England.Specifically,we mapped both the working and daytime population densities by integrating the multi-source data such as residential population density,point-of-interest density,point-of-interest category mix,and nighttime light intensity.It is demonstrated that combining remote sensing and social sensing data provides a plausible way to map annual working or daytime population densities.In this paper,we trained models with England-wide data and subsequently tested these models with Wales-wide data.In addition,we further tested the models with England-wide data at a higher level of spatial granularity.Particularly,the random forest and convolutional neural network models were adopted to map population density.The estimated results and validation suggest that the three built models have high prediction accuracies at the local authority district level.It is shown that the convolutional neural network models have the greatest prediction accuracies at the local authority district level though they are most time-consuming.The models trained with the data at the local authority district level are less appropriately applicable to test data at a higher level of spatial granularity.The proposed multi-source approach performs well in mapping local-scale population density.It indicates that combining remote sensing and social sensing data is advantageous to mapping socioeconomic variables.