Rank Histograms are suitable tools to assess the quality of ensembles within an ensemble prediction system or framework. By counting the rank of a given variable in the ensemble, we are basically making a sample analy...Rank Histograms are suitable tools to assess the quality of ensembles within an ensemble prediction system or framework. By counting the rank of a given variable in the ensemble, we are basically making a sample analysis, which does not allow us to distinguish if the origin of its variability is external noise or comes from chaotic sources. The recently introduced Mean to Variance Logarithmic (MVL) Diagram accounts for the spatial variability, being very sensitive to the spatial localization produced by infinitesimal perturbations of spatiotemporal chaotic systems. By using as a benchmark a simple model subject to noise, we show the distinct information given by Rank Histograms and MVL Diagrams. Hence, the main effects of the external noise can be visualized in a graphic. From the MVL diagram we clearly observe a reduction of the amplitude growth rate and of the spatial localization (chaos suppression), while from the Rank Histogram we observe changes in the reliability of the ensemble. We conclude that in a complex framework including spatiotemporal chaos and noise, both provide a more complete forecasting picture.展开更多
Content based image retrieval(CBIR)techniques have been widely deployed in many applications for seeking the abundant information existed in images.Due to large amounts of storage and computational requirements of CBI...Content based image retrieval(CBIR)techniques have been widely deployed in many applications for seeking the abundant information existed in images.Due to large amounts of storage and computational requirements of CBIR,outsourcing image search work to the cloud provider becomes a very attractive option for many owners with small devices.However,owing to the private content contained in images,directly outsourcing retrieval work to the cloud provider apparently bring about privacy problem,so the images should be protected carefully before outsourcing.This paper presents a secure retrieval scheme for the encrypted images in the YUV color space.With this scheme,the discrete cosine transform(DCT)is performed on the Y component.The resulting DC coefficients are encrypted with stream cipher technology and the resulting AC coefficients as well as other two color components are encrypted with value permutation and position scrambling.Then the image owner transmits the encrypted images to the cloud server.When receiving a query trapdoor form on query user,the server extracts AC-coefficients histogram from the encrypted Y component and extracts two color histograms from the other two color components.The similarity between query trapdoor and database image is measured by calculating the Manhattan distance of their respective histograms.Finally,the encrypted images closest to the query image are returned to the query user.展开更多
Image segmentation denotes a process for partitioning an image into distinct regions, it plays an important role in interpretation and decision making. A large variety of segmentation methods has been developed;among ...Image segmentation denotes a process for partitioning an image into distinct regions, it plays an important role in interpretation and decision making. A large variety of segmentation methods has been developed;among them, multidimensional histogram methods have been investigated but their implementation stays difficult due to the big size of histograms. We present an original method for segmenting n-D (where n is the number of components in image) images or multidimensional images in an unsupervised way using a fuzzy neighbourhood model. It is based on the hierarchical analysis of full n-D compact histograms integrating a fuzzy connected components labelling algorithm that we have realized in this work. Each peak of the histo- gram constitutes a class kernel, as soon as it encloses a number of pixels greater than or equal to a secondary arbitrary threshold knowing that a first threshold was set to define the degree of binary fuzzy similarity be- tween pixels. The use of a lossless compact n-D histogram allows a drastic reduction of the memory space necessary for coding it. As a consequence, the segmentation can be achieved without reducing the colors population of images in the classification step. It is shown that using n-D compact histograms, instead of 1-D and 2-D ones, leads to better segmentation results. Various images were segmented;the evaluation of the quality of segmentation in supervised and unsupervised of segmentation method proposed compare to the classification method k-means gives better results. It thus highlights the relevance of our approach, which can be used for solving many problems of segmentation.展开更多
A new gray-spatial histogram is proposed, which incorporates spatial informatio n with gray compositions without sacrificing the robustness of traditional gray histograms. The purpose is to consider the representation...A new gray-spatial histogram is proposed, which incorporates spatial informatio n with gray compositions without sacrificing the robustness of traditional gray histograms. The purpose is to consider the representation role of gray compositi ons and spatial information simultaneously. Each entry in the gray-spatial hist ogram is the gray frequency and corresponding position information of images. In the experiments of sonar image recognition, the results show that the gray-spa tial histogram is effective in practical use.展开更多
A new algorithm taking the spatial context of local features into account by utilizing contextualized histograms was proposed to recognize facial expression. The contextualized histograms were extracted fromtwo widely...A new algorithm taking the spatial context of local features into account by utilizing contextualized histograms was proposed to recognize facial expression. The contextualized histograms were extracted fromtwo widely used descriptors—the local binary pattern( LBP) and weber local descriptor( WLD). The LBP and WLD feature histograms were extracted separately fromeach facial image,and contextualized histogram was generated as feature vectors to feed the classifier. In addition,the human face was divided into sub-blocks and each sub-block was assigned different weights by their different contributions to the intensity of facial expressions to improve the recognition rate. With the support vector machine(SVM) as classifier,the experimental results on the 2D texture images fromthe 3D-BU FE dataset indicated that contextualized histograms improved facial expression recognition performance when local features were employed.展开更多
Alzheimer’s Disease(AD)is a progressive neurodegenerative disorder that significantly affects cognitive function,making early and accurate diagnosis essential.Traditional Deep Learning(DL)-based approaches often stru...Alzheimer’s Disease(AD)is a progressive neurodegenerative disorder that significantly affects cognitive function,making early and accurate diagnosis essential.Traditional Deep Learning(DL)-based approaches often struggle with low-contrast MRI images,class imbalance,and suboptimal feature extraction.This paper develops a Hybrid DL system that unites MobileNetV2 with adaptive classification methods to boost Alzheimer’s diagnosis by processing MRI scans.Image enhancement is done using Contrast-Limited Adaptive Histogram Equalization(CLAHE)and Enhanced Super-Resolution Generative Adversarial Networks(ESRGAN).A classification robustness enhancement system integrates class weighting techniques and a Matthews Correlation Coefficient(MCC)-based evaluation method into the design.The trained and validated model gives a 98.88%accuracy rate and 0.9614 MCC score.We also performed a 10-fold cross-validation experiment with an average accuracy of 96.52%(±1.51),a loss of 0.1671,and an MCC score of 0.9429 across folds.The proposed framework outperforms the state-of-the-art models with a 98%weighted F1-score while decreasing misdiagnosis results for every AD stage.The model demonstrates apparent separation abilities between AD progression stages according to the results of the confusion matrix analysis.These results validate the effectiveness of hybrid DL models with adaptive preprocessing for early and reliable Alzheimer’s diagnosis,contributing to improved computer-aided diagnosis(CAD)systems in clinical practice.展开更多
When am I ever going to use this?ANIMALS The cheetah is the fastest known and animal.The table shows its fastest speed and the top speeds of four other land animals.1.What are the fastest and slowest speeds recorded i...When am I ever going to use this?ANIMALS The cheetah is the fastest known and animal.The table shows its fastest speed and the top speeds of four other land animals.1.What are the fastest and slowest speeds recorded in the table?2.How can you create a visual representation to summarize the data?3.Do any of these representations show both the animal name and its speed?Abar graphis one method of comparing data by using solid bars to represent quantities.展开更多
Histogram and kernel estimators are usually regarded as the two main classical data-based nonparametric tools to estimate the underlying density functions for some given data sets. In this paper we will integrate them...Histogram and kernel estimators are usually regarded as the two main classical data-based nonparametric tools to estimate the underlying density functions for some given data sets. In this paper we will integrate them and define a histogram-kernel error based on the integrated square error between histogram and binned kernel density estimator, and then exploit its asymptotic properties. 3ust as indicated in this paper, the histogram-kernel error only depends on the choice of bin width and the data for the given prior kernel densities. The asymptotic optimal bin width is derived by minimizing the mean histogram-kernel error. By comparing with Scott's optimal bin width formula for a histogram, a new method is proposed to construct the data-based histogram without knowledge of the underlying density function. Monte Carlo study is used to verify the usefulness of our method for different kinds of density functions and sample sizes.展开更多
Gabor features have been shown to be effective for palm vein recognition. This paper presents a novel feature representation method, implementing the fusion of local Gabor histograms (FLGH), in order to improve the ...Gabor features have been shown to be effective for palm vein recognition. This paper presents a novel feature representation method, implementing the fusion of local Gabor histograms (FLGH), in order to improve the accuracy of palm vein recognition systems. A new local descriptor called local Gabor principal differences patterns (LGPDP) encodes the Gabor magnitude using the local maximum difference (LMD) operator. The corresponding Gabor phase patterns are encoded by local Gabor exclusive OR (XOR) patterns (LGXP). Fisher's linear discriminant (FLD) method is then implemented to reduce the dimensionality of the feature representation. Low-dimensional Gabor magnitude and phase feature vectors are finally fused to enhance accuracy. Experimental results from Institute of Automation, Chinese Academy of sciences (CASIA) database show that the proposed FLGH method achieves better performance by utilizing score-level fusion. The equal error rate (EER) is 0.08%, which outperforms other conventional palm vein recognition methods (EER range from 2.87% to 0.16%), e.g., the Laplacian palm, minutiae feature, Hessian phase, Eigenvein, local invariant features, mutual foreground local binary patterns (LBP), and multi-sampling feature fusion methods.展开更多
An in-pixel histogramming time-to-digital converter(hTDC)based on octonary search and 4-tap phase detection is presented,aiming to improve frame rate while ensuring high precicion.The proposed hTDC is a 12-bit two-ste...An in-pixel histogramming time-to-digital converter(hTDC)based on octonary search and 4-tap phase detection is presented,aiming to improve frame rate while ensuring high precicion.The proposed hTDC is a 12-bit two-step converter consisting of a 6-bit coarse quantization and a 6-bit fine quantization,which supports a time resolution of 120 ps and multiphoton counting up to 2 GHz without a GHz reference frequency.The proposed hTDC is designed in 0.11μm CMOS process with an area consumption of 6900μm^(2).The data from a behavioral-level model is imported into the designed hTDC circuit for simulation verification.The post-simulation results show that the proposed hTDC achieves 0.8%depth precision in 9 m range for short-range system design specifications and 0.2%depth precision in 48 m range for long-range system design specifications.Under 30×10^(3) lux background light conditions,the proposed hTDC can be used for SPAD-based flash LiDAR sensor to achieve a frame rate to 40 fps with 200 ps resolution in 9 m range.展开更多
Ground penetrating radar(GPR),as a fast,efficient,and non-destructive detection device,holds great potential for the detection of shallow subsurface environments,such as urban road subsurface monitoring.However,the in...Ground penetrating radar(GPR),as a fast,efficient,and non-destructive detection device,holds great potential for the detection of shallow subsurface environments,such as urban road subsurface monitoring.However,the interpretation of GPR echo images often relies on manual recognition by experienced engineers.In order to address the automatic interpretation of cavity targets in GPR echo images,a recognition-algorithm based on Gaussian mixed model-hidden Markov model(GMM-HMM)is proposed,which can recognize three dimensional(3D)underground voids automatically.First,energy detection on the echo images is performed,whereby the data is preprocessed and pre-filtered.Then,edge histogram descriptor(EHD),histogram of oriented gradient(HOG),and Log-Gabor filters are used to extract features from the images.The traditional method can only be applied to 2D images and pre-processing is required for C-scan images.Finally,the aggregated features are fed into the GMM-HMM for classification and compared with two other methods,long short-term memory(LSTM)and gate recurrent unit(GRU).By testing on a simulated dataset,an accuracy rate of 90%is obtained,demonstrating the effectiveness and efficiency of our proposed method.展开更多
Skin cancer,a severe health threat,can spread rapidly if undetected.Therefore,early detection can lead to an advanced and efficient diagnosis,thus reducing mortality.Unsupervised classification techniques analyse exte...Skin cancer,a severe health threat,can spread rapidly if undetected.Therefore,early detection can lead to an advanced and efficient diagnosis,thus reducing mortality.Unsupervised classification techniques analyse extensive skin image datasets,identifying patterns and anomalies without prior labelling,facilitating early detection and effective diagnosis and potentially saving lives.In this study,the authors aim to explore the potential of unsupervised learning methods in classifying different types of skin lesions in dermatoscopic images.The authors aim to bridge the gap in dermatological research by introducing innovative techniques that enhance image quality and improve feature extraction.To achieve this,enhanced super-resolution generative adversarial networks(ESRGAN)was fine-tuned to strengthen the resolution of skin lesion images,making critical features more visible.The authors extracted histogram features to capture essential colour characteristics and used the Davies-Bouldin index and silhouette score to determine optimal clusters.Fine-tuned k-means clustering with Euclidean distance in the histogram feature space achieved 87.77% and 90.5% test accuracies on the ISIC2019 and HAM10000 datasets,respectively.The unsupervised approach effectively categorises skin lesions,indicating that unsupervised learning can significantly advance dermatology by enabling early detection and classification without extensive manual annotation.展开更多
Background:Pneumonia remains a critical global health challenge,manifesting as a severe respiratory infection caused by viruses,bacteria,and fungi.Early detection is paramount for effective treatment,potentially reduc...Background:Pneumonia remains a critical global health challenge,manifesting as a severe respiratory infection caused by viruses,bacteria,and fungi.Early detection is paramount for effective treatment,potentially reducing mortality rates and optimizing healthcare resource allocation.Despite the importance of chest X-ray diagnosis,image analysis presents significant challenges,particularly in regions with limited medical expertise.This study addresses these challenges by proposing a computer-aided diagnosis system leveraging targeted image preprocessing and optimized deep learning techniques.Methods:We systematically evaluated contrast limited adaptive histogram equalization with varying clip limits for preprocessing chest X-ray images,demonstrating its effectiveness in enhancing feature visibility for diagnostic accuracy.Employing a comprehensive dataset of 5,863 X-ray images(1,583 pneumonia-negative,4,280 pneumonia-positive)collected from multiple healthcare facilities,we conducted a comparative analysis of transfer learning with pre-trained models including ResNet50v2,VGG-19,and MobileNetV2.Statistical validation was performed through 5-fold cross-validation.Results:Our results show that the contrast limited adaptive histogram equalization-enhanced approach with ResNet50v2 achieves 93.40%accuracy,outperforming VGG-19(84.90%)and MobileNetV2(89.70%).Statistical validation confirms the significance of these improvements(P<0.01).The development and optimization resulted in a lightweight mobile application(74 KB)providing rapid diagnostic support(1-2 s response time).Conclusion:The proposed approach demonstrates practical applicability in resource-constrained settings,balancing diagnostic accuracy with deployment efficiency,and offers a viable solution for computer-aided pneumonia diagnosis in areas with limited medical expertise.展开更多
Esophageal cancer is one of the most difficult cancers to treat since it is often at an advanced stage at the time of symptom presentation.For locally advanced esophageal cancer,treatment options include multidiscipli...Esophageal cancer is one of the most difficult cancers to treat since it is often at an advanced stage at the time of symptom presentation.For locally advanced esophageal cancer,treatment options include multidisciplinary treatment such as surgery or definitive chemoradiotherapy.Surgery has a high local control rate because it involves excision of the cancer along with the surrounding organs;however,it is still highly invasive,although advances in surgery have reduced the burden on patients.On the other hand,chemoradiotherapy may also be applicable in cases in which surgery is inoperable owing to complications or distant lymph node metastasis.However,chemoradiotherapy using X-ray irradiation can cause late toxicities,including those to the heart.Proton beam therapy is widely used to treat esophageal cancer because of its characteristics,and some comparisons between proton beam therapy and X-ray therapy or surgery have recently been reported.This review discusses the role of proton beam therapy in esophageal cancer in comparison to X-ray therapy and surgery.展开更多
Early detection of Forest and Land Fires(FLF)is essential to prevent the rapid spread of fire as well as minimize environmental damage.However,accurate detection under real-world conditions,such as low light,haze,and ...Early detection of Forest and Land Fires(FLF)is essential to prevent the rapid spread of fire as well as minimize environmental damage.However,accurate detection under real-world conditions,such as low light,haze,and complex backgrounds,remains a challenge for computer vision systems.This study evaluates the impact of three image enhancement techniques—Histogram Equalization(HE),Contrast Limited Adaptive Histogram Equalization(CLAHE),and a hybrid method called DBST-LCM CLAHE—on the performance of the YOLOv11 object detection model in identifying fires and smoke.The D-Fire dataset,consisting of 21,527 annotated images captured under diverse environmental scenarios and illumination levels,was used to train and evaluate the model.Each enhancement method was applied to the dataset before training.Model performance was assessed using multiple metrics,including Precision,Recall,mean Average Precision at 50%IoU(mAP50),F1-score,and visual inspection through bounding box results.Experimental results show that all three enhancement techniques improved detection performance.HE yielded the highest mAP50 score of 0.771,along with a balanced precision of 0.784 and recall of 0.703,demonstrating strong generalization across different conditions.DBST-LCM CLAHE achieved the highest Precision score of 79%,effectively reducing false positives,particularly in scenes with dispersed smoke or complex textures.CLAHE,with slightly lower overall metrics,contributed to improved local feature detection.Each technique showed distinct advantages:HE enhanced global contrast;CLAHE improved local structure visibility;and DBST-LCM CLAHE provided an optimal balance through dynamic block sizing and local contrast preservation.These results underline the importance of selecting preprocessing methods according to detection priorities,such as minimizing false alarms or maximizing completeness.This research does not propose a new model architecture but rather benchmarks a recent lightweight detector,YOLOv11,combined with image enhancement strategies for practical deployment in FLF monitoring.The findings support the integration of preprocessing techniques to improve detection accuracy,offering a foundation for real-time FLF detection systems on edge devices or drones,particularly in regions like Indonesia.展开更多
AIM:To find the effective contrast enhancement method on retinal images for effective segmentation of retinal features.METHODS:A novel image preprocessing method that used neighbourhood-based improved contrast limited...AIM:To find the effective contrast enhancement method on retinal images for effective segmentation of retinal features.METHODS:A novel image preprocessing method that used neighbourhood-based improved contrast limited adaptive histogram equalization(NICLAHE)to improve retinal image contrast was suggested to aid in the accurate identification of retinal disorders and improve the visibility of fine retinal structures.Additionally,a minimal-order filter was applied to effectively denoise the images without compromising important retinal structures.The novel NICLAHE algorithm was inspired by the classical CLAHE algorithm,but enhanced it by selecting the clip limits and tile sized in a dynamical manner relative to the pixel values in an image as opposed to using fixed values.It was evaluated on the Drive and high-resolution fundus(HRF)datasets on conventional quality measures.RESULTS:The new proposed preprocessing technique was applied to two retinal image databases,Drive and HRF,with four quality metrics being,root mean square error(RMSE),peak signal to noise ratio(PSNR),root mean square contrast(RMSC),and overall contrast.The technique performed superiorly on both the data sets as compared to the traditional enhancement methods.In order to assess the compatibility of the method with automated diagnosis,a deep learning framework named ResNet was applied in the segmentation of retinal blood vessels.Sensitivity,specificity,precision and accuracy were used to analyse the performance.NICLAHE–enhanced images outperformed the traditional techniques on both the datasets with improved accuracy.CONCLUSION:NICLAHE provides better results than traditional methods with less error and improved contrastrelated values.These enhanced images are subsequently measured by sensitivity,specificity,precision,and accuracy,which yield a better result in both datasets.展开更多
Perceptual quality assessment for point cloud is critical for immersive metaverse experience and is a challenging task.Firstly,because point cloud is formed by unstructured 3D points that makes the topology more compl...Perceptual quality assessment for point cloud is critical for immersive metaverse experience and is a challenging task.Firstly,because point cloud is formed by unstructured 3D points that makes the topology more complex.Secondly,the quality impairment generally involves both geometric attributes and color properties,where the measurement of the geometric distortion becomes more complex.We propose a perceptual point cloud quality assessment model that follows the perceptual features of Human Visual System(HVS)and the intrinsic characteristics of the point cloud.The point cloud is first pre-processed to extract the geometric skeleton keypoints with graph filtering-based re-sampling,and local neighboring regions around the geometric skeleton keypoints are constructed by K-Nearest Neighbors(KNN)clustering.For geometric distortion,the Point Feature Histogram(PFH)is extracted as the feature descriptor,and the Earth Mover’s Distance(EMD)between the PFHs of the corresponding local neighboring regions in the reference and the distorted point clouds is calculated as the geometric quality measurement.For color distortion,the statistical moments between the corresponding local neighboring regions are computed as the color quality measurement.Finally,the global perceptual quality assessment model is obtained as the linear weighting aggregation of the geometric and color quality measurement.The experimental results on extensive datasets show that the proposed method achieves the leading performance as compared to the state-of-the-art methods with less computing time.Meanwhile,the experimental results also demonstrate the robustness of the proposed method across various distortion types.The source codes are available at https://github.com/llsurreal919/Point Cloud Quality Assessment.展开更多
Frequency-modulated continuous-wave radar enables the non-contact and privacy-preserving recognition of human behavior.However,the accuracy of behavior recognition is directly influenced by the spatial relationship be...Frequency-modulated continuous-wave radar enables the non-contact and privacy-preserving recognition of human behavior.However,the accuracy of behavior recognition is directly influenced by the spatial relationship between human posture and the radar.To address the issue of low accuracy in behavior recognition when the human body is not directly facing the radar,a method combining local outlier factor with Doppler information is proposed for the correction of multi-classifier recognition results.Initially,the information such as distance,velocity,and micro-Doppler spectrogram of the target is obtained using the fast Fourier transform and histogram of oriented gradients-support vector machine methods,followed by preliminary recognition.Subsequently,Platt scaling is employed to transform recognition results into confidence scores,and finally,the Doppler-local outlier factor method is utilized to calibrate the confidence scores,with the highest confidence classifier result considered as the recognition outcome.Experimental results demonstrate that this approach achieves an average recognition accuracy of 96.23%for comprehensive human behavior recognition in various orientations.展开更多
To improve image quality under low illumination conditions,a novel low-light image enhancement method is proposed in this paper based on multi-illumination estimation and multi-scale fusion(MIMS).Firstly,the illuminat...To improve image quality under low illumination conditions,a novel low-light image enhancement method is proposed in this paper based on multi-illumination estimation and multi-scale fusion(MIMS).Firstly,the illumination is processed by contrast-limited adaptive histogram equalization(CLAHE),adaptive complementary gamma function(ACG),and adaptive detail preserving S-curve(ADPS),respectively,to obtain three components.Then,the fusion-relevant features,exposure,and color contrast are selected as the weight maps.Subsequently,these components and weight maps are fused through multi-scale to generate enhanced illumination.Finally,the enhanced images are obtained by multiplying the enhanced illumination and reflectance.Compared with existing approaches,this proposed method achieves an average increase of 0.81%and 2.89%in the structural similarity index measurement(SSIM)and peak signal-to-noise ratio(PSNR),and a decrease of 6.17%and 32.61%in the natural image quality evaluator(NIQE)and gradient magnitude similarity deviation(GMSD),respectively.展开更多
Eigenstructure-based coherence attributes are efficient and mature techniques for large-scale fracture detection. However, in horizontally bedded and continuous strata, buried fractures in high grayscale value zones a...Eigenstructure-based coherence attributes are efficient and mature techniques for large-scale fracture detection. However, in horizontally bedded and continuous strata, buried fractures in high grayscale value zones are difficult to detect. Furthermore, middleand small-scale fractures in fractured zones where migration image energies are usually not concentrated perfectly are also hard to detect because of the fuzzy, clouded shadows owing to low grayscale values. A new fracture enhancement method combined with histogram equalization is proposed to solve these problems. With this method, the contrast between discontinuities and background in coherence images is increased, linear structures are highlighted by stepwise adjustment of the threshold of the coherence image, and fractures are detected at different scales. Application of the method shows that it can also improve fracture cognition and accuracy.展开更多
基金support from MEC,Spain,through Grant No.CGL2007-64387/CLIthe AECID,Spain,for support through projects A/013666/07 and A/018685/08
文摘Rank Histograms are suitable tools to assess the quality of ensembles within an ensemble prediction system or framework. By counting the rank of a given variable in the ensemble, we are basically making a sample analysis, which does not allow us to distinguish if the origin of its variability is external noise or comes from chaotic sources. The recently introduced Mean to Variance Logarithmic (MVL) Diagram accounts for the spatial variability, being very sensitive to the spatial localization produced by infinitesimal perturbations of spatiotemporal chaotic systems. By using as a benchmark a simple model subject to noise, we show the distinct information given by Rank Histograms and MVL Diagrams. Hence, the main effects of the external noise can be visualized in a graphic. From the MVL diagram we clearly observe a reduction of the amplitude growth rate and of the spatial localization (chaos suppression), while from the Rank Histogram we observe changes in the reliability of the ensemble. We conclude that in a complex framework including spatiotemporal chaos and noise, both provide a more complete forecasting picture.
基金This work is supported in part by the National Natural Science Foundation of China under grant numbers 61672294,61502242,61702276,U1536206,U1405254,61772283,61602253,61601236 and 61572258,in part by Six peak talent project of Jiangsu Province(R2016L13),in part by National Key R&D Program of China under grant 2018YFB1003205,in part by NRF-2016R1D1A1B03933294,in part by the Jiangsu Basic Research Programs-Natural Science Foundation under grant numbers BK20150925 and BK20151530,in part by the Priority Academic Program Development of Jiangsu Higher Education Institutions(PAPD)fund,in part by the Collaborative Innovation Center of Atmospheric Environment and Equipment Technology(CICAEET)fund,China.Zhihua Xia is supported by BK21+program from the Ministry of Education of Korea.
文摘Content based image retrieval(CBIR)techniques have been widely deployed in many applications for seeking the abundant information existed in images.Due to large amounts of storage and computational requirements of CBIR,outsourcing image search work to the cloud provider becomes a very attractive option for many owners with small devices.However,owing to the private content contained in images,directly outsourcing retrieval work to the cloud provider apparently bring about privacy problem,so the images should be protected carefully before outsourcing.This paper presents a secure retrieval scheme for the encrypted images in the YUV color space.With this scheme,the discrete cosine transform(DCT)is performed on the Y component.The resulting DC coefficients are encrypted with stream cipher technology and the resulting AC coefficients as well as other two color components are encrypted with value permutation and position scrambling.Then the image owner transmits the encrypted images to the cloud server.When receiving a query trapdoor form on query user,the server extracts AC-coefficients histogram from the encrypted Y component and extracts two color histograms from the other two color components.The similarity between query trapdoor and database image is measured by calculating the Manhattan distance of their respective histograms.Finally,the encrypted images closest to the query image are returned to the query user.
文摘Image segmentation denotes a process for partitioning an image into distinct regions, it plays an important role in interpretation and decision making. A large variety of segmentation methods has been developed;among them, multidimensional histogram methods have been investigated but their implementation stays difficult due to the big size of histograms. We present an original method for segmenting n-D (where n is the number of components in image) images or multidimensional images in an unsupervised way using a fuzzy neighbourhood model. It is based on the hierarchical analysis of full n-D compact histograms integrating a fuzzy connected components labelling algorithm that we have realized in this work. Each peak of the histo- gram constitutes a class kernel, as soon as it encloses a number of pixels greater than or equal to a secondary arbitrary threshold knowing that a first threshold was set to define the degree of binary fuzzy similarity be- tween pixels. The use of a lossless compact n-D histogram allows a drastic reduction of the memory space necessary for coding it. As a consequence, the segmentation can be achieved without reducing the colors population of images in the classification step. It is shown that using n-D compact histograms, instead of 1-D and 2-D ones, leads to better segmentation results. Various images were segmented;the evaluation of the quality of segmentation in supervised and unsupervised of segmentation method proposed compare to the classification method k-means gives better results. It thus highlights the relevance of our approach, which can be used for solving many problems of segmentation.
文摘A new gray-spatial histogram is proposed, which incorporates spatial informatio n with gray compositions without sacrificing the robustness of traditional gray histograms. The purpose is to consider the representation role of gray compositi ons and spatial information simultaneously. Each entry in the gray-spatial hist ogram is the gray frequency and corresponding position information of images. In the experiments of sonar image recognition, the results show that the gray-spa tial histogram is effective in practical use.
基金Supported by the National Natural Science Foundation of China(60772066)
文摘A new algorithm taking the spatial context of local features into account by utilizing contextualized histograms was proposed to recognize facial expression. The contextualized histograms were extracted fromtwo widely used descriptors—the local binary pattern( LBP) and weber local descriptor( WLD). The LBP and WLD feature histograms were extracted separately fromeach facial image,and contextualized histogram was generated as feature vectors to feed the classifier. In addition,the human face was divided into sub-blocks and each sub-block was assigned different weights by their different contributions to the intensity of facial expressions to improve the recognition rate. With the support vector machine(SVM) as classifier,the experimental results on the 2D texture images fromthe 3D-BU FE dataset indicated that contextualized histograms improved facial expression recognition performance when local features were employed.
基金funded by the Deanship of Graduate Studies and Scientific Research at Jouf University under grant No.(DGSSR-2025-02-01295).
文摘Alzheimer’s Disease(AD)is a progressive neurodegenerative disorder that significantly affects cognitive function,making early and accurate diagnosis essential.Traditional Deep Learning(DL)-based approaches often struggle with low-contrast MRI images,class imbalance,and suboptimal feature extraction.This paper develops a Hybrid DL system that unites MobileNetV2 with adaptive classification methods to boost Alzheimer’s diagnosis by processing MRI scans.Image enhancement is done using Contrast-Limited Adaptive Histogram Equalization(CLAHE)and Enhanced Super-Resolution Generative Adversarial Networks(ESRGAN).A classification robustness enhancement system integrates class weighting techniques and a Matthews Correlation Coefficient(MCC)-based evaluation method into the design.The trained and validated model gives a 98.88%accuracy rate and 0.9614 MCC score.We also performed a 10-fold cross-validation experiment with an average accuracy of 96.52%(±1.51),a loss of 0.1671,and an MCC score of 0.9429 across folds.The proposed framework outperforms the state-of-the-art models with a 98%weighted F1-score while decreasing misdiagnosis results for every AD stage.The model demonstrates apparent separation abilities between AD progression stages according to the results of the confusion matrix analysis.These results validate the effectiveness of hybrid DL models with adaptive preprocessing for early and reliable Alzheimer’s diagnosis,contributing to improved computer-aided diagnosis(CAD)systems in clinical practice.
文摘When am I ever going to use this?ANIMALS The cheetah is the fastest known and animal.The table shows its fastest speed and the top speeds of four other land animals.1.What are the fastest and slowest speeds recorded in the table?2.How can you create a visual representation to summarize the data?3.Do any of these representations show both the animal name and its speed?Abar graphis one method of comparing data by using solid bars to represent quantities.
基金Supported by the National Natural Science Foundation of China (No. 70371018, 70572074)
文摘Histogram and kernel estimators are usually regarded as the two main classical data-based nonparametric tools to estimate the underlying density functions for some given data sets. In this paper we will integrate them and define a histogram-kernel error based on the integrated square error between histogram and binned kernel density estimator, and then exploit its asymptotic properties. 3ust as indicated in this paper, the histogram-kernel error only depends on the choice of bin width and the data for the given prior kernel densities. The asymptotic optimal bin width is derived by minimizing the mean histogram-kernel error. By comparing with Scott's optimal bin width formula for a histogram, a new method is proposed to construct the data-based histogram without knowledge of the underlying density function. Monte Carlo study is used to verify the usefulness of our method for different kinds of density functions and sample sizes.
文摘Gabor features have been shown to be effective for palm vein recognition. This paper presents a novel feature representation method, implementing the fusion of local Gabor histograms (FLGH), in order to improve the accuracy of palm vein recognition systems. A new local descriptor called local Gabor principal differences patterns (LGPDP) encodes the Gabor magnitude using the local maximum difference (LMD) operator. The corresponding Gabor phase patterns are encoded by local Gabor exclusive OR (XOR) patterns (LGXP). Fisher's linear discriminant (FLD) method is then implemented to reduce the dimensionality of the feature representation. Low-dimensional Gabor magnitude and phase feature vectors are finally fused to enhance accuracy. Experimental results from Institute of Automation, Chinese Academy of sciences (CASIA) database show that the proposed FLGH method achieves better performance by utilizing score-level fusion. The equal error rate (EER) is 0.08%, which outperforms other conventional palm vein recognition methods (EER range from 2.87% to 0.16%), e.g., the Laplacian palm, minutiae feature, Hessian phase, Eigenvein, local invariant features, mutual foreground local binary patterns (LBP), and multi-sampling feature fusion methods.
基金National Key Research and Development Program of China(2022YFB2804401)。
文摘An in-pixel histogramming time-to-digital converter(hTDC)based on octonary search and 4-tap phase detection is presented,aiming to improve frame rate while ensuring high precicion.The proposed hTDC is a 12-bit two-step converter consisting of a 6-bit coarse quantization and a 6-bit fine quantization,which supports a time resolution of 120 ps and multiphoton counting up to 2 GHz without a GHz reference frequency.The proposed hTDC is designed in 0.11μm CMOS process with an area consumption of 6900μm^(2).The data from a behavioral-level model is imported into the designed hTDC circuit for simulation verification.The post-simulation results show that the proposed hTDC achieves 0.8%depth precision in 9 m range for short-range system design specifications and 0.2%depth precision in 48 m range for long-range system design specifications.Under 30×10^(3) lux background light conditions,the proposed hTDC can be used for SPAD-based flash LiDAR sensor to achieve a frame rate to 40 fps with 200 ps resolution in 9 m range.
基金National Natural Science Foundation of China(62071147)。
文摘Ground penetrating radar(GPR),as a fast,efficient,and non-destructive detection device,holds great potential for the detection of shallow subsurface environments,such as urban road subsurface monitoring.However,the interpretation of GPR echo images often relies on manual recognition by experienced engineers.In order to address the automatic interpretation of cavity targets in GPR echo images,a recognition-algorithm based on Gaussian mixed model-hidden Markov model(GMM-HMM)is proposed,which can recognize three dimensional(3D)underground voids automatically.First,energy detection on the echo images is performed,whereby the data is preprocessed and pre-filtered.Then,edge histogram descriptor(EHD),histogram of oriented gradient(HOG),and Log-Gabor filters are used to extract features from the images.The traditional method can only be applied to 2D images and pre-processing is required for C-scan images.Finally,the aggregated features are fed into the GMM-HMM for classification and compared with two other methods,long short-term memory(LSTM)and gate recurrent unit(GRU).By testing on a simulated dataset,an accuracy rate of 90%is obtained,demonstrating the effectiveness and efficiency of our proposed method.
文摘Skin cancer,a severe health threat,can spread rapidly if undetected.Therefore,early detection can lead to an advanced and efficient diagnosis,thus reducing mortality.Unsupervised classification techniques analyse extensive skin image datasets,identifying patterns and anomalies without prior labelling,facilitating early detection and effective diagnosis and potentially saving lives.In this study,the authors aim to explore the potential of unsupervised learning methods in classifying different types of skin lesions in dermatoscopic images.The authors aim to bridge the gap in dermatological research by introducing innovative techniques that enhance image quality and improve feature extraction.To achieve this,enhanced super-resolution generative adversarial networks(ESRGAN)was fine-tuned to strengthen the resolution of skin lesion images,making critical features more visible.The authors extracted histogram features to capture essential colour characteristics and used the Davies-Bouldin index and silhouette score to determine optimal clusters.Fine-tuned k-means clustering with Euclidean distance in the histogram feature space achieved 87.77% and 90.5% test accuracies on the ISIC2019 and HAM10000 datasets,respectively.The unsupervised approach effectively categorises skin lesions,indicating that unsupervised learning can significantly advance dermatology by enabling early detection and classification without extensive manual annotation.
文摘Background:Pneumonia remains a critical global health challenge,manifesting as a severe respiratory infection caused by viruses,bacteria,and fungi.Early detection is paramount for effective treatment,potentially reducing mortality rates and optimizing healthcare resource allocation.Despite the importance of chest X-ray diagnosis,image analysis presents significant challenges,particularly in regions with limited medical expertise.This study addresses these challenges by proposing a computer-aided diagnosis system leveraging targeted image preprocessing and optimized deep learning techniques.Methods:We systematically evaluated contrast limited adaptive histogram equalization with varying clip limits for preprocessing chest X-ray images,demonstrating its effectiveness in enhancing feature visibility for diagnostic accuracy.Employing a comprehensive dataset of 5,863 X-ray images(1,583 pneumonia-negative,4,280 pneumonia-positive)collected from multiple healthcare facilities,we conducted a comparative analysis of transfer learning with pre-trained models including ResNet50v2,VGG-19,and MobileNetV2.Statistical validation was performed through 5-fold cross-validation.Results:Our results show that the contrast limited adaptive histogram equalization-enhanced approach with ResNet50v2 achieves 93.40%accuracy,outperforming VGG-19(84.90%)and MobileNetV2(89.70%).Statistical validation confirms the significance of these improvements(P<0.01).The development and optimization resulted in a lightweight mobile application(74 KB)providing rapid diagnostic support(1-2 s response time).Conclusion:The proposed approach demonstrates practical applicability in resource-constrained settings,balancing diagnostic accuracy with deployment efficiency,and offers a viable solution for computer-aided pneumonia diagnosis in areas with limited medical expertise.
文摘Esophageal cancer is one of the most difficult cancers to treat since it is often at an advanced stage at the time of symptom presentation.For locally advanced esophageal cancer,treatment options include multidisciplinary treatment such as surgery or definitive chemoradiotherapy.Surgery has a high local control rate because it involves excision of the cancer along with the surrounding organs;however,it is still highly invasive,although advances in surgery have reduced the burden on patients.On the other hand,chemoradiotherapy may also be applicable in cases in which surgery is inoperable owing to complications or distant lymph node metastasis.However,chemoradiotherapy using X-ray irradiation can cause late toxicities,including those to the heart.Proton beam therapy is widely used to treat esophageal cancer because of its characteristics,and some comparisons between proton beam therapy and X-ray therapy or surgery have recently been reported.This review discusses the role of proton beam therapy in esophageal cancer in comparison to X-ray therapy and surgery.
基金funded by the Directorate of Research,Technology,and Community Service,Ministry of Higher Education,Science,and Technology of the Republic of Indonesia the Regular Fundamental Research scheme,with grant numbers 001/LL6/PL/AL.04/2025,011/SPK-PFR/RIK/05/2025.
文摘Early detection of Forest and Land Fires(FLF)is essential to prevent the rapid spread of fire as well as minimize environmental damage.However,accurate detection under real-world conditions,such as low light,haze,and complex backgrounds,remains a challenge for computer vision systems.This study evaluates the impact of three image enhancement techniques—Histogram Equalization(HE),Contrast Limited Adaptive Histogram Equalization(CLAHE),and a hybrid method called DBST-LCM CLAHE—on the performance of the YOLOv11 object detection model in identifying fires and smoke.The D-Fire dataset,consisting of 21,527 annotated images captured under diverse environmental scenarios and illumination levels,was used to train and evaluate the model.Each enhancement method was applied to the dataset before training.Model performance was assessed using multiple metrics,including Precision,Recall,mean Average Precision at 50%IoU(mAP50),F1-score,and visual inspection through bounding box results.Experimental results show that all three enhancement techniques improved detection performance.HE yielded the highest mAP50 score of 0.771,along with a balanced precision of 0.784 and recall of 0.703,demonstrating strong generalization across different conditions.DBST-LCM CLAHE achieved the highest Precision score of 79%,effectively reducing false positives,particularly in scenes with dispersed smoke or complex textures.CLAHE,with slightly lower overall metrics,contributed to improved local feature detection.Each technique showed distinct advantages:HE enhanced global contrast;CLAHE improved local structure visibility;and DBST-LCM CLAHE provided an optimal balance through dynamic block sizing and local contrast preservation.These results underline the importance of selecting preprocessing methods according to detection priorities,such as minimizing false alarms or maximizing completeness.This research does not propose a new model architecture but rather benchmarks a recent lightweight detector,YOLOv11,combined with image enhancement strategies for practical deployment in FLF monitoring.The findings support the integration of preprocessing techniques to improve detection accuracy,offering a foundation for real-time FLF detection systems on edge devices or drones,particularly in regions like Indonesia.
文摘AIM:To find the effective contrast enhancement method on retinal images for effective segmentation of retinal features.METHODS:A novel image preprocessing method that used neighbourhood-based improved contrast limited adaptive histogram equalization(NICLAHE)to improve retinal image contrast was suggested to aid in the accurate identification of retinal disorders and improve the visibility of fine retinal structures.Additionally,a minimal-order filter was applied to effectively denoise the images without compromising important retinal structures.The novel NICLAHE algorithm was inspired by the classical CLAHE algorithm,but enhanced it by selecting the clip limits and tile sized in a dynamical manner relative to the pixel values in an image as opposed to using fixed values.It was evaluated on the Drive and high-resolution fundus(HRF)datasets on conventional quality measures.RESULTS:The new proposed preprocessing technique was applied to two retinal image databases,Drive and HRF,with four quality metrics being,root mean square error(RMSE),peak signal to noise ratio(PSNR),root mean square contrast(RMSC),and overall contrast.The technique performed superiorly on both the data sets as compared to the traditional enhancement methods.In order to assess the compatibility of the method with automated diagnosis,a deep learning framework named ResNet was applied in the segmentation of retinal blood vessels.Sensitivity,specificity,precision and accuracy were used to analyse the performance.NICLAHE–enhanced images outperformed the traditional techniques on both the datasets with improved accuracy.CONCLUSION:NICLAHE provides better results than traditional methods with less error and improved contrastrelated values.These enhanced images are subsequently measured by sensitivity,specificity,precision,and accuracy,which yield a better result in both datasets.
基金supported in part by the National Natural Science Foundation of China under Grant(62171257,U22B2001,U19A2052,62020106011,62061015)in part by the Natural Science Foundation of Chongqing under Grant(2023NSCQMSX2930)+1 种基金in part by the Youth Innovation Group Support Program of ICE Discipline of CQUPT under Grant(SCIE-QN-2022-05)in part by the Graduate Scientifc Research and Innovation Project of Chongqing under Grant(CYS22469).
文摘Perceptual quality assessment for point cloud is critical for immersive metaverse experience and is a challenging task.Firstly,because point cloud is formed by unstructured 3D points that makes the topology more complex.Secondly,the quality impairment generally involves both geometric attributes and color properties,where the measurement of the geometric distortion becomes more complex.We propose a perceptual point cloud quality assessment model that follows the perceptual features of Human Visual System(HVS)and the intrinsic characteristics of the point cloud.The point cloud is first pre-processed to extract the geometric skeleton keypoints with graph filtering-based re-sampling,and local neighboring regions around the geometric skeleton keypoints are constructed by K-Nearest Neighbors(KNN)clustering.For geometric distortion,the Point Feature Histogram(PFH)is extracted as the feature descriptor,and the Earth Mover’s Distance(EMD)between the PFHs of the corresponding local neighboring regions in the reference and the distorted point clouds is calculated as the geometric quality measurement.For color distortion,the statistical moments between the corresponding local neighboring regions are computed as the color quality measurement.Finally,the global perceptual quality assessment model is obtained as the linear weighting aggregation of the geometric and color quality measurement.The experimental results on extensive datasets show that the proposed method achieves the leading performance as compared to the state-of-the-art methods with less computing time.Meanwhile,the experimental results also demonstrate the robustness of the proposed method across various distortion types.The source codes are available at https://github.com/llsurreal919/Point Cloud Quality Assessment.
基金the National Key Research and Development Program of China(No.2022YFC3601400)。
文摘Frequency-modulated continuous-wave radar enables the non-contact and privacy-preserving recognition of human behavior.However,the accuracy of behavior recognition is directly influenced by the spatial relationship between human posture and the radar.To address the issue of low accuracy in behavior recognition when the human body is not directly facing the radar,a method combining local outlier factor with Doppler information is proposed for the correction of multi-classifier recognition results.Initially,the information such as distance,velocity,and micro-Doppler spectrogram of the target is obtained using the fast Fourier transform and histogram of oriented gradients-support vector machine methods,followed by preliminary recognition.Subsequently,Platt scaling is employed to transform recognition results into confidence scores,and finally,the Doppler-local outlier factor method is utilized to calibrate the confidence scores,with the highest confidence classifier result considered as the recognition outcome.Experimental results demonstrate that this approach achieves an average recognition accuracy of 96.23%for comprehensive human behavior recognition in various orientations.
基金supported by the National Key R&D Program of China(No.2022YFB3205101)NSAF(No.U2230116)。
文摘To improve image quality under low illumination conditions,a novel low-light image enhancement method is proposed in this paper based on multi-illumination estimation and multi-scale fusion(MIMS).Firstly,the illumination is processed by contrast-limited adaptive histogram equalization(CLAHE),adaptive complementary gamma function(ACG),and adaptive detail preserving S-curve(ADPS),respectively,to obtain three components.Then,the fusion-relevant features,exposure,and color contrast are selected as the weight maps.Subsequently,these components and weight maps are fused through multi-scale to generate enhanced illumination.Finally,the enhanced images are obtained by multiplying the enhanced illumination and reflectance.Compared with existing approaches,this proposed method achieves an average increase of 0.81%and 2.89%in the structural similarity index measurement(SSIM)and peak signal-to-noise ratio(PSNR),and a decrease of 6.17%and 32.61%in the natural image quality evaluator(NIQE)and gradient magnitude similarity deviation(GMSD),respectively.
基金sponsored by the National Science&Technology Major Special Project(Grant No.2011ZX05025-001-04)
文摘Eigenstructure-based coherence attributes are efficient and mature techniques for large-scale fracture detection. However, in horizontally bedded and continuous strata, buried fractures in high grayscale value zones are difficult to detect. Furthermore, middleand small-scale fractures in fractured zones where migration image energies are usually not concentrated perfectly are also hard to detect because of the fuzzy, clouded shadows owing to low grayscale values. A new fracture enhancement method combined with histogram equalization is proposed to solve these problems. With this method, the contrast between discontinuities and background in coherence images is increased, linear structures are highlighted by stepwise adjustment of the threshold of the coherence image, and fractures are detected at different scales. Application of the method shows that it can also improve fracture cognition and accuracy.