In healthcare sector,image classification is one of the crucial problems that impact the quality output from image processing domain.The purpose of image classification is to categorize different healthcare images under...In healthcare sector,image classification is one of the crucial problems that impact the quality output from image processing domain.The purpose of image classification is to categorize different healthcare images under various class labels which in turn helps in the detection and management of diseases.Magnetic Resonance Imaging(MRI)is one of the effective non-invasive strate-gies that generate a huge and distinct number of tissue contrasts in every imaging modality.This technique is commonly utilized by healthcare professionals for Brain Tumor(BT)diagnosis.With recent advancements in Machine Learning(ML)and Deep Learning(DL)models,it is possible to detect the tumor from images automatically,using a computer-aided design.The current study focuses on the design of automated Deep Learning-based BT Detection and Classification model using MRI images(DLBTDC-MRI).The proposed DLBTDC-MRI techni-que aims at detecting and classifying different stages of BT.The proposed DLBTDC-MRI technique involves medianfiltering technique to remove the noise and enhance the quality of MRI images.Besides,morphological operations-based image segmentation approach is also applied to determine the BT-affected regions in brain MRI image.Moreover,a fusion of handcrafted deep features using VGGNet is utilized to derive a valuable set of feature vectors.Finally,Artificial Fish Swarm Optimization(AFSO)with Artificial Neural Network(ANN)model is utilized as a classifier to decide the presence of BT.In order to assess the enhanced BT classification performance of the proposed model,a comprehensive set of simulations was performed on benchmark dataset and the results were vali-dated under several measures.展开更多
Brain cancer detection and classification is done utilizing distinct medical imaging modalities like computed tomography(CT),or magnetic resonance imaging(MRI).An automated brain cancer classification using computer a...Brain cancer detection and classification is done utilizing distinct medical imaging modalities like computed tomography(CT),or magnetic resonance imaging(MRI).An automated brain cancer classification using computer aided diagnosis(CAD)models can be designed to assist radiologists.With the recent advancement in computer vision(CV)and deep learning(DL)models,it is possible to automatically detect the tumor from images using a computer-aided design.This study focuses on the design of automated Henry Gas Solubility Optimization with Fusion of Handcrafted and Deep Features(HGSO-FHDF)technique for brain cancer classification.The proposed HGSO-FHDF technique aims for detecting and classifying different stages of brain tumors.The proposed HGSO-FHDF technique involves Gabor filtering(GF)technique for removing the noise and enhancing the quality of MRI images.In addition,Tsallis entropy based image segmentation approach is applied to determine injured brain regions in the MRI image.Moreover,a fusion of handcrafted with deep features using Residual Network(ResNet)is utilized as feature extractors.Finally,HGSO algorithm with kernel extreme learning machine(KELM)model was utilized for identifying the presence of brain tumors.For examining the enhanced brain tumor classification performance,a comprehensive set of simulations take place on the BRATS 2015 dataset.展开更多
Histopathology is the investigation of tissues to identify the symptom of abnormality.The histopathological procedure comprises gathering samples of cells/tissues,setting them on the microscopic slides,and staining th...Histopathology is the investigation of tissues to identify the symptom of abnormality.The histopathological procedure comprises gathering samples of cells/tissues,setting them on the microscopic slides,and staining them.The investigation of the histopathological image is a problematic and laborious process that necessitates the expert’s knowledge.At the same time,deep learning(DL)techniques are able to derive features,extract data,and learn advanced abstract data representation.With this view,this paper presents an ensemble of handcrafted with deep learning enabled histopathological image classification(EHCDL-HIC)model.The proposed EHCDLHIC technique initially performs Weiner filtering based noise removal technique.Once the images get smoothened,an ensemble of deep features and local binary pattern(LBP)features are extracted.For the classification process,the bidirectional gated recurrent unit(BGRU)model can be employed.At the final stage,the bacterial foraging optimization(BFO)algorithm is utilized for optimal hyperparameter tuning process which leads to improved classification performance,shows the novelty of the work.For validating the enhanced execution of the proposed EHCDL-HIC method,a set of simulations is performed.The experimentation outcomes highlighted the betterment of the EHCDL-HIC approach over the existing techniques with maximum accuracy of 94.78%.Therefore,the EHCDL-HIC model can be applied as an effective approach for histopathological image classification.展开更多
On a global scale,lung cancer is responsible for around 27%of all cancer fatalities.Even though there have been great strides in diagnosis and therapy in recent years,the five-year cure rate is just 19%.Classification...On a global scale,lung cancer is responsible for around 27%of all cancer fatalities.Even though there have been great strides in diagnosis and therapy in recent years,the five-year cure rate is just 19%.Classification is crucial for diagnosing lung nodules.This is especially true today that automated categorization may provide a professional opinion that can be used by doctors.New computer vision and machine learning techniques have made possible accurate and quick categorization of CT images.This field of research has exploded in popularity in recent years because of its high efficiency and ability to decrease labour requirements.Here,they want to look carefully at the current state of automated categorization of lung nodules.Generalpurpose structures are briefly discussed,and typical algorithms are described.Our results show deep learning-based lung nodule categorization quickly becomes the industry standard.Therefore,it is critical to pay greater attention to the coherence of the data inside the study and the consistency of the research topic.Furthermore,there should be greater collaboration between designers,medical experts,and others in the field.展开更多
I've always been interested in traditional crafts,so when my school organized a workshop on wood carving,an ancient Chinese art form,I signed up right away.The workshop was held in a small studio in the old town.A...I've always been interested in traditional crafts,so when my school organized a workshop on wood carving,an ancient Chinese art form,I signed up right away.The workshop was held in a small studio in the old town.As I walked in,I was greeted by the smell of wood and the sound made by carving wood.The teacher,Mr Zhang,was a master carver with over 30 years of experience.He showed us some of his works-beautiful sculptures of animals and flowers.I was amazed by the skill and patience it must have taken to create them.展开更多
Recognizing road scene context from a single image remains a critical challenge for intelligent autonomous driving systems,particularly in dynamic and unstructured environments.While recent advancements in deep learni...Recognizing road scene context from a single image remains a critical challenge for intelligent autonomous driving systems,particularly in dynamic and unstructured environments.While recent advancements in deep learning have significantly enhanced road scene classification,simultaneously achieving high accuracy,computational efficiency,and adaptability across diverse conditions continues to be difficult.To address these challenges,this study proposes HybridLSTM,a novel and efficient framework that integrates deep learning-based,object-based,and handcrafted feature extraction methods within a unified architecture.HybridLSTM is designed to classify four distinct road scene categories—crosswalk(CW),highway(HW),overpass/tunnel(OP/T),and parking(P)—by leveraging multiple publicly available datasets,including Places-365,BDD100K,LabelMe,and KITTI,thereby promoting domain generalization.The framework fuses object-level features extracted using YOLOv5 and VGG19,scene-level global representations obtained from a modified VGG19,and fine-grained texture features captured through eight handcrafted descriptors.This hybrid feature fusion enables the model to capture both semantic context and low-level visual cues,which are critical for robust scene understanding.To model spatial arrangements and latent sequential dependencies present even in static imagery,the combined features are processed through a Long Short-Term Memory(LSTM)network,allowing the extraction of discriminative patterns across heterogeneous feature spaces.Extensive experiments conducted on 2725 annotated road scene images,with an 80:20 training-to-testing split,validate the effectiveness of the proposed model.HybridLSTM achieves a classification accuracy of 96.3%,a precision of 95.8%,a recall of 96.1%,and an F1-score of 96.0%,outperforming several existing state-of-the-art methods.These results demonstrate the robustness,scalability,and generalization capability of HybridLSTM across varying environments and scene complexities.Moreover,the framework is optimized to balance classification performance with computational efficiency,making it highly suitable for real-time deployment in embedded autonomous driving systems.Future work will focus on extending the model to multi-class detection within a single frame and optimizing it further for edge-device deployments to reduce computational overhead in practical applications.展开更多
A design of super-orthogonal space-time trellis codes (SOSTTCs) based on the trace criterion (TC) is proposed for improving the design of SOSTTCs. The shortcomings of the rank and determinant criteria based design...A design of super-orthogonal space-time trellis codes (SOSTTCs) based on the trace criterion (TC) is proposed for improving the design of SOSTTCs. The shortcomings of the rank and determinant criteria based design and the advantages of the TC-based design are analyzed. The optimization principle of four factors is presented, which includes the space-time block coding (STBC) scheme, set partitioning, trellis structure, and the assignment of signal subsets and STBC schemes in the trellis. According to this principle, systematical and handcrafted design steps are given in detail. By constellation expansion, the code performance can be further improved. The code design results are given, and the new codes outperform others in the simulation.展开更多
Spectral and spatial features in remotely sensed data play an irreplaceable role in classifying crop types for precision agriculture. Despite the thriving establishment of the handcrafted features, designing or select...Spectral and spatial features in remotely sensed data play an irreplaceable role in classifying crop types for precision agriculture. Despite the thriving establishment of the handcrafted features, designing or selecting such features valid for specific crop types requires prior knowledge and thus remains an open challenge. Convolutional neural networks(CNNs) can effectively overcome this issue with their advanced ability to generate high-level features automatically but are still inadequate in mining spectral features compared to mining spatial features. This study proposed an enhanced spectral feature called Stacked Spectral Feature Space Patch(SSFSP) for CNN-based crop classification. SSFSP is a stack of twodimensional(2 D) gridded spectral feature images that record various crop types’ spatial and intensity distribution characteristics in a 2 D feature space consisting of two spectral bands. SSFSP can be input into2 D-CNNs to support the simultaneous mining of spectral and spatial features, as the spectral features are successfully converted to 2 D images that can be processed by CNN. We tested the performance of SSFSP by using it as the input to seven CNN models and one multilayer perceptron model for crop type classification compared to using conventional spectral features as input. Using high spatial resolution hyperspectral datasets at three sites, the comparative study demonstrated that SSFSP outperforms conventional spectral features regarding classification accuracy, robustness, and training efficiency. The theoretical analysis summarizes three reasons for its excellent performance. First, SSFSP mines the spectral interrelationship with feature generality, which reduces the required number of training samples.Second, the intra-class variance can be largely reduced by grid partitioning. Third, SSFSP is a highly sparse feature, which reduces the dependence on the CNN model structure and enables early and fast convergence in model training. In conclusion, SSFSP has great potential for practical crop classification in precision agriculture.展开更多
Automatic gastrointestinal(GI)tract disease recognition is an important application of biomedical image processing.Conventionally,microscopic analysis of pathological tissue is used to detect abnormal areas of the GI ...Automatic gastrointestinal(GI)tract disease recognition is an important application of biomedical image processing.Conventionally,microscopic analysis of pathological tissue is used to detect abnormal areas of the GI tract.The procedure is subjective and results in significant inter-/intraobserver variations in disease detection.Moreover,a huge frame rate in video endoscopy is an overhead for the pathological findings of gastroenterologists to observe every frame with a detailed examination.Consequently,there is a huge demand for a reliable computer-aided diagnostic system(CADx)for diagnosing GI tract diseases.In this work,a CADx was proposed for the diagnosis and classification of GI tract diseases.A novel framework is presented where preprocessing(LAB color space)is performed first;then local binary patterns(LBP)or texture and deep learning(inceptionNet,ResNet50,and VGG-16)features are fused serially to improve the prediction of the abnormalities in the GI tract.Additionally,principal component analysis(PCA),entropy,and minimum redundancy and maximum relevance(mRMR)feature selection methods were analyzed to acquire the optimized characteristics,and various classifiers were trained using the fused features.Open-source color image datasets(KVASIR,NERTHUS,and stomach ULCER)were used for performance evaluation.The study revealed that the subspace discriminant classifier provided an efficient result with 95.02%accuracy on the KVASIR dataset,which proved to be better than the existing state-of-the-art approaches.展开更多
Gold clay is a crafting medium consisting of gold particles mixed with an organic binder and water for making jewelry or decoration.The clay can be shaped by hand,textured,carved,formed or using molds.After drying and...Gold clay is a crafting medium consisting of gold particles mixed with an organic binder and water for making jewelry or decoration.The clay can be shaped by hand,textured,carved,formed or using molds.After drying and burning,the organic binder and water were decomposed and the gold particles were transformed to its final metal state.Although,gold clay is very expensive,it is useful to decorate the silver clay designed jewelry or small sculptures.In this research,nano-microplate gold and specific organic binder was used for producing nano-microplate gold clay.The objectives of this research are to study binder's type and ratios for optimum producing gold clay,and to study the heating condition for making silver and gold clay jewelry.The result showed that the clay can be fired with heating temperature at 900°C for an hour by electric kiln.The physical properties of the gold clay at different heating temperatures were determined.Furthermore,prototype of jewelry using the clay was展开更多
Digital low-density parity-check(LDPC) decoders can hardly meet the power-limits brought by the new application scenarios. The analog LDPC decoder, which is an application of the analog computation technology, is cons...Digital low-density parity-check(LDPC) decoders can hardly meet the power-limits brought by the new application scenarios. The analog LDPC decoder, which is an application of the analog computation technology, is considered to have the potential to address this issue to some extent. However, due to the lack of automation tools and analog stopping criteria, the analog LDPC decoders suffer from costly handcraft design and additional decoding delay, and are not feasible to practical applications. To address these issues, a decoder architecture using reusable building blocks is designed to lower the handcraft design, and a probability stopping criterion that is specially designed for analog decoder is further planned and implemented to reduce the decoding delay. Then, a(480,240) CMOS analog LDPC decoder is designed and fabricated in a 0.35-μm CMOS technology. Experimental results show that the decoder prototype can achieve 50 Mbps throughput when the power consumption is about 86.3m W, and the decoding delay can be reduced by at most 93% compared with using the preset maximum decoding delay in existing works.展开更多
Competition in modern society requires teachers to cultivate creative and pioneering talents, which indicates that teachers are required to cultivate students' creative ability. In the primary school stage, art ha...Competition in modern society requires teachers to cultivate creative and pioneering talents, which indicates that teachers are required to cultivate students' creative ability. In the primary school stage, art handicraft is to cultivate students' practical ability and interest, while in the primary school stage, and art handicraft teaching mostly belongs to the teaching of cultivating students' creative ability. This article will discuss the art handicraft teaching and the primary school students' creative ability cultivation countermeasure.展开更多
Medical imaging is crucial in modern clinics to provide guidance to the diagnosis and treatment of diseases.Medical image reconstruction is one of the most fundamental and important components of medical imaging,whose...Medical imaging is crucial in modern clinics to provide guidance to the diagnosis and treatment of diseases.Medical image reconstruction is one of the most fundamental and important components of medical imaging,whose major objective is to acquire high-quality medical images for clinical usage at the minimal cost and risk to the patients.Mathematical models in medical image reconstruction or,more generally,image restoration in computer vision have been playing a prominent role.Earlier mathematical models are mostly designed by human knowledge or hypothesis on the image to be reconstructed,and we shall call these models handcrafted models.Later,handcrafted plus data-driven modeling started to emerge which still mostly relies on human designs,while part of the model is learned from the observed data.More recently,as more data and computation resources are made available,deep learning based models(or deep models)pushed the data-driven modeling to the extreme where the models are mostly based on learning with minimal human designs.Both handcrafted and data-driven modeling have their own advantages and disadvantages.Typical handcrafted models are well interpretable with solid theoretical supports on the robustness,recoverability,complexity,etc.,whereas they may not be flexible and sophisticated enough to fully leverage large data sets.Data-driven models,especially deep models,on the other hand,are generally much more flexible and effective in extracting useful information from large data sets,while they are currently still in lack of theoretical foundations.Therefore,one of the major research trends in medical imaging is to combine handcrafted modeling with deep modeling so that we can enjoy benefits from both approaches.The major part of this article is to provide a conceptual review of some recent works on deep modeling from the unrolling dynamics viewpoint.This viewpoint stimulates new designs of neural network architectures with inspirations from optimization algorithms and numerical differential equations.Given the popularity of deep modeling,there are still vast remaining challenges in the field,as well as opportunities which we shall discuss at the end of this article.展开更多
Three-dimensional skeleton-based action recognition(3D SAR)has gained important attention within the computer vision community,owing to the inherent advantages offered by skeleton data.As a result,a plethora of impres...Three-dimensional skeleton-based action recognition(3D SAR)has gained important attention within the computer vision community,owing to the inherent advantages offered by skeleton data.As a result,a plethora of impressive works,including those based on conventional handcrafted features and learned feature extraction methods,have been conducted over the years.However,prior surveys on action recognition have primarily focused on video or red-green-blue(RGB)data-dominated approaches,with limited coverage of reviews related to skeleton data.Furthermore,despite the extensive application of deep learning methods in this field,there has been a notable absence of research that provides an introductory or comprehensive review from the perspective of deep learning architectures.To address these limitations,this survey first underscores the importance of action recognition and emphasizes the significance of 3-dimensional(3D)skeleton data as a valuable modality.Subsequently,we provide a comprehensive introduction to mainstream action recognition techniques based on 4 fundamental deep architectures,i.e.,recurrent neural networks,convolutional neural networks,graph convolutional network,and Transformers.All methods with the corresponding architectures are then presented in a data-driven manner with detailed discussion.Finally,we offer insights into the current largest 3D skeleton dataset,NTU-RGB+D,and its new edition,NTU-RGB+D 120,along with an overview of several top-performing algorithms on these datasets.To the best of our knowledge,this research represents the first comprehensive discussion of deep learning-based action recognition using 3D skeleton data.展开更多
基金supported through the Annual Funding track by the Deanship of Scientific Research,Vice Presidency for Graduate Studies and Scientific Research,King Faisal University,Saudi Arabia[Project No.AN000684].
文摘In healthcare sector,image classification is one of the crucial problems that impact the quality output from image processing domain.The purpose of image classification is to categorize different healthcare images under various class labels which in turn helps in the detection and management of diseases.Magnetic Resonance Imaging(MRI)is one of the effective non-invasive strate-gies that generate a huge and distinct number of tissue contrasts in every imaging modality.This technique is commonly utilized by healthcare professionals for Brain Tumor(BT)diagnosis.With recent advancements in Machine Learning(ML)and Deep Learning(DL)models,it is possible to detect the tumor from images automatically,using a computer-aided design.The current study focuses on the design of automated Deep Learning-based BT Detection and Classification model using MRI images(DLBTDC-MRI).The proposed DLBTDC-MRI techni-que aims at detecting and classifying different stages of BT.The proposed DLBTDC-MRI technique involves medianfiltering technique to remove the noise and enhance the quality of MRI images.Besides,morphological operations-based image segmentation approach is also applied to determine the BT-affected regions in brain MRI image.Moreover,a fusion of handcrafted deep features using VGGNet is utilized to derive a valuable set of feature vectors.Finally,Artificial Fish Swarm Optimization(AFSO)with Artificial Neural Network(ANN)model is utilized as a classifier to decide the presence of BT.In order to assess the enhanced BT classification performance of the proposed model,a comprehensive set of simulations was performed on benchmark dataset and the results were vali-dated under several measures.
基金This research work was funded by Institutional fund projects under grant no.(IFPHI-180-612-2020)Therefore,the authors gratefully acknowledge technical and financial support from the Ministry of Education and King Abdulaziz University,DSR,Jeddah,Saudi Arabia.
文摘Brain cancer detection and classification is done utilizing distinct medical imaging modalities like computed tomography(CT),or magnetic resonance imaging(MRI).An automated brain cancer classification using computer aided diagnosis(CAD)models can be designed to assist radiologists.With the recent advancement in computer vision(CV)and deep learning(DL)models,it is possible to automatically detect the tumor from images using a computer-aided design.This study focuses on the design of automated Henry Gas Solubility Optimization with Fusion of Handcrafted and Deep Features(HGSO-FHDF)technique for brain cancer classification.The proposed HGSO-FHDF technique aims for detecting and classifying different stages of brain tumors.The proposed HGSO-FHDF technique involves Gabor filtering(GF)technique for removing the noise and enhancing the quality of MRI images.In addition,Tsallis entropy based image segmentation approach is applied to determine injured brain regions in the MRI image.Moreover,a fusion of handcrafted with deep features using Residual Network(ResNet)is utilized as feature extractors.Finally,HGSO algorithm with kernel extreme learning machine(KELM)model was utilized for identifying the presence of brain tumors.For examining the enhanced brain tumor classification performance,a comprehensive set of simulations take place on the BRATS 2015 dataset.
文摘Histopathology is the investigation of tissues to identify the symptom of abnormality.The histopathological procedure comprises gathering samples of cells/tissues,setting them on the microscopic slides,and staining them.The investigation of the histopathological image is a problematic and laborious process that necessitates the expert’s knowledge.At the same time,deep learning(DL)techniques are able to derive features,extract data,and learn advanced abstract data representation.With this view,this paper presents an ensemble of handcrafted with deep learning enabled histopathological image classification(EHCDL-HIC)model.The proposed EHCDLHIC technique initially performs Weiner filtering based noise removal technique.Once the images get smoothened,an ensemble of deep features and local binary pattern(LBP)features are extracted.For the classification process,the bidirectional gated recurrent unit(BGRU)model can be employed.At the final stage,the bacterial foraging optimization(BFO)algorithm is utilized for optimal hyperparameter tuning process which leads to improved classification performance,shows the novelty of the work.For validating the enhanced execution of the proposed EHCDL-HIC method,a set of simulations is performed.The experimentation outcomes highlighted the betterment of the EHCDL-HIC approach over the existing techniques with maximum accuracy of 94.78%.Therefore,the EHCDL-HIC model can be applied as an effective approach for histopathological image classification.
文摘On a global scale,lung cancer is responsible for around 27%of all cancer fatalities.Even though there have been great strides in diagnosis and therapy in recent years,the five-year cure rate is just 19%.Classification is crucial for diagnosing lung nodules.This is especially true today that automated categorization may provide a professional opinion that can be used by doctors.New computer vision and machine learning techniques have made possible accurate and quick categorization of CT images.This field of research has exploded in popularity in recent years because of its high efficiency and ability to decrease labour requirements.Here,they want to look carefully at the current state of automated categorization of lung nodules.Generalpurpose structures are briefly discussed,and typical algorithms are described.Our results show deep learning-based lung nodule categorization quickly becomes the industry standard.Therefore,it is critical to pay greater attention to the coherence of the data inside the study and the consistency of the research topic.Furthermore,there should be greater collaboration between designers,medical experts,and others in the field.
文摘I've always been interested in traditional crafts,so when my school organized a workshop on wood carving,an ancient Chinese art form,I signed up right away.The workshop was held in a small studio in the old town.As I walked in,I was greeted by the smell of wood and the sound made by carving wood.The teacher,Mr Zhang,was a master carver with over 30 years of experience.He showed us some of his works-beautiful sculptures of animals and flowers.I was amazed by the skill and patience it must have taken to create them.
文摘Recognizing road scene context from a single image remains a critical challenge for intelligent autonomous driving systems,particularly in dynamic and unstructured environments.While recent advancements in deep learning have significantly enhanced road scene classification,simultaneously achieving high accuracy,computational efficiency,and adaptability across diverse conditions continues to be difficult.To address these challenges,this study proposes HybridLSTM,a novel and efficient framework that integrates deep learning-based,object-based,and handcrafted feature extraction methods within a unified architecture.HybridLSTM is designed to classify four distinct road scene categories—crosswalk(CW),highway(HW),overpass/tunnel(OP/T),and parking(P)—by leveraging multiple publicly available datasets,including Places-365,BDD100K,LabelMe,and KITTI,thereby promoting domain generalization.The framework fuses object-level features extracted using YOLOv5 and VGG19,scene-level global representations obtained from a modified VGG19,and fine-grained texture features captured through eight handcrafted descriptors.This hybrid feature fusion enables the model to capture both semantic context and low-level visual cues,which are critical for robust scene understanding.To model spatial arrangements and latent sequential dependencies present even in static imagery,the combined features are processed through a Long Short-Term Memory(LSTM)network,allowing the extraction of discriminative patterns across heterogeneous feature spaces.Extensive experiments conducted on 2725 annotated road scene images,with an 80:20 training-to-testing split,validate the effectiveness of the proposed model.HybridLSTM achieves a classification accuracy of 96.3%,a precision of 95.8%,a recall of 96.1%,and an F1-score of 96.0%,outperforming several existing state-of-the-art methods.These results demonstrate the robustness,scalability,and generalization capability of HybridLSTM across varying environments and scene complexities.Moreover,the framework is optimized to balance classification performance with computational efficiency,making it highly suitable for real-time deployment in embedded autonomous driving systems.Future work will focus on extending the model to multi-class detection within a single frame and optimizing it further for edge-device deployments to reduce computational overhead in practical applications.
文摘A design of super-orthogonal space-time trellis codes (SOSTTCs) based on the trace criterion (TC) is proposed for improving the design of SOSTTCs. The shortcomings of the rank and determinant criteria based design and the advantages of the TC-based design are analyzed. The optimization principle of four factors is presented, which includes the space-time block coding (STBC) scheme, set partitioning, trellis structure, and the assignment of signal subsets and STBC schemes in the trellis. According to this principle, systematical and handcrafted design steps are given in detail. By constellation expansion, the code performance can be further improved. The code design results are given, and the new codes outperform others in the simulation.
基金supported by the National Natural Science Foundation of China (67441830108 and 41871224)。
文摘Spectral and spatial features in remotely sensed data play an irreplaceable role in classifying crop types for precision agriculture. Despite the thriving establishment of the handcrafted features, designing or selecting such features valid for specific crop types requires prior knowledge and thus remains an open challenge. Convolutional neural networks(CNNs) can effectively overcome this issue with their advanced ability to generate high-level features automatically but are still inadequate in mining spectral features compared to mining spatial features. This study proposed an enhanced spectral feature called Stacked Spectral Feature Space Patch(SSFSP) for CNN-based crop classification. SSFSP is a stack of twodimensional(2 D) gridded spectral feature images that record various crop types’ spatial and intensity distribution characteristics in a 2 D feature space consisting of two spectral bands. SSFSP can be input into2 D-CNNs to support the simultaneous mining of spectral and spatial features, as the spectral features are successfully converted to 2 D images that can be processed by CNN. We tested the performance of SSFSP by using it as the input to seven CNN models and one multilayer perceptron model for crop type classification compared to using conventional spectral features as input. Using high spatial resolution hyperspectral datasets at three sites, the comparative study demonstrated that SSFSP outperforms conventional spectral features regarding classification accuracy, robustness, and training efficiency. The theoretical analysis summarizes three reasons for its excellent performance. First, SSFSP mines the spectral interrelationship with feature generality, which reduces the required number of training samples.Second, the intra-class variance can be largely reduced by grid partitioning. Third, SSFSP is a highly sparse feature, which reduces the dependence on the CNN model structure and enables early and fast convergence in model training. In conclusion, SSFSP has great potential for practical crop classification in precision agriculture.
基金This research was supported by Korea Institute for Advancement of Technology(KIAT)grant funded by the Korea Government(MOTIE)(P0012724,The Competency Development Program for Industry Specialist)and the Soonchunhyang University Research Fund.
文摘Automatic gastrointestinal(GI)tract disease recognition is an important application of biomedical image processing.Conventionally,microscopic analysis of pathological tissue is used to detect abnormal areas of the GI tract.The procedure is subjective and results in significant inter-/intraobserver variations in disease detection.Moreover,a huge frame rate in video endoscopy is an overhead for the pathological findings of gastroenterologists to observe every frame with a detailed examination.Consequently,there is a huge demand for a reliable computer-aided diagnostic system(CADx)for diagnosing GI tract diseases.In this work,a CADx was proposed for the diagnosis and classification of GI tract diseases.A novel framework is presented where preprocessing(LAB color space)is performed first;then local binary patterns(LBP)or texture and deep learning(inceptionNet,ResNet50,and VGG-16)features are fused serially to improve the prediction of the abnormalities in the GI tract.Additionally,principal component analysis(PCA),entropy,and minimum redundancy and maximum relevance(mRMR)feature selection methods were analyzed to acquire the optimized characteristics,and various classifiers were trained using the fused features.Open-source color image datasets(KVASIR,NERTHUS,and stomach ULCER)were used for performance evaluation.The study revealed that the subspace discriminant classifier provided an efficient result with 95.02%accuracy on the KVASIR dataset,which proved to be better than the existing state-of-the-art approaches.
文摘Gold clay is a crafting medium consisting of gold particles mixed with an organic binder and water for making jewelry or decoration.The clay can be shaped by hand,textured,carved,formed or using molds.After drying and burning,the organic binder and water were decomposed and the gold particles were transformed to its final metal state.Although,gold clay is very expensive,it is useful to decorate the silver clay designed jewelry or small sculptures.In this research,nano-microplate gold and specific organic binder was used for producing nano-microplate gold clay.The objectives of this research are to study binder's type and ratios for optimum producing gold clay,and to study the heating condition for making silver and gold clay jewelry.The result showed that the clay can be fired with heating temperature at 900°C for an hour by electric kiln.The physical properties of the gold clay at different heating temperatures were determined.Furthermore,prototype of jewelry using the clay was
基金supported in part by the National Natural Science Foundation of China(No.61601027)the Opening Fund of the Space Objective Measure Key Laboratory(No.2016011)
文摘Digital low-density parity-check(LDPC) decoders can hardly meet the power-limits brought by the new application scenarios. The analog LDPC decoder, which is an application of the analog computation technology, is considered to have the potential to address this issue to some extent. However, due to the lack of automation tools and analog stopping criteria, the analog LDPC decoders suffer from costly handcraft design and additional decoding delay, and are not feasible to practical applications. To address these issues, a decoder architecture using reusable building blocks is designed to lower the handcraft design, and a probability stopping criterion that is specially designed for analog decoder is further planned and implemented to reduce the decoding delay. Then, a(480,240) CMOS analog LDPC decoder is designed and fabricated in a 0.35-μm CMOS technology. Experimental results show that the decoder prototype can achieve 50 Mbps throughput when the power consumption is about 86.3m W, and the decoding delay can be reduced by at most 93% compared with using the preset maximum decoding delay in existing works.
文摘Competition in modern society requires teachers to cultivate creative and pioneering talents, which indicates that teachers are required to cultivate students' creative ability. In the primary school stage, art handicraft is to cultivate students' practical ability and interest, while in the primary school stage, and art handicraft teaching mostly belongs to the teaching of cultivating students' creative ability. This article will discuss the art handicraft teaching and the primary school students' creative ability cultivation countermeasure.
基金The work of Hai-Miao Zhang was funded by China Postdoctoral Science Foundation(No.2018M641056)The work of Bin Dong was supported in part by the National Natural Science Foundation of China(No.11831002)Natural Science Foundation of Beijing(No.Z180001).
文摘Medical imaging is crucial in modern clinics to provide guidance to the diagnosis and treatment of diseases.Medical image reconstruction is one of the most fundamental and important components of medical imaging,whose major objective is to acquire high-quality medical images for clinical usage at the minimal cost and risk to the patients.Mathematical models in medical image reconstruction or,more generally,image restoration in computer vision have been playing a prominent role.Earlier mathematical models are mostly designed by human knowledge or hypothesis on the image to be reconstructed,and we shall call these models handcrafted models.Later,handcrafted plus data-driven modeling started to emerge which still mostly relies on human designs,while part of the model is learned from the observed data.More recently,as more data and computation resources are made available,deep learning based models(or deep models)pushed the data-driven modeling to the extreme where the models are mostly based on learning with minimal human designs.Both handcrafted and data-driven modeling have their own advantages and disadvantages.Typical handcrafted models are well interpretable with solid theoretical supports on the robustness,recoverability,complexity,etc.,whereas they may not be flexible and sophisticated enough to fully leverage large data sets.Data-driven models,especially deep models,on the other hand,are generally much more flexible and effective in extracting useful information from large data sets,while they are currently still in lack of theoretical foundations.Therefore,one of the major research trends in medical imaging is to combine handcrafted modeling with deep modeling so that we can enjoy benefits from both approaches.The major part of this article is to provide a conceptual review of some recent works on deep modeling from the unrolling dynamics viewpoint.This viewpoint stimulates new designs of neural network architectures with inspirations from optimization algorithms and numerical differential equations.Given the popularity of deep modeling,there are still vast remaining challenges in the field,as well as opportunities which we shall discuss at the end of this article.
基金supported by the National Natural Science Foundation of China(No.62203476)the Natural Science Foundation of Shenzhen(No.JCYJ20230807120801002).
文摘Three-dimensional skeleton-based action recognition(3D SAR)has gained important attention within the computer vision community,owing to the inherent advantages offered by skeleton data.As a result,a plethora of impressive works,including those based on conventional handcrafted features and learned feature extraction methods,have been conducted over the years.However,prior surveys on action recognition have primarily focused on video or red-green-blue(RGB)data-dominated approaches,with limited coverage of reviews related to skeleton data.Furthermore,despite the extensive application of deep learning methods in this field,there has been a notable absence of research that provides an introductory or comprehensive review from the perspective of deep learning architectures.To address these limitations,this survey first underscores the importance of action recognition and emphasizes the significance of 3-dimensional(3D)skeleton data as a valuable modality.Subsequently,we provide a comprehensive introduction to mainstream action recognition techniques based on 4 fundamental deep architectures,i.e.,recurrent neural networks,convolutional neural networks,graph convolutional network,and Transformers.All methods with the corresponding architectures are then presented in a data-driven manner with detailed discussion.Finally,we offer insights into the current largest 3D skeleton dataset,NTU-RGB+D,and its new edition,NTU-RGB+D 120,along with an overview of several top-performing algorithms on these datasets.To the best of our knowledge,this research represents the first comprehensive discussion of deep learning-based action recognition using 3D skeleton data.