In this Paper, a classification method based on neural networks is presented for recognition of 3D objects. Indeed, the objective of this paper is to classify an object query against objects in a database, which leads...In this Paper, a classification method based on neural networks is presented for recognition of 3D objects. Indeed, the objective of this paper is to classify an object query against objects in a database, which leads to recognition of the former. 3D objects of this database are transformations of other objects by one element of the overall transformation. The set of transformations considered in this work is the general affine group.展开更多
Action recognition is important for understanding the human behaviors in the video,and the video representation is the basis for action recognition.This paper provides a new video representation based on convolution n...Action recognition is important for understanding the human behaviors in the video,and the video representation is the basis for action recognition.This paper provides a new video representation based on convolution neural networks(CNN).For capturing human motion information in one CNN,we take both the optical flow maps and gray images as input,and combine multiple convolutional features by max pooling across frames.In another CNN,we input single color frame to capture context information.Finally,we take the top full connected layer vectors as video representation and train the classifiers by linear support vector machine.The experimental results show that the representation which integrates the optical flow maps and gray images obtains more discriminative properties than those which depend on only one element.On the most challenging data sets HMDB51 and UCF101,this video representation obtains competitive performance.展开更多
Scene recognition is a popular open problem in the computer vision field.Among lots of methods proposed in recent years,Convolutional Neural Network(CNN)based approaches achieve the best performance in scene recogniti...Scene recognition is a popular open problem in the computer vision field.Among lots of methods proposed in recent years,Convolutional Neural Network(CNN)based approaches achieve the best performance in scene recognition.We propose in this paper an advanced feature fusion algorithm using Multiple Convolutional Neural Network(Multi-CNN)for scene recognition.Unlike existing works that usually use individual convolutional neural network,a fusion of multiple different convolutional neural networks is applied for scene recognition.Firstly,we split training images in two directions and apply to three deep CNN model,and then extract features from the last full-connected(FC)layer and probabilistic layer on each model.Finally,feature vectors are fused with different fusion strategies in groups forwarded into SoftMax classifier.Our proposed algorithm is evaluated on three scene datasets for scene recognition.The experimental results demonstrate the effectiveness of proposed algorithm compared with other state-of-art approaches.展开更多
In order to accomplish the task of object recognition in natural scenes,a new object recognition algorithm based on an improved convolutional neural network(CNN)is proposed.First,candidate object windows are extracted...In order to accomplish the task of object recognition in natural scenes,a new object recognition algorithm based on an improved convolutional neural network(CNN)is proposed.First,candidate object windows are extracted from the original image.Then,candidate object windows are input into the improved CNN model to obtain deep features.Finally,the deep features are input into the Softmax and the confidence scores of classes are obtained.The candidate object window with the highest confidence score is selected as the object recognition result.Based on AlexNet,Inception V1 is introduced into the improved CNN and the fully connected layer is replaced by the average pooling layer,which widens the network and deepens the network at the same time.Experimental results show that the improved object recognition algorithm can obtain better recognition results in multiple natural scene images,and has a higher degree of accuracy than the classical algorithms in the field of object recognition.展开更多
Human gait recognition(HGR)is the process of identifying a sub-ject(human)based on their walking pattern.Each subject is a unique walking pattern and cannot be simulated by other subjects.But,gait recognition is not e...Human gait recognition(HGR)is the process of identifying a sub-ject(human)based on their walking pattern.Each subject is a unique walking pattern and cannot be simulated by other subjects.But,gait recognition is not easy and makes the system difficult if any object is carried by a subject,such as a bag or coat.This article proposes an automated architecture based on deep features optimization for HGR.To our knowledge,it is the first architecture in which features are fused using multiset canonical correlation analysis(MCCA).In the proposed method,original video frames are processed for all 11 selected angles of the CASIA B dataset and utilized to train two fine-tuned deep learning models such as Squeezenet and Efficientnet.Deep transfer learning was used to train both fine-tuned models on selected angles,yielding two new targeted models that were later used for feature engineering.Features are extracted from the deep layer of both fine-tuned models and fused into one vector using MCCA.An improved manta ray foraging optimization algorithm is also proposed to select the best features from the fused feature matrix and classified using a narrow neural network classifier.The experimental process was conducted on all 11 angles of the large multi-view gait dataset(CASIA B)dataset and obtained improved accuracy than the state-of-the-art techniques.Moreover,a detailed confidence interval based analysis also shows the effectiveness of the proposed architecture for HGR.展开更多
The complicated electromagnetic environment of the BeiDou satellites introduces vari-ous types of external jamming to communication links,in which recognition of jamming signals with uncertainties is essential.In this...The complicated electromagnetic environment of the BeiDou satellites introduces vari-ous types of external jamming to communication links,in which recognition of jamming signals with uncertainties is essential.In this work,the jamming recognition framework proposed consists of fea-ture fusion and a convolutional neural network(CNN).Firstly,the recognition inputs are obtained by prepossessing procedure,in which the 1-D power spectrum and 2-D time-frequency image are ac-cessed through the Welch algorithm and short-time Fourier transform(STFT),respectively.Then,the 1D-CNN and residual neural network(ResNet)are introduced to extract the deep features of the two prepossessing inputs,respectively.Finally,the two deep features are concatenated for the following three fully connected layers and output the jamming signal classification results through the softmax layer.Results show the proposed method could reduce the impacts of potential feature loss,therefore improving the generalization ability on dealing with uncertainties.展开更多
This article discusses vision recognition process and finds out that human recognizes objects not by their isolated features, but by their main difference features which people get by contrasting them. According to th...This article discusses vision recognition process and finds out that human recognizes objects not by their isolated features, but by their main difference features which people get by contrasting them. According to the resolving character of difference features for vision recognition, the difference feature neural network(DFNN) which is the improved auto-associative neural network is proposed.Using ORL database, the comparative experiment for face recognition with face images and the ones added Gaussian noise is performed, and the result shows that DFNN is better than the auto-associative neural network and it proves DFNN is more efficient.展开更多
To improve the recognition rate of signal modulation recognition methods based on the clustering algorithm under the low SNR, a modulation recognition method is proposed. The characteristic parameter of the signal is ...To improve the recognition rate of signal modulation recognition methods based on the clustering algorithm under the low SNR, a modulation recognition method is proposed. The characteristic parameter of the signal is extracted by using a clustering algorithm, the neural network is trained by using the algorithm of variable gradient correction (Polak-Ribiere) so as to enhance the rate of convergence, improve the performance of recognition under the low SNR and realize modulation recognition of the signal based on the modulation system of the constellation diagram. Simulation results show that the recognition rate based on this algorithm is enhanced over 30% compared with the methods that adopt clustering algorithm or neural network based on the back propagation algorithm alone under the low SNR. The recognition rate can reach 90% when the SNR is 4 dB, and the method is easy to be achieved so that it has a broad application prospect in the modulating recognition.展开更多
Plant recognition has great potential in forestry research and management.A new method combined back propagation neural network and radial basis function neural network to identify tree species using a few features an...Plant recognition has great potential in forestry research and management.A new method combined back propagation neural network and radial basis function neural network to identify tree species using a few features and samples.The process was carried out in three steps:image pretreatment,feature extraction,and leaf recognition.In the image pretreatment processing,an image segmentation method based on hue,saturation and value color space and connected component labeling was presented,which can obtain the complete leaf image without veins and back-ground.The BP-RBF hybrid neural network was used to test the influence of shape and texture on species recogni-tion.The recognition accuracy of different classifiers was used to compare classification performance.The accuracy of the BP-RBF hybrid neural network using nine dimensional features was 96.2%,highest among all the classifiers.展开更多
Objective In tongue diagnosis,the location,color,and distribution of spots can be used to speculate on the viscera and severity of the heat evil.This work focuses on the image analysis method of artificial intelligenc...Objective In tongue diagnosis,the location,color,and distribution of spots can be used to speculate on the viscera and severity of the heat evil.This work focuses on the image analysis method of artificial intelligence(AI)to study the spotted tongue recognition of traditional Chinese medicine(TCM).Methods A model of spotted tongue recognition and extraction is designed,which is based on the principle of image deep learning and instance segmentation.This model includes multiscale feature map generation,region proposal searching,and target region recognition.Firstly,deep convolution network is used to build multiscale low-and high-abstraction feature maps after which,target candidate box generation algorithm and selection strategy are used to select high-quality target candidate regions.Finally,classification network is used for classifying target regions and calculating target region pixels.As a result,the region segmentation of spotted tongue is obtained.Under non-standard illumination conditions,various tongue images were taken by mobile phones,and experiments were conducted.Results The spotted tongue recognition achieved an area under curve(AUC)of 92.40%,an accuracy of 84.30%with a sensitivity of 88.20%,a specificity of 94.19%,a recall of 88.20%,a regional pixel accuracy index pixel accuracy(PA)of 73.00%,a mean pixel accuracy(m PA)of73.00%,an intersection over union(Io U)of 60.00%,and a mean intersection over union(mIo U)of 56.00%.Conclusion The results of the study verify that the model is suitable for the application of the TCM tongue diagnosis system.Spotted tongue recognition via multiscale convolutional neural network(CNN)would help to improve spot classification and the accurate extraction of pixels of spot area as well as provide a practical method for intelligent tongue diagnosis of TCM.展开更多
With the rapid development of flexible electronics,the tactile systems for object recognition are becoming increasingly delicate.This paper presents the design of a tactile glove for object recognition,integrating 243...With the rapid development of flexible electronics,the tactile systems for object recognition are becoming increasingly delicate.This paper presents the design of a tactile glove for object recognition,integrating 243 palm pressure units and 126 finger joint strain units that are implemented by piezoresistive Velostat film.The palm pressure and joint bending strain data from the glove were collected using a two-dimensional resistance array scanning circuit and further converted into tactile images with a resolution of 32×32.To verify the effect of tactile data types on recognition precision,three datasets of tactile images were respectively built by palm pressure data,joint bending strain data,and a tactile data combing of both palm pressure and joint bending strain.An improved residual convolutional neural network(CNN)model,SP-ResNet,was developed by light-weighting ResNet-18 to classify these tactile images.Experimental results show that the data collection method combining palm pressure and joint bending strain demonstrates a 4.33%improvement in recognition precision compared to the best results obtained by using only palm pressure or joint bending strain.The recognition precision of 95.50%for 16 objects can be achieved by the presented tactile glove with SP-ResNet of less computation cost.The presented tactile system can serve as a sensing platform for intelligent prosthetics and robot grippers.展开更多
A neural network integrated classifier(NNIC) designed with a new modulation recognition algorithm based on the decision-making tree is proposed in this paper.Firstly,instantaneous parameters are extracted in the time ...A neural network integrated classifier(NNIC) designed with a new modulation recognition algorithm based on the decision-making tree is proposed in this paper.Firstly,instantaneous parameters are extracted in the time domain by the coordinated rotation digital computer(CORDIC) algorithm based on the extended convergence domain and feature parameters of frequency spectrum and power spectrum are extracted by the time-frequency analysis method.All pattern identification parameters are calculated under the I/Q orthogonal two-channel structure,and constructed into the feature vector set.Next,the classifier is designed according to the modulation pattern and recognition performance of the feature parameter set,the optimum threshold is selected for each feature parameter based on the decision-making mechanism in a single classifier,multi-source information fusion and modulation recognition are realized based on feature parameter judge process in the NNIC.Simulation results show NNIC is competent for all modulation recognitions,8 kinds of digital modulated signals are effectively identified,which shows the recognition rate and anti-interference capability at low SNR are improved greatly,the overall recognition rate can reach 100%when SNR is12dB.展开更多
Gait is an essential biomedical feature that distinguishes individuals through walking.This feature automatically stimulates the need for remote human recognition in security-sensitive visual monitoring applications.H...Gait is an essential biomedical feature that distinguishes individuals through walking.This feature automatically stimulates the need for remote human recognition in security-sensitive visual monitoring applications.However,there is still a lack of sufficient accuracy of gait recognition at night,in addition to taking some critical factors that affect the performances of the recognition algorithm.Therefore,a novel approach is proposed to automatically identify individuals from thermal infrared(TIR)images according to their gaits captured at night.This approach uses a new night gait network(NGaitNet)based on similarity deep convolutional neural networks(CNNs)method to enhance gait recognition at night.First,the TIR image is represented via personal movements and enhanced body skeleton segments.Then,the state-space method with a Hough transform is used to extract gait features to obtain skeletal joints′angles.These features are trained to identify the most discriminating gait patterns that indicate a change in human identity.To verify the proposed method,the experimental results are performed by using learning and validation curves via being connected by the Visdom website.The proposed thermal infrared imaging night gait recognition(TIRNGaitNet)approach has achieved the highest gait recognition accuracy rates(99.5%,97.0%),especially under normal walking conditions on the Chinese Academy of Sciences Institute of Automation infrared night gait dataset(CASIA C)and Donghua University thermal infrared night gait database(DHU night gait dataset).On the same dataset,the results of the TIRNGaitNet approach provide the record scores of(98.0%,87.0%)under the slow walking condition and(94.0%,86.0%)for the quick walking condition.展开更多
This paper presents work on modulated signal recognition using an artificial neural network (ANN) developed using the Python programme language. The study is basically on the analysis of analog modulated signals. Fo...This paper presents work on modulated signal recognition using an artificial neural network (ANN) developed using the Python programme language. The study is basically on the analysis of analog modulated signals. Four of the best-known analog modulation types are considered namely: amplitude modulation (AM), double sideband (DSB) modulation, single sideband (SSB) modulation and frequency modulation (FM). Computer simulations of the four modulated signals are carried out using MATLAB. MATLAB code is used in simulating the analog signals as well as the power spectral density of each of the analog modulated signals. In achieving an accurate classification of each of the modulated signals, extensive simulations are performed for the training of the artificial neural network. The results of the study show accurate and correct performance of the developed automatic modulation recognition with average success rate above 99.5%.展开更多
To solve the problem of mistake recognition among rice diseases, automatic recognition methods based on BP(back propagation) neural network were studied in this paper for blast, sheath blight and bacterial blight. Cho...To solve the problem of mistake recognition among rice diseases, automatic recognition methods based on BP(back propagation) neural network were studied in this paper for blast, sheath blight and bacterial blight. Chose mobile terminal equipment as image collecting tool and built database of rice leaf images with diseases under threshold segmentation method. Characteristic parameters were extracted from color, shape and texture. Furthermore, parameters were optimized using the single-factor variance analysis and the effects of BP neural network model. The optimization would simplify BP neural network model without reducing the recognition accuracy. The finally model could successfully recognize 98%, 96% and 98% of rice blast, sheath blight and white leaf blight, respectively.展开更多
With the development of the economy and the surge in car ownership, the sale of used cars has been welcomed by more and more people, and the information of the vehicle condition is the focus information of them. The f...With the development of the economy and the surge in car ownership, the sale of used cars has been welcomed by more and more people, and the information of the vehicle condition is the focus information of them. The frame number is a unique number used in the vehicle, and by identifying it can quickly find out the vehicle models and manufacturers. The traditional character recognition method has the problem of complex feature extraction, and the convolutional neural network has unique advantages in processing two-dimensional images. This paper analyzed the key techniques of convolutional neural networks compared with traditional neural networks, and proposed improved methods for key technologies, thus increasing the recognition of characters and applying them to the recognition of frame number characters.展开更多
Handwritten signature and character recognition has become challenging research topic due to its numerous applications. In this paper, we proposed a system that has three sub-systems. The three subsystems focus on off...Handwritten signature and character recognition has become challenging research topic due to its numerous applications. In this paper, we proposed a system that has three sub-systems. The three subsystems focus on offline recognition of handwritten English alphabetic characters (uppercase and lowercase), numeric characters (0 - 9) and individual signatures respectively. The system includes several stages like image preprocessing, the post-processing, the segmentation, the detection of the required amount of the character and signature, feature extraction and finally Neural Network recognition. At first, the scanned image is filtered after conversion of the scanned image into a gray image. Then image cropping method is applied to detect the signature. Then an accurate recognition is ensured by post-processing the cropped images. MATLAB has been used to design the system. The subsystems are then tested for several samples and the results are found satisfactory at about 97% success rate. The quality of the image plays a vital role as the images of poor or mediocre quality may lead to unsuccessful recognition and verification.展开更多
An efficient face recognition system with face image representation using averaged wavelet packet coefficients, compact and meaningful feature vectors dimensional reduction and recognition using radial basis function ...An efficient face recognition system with face image representation using averaged wavelet packet coefficients, compact and meaningful feature vectors dimensional reduction and recognition using radial basis function (RBF) neural network is presented. The face images are decomposed by 2-level two-dimensional (2-D) wavelet packet transformation. The wavelet packet coefficients obtained from the wavelet packet transformation are averaged using two different proposed methods. In the first method, wavelet packet coefficients of individual samples of a class are averaged then decomposed. The wavelet packet coefficients of all the samples of a class are averaged in the second method. The averaged wavelet packet coefficients are recognized by a RBF network. The proposed work tested on three face databases such as Olivetti-Oracle Research Lab (ORL), Japanese Female Facial Expression (JAFFE) and Essexface database. The proposed methods result in dimensionality reduction, low computational complexity and provide better recognition rates. The computational complexity is low as the dimensionality of the input pattern is reduced.展开更多
Object recognition in very high-resolution remote sensing images is a basic problem in the field of aerial and satellite image analysis.With the development of sensor technology and aerospace remote sensing technology...Object recognition in very high-resolution remote sensing images is a basic problem in the field of aerial and satellite image analysis.With the development of sensor technology and aerospace remote sensing technology,the quality and quantity of remote sensing images are improved.Traditional recognition methods have a certain limitation in describing higher-level features,but object recognition method based on convolutional neural network(CNN)can not only deal with large scale images,but also train features automatically with high efficiency.It is mainly used on object recognition for remote sensing images.In this paper,an AlexNet CNN model is trained using 2100 remote sensing images,and correction rate can reach 97.6%after 2000 iterations.Then based on trained model,a parallel design of CNN for remote sensing images object recognition based on data-driven array processor(DDAP)is proposed.The consuming cycles are counted.Simultaneously,the proposed architecture is realized on Xilinx V6 development board,and synthesized based on SMIC 130 nm complementary metal oxid semiconductor(CMOS)technology.The experimental results show that the proposed architecture has a certain degree of parallelism to achieve the purpose of accelerating calculations.展开更多
With the development of deep learning and Convolutional Neural Networks(CNNs),the accuracy of automatic food recognition based on visual data have significantly improved.Some research studies have shown that the deepe...With the development of deep learning and Convolutional Neural Networks(CNNs),the accuracy of automatic food recognition based on visual data have significantly improved.Some research studies have shown that the deeper the model is,the higher the accuracy is.However,very deep neural networks would be affected by the overfitting problem and also consume huge computing resources.In this paper,a new classification scheme is proposed for automatic food-ingredient recognition based on deep learning.We construct an up-to-date combinational convolutional neural network(CBNet)with a subnet merging technique.Firstly,two different neural networks are utilized for learning interested features.Then,a well-designed feature fusion component aggregates the features from subnetworks,further extracting richer and more precise features for image classification.In order to learn more complementary features,the corresponding fusion strategies are also proposed,including auxiliary classifiers and hyperparameters setting.Finally,CBNet based on the well-known VGGNet,ResNet and DenseNet is evaluated on a dataset including 41 major categories of food ingredients and 100 images for each category.Theoretical analysis and experimental results demonstrate that CBNet achieves promising accuracy for multi-class classification and improves the performance of convolutional neural networks.展开更多
文摘In this Paper, a classification method based on neural networks is presented for recognition of 3D objects. Indeed, the objective of this paper is to classify an object query against objects in a database, which leads to recognition of the former. 3D objects of this database are transformations of other objects by one element of the overall transformation. The set of transformations considered in this work is the general affine group.
基金Supported by the National High Technology Research and Development Program of China(863 Program,2015AA016306)National Nature Science Foundation of China(61231015)+2 种基金Internet of Things Development Funding Project of Ministry of Industry in 2013(25)Technology Research Program of Ministry of Public Security(2016JSYJA12)the Nature Science Foundation of Hubei Province(2014CFB712)
文摘Action recognition is important for understanding the human behaviors in the video,and the video representation is the basis for action recognition.This paper provides a new video representation based on convolution neural networks(CNN).For capturing human motion information in one CNN,we take both the optical flow maps and gray images as input,and combine multiple convolutional features by max pooling across frames.In another CNN,we input single color frame to capture context information.Finally,we take the top full connected layer vectors as video representation and train the classifiers by linear support vector machine.The experimental results show that the representation which integrates the optical flow maps and gray images obtains more discriminative properties than those which depend on only one element.On the most challenging data sets HMDB51 and UCF101,this video representation obtains competitive performance.
文摘Scene recognition is a popular open problem in the computer vision field.Among lots of methods proposed in recent years,Convolutional Neural Network(CNN)based approaches achieve the best performance in scene recognition.We propose in this paper an advanced feature fusion algorithm using Multiple Convolutional Neural Network(Multi-CNN)for scene recognition.Unlike existing works that usually use individual convolutional neural network,a fusion of multiple different convolutional neural networks is applied for scene recognition.Firstly,we split training images in two directions and apply to three deep CNN model,and then extract features from the last full-connected(FC)layer and probabilistic layer on each model.Finally,feature vectors are fused with different fusion strategies in groups forwarded into SoftMax classifier.Our proposed algorithm is evaluated on three scene datasets for scene recognition.The experimental results demonstrate the effectiveness of proposed algorithm compared with other state-of-art approaches.
基金Supported by the National Natural Science Foundation of China(61701029)Basic Research Foundation of Beijing Institute of Technology(20170542008)Industry-University Research Innovation Foundation of the Science and Technology Development Center of the Ministry of Education(2018A02012)。
文摘In order to accomplish the task of object recognition in natural scenes,a new object recognition algorithm based on an improved convolutional neural network(CNN)is proposed.First,candidate object windows are extracted from the original image.Then,candidate object windows are input into the improved CNN model to obtain deep features.Finally,the deep features are input into the Softmax and the confidence scores of classes are obtained.The candidate object window with the highest confidence score is selected as the object recognition result.Based on AlexNet,Inception V1 is introduced into the improved CNN and the fully connected layer is replaced by the average pooling layer,which widens the network and deepens the network at the same time.Experimental results show that the improved object recognition algorithm can obtain better recognition results in multiple natural scene images,and has a higher degree of accuracy than the classical algorithms in the field of object recognition.
基金supported by the MSIT(Ministry of Science and ICT),Korea,under the ICAN(ICT Challenge and Advanced Network of HRD)program(IITP-2022-2020-0-01832)supervised by the IITP(Institute of Information&Communications Technology Planning&Evaluation)and the Soonchunhyang University Research Fund.
文摘Human gait recognition(HGR)is the process of identifying a sub-ject(human)based on their walking pattern.Each subject is a unique walking pattern and cannot be simulated by other subjects.But,gait recognition is not easy and makes the system difficult if any object is carried by a subject,such as a bag or coat.This article proposes an automated architecture based on deep features optimization for HGR.To our knowledge,it is the first architecture in which features are fused using multiset canonical correlation analysis(MCCA).In the proposed method,original video frames are processed for all 11 selected angles of the CASIA B dataset and utilized to train two fine-tuned deep learning models such as Squeezenet and Efficientnet.Deep transfer learning was used to train both fine-tuned models on selected angles,yielding two new targeted models that were later used for feature engineering.Features are extracted from the deep layer of both fine-tuned models and fused into one vector using MCCA.An improved manta ray foraging optimization algorithm is also proposed to select the best features from the fused feature matrix and classified using a narrow neural network classifier.The experimental process was conducted on all 11 angles of the large multi-view gait dataset(CASIA B)dataset and obtained improved accuracy than the state-of-the-art techniques.Moreover,a detailed confidence interval based analysis also shows the effectiveness of the proposed architecture for HGR.
基金supported by the National Key Research and De-velopment Program of China(No.2020YFB0505601).
文摘The complicated electromagnetic environment of the BeiDou satellites introduces vari-ous types of external jamming to communication links,in which recognition of jamming signals with uncertainties is essential.In this work,the jamming recognition framework proposed consists of fea-ture fusion and a convolutional neural network(CNN).Firstly,the recognition inputs are obtained by prepossessing procedure,in which the 1-D power spectrum and 2-D time-frequency image are ac-cessed through the Welch algorithm and short-time Fourier transform(STFT),respectively.Then,the 1D-CNN and residual neural network(ResNet)are introduced to extract the deep features of the two prepossessing inputs,respectively.Finally,the two deep features are concatenated for the following three fully connected layers and output the jamming signal classification results through the softmax layer.Results show the proposed method could reduce the impacts of potential feature loss,therefore improving the generalization ability on dealing with uncertainties.
文摘This article discusses vision recognition process and finds out that human recognizes objects not by their isolated features, but by their main difference features which people get by contrasting them. According to the resolving character of difference features for vision recognition, the difference feature neural network(DFNN) which is the improved auto-associative neural network is proposed.Using ORL database, the comparative experiment for face recognition with face images and the ones added Gaussian noise is performed, and the result shows that DFNN is better than the auto-associative neural network and it proves DFNN is more efficient.
基金supported by the National Natural Science Foundation of China(6107207061301179)the National Science and Technology Major Project(2010ZX03006-002-04)
文摘To improve the recognition rate of signal modulation recognition methods based on the clustering algorithm under the low SNR, a modulation recognition method is proposed. The characteristic parameter of the signal is extracted by using a clustering algorithm, the neural network is trained by using the algorithm of variable gradient correction (Polak-Ribiere) so as to enhance the rate of convergence, improve the performance of recognition under the low SNR and realize modulation recognition of the signal based on the modulation system of the constellation diagram. Simulation results show that the recognition rate based on this algorithm is enhanced over 30% compared with the methods that adopt clustering algorithm or neural network based on the back propagation algorithm alone under the low SNR. The recognition rate can reach 90% when the SNR is 4 dB, and the method is easy to be achieved so that it has a broad application prospect in the modulating recognition.
基金This work is supported by the Fundamental Research Funds for the Central Universities(No.2572020BC07)the Project of National Science Foundation of China(No.31570712).
文摘Plant recognition has great potential in forestry research and management.A new method combined back propagation neural network and radial basis function neural network to identify tree species using a few features and samples.The process was carried out in three steps:image pretreatment,feature extraction,and leaf recognition.In the image pretreatment processing,an image segmentation method based on hue,saturation and value color space and connected component labeling was presented,which can obtain the complete leaf image without veins and back-ground.The BP-RBF hybrid neural network was used to test the influence of shape and texture on species recogni-tion.The recognition accuracy of different classifiers was used to compare classification performance.The accuracy of the BP-RBF hybrid neural network using nine dimensional features was 96.2%,highest among all the classifiers.
基金Anhui Province College Natural Science Fund Key Project of China(KJ2020ZD77)the Project of Education Department of Anhui Province(KJ2020A0379)。
文摘Objective In tongue diagnosis,the location,color,and distribution of spots can be used to speculate on the viscera and severity of the heat evil.This work focuses on the image analysis method of artificial intelligence(AI)to study the spotted tongue recognition of traditional Chinese medicine(TCM).Methods A model of spotted tongue recognition and extraction is designed,which is based on the principle of image deep learning and instance segmentation.This model includes multiscale feature map generation,region proposal searching,and target region recognition.Firstly,deep convolution network is used to build multiscale low-and high-abstraction feature maps after which,target candidate box generation algorithm and selection strategy are used to select high-quality target candidate regions.Finally,classification network is used for classifying target regions and calculating target region pixels.As a result,the region segmentation of spotted tongue is obtained.Under non-standard illumination conditions,various tongue images were taken by mobile phones,and experiments were conducted.Results The spotted tongue recognition achieved an area under curve(AUC)of 92.40%,an accuracy of 84.30%with a sensitivity of 88.20%,a specificity of 94.19%,a recall of 88.20%,a regional pixel accuracy index pixel accuracy(PA)of 73.00%,a mean pixel accuracy(m PA)of73.00%,an intersection over union(Io U)of 60.00%,and a mean intersection over union(mIo U)of 56.00%.Conclusion The results of the study verify that the model is suitable for the application of the TCM tongue diagnosis system.Spotted tongue recognition via multiscale convolutional neural network(CNN)would help to improve spot classification and the accurate extraction of pixels of spot area as well as provide a practical method for intelligent tongue diagnosis of TCM.
基金supported by the Key Research and Development Program of Shaanxi Province(No.2024 GX-YBXM-178)the Shaanxi Province Qinchuangyuan“Scientists+Engineers”Team Development(No.2022KXJ032)。
文摘With the rapid development of flexible electronics,the tactile systems for object recognition are becoming increasingly delicate.This paper presents the design of a tactile glove for object recognition,integrating 243 palm pressure units and 126 finger joint strain units that are implemented by piezoresistive Velostat film.The palm pressure and joint bending strain data from the glove were collected using a two-dimensional resistance array scanning circuit and further converted into tactile images with a resolution of 32×32.To verify the effect of tactile data types on recognition precision,three datasets of tactile images were respectively built by palm pressure data,joint bending strain data,and a tactile data combing of both palm pressure and joint bending strain.An improved residual convolutional neural network(CNN)model,SP-ResNet,was developed by light-weighting ResNet-18 to classify these tactile images.Experimental results show that the data collection method combining palm pressure and joint bending strain demonstrates a 4.33%improvement in recognition precision compared to the best results obtained by using only palm pressure or joint bending strain.The recognition precision of 95.50%for 16 objects can be achieved by the presented tactile glove with SP-ResNet of less computation cost.The presented tactile system can serve as a sensing platform for intelligent prosthetics and robot grippers.
基金Supported by the National Natural Science Foundation of China(No.61001049)Key Laboratory of Computer Architecture Opening Topic Fund Subsidization(CARCH201103)Beijing Natural Science Foundation(No.Z2002012201101)
文摘A neural network integrated classifier(NNIC) designed with a new modulation recognition algorithm based on the decision-making tree is proposed in this paper.Firstly,instantaneous parameters are extracted in the time domain by the coordinated rotation digital computer(CORDIC) algorithm based on the extended convergence domain and feature parameters of frequency spectrum and power spectrum are extracted by the time-frequency analysis method.All pattern identification parameters are calculated under the I/Q orthogonal two-channel structure,and constructed into the feature vector set.Next,the classifier is designed according to the modulation pattern and recognition performance of the feature parameter set,the optimum threshold is selected for each feature parameter based on the decision-making mechanism in a single classifier,multi-source information fusion and modulation recognition are realized based on feature parameter judge process in the NNIC.Simulation results show NNIC is competent for all modulation recognitions,8 kinds of digital modulated signals are effectively identified,which shows the recognition rate and anti-interference capability at low SNR are improved greatly,the overall recognition rate can reach 100%when SNR is12dB.
文摘Gait is an essential biomedical feature that distinguishes individuals through walking.This feature automatically stimulates the need for remote human recognition in security-sensitive visual monitoring applications.However,there is still a lack of sufficient accuracy of gait recognition at night,in addition to taking some critical factors that affect the performances of the recognition algorithm.Therefore,a novel approach is proposed to automatically identify individuals from thermal infrared(TIR)images according to their gaits captured at night.This approach uses a new night gait network(NGaitNet)based on similarity deep convolutional neural networks(CNNs)method to enhance gait recognition at night.First,the TIR image is represented via personal movements and enhanced body skeleton segments.Then,the state-space method with a Hough transform is used to extract gait features to obtain skeletal joints′angles.These features are trained to identify the most discriminating gait patterns that indicate a change in human identity.To verify the proposed method,the experimental results are performed by using learning and validation curves via being connected by the Visdom website.The proposed thermal infrared imaging night gait recognition(TIRNGaitNet)approach has achieved the highest gait recognition accuracy rates(99.5%,97.0%),especially under normal walking conditions on the Chinese Academy of Sciences Institute of Automation infrared night gait dataset(CASIA C)and Donghua University thermal infrared night gait database(DHU night gait dataset).On the same dataset,the results of the TIRNGaitNet approach provide the record scores of(98.0%,87.0%)under the slow walking condition and(94.0%,86.0%)for the quick walking condition.
文摘This paper presents work on modulated signal recognition using an artificial neural network (ANN) developed using the Python programme language. The study is basically on the analysis of analog modulated signals. Four of the best-known analog modulation types are considered namely: amplitude modulation (AM), double sideband (DSB) modulation, single sideband (SSB) modulation and frequency modulation (FM). Computer simulations of the four modulated signals are carried out using MATLAB. MATLAB code is used in simulating the analog signals as well as the power spectral density of each of the analog modulated signals. In achieving an accurate classification of each of the modulated signals, extensive simulations are performed for the training of the artificial neural network. The results of the study show accurate and correct performance of the developed automatic modulation recognition with average success rate above 99.5%.
基金Supported by Quality and Brand Construction of"Internet+County Characteristic Agricultural Products"(ZY17C06)
文摘To solve the problem of mistake recognition among rice diseases, automatic recognition methods based on BP(back propagation) neural network were studied in this paper for blast, sheath blight and bacterial blight. Chose mobile terminal equipment as image collecting tool and built database of rice leaf images with diseases under threshold segmentation method. Characteristic parameters were extracted from color, shape and texture. Furthermore, parameters were optimized using the single-factor variance analysis and the effects of BP neural network model. The optimization would simplify BP neural network model without reducing the recognition accuracy. The finally model could successfully recognize 98%, 96% and 98% of rice blast, sheath blight and white leaf blight, respectively.
文摘With the development of the economy and the surge in car ownership, the sale of used cars has been welcomed by more and more people, and the information of the vehicle condition is the focus information of them. The frame number is a unique number used in the vehicle, and by identifying it can quickly find out the vehicle models and manufacturers. The traditional character recognition method has the problem of complex feature extraction, and the convolutional neural network has unique advantages in processing two-dimensional images. This paper analyzed the key techniques of convolutional neural networks compared with traditional neural networks, and proposed improved methods for key technologies, thus increasing the recognition of characters and applying them to the recognition of frame number characters.
文摘Handwritten signature and character recognition has become challenging research topic due to its numerous applications. In this paper, we proposed a system that has three sub-systems. The three subsystems focus on offline recognition of handwritten English alphabetic characters (uppercase and lowercase), numeric characters (0 - 9) and individual signatures respectively. The system includes several stages like image preprocessing, the post-processing, the segmentation, the detection of the required amount of the character and signature, feature extraction and finally Neural Network recognition. At first, the scanned image is filtered after conversion of the scanned image into a gray image. Then image cropping method is applied to detect the signature. Then an accurate recognition is ensured by post-processing the cropped images. MATLAB has been used to design the system. The subsystems are then tested for several samples and the results are found satisfactory at about 97% success rate. The quality of the image plays a vital role as the images of poor or mediocre quality may lead to unsuccessful recognition and verification.
文摘An efficient face recognition system with face image representation using averaged wavelet packet coefficients, compact and meaningful feature vectors dimensional reduction and recognition using radial basis function (RBF) neural network is presented. The face images are decomposed by 2-level two-dimensional (2-D) wavelet packet transformation. The wavelet packet coefficients obtained from the wavelet packet transformation are averaged using two different proposed methods. In the first method, wavelet packet coefficients of individual samples of a class are averaged then decomposed. The wavelet packet coefficients of all the samples of a class are averaged in the second method. The averaged wavelet packet coefficients are recognized by a RBF network. The proposed work tested on three face databases such as Olivetti-Oracle Research Lab (ORL), Japanese Female Facial Expression (JAFFE) and Essexface database. The proposed methods result in dimensionality reduction, low computational complexity and provide better recognition rates. The computational complexity is low as the dimensionality of the input pattern is reduced.
基金This work was supported by the National Natural Science Foundation of China(61802304,61834005,61772417,61634004,61602377)the Shaanxi Provincial Co-ordination Innovation Project of Science and Technology(20I6KTZDGY02-04-02)the Shaanxi Provincial Key Research and Development Plan(2017GY-060).
文摘Object recognition in very high-resolution remote sensing images is a basic problem in the field of aerial and satellite image analysis.With the development of sensor technology and aerospace remote sensing technology,the quality and quantity of remote sensing images are improved.Traditional recognition methods have a certain limitation in describing higher-level features,but object recognition method based on convolutional neural network(CNN)can not only deal with large scale images,but also train features automatically with high efficiency.It is mainly used on object recognition for remote sensing images.In this paper,an AlexNet CNN model is trained using 2100 remote sensing images,and correction rate can reach 97.6%after 2000 iterations.Then based on trained model,a parallel design of CNN for remote sensing images object recognition based on data-driven array processor(DDAP)is proposed.The consuming cycles are counted.Simultaneously,the proposed architecture is realized on Xilinx V6 development board,and synthesized based on SMIC 130 nm complementary metal oxid semiconductor(CMOS)technology.The experimental results show that the proposed architecture has a certain degree of parallelism to achieve the purpose of accelerating calculations.
基金This paper is partially supported by National Natural Foundation of China(Grant No.61772561)the Key Research&Development Plan of Hunan Province(Grant No.2018NK2012)+2 种基金Postgraduate Research and Innovative Project of Central South University of Forestry and Technology(Grant No.20183012)Graduate Education and Teaching Reform Project of Central South University of Forestry and Technology(Grant No.2018JG005)Teaching Reform Project of Central South University of Forestry and Technology(Grant No.20180682).
文摘With the development of deep learning and Convolutional Neural Networks(CNNs),the accuracy of automatic food recognition based on visual data have significantly improved.Some research studies have shown that the deeper the model is,the higher the accuracy is.However,very deep neural networks would be affected by the overfitting problem and also consume huge computing resources.In this paper,a new classification scheme is proposed for automatic food-ingredient recognition based on deep learning.We construct an up-to-date combinational convolutional neural network(CBNet)with a subnet merging technique.Firstly,two different neural networks are utilized for learning interested features.Then,a well-designed feature fusion component aggregates the features from subnetworks,further extracting richer and more precise features for image classification.In order to learn more complementary features,the corresponding fusion strategies are also proposed,including auxiliary classifiers and hyperparameters setting.Finally,CBNet based on the well-known VGGNet,ResNet and DenseNet is evaluated on a dataset including 41 major categories of food ingredients and 100 images for each category.Theoretical analysis and experimental results demonstrate that CBNet achieves promising accuracy for multi-class classification and improves the performance of convolutional neural networks.