期刊文献+
共找到5,003篇文章
< 1 2 250 >
每页显示 20 50 100
Method for Behavior Recognition of Hu Sheep in Intensive Farming Based on HLNC-YOLO
1
作者 JI Ronghua CHANG Hongrui +2 位作者 ZHANG Suoxiang LIU Zhongying WU Zhonghong 《农业机械学报》 北大核心 2026年第2期265-275,共11页
Behavior recognition of Hu sheep contributes to their intensive and intelligent farming.Due to the generally high density of Hu sheep farming,severe occlusion occurs among different behaviors and even among sheep perf... Behavior recognition of Hu sheep contributes to their intensive and intelligent farming.Due to the generally high density of Hu sheep farming,severe occlusion occurs among different behaviors and even among sheep performing the same behavior,leading to missing and false detection issues in existing behavior recognition methods.A high-low frequency aggregated attention and negative sample comprehensive score loss and comprehensive score soft non-maximum suppression-YOLO(HLNC-YOLO)was proposed for identifying the behavior of Hu sheep,addressing the issues of missed and erroneous detections caused by occlusion between Hu sheep in intensive farming.Firstly,images of four typical behaviors-standing,lying,eating,and drinking-were collected from the sheep farm to construct the Hu sheep behavior dataset(HSBD).Next,to solve the occlusion issues,during the training phase,the C2F-HLAtt module was integrated,which combined high-low frequency aggregation attention,into the YOLO v8 Backbone to perceive occluded objects and introduce an auxiliary reversible branch to retain more effective features.Using comprehensive score regression loss(CSLoss)to reduce the scores of suboptimal boxes and enhance the comprehensive scores of occluded object boxes.Finally,the soft comprehensive score non-maximal suppression(Soft-CS-NMS)algorithm filtered prediction boxes during the inferencing.Testing on the HSBD,HLNC-YOLO achieved a mean average precision(mAP@50)of 87.8%,with a memory footprint of 17.4 MB.This represented an improvement of 7.1,2.2,4.6,and 11 percentage points over YOLO v8,YOLO v9,YOLO v10,and Faster R-CNN,respectively.Research indicated that the HLNC-YOLO accurately identified the behavior of Hu sheep in intensive farming and possessed generalization capabilities,providing technical support for smart farming. 展开更多
关键词 behavior recognition YOLO loss function attention mechanism
在线阅读 下载PDF
Human Activity Recognition Using Weighted Average Ensemble by Selected Deep Learning Models
2
作者 Waseem Akhtar Mahwish Ilyas +3 位作者 Romana Aziz Ghadah Aldehim Tassawar Iqbal Muhammad Ramzan 《Computer Modeling in Engineering & Sciences》 2026年第2期971-989,共19页
Human Activity Recognition(HAR)is a novel area for computer vision.It has a great impact on healthcare,smart environments,and surveillance while is able to automatically detect human behavior.It plays a vital role in ... Human Activity Recognition(HAR)is a novel area for computer vision.It has a great impact on healthcare,smart environments,and surveillance while is able to automatically detect human behavior.It plays a vital role in many applications,such as smart home,healthcare,human computer interaction,sports analysis,and especially,intelligent surveillance.In this paper,we propose a robust and efficient HAR system by leveraging deep learning paradigms,including pre-trained models,CNN architectures,and their average-weighted fusion.However,due to the diversity of human actions and various environmental influences,as well as a lack of data and resources,achieving high recognition accuracy remain elusive.In this work,a weighted average ensemble technique is employed to fuse three deep learning models:EfficientNet,ResNet50,and a custom CNN.The results of this study indicate that using a weighted average ensemble strategy for developing more effective HAR models may be a promising idea for detection and classification of human activities.Experiments by using the benchmark dataset proved that the proposed weighted ensemble approach outperformed existing approaches in terms of accuracy and other key performance measures.The combined average-weighted ensemble of pre-trained and CNN models obtained an accuracy of 98%,compared to 97%,96%,and 95%for the customized CNN,EfficientNet,and ResNet50 models,respectively. 展开更多
关键词 Artificial intelligence computer vision deep learning recognition human activity classification image processing
在线阅读 下载PDF
MFCCT:A Robust Spectral-Temporal Fusion Method with DeepConvLSTM for Human Activity Recognition
3
作者 Rashid Jahangir Nazik Alturki +1 位作者 Muhammad Asif Nauman Faiqa Hanif 《Computers, Materials & Continua》 2026年第2期852-871,共20页
Human activity recognition(HAR)is a method to predict human activities from sensor signals using machine learning(ML)techniques.HAR systems have several applications in various domains,including medicine,surveillance,... Human activity recognition(HAR)is a method to predict human activities from sensor signals using machine learning(ML)techniques.HAR systems have several applications in various domains,including medicine,surveillance,behavioral monitoring,and posture analysis.Extraction of suitable information from sensor data is an important part of the HAR process to recognize activities accurately.Several research studies on HAR have utilizedMel frequency cepstral coefficients(MFCCs)because of their effectiveness in capturing the periodic pattern of sensor signals.However,existing MFCC-based approaches often fail to capture sufficient temporal variability,which limits their ability to distinguish between complex or imbalanced activity classes robustly.To address this gap,this study proposes a feature fusion strategy that merges time-based and MFCC features(MFCCT)to enhance activity representation.The merged features were fed to a convolutional neural network(CNN)integrated with long shortterm memory(LSTM)—DeepConvLSTM to construct the HAR model.The MFCCT features with DeepConvLSTM achieved better performance as compared to MFCCs and time-based features on PAMAP2,UCI-HAR,and WISDM by obtaining an accuracy of 97%,98%,and 97%,respectively.In addition,DeepConvLSTM outperformed the deep learning(DL)algorithms that have recently been employed in HAR.These results confirm that the proposed hybrid features are not only practical but also generalizable,making them applicable across diverse HAR datasets for accurate activity classification. 展开更多
关键词 DeepConvLSTM human activity recognition(HAR) MFCCT feature fusion wearable sensors
在线阅读 下载PDF
Intelligent Human Interaction Recognition with Multi-Modal Feature Extraction and Bidirectional LSTM
4
作者 Muhammad Hamdan Azhar Yanfeng Wu +4 位作者 Nouf Abdullah Almujally Shuaa S.Alharbi Asaad Algarni Ahmad Jalal Hui Liu 《Computers, Materials & Continua》 2026年第4期1632-1649,共18页
Recognizing human interactions in RGB videos is a critical task in computer vision,with applications in video surveillance.Existing deep learning-based architectures have achieved strong results,but are computationall... Recognizing human interactions in RGB videos is a critical task in computer vision,with applications in video surveillance.Existing deep learning-based architectures have achieved strong results,but are computationally intensive,sensitive to video resolution changes and often fail in crowded scenes.We propose a novel hybrid system that is computationally efficient,robust to degraded video quality and able to filter out irrelevant individuals,making it suitable for real-life use.The system leverages multi-modal handcrafted features for interaction representation and a deep learning classifier for capturing complex dependencies.Using Mask R-CNN and YOLO11-Pose,we extract grayscale silhouettes and keypoint coordinates of interacting individuals,while filtering out irrelevant individuals using a proposed algorithm.From these,we extract silhouette-based features(local ternary pattern and histogram of optical flow)and keypoint-based features(distances,angles and velocities)that capture distinct spatial and temporal information.A Bidirectional Long Short-Term Memory network(BiLSTM)then classifies the interactions.Extensive experiments on the UT Interaction,SBU Kinect Interaction and the ISR-UOL 3D social activity datasets demonstrate that our system achieves competitive accuracy.They also validate the effectiveness of the chosen features and classifier,along with the proposed system’s computational efficiency and robustness to occlusion. 展开更多
关键词 human interaction recognition keypoint coordinates grayscale silhouettes bidirectional long shortterm memory network
在线阅读 下载PDF
YOLO-Drive:Robust Driver Distraction Recognition under Fine-Grained and Overlapping Behaviors
5
作者 Zhichao Yu Jiahui Yu +1 位作者 Simon James Fong Yaoyang Wu 《Computers, Materials & Continua》 2026年第5期621-638,共18页
Accurately recognizing driver distraction is critical for preventing traffic accidents,yet current detection models face two persistent challenges.First,distractions are often fine-grained,involving subtle cues such a... Accurately recognizing driver distraction is critical for preventing traffic accidents,yet current detection models face two persistent challenges.First,distractions are often fine-grained,involving subtle cues such as brief eye closures or partial yawns,which are easily missed by conventional detectors.Second,in real-world scenarios,drivers frequently exhibit overlapping behaviors,such as simultaneously holding a cup,closing their eyes,and yawning,leading tomultiple detection boxes and degradedmodel performance.Existing approaches fail to robustly address these complexities,resulting in limited reliability in safety critical applications.To overcome these pain points,we propose YOLO-Drive,a novel framework that enhances YOLO-based driver monitoring with EfficientViM and Polarized Spectral–Spatial Attention(PSSA)modules.Efficient ViMprovides lightweight yet powerful global–local feature extraction,enabling accurate recognition of subtle driver states.PSSA further amplifies discriminative features across spatial and spectral domains,ensuring robust separation of concurrent distraction cues.By explicitly modeling fine-grained and overlapping behaviors,our approach delivers significant improvements in both precision and robustness.Extensive experiments on benchmark driver distraction datasets demonstrate that YOLO-Drive consistently out-performs stateof-the-art models,achieving higher detection accuracy while maintaining real-time efficiency.These results validate YOLO-Drive as a practical and reliable solution for advanced driver monitoring systems,addressing long-standing challenges of subtle cue recognition and multi-cue distraction detection. 展开更多
关键词 Driver distraction recognition attention mechanism fine-grained feature modeling object detection overlapping behavior detection state space model YOLO extensions
在线阅读 下载PDF
Human Activity Recognition Using a CNN with an Enhanced Convolutional Block Attention Module
6
作者 HU Biling TONG Yu 《Wuhan University Journal of Natural Sciences》 2026年第1期10-24,共15页
WiFi-based human activity recognition(HAR)provides a non-intrusive approach for ubiquitous monitoring;however,achieving both high accuracy and robustness simultaneously remains a significant challenge.This paper propo... WiFi-based human activity recognition(HAR)provides a non-intrusive approach for ubiquitous monitoring;however,achieving both high accuracy and robustness simultaneously remains a significant challenge.This paper proposes a Convolutional Neural Network with Enhanced Convolutional Block Attention Module(CNN-ECBAM)framework.The approach systematically converts raw Channel State Information(CSI)into pseudo-color images,effectively preserving essential signal characteristics for deep neural network processing.The core innovation is an Enhanced Convolutional Block Attention Module(ECBAM),tailored to CSI data characteristics,which integrates Efficient Channel Attention(ECA)and Multi-Scale Spatial Attention(MSSA).By employing learnable adaptive fusion weights,it achieves dynamic synergy between channel and spatial features,enabling the network to capture highly discriminative spatiotemporal patterns.The ECBAM module is integrated into a unified Convolutional Neural Network(CNN)to form the overall CNN-ECBAM model.Experimental results on the UT-HAR and NTU-Fi_HAR datasets demonstrate that CNN-ECBAM achieves competitive performance in recognition accuracy and outperforms mainstream baseline models.Specifically,it attains 99.20%accuracy on UT-HAR(surpassing ResNet-18 at 98.60%)and achieves 100%accuracy on NTU-Fi_HAR(exceeding GAF-CNN at 99.62%).These results validate the effectiveness of the proposed method for high-precision and reliable WiFi-based HAR. 展开更多
关键词 human activity recognition deep learning channel state information Enhanced Convolutional Block Attention Module(ECBAM) pseudo-color images
原文传递
Transformer-Driven Multimodal for Human-Object Detection and Recognition for Intelligent Robotic Surveillance
7
作者 Aman Aman Ullah Yanfeng Wu +3 位作者 Shaheryar Najam Nouf Abdullah Almujally Ahmad Jalal Hui Liu 《Computers, Materials & Continua》 2026年第4期1364-1383,共20页
Human object detection and recognition is essential for elderly monitoring and assisted living however,models relying solely on pose or scene context often struggle in cluttered or visually ambiguous settings.To addre... Human object detection and recognition is essential for elderly monitoring and assisted living however,models relying solely on pose or scene context often struggle in cluttered or visually ambiguous settings.To address this,we present SCENET-3D,a transformer-drivenmultimodal framework that unifies human-centric skeleton features with scene-object semantics for intelligent robotic vision through a three-stage pipeline.In the first stage,scene analysis,rich geometric and texture descriptors are extracted from RGB frames,including surface-normal histograms,angles between neighboring normals,Zernike moments,directional standard deviation,and Gabor-filter responses.In the second stage,scene-object analysis,non-human objects are segmented and represented using local feature descriptors and complementary surface-normal information.In the third stage,human-pose estimation,silhouettes are processed through an enhanced MoveNet to obtain 2D anatomical keypoints,which are fused with depth information and converted into RGB-based point clouds to construct pseudo-3D skeletons.Features from all three stages are fused and fed in a transformer encoder with multi-head attention to resolve visually similar activities.Experiments on UCLA(95.8%),ETRI-Activity3D(89.4%),andCAD-120(91.2%)demonstrate that combining pseudo-3D skeletonswith rich scene-object fusion significantly improves generalizable activity recognition,enabling safer elderly care,natural human–robot interaction,and robust context-aware robotic perception in real-world environments. 展开更多
关键词 human object detection elderly care RGB-based pose estimation scene context analysis object recognition Gabor features point cloud reconstruction
在线阅读 下载PDF
A Hybrid Deep Learning Approach for IoT-Enabled Human Activity Recognition and Advanced Analytics
8
作者 Shtwai Alsubai Abdullah Al Hejaili +2 位作者 Najib Ben Aoun Amina Salhi Vincent Karovic 《Computers, Materials & Continua》 2026年第5期1546-1563,共18页
The concept of Human Activity Recognition(HAR)is integral to applications based on Internet of Things(IoT)-enabled devices,particularly in healthcare,fitness tracking,and smart environments.The streams of data from we... The concept of Human Activity Recognition(HAR)is integral to applications based on Internet of Things(IoT)-enabled devices,particularly in healthcare,fitness tracking,and smart environments.The streams of data from wearable sensors are rich in information,yet their high dimensionality and variability pose a significant challenge to proper classification.To address this problem,this paper proposes hybrid architectures that integrate traditional machine learning models with a deep neural network(DNN)to deliver improved performance and enhanced capabilities for HAR tasks.Multi-sensor HAR data were used to systematically test several hybrid models,including:RF+DNN(Random Forest+Deep Neural Network),XGB+DNN(XGBoost+DNN),GB+DNN(Gradient Boosting+DNN),KNN+DNN(K-Nearest Neighbors+DNN),and DT+DNN(Decision Tree+DNN).The RF+DNN model was the most accurate,achieving a 97.03%score with excellent precision,recall,and F1-score.These findings demonstrate that hybrid machine learning and deep learning systems have a promising future in IoT-based HAR applications.The model provides a novel solution for developing smart and trustworthy monitoring systems that support real-time analytics,patient surveillance,and other IoT applications. 展开更多
关键词 human activity recognition(HAR) Internet of Things(IoT) wearable sensors hybrid models deep neural networks(DNN)
在线阅读 下载PDF
Omnidirectional Human Behavior Recognition Method Based on Frequency-Modulated Continuous-Wave Radar
9
作者 SUN Chang WANG Shaohong LIN Yanping 《Journal of Shanghai Jiaotong university(Science)》 2025年第4期637-645,共9页
Frequency-modulated continuous-wave radar enables the non-contact and privacy-preserving recognition of human behavior.However,the accuracy of behavior recognition is directly influenced by the spatial relationship be... Frequency-modulated continuous-wave radar enables the non-contact and privacy-preserving recognition of human behavior.However,the accuracy of behavior recognition is directly influenced by the spatial relationship between human posture and the radar.To address the issue of low accuracy in behavior recognition when the human body is not directly facing the radar,a method combining local outlier factor with Doppler information is proposed for the correction of multi-classifier recognition results.Initially,the information such as distance,velocity,and micro-Doppler spectrogram of the target is obtained using the fast Fourier transform and histogram of oriented gradients-support vector machine methods,followed by preliminary recognition.Subsequently,Platt scaling is employed to transform recognition results into confidence scores,and finally,the Doppler-local outlier factor method is utilized to calibrate the confidence scores,with the highest confidence classifier result considered as the recognition outcome.Experimental results demonstrate that this approach achieves an average recognition accuracy of 96.23%for comprehensive human behavior recognition in various orientations. 展开更多
关键词 frequency-modulated continuous-wave radar omnidirectional human behavior recognition histogram of oriented gradients support vector machine micro-Doppler spectrogram Doppler-local outlier factor
原文传递
Study on Local Optical Flow Method Based on YOLOv3 in Human Behavior Recognition 被引量:3
10
作者 Hao Zheng Jianfang Liu Mengyi Liao 《Journal of Computer and Communications》 2021年第1期10-18,共9页
In the process of human behavior recognition, the traditional dense optical flow method has too many pixels and too much overhead, which limits the running speed. This paper proposed a method combing YOLOv3 (You Only ... In the process of human behavior recognition, the traditional dense optical flow method has too many pixels and too much overhead, which limits the running speed. This paper proposed a method combing YOLOv3 (You Only Look Once v3) and local optical flow method. Based on the dense optical flow method, the optical flow modulus of the area where the human target is detected is calculated to reduce the amount of computation and save the cost in terms of time. And then, a threshold value is set to complete the human behavior identification. Through design algorithm, experimental verification and other steps, the walking, running and falling state of human body in real life indoor sports video was identified. Experimental results show that this algorithm is more advantageous for jogging behavior recognition. 展开更多
关键词 YOLOv3 Local Optical Flow Method human behavior recognition
在线阅读 下载PDF
IoT-Based Real-Time Medical-Related Human Activity Recognition Using Skeletons and Multi-Stage Deep Learning for Healthcare 被引量:1
11
作者 Subrata Kumer Paul Abu Saleh Musa Miah +3 位作者 Rakhi Rani Paul Md.EkramulHamid Jungpil Shin Md Abdur Rahim 《Computers, Materials & Continua》 2025年第8期2513-2530,共18页
The Internet of Things(IoT)and mobile technology have significantly transformed healthcare by enabling real-time monitoring and diagnosis of patients.Recognizing Medical-Related Human Activities(MRHA)is pivotal for he... The Internet of Things(IoT)and mobile technology have significantly transformed healthcare by enabling real-time monitoring and diagnosis of patients.Recognizing Medical-Related Human Activities(MRHA)is pivotal for healthcare systems,particularly for identifying actions critical to patient well-being.However,challenges such as high computational demands,low accuracy,and limited adaptability persist in Human Motion Recognition(HMR).While some studies have integrated HMR with IoT for real-time healthcare applications,limited research has focused on recognizing MRHA as essential for effective patient monitoring.This study proposes a novel HMR method tailored for MRHA detection,leveraging multi-stage deep learning techniques integrated with IoT.The approach employs EfficientNet to extract optimized spatial features from skeleton frame sequences using seven Mobile Inverted Bottleneck Convolutions(MBConv)blocks,followed by Convolutional Long Short Term Memory(ConvLSTM)to capture spatio-temporal patterns.A classification module with global average pooling,a fully connected layer,and a dropout layer generates the final predictions.The model is evaluated on the NTU RGB+D 120 and HMDB51 datasets,focusing on MRHA such as sneezing,falling,walking,sitting,etc.It achieves 94.85%accuracy for cross-subject evaluations and 96.45%for cross-view evaluations on NTU RGB+D 120,along with 89.22%accuracy on HMDB51.Additionally,the system integrates IoT capabilities using a Raspberry Pi and GSM module,delivering real-time alerts via Twilios SMS service to caregivers and patients.This scalable and efficient solution bridges the gap between HMR and IoT,advancing patient monitoring,improving healthcare outcomes,and reducing costs. 展开更多
关键词 Real-time human motion recognition(HMR) ENConvLSTM EfficientNet ConvLSTM skeleton data NTU RGB+D 120 dataset MRHA
在线阅读 下载PDF
Video action recognition meets vision-language models exploring human factors in scene interaction: a review
12
作者 GUO Yuping GAO Hongwei +3 位作者 YU Jiahui GE Jinchao HAN Meng JU Zhaojie 《Optoelectronics Letters》 2025年第10期626-640,共15页
Video action recognition(VAR)aims to analyze dynamic behaviors in videos and achieve semantic understanding.VAR faces challenges such as temporal dynamics,action-scene coupling,and the complexity of human interactions... Video action recognition(VAR)aims to analyze dynamic behaviors in videos and achieve semantic understanding.VAR faces challenges such as temporal dynamics,action-scene coupling,and the complexity of human interactions.Existing methods can be categorized into motion-level,event-level,and story-level ones based on spatiotemporal granularity.However,single-modal approaches struggle to capture complex behavioral semantics and human factors.Therefore,in recent years,vision-language models(VLMs)have been introduced into this field,providing new research perspectives for VAR.In this paper,we systematically review spatiotemporal hierarchical methods in VAR and explore how the introduction of large models has advanced the field.Additionally,we propose the concept of“Factor”to identify and integrate key information from both visual and textual modalities,enhancing multimodal alignment.We also summarize various multimodal alignment methods and provide in-depth analysis and insights into future research directions. 展开更多
关键词 human factors video action recognition vision language models analyze dynamic behaviors spatiotemporal granularity video action recognition var aims multimodal alignment scene interaction
原文传递
Design of Fish School Behavior Pattern Recognition Model SPD-YOLOv10n
13
作者 Hanlin XU Shiyu WU Guochao DING 《Agricultural Biotechnology》 2025年第1期77-79,共3页
A common but flawed design in existing CNN architectures is using strided convolutions and/or pooling layer,which will result in the loss of fine-grained feature information,especially for low-resolution images and sm... A common but flawed design in existing CNN architectures is using strided convolutions and/or pooling layer,which will result in the loss of fine-grained feature information,especially for low-resolution images and small objects.In this paper,a new CNN building block named SPD-Conv was used,which completely eliminated stride and pooling operations and replaced them with a space-to-depth convolution and a non-strided convolution.Such new design has the advantage of downsampling feature maps while retaining discriminant feature information.It also represents a general unified method,which can be easily applied to any CNN architectures,and can also be applied to strided conversion and pooling in the same way. 展开更多
关键词 FISH Group behavior behavior recognition Deep learning YOLOv10
在线阅读 下载PDF
Dynamic behavior recognition in aerial deployment of multi-segmented foldable-wing drones using variational autoencoders
14
作者 Yilin DOU Zhou ZHOU Rui WANG 《Chinese Journal of Aeronautics》 2025年第6期143-165,共23页
The aerial deployment method enables Unmanned Aerial Vehicles(UAVs)to be directly positioned at the required altitude for their mission.This method typically employs folding technology to improve loading efficiency,wi... The aerial deployment method enables Unmanned Aerial Vehicles(UAVs)to be directly positioned at the required altitude for their mission.This method typically employs folding technology to improve loading efficiency,with applications such as the gravity-only aerial deployment of high-aspect-ratio solar-powered UAVs,and aerial takeoff of fixed-wing drones in Mars research.However,the significant morphological changes during deployment are accompanied by strong nonlinear dynamic aerodynamic forces,which result in multiple degrees of freedom and an unstable character.This hinders the description and analysis of unknown dynamic behaviors,further leading to difficulties in the design of deployment strategies and flight control.To address this issue,this paper proposes an analysis method for dynamic behaviors during aerial deployment based on the Variational Autoencoder(VAE).Focusing on the gravity-only deployment problem of highaspect-ratio foldable-wing UAVs,the method encodes the multi-degree-of-freedom unstable motion signals into a low-dimensional feature space through a data-driven approach.By clustering in the feature space,this paper identifies and studies several dynamic behaviors during aerial deployment.The research presented in this paper offers a new method and perspective for feature extraction and analysis of complex and difficult-to-describe extreme flight dynamics,guiding the research on aerial deployment drones design and control strategies. 展开更多
关键词 Dynamic behavior recognition Aerial deployment technology Variational autoencoder Pattern recognition Multi-rigid-bodydynamics
原文传递
Research on Multi-target Cow Behavior Recognition Method Based on Deep Learning
15
作者 Jizhen WU Jianfei SHI Zhiyuan JING 《Agricultural Biotechnology》 2025年第6期36-39,共4页
To address the issue of low recognition accuracy for eight types of behaviors including standing,walking,drinking,lying,eating,mounting,fighting and limping in complex multi-cow farm environments,a multi-target cow be... To address the issue of low recognition accuracy for eight types of behaviors including standing,walking,drinking,lying,eating,mounting,fighting and limping in complex multi-cow farm environments,a multi-target cow behavior recognition method based on an improved YOLOv11n algorithm was proposed.The detection capability for small targets in images was enhanced by incorporating a DASI module into the backbone network and a MDCR module into the neck network,based on YOLOv11.The improved YOLOv11 algorithm increased the mean average precision from the original 89.5%to 93%,with particularly notable improvements of 8.7%and 6.3%in the average precision for recognizing drinking and walking behaviors,respectively.These results fully demonstrate that the proposed method enhances the model s ability to recognize cow behaviors. 展开更多
关键词 Image recognition YOLOv11n Cow behavior recognition Deep learning
在线阅读 下载PDF
Knowledge,attitude,and behavior of human papillomavirus vaccination among undergraduate students in Shanghai,China
16
作者 Rong-Rong Zhang Yan Xu 《Frontiers of Nursing》 2025年第1期47-56,共10页
Objective:This study aimed to explore undergraduates’knowledge,attitude,and practice/behavior of human papillomavirus(HPV)vaccination,as well as the essential influencing factors for vaccination decision-making.Metho... Objective:This study aimed to explore undergraduates’knowledge,attitude,and practice/behavior of human papillomavirus(HPV)vaccination,as well as the essential influencing factors for vaccination decision-making.Methods:Through cluster and convenience sampling,2000 undergraduates from the Nursing and Language department of a university in Shanghai were sent a self-designed questionnaire.Chi-square tests,independent sample t-test/ANOVE,and multiple linear regression were used to investigate participants’knowledge and attitude on HPV vaccination,as well as the factors that predicted potential action to receive and promote HPV vaccination in the future.Results:The mean HPV knowledge score was 5.027 out of 10.Health science students showed a significantly higher knowledge mean score than the non-health science students(P<0.000).There was a statistically difference in HPV vaccination attitude among undergraduates in different grades(P<0.05).Awareness of cervical cancer and worries about the risk of cervical cancer were the significant predictors of willingness to receive and promote HPV vaccination in the future.Conclusions:It would take time for a new health product to be aware,understood,accepted,and received.Education providing and information sharing are expected to break the dawn and make the procedure processed. 展开更多
关键词 human papillomavirus UNDERGRADUATE VACCINATION KNOWLEDGE ATTITUDE behavior
暂未订购
ideo-Based Human Activity Recognition Using Hybrid Deep Learning Model
17
作者 Jungpil Shin Md.Al Mehedi Hasan +2 位作者 Md.Maniruzzaman Satoshi Nishimura Sultan Alfarhood 《Computer Modeling in Engineering & Sciences》 2025年第6期3615-3638,共24页
Activity recognition is a challenging topic in the field of computer vision that has various applications,including surveillance systems,industrial automation,and human-computer interaction.Today,the demand for automa... Activity recognition is a challenging topic in the field of computer vision that has various applications,including surveillance systems,industrial automation,and human-computer interaction.Today,the demand for automation has greatly increased across industries worldwide.Real-time detection requires edge devices with limited computational time.This study proposes a novel hybrid deep learning system for human activity recognition(HAR),aiming to enhance the recognition accuracy and reduce the computational time.The proposed system combines a pretrained image classification model with a sequence analysis model.First,the dataset was divided into a training set(70%),validation set(10%),and test set(20%).Second,all the videos were converted into frames and deep-based features were extracted from each frame using convolutional neural networks(CNNs)with a vision transformer.Following that,bidirectional long short-term memory(BiLSTM)-and temporal convolutional network(TCN)-based models were trained using the training set,and their performances were evaluated using the validation set and test set.Four benchmark datasets(UCF11,UCF50,UCF101,and JHMDB)were used to evaluate the performance of the proposed HAR-based system.The experimental results showed that the combination of ConvNeXt and the TCN-based model achieved a recognition accuracy of 97.73%for UCF11,98.81%for UCF50,98.46%for UCF101,and 83.38%for JHMDB,respectively.This represents improvements in the recognition accuracy of 4%,2.67%,3.67%,and 7.08%for the UCF11,UCF50,UCF101,and JHMDB datasets,respectively,over existing models.Moreover,the proposed HAR-based system obtained superior recognition accuracy,shorter computational times,and minimal memory usage compared to the existing models. 展开更多
关键词 human activity recognition BiLSTM ConvNeXt temporal convolutional network deep learning
在线阅读 下载PDF
A Novel Attention-Based Parallel Blocks Deep Architecture for Human Action Recognition
18
作者 Yasir Khan Jadoon Yasir Noman Khalid +4 位作者 Muhammad Attique Khan Jungpil Shin Fatimah Alhayan Hee-Chan Cho Byoungchol Chang 《Computer Modeling in Engineering & Sciences》 2025年第7期1143-1164,共22页
Real-time surveillance is attributed to recognizing the variety of actions performed by humans.Human Action Recognition(HAR)is a technique that recognizes human actions from a video stream.A range of variations in hum... Real-time surveillance is attributed to recognizing the variety of actions performed by humans.Human Action Recognition(HAR)is a technique that recognizes human actions from a video stream.A range of variations in human actions makes it difficult to recognize with considerable accuracy.This paper presents a novel deep neural network architecture called Attention RB-Net for HAR using video frames.The input is provided to the model in the form of video frames.The proposed deep architecture is based on the unique structuring of residual blocks with several filter sizes.Features are extracted from each frame via several operations with specific parameters defined in the presented novel Attention-based Residual Bottleneck(Attention-RB)DCNN architecture.A fully connected layer receives an attention-based features matrix,and final classification is performed.Several hyperparameters of the proposed model are initialized using Bayesian Optimization(BO)and later utilized in the trained model for testing.In testing,features are extracted from the self-attention layer and passed to neural network classifiers for the final action classification.Two highly cited datasets,HMDB51 and UCF101,were used to validate the proposed architecture and obtained an average accuracy of 87.70%and 97.30%,respectively.The deep convolutional neural network(DCNN)architecture is compared with state-of-the-art(SOTA)methods,including pre-trained models,inside blocks,and recently published techniques,and performs better. 展开更多
关键词 human action recognition self-attention video streams residual bottleneck classification neural networks
在线阅读 下载PDF
A YOLOv11-Based Deep Learning Framework for Multi-Class Human Action Recognition
19
作者 Nayeemul Islam Nayeem Shirin Mahbuba +4 位作者 Sanjida Islam Disha Md Rifat Hossain Buiyan Shakila Rahman M.Abdullah-Al-Wadud Jia Uddin 《Computers, Materials & Continua》 2025年第10期1541-1557,共17页
Human activity recognition is a significant area of research in artificial intelligence for surveillance,healthcare,sports,and human-computer interaction applications.The article benchmarks the performance of You Only... Human activity recognition is a significant area of research in artificial intelligence for surveillance,healthcare,sports,and human-computer interaction applications.The article benchmarks the performance of You Only Look Once version 11-based(YOLOv11-based)architecture for multi-class human activity recognition.The article benchmarks the performance of You Only Look Once version 11-based(YOLOv11-based)architecture for multi-class human activity recognition.The dataset consists of 14,186 images across 19 activity classes,from dynamic activities such as running and swimming to static activities such as sitting and sleeping.Preprocessing included resizing all images to 512512 pixels,annotating them in YOLO’s bounding box format,and applying data augmentation methods such as flipping,rotation,and cropping to enhance model generalization.The proposed model was trained for 100 epochs with adaptive learning rate methods and hyperparameter optimization for performance improvement,with a mAP@0.5 of 74.93%and a mAP@0.5-0.95 of 64.11%,outperforming previous versions of YOLO(v10,v9,and v8)and general-purpose architectures like ResNet50 and EfficientNet.It exhibited improved precision and recall for all activity classes with high precision values of 0.76 for running,0.79 for swimming,0.80 for sitting,and 0.81 for sleeping,and was tested for real-time deployment with an inference time of 8.9 ms per image,being computationally light.Proposed YOLOv11’s improvements are attributed to architectural advancements like a more complex feature extraction process,better attention modules,and an anchor-free detection mechanism.While YOLOv10 was extremely stable in static activity recognition,YOLOv9 performed well in dynamic environments but suffered from overfitting,and YOLOv8,while being a decent baseline,failed to differentiate between overlapping static activities.The experimental results determine proposed YOLOv11 to be the most appropriate model,providing an ideal balance between accuracy,computational efficiency,and robustness for real-world deployment.Nevertheless,there exist certain issues to be addressed,particularly in discriminating against visually similar activities and the use of publicly available datasets.Future research will entail the inclusion of 3D data and multimodal sensor inputs,such as depth and motion information,for enhancing recognition accuracy and generalizability to challenging real-world environments. 展开更多
关键词 human activity recognition YOLOv11 deep learning real-time detection anchor-free detection attention mechanisms object detection image classification multi-class recognition surveillance applications
在线阅读 下载PDF
Investigation of dynamic behavior of human skin tissue under micro-particle impact in the context of transdermal drug delivery:numerical and analytical perspectives
20
作者 Jianbo Shen Jiacai Huang +1 位作者 Yaoke Wen Sebastien Roth 《Theoretical & Applied Mechanics Letters》 2025年第2期188-193,共6页
The understanding of the impact of high-velocity microparticles on human skin tissue is important for the ad-ministration of drugs during transdermal drug delivery.This paper aims to numerically investigate the dynami... The understanding of the impact of high-velocity microparticles on human skin tissue is important for the ad-ministration of drugs during transdermal drug delivery.This paper aims to numerically investigate the dynamic behavior of human skin tissue under micro-particle impact in transdermal drug delivery.The numerical model was developed based on a coupled smoothed particle hydrodynamics(SPH)and FEM method via commercial FE software RADIOSS.Analytical analysis was conducted applying the Poncelet model and was used as validation data.A hyperelastic one-term Ogden model with one pair of material parameters(μ,α)was implemented for the skin tissue.Sensitivity studies reveal that the effect of parameter α on the penetration process is much more significant than μ.Numerical results correlate well with the analytical curves with various particle diameters and impact velocities,its capability of predicting the penetration process of micro-particle impacts into skin tissues.This work can be further investigated to guide the design of transdermal drug delivery equipment. 展开更多
关键词 Micro-particle impact Dynamic behavior human skin tissue Smoothed particle hydrodynamics FEM penetration
暂未订购
上一页 1 2 250 下一页 到第
使用帮助 返回顶部