The Internet of Things(IoT)and mobile technology have significantly transformed healthcare by enabling real-time monitoring and diagnosis of patients.Recognizing Medical-Related Human Activities(MRHA)is pivotal for he...The Internet of Things(IoT)and mobile technology have significantly transformed healthcare by enabling real-time monitoring and diagnosis of patients.Recognizing Medical-Related Human Activities(MRHA)is pivotal for healthcare systems,particularly for identifying actions critical to patient well-being.However,challenges such as high computational demands,low accuracy,and limited adaptability persist in Human Motion Recognition(HMR).While some studies have integrated HMR with IoT for real-time healthcare applications,limited research has focused on recognizing MRHA as essential for effective patient monitoring.This study proposes a novel HMR method tailored for MRHA detection,leveraging multi-stage deep learning techniques integrated with IoT.The approach employs EfficientNet to extract optimized spatial features from skeleton frame sequences using seven Mobile Inverted Bottleneck Convolutions(MBConv)blocks,followed by Convolutional Long Short Term Memory(ConvLSTM)to capture spatio-temporal patterns.A classification module with global average pooling,a fully connected layer,and a dropout layer generates the final predictions.The model is evaluated on the NTU RGB+D 120 and HMDB51 datasets,focusing on MRHA such as sneezing,falling,walking,sitting,etc.It achieves 94.85%accuracy for cross-subject evaluations and 96.45%for cross-view evaluations on NTU RGB+D 120,along with 89.22%accuracy on HMDB51.Additionally,the system integrates IoT capabilities using a Raspberry Pi and GSM module,delivering real-time alerts via Twilios SMS service to caregivers and patients.This scalable and efficient solution bridges the gap between HMR and IoT,advancing patient monitoring,improving healthcare outcomes,and reducing costs.展开更多
Advanced artificial intelligence technologies such as ChatGPT and other large language models(LLMs)have significantly impacted fields such as education and research in recent years.ChatGPT benefits students and educat...Advanced artificial intelligence technologies such as ChatGPT and other large language models(LLMs)have significantly impacted fields such as education and research in recent years.ChatGPT benefits students and educators by providing personalized feedback,facilitating interactive learning,and introducing innovative teaching methods.While many researchers have studied ChatGPT across various subject domains,few analyses have focused on the engineering domain,particularly in addressing the risks of academic dishonesty and potential declines in critical thinking skills.To address this gap,this study explores both the opportunities and limitations of ChatGPT in engineering contexts through a two-part analysis.First,we conducted experiments with ChatGPT to assess its effectiveness in tasks such as code generation,error checking,and solution optimization.Second,we surveyed 125 users,predominantly engineering students,to analyze ChatGPTs role in academic support.Our findings reveal that 93.60%of respondents use ChatGPT for quick academic answers,particularly among early-stage university students,and that 84.00%find it helpful for sourcing research materials.The study also highlights ChatGPT’s strengths in programming assistance,with 84.80%of users utilizing it for debugging and 86.40%for solving coding problems.However,limitations persist,with many users reporting inaccuracies in mathematical solutions and occasional false citations.Furthermore,the reliance on the free version by 96%of users underscores its accessibility but also suggests limitations in resource availability.This work provides key insights into ChatGPT’s strengths and limitations,establishing a framework for responsible AI use in education.Highlighting areas for improvement marks a milestone in understanding and optimizing AI’s role in academia for sustainable future use.展开更多
Communication between people with disabilities and people who do not understand sign language is a growing social need and can be a tedious task.One of the main functions of sign language is to communicate with each o...Communication between people with disabilities and people who do not understand sign language is a growing social need and can be a tedious task.One of the main functions of sign language is to communicate with each other through hand gestures.Recognition of hand gestures has become an important challenge for the recognition of sign language.There are many existing models that can produce a good accuracy,but if the model test with rotated or translated images,they may face some difficulties to make good performance accuracy.To resolve these challenges of hand gesture recognition,we proposed a Rotation,Translation and Scale-invariant sign word recognition system using a convolu-tional neural network(CNN).We have followed three steps in our work:rotated,translated and scaled(RTS)version dataset generation,gesture segmentation,and sign word classification.Firstly,we have enlarged a benchmark dataset of 20 sign words by making different amounts of Rotation,Translation and Scale of the ori-ginal images to create the RTS version dataset.Then we have applied the gesture segmentation technique.The segmentation consists of three levels,i)Otsu Thresholding with YCbCr,ii)Morphological analysis:dilation through opening morphology and iii)Watershed algorithm.Finally,our designed CNN model has been trained to classify the hand gesture as well as the sign word.Our model has been evaluated using the twenty sign word dataset,five sign word dataset and the RTS version of these datasets.We achieved 99.30%accuracy from the twenty sign word dataset evaluation,99.10%accuracy from the RTS version of the twenty sign word evolution,100%accuracy from thefive sign word dataset evaluation,and 98.00%accuracy from the RTS versionfive sign word dataset evolution.Furthermore,the influence of our model exists in competitive results with state-of-the-art methods in sign word recognition.展开更多
Sesame(Sesamum indicum L.)is an ancient oilseed crop of the Pedaliaceae family with high oil content and potential health benefits.SHI RELATED SEQUENCE(SRS)proteins are the transcription factors(TFs)specific to plants...Sesame(Sesamum indicum L.)is an ancient oilseed crop of the Pedaliaceae family with high oil content and potential health benefits.SHI RELATED SEQUENCE(SRS)proteins are the transcription factors(TFs)specific to plants that contain RING-like zinc finger domain and are associated with the regulation of several physiological and biochemical processes.They also play vital roles in plant growth and development such as root formation,leaf development,floral development,hormone biosynthesis,signal transduction,and biotic and abiotic stress responses.Nevertheless,the SRS gene family was not reported in sesame yet.In this study,identification,molecular characterization,phylogenetic relationship,cis-acting regulatory elements,protein-protein interaction,syntenic relationship,duplication events and expression pattern of SRS genes were analyzed in S.indicum.We identified total six SiSRS genes on seven different linkage groups in the S.indicum genome by comparing with the other species,including the model plant Arabidopsis thaliana.The SiSRS genes showed variation in their structure like2–5 exons and 1–4 introns.Like other species,SiSRS proteins also contained‘RING-like zinc finger'and‘LRP1'domains.Then,the SiSRS genes were clustered into subclasses via phylogenetic analysis with proteins of S.indicum,A.thaliana,and some other plant species.The cis-acting regulatory elements analysis revealed that the promoter region of SiSRS4(SIN_1011561)showed the highest 13 and 16 elements for light-and phytohormone-responses whereas,SiSRS1(SIN_1015187)showed the highest 15 elements for stress-response.The ABREs,or ABA-responsive elements,were found in a maximum of 8 copies in the SiSRS3(SIN 1009100).Moreover,the available RNA-seq based expression of SiSRS genes revealed variation in expression patterns between stress-treated and non-treated samples,especially in drought and salinity conditions in.S.indicum.Two SiSRS genes like SiSRS1(SIN_1015187)and SiSRS5(SIN_1021065),also exhibited variable expression patterns between control vs PEG-treated sesame root samples and three SiSRS genes,including SiSRS1(SIN_1015187),SiSRS2(SIN_1003328)and SiSRS5(SIN_1021065)were responsive to salinity treatments.The present outcomes will encourage more research into the gene expression and functionality analysis of SiSRS genes in S.indicum and other related species.展开更多
In the digital age,non-touch communication technologies are reshaping human-device interactions and raising security concerns.A major challenge in current technology is the misinterpretation of gestures by sensors and...In the digital age,non-touch communication technologies are reshaping human-device interactions and raising security concerns.A major challenge in current technology is the misinterpretation of gestures by sensors and cameras,often caused by environmental factors.This issue has spurred the need for advanced data processing methods to achieve more accurate gesture recognition and predictions.Our study presents a novel virtual keyboard allowing character input via distinct hand gestures,focusing on two key aspects:hand gesture recognition and character input mechanisms.We developed a novel model with LSTM and fully connected layers for enhanced sequential data processing and hand gesture recognition.We also integrated CNN,max-pooling,and dropout layers for improved spatial feature extraction.This model architecture processes both temporal and spatial aspects of hand gestures,using LSTM to extract complex patterns from frame sequences for a comprehensive understanding of input data.Our unique dataset,essential for training the model,includes 1,662 landmarks from dynamic hand gestures,33 postures,and 468 face landmarks,all captured in real-time using advanced pose estimation.The model demonstrated high accuracy,achieving 98.52%in hand gesture recognition and over 97%in character input across different scenarios.Its excellent performance in real-time testing underlines its practicality and effectiveness,marking a significant advancement in enhancing human-device interactions in the digital age.展开更多
基金funded by the ICT Division of theMinistry of Posts,Telecommunications,and Information Technology of Bangladesh under Grant Number 56.00.0000.052.33.005.21-7(Tracking No.22FS15306)support from the University of Rajshahi.
文摘The Internet of Things(IoT)and mobile technology have significantly transformed healthcare by enabling real-time monitoring and diagnosis of patients.Recognizing Medical-Related Human Activities(MRHA)is pivotal for healthcare systems,particularly for identifying actions critical to patient well-being.However,challenges such as high computational demands,low accuracy,and limited adaptability persist in Human Motion Recognition(HMR).While some studies have integrated HMR with IoT for real-time healthcare applications,limited research has focused on recognizing MRHA as essential for effective patient monitoring.This study proposes a novel HMR method tailored for MRHA detection,leveraging multi-stage deep learning techniques integrated with IoT.The approach employs EfficientNet to extract optimized spatial features from skeleton frame sequences using seven Mobile Inverted Bottleneck Convolutions(MBConv)blocks,followed by Convolutional Long Short Term Memory(ConvLSTM)to capture spatio-temporal patterns.A classification module with global average pooling,a fully connected layer,and a dropout layer generates the final predictions.The model is evaluated on the NTU RGB+D 120 and HMDB51 datasets,focusing on MRHA such as sneezing,falling,walking,sitting,etc.It achieves 94.85%accuracy for cross-subject evaluations and 96.45%for cross-view evaluations on NTU RGB+D 120,along with 89.22%accuracy on HMDB51.Additionally,the system integrates IoT capabilities using a Raspberry Pi and GSM module,delivering real-time alerts via Twilios SMS service to caregivers and patients.This scalable and efficient solution bridges the gap between HMR and IoT,advancing patient monitoring,improving healthcare outcomes,and reducing costs.
基金supported by Competitive Research by the University of Aizu.
文摘Advanced artificial intelligence technologies such as ChatGPT and other large language models(LLMs)have significantly impacted fields such as education and research in recent years.ChatGPT benefits students and educators by providing personalized feedback,facilitating interactive learning,and introducing innovative teaching methods.While many researchers have studied ChatGPT across various subject domains,few analyses have focused on the engineering domain,particularly in addressing the risks of academic dishonesty and potential declines in critical thinking skills.To address this gap,this study explores both the opportunities and limitations of ChatGPT in engineering contexts through a two-part analysis.First,we conducted experiments with ChatGPT to assess its effectiveness in tasks such as code generation,error checking,and solution optimization.Second,we surveyed 125 users,predominantly engineering students,to analyze ChatGPTs role in academic support.Our findings reveal that 93.60%of respondents use ChatGPT for quick academic answers,particularly among early-stage university students,and that 84.00%find it helpful for sourcing research materials.The study also highlights ChatGPT’s strengths in programming assistance,with 84.80%of users utilizing it for debugging and 86.40%for solving coding problems.However,limitations persist,with many users reporting inaccuracies in mathematical solutions and occasional false citations.Furthermore,the reliance on the free version by 96%of users underscores its accessibility but also suggests limitations in resource availability.This work provides key insights into ChatGPT’s strengths and limitations,establishing a framework for responsible AI use in education.Highlighting areas for improvement marks a milestone in understanding and optimizing AI’s role in academia for sustainable future use.
基金This work was supported by the Competitive Research Fund of The University of Aizu,Japan.
文摘Communication between people with disabilities and people who do not understand sign language is a growing social need and can be a tedious task.One of the main functions of sign language is to communicate with each other through hand gestures.Recognition of hand gestures has become an important challenge for the recognition of sign language.There are many existing models that can produce a good accuracy,but if the model test with rotated or translated images,they may face some difficulties to make good performance accuracy.To resolve these challenges of hand gesture recognition,we proposed a Rotation,Translation and Scale-invariant sign word recognition system using a convolu-tional neural network(CNN).We have followed three steps in our work:rotated,translated and scaled(RTS)version dataset generation,gesture segmentation,and sign word classification.Firstly,we have enlarged a benchmark dataset of 20 sign words by making different amounts of Rotation,Translation and Scale of the ori-ginal images to create the RTS version dataset.Then we have applied the gesture segmentation technique.The segmentation consists of three levels,i)Otsu Thresholding with YCbCr,ii)Morphological analysis:dilation through opening morphology and iii)Watershed algorithm.Finally,our designed CNN model has been trained to classify the hand gesture as well as the sign word.Our model has been evaluated using the twenty sign word dataset,five sign word dataset and the RTS version of these datasets.We achieved 99.30%accuracy from the twenty sign word dataset evaluation,99.10%accuracy from the RTS version of the twenty sign word evolution,100%accuracy from thefive sign word dataset evaluation,and 98.00%accuracy from the RTS versionfive sign word dataset evolution.Furthermore,the influence of our model exists in competitive results with state-of-the-art methods in sign word recognition.
文摘Sesame(Sesamum indicum L.)is an ancient oilseed crop of the Pedaliaceae family with high oil content and potential health benefits.SHI RELATED SEQUENCE(SRS)proteins are the transcription factors(TFs)specific to plants that contain RING-like zinc finger domain and are associated with the regulation of several physiological and biochemical processes.They also play vital roles in plant growth and development such as root formation,leaf development,floral development,hormone biosynthesis,signal transduction,and biotic and abiotic stress responses.Nevertheless,the SRS gene family was not reported in sesame yet.In this study,identification,molecular characterization,phylogenetic relationship,cis-acting regulatory elements,protein-protein interaction,syntenic relationship,duplication events and expression pattern of SRS genes were analyzed in S.indicum.We identified total six SiSRS genes on seven different linkage groups in the S.indicum genome by comparing with the other species,including the model plant Arabidopsis thaliana.The SiSRS genes showed variation in their structure like2–5 exons and 1–4 introns.Like other species,SiSRS proteins also contained‘RING-like zinc finger'and‘LRP1'domains.Then,the SiSRS genes were clustered into subclasses via phylogenetic analysis with proteins of S.indicum,A.thaliana,and some other plant species.The cis-acting regulatory elements analysis revealed that the promoter region of SiSRS4(SIN_1011561)showed the highest 13 and 16 elements for light-and phytohormone-responses whereas,SiSRS1(SIN_1015187)showed the highest 15 elements for stress-response.The ABREs,or ABA-responsive elements,were found in a maximum of 8 copies in the SiSRS3(SIN 1009100).Moreover,the available RNA-seq based expression of SiSRS genes revealed variation in expression patterns between stress-treated and non-treated samples,especially in drought and salinity conditions in.S.indicum.Two SiSRS genes like SiSRS1(SIN_1015187)and SiSRS5(SIN_1021065),also exhibited variable expression patterns between control vs PEG-treated sesame root samples and three SiSRS genes,including SiSRS1(SIN_1015187),SiSRS2(SIN_1003328)and SiSRS5(SIN_1021065)were responsive to salinity treatments.The present outcomes will encourage more research into the gene expression and functionality analysis of SiSRS genes in S.indicum and other related species.
文摘In the digital age,non-touch communication technologies are reshaping human-device interactions and raising security concerns.A major challenge in current technology is the misinterpretation of gestures by sensors and cameras,often caused by environmental factors.This issue has spurred the need for advanced data processing methods to achieve more accurate gesture recognition and predictions.Our study presents a novel virtual keyboard allowing character input via distinct hand gestures,focusing on two key aspects:hand gesture recognition and character input mechanisms.We developed a novel model with LSTM and fully connected layers for enhanced sequential data processing and hand gesture recognition.We also integrated CNN,max-pooling,and dropout layers for improved spatial feature extraction.This model architecture processes both temporal and spatial aspects of hand gestures,using LSTM to extract complex patterns from frame sequences for a comprehensive understanding of input data.Our unique dataset,essential for training the model,includes 1,662 landmarks from dynamic hand gestures,33 postures,and 468 face landmarks,all captured in real-time using advanced pose estimation.The model demonstrated high accuracy,achieving 98.52%in hand gesture recognition and over 97%in character input across different scenarios.Its excellent performance in real-time testing underlines its practicality and effectiveness,marking a significant advancement in enhancing human-device interactions in the digital age.