期刊文献+
共找到3篇文章
< 1 >
每页显示 20 50 100
Facial Expression Generation from Text with FaceCLIP
1
作者 Wen-Wen Fu Wen-Juan Gong +2 位作者 Chen-Yang Yu Wei Wang Jordi Gonzàlez 《Journal of Computer Science & Technology》 2025年第2期359-377,共19页
Facial expression generation from pure textual descriptions is widely applied in human-computer interaction,computer-aided design,assisted education,etc.However,this task is challenging due to the intricate facial str... Facial expression generation from pure textual descriptions is widely applied in human-computer interaction,computer-aided design,assisted education,etc.However,this task is challenging due to the intricate facial structure and the complex mapping between texts and images.Existing methods face limitations in generating high-resolution images or capturing diverse facial expressions.In this study,we propose a novel generation approach,named FaceCLIP,to tackle these problems.The proposed method utilizes a CLIP-based multi-stage generative adversarial model to produce vivid facial expressions with high resolutions.With strong semantic priors from multi-modal textual and visual cues,the proposed method effectively disentangles facial attributes,enabling attribute editing and semantic reasoning.To facilitate text-toexpression generation,we build a new dataset called the FET dataset,which contains facial expression images and corresponding textual descriptions.Experiments on the dataset demonstrate improved image quality and semantic consistency compared with state-of-the-art methods. 展开更多
关键词 facial expression generation contrastive language-image pre-training(CLIP) MULTI-STAGE generative adversarial network(GAN)
原文传递
Hyperparameter optimization of YOLOv8 for smoke and wildfire detection:Implications for agricultural and environmental safety
2
作者 Leo Ramos Edmundo Casas +2 位作者 Eduardo Bendek Cristian Romero Francklin Rivas-Echeverría 《Artificial Intelligence in Agriculture》 2024年第2期109-126,共18页
In this study,we extensively evaluated the viability of the state-of-the-art YOLOv8 architecture for object detec-tion tasks,specifically tailored for smoke and wildfire identification with a focus on agricultural and... In this study,we extensively evaluated the viability of the state-of-the-art YOLOv8 architecture for object detec-tion tasks,specifically tailored for smoke and wildfire identification with a focus on agricultural and environmen-tal safety.All available versions of YOLOv8 were initially fine-tuned on a domain-specific dataset that included a variety of scenarios,crucial for comprehensive agricultural monitoring.The‘large’version(YOLOv8l)was se-lected for further hyperparameter tuning based on its performance metrics.This model underwent a detailed hyperparameter optimization using the One Factor At a Time(OFAT)methodology,concentrating on key param-eters such as learning rate,batch size,weight decay,epochs,and optimizer.Insights from the OFAT study were used to define search spaces for a subsequent Random Search(RS).The final model derived from RS demon-strated significant improvements over the initial fine-tuned model,increasing overall precision by 1.39%,recall by 1.48%,F1-score by 1.44%,mAP@0.50 by 0.70%,and mAP@0.50:0.95 by 5.09%.We validated the enhanced model's efficacy on a diverse set of real-world images,reflecting various agricultural settings,to confirm its ro-bustness in detecting smoke and fire.These results underscore the model's reliability and effectiveness in scenar-ios critical to agricultural safety and environmental monitoring.This work,representing a significant advancement in the field of fire and smoke detection through machine learning,lays a strong foundation for fu-ture research and solutions aimed at safeguarding agricultural areas and natural environments. 展开更多
关键词 Agricultural safety Wildfire detection Smoke detection Object detection Computer vision YOL
原文传递
Corn kernel classification from few training samples
3
作者 Patricia L.Suárez Henry O.Velesaca +1 位作者 Dario Carpio Angel D.Sappa 《Artificial Intelligence in Agriculture》 2023年第3期89-99,共11页
This article presents an efficient approach to classify a set of corn kernels in contact,which may contain good,or defective kernels along with impurities.The proposed approach consists of two stages,the first one is ... This article presents an efficient approach to classify a set of corn kernels in contact,which may contain good,or defective kernels along with impurities.The proposed approach consists of two stages,the first one is a next-generation segmentation network,trained by using a set of synthesized images that is applied to divide the given image into a set of individual instances.An ad-hoc lightweight CNN architecture is then proposed to classify each instance into one of three categories(ie good,defective,and impurities).The segmentation network is trained using a strategy that avoids the time-consuming and human-error-prone task of manual data annotation.Regarding the classification stage,the proposed ad-hoc network is designed with only a few sets of layers to result in a lightweight architecture capable of being used in integrated solutions.Experimental results and comparisons with previous approaches showing both the improvement in accuracy and the reduction in time are provided.Finally,the segmentation and classification approach proposed can be easily adapted for use with other cereal types. 展开更多
关键词 Corn kernel classification Computer vision approaches Quality inspection Food grain identification Machine vision Instance segmentation Synthesized dataset generation
原文传递
上一页 1 下一页 到第
使用帮助 返回顶部