期刊文献+
共找到29篇文章
< 1 2 >
每页显示 20 50 100
Digital image inpainting by example-based image synthesis method 被引量:1
1
作者 聂栋栋 Ma Lizhuang Xiao Shuangjiu 《High Technology Letters》 EI CAS 2006年第3期276-282,共7页
A simple and effective image inpainting method is proposed in this paper, which is proved to be suitable for different kinds of target regions with shapes from little scraps to large unseemly objects in a wide range o... A simple and effective image inpainting method is proposed in this paper, which is proved to be suitable for different kinds of target regions with shapes from little scraps to large unseemly objects in a wide range of images. It is an important improvement upon the traditional image inpainting techniques. By introducing a new bijeetive-mapping term into the matching cost function, the artificial repetition problem in the final inpainting image is practically solved. In addition, by adopting an inpainting error map, not only the target pixels are refined gradually during the inpainting process but also the overlapped target patches are combined more seamlessly than previous method. Finally, the inpainting time is dramatically decreased by using a new acceleration method in the matching process. 展开更多
关键词 INPAINTING image synthesis texture synthesis prority matching cost function example patch isophote DIFFUSION
在线阅读 下载PDF
A Survey of GAN Based Image Synthesis
2
作者 Jiahe Ni 《Journal of Information Hiding and Privacy Protection》 2022年第2期79-88,共10页
Image generation is a hot topic in the academic recently,and has been applied to AI drawing,which can bring Vivid AI paintings without labor costs.In image generation,we represent the image as a random vector,assuming... Image generation is a hot topic in the academic recently,and has been applied to AI drawing,which can bring Vivid AI paintings without labor costs.In image generation,we represent the image as a random vector,assuming that the images of the natural scene obey an unknown distribution,we hope to estimate its distribution through some observation samples.Especially,with the development of GAN(Generative Adversarial Network),The generator and discriminator improve the model capability through adversarial,the quality of the generated image is also increasing.The image quality generated by the existing GAN based image generation model is so well-paint that it can be passed for genuine one.Based on the brief introduction of the concept ofGAN,this paper analyzes themain ideas of image synthesis,studies the representative SOTA GAN based Image synthesis method. 展开更多
关键词 Deep learning image synthesis SOTA generative adversarial network
在线阅读 下载PDF
IIDM:Image-to-image diffusion model for semantic image synthesis
3
作者 Feng Liu Xiaobin Chang 《Computational Visual Media》 2025年第2期423-429,共7页
Semantic image synthesis aims to generate highquality images given semantic conditions,i.e.,segmentation masks and style reference images.Existing methods widely adopt generative adversarial networks(GANs).GANs take a... Semantic image synthesis aims to generate highquality images given semantic conditions,i.e.,segmentation masks and style reference images.Existing methods widely adopt generative adversarial networks(GANs).GANs take all conditional inputs and directly synthesize images in a single forward step.In this paper,semantic image synthesis is treated as an image denoising task and is handled with a novel image-to-image diffusion model(IIDM). 展开更多
关键词 generative adversarial networks semantic image synthesis image synthesis directly synthesize images image image diffusion model style reference imagesexisting generative adversarial networks gans gans image denoising task
原文传递
HAGAN:Hybrid Augmented Generative Adversarial Network for Medical Image Synthesis
4
作者 Zhihan Ju Wanting Zhou +4 位作者 Longteng Kong Yu Chen Yi Li Zhenan Sun Caifeng Shan 《Machine Intelligence Research》 2025年第5期969-982,共14页
Medical image synthesis(MIS)can greatly save the economic and time costs of medical diagnosis.However,due to the complexity of medical images and similar characteristics of different tissue cells,existing methods face... Medical image synthesis(MIS)can greatly save the economic and time costs of medical diagnosis.However,due to the complexity of medical images and similar characteristics of different tissue cells,existing methods face great challenges in meeting their biological consistency.To this end,we propose the hybrid augmented generative adversarial network(HAGAN)to maintain the authenticity of structural texture and tissue cells.HAGAN contains attention mixed(AttnMix)generator,hierarchical discriminator and reverse skip connection between discriminator and generator.The AttnMix consistency differentiable regularization encourages the perception in structural and textural variations between real and fake images,which improves the pathological integrity of synthetic images and the accuracy of features in local areas.The hierarchical discriminator introduces pixel-by-pixel discriminant feedback to generator for enhancing the saliency and discriminance of global and local details simultaneously.The reverse skip connection further improves the accuracy for fine details by fusing real and synthetic distribution features.Our experimental evaluations on two datasets of different scales,i.e.,ACDC and BraTS2018,demonstrate that HAGAN outperforms the existing methods and achieves state-of-the-art performance in both high-resolution and low-resolution. 展开更多
关键词 Medical image synthesis generative adversarial network hybrid augmentation consistency differentiable regularization localdiscrimination
原文传递
Enhancing vehicle Re-identification by pair-flexible pose guided vehicle image synthesis
5
作者 Baolu Li Ping Liu +4 位作者 Lan Fu Jinlong Li Jianwu Fang Zhigang Xu Hongkai Yu 《Green Energy and Intelligent Transportation》 2025年第5期15-25,共11页
Vehicle Re-identification(Re-ID)has drawn extensive exploration recently;nevertheless,the issue of accurately distinguishing features in latent space across varying vehicle poses,remains a challenging hurdle for real-... Vehicle Re-identification(Re-ID)has drawn extensive exploration recently;nevertheless,the issue of accurately distinguishing features in latent space across varying vehicle poses,remains a challenging hurdle for real-world application of Vehicle Re-ID.To address this challenge,we supply a novel idea which projects the various-pose vehicle images into a unified target pose so as to promote the discriminative capability of vehicle Re-ID model.Acknowledging the labor and cost of paired data for the same vehicle images across different traffic surveillance cameras in practical scenarios,we propose the pioneering Pair-flexible Pose Guided Image Synthesis for vehicle Re-ID,denominated as VehicleGAN.Our method is adept at both supervised(paired images of same vehicle)and unsupervised(unpaired images of any vehicle)settings,and bypasses the need of geometric 3D model information.Furthermore,we propose a novel Joint Metric Learning(JML)method to facilitate the effective fusion of both real and synthetic data.Comprehensive experimental analyses conducted on the public VeRi-776 and VehicleID datasets substantiate the precision and efficacy of our proposed VehicleGAN and JML. 展开更多
关键词 Vehicle Re-identification Metric learning image synthesis
原文传递
A Survey of Image Synthesis and Editing with Generative Adversarial Networks 被引量:20
6
作者 Xian Wu Kun Xu Peter Hall 《Tsinghua Science and Technology》 SCIE EI CAS CSCD 2017年第6期660-674,共15页
This paper presents a survey of image synthesis and editing with Generative Adversarial Networks(GANs). GANs consist of two deep networks, a generator and a discriminator, which are trained in a competitive way. Due... This paper presents a survey of image synthesis and editing with Generative Adversarial Networks(GANs). GANs consist of two deep networks, a generator and a discriminator, which are trained in a competitive way. Due to the power of deep networks and the competitive training manner, GANs are capable of producing reasonable and realistic images, and have shown great capability in many image synthesis and editing applications.This paper surveys recent GAN papers regarding topics including, but not limited to, texture synthesis, image inpainting, image-to-image translation, and image editing. 展开更多
关键词 image synthesis image editing constrained image synthesis generative adversarial networks imageto-image translation
原文传递
Deep image synthesis from intuitive user input:A review and perspectives 被引量:2
7
作者 Yuan Xue Yuan-Chen Guo +3 位作者 Han Zhang Tao Xu Song-Hai Zhang Xiaolei Huang 《Computational Visual Media》 SCIE EI CSCD 2022年第1期3-31,共29页
In many applications of computer graphics,art,and design,it is desirable for a user to provide intuitive non-image input,such as text,sketch,stroke,graph,or layout,and have a computer system automatically generate pho... In many applications of computer graphics,art,and design,it is desirable for a user to provide intuitive non-image input,such as text,sketch,stroke,graph,or layout,and have a computer system automatically generate photo-realistic images according to that input.While classically,works that allow such automatic image content generation have followed a framework of image retrieval and composition,recent advances in deep generative models such as generative adversarial networks(GANs),variational autoencoders(VAEs),and flow-based methods have enabled more powerful and versatile image generation approaches.This paper reviews recent works for image synthesis given intuitive user input,covering advances in input versatility,image generation methodology,benchmark datasets,and evaluation metrics.This motivates new perspectives on input representation and interactivity,cross fertilization between major image generation paradigms,and evaluation and comparison of generation methods. 展开更多
关键词 image synthesis intuitive user input deep generative models synthesized image quality evaluation
原文传递
A Comprehensive Pipeline for Complex Text-to-Image Synthesis
8
作者 Fei Fang Fei Luo +3 位作者 Hong-Pan Zhang Hua-Jian Zhou Alix L.H.Chow Chun-Xia Xiao 《Journal of Computer Science & Technology》 SCIE EI CSCD 2020年第3期522-537,共16页
Synthesizing a complex scene image with multiple objects and background according to text description is a challenging problem.It needs to solve several difficult tasks across the fields of natural language processing... Synthesizing a complex scene image with multiple objects and background according to text description is a challenging problem.It needs to solve several difficult tasks across the fields of natural language processing and computer vision.We model it as a combination of semantic entity recognition,object retrieval and recombination,and objects’status optimization.To reach a satisfactory result,we propose a comprehensive pipeline to convert the input text to its visual counterpart.The pipeline includes text processing,foreground objects and background scene retrieval,image synthesis using constrained MCMC,and post-processing.Firstly,we roughly divide the objects parsed from the input text into foreground objects and background scenes.Secondly,we retrieve the required foreground objects from the foreground object dataset segmented from Microsoft COCO dataset,and retrieve an appropriate background scene image from the background image dataset extracted from the Internet.Thirdly,in order to ensure the rationality of foreground objects’positions and sizes in the image synthesis step,we design a cost function and use the Markov Chain Monte Carlo(MCMC)method as the optimizer to solve this constrained layout problem.Finally,to make the image look natural and harmonious,we further use Poisson-based and relighting-based methods to blend foreground objects and background scene image in the post-processing step.The synthesized results and comparison results based on Microsoft COCO dataset prove that our method outperforms some of the state-of-the-art methods based on generative adversarial networks(GANs)in visual quality of generated scene images. 展开更多
关键词 image synthesis scene generation text-to-image conversion Markov Chain Monte Carlo(MCMC)
原文传递
Multi3D:3D-aware multimodal image synthesis
9
作者 Wenyang Zhou Lu Yuan Taijiang Mu 《Computational Visual Media》 CSCD 2024年第6期1205-1217,共13页
3D-aware image synthesis has attained high quality and robust 3D consistency.Existing 3D controllable generative models are designed to synthesize 3D-aware images through a single modality,such as 2D segmentation or s... 3D-aware image synthesis has attained high quality and robust 3D consistency.Existing 3D controllable generative models are designed to synthesize 3D-aware images through a single modality,such as 2D segmentation or sketches,but lack the ability to finely control generated content,such as texture and age.In pursuit of enhancing user-guided controllability,we propose Multi3D,a 3D-aware controllable image synthesis model that supports multi-modal input.Our model can govern the geometry of the generated image using a 2D label map,such as a segmentation or sketch map,while concurrently regulating the appearance of the generated image through a textual description.To demonstrate the effectiveness of our method,we have conducted experiments on multiple datasets,including CelebAMask-HQ,AFHQ-cat,and shapenet-car.Qualitative and quantitative evaluations show that our method outperforms existing state-of-the-art methods. 展开更多
关键词 generate adversarial networks(GANs) neural radiation field(NeRF) 3D-aware image synthesis controllable generation
原文传递
EACNet:Ensemble adversarial co-training neural network for handling missing modalities in MRI images for brain tumor segmentation
10
作者 RAMADHAN Amran Juma CHEN Jing PENG Junlan 《Journal of Measurement Science and Instrumentation》 2025年第1期11-25,共15页
Brain tumor segmentation is critical in clinical diagnosis and treatment planning.Existing methods for brain tumor segmentation with missing modalities often struggle when dealing with multiple missing modalities,a co... Brain tumor segmentation is critical in clinical diagnosis and treatment planning.Existing methods for brain tumor segmentation with missing modalities often struggle when dealing with multiple missing modalities,a common scenario in real-world clinical settings.These methods primarily focus on handling a single missing modality at a time,making them insufficiently robust for the additional complexity encountered with incomplete data containing various missing modality combinations.Additionally,most existing methods rely on single models,which may limit their performance and increase the risk of overfitting the training data.This work proposes a novel method called the ensemble adversarial co-training neural network(EACNet)for accurate brain tumor segmentation from multi-modal magnetic resonance imaging(MRI)scans with multiple missing modalities.The proposed method consists of three key modules:the ensemble of pre-trained models,which captures diverse feature representations from the MRI data by employing an ensemble of pre-trained models;adversarial learning,which leverages a competitive training approach involving two models;a generator model,which creates realistic missing data,while sub-networks acting as discriminators learn to distinguish real data from the generated“fake”data.Co-training framework utilizes the information extracted by the multimodal path(trained on complete scans)to guide the learning process in the path handling missing modalities.The model potentially compensates for missing information through co-training interactions by exploiting the relationships between available modalities and the tumor segmentation task.EACNet was evaluated on the BraTS2018 and BraTS2020 challenge datasets and achieved state-of-the-art and competitive performance respectively.Notably,the segmentation results for the whole tumor(WT)dice similarity coefficient(DSC)reached 89.27%,surpassing the performance of existing methods.The analysis suggests that the ensemble approach offers potential benefits,and the adversarial co-training contributes to the increased robustness and accuracy of EACNet for brain tumor segmentation of MRI scans with missing modalities.The experimental results show that EACNet has promising results for the task of brain tumor segmentation of MRI scans with missing modalities and is a better candidate for real-world clinical applications. 展开更多
关键词 deep learning magnetic resonance imaging(MRI) medical image analysis semantic segmentation segmentation accuracy image synthesis
在线阅读 下载PDF
A Perspective-Aware Cyclist Image Generation Method for Perception Development of Autonomous Vehicles
11
作者 Beike Yu Dafang Wang +1 位作者 Xing Cui Bowen Yang 《Computers, Materials & Continua》 2025年第2期2687-2702,共16页
Realistic urban scene generation has been extensively studied for the sake of the development of autonomous vehicles. However, the research has primarily focused on the synthesis of vehicles and pedestrians, while the... Realistic urban scene generation has been extensively studied for the sake of the development of autonomous vehicles. However, the research has primarily focused on the synthesis of vehicles and pedestrians, while the generation of cyclists is rarely presented due to its complexity. This paper proposes a perspective-aware and realistic cyclist generation method via object retrieval. Images, semantic maps, and depth labels of objects are first collected from existing datasets, categorized by class and perspective, and calculated by an algorithm newly designed according to imaging principles. During scene generation, objects with the desired class and perspective are retrieved from the collection and inserted into the background, which is then sent to the modified 2D synthesis model to generate images. This pipeline introduces a perspective computing method, utilizes object retrieval to control the perspective accurately, and modifies a diffusion model to achieve high fidelity. Experiments show that our proposal gets a 2.36 Fréchet Inception Distance, which is lower than the competitive methods, indicating a superior realistic expression ability. When these images are used for augmentation in the semantic segmentation task, the performance of ResNet-50 on the target class can be improved by 4.47%. These results demonstrate that the proposed method can be used to generate cyclists in corner cases to augment model training data, further enhancing the perception capability of autonomous vehicles and improving the safety performance of autonomous driving technology. 展开更多
关键词 Realistic cyclist generation perspective-aware image synthesis autonomous vehicle artificial intelligence
在线阅读 下载PDF
An Approach to Synthesize Diverse Underwater Image Dataset 被引量:4
12
作者 Xiaodong LIU Ben M.CHEN 《Instrumentation》 2019年第3期67-75,共9页
Images that are taken underwater mostly present color shift with hazy effects due to the special property of water.Underwater image enhancement methods are proposed to handle this issue.However,their enhancement resul... Images that are taken underwater mostly present color shift with hazy effects due to the special property of water.Underwater image enhancement methods are proposed to handle this issue.However,their enhancement results are only evaluated on a small number of underwater images.The lack of a sufficiently large and diverse dataset for efficient evaluation of underwater image enhancement methods provokes the present paper.The present paper proposes an organized method to synthesize diverse underwater images,which can function as a benchmark dataset.The present synthesis is based on the underwater image formation model,which describes the physical degradation process.The indoor RGB-D image dataset is used as the seed for underwater style image generation.The ambient light is simulated based on the statistical mean value of real-world underwater images.Attenuation coefficients for diverse water types are carefully selected.Finally,in total 14490 underwater images of 10 water types are synthesized.Based on the synthesized database,state-of-the-art image enhancement methods are appropriately evaluated.Besides,the large diverse underwater image database is beneficial in the development of learning-based methods. 展开更多
关键词 image Processing Underwater image Enhancement Underwater image synthesis
原文传递
A Novel Unsupervised MRI Synthetic CT Image Generation Framework with Registration Network
13
作者 Liwei Deng Henan Sun +2 位作者 Jing Wang Sijuan Huang Xin Yang 《Computers, Materials & Continua》 SCIE EI 2023年第11期2271-2287,共17页
In recent years,radiotherapy based only on Magnetic Resonance(MR)images has become a hot spot for radiotherapy planning research in the current medical field.However,functional computed tomography(CT)is still needed f... In recent years,radiotherapy based only on Magnetic Resonance(MR)images has become a hot spot for radiotherapy planning research in the current medical field.However,functional computed tomography(CT)is still needed for dose calculation in the clinic.Recent deep-learning approaches to synthesized CT images from MR images have raised much research interest,making radiotherapy based only on MR images possible.In this paper,we proposed a novel unsupervised image synthesis framework with registration networks.This paper aims to enforce the constraints between the reconstructed image and the input image by registering the reconstructed image with the input image and registering the cycle-consistent image with the input image.Furthermore,this paper added ConvNeXt blocks to the network and used large kernel convolutional layers to improve the network’s ability to extract features.This research used the collected head and neck data of 180 patients with nasopharyngeal carcinoma to experiment and evaluate the training model with four evaluation metrics.At the same time,this research made a quantitative comparison of several commonly used model frameworks.We evaluate the model performance in four evaluation metrics which achieve Mean Absolute Error(MAE),Root Mean Square Error(RMSE),Peak Signal-to-Noise Ratio(PSNR),and Structural Similarity(SSIM)are 18.55±1.44,86.91±4.31,33.45±0.74 and 0.960±0.005,respectively.Compared with other methods,MAE decreased by 2.17,RMSE decreased by 7.82,PSNR increased by 0.76,and SSIM increased by 0.011.The results show that the model proposed in this paper outperforms other methods in the quality of image synthesis.The work in this paper is of guiding significance to the study of MR-only radiotherapy planning. 展开更多
关键词 MRI-CT image synthesis variational auto-encoder medical image translation MRI-only based radiotherapy
在线阅读 下载PDF
Tight Sandstone Image Augmentation for Image Identification Using Deep Learning
14
作者 Dongsheng Li Chunsheng Li +4 位作者 Kejia Zhang Tao Liu Fang Liu Jingsong Yin Mingyue Liao 《Computer Systems Science & Engineering》 SCIE EI 2023年第10期1209-1231,共23页
Intelligent identification of sandstone slice images using deep learning technology is the development trend of mineral identification,and accurate mineral particle segmentation is the most critical step for intellige... Intelligent identification of sandstone slice images using deep learning technology is the development trend of mineral identification,and accurate mineral particle segmentation is the most critical step for intelligent identification.A typical identification model requires many training samples to learn as many distinguishable features as possible.However,limited by the difficulty of data acquisition,the high cost of labeling,and privacy protection,this has led to a sparse sample number and cannot meet the training requirements of deep learning image identification models.In order to increase the number of samples and improve the training effect of deep learning models,this paper proposes a tight sandstone image data augmentation method by combining the advantages of the data deformation method and the data oversampling method in the Putaohua reservoir in the Sanzhao Sag of the Songliao Basin as the target area.First,the Style Generative Adversarial Network(StyleGAN)is improved to generate high-resolution tight sandstone images to improve data diversity.Second,we improve the Automatic Data Augmentation(AutoAugment)algorithm to search for the optimal augmentation strategy to expand the data scale.Finally,we design comparison experiments to demonstrate that this method has obvious advantages in generating image quality and improving the identification effect of deep learning models in real application scenarios. 展开更多
关键词 Tight sandstone image synthesis generative adversarial networks data augmentation image segmentation
在线阅读 下载PDF
Anime Generation through Diffusion and Language Models:A Comprehensive Survey of Techniques and Trends
15
作者 Yujie Wu Xing Deng +4 位作者 Haijian Shao Ke Cheng Ming Zhang Yingtao Jiang Fei Wang 《Computer Modeling in Engineering & Sciences》 2025年第9期2709-2778,共70页
The application of generative artificial intelligence(AI)is bringing about notable changes in anime creation.This paper surveys recent advancements and applications of diffusion and language models in anime generation... The application of generative artificial intelligence(AI)is bringing about notable changes in anime creation.This paper surveys recent advancements and applications of diffusion and language models in anime generation,focusing on their demonstrated potential to enhance production efficiency through automation and personalization.Despite these benefits,it is crucial to acknowledge the substantial initial computational investments required for training and deploying these models.We conduct an in-depth survey of cutting-edge generative AI technologies,encompassing models such as Stable Diffusion and GPT,and appraise pivotal large-scale datasets alongside quantifiable evaluation metrics.Review of the surveyed literature indicates the achievement of considerable maturity in the capacity of AI models to synthesize high-quality,aesthetically compelling anime visual images from textual prompts,alongside discernible progress in the generation of coherent narratives.However,achieving perfect long-form consistency,mitigating artifacts like flickering in video sequences,and enabling fine-grained artistic control remain critical ongoing challenges.Building upon these advancements,research efforts have increasingly pivoted towards the synthesis of higher-dimensional content,such as video and three-dimensional assets,with recent studies demonstrating significant progress in this burgeoning field.Nevertheless,formidable challenges endure amidst these advancements.Foremost among these are the substantial computational exigencies requisite for training and deploying these sophisticated models,particularly pronounced in the realm of high-dimensional generation such as video synthesis.Additional persistent hurdles include maintaining spatial-temporal consistency across complex scenes and mitigating ethical considerations surrounding bias and the preservation of human creative autonomy.This research underscores the transformative potential and inherent complexities of AI-driven synergy within the creative industries.We posit that future research should be dedicated to the synergistic fusion of diffusion and autoregressive models,the integration of multimodal inputs,and the balanced consideration of ethical implications,particularly regarding bias and the preservation of human creative autonomy,thereby establishing a robust foundation for the advancement of anime creation and the broader landscape of AI-driven content generation. 展开更多
关键词 Diffusion models language models anime generation image synthesis video generation stable diffusion AIGC
在线阅读 下载PDF
TerraFusion:Joint generation of terrain geometry and texture using latent diffusion models
16
作者 Kazuki HIGO Toshiki KANAI +1 位作者 Yuki ENDO Yoshihiro KANAMORI 《虚拟现实与智能硬件(中英文)》 2025年第6期560-576,共17页
Background Three-dimensional terrain models are essential in domains such as video game development and film production.Because surface color is often correlated with terrain geometry,capturing this relationship is cr... Background Three-dimensional terrain models are essential in domains such as video game development and film production.Because surface color is often correlated with terrain geometry,capturing this relationship is critical for generating realistic results.However,most existing methods synthesize either a heightmap or a texture without adequately modeling their inherent correlation.Methods We propose a method that jointly generates terrain heightmaps and textures using a latent diffusion model.First,we train the model in an unsupervised manner to randomly generate paired heightmaps and textures.Then,we perform supervised learning on an external adapter to enable user control via hand-drawn sketches.Results Experiments demonstrate that our approach supports intuitive terrain generation while preserving the correlation between heightmaps and textures.Conclusion Our method outperforms the two-stage and GAN-based baselines by ensuring structural coherence,in which textures naturally align with geometry,successfully accommodating both realistic landscapes and extreme user-defined shapes. 展开更多
关键词 TERRAIN TEXTURE Heightmap image synthesis Deep learning
在线阅读 下载PDF
A Study on Polyp Dataset Expansion Algorithm Based on Improved Pix2Pix
17
作者 Ziji Xiao Kaibo Yang +3 位作者 Mingen Zhong Kang Fan Jiawei Tan Zhiying Deng 《Computers, Materials & Continua》 2025年第2期2665-2686,共22页
The polyp dataset involves the confidentiality of medical records, so it might be difficult to obtain datasets with accurate annotations. This problem can be effectively solved by expanding the polyp data set with alg... The polyp dataset involves the confidentiality of medical records, so it might be difficult to obtain datasets with accurate annotations. This problem can be effectively solved by expanding the polyp data set with algorithms. The traditional polyp dataset expansion scheme usually requires the use of two models or traditional visual methods. These methods are both tedious and difficult to provide new polyp features for training data. Therefore, our research aims to efficiently generate high-quality polyp samples, so as to effectively expand the polyp dataset. In this study, we first added the attention mechanism to the generation model and improved the loss function to reduce the interference caused by reflection in the image generation process. Meanwhile, we used the improved generation model to remove polyps from the original image. In addition, we used masks of different shapes generated by random combinations to generate polyps with more characteristic information. The same generation model was used for the removal and generation of polyps. The generated polyp image has its own annotation, which is conducive to us directly using the expanded data set for training. Finally, we verified the effectiveness of the improved model and the dataset expansion scheme through a series of comparative experiments on the public dataset. The results showed that using the dataset we generate for training can significantly optimize the main performance indicators. 展开更多
关键词 Polyp formation polyp detection image synthesis generative adversarial network Pix2Pix
在线阅读 下载PDF
A scale determination method for MSMFS CLEAN based on gradient descent optimizer
18
作者 Xueying He Lei Tan +1 位作者 Ying Mei Hui Deng 《Astronomical Techniques and Instruments》 2025年第4期219-225,共7页
The performance of the deconvolution algorithm plays a crucial role in data processing of radio interferometers.The multi-scale multi-frequency synthesis(MSMFS)CLEAN is a widely used deconvolution algorithm for radio ... The performance of the deconvolution algorithm plays a crucial role in data processing of radio interferometers.The multi-scale multi-frequency synthesis(MSMFS)CLEAN is a widely used deconvolution algorithm for radio interferometric imaging,which combines the advantages of both wide-band synthesis imaging and multi-scale imaging and can substantially improve performance.However,how best to effectively determine the optimal scale is an important problem when implementing the MSMFS CLEAN algorithm.In this study,we proposed a Gaussian fitting method for multiple sources based on the gradient descent algorithm,with consideration of the influence of the point spread function(PSF).After fitting,we analyzed the fitting components using statistical analysis to derive reasonable scale information through the model parameters.A series of simulation validations demonstrated that the scales extracted by our proposed algorithm are accurate and reasonable.The proposed method can be applied to the deconvolution algorithm and provide modeling analysis for Gaussian sources,offering data support for source extraction algorithms. 展开更多
关键词 Radio astronomy DECONVOLUTION synthesis imaging
在线阅读 下载PDF
Evolution and Effectiveness of Loss Functions in Generative Adversarial Networks 被引量:1
19
作者 Ali Syed Saqlain Fang Fang +2 位作者 Tanvir Ahmad Liyun Wang Zain-ul Abidin 《China Communications》 SCIE CSCD 2021年第10期45-76,共32页
Recently,the evolution of Generative Adversarial Networks(GANs)has embarked on a journey of revolutionizing the field of artificial and computational intelligence.To improve the generating ability of GANs,various loss... Recently,the evolution of Generative Adversarial Networks(GANs)has embarked on a journey of revolutionizing the field of artificial and computational intelligence.To improve the generating ability of GANs,various loss functions are introduced to measure the degree of similarity between the samples generated by the generator and the real data samples,and the effectiveness of the loss functions in improving the generating ability of GANs.In this paper,we present a detailed survey for the loss functions used in GANs,and provide a critical analysis on the pros and cons of these loss functions.First,the basic theory of GANs along with the training mechanism are introduced.Then,the most commonly used loss functions in GANs are introduced and analyzed.Third,the experimental analyses and comparison of these loss functions are presented in different GAN architectures.Finally,several suggestions on choosing suitable loss functions for image synthesis tasks are given. 展开更多
关键词 loss functions deep learning machine learning unsupervised learning generative adversarial networks(GANs) image synthesis
在线阅读 下载PDF
Explicitly Color-Inspired Neural Style Transfer Using
20
作者 Bumsoo Kim Wonseop Shin +2 位作者 Yonghoon Jung Youngsup Park Sanghyun Seo 《Computer Modeling in Engineering & Sciences》 SCIE EI 2024年第12期2143-2164,共22页
Arbitrary style transfer aims to perceptually reflect the style of a reference image in artistic creations with visual aesthetics.Traditional style transfer models,particularly those using adaptive instance normalizat... Arbitrary style transfer aims to perceptually reflect the style of a reference image in artistic creations with visual aesthetics.Traditional style transfer models,particularly those using adaptive instance normalization(AdaIN)layer,rely on global statistics,which often fail to capture the spatially local color distribution,leading to outputs that lack variation despite geometric transformations.To address this,we introduce Patchified AdaIN,a color-inspired style transfer method that applies AdaIN to localized patches,utilizing local statistics to capture the spatial color distribution of the reference image.This approach enables enhanced color awareness in style transfer,adapting dynamically to geometric transformations by leveraging local image statistics.Since Patchified AdaIN builds on AdaIN,it integrates seamlessly into existing frameworks without the need for additional training,allowing users to control the output quality through adjustable blending parameters.Our comprehensive experiments demonstrate that Patchified AdaIN can reflect geometric transformations(e.g.,translation,rotation,flipping)of images for style transfer,thereby achieving superior results compared to state-of-the-art methods.Additional experiments show the compatibility of Patchified AdaIN for integration into existing networks to enable spatial color-aware arbitrary style transfer by replacing the conventional AdaIN layer with the Patchified AdaIN layer. 展开更多
关键词 Neural style transfer image synthesis image stylization
在线阅读 下载PDF
上一页 1 2 下一页 到第
使用帮助 返回顶部