首页> 外文期刊>IEEE transactions on multimedia >FuseGAN: Learning to Fuse Multi-Focus Image via Conditional Generative Adversarial Network
【24h】

FuseGAN: Learning to Fuse Multi-Focus Image via Conditional Generative Adversarial Network

机译:FuseGAN:通过有条件的生成对抗网络学习融合多焦点图像

获取原文
获取原文并翻译 | 示例
获取外文期刊封面目录资料

摘要

We study the problem of multi-focus image fusion, where the key challenge is detecting the focused regions accurately among multiple partially focused source images. Inspired by the conditional generative adversarial network (cGAN) to image-to-image task, we propose a novel FuseGAN to fulfill the images-to-image for multi-focus image fusion. To satisfy the requirement of dual input-to-one output, the encoder of the generator in FuseGAN is designed as a Siamese network. The least square GAN objective is employed to enhance the training stability of FuseGAN, resulting in an accurate confidence map for focus region detection. Also, we exploit the convolutional conditional random fields technique on the confidence map to reach a refined final decision map for better focus region detection. Moreover, due to the lack of a large-scale standard dataset, we synthesize a large enough multi-focus image dataset based on a public natural image dataset PASCAL VOC 2012, where we utilize a normalized disk point spread function to simulate the defocus and separate the background and foreground in the synthesis for each image. We conduct extensive experiments on two public datasets to verify the effectiveness of the proposed method. Results demonstrate that the proposed method presents accurate decision maps for focus regions in multi-focus images, such that the fused images are superior to 11 recent state-of-the-art algorithms, not only in visual perception, but also in quantitative analysis in terms of five metrics.
机译:我们研究了多焦点图像融合的问题,其中的关键挑战是在多个部分聚焦的源图像之间准确检测聚焦区域。受条件生成对抗网络(cGAN)启发进行图像到图像的任务,我们提出了一种新颖的FuseGAN来实现图像到图像的多焦点图像融合。为了满足双输入一对一输出的要求,FuseGAN中发电机的编码器被设计为连体网络。最小二乘GAN物镜用于增强FuseGAN的训练稳定性,从而获得用于焦点区域检测的准确置信度图。此外,我们在置信度图中利用卷积条件随机场技术来达到精炼的最终决策图,以更好地进行焦点区域检测。此外,由于缺乏大规模的标准数据集,我们基于公共自然图像数据集PASCAL VOC 2012合成了足够大的多焦点图像数据集,其中我们使用归一化的磁盘点扩展函数来模拟散焦并分离每个图像合成的背景和前景。我们在两个公共数据集上进行了广泛的实验,以验证该方法的有效性。结果表明,所提出的方法为多焦点图像中的焦点区域提供了准确的决策图,从而使融合图像不仅在视觉感知方面而且在定量分析方面均优于11种最新的算法。五个指标的术语。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号