超解像のための画像及び言語の統合特徴を利用したPerceptual Lossの改善

書誌事項

タイトル別名
  • Improving Perceptual Loss with CLIP for Super-Resolution

抄録

<p>Perceptual loss, calculated by VGG network pre-trained on ImageNet, has been widely employed in the past for super-resolution tasks, enabling the generation of photo-realistic images. However, it has been reported that grid-like artifacts frequently appear in the generated images. To address this problem, we consider that large-scale pre-trained models can make significant contributions to super-resolution across different scenes. In particular, by combining language, those models can exhibit a strong capability to comprehend complex scenes, potentially enhancing super-resolution performance. Therefore, this paper proposes new perceptual loss with Contrastive Language-Image Pre-training (CLIP) based on Vision Transformer (ViT) instead of VGG network. The results demonstrate our proposed perceptual loss can generate photo-realistic images without grid-like artifacts.</p>

収録刊行物

  • 精密工学会誌

    精密工学会誌 90 (2), 217-223, 2024-02-05

    公益社団法人 精密工学会

参考文献 (13)*注記

もっと見る

詳細情報 詳細情報について

問題の指摘

ページトップへ