Download PDFOpen PDF in browser

Tensor Super-Resolution with Generative Adversarial Nets: A Large Image Generation Approach

EasyChair Preprint no. 1395

15 pagesDate: August 12, 2019


Deep generative models have been successfully applied to many applications. However, existing methods experience limitations when generating large images (the literature usually generates small images, e.g., 32*32 or 128*128). In this paper, we propose a novel scheme using tensor super-resolution with adversarial generative nets (TSRGAN), to generate large high-quality images by exploring tensor structures. Essentially, the super resolution process of TSRGAN is based on tensor representation. First, we impose tensor structures for concise image representation, which is superior in capturing the pixel proximity information and the spatial patterns of elementary objects in images, over the vectorization preprocess in existing works. Secondly, we propose TSRGAN that integrates deep convolutional generative adversarial networks and tensor super-resolution in a cascading manner, to generate high-quality images from random distributions. More specifically, we design a tensor super-resolution process that consists of tensor dictionary learning and tensor coefficients learning. Finally, on three datasets, the proposed TSRGAN generates images with more realistic textures, compared with state-of-the-art adversarial autoencoders and super-resolution methods. The size of the generated images is increased by over 8.5 times, namely 374*374 in PASCAL2.

Keyphrases: GAN, generative model, super-resolution, tensor representation, tensor sparse coding

BibTeX entry
BibTeX does not have the right entry for preprints. This is a hack for producing the correct reference:
  author = {Zihan Ding and Xiao-Yang Liu and Miao Yin and Linghe Kong},
  title = {Tensor Super-Resolution with Generative Adversarial Nets: A Large Image Generation Approach},
  howpublished = {EasyChair Preprint no. 1395},

  year = {EasyChair, 2019}}
Download PDFOpen PDF in browser