Authors:
Clément Peyrard
1
;
Franck Mamalet
2
and
Christophe Garcia
3
Affiliations:
1
Orange Labs and INSA Lyon, France
;
2
Orange Labs, France
;
3
INSA Lyon, France
Keyword(s):
Super-Resolution, Text Image, Multi-Layer Perceptron, Convolutional Neural Network, OCR.
Related
Ontology
Subjects/Areas/Topics:
Computer Vision, Visualization and Computer Graphics
;
Image Enhancement and Restoration
;
Image Formation and Preprocessing
Abstract:
We compare the performances of several Multi-Layer Perceptrons (MLPs) and Convolutional Neural Networks (ConvNets) for single text image Super-Resolution. We propose an example-based framework for both MLP and ConvNet, where a non-linear mapping between pairs of patches and high-frequency pixel values is learned. We then demonstrate that for equivalent complexity, ConvNets are better than MLPs at predicting missing details in upsampled text images. To evaluate the performances, we make use of a recent database (ULR-textSISR-2013a) along with different quality measures. We show that the proposed methods outperforms sparse coding-based methods for this database.