Core Concepts
Distortion rectification can be effectively achieved by learning an ordinal distortion representation from a single distorted image, which provides a more explicit and homogeneous learning target compared to the traditional implicit and heterogeneous distortion parameters.
Abstract
The key insights of this work are:
Distortion rectification can be cast as a problem of learning an ordinal distortion from a single distorted image. The ordinal distortion indicates the distortion levels of a series of pixels, which extend outward from the principal point.
The ordinal distortion is more explicit to image features and homogeneous in representation compared to the traditional distortion parameters. This enables neural networks to gain sufficient distortion perception and achieve faster convergence without extra feature guidance or pixel-wise supervision.
The authors design a local-global associated estimation network that learns the ordinal distortion to approximate the realistic distortion distribution. A distortion-aware perception layer is exploited to boost the feature extraction of different degrees of distortion.
The estimated ordinal distortion can be easily converted to the distortion parameters for various camera models, enabling efficient and accurate distortion rectification.
Extensive experiments demonstrate that the proposed approach outperforms state-of-the-art methods by a significant margin, with approximately 23% improvement on the quantitative evaluation while using fewer input images.
Stats
The distortion coefficients k1, k2, k3, k4 are randomly generated from their corresponding ranges: k1 ∈ [-1e-3, -1e-8], k2 ∈ [-1e-7, -1e-12] or [1e-12, 1e-7], k3 ∈ [-1e-11, -1e-16] or [1e-16, 1e-11], and k4 ∈ [-1e-15, -1e-20] or [1e-20, 1e-15].
The synthetic dataset contains 20,000 training images, 2,000 test images, and 2,000 validation images.
Quotes
"Our key insight is that distortion rectification can be cast as a problem of learning an ordinal distortion from a single distorted image."
"The ordinal distortion is homogeneous as all its elements share a similar magnitude and description. Therefore, the imbalanced optimization problem no longer exists during the training process, and we do not need to focus on the cumbersome factor-balancing task anymore."
"The ordinal distortion can be estimated using only a part of a distorted image. Unlike the semantic information, the distortion information is redundant in images, showing the central symmetry and mirror symmetry to the principal point."