IPOLIPOLhttp://www.ipol.im/feed/IPOL Articles — Latest articles published in IPOL.ikiwiki2018-12-15T22:35:42ZAn Analysis and Implementation of the Shape Preserving Local Histogram
Modification Algorithmhttp://www.ipol.im/pub/art/2018/236/Jose-Luis Lisani2018-12-06T23:02:26Z2018-12-06T23:00:00Z
In this paper we describe the implementation of the
algorithm for local contrast enhancement published by Caselles et al. in 1999.
This algorithm was the first designed explicitly to increase the contrast
while preserving the so-called 'shape structure' of the image, that is, its
set of level sets. According to the mathematical morphology school,
artifacts are created when this structure is modified.
The original algorithm is described and also two alternative implementations
are proposed, which limit the over-enhancement of noise.
Study of the Principal Component Analysis Method for the
Correction of Images Degraded by Turbulencehttp://www.ipol.im/pub/art/2018/47/Tristan Dagobert,
Yohann Tendero,
Stéphane Landeau2018-12-15T22:35:42Z2018-11-22T23:00:00Z
This article analyzes and discusses a well-known paper
[D. Li, R.M. Mersereau and S. Simske, IEEE Letters on Geoscience and Remote Sensing, 3:4 (2007), pp. 340-344]
that applies principal component analysis
in order to restore image sequences degraded by atmospheric turbulence.
We propose a variant of this
method and its ANSI C implementation.
The proposed variant applies to image sequences acquired with short as well
as long exposure times.
Examples of restored images
using sequences of real atmospheric turbulence are presented.
The acquisition of a dataset of image sequences with real atmospheric turbulence is described
and the dataset is made available for download.
An Implementation of the Exposure Fusion Algorithmhttp://www.ipol.im/pub/art/2018/230/Charles Hessel2018-11-17T20:35:04Z2018-11-16T23:00:00Z
Exposure Fusion is a high dynamic range imaging technique to fuse a bracketed
exposure sequence into a high quality image, introduced in 2009 by Mertens et al.
Contrarily to most HDR imaging methods, exposure fusion does not construct an
intermediate HDR image but directly constructs the final LDR one by seamlessly
fusing the best regions of the input sequence, using the Laplacian pyramid.
Since its publication, this method received considerable attention, being both
effective and efficient. We propose in this paper a precise description of the method
and an analysis of its main limitation, an out-of-range artifact.
Structural Similarity Metrics for Quality Image Fusion Assessment: Algorithmshttp://www.ipol.im/pub/art/2018/196/Silvina Pistonesi,
Jorge Martinez,
Silvia Maria Ojeda,
Ronny Vallejos2018-10-25T22:16:44Z2018-10-25T22:00:00Z
The wide use of image fusion techniques in different fields such as medical diagnostics, digital camera vision, military and surveillance applications, among others, has motivated the development of various image quality fusion metrics, in order to evaluate them. In this paper, we study and implement the algorithms of non-reference image structural similarity based metrics for fusion assessment: Piella's metric, Cvejic's metric, Yang's metric, and Codispersion Fusion Quality metric. We conduct the comparative experiment of the selected image fusion metrics over four multiresolution image fusion algorithms, performed on different pairs of images used in different applications.
Interpolation of Missing Samples in Sound Signals Based on Autoregressive Modelinghttp://www.ipol.im/pub/art/2018/23/Laurent Oudre2018-10-17T21:56:33Z2018-10-16T22:00:00Z
This article proposes an implementation and a study of the paper 'Adaptive Interpolation of Discrete-Time Signals That Can Be Modeled as Autoregressive Processes' by Janssen et al. The algorithm presented in this paper allows one to reconstruct an audio signal which presents localized degradations by interpolating the missing samples. This method assumes that the signal can locally be modeled as a realization of an autoregressive process and iteratively estimates the model parameters and the interpolated samples by minimizing a quadratic criterion. We investigate the limits and the algorithmic aspects of this method on several audio examples.
An Analysis and Implementation of the Harris Corner Detectorhttp://www.ipol.im/pub/art/2018/229/Javier Sánchez,
Nelson Monzón,
Agustín Salgado2018-10-03T10:26:21Z2018-10-02T22:00:00Z
In this work, we present an implementation and thorough study of the Harris corner detector. This feature detector relies on the analysis of the eigenvalues of the autocorrelation matrix. The algorithm comprises seven steps, including several measures for the classification of corners, a generic non-maximum suppression method for selecting interest points, and the possibility to obtain the corners position with subpixel accuracy. We study each step in detail and propose several alternatives for improving the precision and speed. The experiments analyze the repeatability rate of the detector using different types of transformations.
Estimating an Image's Blur Kernel Using Natural Image Statistics, and Deblurring it: An Analysis of the Goldstein-Fattal Methodhttp://www.ipol.im/pub/art/2018/211/Jérémy Anger,
Gabriele Facciolo,
Mauricio Delbracio2018-09-26T21:47:21Z2018-09-25T22:00:00Z
Despite the significant improvement in image quality resulting from improvement in optical sensors and general electronics, camera shake blur significantly undermines the quality of hand-held photographs. In this work, we present a detailed description and implementation of the blur kernel estimation algorithm introduced by Goldstein and Fattal in 2012. Unlike most methods that attempt to solve an inverse problem through a variational formulation (e.g. through a Maximum A Posteriori estimation), this method directly estimates the blur kernel by modeling statistical irregularities in the power spectrum of blurred natural images. The adopted mathematical model extends the well-known power-law by contemplating the presence of dominant strong edges in particular directions. The blur kernel is retrieved from an estimation of its power spectrum, by solving a phase retrieval problem using additional constraints associated with the particular nature of camera shake blur kernels (e.g. non-negativity and small spatial support). Although the algorithm is conceptually simple, its numerical implementation presents several challenges. This work contributes to a detailed anatomy of the Goldstein and Fattal method, its algorithmic description, and its parameters.
Fast Affine Invariant Image Matchinghttp://www.ipol.im/pub/art/2018/225/Mariano Rodríguez,
Julie Delon,
Jean-Michel Morel2018-09-30T21:43:24Z2018-09-23T22:00:00Z
Methods performing Image Matching by Affine Simulation (IMAS) attain affine invariance by applying a finite set of affine transforms to the images before comparing them with a Scale Invariant Image Matching (SIIM) method like SIFT or SURF. We describe here how to optimize IMAS methods. First, we detail an algorithm computing a minimal discrete set of affine transforms to be applied to each image before comparison. It yields a full practical affine invariance at the lowest computational cost. The matching complexity of current IMAS algorithms is divided by about 4. Our approach also associates to each image an affine invariant set of descriptors, which is twice smaller than the set of descriptors usually used in IMAS methods, and only 6.4 times larger than the set of similarity invariant descriptors of SIIM methods. In order to reduce the number of false matches, which are inherently more frequent in IMAS approaches than in SIIM, we introduce the notion of hyper-descriptor, which groups descriptors whose keypoints are spatially close. Finally, we also propose a matching criterion allowing each keypoint of the query image to be matched with several keypoints of the target image, in order to deal with situations where an object is repeated several times in the target image.
Numerical Simulation of Landscape Evolution Modelshttp://www.ipol.im/pub/art/2018/205/Marc Lebrun,
Miguel Colom,
Jérôme Darbon,
Jean-Michel Morel2018-09-03T19:51:45Z2018-09-02T22:00:00Z
This paper gives the complete numerical schemes implementing the main physical laws proposed in landscape evolution models (LEMs). These laws can be modeled by a system of three partial differential equations governing water runoff, stream incision, hill slope evolution and sedimentation. The goal of the presented algorithm, code and online demo is to be able to test these equations on digital elevation models (DEMs) of any resolution, and to illustrate its potential to simulate the fine structure of the river network, and to understand the landscape morphology and its causes. The equations simulate plausible evolutions. We illustrate experiments on DEMs of several sites, including one site, La Réunion where the DEM is given at three different resolutions: the SRTM resolution (90m), and then 12m and 4m on DEMs derived from several Pléiades pairs. Other many DEMs are proposed in the online demo, which allows to upload and tests other DEMs.
An Analysis and Implementation of Multigrid Poisson Solvers With Verified Linear
Complexityhttp://www.ipol.im/pub/art/2018/228/Matías Di Martino,
Gabriele Facciolo2018-09-04T21:52:26Z2018-07-25T22:00:00Z
The Poisson equation is the most studied partial differential equation, and it
allows to formulate many useful image processing methods in an elegant and efficient
mathematical framework. Using different variations of data terms and boundary conditions,
Poisson-like problems can be developed, e.g., for local contrast enhancement, inpainting, or
image seamless cloning among many other applications. Multigrid solvers are among the most
efficient numerical solvers for discrete Poisson-like equations. However, their correct
implementation relies on: (i) the proper definition of the discrete problem, (ii) the right
choice of interpolation and restriction operators, and (iii) the adequate
formulation of the problem across different scales and layers. In the present work we address
these aspects, and we provide a mathematical and practical description of multigrid methods. In
addition, we present an alternative to the extended formulation of Poisson equation
proposed in 2011 by Mainberger et al. The proposed formulation of the problem suits better
multigrid methods, in particular, because it has important mathematical properties that can be
exploited to define the problem at different scales in a intuitive and natural way. In
addition, common iterative solvers and Poisson-like problems are empirically analyzed and
compared. For example, the complexity of problems is compared when the topology of Dirichlet
boundary conditions changes in the interior of the regular domain of the image. The main
contribution of this work is the development and detailed description of an implementation of a
multigrid numerical solver which converges in linear time.
Automatic Detection of Internal Copy-Move Forgeries in Imageshttp://www.ipol.im/pub/art/2018/213/Thibaud Ehret2018-09-04T21:52:26Z2018-07-24T22:00:00Z
This article presents an implementation and discussion of the recently proposed 'Efficient Dense-Field Copy-Move Forgery Detection' by Cozzolino et al. This method is a forgery detection based on a dense field of descriptors chosen to be invariant by rotation. Zernike moments were suggested in the original article. An efficient matching of the descriptors is then performed using PatchMatch, which is extremely efficient to find duplicate regions. Regions matched by PatchMatch are processed to find the final detections. This allows a precise and accurate detection of copy-move forgeries inside a single suspicious image. We also extend successfully the method to the use of dense SIFT descriptors and show that they are better at detecting forgeries using Poisson editing.
Video Denoising with Optical Flow Estimationhttp://www.ipol.im/pub/art/2018/224/Antoni Buades,
Jose-Luis Lisani2018-09-04T21:52:26Z2018-07-22T22:00:00Z
In this paper we describe the implementation of state-of-the-art video denoising algorithm SPTWO [A. Buades, J.L. Lisani, M. Miladinovic, Patch Based Video Denoising with Optical Flow Estimation, IEEE Transactions on Image Processing 25 (6), 2573--2586]. This algorithm, inspired by image fusion techniques, uses motion compensation by regularized optical flow methods, which permits robust patch comparison in spatiotemporal volumes. Groups of similar patches are denoised using Principal Component Analysis, which ensures the correct preservation of fine texture and details.
Theory and Practice of Image B-Spline Interpolationhttp://www.ipol.im/pub/art/2018/221/Thibaud Briand,
Pascal Monasse2018-09-04T21:52:26Z2018-07-22T22:00:00Z
We explain how the B-spline interpolation of signals and, in particular, of images can be efficiently performed by linear filtering. Based on the seminal two-step method proposed by Unser et al. in 1991, we propose two slightly different prefiltering algorithms whose precisions are proven to be controlled thanks to a rigorous boundary handling. This paper contains all the information, theoretical and practical, required to perform efficiently B-spline interpolation for any order and any boundary extension. We describe precisely how to evaluate the kernel and to compute the B-spline interpolator parameters. We show experimentally that increasing the order improves the interpolation quality. As a fundamental application we also provide an implementation of homographic transformation of images using B-spline interpolation.
Efficient Large-scale Image Search With a Vocabulary Treehttp://www.ipol.im/pub/art/2018/199/Esteban Uriza,
Francisco Gómez Fernández,
Martín Rais2018-07-22T11:40:36Z2018-06-02T22:00:00Z
The task of searching and recognizing objects in images has become an important research
topic in the area of image processing and computer vision. Looking for similar images in large
data sets given an input query and responding as fast as possible is a very challenging task.
In this work the Bag of Features approach is studied, and an implementation of the visual
vocabulary tree method from Nistér and Stewénius is presented. Images are described using
local invariant descriptor techniques and then indexed in a database using an inverted index
for further queries. The descriptors are quantized according to a visual vocabulary, creating
sparse vectors, which allows to compute very efficiently, for each query, a ranking of similarity
for indexed images. The performance of the method is analyzed varying different
factors, such as the parameters for the vocabulary tree construction, different techniques
of local descriptors extraction and dimensionality reduction with PCA.
It can be observed that the retrieval performance increases
with a richer vocabulary and decays very slowly as the size of the dataset grows.
A MATLAB SMO Implementation to Train a SVM Classifier: Application to Multi-Style License Plate Numbers Recognitionhttp://www.ipol.im/pub/art/2018/173/Pablo Negri2018-07-22T11:40:36Z2018-05-21T22:00:00Z
This paper implements the Support Vector Machine (SVM) training procedure proposed by John Platt denominated Sequential Minimimal Optimization (SMO).
The application of this system involves a multi-style license plate characters recognition identifying numbers from '0' to '9'.
In order to be robust against license plates with different character/background colors, the characters (numbers) visual information is encoded using Histograms of Oriented Gradients (HOG).
A reliability measure to validate the system outputs is also proposed.
Several tests are performed to evaluate the sensitivity of the algorithm to different parameters and kernel functions.
Gestaltic Grouping of Line Segmentshttp://www.ipol.im/pub/art/2018/194/Boshra Rajaei,
Rafael Grompone von Gioi2018-11-02T09:25:40Z2018-03-21T23:00:00Z
Using simple grouping rules in Gestalt theory, one may detect higher level features (geometric
structures) in an image from elementary features. By recursive grouping of already detected
geometric structures a bottom-up pyramid could be built that extracts increasingly complex
geometric features from the input image. Taking advantage of the (recent) advances in
reliable line segment detectors, in this paper, we propose three feature detectors along with their
corresponding detailed algorithms that constitute one step up in this pyramid. For any digital
image, our unsupervised algorithm computes three classic Gestalts from the set of predetected
line segments: good continuations, non-local alignments, and bars. The methodology is based
on a common stochastic a contrario model yielding three simple detection formulas,
characterized by their number of false alarms.
This detection algorithm is illustrated on several digital images.
Contours, Corners and T-Junctions Detection Algorithmhttp://www.ipol.im/pub/art/2018/218/Antoni Buades,
Rafael Grompone von Gioi,
Julia Navarro2018-09-04T21:52:26Z2018-02-26T23:00:00Z
This article describes the implementation of the method by Buades, Grompone and Navarro in 2017 for the detection of line segments, contours, corners and T-junctions. The method is inspired by the mammal visual system. The detection of corners and T-junctions plays a role as part of the process in contour detection. An a contrario validation is applied to select the most meaningful contours without the need of fixing any critical parameter.
The Production of Ground Truths for Evaluating Highly Accurate Stereovision Algorithmshttp://www.ipol.im/pub/art/2018/187/Tristan Dagobert2018-07-22T11:40:36Z2018-01-01T23:00:00Z
The conception and improvement of algorithms for subpixel stereovision requires very precise test databases. The state of the art on the sets of images used extensively by the scientific community shows that they are often incomplete and imprecise compared to the dataset goals. We will present a method based on image synthesis to produce stereoscopic pairs with ground truths such as disparity and occlusion maps reaching an accuracy of about 1e-6 pixels. The a priori noise estimate is also taken into account. This process allows us to deliver a new image database consisting of 66 stereo pairs together with their ground truths.
A Contrario 3D Point Alignment Detection Algorithmhttp://www.ipol.im/pub/art/2017/214/Álvaro Gómez,
Gregory Randall,
Rafael Grompone von Gioi2018-07-22T11:40:36Z2017-12-29T23:00:00Z
In this article we present an algorithm for the detection of perceptually relevant alignments in 3D point clouds. The algorithm is an extension of the algorithm developed by Lezama et al. [J. Lezama, J-M. Morel, G. Randall, R. Grompone von Gioi, A Contrario 2D Point Alignment Detection, IEEE Transactions on Pattern Analysis and Machine Intelligence, 37 (3), pp. 499-512, 2015] for the case of sets of 2D points. The algorithm is based on the a contrario detection theory that mathematically formalizes the non-accidentalness principle proposed for perception: an observed structure is relevant if it rarely occurs by chance. This framework has been widely used in different detection tasks and leads to algorithms with a single critical parameter to control the number of false detections.
Local Region Expansion: a Method for Analyzing and Refining Image Matcheshttp://www.ipol.im/pub/art/2017/154/Erez Farhan,
Elad Meir,
Rami Hagege2018-07-22T11:40:36Z2017-12-22T23:00:00Z
We present a novel method for locating large amounts of local matches between images, with highly accurate localization. Point matching is one of the most fundamental tasks in computer vision, extensively used in applications such as object detection, object tracking and structure from motion. The major challenge in point matching is to preserve large numbers of accurate matches between corresponding scene locations under different geometric and radiometric conditions, while keeping the number of false positives low. Recent publications have shown that applying an affine transformation model on local regions is a particularly suitable approach for point matching. Yet, affine invariant methods are not used extensively for two reasons: first, because these methods are computationally demanding; and second because the derived affine estimations have limited accuracy. In this work, we propose a novel method of region expansion that enhances region matches detected by any state-of-the-art method. The method is based on accurate estimation of affine transformations, which are used to predict matching locations beyond initially detected matches. We use the improved estimations of affine transformations to locally verify tentative matches in an efficient way. We systematically reject false matches, while improving the localization of correct matches that are usually rejected by state-of-the-art methods.
Non-Local Patch-Based Image Inpaintinghttp://www.ipol.im/pub/art/2017/189/Alasdair Newson,
Andrés Almansa,
Yann Gousseau,
Patrick Pérez2018-07-22T11:40:36Z2017-12-12T23:00:00Z
Image inpainting is the process of filling in missing regions in an image in a plausible way.
In this contribution, we propose and describe an implementation of a patch-based image inpainting algorithm. The method is actually a two-dimensional version of our video inpainting algorithm proposed in
[A. Newson et al., Video inpainting of complex scenes, SIAM Journal of Imaging Sciences, 7 (2014)].
The algorithm attempts to minimize a highly non-convex functional, first introducted by Wexler et al. in
[Wexler et al., Space-time video completion, CCVPR (2004)].
The functional
specifies that a good solution to the inpainting problem should be an image where each patch is very similar
to its nearest neighbor in the unoccluded area.
Iterations are performed in a multi-scale framework which yields globally coherent results. In this manner two of the major goals of image inpainting, the correct reconstruction
of textures and structures, are addressed.
We address a series of important practical issues which arise when using such an
approach. In particular, we reduce execution times by using the PatchMatch
[C. Barnes, PatchMatch: a randomized correspondence algorithm for structural image editing, ACM Transactions on Graphics, (2009)]
algorithm for nearest neighbor searches, and we propose a modified
patch distance which improves the comparison of textured patches.
We address the crucial issue of initialization and the choice of the
number of pyramid levels, two points which are rarely discussed in
such approaches.
We provide several examples which illustrate the advantages of our algorithm,
and compare our results with those of state-of-the-art methods.
A Sub-Pixel Edge Detector: an Implementation of the Canny/Devernay Algorithmhttp://www.ipol.im/pub/art/2017/216/Rafael Grompone von Gioi,
Gregory Randall2018-07-22T11:40:36Z2017-11-27T23:00:00Z
An image edge detector is described which produces chained edge points with sub-pixel accuracy. The method incorporates the main ideas of the classic Canny and Devernay algorithms. The analysis shows that a slight modification to the original formulation improves the accuracy of the edge points.
Comparison of Motion Smoothing Strategies for Video Stabilization using Parametric Modelshttp://www.ipol.im/pub/art/2017/209/Javier Sánchez2018-07-22T11:40:36Z2017-11-25T23:00:00Z
This paper is devoted to a rigorous implementation and to an exhaustive comparison of video stabilization techniques. These techniques aim at removing the undesirable effects of camera shake. They first estimate a global transform from frame to frame, which can be a translation, a similarity, an affine map or a homography. This generates a signal that can be smoothed and used to compensate the noisy transform signal. This paper compares all classic smoothing methods and their boundary conditions. It also analyzes two algorithms to crop the video after stabilization.
The stabilization results are displayed in a scale-space form permitting to extract valuable information about ego-motion such as its frequencies and its general tendencies.
Multi-Scale DCT Denoisinghttp://www.ipol.im/pub/art/2017/201/Nicola Pierazzo,
Jean-Michel Morel,
Gabriele Facciolo2018-07-22T11:40:36Z2017-10-28T22:00:00Z
DCT denoising is a classic low complexity method built in the JPEG compression norm. Once made translation invariant, this algorithm was still proven to be competitive at the beginning of this century. Since then, it has been outperformed by patch based methods, which are far more complex. This paper proposes a two-step multi-scale version of the algorithm that boosts its performance and reduces its artifacts.
The multi-scale strategy decomposes the image in a dyadic DCT pyramid, which keeps noise white at all scales. The single scale denoising is then applied to all scales, thus giving multiple denoised versions of the low frequency coefficients of the denoised image. A 'multi-scale fusion' of these multiple estimates avoids the ringing artifacts resulting from the pyramid recomposition. The final algorithm attains a good PNSR and much improved visual image quality. It is shown to have a deficit of only 1dB with respect to state of the art algorithms, but its complexity is two orders of magnitude lower.
The Bilateral Filter for Point Cloudshttp://www.ipol.im/pub/art/2017/179/Julie Digne,
Carlo de Franchis2018-07-22T11:40:36Z2017-10-28T22:00:00Z
Point sets obtained by 3D scanners are often corrupted with noise, that can have several causes, such as a tangential acquisition direction, changing environmental lights or a reflective object material. It is thus crucial to design efficient tools to remove noise from the acquired data without removing important information such as sharp edges or shape details. To do so, Fleishman et al. introduced a bilateral filter for meshes adapted from the bilateral filter for gray level images. This anisotropic filter denoises a point with respect to its neighbors by considering not only the distance from the neighbors to the point but also the distance along a normal direction. This simple fact allows for a much better preservation of sharp edges. In this paper, we analyze a parallel implementation of the bilateral filter adapted for point clouds.