Next Article in Journal
Method for Determining the Plasmon Resonance Wavelength in Fiber Sensors Based on Tilted Fiber Bragg Gratings
Next Article in Special Issue
Cooperative Traffic Signal Control with Traffic Flow Prediction in Multi-Intersection
Previous Article in Journal
A Two-Level WiFi Fingerprint-Based Indoor Localization Method for Dangerous Area Monitoring
Previous Article in Special Issue
An Adaptation Multi-Group Quasi-Affine Transformation Evolutionary Algorithm for Global Optimization and Its Application in Node Localization in Wireless Sensor Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Feature Point Matching Based on Distinct Wavelength Phase Congruency and Log-Gabor Filters in Infrared and Visible Images

1
School of Electronics and Information Engineering, Harbin Institute of Technology, Harbin 150001, China
2
Information and Electronic Technology Institute, Jiamusi University, Jiamusi 154002, China
3
College of Computer Science and Engineering, Shandong University of Science and Technology, Qingdao 266510, China
4
Fujian Provincial Key Laboratory of Big Data Minning and Applications, Fujian University of Technology, Fuzhou 350118, China
5
College of Informatics, Chaoyang University of Science and Technology, Taichung 413, Taiwan
*
Author to whom correspondence should be addressed.
Sensors 2019, 19(19), 4244; https://doi.org/10.3390/s19194244
Submission received: 29 July 2019 / Revised: 20 September 2019 / Accepted: 24 September 2019 / Published: 29 September 2019
(This article belongs to the Special Issue Intelligent Systems in Sensor Networks and Internet of Things)

Abstract

:
Infrared and visible image matching methods have been rising in popularity with the emergence of more kinds of sensors, which provide more applications in visual navigation, precision guidance, image fusion, and medical image analysis. In such applications, image matching is utilized for location, fusion, image analysis, and so on. In this paper, an infrared and visible image matching approach, based on distinct wavelength phase congruency (DWPC) and log-Gabor filters, is proposed. Furthermore, this method is modified for non-linear image matching with different physical wavelengths. Phase congruency (PC) theory is utilized to obtain PC images with intrinsic and affluent image features for images containing complex intensity changes or noise. Then, the maximum and minimum moments of the PC images are computed to obtain the corners in the matched images. In order to obtain the descriptors, log-Gabor filters are utilized and overlapping subregions are extracted in a neighborhood of certain pixels. In order to improve the accuracy of the algorithm, the moments of PCs in the original image and a Gaussian smoothed image are combined to detect the corners. Meanwhile, it is improper that the two matched images have the same PC wavelengths, due to the images having different physical wavelengths. Thus, in the experiment, the wavelength of the PC is changed for different physical wavelengths. For realistic application, BiDimRegression method is proposed to compute the similarity between two points set in infrared and visible images. The proposed approach is evaluated on four data sets with 237 pairs of visible and infrared images, and its performance is compared with state-of-the-art approaches: the edge-oriented histogram descriptor (EHD), phase congruency edge-oriented histogram descriptor (PCEHD), and log-Gabor histogram descriptor (LGHD) algorithms. The experimental results indicate that the accuracy rate of the proposed approach is 50% higher than the traditional approaches in infrared and visible images.

1. Introduction

Image matching is the process of aligning images of the same scene which have been acquired under different conditions, such as with a different field of view, at different scales, different resolutions, different times, or using different sensors, and so on. It is a prerequisite step for many applications [1,2,3], such as visual navigation, precision guidance, image fusion, medical image analysis, and target location. Conte et al. proposed a vision-based navigation system which uses an image matching method to obtain useful position information from aerial imagery [4]. Zhang et al. employed the coarse-to-fine image matching methodology to achieve scene matching guidance [5]. In [6], an effective visible light and infrared image fusion algorithm, using feature residuals and statistical matching, has been proposed. Yang et al. utilized learned non-linear local descriptors and feature matching to predict pseudo-computed tomography (PCT) images from T1- and T2-weighted magnetic resonance imaging (MRI) data [7]. Recently, image matching methodologies have been widely and effectively applied in the field of the target location, presenting unique advantages [8]. For the purpose of determining the best matching relations between two images with such differences, an effective feature point detection approach is of significant use.
Infrared (IR) and visible image matching, which is utilized commonly in the fields of computer vision, target identification, and military ground vehicle identification [9], has recently attracted the attention of the research community. Huang et al. proposed an improved registration method for infrared and visible remote sensing images using NSCT and SIFT [10]. Hariharan et al. presented a new image fusion method, which utilized Empirical Mode Decomposition (EMD), for improved face recognition [11]. Cheng et al. employed infrared and visible image matching methods to complete automatic target recognition [12]. In the military field, 24 h all-weather functionality is required for the military tasks of target/threat detection and identification. In target identification, infrared image matching has been used as a supplement for visible images, in order to reduce the effects of poor illumination. Infrared and visible image matching is divided into the categories of multispectral or multiband image analysis [13]. Infrared and visible light images, captured from two sensors working in different bands, present significant distinctions from single-band methods. These distinctions include the following points: (1) different imaging mechanisms make the infrared and visible images able to deliver properties of objects in different bands. (2) Different imaging conditions cause the gray distortion and the geometric deformation of images to be easily affected by shooting time, season, environment, light intensity, and so on, which causes the infrared and visible images to differ in many characteristics. Pan et al. studied wireless sensor technology concerning deep algorithms [14,15,16]. Wang et al. proposed many approaches for heterogeneous wireless sensor networks [17,18,19]. According to the analysis of the physical wavelengths for the images, visible images possess more information than infrared.
The wavelet transform has been utilized to obtain local frequency information at a point in a signal [20]. In [21], a considerable amount of visual processing models have been shown to be quite effective in accounting for a wide range of physiological observations. Theories of cortical neuron behavior have varied widely, from Fourier analysis to edge detection. Edge detection has proven to be an effective means of describing image features. However, it has been shown that cortical neurons lack an edge detection mechanism. Thus, a new approach, based on Gabor’s theory of communication, was proposed. Gabor showed how to represent time-varying signals in terms of functions that are localized in both time and frequency. In 1985, Daugman proposed the 2D Gabor filter family, which can describe the two-dimensional receptive field profiles of simple cells in the mammalian visual cortex and is the best choice for discerning image textures [22]. Following that, Daugman extracted phase information of iris patterns to encode iris features, in order to make iris recognition more accurate [23]. Recently, the phase congruency-based Gabor filter family has been proposed for the detection of corners and edges in the images with varying contrast [24]. David showed that a feature of this model is that, on a linear scale, the bandwidths of the different sensors are proportional to their optimal bandwidths [21]. However, the distinction of images from different sensors reflects the physical wavelength difference, as was analyzed above. Similar to the physical wavelength of the light, the wavelength of the wavelet influences image feature extraction. A shorter wavelength will describe detailed image information and a longer wavelength will describe coarse image information. Thus, the wavelength should be adjusted, according to the physical wavelengths of the matched images. The wavelengths utilized for the visible range should be set to be larger, in order to extract coarse information; and that for the infrared should be smaller, in order to extract detailed information. In other words, the two images are decomposed by the wavelet, and the most similar parts are matched to obtain better matching results. Inspired by these ideas, distinct wavelength log-Gabor filters are used for feature detection of infrared and visible images in this paper.
The rest of this paper is arranged as follows. Section 2 details the related works. Then, the proposed approach is illustrated in Section 3. The experiments are presented in Section 4. Finally, the conclusions are provided in Section 5.

2. Related Work

A classical automatic image matching process includes the following four steps: (1) feature points detection; (2) feature points description and matching; (3) transformation model estimation; and (4) image resampling. According to the above process, most multimodal image matching methods can be broadly classified into two categories: feature-based and area-based.
Area-based methods commonly utilize a template window of a given size to detect the feature points between two images. After the template window in an image is defined, the corresponding window in the other image is found by computing the matching information according to a predefined similarity measure, such as cross-correlation [25], phase correlation [26], or mutual information [27]. The centers of the matching windows are regarded as the feature points, which are then used to align the two images.
Feature-based methods first extract the significant features from both matched images and then match them using their similarities. These significant features can be regions [28], lines [29] or points [30,31], and are regarded as features with distinctiveness and stability at fixed locations, regardless of changes of image geometry, scanned scene changes, non-linear intensity changes, and so on. Recently, local invariant features have been widely applied in image matching, such as SIFT [32], SURF [33], BRISKLeu [34], ORB [35], WLD [36], WLBP [37] and so on, which have been broadly used in applications for visible image matching due to their robustness to geometric and illumination changes [38]. However, inferior matching performance has been reported for these methods when used with multispectral images.
A host of researchers have made significant efforts towards matching infrared and optical images. In the 1990s, a visible–infrared matching method using a multi-scale edge detection algorithm to obtain surface boundaries was first presented, in which the matching system used a hierarchical estimation process [39]. Following that, many variants of SIFT have been advocated for visible–infrared image matching. Firmenich et al. recommended the gradient-invariant SIFT (GDSIFT) as an RGB–NIR image matching method [40]. The gradient orientation modification SIFT (GOM-SIFT) method has been employed by Yi et al [41]. On the other hand, the gradient information differs between the visible and infrared images, which can cause these variants to perform worse than the SIFT method [42]. The discrete curve evolution (DCE) method has also been advocated, which gives better results for planar scenes. However, it cannot be well-adapted to non-planar scenes [43].
Recently, the local self-similarity frequency descriptor has been put forward [44], which allows for robust descriptors in multispectral image matching [45]. More studies have detailed the advances in visible and infrared image matching [46,47]. Concerning the methods which deal with the special application of handling multispectral images, the descriptor edge-oriented histogram (EOH) has been considered to be a dramatic baseline in scientific studies [46,47,48,49]. However, it has been shown that it is difficult to select an appropriate threshold for the EOH descriptor, in order to ensure that the edges extracted from multispectral images are similar [48]. Thus, Fu et al. advocated a local feature descriptor with a combination of structural and textural information for multispectral image matching, which is perfect for the non-linear intensity changes of multispectral images [50].
In the early 1970s, a number of attempts were made to reveal the purpose of the rather mysterious behavior of cortical neurons. It is evident that cortical neurons are selective to spatial frequency and orientation, which directed a number of researchers to produce something like a Fourier transform. In [21], two types of feature descriptions—Gabor and log-Gabor—have been considered as good methods for transforming redundancy. Gabor filtering is similar to the operations of simple cells in the visual cortex, which was presented by Daugman, in 1980, as a framework for understanding the orientation- and spatial frequency-selective receptive field properties of neurons in the brain’s visual cortex, which outperformed the classical dyadic wavelet transform for texture classifications in many tasks [51]. Recently, log-Gabor filters have been advocated for replacing EOH filters, as they have shown better performance than other methods [46,52]. In these approaches, fast feature detection is utilized to detect the feature point [53], which has no adaptability for the descriptor, based on the log-Gabor filters. However, feature detection is an important step in the image matching process. To address this problem, a corner and edge detector have been developed from the phase congruency model of feature detection, which provides information invariant to image contrast [24]. Recently, many researchers have used phase congruency methods for feature detection [54,55,56,57].

3. Feature Point Matching Based on Distinct Wavelength Phase Congruency and Log-Gabor Filters for Infrared and Visible Images

Inspired by the theory of phase congruency and Gabor filters, we present a feature point matching method based on distinct wavelength phase congruency and log-Gabor filters for infrared and visible images.
Figure 1 gives an overview of the proposed approach. First, the original target and reference images are input, as shown as Figure 1a,b, which are dealt with, using phase congruency (PC) theory, to obtain the PC images, as shown in Figure 1c,d. Then, the maximum and minimum moments of the PC images, as shown as Figure 1e–h, are computed to obtain the corners in the images, as shown in Figure 1i,j. Next, the complex images in the real domain, as shown in Figure 1l, are computed by the log-Gabor filter, using four scales and six orientations. The phase information with maximum energy is then utilized to form the phase images at each scale, as shown in Figure 1m, and the orientation histograms of the phase images are utilized to describe the image features, as shown in Figure 1n. Finally, the nearest neighbor and RANSAC algorithms are used to determine the corresponding keypoints, as shown in Figure 1k.

3.1. Corner Detection Based on Distinct Wavelength Phase Congruency

Keypoints are defined as a class of important image features with large changes in intensity, such as line endings, corners, junctions, and so on, where the Fourier components of the image are typically maximum in phase. The quality of the keypoints has a direct impact on image matching. Phase congruency is a dimensionless quantity which finds keypoints in images with multispectral changes, which presents significant advantages over gradient-based methods. The local energy model of feature detection suggests that keypoints are determined at points of maximum phase congruency in an image. In images obtained from different sensors, there will be similar spectral qualities in two matched images. Thus, in our proposed method, the wavelength of the wavelet is adjusted according to the physical wavelength ranges of images from different sensors.
Morrone and Owens defined the phase congruency function, in terms of the Fourier series expansion of a signal at a point, as [58]:
P C ( x ) = m a x ϕ ¯ ( x ) [ 0 , 2 π ] n A n c o s ( ϕ n ( x ) ϕ ¯ ( x ) ) n A n ,
where A n represents the amplitude of the nth Fourier component and ϕ n ( x ) represents the local phase of the Fourier component at location x. The value of ϕ ¯ ( x ) which maximizes this equation is the amplitude weighted mean local phase angle of all Fourier coefficients at the considered point. For the sake of obtaining better localization without sensitivity to noise, a new measure [24] is defined as
P C n ( x ) = n W ( x ) A n ( x ) ( c o s ( ϕ n ( x ) ϕ ¯ ( x ) ) s i n ( ϕ n ( x ) ϕ ¯ ( x ) ) ) T n A n ( x ) + ϵ ,
where W ( x ) is a weighting factor for frequency spread and ϵ is used to avoid division by zero. Only energy values exceeding T are counted in the result. As shown in Figure 2, given an input image (Figure 2a), PC images for six orientations (Figure 2b) are computed using Equation (2). Then, the covariance matrix is computed, and its minimum and maximum moments are calculated in order to produce a highly localized operator, which is then used to identify both edges and corners in a contrast-invariant manner. The covariance data is built as follows:
C o v x ( θ ) = P C ( θ ) c o s ( θ ) ,
C o v y ( θ ) = P C ( θ ) s i n ( θ ) ,
a = C o v x ( θ ) 2 ,
b = 2 C o v x ( θ ) C o v y ( θ ) ,
c = C o v y ( θ ) 2 ,
where P C ( θ ) is the phase congruency value determined at orientation θ . The maximum and minimum moments—M and m, respectively—are computed as
M = 1 2 ( c + a + b 2 + ( a c ) 2 ) ,
m = 1 2 ( c + a b 2 + ( a c ) 2 ) .
Figure 2c,d show the maximum and minimum moment images, respectively. When M is greater than the threshold T M , an edge will be marked; when m is greater than the threshold T m , a corner of an edge in the image will be detected, as shown in Figure 2e.
In our work, log-Gabor filters are used to obtain frequency information local to a point in an image. The log-Gabor transform in the frequency domain, using polar co-ordinates, is given as follows:
G ( r , θ , ψ , n ) = e x p ( [ l o g ( r λ k n ) ] 2 2 σ 2 ) e x p ( ( θ ψ ) 2 2 ) ,
where r and θ denote the radius and filter angle (in polar coordinates), n represents the orientation angle, σ represents the Gaussian standard deviation of the filter, λ represents the smallest wavelength of the filter, and k denotes the scaling factor between successive filters, which controls the wavelength of the log-Gabor filter. Images smoothed by the log-Gabor filter at different wavelengths can be considered to be images including special spectral information. Thus, in the experiment, the wavelengths of the log-Gabor filter for the matched images are adjusted respectively, in order to obtain similar spectral information from the visible and infrared images. For the visible image matching, it has better performance, in traditional approaches, when the value of k is set to 2.1. As shown in Figure 3, the matching accuracy rate will change when k is adjusted for the infrared images while k = 2.1 for the visible images. Figure 3 shows that k = 1.4 gave the highest matching accuracy rate. Thus, given the physical wavelength range of the two matching images, we can determine certain wavelet wavelengths for the images from different sensors. Thus, the proposed method can be generalized to images from heterologous sensors.

3.2. Corners Detection Combining Distinct Wavelength Phase Congruency of Original Images and Gaussian Smoothing Images

It is not enough to only use the phase congruency of the original images. In [59], a corner detection algorithm based on Gaussian smoothing has been proposed and an improved corner detection algorithm based on Gaussian smoothing has been put forward by Wang et al. [60]. The proposed approaches combined the phase congruency of the original images and the Gaussian smoothed images to obtain more accurate feature points, in which maximum and minimum moments of the phase congruency for the Gaussian smoothed images were considered as the weighted image added to the corresponding moments of the original image. A two-dimensional Gaussian function is used as a tool for analysis in the linear scale-space and is defined as
G ( x , y , σ ) = 1 2 π σ 2 e x 2 + y 2 2 σ 2 .
In Figure 4a is the maximum moment of the phase congruency for the original image and Figure 4b is the maximum moment of the phase congruency for the smoothed image; Figure 4c is the enhanced maximum moment of the phase congruency obtained by adding Figure 4a,b; Meanwhile, Figure 4d,e are the minimum moment of the phase congruency for the original and smoothed images, respectively; And Figure 4f is the enhanced minimum moment of the phase congruency. Finally, the corners are detected by the enhanced maximum and minimum moment images.

3.3. Feature Descriptor Based on Log-Gabor Filters and the Corresponding Keypoint Detection Using the RANSAC Algorithm

A Gabor wavelet is a series of Gaussian envelopes of plane waves, which can extract spatial frequencies and local structural characteristics within the local area of the images in multiple directions and has some tolerance to changes in the spectrum, displacement, deformation, rotation, scaling, and illumination. In this paper, log-Gabor filters are proposed to describe the image features which exceed that of Gabor filters, due to the independence of the DC component [61]. As shown in Figure 1l, we obtain 24 energy images corresponding to four scales and six orientations and, then, the orientation information with maximum energy for a pixel is determined, in order to obtain the orientation image (Figure 1m). Finally, the orientation histogram is calculated (with six bins) for the orientation image, which is regarded as representing the feature descriptors, as shown in Figure 1n. For the four scales, we obtain a 4 × 6 dimensional vector, which can be regarded as the feature descriptor. As shown in Figure 5, we usually divide the 100 × 100 region centered on a given pixel into 16 subregions of size 25 × 25 , from which we can obtain 16 histograms with 4 × 6 bins (i.e., giving a 4 × 6 × 4 × 4 dimensional vector) as the feature descriptors. However, in our approach, we selected a 40 × 40 subregion with an interval of 20 pixels for rows and columns in a given 100 × 100 region, which allows us to obtain the same 16 subregions and obtain a 4 × 6 × 4 × 4 dimensional vector as a feature descriptor for a given pixel. The experiment illustrated that large, overlapping subregions include more information, increasing the performance of image matching.
In our experiment, the nearest neighbor distance ratio matching similarity metric was used to find corresponding keypoints. Two keypoints from the reference image and object image were considered coincident when they satisfied the following equation:
D ( d i , d j ) < t h D ( d i , d k ) ,
where D ( . , . ) is the Euclidean distance, d i is the keypoint in the reference image, d j and d k are the first and the second closest keypoints to d i in the target image, respectively, and t h is the threshold of the distance, which was obtained based on experience. Finally, the RANSAC algorithm was used to reduce the outliers; the result of the corresponding keypoint pairs is shown in Figure 6. The experimental results showed that the proposed partition approach had a perfect effect.

3.4. Similarity Computation of the Points Sets from the Visible and Infrared Images by BiDimRegressional Regression Modeling

For realistic application, we need to know about the similarity between two points sets from infrared and visible images to determine whether the image matching is successful or not. Thus, BiDimRegressional regression modeling was proposed to compute the similarity for geometrics of two points sets in the infrared and visible images. The main purpose of the bidimensional regression was to estimate the degree of correspondence between two plane patterns of point locations as the points sets shown in the infrared and visible images [62]. In our experiments, BiDimRegression package has been implemented and tested in ordinary R, whose input parameter is coord containing the coordinates of the independent (A,B) and the dependent configurations (X,Y).
  • Vectors A,B represent the coordinates of the visible image, which are extracted from the independent image possessing the relationships with the corresponding infrared image from the same scene. A and B are known as the first and the second dimension, respectively.
  • Vectors X,Y represent the coordinates of the infrared image, which are extracted from the independent image possessing the relationships with the corresponding visible image from the same scene. X and Y are known as the first and the second dimension, respectively.
In this algorithm, the main output values include: R 2 , F, p.
  • R 2 is the squared regression coefficient.
  • F represents F statistics for the overall regression model including appendant degrees of freedom(df1,df2).
  • p value is the accordant significance level.
By comparing the coordinates of the points sets, we found a very high correlation for the images with higher AR (e.g., for the Euclidean r = 0.999 , F = 7253.802 , p = 2 × 10 16 ) and a very low correlation for the images with lower AR (e.g., for the Euclidean r = 0.651 , F = 2.987 , p = 0.16 ). Thus, we can determine the similarity of the infrared and visible images by the range of the parameters in realistic applications.

4. Experiments

In this section, we first describe the experimental data and evaluation measures. Then, four algorithms are compared and analyzed. The performance of the proposed approach is illustrated on visible and infrared images. The experiments show encouraging results.

4.1. Experiment Data

In order to demonstrate the advantages of the proposed approach for visible and infrared image matching, it was tested on four different data sets which have been widely used in the literature. The data sets used in the experiments are freely available, which guarantees the reproducibility of the results, and can also be utilized to improve other approaches. The Potsdam, NIR, and RGB-LWIR data sets can be found in [46] and the Multimodel Stereo Data set 2 can be obtained in [63].
As shown in Figure 7a, the first data set is the Potsdam data set, which is a data set of remote sensing images distributed by the International Society for Photogrammetry and Remote Sensing. It includes 38 aerial images extracted from a larger TOP (true orthophoto) mosaic, in the form of TIFF with different channel compositions. In each channel, it has a spectral resolution of 8 bits and all images have dimension 6000 × 6000 pixels, including the IRRG, RGB, and RGBIR forms. In our experiment, the RGBIR forms were used to measure the proposed algorithms.
The second data set is the RGB-near infrared (NIR) dataset, shown in Figure 7b, which consists of 477 images in nine categories obtained in RGB and near-infrared. The images were captured with separate exposures from modified SLR cameras, using visible and NIR filters. The scenes include country, field, forest indoor, mountain, old buildings, street, urban, and water, and all images have dimension 1024 × 768 pixels. The country data set was utilized in the algorithm comparison experiment.
The third data set is the RGB-LWIR data set, as shown in Figure 7c, which was proposed in a previous study. It consists of 44 pairs of visible spectrum images and long-wave infrared images (LWIR), where all images are of size 639 × 431 pixels. The LWIR band is the most distant infrared band from the visible spectrum, and the image pairs mostly include common shape information, while most of the texture information is missing.
The fourth data set is the Multimodal Stereo Data set 2, which was used to supplement the evaluation of the proposed algorithm, as shown in Figure 7d. This data set contains outdoor images of different urban scenes, consisting of 100 pairs of VS-LWIR images of different outdoor urban scenarios. All images have size 506 × 408 pixels. All the images were rectified and aligned, so that matches could be obtained in horizontal lines. Using this data set, it was further proved that the proposed algorithm was superior to the state-of-the-art algorithms.
In these data set, the matching precision for the first two data set was higher than that in the other two data sets. This fact was principally due to the spectral band closeness of the image pairs. The NIR spectrum is the closest infrared band to the visible spectrum; however, the LWIR band is the most distant infrared band from the visible spectrum.

4.2. Evaluation Measures

In some sense, an ideal local feature should be a point defined geometrically with a location in space, but no spatial extent. In practice, images are discrete with a smallest spatial unit, known as a pixel, whose discrete property plays an important role. Ideally, one needs such local features with semantically meaningful object parts. However, this is unfeasible in practice, since this would demand a a deep interpretation of the scene content, which is not available at an early stage. On the contrary, local features can be employed to represent intensity patterns directly. At present, many evaluation measures have been proposed for the evaluation of the performance of image matching [64,65,66]. However, these evaluation measures are typically incomplete. Good features usually possess six properties [67]: repeatability, distinctiveness/informativeness, locality, quantity, accuracy, and efficiency. For measuring detector performance, according to these six properties, we advocate the following five evaluation measures:
  • Repeatable rate (RPR): as shown in Figure 8, the red points show repeatable keypoints and the green points show non-matching keypoints. In two images (the infrared image in Figure 8a and the visible image in Figure 8b), the high percentage of repeatable keypoints in both images is called repeatability. Thus, the repeatable rate (RPR) is defined as follows:
    R P R = C K N K N ,
    where C K N and K N are the numbers of true corresponding keypoints and total keypoints, respectively. A higher R P R value denotes that the keypoint detection approach has a better performance.
  • Recall rate (RR): this is shown as Figure 9, he red keypoints are the true matched corresponding keypoints and the yellow ones are non-matched keypoints. A better keypoint detection algorithm should detect more corresponding keypoints accurately over the repeatable keypoints set. Thus, the recall rate (RR) can be defined as:
    R R = D T M P N D T M P N + U D T M P N ,
    where D T M P N and U D T M P N are the detected true matched point number and the undetected true matched point number, respectively. The higher the value of the RR is, the better the performance of the matching approach.
  • Accuracy rate (AR): as shown in Figure 10, the detected corresponding keypoints should be accurate; the red lines show accurately detected corresponding keypoints and the green lines represent those which were not accurate. More red lines denote better performance of the image matching. Thus, the accuracy rate (AR) is defined as:
    A R = D T M P N D C P N ,
    where D T M P N and D C P N are the numbers of detected true matched points and detected corresponding points, respectively. The higher the AR is, the better the performance of the matching approach.
  • Quantity rate (QR): Figure 11a demonstrates that a larger number of keypoints will generate more corresponding keypoints favorable to image matching. Figure 11b, in contrast, shows that a smaller number of keypoints can cause failure in image matching. Thus, the number of detected keypoints should be sufficiently large; for example, a reasonable number of keypoints should be detected, even on small objects. Nevertheless, the optimal number of features depends on the application. Thus, we define the QR as:
    Q R = D K N I P N ,
    where D K N and I P N are the numbers of detected keypoints and image pixels, respectively. A higher QR leads to better performance of the image matching, but also results in a slower speed of image matching. Thus, an adaptive QR should be determined for special applications.
  • Efficiency (EF): The detection of features in an image should be considered as time-critical application. Thus, we defined the E F as:
    E F = T C K D + T C K D M + T O R ,
    where T C K D , T C K D M , and T R O are known as the times of candidate keypoint detection, corresponding keypoint detection, and outlier removal, respectively. In order to satisfy time-critical applications, we must make E F small enough.

4.3. Experiment Results Comparison and Discussion

The proposed algorithm was evaluated on the four data sets using the five evaluation measures and was compared with five existing algorithms, where the results showed that the proposed algorithm had better matching performance. The compared algorithms were the edge-oriented histogram descriptor (EHD) [68], phase congruency edge-oriented histogram descriptor (PCEHD) [9], and log-Gabor histogram descriptor (LGHD) [46] algorithms.
  • Edge-oriented histogram descriptor (EHD): this algorithm first detects the contour of the image and, then, the edge histogram descriptor is obtained by using the MPEG-7 standard [69].
  • Phase congruency edge-oriented histogram descriptor (PCEHD): This algorithm uses phase congruency to detect corners and edges in the image. In the corners, the EHD algorithm is utilized to obtain the feature descriptors.
  • Log-Gabor feature descriptor (LGHD): this algorithm uses a fast algorithm to detect corners and log-Gabor filters to generate feature descriptors.
  • Phase congruency log-Gabor image matching (PCLGM): This algorithm uses phase congruency to detect corners and edges in the image. In these corners, log-Gabor histograms are utilized to obtain the feature descriptors in overlapping subregions.
  • Modified phase congruency log-Gabor image matching(MPCLGM): This algorithm modified corners detection based on phase congruency by combining the original images and Gaussian smoothed images. In these corners, log-Gabor feature descriptors are utilized to obtain the feature descriptors.
  • Distinct modified phase congruency log-Gabor image matching(DMPCLGM): Distinct wavelengths are employed in the DMPCLGM for the visible and infrared images.
Figure 12 demonstrates the results obtained with one pair of VS-LWIR images using EHD, PCEHD, LGHD, and the proposed approach. It can be observed that the AR of EHD and PCEHD were 0.015 and 0.01, respectively, and the AR of the proposed approach was 91.5%. This result shows that the proposed approach must have advantages over the other approaches. Table 1, Table 2, Table 3 and Table 4 present the comparison results for the RGB-LWIR, Multimodal Stereo 2, RGB-NIR, and Potsdam data sets respectively.
In Table 1, it can be seen that PCLGM, MPCLGM, and DMPCLGM had higher RPR, AR, and QR, but the RR was lower. In image matching methods, AR is more necessary than RR, which means that higher AR was more meaningful. The results of the comparisons for the six algorithms illustrate that combining the phase congruency of the original images and the smoothed images to detect corners and extracting the feature histogram through subregion extraction played an important role in image matching using different wavelengths of the wavelet for the visible and infrared images. For DMPCLGM, the wavelengths of the wavelets for the visible and infrared images were set to 2.1 and 1.4, which advanced the performance of the algorithm.
In Table 2, the results were similar to those for the RGB-LWIR data set, which presented higher RPR, AR, and QR, while the RR was lower. It was, thus, also shown that the proposed methods presented advantages over the state-of-the-art algorithms when used for visible and infrared images.
In Table 3 and Table 4, for DMPCLGM, all wavelengths of the wavelet for the visible and infrared images were set to 2.1, due to the similarities of the physical wavelengths. The compared results show that all the approaches had higher AR. However, the proposed algorithm showed no special advantages. Therefore, our proposed algorithm still needs to be improved to suit images with different physical wavelengths.
As shown in Table 1, Table 2, Table 3 and Table 4, the proposed algorithm demonstrated perfect performance. The lower QR at the corners caused the state-of-the-art methods to present lower performance. However, the higher QR at the corners also caused lower speeds. Thus, it is necessary to guarantee a balance between QR and AR in the proposed algorithm.

5. Conclusions

This paper presented a feature point matching approach based on distinct wavelength phase congruency and log-Gabor filters for infrared and visible images. First, the use of phase congruency was proposed to determine corners in the images. Generally, it is not proper to set the same wavelengths for PC when using two matched images with different physical wavelengths. Thus, the PC wavelengths for the two images were determined experimentally. Next, the moments of the PCs for the original image and Gaussian smoothed image were combined to detect the corners. Finally, log-Gabor filters were used and overlapping subregions were extracted for generating the descriptors. Five evaluation measures were employed for testing the performance of the algorithm. The experimental results show that the proposed algorithm with different wavelengths for infrared and visible images showed superior performance than other state-of-the-art approaches. Combining the original image and Gaussian smoothed image for computing the moments of the PCs also advances the performance of the algorithm. The BiDimRegression regression modeling is proposed to determine the similarity of the infrared and visible images in realistic applications.
Image matching for visible and infrared images presents many challenges for images captured using different sensors, which requires more effort to increase the performance of the approaches. The proposed approaches still have two main problems: (1) although the proposed approach shows some advantages over other methodologies, it still has a lower accuracy rate, and (2) due to the complexity of the algorithm, its runtime is too long. In the future, we will study how to remove the outlier points in order to increase the accuracy rate and adjust the number of keypoints to reduce the runtime. Meanwhile, we will extend the proposed approach to multimodal image matching.

Author Contributions

For research articleswith three authors, conceptualization, X.L. and J.-B.L.; methodology, X.L.; software, X.L.; validation, X.L., J.-B.L. and J.-S.P.; formal analysis, X.L.; investigation, X.L.; resources, X.L.; data curation, X.L.; writing—original draft preparation, X.L.;writing—review and editing, X.L.; visualization, X.L.; supervision, J.-B.L.; project administration, J.-B.L.; funding acquisition, J.-B.L., J.-S.P., X.L.

Funding

This research was funded by National Science Foundation of China grant number 61671170 and 61872085, the Open Projects Program of National Laboratory of Pattern Recognition grant number 201700019,Heilongjiang Provincial Natural Science Foundation of China grant grant number QC2015072, and Jiamusi University Young Innovative Talents Training Program grant number 22Zq201506.

Acknowledgments

First and foremost, I would like to show my deepest gratitude to my supervisor, Jun-Bao Li, a respectable, responsible and resourceful scholar, who has provided me with valuable guidance in every stage of the writing of this paper. I shall extend my thanks to Jeng-Shyang Pan for all his kindness and help. I would also like to thank all the who have helped me to develop the fundamental and essential academic competence.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zheng, H.; Li, S.Y.; Shao, Y.Y.; Yang, S. Typical Building of Multi-Sensor Image Feature Extraction and Recognition. Artif. Intell. Sci. Technol. 2017, 259–272. [Google Scholar] [CrossRef]
  2. Son, J.; Kim, S.; Sohn, K. A Multi-vision Sensor-based Fast Localization System with Image Matching for Challenging Outdoor Environments. Expert Syst. Appl. 2015, 42, 8830–8839. [Google Scholar] [CrossRef]
  3. Xu, Y.; Zhou, J.; Zhuang, L. Binary auto encoding feature for multi-sensor image matching. In Proceedings of the 2016 Fourth International Conference on Ubiquitous Positioning, Indoor Navigation and Location Based Services (UPINLBS), Shanghai, China, 2–4 November 2016; pp. 278–282. [Google Scholar] [CrossRef]
  4. Conte, G.; Doherty, P. An Integrated UAV Navigation System Based on Aerial Image Matching. In Proceedings of the 2008 IEEE Aerospace Conference, Big Sky, MT, USA, 1–8 March 2008; pp. 1–10. [Google Scholar] [CrossRef]
  5. Zhang, Y.F.; Hui, Y. Study on Guidance Algorithm of Scene Matching based on Different Source Images for Cruise Missile. In Proceedings of the 2018 IEEE International Conference on Mechatronics and Automation (ICMA), Changchun, China, 5–8 August 2018; pp. 912–917. [Google Scholar] [CrossRef]
  6. Wang, L.; Han, J.; Zhang, Y.; Bai, L. Image fusion via feature residual and statistical matching. IET Comput. Vis. 2016, 10, 551–558. [Google Scholar] [CrossRef]
  7. Yang, W.; Zhong, L.; Chen, Y.; Lin, L.; Lu, Z.; Liu, S.; Wu, Y.; Feng, Q.; Chen, W. Predicting CT Image From MRI Data Through Feature Matching With Learned Nonlinear Local Descriptors. IEEE Trans. Med. Imaging 2018, 37, 977–987. [Google Scholar] [CrossRef] [PubMed]
  8. Pan, P.; Liu, J.; Zeng, Q.; Wang, Y.; Liu, S. Image matching for target location in airborne optoelectronic pod. In Proceedings of the 2016 IEEE Chinese Guidance, Navigation and Control Conference (CGNCC), Nanjing, China, 12–14 August 2016; pp. 1330–1332. [Google Scholar] [CrossRef]
  9. Mouats, T.; Aouf, N.; Sappa, A.D.; Aguilera, C.; Toledo, R. Multispectral Stereo Odometry. IEEE Trans. Intell. Transp. Syst. 2015, 16, 1210–1224. [Google Scholar] [CrossRef]
  10. Huang, Q.; Yang, J.; Wang, C.; Chen, J.; Meng, Y. Improved registration method for infrared and visible remote sensing image using NSCT and SIFT. In Proceedings of the 2012 IEEE International Geoscience and Remote Sensing Symposium, Munich, Germany, 22–27 July 2012; pp. 2360–2363. [Google Scholar] [CrossRef]
  11. Hariharan, H.; Koschan, A.; Abidi, B.; Gribok, A.; Abidi, M. Fusion of Visible and Infrared Images using Empirical Mode Decomposition to Improve Face Recognition. In Proceedings of the 2006 International Conference on Image Processing, Atlanta, GA, USA, 8–11 October 2006; pp. 2049–2052. [Google Scholar] [CrossRef]
  12. Cheng, K.; Lin, H. Automatic target recognition by infrared and visible image matching. In Proceedings of the 2015 14th IAPR International Conference on Machine Vision Applications (MVA), Tokyo, Japan, 18–22 May 2015; pp. 312–315. [Google Scholar] [CrossRef]
  13. Aguilera, C.; Barrera, F.; Lumbreras, F.; Sappa, A.D.; Toledo, R. Multispectral Image Feature Points. Sensors 2012, 12, 12661–12672. [Google Scholar] [CrossRef] [Green Version]
  14. Pan, J.S.; Kong, L.; Sung, T.W.; Tsai, P.W.; Snasel, V. α-Fraction First Strategy for Hierarchical Model in Wireless Sensor Networks. J. Internet Technol. 2018, 19, 1717–1726. [Google Scholar] [CrossRef]
  15. Pan, J.S.; Kong, L.; Sung, T.W.; Tsai, P.W.; Snášel, V. A Clustering Scheme for Wireless Sensor Networks Based on Genetic Algorithm and Dominating Set. J. Internet Technol. 2018, 19, 1111–1118. [Google Scholar] [CrossRef]
  16. Pan, J.S.; Lee, C.Y.; Sghaier, A.; Zeghid, M.; Xie, J. Novel Systolization of Subquadratic Space Complexity Multipliers Based on Toeplitz Matrix–Vector Product Approach. IEEE Trans. Large Scale Integr. Syst. 2019, 27, 1614–1622. [Google Scholar] [CrossRef]
  17. Wang, J.; Gao, Y.; Liu, W.; Sangaiah, A.K.; Kim, H.J. An Improved Routing Schema with Special Clustering using PSO Algorithm for Heterogeneous Wireless Sensor Network. Sensors 2019, 19, 671. [Google Scholar] [CrossRef]
  18. Wang, J.; Cao, J.; Sherratt, R.S.; Park, J.H. An improved ant colony optimization-based approach with mobile sink for wireless sensor networks. J. Supercomput. 2018, 74, 6633–6645. [Google Scholar] [CrossRef]
  19. Wang, J.; Gao, Y.; Liu, W.; Sangaiah, A.K.; Kim, H.J. Energy Efficient Routing Algorithm with Mobile Sink Support for Wireless Sensor Networks. Sensors 2019, 19, 1494. [Google Scholar] [CrossRef] [PubMed]
  20. Kovesi, P. Image Features From Phase Congruency. Videre J. Comput. Vision Res. 1999, 1, 1–26. [Google Scholar]
  21. Field, D.J. Relations between the statistics of natural images and the response properties of cortical cells. J. Opt. Soc. Am. A 1987, 4, 2379–2394. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  22. Daugman, J. Uncertainty relation for resolution in space, spatial frequency, and orientation optimized by two-dimensional visual cortical filters. J. Opt. Soc. Am. Opt. Image Sci. 1985, 2, 1160–1169. [Google Scholar] [CrossRef]
  23. Daugman, J. Statistical Richness of Visual Phase Information: Update on Recognizing Persons by Iris Patterns. Int. J. Comput. Vis. 2001, 45, 25–38. [Google Scholar] [CrossRef]
  24. Kovesi, P.D. Phase congruency detects corners and edges. In Proceedings of the The Australian Pattern Recognition Society Conference, Sydney, Australia, 10–12 December 2003; pp. 309–318. [Google Scholar]
  25. Avants, B.B.; Epstein, C.; Grossman, M.; Gee, J. Symmetric Diffeomorphic Image Registration with Cross-Correlation: Evaluating Automated Labeling of Elderly and Neurodegenerative Brain. Med. Image Anal. 2008, 12, 26–41. [Google Scholar] [CrossRef]
  26. Yi, X.; Wang, B.; Fang, Y.; Liu, S. Registration of infrared and visible images based on the correlation of the edges. In Proceedings of the 2013 6th International Congress on Image and Signal Processing (CISP), Hangzhou, China, 16–18 December 2013; pp. 990–994. [Google Scholar] [CrossRef]
  27. Zhuang, Y.; Gao, K.; Miu, X.; Han, L.; Gong, X. Equation Infrared and Visual Image Registration Based on Mutual Information with a Combined Particle Swarm Optimization-Powell Search Algorithm. Opt. Int. J. Light Electron. Opt. 2015, 127, 188–191. [Google Scholar] [CrossRef]
  28. Sinisa, T.; Narendra, A. Region-based hierarchical image matching. Int. J. Comput. Vis. 2008, 78, 47–66. [Google Scholar] [CrossRef]
  29. Bhat, K.K.S.; Heikkilä, J. Line Matching and Pose Estimation for Unconstrained Model-to-Image Alignment. In Proceedings of the 2nd International Conference on 3D Vision, Tokyo, Japan, 8–11 December 2014; pp. 155–162. [Google Scholar] [CrossRef]
  30. Senthilnath, J.; Kalro, N.P. Accurate point matching based on multi-objective Genetic Algorithm for multi-sensor satellite imagery. Appl. Math. Comput. 2014, 236, 546–564. [Google Scholar] [CrossRef]
  31. Saleem, S.; Bais, A.; Sablatnig, R.; Ahmad, A.; Naseer, N. Feature points for multisensor images. Comput. Electr. Eng. 2017, 62, 511–523. [Google Scholar] [CrossRef]
  32. Lowe, D.G. Distinctive Image Features from Scale-Invariant Keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
  33. Bay, H.; Ess, A.; Tuytelaars, T.; Goolab, L.V. SURF: Speed-Up Robust Features. Comput. Vis. Image Underst. 2007, 110, 346–359. [Google Scholar] [CrossRef]
  34. Leutenegger, S.; Chli, M.; Siegwart, R.Y. BRISK: Binary Robust invariant scalable keypoints. In Proceedings of the 2011 International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011; pp. 2548–2555. [Google Scholar] [CrossRef]
  35. Rublee, E.; Rabaud, V.; Konolige, K.; Bradski, G. ORB: An efficient alternative to SIFT or SURF. In Proceedings of the 2011 International Conference on Computer Vision, ICCV2012, Barcelona, Spain, 6–13 November 2011; pp. 1–8. [Google Scholar] [CrossRef]
  36. Chen, J.; Shan, S.; He, C.; Zhao, G.; Pietikainen, M.; Chen, X.; Gao, W. WLD: A Robust Local Image Descriptor. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 32, 1705–1720. [Google Scholar] [CrossRef] [PubMed]
  37. Tang, F.L.T. A Webber local binary pattern descriptor for pancreas endoscopic ultrasound image classification. In Proceedings of the 2014 IEEE Workshop on Electronics, Computer and Applications, Ottawa, ON, Canada, 8–9 May 2014; pp. 836–839. [Google Scholar] [CrossRef]
  38. Li, Q.; Wang, G.; Liu, J.; Chen, S. Robust Scale-Invariant Feature Matching for Remote Sensing Image Registration. IEEE Geosci. Remote Sens. Lett. 2009, 6, 287–291. [Google Scholar] [CrossRef]
  39. Dana, J.; Anandan, P. Registration of visible and infrared images. Archit. Hardware Forw. Look. Infrared Issues Autom. Object Recognit. 1993, 1957, 2–13. [Google Scholar] [CrossRef]
  40. Firmenichy, D.; Brown, M.; Süsstrunk, S. Multispectral interest points for RGB-NIR image registration. In Proceedings of the 2011 18th IEEE International Conference on Image Processing, Brussels, Belgium, 11–14 September 2011; pp. 181–184. [Google Scholar] [CrossRef]
  41. Yi, Z.; Zhiguo, C.; Yang, X. Multi-spectral remote image registration based on SIFT. Electron. Lett. 2008, 44, 107–108. [Google Scholar] [CrossRef]
  42. Saleem, S.; Bais, A.; Sablatnig, R. A Performance Evaluation of SIFT and SURF for Multispectral Image Matching. In Image Analysis and Recognition; Campilho, A., Kamel, M., Eds.; Springer: Berlin/Heidelberg, Germany, 2012; pp. 166–173. [Google Scholar]
  43. Sonn, S.; Bilodeau, G.; Galinier, P. Fast and Accurate Registration of Visible and Infrared Videos. In Proceedings of the 2013 IEEE Conference on Computer Vision and Pattern Recognition Workshops, Portland, OR, USA, 23–28 June 2013; pp. 308–313. [Google Scholar] [CrossRef]
  44. Kim, S.; Ryu, S.; Ham, B.; Kim, J.; Sohn, K. Local self-similarity frequency descriptor for multispectral feature matching. In Proceedings of the 2014 IEEE International Conference on Image Processing (ICIP), Paris, France, 27–30 October 2014; pp. 5746–5750. [Google Scholar] [CrossRef]
  45. Torabi, A.; Bilodeau, G. Local self-similarity-based registration of human ROIs in pairs of stereo thermal-visible videos. Pattern Recognit. 2013, 46, 578–589. [Google Scholar] [CrossRef]
  46. Cristhian, A.; Angel, S.; Ricardo, T. LGHD: A feature descriptor for matching across non-linear intensity variations. In Proceedings of the 2015 IEEE International Conference on Image Processing (ICIP), Quebec City, QC, Canada, 27–30 September 2015. [Google Scholar] [CrossRef]
  47. Li, Y.; Zou, J.; Jing, J.; Jin, H.; Yu, H. Establish keypoint matches on multispectral images utilizing descriptor and global information over entire image. Infrared Phys. Technol. 2016, 76, 1–10. [Google Scholar] [CrossRef]
  48. Qin, Y.; Cao, Z.; Zhuo, W.; Yu, Z. Robust key point descriptor for multi-spectral image matching. J. Syst. Eng. Electron. 2014, 25, 681–687. [Google Scholar] [CrossRef]
  49. Li, Y.; Jin, H.; Qiao, W. Robustly building keypoint mappings with global information on multispectral images. EURASIP J. Adv. Signal Process 2015, 2015, 53. [Google Scholar] [CrossRef]
  50. Fu, Z.; Qin, Q.; Luo, B.; Wu, C.; Sun, H. A Local Feature Descriptor Based on Combination of Structure and Texture Information for Multispectral Image Matching. IEEE Geosci. Remote Sens. Lett. 2019, 16, 100–104. [Google Scholar] [CrossRef]
  51. Daugman, J.G. Two-dimensional spectral analysis of cortical receptive field profiles. Vis. Res. 1980, 20, 847–856. [Google Scholar] [CrossRef]
  52. Nunes, C.F.G.; Pádua, F.L.C. A Local Feature Descriptor Based on Log-Gabor Filters for Keypoint Matching in Multispectral Images. IEEE Geosci. Remote Sens. Lett. 2017, 14, 1850–1854. [Google Scholar] [CrossRef]
  53. Rosten, E.; Drummond, T. Fusing points and lines for high performance tracking. In Proceedings of the Tenth IEEE International Conference on Computer Vision, Beijing, China, 17–21 October 2005; pp. 1508–1515. [Google Scholar] [CrossRef]
  54. Gao, W.; Kwong, S.; Zhou, Y.; Jia, Y.; Zhang, J.; Wu, W. Multiscale phase congruency analysis for image edge visual saliency detection. In Proceedings of the 2016 International Conference on Machine Learning and Cybernetics (ICMLC), Jeju, Korea, 10–13 July 2016; pp. 75–80. [Google Scholar] [CrossRef]
  55. Fan, J.; Wu, Y.; Li, M.; Liang, W.; Cao, Y. SAR and Optical Image Registration Using Nonlinear Diffusion and Phase Congruency Structural Descriptor. IEEE Trans. Geosci. Remote Sens. 2018, 56, 5368–5379. [Google Scholar] [CrossRef]
  56. Ma, W.; Wu, Y.; Liu, S.; Su, Q.; Zhong, Y. Remote Sensing Image Registration Based on Phase Congruency Feature Detection and Spatial Constraint Matching. IEEE Access 2018, 6, 77554–77567. [Google Scholar] [CrossRef]
  57. Zhu, Z.; Zheng, M.; Qi, G.; Wang, D.; Xiang, Y. A Phase Congruency and Local Laplacian Energy Based Multi-Modality Medical Image Fusion Method in NSCT Domain. IEEE Access 2019, 7, 20811–20824. [Google Scholar] [CrossRef]
  58. Morrone, M.; Owens, R. Feature detection from local energy. Pattern Recognit. Lett. 1987, 6, 303–313. [Google Scholar] [CrossRef]
  59. Mokhtarian, F.; Suomela, R. Robust image corner detection through curvature scale space. IEEE Trans. Pattern Anal. Mach. Intell. 1998, 20, 1376–1381. [Google Scholar] [CrossRef]
  60. Wang, X.; Shi, B.E. GPU implemention of fast Gabor filters. In Proceedings of the 2010 IEEE International Symposium on Circuits and Systems, Paris, France, 30 May–2 June 2010; pp. 373–376. [Google Scholar] [CrossRef]
  61. Nava, R.; Escalante, B.; Cristobal., G. Texture image retrieval based on log-Gabor features. In Progress in Pattern Recognition, Image Analysis, Computer Vision, and Applications; Springer: Berlin/Heidelberg, Germany, 2012; pp. 414–421. [Google Scholar] [CrossRef]
  62. Carbon, C.C. BiDimRegression: Bidimensional Regression Modeling Using R. J. Stat. Softw. 2013, 52, 1–11. [Google Scholar] [CrossRef]
  63. Barrera, F.; Lumbreras, F.; Sappa, A.D. Multispectral piecewise planar stereo using Manhattan-world assumption. Pattern Recognit. Lett. 2013, 34, 52–61. [Google Scholar] [CrossRef]
  64. Mikolajczyk, K.; Schmid, C. A performance evaluation of local descriptors. In Proceedings of the 2003 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Madison, WI, USA, 18–20 June 2003. [Google Scholar] [CrossRef]
  65. Zhu, Z.; Davari, K. Comparison of Local Visual Feature Detectors and Descriptors for the Registration of 3D Building Scenes. J. Comput. Civ. Eng. 2014, 29, 04014071. [Google Scholar] [CrossRef]
  66. Mikolajczyk, K.; Tuytelaars, T.; Schmid, C.; Zisserman, A.; Matas, J.; Schaffalitzky, F.; Kadir, T.; Gool, L.V. A Comparison of Affine Region Detectors. Int. J. Comput. Vis. 2005, 65, 43–72. [Google Scholar] [CrossRef]
  67. Tuytelaars, T.; Mikolajczyk, K. Local Invariant Feature Detectors: A Survey. Found. Trends Comput. Graph. Vis. 2008, 3, 177–280. [Google Scholar] [CrossRef]
  68. Manjunath, B.S.; Ohm, J.; Vasudevan, V.V.; Yamada, A. Color and texture descriptors. IEEE Trans. Circuits Syst. Video Technol. 2001, 11, 703–715. [Google Scholar] [CrossRef] [Green Version]
  69. Sikora, T.; Zhong, L.; Chen, Y.; Lin, L.; Lu, Z.; Liu, S.; Wu, Y.; Feng, Q.; Chen, W. The MPEG-7 visual standard for content description-an overview. IEEE Trans. Circuits Syst. Video Technol. 2001, 11, 696–702. [Google Scholar] [CrossRef]
Figure 1. Overview of feature point matching based on phase congruency and log-Gabor filters for infrared and visible images: (a,b) show the target image and the reference image; (c,d) show phase congruency (PC) images for six orientations in the visible and infrared images; (eh) show maximum and minimum moments of the phase congruency for the visible and infrared images; (i,j) show the corners in the visible and infrared images; (k) shows the corresponding pairs; (l) shows the complex images in the real domain obtained with log-Gabor filters; (m) shows the phase image with maximum energy; and (n) shows the orientation histograms of the phase images.
Figure 1. Overview of feature point matching based on phase congruency and log-Gabor filters for infrared and visible images: (a,b) show the target image and the reference image; (c,d) show phase congruency (PC) images for six orientations in the visible and infrared images; (eh) show maximum and minimum moments of the phase congruency for the visible and infrared images; (i,j) show the corners in the visible and infrared images; (k) shows the corresponding pairs; (l) shows the complex images in the real domain obtained with log-Gabor filters; (m) shows the phase image with maximum energy; and (n) shows the orientation histograms of the phase images.
Sensors 19 04244 g001
Figure 2. Illustration of corner detection.
Figure 2. Illustration of corner detection.
Sensors 19 04244 g002
Figure 3. Trend of matching accuracy rate at different values of k.
Figure 3. Trend of matching accuracy rate at different values of k.
Sensors 19 04244 g003
Figure 4. Enhanced moments of the phase congruency images.
Figure 4. Enhanced moments of the phase congruency images.
Sensors 19 04244 g004
Figure 5. Partition of the neighborhood of a certain pixel.
Figure 5. Partition of the neighborhood of a certain pixel.
Sensors 19 04244 g005
Figure 6. Illustration of corresponding feature point matching in the visible and infrared images.
Figure 6. Illustration of corresponding feature point matching in the visible and infrared images.
Sensors 19 04244 g006
Figure 7. Examples from the data set.
Figure 7. Examples from the data set.
Sensors 19 04244 g007
Figure 8. Illustration of the repeatable keypoints.
Figure 8. Illustration of the repeatable keypoints.
Sensors 19 04244 g008
Figure 9. Illustration of the recall rate (RR) for the keypoints.
Figure 9. Illustration of the recall rate (RR) for the keypoints.
Sensors 19 04244 g009
Figure 10. Illustration of the accuracy rate (AR) for the keypoints.
Figure 10. Illustration of the accuracy rate (AR) for the keypoints.
Sensors 19 04244 g010
Figure 11. Illustration of the quantity rate (QR) for the keypoints.
Figure 11. Illustration of the quantity rate (QR) for the keypoints.
Sensors 19 04244 g011
Figure 12. Keypoints matched in two image pairs.
Figure 12. Keypoints matched in two image pairs.
Sensors 19 04244 g012
Table 1. Comparison of the six algorithms using the RGB-LWIR data set.
Table 1. Comparison of the six algorithms using the RGB-LWIR data set.
DescriptorRPRRRARQREF
EHD0.020.040.140.0050.28
PCEHD0.020.020.010.0051.05
LGHD0.020.020.010.0051.55
PCLGM0.190.0140.590.2740.3
MPCLGM0.1980.0130.620.3165.54
DMPCLGM0.270.010.640.4682
Table 2. Comparison of the six algorithms using the Multimodal Stereo Data set 2.
Table 2. Comparison of the six algorithms using the Multimodal Stereo Data set 2.
DescriptorRPRRRARQREF
EHD0.0250.0540.10.00180.4
PCEHD0.0340.0370.1230.00251.87
LGHD0.040.080.250.00311.56
PCLGM0.240.0030.210.4367.68
MPCLGM0.240.0130.260.4371.62
DMPCLGM0.240.00750.320.4378.01
Table 3. Comparison of the six algorithms using the RGB-NIR data set.
Table 3. Comparison of the six algorithms using the RGB-NIR data set.
DescriptorRPRRRARQREF
EHD0.0870.30.910.118.65
PCEHD0.0870.30.90.1111.41
LGHD0.0870.40.930.1130.3
PCLGM0.29760.02820.89910.227930.51
MPCLGM0.29760.02780.89190.2234130.40
DMPCLGM-----
Table 4. Comparison of the six algorithms using the Potsdam data set.
Table 4. Comparison of the six algorithms using the Potsdam data set.
DescriptorRPRRRARQREF
EHD0.220.270.990.064.48
PCEHD0.220.270.990.065.73
LGHD0.220.2810.0615.16
PCLGM0.226390.05280.72340.313469.05
MPCLGM0.226390.06040.72610.417096.1237
DMPCLGM-----

Share and Cite

MDPI and ACS Style

Liu, X.; Li, J.-B.; Pan, J.-S. Feature Point Matching Based on Distinct Wavelength Phase Congruency and Log-Gabor Filters in Infrared and Visible Images. Sensors 2019, 19, 4244. https://doi.org/10.3390/s19194244

AMA Style

Liu X, Li J-B, Pan J-S. Feature Point Matching Based on Distinct Wavelength Phase Congruency and Log-Gabor Filters in Infrared and Visible Images. Sensors. 2019; 19(19):4244. https://doi.org/10.3390/s19194244

Chicago/Turabian Style

Liu, Xiaomin, Jun-Bao Li, and Jeng-Shyang Pan. 2019. "Feature Point Matching Based on Distinct Wavelength Phase Congruency and Log-Gabor Filters in Infrared and Visible Images" Sensors 19, no. 19: 4244. https://doi.org/10.3390/s19194244

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop