Satish Nirala, Deepak Mishra*, K. Martin Sagayam,
D. Narain Ponraj, X. Ajay Vasanth, Lawrence Henesey, and
Chiung Ching Ho

6Image fusion in remote sensing based on sparse sampling method and PCNN techniques

Satish Nirala, Deepak Mishra, Department of Avionics, Indian Institute of Space and Technology, Trivandrum, India

K. Martin Sagayam, D. Narain Ponraj, X. Ajay Vasanth, School of Engineering and Technology, Karunya University, Coimbatore, India

Lawrence Henesey, Department of Systems and Software Engineering, Blekine Institute of Technology, Sweden

Chiung Ching Ho, Department of Computing and Informatics, Multimedia University, Malaysia

Abstract: Image fusion is a combination of multiple images that results in a fused image. It provides more information than any other input images and is based on discrete wavelet transformation (DWT) and a sparse sampling method. The sparse sampling method offers better performance than Nyquist theorem as a signal processing technique. Among the various techniques, DWT offers many advantages, such as yielding higher quality, requiring less storage and incurring lower costs, which is very useful in image applications. Image-related applications have few constraints, such as minimum data storage and lower bandwidth for communication that takes place through satellites, which actually results in the capture of low-quality images. Major technical constraints, such as minimum data storage for satellite imaging in space and lower bandwidth for communication with earth-based stations, etc., limit the ability of satellite sensors to capture images with high spatial and high spectral resolution simultaneously. To overcome this problem, image fusion has proved to be a potential tool for remote sensing applications that incorporate data from combinations of panchromatic, multispectral images and that aim to produce a composite image having both higher spatial and spectral resolutions. Research in this area can be traced back to the most recent couple of decades; unfortunately, the diverse methodological approaches proposed so far by different researchers have not been discussed. Experimental results indicate that a comparative analysis of two methods used in image fusion, such as compressive sensing and pulse-coupled neural network (PCNN), for a few remote sensing location data, e.g., England, forest land, Egypt, island, Balboa and bare soil, is used in this work. The following performance metrics have two assessment procedures, i.e., at full- and reduced-scale resolutions, to evaluate the performance of these algorithms.

Keywords: Compressive sensing, sparse representation, discrete wavelet transform, PCNN

6.1Introduction

Nowadays, image fusion has become a very interesting research topic involving many related areas such as automatic object detection, computer vision, medical imaging, remote sensing, robotics and night vision applications. Furthermore, the need for image fusion in computer vision systems is increasing dramatically. Image data must be extracted and combined from multiple images from the same scene. The result of image fusion is a single image, which is more appropriate and feasible for both humans and machines to observe for further image processing tasks. Multifocus and multisensor image fusion are the two important branches of image fusion. Due to the limited depth of field of optical lenses, it is difficult to obtain an image with all objects in focus. Multifocus image fusion attempts to increase the apparent depth of field through the fusion of objects within several different fields of focus. In many situations, a single sensor is unable to provide an exact view of real-time applications. Thus, multisensor data fusion has emerged as an extraordinary research area in recent years. Multisensor fusion is a process of binding data from various sensors into a solitary representational format. Image fusion methods are divided into two major classes: spatial domain image fusion and transform domain image fusion. Spatial domain image fusion includes fusion methods like principal component analysis (PCA), which is based on a fusion method, averaging-based image fusion, and minimum and maximum information-based image fusion methods. Transform domain image fusion includes fusion methods based on multiscale transform such as dual-tree complex wavelet transform (DTCWT) and discrete wavelet transform (DWT). PCNN was developed by Eckhorn in 1990 and based on experimental observations of synchronous pulse bursts in cat and monkey visual cortex. It is characterized by the global coupling and pulse synchronization of neurons. These features facilitate image fusion, which makes use of local image information. All fusion the aforementioned methods operate on the basis of whole sampling, which requires very high computational and high-complexity computer resources. It is a difficult problem for the increasing real-time demand placed on image fusion systems.Then compressed sensing (CS) is used to provide a way to resolve the problem of image fusion. CS theory enables a sparse or compressible signal to be properly reconstructed from a small number of non-adaptive linear projections. For image fusion at the pixel level, a small amount of measurements can be directly fused in the compressed domain and be reconstructed with high accuracy according to the fused measurements. It will greatly reduce the demand for computational resources and complexity. The implemented image fusion methods include the PCA method and the spatial or simple averaging method in the spatial domain category. DWT- , DTCWT-) and PCNN-based image fusion methods were applied in the transform domain category. Image fusion in a CS framework was also implemented, which includes the aforementioned transformed domain image fusion methods. Integrated image fusion, a technique based on PCNN and PCA with or without a CS framework, is proposed in this chapter. The proposed method gives clearer results methods currently used in the literature since the fusion process is carried out based on the different properties of high-pass- and low-pass-frequency components of images.

We focus on the results of an investigation into an objective evaluation of an image fusion method. The gradient-based objective fusion framework was tested by C. Xydeas and V. Petrovic [1], who make inferences on input gradients of the fused images under evaluation. The results and discussion section presents an analysis of image fusion methods evaluated using quality measures such as entropy (E), standard deviation (SD), mutual information (MI) and others.

6.1.1Motivation and scope

In the field of image processing a lot of developments have taken place recently, where image fusion is playing a vital role in research on night vision, military applications and others. Many imaging devices suffer from a limited depth of field, so there is a need for the development of better-quality fused images in multifocus image fusion, as suggested by S. Wei and W. Ke [2]. The works related to image fusion algorithms indicate that all existing algorithms degrade the brightness of fused images and are time consuming. Since image features differ from each other, we need a fusion method based on multiscale decomposition so that all the information in decomposed images can be acquired and fused.

In this chapter, an integrated image fusion method is developed and implemented. Image fusion using PCNN and PCA in a compressive sensing approach, where PCNN and PCA methods are used for fusing low-pass frequency and high-pass frequency component respectively. The performance measures from an experimental result improves of evaluation and visual effect metrics. It has been compared with the other conventional methods based on the frequency components of the image. The fusion results obtained using the proposed methods show a significant improvement in image quality visually and in terms of the evaluation metrics, since the fusion process is done based on the different properties on high pass and low pass frequency components of image.

6.1.2Literature survey

To date, many image fusion methods have been developed that vary in complexity and applications. The main objective of some of the popular image fusion techniques will be discussed in this section.

C. Polh and J. Van Genderen [3] discussed image fusion using PCA (PCA). PCA is used to extract the weighted feature points in image fusion. The principal components of the images are the eigenvectors corresponding to the largest eigenvalues of the co-variance matrix formed by the images. Hence, it reduces redundancy in the fused image data.

H. Yin et al. [4] presented an image fusion technique based on a sparse sampling method or CS. The execution rate is reduced by increasing the quality of the resultant image fused by the integration of non-zero coefficients present in the image. Many errors occur in the conventional approach, such as reconstruction errors, poor fidelity and blocking artefacts that produce uncertain results. To overcome these problems, another fusion-based approach was proposed that includes non-subsampled contourlet transform, which is widely used in decomposition techniques. The CS-based fusion approach can greatly reduce processing speed and guarantee the quality of fused images by integrating fewer non-zero coefficients. Due to the limitations in the conventional CS-based fusion approach, directly fusing sensing measurements may produce results that have greater uncertainty with a high rate of reconstruction errors, and using a single fusion rule may result in problems of blocking artefacts and poor fidelity. An image fusion approach based on CS was developed to solve these problems. The NSCT has been used for image decomposition. A dual-layer PCNN model was used to carry out low-pass sub-band fusion, while high-pass sub-band fusion was used to implement the edge-based fusion rule. For the reconstruction of fused images, a CS matching pursuit algorithm (CoSaMP) was propsed by D. Needell and J. A. Tropp [5].

The motivation behind using PCNN, used to transfer the spatial frequency domain using NSCT for an image fusion procedure, is discussed in Q. Xiao-Bo [6]. Non-subsampled contourlet transform (NSCT) provides flexible multiresolution, anisotropic and directional expansion of images. Compared with the original contourlet transform, it is shift-invariant and can overcome the pseudo-Gibbs phenomenon around singularities. The PCNN is a visual cortex-inspired neural network characterized by the global coupling pulse synchronization of neurons. It has proven suitable for image processing and successfully used in image fusion. Q. Xiao-Bo et al. [6] have shown through experimentation that NSCT is associated with PCNN and can be used in image fusion to make full use of image characteristics. Spatial frequency in the NSCT domain is used to motivate PCNN, and coefficients in the NSCT domain with large firing times are selected as coefficients of fused images. The results obtained from experiments show that the algorithm has performed well using contourlet-based PCNN and wavelet-based PCNN in terms of visual quality and performance metrics.

Yunxiang Tian and Xiaolin Tian [7] have proposed to extend the work of [6]. This includes the intensity hue saturation (IHS) space component from multispectral (MS) images, then decomposition into an I component using NSCT. This method reduces computation time and complexity. Yang Chen and Zheng Qin [8] have presented PCNN-based image fusion in a compression domain approach. In this experiment, the researchers validated the performance measure for three different fusion applications. This produced results in both qualitative and quantitative analysis carried out by Vaibhav R. Pandit and R. J. Bhiwani [9] for applications in image fusion. It has been compared with various conventional image fusion techniques. Hongguang Li et al. [10] have proposed a novel method for image registration and image fusion that has been integrated with cameras in both visible and infrared light on unmanned aerial vehicles (UAVs). This system is very efficient and uses a particle swarm optimization (PSO) hybrid with PCNN and NSCT. Thus, it can be applied to medium-or high-altitude UAVs. Yong Yang et al. [11] have presented two major issues related to multifocus image fusion using fast discrete curvelet transform (FDCT) texture selection and block effects. Experimental results show a reduction in computational complexity and improvement in performance metrics, which has been compared with various state-of-the-art frequency-based fusion and spatial-based fusion methods. B. Rajalingam and R. Priya [12] have implemented, using PCA-DWT, a technique to improve medical image fusion. It is superior to other existing techniques in terms of image quality, visual perception and computational time. Biswas et al. [13] proposed an idea for remote sensing image fusion based on shearlet transform (ST) and PCNN. It can be used to vary the decomposition level of source information in order to maintain geometrical and spectral information. Jun Li et al. [14] discussed image fusion methods applied to infrared and visible images using robust PCA (RPCA) and CS. It preserves background information from infrared and visible images. It computes in parallel in less computation time.

6.2Methodology used

The integration of various data that could provide valuable information is the major objective of image fusion. The major objective of image fusion is to integrate different data in order to obtain more information than can be derived from each of the single data. The integration of images from different scenarios will lead to additional information that should not be present in a separate imaging scenario. But owing to design or observational constraints, imaging devices are not capable of providing such information. For this reason several image fusion methods have been developed. Individual images are integrated to obtain a single image, so that all information can be obtained. The objective of the method is to design a system that satisfies the following criteria:

  1. Extract all useful information from the source information.
  2. Avoid artificial facts or inconsistencies that perturb human observers.
  3. Be reliable and robust to imperfections such as misregistration.
  4. Suppress noise and irrelevant paths.
  5. Reduce inconsistencies or distortions.
  6. Limit bandwidth and storage capacity.

6.2.1Principal component analysis theory

The aim of PCA is to convert a set of observations of correlated variables into a set of values that are linearly uncorrelated variables in a systematic process called principal components. Consider X represents an m-dimensional random vector, where V represents an orthonormal projection matrix such that Y = VTX. The covariance of Y is cov(Y), which is a diagonal matrix, and V−1 = VT:

cov(Y)=E(YYT)=E[(VTX)(VTXT)]=E[(VTX)(XTV)]=VTE(XXT)Vcov(Y)=VTcov(X)V.(6.1)

Multiplying both sides of Equation (6.1) by V we get

Vcov(Y)=VVTcov(X)V=cov(X)V.(6.2)

V can be written V = [V1, V2, . . .Vd] and

[λ10000λd](6.3)

Substituting Equation (6.2) into Equation (6.3)

[λ1V1,λ2V2λdVd]=[cov(X)V1,cov(X)V2cov(X)Vd],(6.4)

which can be written

λiVi=cov(X)Vi,(6.5)

where i = 1, 2. . .d, and Vi is an eigenvector of cov(X).

6.2.1.1PCA image fusion algorithm

Step 1:First we will generate a two-column matrix A of dimension 2 × N using the two source images and calculate the empirical mean vector along each column Me of dimension 1 × 2.
Step 2:Subtract M e from each column of A; the resulting matrix X will be of dimension 2 × N.
Step 3:Find the covariance matrix C of X and calculate the eigenvectors V and eigenvalue D of C and arrange them in decreasing order of their eigenvalues.

Step 4: Calculate the normalized principal components P1 and P2 as follows:

P1=V1ΣVandP2=V2ΣV.(6.6)

Step 5: The fused image is

Ifuse(x,y)=P1I1(x,y)+P2I2(x,y).(6.7)
Fig. 6.1: Information flow diagram in image fusion scheme employing PCA [12]

6.2.2Spatial or simple averaging

The average value of the input image is taken pixel by pixel using a simple image fusion method. The averaging-based fusion rule is as follows:

IF(i,j)=IA(i,j)2+IB(i,j)2,(6.8)

where IF is the fused image and IA, IB are source components of the image. Each image has three basic multispectral bands of R, G and B.

The relative averaging of two groups, C X and C Y, whichproduced C F, has a single band for all individual pixels, represented in Equation (6.9):

CF(m,n,i)=CX(m,n,i)2+CY(m,n,i)2,(6.9)

where (m, n) the denotes pixel and spatial domain and i denotes the R, G, B of an image.

Fig. 6.2: Block diagram of spatial averaging method [12]

6.2.3Image fusion–based wavelet transform

6.2.3.1Discrete wavelet transform

Signals can be represented with higher efficiency using wavelet transforms, which is one of the popular existing techniques. A signal is transformed into a set of basis functions, referred to as a wavelet. The wavelet is chosen and the level of decomposition is chosen at the initial stage. A threshold value is determined after the decomposition of the input data. After applying DWT, the size of the resultant output data vector is the same as that of the applied input data vector. Filters are used to reconstruct the original data. The filtered data are compared to DWT, the continuously transformed data are returned that were larger than the input data vector. The wavelet transform is a well-known signal processing technique that can be used to represent signals with high efficiency. The wavelet transform changes signals into a set of basis functions. These basis functions are called wavelets. A waveform of finite duration having an average value of zero is called a wavelet. A multiscale representation of a signal is given by a DWT. Most applications are not directly associated with scaling factors or wavelet coefficients. A filter band coefficient filters input signals using high-pass and low-pass filters. Let the low pass be h(n) and high pass be g(n). Many sets of filters satisfy the perfect reconstruction conditions, which are approximately symmetric. To use a two-dimensional transform image for analysis, synthesis filter bands can be implemented. In the 2D case, a 1D analysis filter band is initially connected to a segment of the image and then connected to the rows. Let the image representation have N1 rows and N 2 columns; then, after applying the 1D analysis filter band to each column, we have two subband images, each with N1/2 rows and N 2 columns; after applying the 1D analysis filter band to each row of both subband images, we have four subband images, each having N1/2 rows and N 2/2 columns. This is illustrated in Figure 6.3, which shows the output of the DWT of the image and its analysis block.

Fig. 6.3: DWT and its output image

6.2.3.2Dual-tree complex wavelet transform

Selesnick et al. [15] have shown how DTCWT is a more advanced form of wavelet transform compared to DWT. There are two major advantages of DTCWT over DWT, shift invariant and selective directional into higher dimensions. The major shortcomings in DWT are as follows.

  1. Wavelet coefficients tend to oscillate between positive and negative around singularities because of the wavelet bandpass function.
  2. A small shift in a signal perturbs the wavelet coefficient oscillation pattern around singularities.
  3. The lack of directional selectivity greatly complicates the modelling and processing of geometric image features.

One of the advantages of the DTCWT is that it can be used to implement 2D wavelet transforms, which are more selective with respect to the orientations than in separable 2D DWTs. There are two types of 2D DTCWT: the real 2D DTCWT tree is twice as expensive, while the complex 2D dual-tree DWT is four times as expensive. Both types of complex wavelet transforms have wavelets that are oriented in six distinct directions. Six wavelets are spread in distinct directions. Each subband of the 2D dual-tree transform indicates a specific orientation. A complex 2D dual tree is implemented as four critically sampled and separable 2D DWTs operating in parallel. However, different filter sets are used in the rows and columns. As in real cases, the sum and difference of subband images are determined to obtain the oriented wavelets. A real 2D dual-tree transform of an image is implemented using two critically sampled and separable DWTs operating in parallel. The sum and difference for each subband are taken into consideration. As shown in Figure 6.4, a complex 2D DTCWT is implemented as four critically sampled and separable 2D DWTs operating in parallel. The pair of conjugate filters applied to image decomposition can be expressed as

(Hx+jGx)(Hy+jGy)=(HxHyGxGy)+j(HxGy+GxHy).(6.10)

The complex conjugate of the wavelet transform has positive and negative coefficients of subbands. All corner-to-corner subbands are partitioned in both vertical and horizontal scale of six subbands in the block.

Fig. 6.4: Analysis block for DTCWT
Fig. 6.5: Block diagram of wavelet transform–based image fusion method

Figure 6.5 shows a general block diagram of a wavelet transform–based image fusion method. It consists of low-pass and high-pass subbands for the fusion process in terms of averaging-based and average mean square–based fusion techniques. The fused images must be obtained by taking the inverse wavelet transform from the previous step.

6.2.4Pulse-coupled neural network

This works on the principle of optic nerves, which transmit data observed by the eyes to the brain and the visual cortex. Generally, the input data are then transferred to fiber pathways and finally processed. The most predominant factor of the visual cortical area in the visual cortex is visual processing, and nearly all visual signals reach the cortex via the primary visual cortex. Eckhorn et al. [16] in the late 1980s discovered that binary images are created in an oscillating manner by the mid-brain, which can extract different features from the visual impression when they have to read the cat visual cortex [16] An actual image is then created in the cat’s brain by the help of these binary images. Inspired by this biological system, machines are also trained to perform the functions of the human brain. This leads to the new structure of a neural network model, called Eckhorn’s model, to simulate the behaviour of the visual cortex. From the case study report on the linking field model H. Ranganath et al. [17] discussed PCNN based on the Eckhorn model with some modifications.

Figure 6.6 shows a PCNN model that consists of two major components: feeding and linking. The link between neighbourhood neurons is established by the symmetric weights M and W. The input stimulus S is given to the feeding compartment. The PCNN model and its components are determined by 6.11

Fij[n]=eαFFij[n1]+Sij+VF(MY),(6.11)
Lij[n]=eαLLij[n1]+[VL](WY),(6.12)
Fig. 6.6: Schematic representation of PCNN processing element [18]

where ∗ represents the convolution operator, ‘n’ represents the number of iterations, ‘Y’ represents the output of neurons from (n − 1) iterations. The compartments have a memory of their previous state, which decays over time by the exponent term. The constants VF and VL are normalizing constants. After receiving the inputs the feeding and linking compartments are combined together to generate the internal state of the neurons. The internal activity U is expressed as in Equation (6.13), where β is the controlling parameter:

Uij[n]=Fij[n](1+βLij[n]).(6.13)

Thus, the internal activity of neuron U has various states mapped with the dynamic threshold, θ, to produce output to another neuron ‘Y’ as given in Equations (6.14) and (6.15):

Yij[n]=1;if Uij[n]>θij[n],(6.14)
Yij[n]=0;Otherwise.(6.15)

If (U > θ), then the neuron fires very high to the threshold value of Y and converges until a neuron gets fired. The dynamic activity of the threshold property is given in Equation (6.16),

θij[n]=eαθθij[n1]+VθYij[n],(6.16)

where Vθ is a constant with a significantly larger value than the average value U and order of magnitude. Initially, values of F, L, U and Y are all set to zero. The values of the θ elements are initially 0 or some larger value depending upon the application. From this activity, in both the feeding and linking compartments, the neuron receives some source stimulus S and stimulus from the neighbourhood state. The pulsing nature of the PCNN can be observed by looking at the rise and fall of the threshold value with the corresponding rise and fall of the internal activity value.

6.2.4.1General image fusion algorithm based on PCNN

A data flow diagram of the PCNN algorithm is shown in Figure 6.7 and follow the steps given below,

Fig. 6.7: Data flow diagram of PCNN based image fusion
Step 1:Segment input data into subbands of low pass and high pass and are called source data.
Step 2:The focus measure for the low-pass subband is given as an input to the PCNN to activate the neural activity using spatial focusing.
Step 3:Calculate the firing times Tij using the following equations:
Fij[n]=Sij,(6.17)
Lij[n]=eαlLij[n1]+VL(WXY),(6.18)
Uij[n]=Fij[n](1+βLij[n]),(6.19)
θij[n]=eijα[n1]+VYij[n],(6.20)
Yij[n]=1;ifUij[n]>θij[n],(6.21)
Yij[n]=0;Otherwise,(6.22)
Tij=Tij[n1]+Yij[n].(6.23)

We can take the parameters as

αL=0.06931,αθ=0.2,β=0.2,VL=1,V=20,W=[0.70710.7071010.70700.707]

and the iteration is n = 200.

Step 4: Obtain the decision map D F,ij and select the coefficients xF,ij which means that the coefficients with large firing times (large T values) are selected as coefficients of the fused image, since the value of each pixel in the firing map images is equal to the firing times, and a longer firing time corresponds to more information, based on the following equations:

DF,ij[n]=1;ifT1,ij[n]T2,ij[n],(6.24)
DF,ij[n]=0;ifT1,ij[n]<T2,ij[n],(6.25)
XF,ij=x1,ij[n];Dij[n]=1,(6.26)
..................Content has been hidden....................

You can't read the all page of ebook, please click here login for view all page.
Reset