11,017
Views
188
CrossRef citations to date
0
Altmetric
Research Articles

Multi-sensor image fusion for pansharpening in remote sensing

, , &
Pages 25-45 | Received 30 Aug 2009, Accepted 28 Oct 2009, Published online: 17 Feb 2010

Abstract

The main objective of this article is quality assessment of pansharpening fusion methods. Pansharpening is a fusion technique to combine a panchromatic image of high spatial resolution with multispectral image data of lower spatial resolution to obtain a high-resolution multispectral image. During this process, the significant spectral characteristics of the multispectral data should be preserved. For images acquired at the same time by the same sensor, most algorithms for pansharpening provide very good results, i.e. they retain the high spatial resolution of the panchromatic image and the spectral information from the multispectral image (single-sensor, single-date fusion). For multi-date, multi-sensor fusion, however, these techniques can still create spatially enhanced data sets, but usually at the expense of the spectral consistency. In this study, eight different methods are compared for image fusion to show their ability to fuse multitemporal and multi-sensor image data. A series of eight multitemporal multispectral remote sensing images is fused with a panchromatic Ikonos image and a TerraSAR-X radar image as a panchromatic substitute. The fused images are visually and quantitatively analysed for spectral characteristics preservation and spatial improvement. It can not only be proven that the Ehlers fusion is superior to all other tested algorithms, it is also the only method that guarantees excellent colour preservation for all dates and sensors used in this study.

1. Introduction

Most of the operating earth observation satellites such as Landsat, Spot, Ikonos, Quickbird, Formosat, GeoEye or Orbview provide panchromatic images at a higher spatial resolution than in their multispectral mode. The difference in spatial resolution between the panchromatic and the multispectral mode can be measured by the resolution ratio (RR), i.e. the ratio of their respective ground sampling distances (GSD) that may vary between 1 : 2 and 1 : 5 (Ehlers et al. Citation2009). A large number of fusion or pansharpening techniques have been developed to combine panchromatic and multispectral images to produce an enhanced multispectral image of high spatial resolution. Most of these methods seem to work well with images that were acquired at the same time by one sensor (single-sensor, single-date fusion) (Price Citation1987, Pohl and van Genderen Citation1998, Alparone et al. Citation2007, Ehlers Citation2008).

However, a number of sensors acquire information about the earth in either only panchromatic (EROS-A, EROS-B, Worldview-1) or only multispectral (RapidEye) mode. It becomes therefore increasingly important to fuse image data from different sensors which are usually recorded at different dates. Thus, there is a need to investigate techniques that allow multi-sensor, multi-date image fusion.

With multi-sensor fusion, an additional problem has to be considered. If data from different satellites are used for fusion, the RR can get worse. For example, the RR between RapidEye (6.5 m GSD) and Worldview (0.5 m GSD) is 1 : 13, for Quickbird (0.61 m GSD) and SPOT-5 multispectral (10 m GSD) about 1 : 16. With the advent of high-resolution radar data from the TerraSAR-X or Radarsat 2 sensor, studies are underway to investigate the potential of an active sensor for pansharpening (Klonus Citation2008).

Generally, image fusion methods can be differentiated into three levels: pixel level (iconic), feature level (symbolic) and knowledge or decision level. Of highest relevance for remote sensing are techniques for iconic image fusion, as the data are least altered and most of the pansharpening techniques have been developed for this fusion (Wald et al. Citation1997, Pohl and van Genderen Citation1998, Zhang Citation2004a). Iconic image fusion techniques can be grouped into three classes: colour-related techniques, statistical methods and numerical methods (Pohl and van Genderen Citation1998). The first class includes colour compositions of three image bands in the RGB colour space as well as the more sophisticated colour transformations such as the intensity-hue-saturation (IHS) or hue-saturation-value (HSV) transform. The second class includes band statistics, such as the principal component (PC) or Gram–Schmidt (GS) transforms. The third class includes arithmetic operations such as image multiplication, summation and image rationing as well as sophisticated numerical approaches such as wavelets (Chipman et al. Citation1995), curvelets (Candès Citation1999) or ridgelets (Chen et al. Citation2005).

Wald (Citation2002) proposed another categorisation of image fusion techniques: projection and substitution methods, relative spectral contribution and the ‘spatial improvement by injection of structures’ (améloration de la resolution spatial par injection de structures = ARSIS) concept. In the first class are those methods, which project the image into another coordinate system and substitute one component. The second class is comparable with the second class of Pohl and van Genderen (Citation1998), with the exception that this category is restricted to band rationing and arithmetic combinations. The basis of the ARSIS concept is a multi-scale technique to inject the high spatial information into the multispectral images. Although this classification scheme bears some merits, we selected the one introduced by Pohl and van Genderen (Citation1998) because of its mathematical precision.

Research on image fusion techniques in remote sensing started as early as the late 1980s and early 1990s (Cliche et al. Citation1985, Welch and Ehlers Citation1987, Chavez et al. Citation1991, Ehlers Citation1991) and concentrated on pixel level fusion (pansharpening). With the advent of the very high-resolution satellite programmes, the interest in pansharpening has significantly increased. Many publications focused on how to retain the spectral characteristics of the multispectral data after pansharpening (see, e.g. Alparone et al. Citation2007 or Thomas et al. Citation2008).

Currently, still uncomplex image fusion techniques such as IHS or Brovey are widely used for pansharpening. Often these methods do not retain the spectral characteristics of the multispectral datasets. More sophisticated techniques seem to work well for many applications when used for single-sensor, single-date fusion. When applied to multi-sensor data, they can still create spatially enhanced datasets but usually at the expense of the spectral consistency (Ehlers Citation2004, Klonus and Ehlers Citation2007, Klonus Citation2008).

Over the last few years, a number of improved algorithms have been developed with the promise to minimise colour distortion while retaining the spatial improvement of the standard data fusion algorithms. Especially, wavelet techniques led to a number of new fusion methods (Otazu et al. Citation2005, Lillo-Saavedra and Gonzalo Citation2006, Yunhao et al. Citation2006). We selected six advanced fusion methods for comparison. These algorithms are colour normalisation spectral sharpening (CN), GS fusion, modified IHS fusion, Ehlers fusion, University of New Brunswick (UNB) fusion and a wavelet-based fusion, the proportional additive wavelet fusion (AWLP). For comparison, we also selected the ‘traditional’ Brovey and PC fusion methods.

2. Study area and datasets

The study area is located in the North of Spain representing the region around Santo Domingo de la Calzada (). This area was used as a control site of the Joint Research Centre of the European Commission (JRC) in the project ‘Control with Remote Sensing of Area-Based Subsidies’ (CwRS) (http://agrifish.jrc.it/marspac/DCM/). A series of eight multitemporal, multispectral remote sensing images (seven SPOT scenes and one Formosat scene) was used (). These multispectral images cover a time frame of almost two years and virtually all seasons and thus pose an excellent challenge for a spectral characteristics preserving data fusion. All images were fused with a panchromatic Ikonos image from 30 May 2005 (1 m GSD) and in a second investigation with a TerraSAR-X spotlight image with 1 m spatial resolution as a panchromatic substitute. The TerraSAR-X radar image was recorded on 3 May 2008 and was provided by the German Aerospace Centre DLR. The image was despeckled with a proprietary algorithm developed at the University of Würzburg in Germany.

Figure 1. Study site near Santo Domingo de la Calzada, Spain.

Figure 1. Study site near Santo Domingo de la Calzada, Spain.

Table 1. Multispectral remote sensing datasets for the study site.

This agricultural area was chosen, because most of the colour changes appear in agricultural fields depending on crop phenology and harvesting dates. These spectral changes pose a serious problem for most fusion methods and consequently prevent a correct visual or automated agricultural assessment.

In urban areas, the colour changes are less severe. Here, the dominant challenge is the spatial improvement due to man-made structures (e.g. houses, roads). It is therefore mandatory to investigate not only the spectral fidelity of the multi-date, multi-sensor fusion algorithms but also the actual spatial improvement. Consequently, our test sites also included houses and roads as artificial structures.

3. Methodology

3.1 Preprocessing

The creation of coregistered datasets is the first and an extremely important part of the fusion process. Especially for radar data with its geometric properties such as foreshortening, layover, and shadowing, it is necessary to ensure that the images are in perfect registration. Otherwise, the quality of the fused image will decrease significantly. Even small misalignments can cause colour artefacts in the fusion process. Fortunately, the new German radar satellite TerraSAR-X provides image data in a perfect quality and a spatial resolution of 1 m in its spotlight mode (Schubert et al. Citation2008). This provided an excellent basis for a correct rectification. The other images were orthocorrected and accuracy checked by different European Union (EU) data contractors as part of the standard CwRS routine (JRC Citation2008).

After the orthocorrection of the original multispectral images, they were resampled using a cubic convolution function to the ground resolution of the high-resolution images (1 m). All bands of the original images were used for the fusion process.

3.2 Image fusion algorithms

We used eight different algorithms for fusion of the images. All eight images were fused with the panchromatic Ikonos image resulting in a dataset of 64 fused images. In a second investigation, four of these images (Formosat, SPOT 5 2005 and SPOT 4 2004a, b) were fused with the TerraSAR-X image. All images were visually and statistically evaluated for colour preservation and spatial improvement. Use was made of standard techniques such as Brovey and PC fusion and a number of recently developed algorithms: CN spectral sharpening fusion, GS fusion, modified IHS fusion, Ehlers fusion as well as two methods that participated in the IGARSS image fusion contest, the UNB fusion and the AWLP. The AWLP method ranked first in the fusion contest (Alparone et al. Citation2007). We are aware that other methods have been proposed but most are either in an experimental stage or have not been fully published. We concentrated on those that were readily available in remote sensing image analysis packages and/or were described in detail so that they could be implemented. The algorithms are described in the following sections.

3.2.1 Brovey transform

The Brovey transform is a combination of arithmetic operations. Each spectral band is first divided by the sum of the three chosen bands and then multiplied with the panchromatic image (Hallada and Cox Citation1983). For our study, this method was extended to four bands because the multispectral datasets were all four-band images.

3.2.2 Principal component transform

PC transform is a statistical technique that transforms a multivariate dataset of correlated variables into a dataset of uncorrelated linear combinations of the original variables. For images, it creates an uncorrelated feature space that can be used for further analysis instead of the original multispectral feature space. The PC is applied to the multispectral bands. The panchromatic image is histogram matched to the first PC. It then replaces the selected component and an inverse PC transform takes the fused dataset back into the original multispectral feature space (Chavez et al. Citation1991). The advantage of the PC fusion is that the number of bands is not restricted (such as the original IHS or Brovey fusion). It is, however, a statistical procedure which means that it is sensitive to the area to be sharpened and produces fusion results that may vary depending on the selected image subsets (Ehlers and Klonus Citation2004).

3.2.3 Modified IHS transform

The IHS transform separates spatial (intensity) and spectral (hue and saturation) information from a standard RGB image. The intensity refers to the total brightness of the image, hue to the dominant or average wavelength of the light contributing to the colour and saturation to the purity of colour. This is comparable with the human perception. To fuse the images, three bands of a multispectral image are transformed from the RGB domain into the IHS colour space. The panchromatic component is matched to the intensity of the IHS image and then replaces the intensity component. We make use of the modified IHS fusion from Siddiqui (Citation2003) which was developed for a better fit of the fused multispectral bands to the original data. After the matching, the panchromatic image replaces the intensity in the original IHS image and the fused image is transformed back into the RGB colour space (Ehlers Citation1991). For more than three bands, the IHS transform has to be used more than once depending on the number of bands to be fused.

3.2.4 Additive wavelet proportional fusion

The AWL method (Nunez et al. Citation1999) is one of the existing multi-resolution wavelet-based image fusion techniques. It was originally designed for a three-band RGB multispectral image. In this method, the spectral signature is preserved because the high-resolution panchromatic structure is integrated into the luminance L-band of the original low-resolution multispectral image. Therefore, this method is only defined for three bands. It was extended to an arbitrary number of bands by Otazu et al. (Citation2005). It maintains the spectral signature of an n-band image in the same way as AWL does with RGB images. This generalised method is called AWLP. The algorithm was implemented in Matlab for this study by the authors.

3.2.5 Colour normalisation spectral sharpening

CN spectral sharpening is an extension of the Brovey transform. The input image bands are grouped into spectral segments defined by the spectral range of the panchromatic image. The corresponding band segments are processed in the following manner: each input band is multiplied by the sharpening band and then normalised by dividing it by the sum of the input bands in the segment. Bands outside the spectral range of the panchromatic image are not sharpened (Vrabel et al. Citation2002).

3.2.6 Gram Schmidt fusion

The GS process transforms a set of vectors into a new set of orthogonal and linear independent vectors. By averaging the multispectral bands, the GS fusion simulates a low-resolution panchromatic band. As the next step, a GS transform is performed for the simulated panchromatic band and the multispectral bands with the simulated panchromatic band applied as the first band. Then the high spatial resolution panchromatic band replaces the first GS component. Finally, an inverse GS transform is applied to create the pansharpened multispectral bands (Laben et al. Citation2000).

3.2.7 Ehlers fusion

The Ehlers fusion was developed specifically for a spectral characteristics preserving image merging (Ehlers Citation2004). It is based on an IHS transform coupled with a Fourier domain filtering. The principal idea behind a spectral characteristics preserving image fusion is that the high-resolution image has to sharpen the multispectral image without adding new grey level information to its spectral components. An ideal fusion algorithm would enhance high-frequency changes such as edges and grey level discontinuities in an image without altering the multispectral components in homogeneous regions. To facilitate these demands, two prerequisites have to be addressed. First, colour and spatial information have to be separated. Second, the spatial information content has to be manipulated in a way that allows an adaptive enhancement of the images. This is achieved by a combination of colour and Fourier transforms.

For optimal colour separation, use is made of an IHS transform. This technique is extended to include more than three bands by using multiple IHS transforms until the number of bands is exhausted. If the assumption of spectral characteristics preservation holds true, there is no dependency on the selection or order of bands for the IHS transform. Subsequent Fourier transforms of the intensity component and the panchromatic image allow an adaptive filter design in the frequency domain. Using fast Fourier transform (FFT) techniques, the spatial components to be enhanced or suppressed can be directly accessed. The intensity spectrum is filtered with a low-pass filter (LP) whereas the spectrum of the high-resolution image is filtered with an inverse high-pass filter (HP). After filtering, the images are transformed back into the spatial domain with an inverse FFT and added together to form a fused intensity component with the low-frequency information from the low-resolution multispectral image and the high-frequency information from the high-resolution panchromatic image. This new intensity component and the original hue and saturation components of the multispectral image form a new IHS image. As the last step, an inverse IHS transformation produces a fused RGB image that contains the spatial resolution of the panchromatic image and the spectral characteristics of the multispectral image. These steps can be repeated with successive three band selections until all bands are fused with the panchromatic image. The order of bands and the inclusion of spectral bands for more than one IHS transform are not critical because of the colour preservation of the procedure (Klonus and Ehlers Citation2007).

3.2.8 University of New Brunswick fusion

The UNB fusion algorithm (Zhang Citation2004b) applies first histogram standardisation for the multispectral and panchromatic bands of the input images. The multispectral bands in the spectral range of the panchromatic image are selected and a regression analysis is calculated using a least square algorithm. The results are used as weight factors for the multispectral bands. Via multiplication with the corresponding bands and a following addition, a new synthesised image is produced. To create the fused image, each standardised multispectral image is multiplied with the standardised panchromatic image and divided by the synthesised image. This method was primarily designed for single-sensor, single-date images and is used as the standard method for Quickbird pansharpening. For this study, the implementation in PCI Geomatica was used.

3.3 Evaluation methods

3.3.1 Visual analysis

For visual analysis, colour preservation and spatial improvement were examined. For colour preservation, each band was evaluated for grey level changes and different band combinations were checked for spectral consistency. Although a number of different band combinations were tested, in this article we concentrate on the standard false colour infrared display because this band combination is widely used for many remote-sensing applications. For spatial improvement analysis, we checked for each band, if more details could be recognised.

These visual analysis techniques, however, have the disadvantage that they are subjective and depend heavily on the experience of the respective interpreter. Therefore, we also applied a number of quantitative-statistical tests.

3.3.2 Spectral evaluation

A number of statistical evaluation methods are used to measure after-fusion colour fidelity. These methods should be objective, reproducible, and of quantitative nature. Therefore, we selected the following statistical measures.

The correlation coefficient (CC) between the original multispectral bands and the equivalent fused bands is the most frequently used method to evaluate the spectral value preservation (see, e.g. González-Audícana et al. Citation2005 or Klonus and Ehlers Citation2007). The values range from −1 to 1. The best correspondence between fused and original image data shows the highest correlation value and should be close to 1.

For a per-pixel deviation (PD; Wald Citation2002), the fused image is first degraded to the spatial resolution of the original image and is then subtracted from the original image on a per-pixel basis. In a final step, the average deviation per pixel is calculated using the number of pixel in the scene. Low differences between the original and the fused image are indicated by low values of the PD. The lowest (and best) value is 0. A perfect 0, however, means that original and fused image are identical and no pansharpening is performed. A good correspondence is therefore indicated by values close to 0.

As an additional measurement, the root mean square error (RMSE) was also proposed by Wald (Citation2002). It is computed by the difference of the standard deviation and the mean of the fused and the original image. The best possible value is again 0.

The structure similarity (SSIM) index was proposed by Wang et al. (Citation2004). The SSIM is a further development of the Q index, which is widely used to evaluate the spectral value preservation of pansharped images (Wang and Bovik Citation2002, Alparone et al. Citation2004). This method combines a comparison of luminance, contrast and structure and is applied locally in an 8 × 8 window. This window is moved pixel-by-pixel over the entire image. At each pixel, the local statistics and the SSIM index are calculated within the window. The values vary between 0 and 1. Values close to 1 show the highest correspondence with the original images.

3.3.3 Spatial improvement evaluation

The objective is to find the fused image with the optimal combination of spectral characteristics preservation and spatial improvement. If the image shows no spatial improvement, no pansharpening is performed which makes the fusion obsolete. The ‘fused image’ would score best for spectral evaluation with no spatial improvement, however. Most evaluations have concentrated on spectral consistency with little or no emphasis on spatial improvement. To avoid this shortcoming, we selected two different quantitative methods to measure the quality of spatial improvement.

For the calculation of the high-pass correlation (HCC) coefficient, a HP filter with a 3 × 3 Laplacian kernel is first applied to the panchromatic image and to each band of the fused image. Then the correlation coefficients between the HP filtered bands and the HP filtered panchromatic image are calculated. This analysis was proposed by Zhou et al. (Citation1998). According to Pradhan et al. (Citation2006), there is no reason for the preference of a specific filter, any HP filter can be used for this purpose.

For the second measure, an edge detection (ED) approach was used. A Canny edge detector (Canny Citation1986) is applied to the panchromatic image and each band of the fused multispectral image. The detected edges are then compared to the panchromatic image edges for each individual band. ED correspondence is measured in per cent; 100% means that all the edges in the panchromatic image are detected in the fused image. Only fusion techniques which produce spectral fidelity and spatial improvement should be used, especially for multi-date and multi-sensor fusion.

4. Fusion results and quality evaluation

For the investigation of the electro-optical (E–O) fusion (i.e. pansharpening with the Ikonos image) a multi-date, multi-sensor dataset with 64 fused images was produced from eight original images with eight different fusion methods. For the investigation using TerraSAR-X radar data as panchromatic substitute (radar fusion), 32 fused images were produced from four original images and again the eight different fusion methods. All images were visually and statistically analysed for all fusion methods. Our comprehensive evaluation is based on an individual analysis of all 96 fusion results.

To illustrate our findings in the following section, we use a subset of the SPOT 5 2004 image, because it is very representative for the encountered fusion effects. shows the original multispectral SPOT 5 2004 (band 3, near infrared; band 2, red; and band 1, green) as a standard false colour infrared display and the panchromatic Ikonos image. For the investigation of the radar fusion, a subset of the SPOT 5 2005 image is used to illustrate the differences between the fusion methods. It was not possible to use the same subsets for both fusion tests, because Ikonos and TerraSAR-X had no overlapping ground coverage ().

Figure 2. Original multispectral SPOT 5 2004 image (a) recorded on 24 July 2004 (© CNES 2004) in the band combination 3 (near infrared), 2 (red), 1 (green) resampled to 1m and the original panchromatic Ikonos image and (b) recorded on 30 May 2005 (satellite image by GeoEye). Note the inverse brightness in the images for the fields on the left side.

Figure 2. Original multispectral SPOT 5 2004 image (a) recorded on 24 July 2004 (© CNES 2004) in the band combination 3 (near infrared), 2 (red), 1 (green) resampled to 1m and the original panchromatic Ikonos image and (b) recorded on 30 May 2005 (satellite image by GeoEye). Note the inverse brightness in the images for the fields on the left side.

Figure 3. Original multispectral SPOT 5 2005 (© CNES 2005) image (a) in the band combination 3 (near infrared), 2 (red), 1 (green) resampled to 1m and the original panchromatic TerraSAR-X image and (b) recorded on 3 May 2008(©DLR 2008). Note the differences in the field structure for the agricultural areas in the upper right part.

Figure 3. Original multispectral SPOT 5 2005 (© CNES 2005) image (a) in the band combination 3 (near infrared), 2 (red), 1 (green) resampled to 1m and the original panchromatic TerraSAR-X image and (b) recorded on 3 May 2008(©DLR 2008). Note the differences in the field structure for the agricultural areas in the upper right part.

The results for the panchromatic fusion are shown in Figures (even numbers); the ones for the radar fusion are presented in Figures (odd numbers). All multispectral images are again shown in the band combination near infrared, red, green.

Figure 4. Fused multispectral SPOT 5 2004 image using Brovey (a) and PC (b) in the band combination 3 (near infrared), 2 (red), and 1 (green).

Figure 4. Fused multispectral SPOT 5 2004 image using Brovey (a) and PC (b) in the band combination 3 (near infrared), 2 (red), and 1 (green).

Figure 5. Fused multispectral SPOT 5 2005 image using Brovey (a) and PC (b) in the band combination 3 (near infrared), 2 (red), and 1 (green).

Figure 5. Fused multispectral SPOT 5 2005 image using Brovey (a) and PC (b) in the band combination 3 (near infrared), 2 (red), and 1 (green).

Figure 6. Fused multispectral SPOT 5 2004 image using IHS (a) and AWLP (b) in the band combination 3 (near infrared), 2 (red), and 1 (green). Note the artefacts in homogeneous areas for the ALWP image.

Figure 6. Fused multispectral SPOT 5 2004 image using IHS (a) and AWLP (b) in the band combination 3 (near infrared), 2 (red), and 1 (green). Note the artefacts in homogeneous areas for the ALWP image.

Figure 7. Fused multispectral SPOT 5 2005 image using IHS (a) and AWLP (b) in the band combination 3 (near infrared), 2 (red), and 1 (green).

Figure 7. Fused multispectral SPOT 5 2005 image using IHS (a) and AWLP (b) in the band combination 3 (near infrared), 2 (red), and 1 (green).

Figure 8. Fused multispectral SPOT 5 2004 image using CN (a) and GS (b) in the band combination 3 (near infrared), 2 (red), and 1 (green).

Figure 8. Fused multispectral SPOT 5 2004 image using CN (a) and GS (b) in the band combination 3 (near infrared), 2 (red), and 1 (green).

Figure 9. Fused multispectral SPOT 5 2005 image using CN (a) and GS (b) in the band combination 3 (near infrared), 2 (red), and 1 (green).

Figure 9. Fused multispectral SPOT 5 2005 image using CN (a) and GS (b) in the band combination 3 (near infrared), 2 (red), and 1 (green).

Figure 10. Fused multispectral SPOT 5 2004 image using Ehlers (a) and UNB (b) in the band combination 3 (near infrared), 2 (red), and 1 (green).

Figure 10. Fused multispectral SPOT 5 2004 image using Ehlers (a) and UNB (b) in the band combination 3 (near infrared), 2 (red), and 1 (green).

4.1 Visual analysis

In the fused Brovey images, the improved structure is clearly visible. For the E–O fusion () colour distortions are visible, especially for the fields in the west. The reason for this is the grey value distribution of the panchromatic Ikonos image. The brightness structure of the panchromatic image is directly injected into the multispectral image and causes the alteration in colour. The best result for this fusion technique was achieved for the SPOT 5 2005 image which was recorded in April, i.e. the closest to the acquisition date of the panchromatic Ikonos image. In the fused radar image (), the colour changes are striking; in some parts of the image, colours appear that seem ‘unnatural’ in a false colour infrared display.

The fused PC images show some spatial improvement but not as detailed as for the Brovey fusion. In both images ( and ), dramatic changes in the spectral values appear. The reason for these colour changes is again directly related to the panchromatic/radar information. For the E–O fusion, the fused image falsely suggests that the bare field in the southwest is actually vegetated which renders it inappropriate for any photointerpretation or automated classification process. For the radar fusion, the near infrared band with high grey values in the original image is represented by low grey values in the fused image. These low values come from the radar image.

In general, the spatial structures in the IHS fused images are improved, however, not as much as for the Brovey fusion. The IHS fusion shows sometimes acceptable results but never without serious colour changes (). The best result for the E–O fusion image was achieved for the SPOT 5 2005 image. This image was recorded in April, the shortest time difference to the acquisition date of the panchromatic Ikonos image. In the fused radar images (), the IHS methods show a better colour preservation than most of the other methods. However, there are still significant colour changes.

The worst results for the spatial enhancement are associated with the AWLP fusion ( and ). The spatial structures are only slightly improved. Especially along the edges, artefacts appear which disrupt and blur the information in the whole image. Particularly in homogeneous areas, no spatial improvement could be found at all. This is true for all images of the E–O fusion process. The spectral values are relatively well preserved making it the second best method for colour fidelity.

For the E–O fusion, the CN spectral sharpening shows good to very good spatial improvement in all images, but often associated with serious spectral distortions. The results were often similar to those of the Brovey fusion (). This comes as no surprise as the CN fusion is an extension of the Brovey method. For the radar fusion, the TerraSAR-X header had to be manipulated so that the wavelength of the multitemporal images could be adapted to the radar image. With this, it was possible to use the TerraSAR-X image as a substitute for a high-resolution panchromatic image. This incompatibility was probably the main reason for the enormous change of the spectral values in the fused radar images (). This is especially true for regions where the TerraSAR-X image exhibits extremely high digital numbers (DNs), such as in edges of buildings. These colour artefacts distort also the spatial resolution.

The GS fusion shows some improvement in spatial resolution but never as much as Brovey, PC, CN or IHS. In some cases, the fused images look still blurred (). The correspondence of the spectral values to the original is low. For the E–O fusion, it displays similar colour deficiencies as the PC fusion. This holds also true for the radar fusion (). In the near infrared band, high grey values in the original images are replaced with low values in the fused images. This leads to spectral changes, especially in this band.

It is clearly visible that only the Ehlers fusion retains the spectral characteristics of the original image ( and ). This is true for all images, as there are almost no visible colour changes in comparison to the original. The spatial improvement is evident and provides good to very good results, particularly for the E–O fusion. For the radar fusion, the general colour preservation is satisfactory. Only for the edges of buildings, which act as corner reflectors, high DNs are produced which do not exist in the multispectral data. As the Ehlers fusion injects these edges into the multispectral images, this effect cannot be avoided. On the other hand, the buildings become more distinct, a fact that can support a subsequent photointerpretation.

Figure 11. Fused multispectral SPOT 5 2005 image using Ehlers (a) and UNB (b) in the band combination 3 (near infrared), 2 (red), and 1 (green).

Figure 11. Fused multispectral SPOT 5 2005 image using Ehlers (a) and UNB (b) in the band combination 3 (near infrared), 2 (red), and 1 (green).

The worst results are produced by the UNB fusion (b and ). The spectral information is replaced by a more or less homogeneous structure for all bands which makes the fused image look like a black-and-white display. In addition, the edges are deformed and many artefacts are visible in the image. Therefore, the spatial structure is also not much improved. It has to be noted that this dramatic effect appears only in two cases for the E–O fusion, the SPOT 5 2005 and the SPOT 2 2004. The reason for this is probably that this method uses reference images for the regression, which are in the spectral range of the panchromatic image. However, all the other fused images (including the radar fusion) also show severe colour distortions (). This function was developed for single-sensor, single-date fusion. The reason, why it does not work for multi-sensor fusion, however, is not clear. It may have something to do with its implementation in the used software package. Consequently, it should not be used for multitemporal and/or multi-sensor fusion.

4.2 Quantitative spectral analysis

Although a visual analysis has to be a significant part of all quality evaluation procedures, it has the disadvantage that it is very subjective and depends on the interpreter (Klonus and Ehlers Citation2009). Statistical analysis techniques have to accompany the visual interpretation for objective and reproducible quantitative results. For this purpose, we make use of the quantitative-statistical methods presented in Section 3.3. Comprehensive results of this evaluation are presented in Tables . Best values in each column (scene) in the tables are marked with bold and the worst in grey. In general, the quantitative analyses for colour preservation did confirm the visual inspection findings. The Ehlers fusion scored the highest marks for colour preservation in nearly 90% of the images.

Table 2. CCs for the E–O fusion.

Table 3. CCs for the radar fusion.

Table 4. Deviation per pixel for the E–O fusion.

Table 5. Deviation per pixel for the radar fusion.

Table 6. RMSE for the for the E–O fusion.

Table 7. RMSE for the radar fusion.

Table 8. SSIM for the E–O fusion.

Table 9. SSIM for the radar fusion.

Table 10. HCC coefficients for the E–O fusion.

Table 11. HCC coefficients for the radar fusion.

Table 12. ED results for the E–O fusion.

Table 13. ED results for the radar fusion.

4.2.1 Correlation coefficient

The calculation of the CC has the benefit that a (constant) bias between the fused and the original image does not alter the results. Best results were obtained by the Ehlers fusion for 10 images and by the AWLP method for two images. Brovey, IHS, PC, and GS fusion have their best results for the SPOT 5 2005 image, which is the closest in time to the acquisition date of the panchromatic Ikonos image. In general, however, CN, IHS, Brovey, PC, and GS exhibit low CCs (with the exception of SPOT 5 2005). The lowest CCs were calculated for the UNB fusion. Highest value was only 0.554 for the SPOT 5 2004 scene. Only the Ehlers fusion exceeded 0.9 for all images. presents the correlation results for the E–O fusion. The coefficient values are calculated as average value for all bands.

CCs are lower for most methods for the radar fusion (). One exception is the IHS function where the values are higher on average than for the E–O case. Again, the Ehlers fusion has the highest correlation with coefficients above 0.9 for all images. Second is the AWLP fusion. All other methods produced low correlation values between 0.05 and 0.6. CCs are the lowest for the UNB and PC fusion, thus confirming the visual analysis.

4.2.2 Per-pixel deviation

The calculated PD values are presented in and . The lowest (and best) values are associated with the Ehlers fusion again confirming the results of the visual analysis. Results for AWLP and the modified IHS are in most cases acceptable. For the SPOT 4 2004b image, however, also these two methods fail to produce satisfactory results. The GS fusion produces acceptable results for only three images with deviation values below 5.0. PC, CN, UNB and Brovey fusion produce insufficient results, especially for the radar fusion with a PD above 5.0 for all methods and exceeding 10.0 for most of the images (). The worst results are associated the Brovey fusion with PD values above 50 for all images with the exception of SPOT 2 2004. It should be mentioned that the result of the Brovey method was adapted to the 8-bit range, but no histogram matching was applied. Nevertheless, the visual inspection and the low CCs prove that the spectral quality of the Brovey method is not sufficient for image fusion.

4.2.3 Root mean square error

The best values for the RMSE () are again associated with the Ehlers fusion. Not in correspondence with the visual interpretation are the results for the GS fusion. For the E–O test case, the GS fusion shows the second best results whereas in the visual analysis it was ranked very low. However, all RSME values are still a factor of 4–20 higher than those of the Ehlers fusion. A close third is the AWLP fusion. Unacceptable results were produced by the PC, CN and Brovey methods. Especially, Brovey shows again the highest (worst) values, with the exception of the winter scene (SPOT 2 2004). The UNB fusion shows very different results ranging from 0.01 to 32.6, where 0.01 is the best value for one band of the SPOT 5 2004 scene and 32.6 the worst value for one band of the SPOT 4 2003 scene.

For the radar fusion (), all methods with the exception of the Ehlers fusion produce very poor results for the scene SPOT 4 2004a. For the three other scenes, the two simplest methods, Brovey and PC, show the worst results. AWLP, IHS, and GS results are somewhat better but still a factor of 2–10 worse than the Ehlers fusion. Especially, the values for the GS fusion could not be confirmed by the visual analysis. CN and UNB did not produce acceptable results.

4.2.4 Structure similarity index

Again, best results for the E–O and the radar fusion were achieved by the Ehlers algorithm followed by the AWLP ( and ). The Ehlers fusion was the only one with SSIM values of 0.9 and better for all images.

For the radar case (), no other fusion algorithm scored above 0.9. For the E–O fusion, PC, IHS, CN and GS show some acceptable results with values between 0.4 and 0.9. For the radar fusion, the values are generally lower (with the exception of Ehlers). UNB and Brovey produce unsatisfactory result for both test cases.

4.3 Quantitative spatial analysis

In most fusion papers, the quality assessment is only concerned with spectral quality. The idea of pansharpening, however, is to inject the spatial structure of the high-resolution panchromatic image into the multispectral bands. If only the spectral fidelity is considered, the unaltered multispectral image would reach the best results which would make the fusion obsolete. Consequently, we assessed also the spatial quality of the fusion algorithms. Here, the results are very different from the spectral analysis. From the visual analysis, we could already conclude that the spatial enhancement in the fused images proved to be sufficient for most of the methods.

4.3.1 High-pass correlation

The CCs after HP filtering are to some degree influenced by the different fusion types. The reason for this should be investigated, because visual inspection showed sometimes different results. For the E–O fusion, most of the methods could improve the structure in the image (). Brovey and Ehlers were very consistent and produced average values above 0.9. IHS and CN were lower but also consistent, whereas PC, CN, and UNB were very inconsistent. High values for SPOT 2 2004 (PC and GS) or SPOT 4 2004a (UNB) are contradicted by low values for other dates such as SPOT 5 2005 (PC and GS) or SPOT 4 2005 (UNB). These results make their use for image fusion unreliable. The lowest values are produced by the AWLP with CC around zero meaning that there is no correspondence between the fused bands and the high-resolution image. We suspect that the observed artefacts are the reason for the insufficient results of the HP correlation analysis ().

In general, the radar fusion showed lower CCs than the E–O fusion which can be explained by the special characteristics of radar imagery (). Best results were obtained by the Brovey transform with the only value above 0.9 (SPOT 4 2004b). CCs were above 0.6 for all Brovey fused images. The only other function that was consistently above 0.5 was the Ehlers fusion. All other methods were inconsistent (PC, IHS and GS) or even insufficient (AWLP, CN and UNB).

4.3.2 Edge detection

For the ED evaluation, results were very similar for most of the fusion functions ( and ). With the exception of the AWLP method, all of them show very good results with an edge correspondence above 90%. The only fusion function with results below 90% for all images was the AWLP. This confirms the visual interpretation results that the ALWP does not really improve the spatial resolution. Best results were obtained by UNB and CN, which failed completely in the spectral quality assessment.

For radar fusion, the results differ somewhat from the E–O case. Here, Brovey is the best overall with values between 94 and 97%. The other functions above 90% are PC, CN, and GS followed by UNB, Ehlers and IHS. The latter functions had values between 87 and 95%, which means that they still produce sufficient results. AWLP is the only fusion technique which has values below 90% for all images with a minimum of 78%, making it the only technique that does not produce adequate spatial improvement.

5. Conclusion

In this study, we could prove that standard and most of the advanced fusion methods cannot cope with the demands that are placed on them by multi-sensor/multi-date fusion. The spectral distortions are manifold: brightness reversions, a complete change of spectral characteristics, artificial artefacts or unnatural and artificial colours are the results of many operational fusion techniques that are readily available in many commercial image processing systems. Fusion methods such as PC, CN, GS or UNB should only be used for single-sensor, single-date images. Wavelet-based fusions can retain most of the spectral characteristics which comes unfortunately at the expense of spatial improvement. The AWLP wavelet method produced additional spatial artefacts instead of spatial improvements. This is probably caused by the wavelet characteristics.

For the radar fusion, we can conclude that simple replacement of the panchromatic image by a high-resolution radar image has to be treated with caution. It could be shown that pansharpening techniques could be used to improve the spatial resolution of multispectral images with TerraSAR-X data as high-resolution input. It is also evident that only the Ehlers fusion was capable of preserving the spectral characteristics of the multispectral image making it suitable for further image analysis (photointerpretation or classification). The only technique of those that were applied in our comparative study that delivers pansharpened images with almost no spectral change is the Ehlers fusion.

As discussed in Section 3.3, a combined method for a quantitative assessment of spatial improvement and spectral preservation is needed because, otherwise, the best colour preservation is observed if no pansharpening is performed. This would, however, defeat the purpose of the fusion process. A combined assessment of spatial improvement and spectral characteristics preservation can be used as a comprehensive quality measure for image fusion. So far, we have used a narrative combination of results for different quantitative analyses. A mathematical equation that integrates spectral- and spatial-quality evaluation would most likely be a better and more consistent quality indicator. The development of such an integrative quality indicator should therefore be high on the list for future fusion research.

Acknowledgements

Thanks to the JRC for providing the data for this research. The TerraSAR-X data and project is supported by the German BMWI through the DLR with the contract number 50EE0704.

References

  • Alparone , L . 2004 . A global quality measurement of pan-sharpened multispectral imagery . IEEE Geoscience and Remote Sensing Letters , 1 ( 4 ) : 313 – 317 .
  • Alparone , L . 2007 . Comparison of pansharpening algorithms: outcome of the 2006 GRS-S data-fusion contest . IEEE Transactions on Geoscience and Remote Sensing , 45 ( 10 ) : 3012 – 3021 .
  • Candès , EJ . 1999 . Harmonic analysis of neural networks . Applied and Computational Harmonic Analysis , 6 : 197 – 218 .
  • Canny , J . 1986 . A computational approach to edge detection . IEEE Transactions on Pattern Analysis and Machine Intelligence , 8 ( 6 ) : 679 – 714 .
  • Chavez , WJ , Sides , SC and Anderson , JA . 1991 . Comparison of three different methods to merge multiresolution and multispectral data: TM and SPOT pan . Photogrammetric Engineering and Remote Sensing , 57 ( 3 ) : 295 – 303 .
  • Chen , T , Zhang , J and Zhang , Y . 2005 . Remote sensing image fusion based on ridgelet transform . Proceedings of the International Geoscience and Remote Sensing Symposium , 2 : 1150 – 1153 .
  • Chipman , LJ , Orr , TM and Lewis , LN . 1995 . Wavelets and image fusion . IEEE Transactions on Image Processing , 3 : 248 – 251 .
  • Cliche , G , Bonn , F and Teillet , P . 1985 . Integration of the SPOT pan channel into its multispectral mode for image sharpness enhancement . Photogrammetric Engineering and Remote Sensing , 51 : 311 – 316 .
  • Ehlers , M . 1991 . Multisensor image fusion techniques in remote sensing . ISPRS Journal of Photogrammetry and Remote Sensing , 46 ( 1 ) : 19 – 30 .
  • Ehlers , M . 2004 . Spectral characteristics preserving image fusion based on Fourier domain filtering . Proceedings of SPIE, Maspalomas, Spain , 5574 : 1 – 13 .
  • Ehlers , M . 2008 . “ Multi-image fusion in remote sensing: spatial enhancement vs. spectral characteristics preservation ” . In Advances in visual computing, Part II , Edited by: Bebis , G . 75 – 84 . Berlin : Springer Verlag .
  • Ehlers , M , Jacobsen , K and Schiewe , J . 2009 . “ High resolution image data and GIS ” . In ASPRS manual of GIS , Edited by: Madden , M . 721 – 777 . Bethesda, MD : American Society for Photogrammetry and Remote Sensing .
  • Ehlers , M and Klonus , S . 2004 . Erhalt der spektralen Charakteristika bei der Bildfusion durch FFT basierte Filterung . Photogrammetrie-Fernerkundung-Geoinformation , 6 : 495 – 506 .
  • González-Audícana , M . 2005 . Comparison between Mallat's and the ‘à trous’ discrete wavelet transform based algorithms for the fusion of multispectral and panchromatic images . International Journal of Remote Sensing , 26 ( 3 ) : 595 – 614 .
  • Hallada , WA and Cox , S . 1983 . Image sharpening for mixed spatial and spectral resolution satellite systems . Proceedings of the 17th international symposium on remote sensing of environment . 1983 . pp. 1023 – 1032 . MI : Environmental Research Institute Michigan Ann Arbor .
  • JRC, 2008. Common Technical Specifications for the 2009 Campaign of Remote-Sensing Control of Area-based Subsidies, ITT No. 2008/S 228-3024. Available from: http://mars.jrc.it/bulletins-publications/common-technical-specifications-for-the-2009-cwrs-campaign
  • Klonus , S . 2008 . Comparison of pansharpening algorithms for combining radar and multispectral data . The International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences , XXXVII ( B6b ) : 189 – 194 .
  • Klonus , S and Ehlers , M . 2007 . Image fusion using the Ehlers spectral characteristics preserving algorithm . GIScience and Remote Sensing , 44 ( 2 ) : 93 – 116 .
  • Klonus , S and Ehlers , M . Performance of evaluation methods in image fusion . Proceedings of the 12th International Conference on Information Fusion . Seattle. 6–9 July . Washington : International Society of Information fusion .
  • Laben, C.A., Bernard, V., and Brower, W., 2000. Process for enhancing the spatial resolution of multispectral imagery using pan-sharpening. United States Patent Application No. 6,011,875.
  • Lillo-Saavedra , M and Gonzalo , C . 2006 . Spectral or spatial quality for fused satellite imagery? A trade-off resolution using the wavelet à trous algorithm . International Journal of Remote Sensing , 27 ( 7 ) : 1453 – 1464 .
  • Nunez , E . 1999 . Multiresolution-based image fusion with adaptive wavelet decomposition . IEEE Transactions on Geoscience and Remote Sensing , 37 ( 3 ) : 1204 – 1211 .
  • Otazu , X . 2005 . Introduction of sensor spectral response into image fusion methods. Application to wavelet-based methods . IEEE Transactions on Geoscience and Remote Sensing , 43 ( 10 ) : 2376 – 2385 .
  • Pohl , C and van Genderen , JL . 1998 . Multisensor image fusion in remote sensing: concepts, methods and applications . International Journal of Remote Sensing , 19 ( 5 ) : 823 – 854 .
  • Pradhan , PS . 2006 . Estimation of the number of decomposition levels for a wavelet-based multiresolution multisensor image fusion . IEEE Transactions on Geoscience and Remote Sensing , 44 ( 12 ) : 3674 – 3686 .
  • Price , JC . 1987 . Combining panchromatic and multispectral imagery from dual resolution satellite instruments . Remote Sensing of Environment , 21 : 119 – 128 .
  • Schubert , A . 2008 . Geometric validation of TerraSAR-X high-resolution products . Proceedings of the 3rd TerraSAR-X Science Team Meeting . 2008 . DLR, Oberhaffenhofen. Available from: http://sss.terrasar-x.dlr.de/papers_sci_meet_3/paper/cal0163_schubert.pdf
  • Siddiqui , Y . 2003 . The modified IHS method for fusing satellite imagery . ASPRS 2003 Annual Conference Proceedings . 2003 . American Society for Photogrammetry and Remote Sensing Anchorage, Alaska (CD publication) .
  • Thomas , C . 2008 . Synthesis of multispectral images to high spatial resolution: a critical review of fusion methods based on remote sensing physics . IEEE Transactions on Geoscience and Remote Sensing , 46 ( 5 ) : 1301 – 1312 .
  • Vrabel , J . 2002 . Demonstration of the accuracy of improved resolution hyperspectral imagery . Proceedings of SPIE , 4725 : 556 – 567 .
  • Wald , L . 2002 . Data fusion. Definitions and architectures–fusion of images of different spatial resolutions , Paris : Presses del'Ecole, École de Mines de Paris .
  • Wald , L , Ranchin , T and Magolini , M . 1997 . Fusion of satellite images of different spatial resolutions: assessing the quality of resulting images . Photogrammetric Engineering and Remote Sensing , 63 ( 6 ) : 691 – 699 .
  • Wang , Z and Bovik , AC . 2002 . A universal image quality index . IEEE Signal Processing Letters , 9 : 81 – 84 .
  • Wang , Z . 2004 . Image quality assessment: from error visibility to structural similarity . IEEE Transactions on Image Processing , 13 ( 1 ) : 600 – 612 .
  • Welch , R and Ehlers , M . 1987 . Merging multiresolution SPOT HRV and Landsat TM data . Photogrammetric Engineering and Remote Sensing , 53 : 301 – 303 .
  • Yunhao , C . 2006 . A new wavelet-based image fusion method for remotely sensed data . International Journal of Remote Sensing , 27 ( 7 ) : 1465 – 1476 .
  • Zhang , Y . 2004a . Understanding image fusion . Photogrammetric Engineering and Remote Sensing , 70 ( 6 ) : 657 – 661 .
  • Zhang , Y . 2004b . System and method for image fusion . United States Patent Application No. 20040141659
  • Zhou , J , Civco , DL and Silander , JA . 1998 . A wavelet transform method to merge Landsat TM and SPOT panchromatic data . International Journal of Remote Sensing , 19 ( 4 ) : 743 – 757 .

Reprints and Corporate Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

To request a reprint or corporate permissions for this article, please click on the relevant link below:

Academic Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

Obtain permissions instantly via Rightslink by clicking on the button below:

If you are unable to obtain permissions via Rightslink, please complete and submit this Permissions form. For more information, please visit our Permissions help page.