Open Access

SAR image target detection in complex environments based on improved visual attention algorithm

EURASIP Journal on Wireless Communications and Networking20142014:54

https://doi.org/10.1186/1687-1499-2014-54

Received: 2 January 2014

Accepted: 26 March 2014

Published: 5 April 2014

Abstract

A novel target detection algorithm for synthetic aperture radar (SAR) images based on an improved visual attention method is proposed in this paper. With the development of SAR technology, target detection algorithms are confronted with many difficulties such as a complicated environment and scarcity of target information. Visual attention of the human visual system can make humans easily focus on key points in a complex picture, and the visual attention algorithm has been used in many fields. However, existing algorithms based on visual attention models cannot obtain satisfactory results for SAR image target detection under complex environmental conditions. After analysing the existing visual attention models, we combine the pyramid model of visual attention with singular value decomposition to simulate the human retina, which can make the visual attention model more suitable to the characteristics of SAR images. We introduce variance weighted information entropy into the model to optimize the detection results. The results obtained by the existing visual attention algorithm for target detection in SAR images yield a large number of false alarms and misses. However, the proposed algorithm can improve both the efficiency and accuracy of target detection in a complicated environment and under weak-target conditions. The experimental results validate the performance of our method.

Keywords

Visual attentionSingular value decompositionVariance weighted information entropy (WIE)Target detectionSAR image

1 Introduction

The visual attention model for synthetic aperture radar (SAR) image target detection plays a positive role because the human visual system focuses on the areas of interest and rapidly decides on them [1]. Visual attention greatly improves the ability of the human visual system to deal with images under a complex environment. Therefore, a positive effect can be gained by introducing visual attention into target detection in SAR images.

With the continuous development in military technology, target detection in a SAR image becomes more difficult. The environment around targets becomes more complicated, and information on targets lessens. Traditional detection methods for SAR images such as constant false alarm rate (CFAR) have been improved to adapt different conditions, but they are also difficult to perform outstandingly. The reason is that under a complicated environment and weak-target condition, key pixels are lacking and interference pixels are too much. Limited improvements cannot compensate for defects. Existing visual attention models generally use the Gaussian pyramid model proposed by Burt and Crouely to simulate the feature of the human retina [2]. The feature can be described by considering that the centre of the retina has a smaller receptive field; on the other hand, the receptive field at the periphery of the retina is much bigger [3]. Therefore, we can conclude that as the feature sampling density and visual resolution of a position become smaller with the increase in distance from the centre of the human retina, the peripheral information is compressed [4]. After building a Gaussian pyramid model, the visual attention system guides the attention to an area of interest according to some features of a target, such as the shape, colour and intensity [5]. Current visual attention methods are not suitable for target detection in a complicated environment or under weak-target conditions [6].

The Gaussian pyramid model of visual attention suffers from the difficulty of effectively compressing a SAR image with weak targets. But compressing an image to different rates and keeping important information is of importance for the visual attention algorithm. Therefore, this paper proposes a new visual attention algorithm for target detection. Since the singular value decomposition (SVD) method can keep the important information of a SAR image when the image is compressed [7], combining it with the Gaussian pyramid model can produce images with different compression ratios, which enable the image to retain the target information and well obscure the environment information. After combining SVD with the Gaussian pyramid model, the variance weighted information entropy (WIE) method is used to distinguish the different types of areas and filter out the regions of interest (ROIs) without targets. As a result, the purpose of efficient target detection under a complicated environment and weak targets can be achieved.

The remainder of this paper is structured as follows: Section 2 introduces in detail the classic visual attention model and the pyramid model. Section 3 presents the steps of each detection stage and focuses on the SVD method and the variance WEI. Section 4 presents the experimental results of the SAR target extraction using the proposed techniques, and Section 5 draws the conclusion.

2 Classic visual attention model

The algorithm proposed in this paper is an improvement of Itti. Itti is a classic visual attention model. The Itti model determines the ROI in human eyes, which includes the target to be detected as a set of significant pixels in images, and then extracts the ROI by finding a significant pixel in the image [810]. The model can adaptively detect the ROIs in the image. Compared with most of the traditional algorithms that need to manually specify the ROI, the Itti model enjoys great advantages in target detection and recognition in image processing. Here, we first introduce the calculation process of the Itti model, and the details can be found in the literature [11]. Then, we describe the details on building the pyramid model.

For an image, the first part of the visual attention model is the linear filter, which extracts the colour, intensity and orientation features from the image. For example, the Itti model makes the output of the Gabor filter in four different directions become the orientation feature. The formula for the Gabor filter is expressed as follows:
h x , y , θ i , λ , α , β = 1 2 παβ exp - π x θ i α 2 + y θ i β 2 exp 2 πi x θ i λ
(1)
where α and β are the standard deviation, λ is the wavelength and θ is the direction. The formulas of x θ i and y θ i are shown as follows:
x θ i = x cos θ i + y sin θ i
(2)
y θ i = - x sin θ i + y cos θ i
(3)

Then, we build a Gaussian pyramid for every feature. In the second part, a centre-surround difference module is used to extract the feature map. In the third part, a plurality of different feature maps is merged to form the conspicuity map, which is a saliency map, through an effective feature consolidation strategy. In the fourth part, the focus of the attention area is located based on the saliency map. In the final part of the visual attention model, the winner-take-all competition net is used to find the most significant point from the saliency map, and the inhibition-of-return is used to ensure that the area would not be focused again [1214].

The visual attention model adopts the multi-scale spatial attention to simulate the nonuniform sampling mechanism of the human retina [15]. Burt and Crouely proposed a pyramid structure from the summary of the nonuniform sampling mechanism [16]. Figure 1 shows the pyramid structure of an image.
Figure 1

Pyramid structure of an image.

Two steps are needed to establish the pyramid structure: one is smoothing and the other is downsampling. For a two-dimensional digital image I(i,j), δ represents the pyramid layers. When δ = 0, the formula I δ (i, j) = I0(i, j) = I indicates the bottom of the pyramid structure. Hence, the calculation of the pyramid layers is shown as follows:
I σ i , j = k = - N k = N t = - N t = N ψ k , t I σ - 1 2 i + k , 2 j + t
(4)

The Itti model adopts the linear discrete Gaussian filter to perform the smoothing and downsampling in the horizontal and vertical directions of the input image, respectively, and forms eight different resolution sub-images [17]. Including the original image, nine images are required to build up the Gaussian pyramid structure. The smoothing filter is [1 4 6 4 1], and downsampling is achieved by a convolution with a [1 1]/2 filter. We use the two filters to achieve our objective, which take the average value of every two pixels in the previous image as one pixel value in the next image. The two steps can be combined into a convolution with a filter K = [1 4 6 4 1]*[1 1]/2 in the horizontal and vertical directions.

3 Improved visual attention model

In recent years, the SVD algorithm has been widely studied. This algorithm extracts the algebraic feature from an image [18]. SVD has the characteristics of energy aggregation for an image [19], which makes it a popular technology in the area of image compression. The features represent the essential characteristics of an image; therefore, the SVD algorithm has the advantage of being insensitive to noise and complexity of an image. The variance WEI is a statistical form of the characteristics, which reflects the average information of a figure. It was first used to detect infrared images [20].

3.1 SVD integrated into visual attention

The SVD technique has been paid close attention since its introduction and has been widely used in statistical analysis, signal processing and system theory. The SVD is an extension of the spectral analysis theory. It is essentially an orthogonal transformation. If a matrix has a linearly correlated row or column, it can be changed into a diagonal matrix by multiplying the orthogonal matrix on the left with that on the right [21]. After the change, we can obtain some singular values whose number reflects the independent row (column) vector number of the original matrix. The eigenvalues can construct each individual component of the original signal. The SVD technique has the advantage of fast processing and handing stability. The SVD in modern linear algebra and signal processing theory is defined as follows: if ARm × n exists, then orthogonal matrices URm × n and VRm × n exist. The formula is obtained as follows:
A = V T A = V H
(5)
= 0 0 r 0
(6)
r = diag σ 1 , σ 2 , , σ r
(7)
The diagonal elements are arranged in a descending order:
σ 1 σ 2 σ n 0 , r = rank A
(8)
U and V are m × m and n × n dimensional unitary matrices which respectively satisfy
U U H = I m
(9)
V V H = I n
(10)

To make the visual attention model adjust to the target detection in a complicated-environment SAR image, we combine SVD with the pyramid model. This combination can compress the original image to a number of images with different compression rates. The information about the targets in the resulting images with different compression rates is retained and that on the environment is obscured. This result allows the centre-surround module to find the saliency point more effectively and optimise the detection results under a complicated environment.

The major steps of the proposed SVD-pyramid model method are as follows:

  • For an original image A, it is decomposed into a diagonal matrix B and another two orthogonal matrices using SVD.

  • The number of nonzero elements R is computed, and the elements are arrayed in a descending order to form vector P.

  • The m number of biggest elements from vector P is retained. The remaining elements form vector Q. The value of m is equal to 60% of R.

  • A new diagonal matrix B 1 is constructed using the elements of vector Q.

  • A new image A 1 is generated using the diagonal matrix B 1 and the two orthogonal matrices.

  • The above steps are repeated until the number of nonzero elements is less than 1 or R × 0.02.

3.2 Post-processing based on the WIE algorithm

Entropy reflects the average information of a figure. The one-dimensional graph entropy can be expressed as
H = i = 0 255 pi log pi
(11)
The variance WIE was first used to detect infrared images [18]. Thus, targets in infrared images alway show different radiation and usually have a distinct grey value. The WIE algorithm is defined as follows:
H s = - s = 0 255 s - s ¯ 2 Ps log Ps
(12)
where Ps is the probability of the grey levels in the infrared image and s ¯ is the mean intensity of the infrared image. In particular, when Ps = 0, we let Ps log (Ps) = 0.Considering that the result of the improved visual attention method needs to filter out the false alarms, in this study, we import WIE to achieve this purpose. We test the value of WIE from different areas in a real SAR image. The results show that different types of regions correspond to a large difference in the WIE value, and similar types of regions correspond to a small difference in the WIE value. When the ROI areas are detected, we compute the values of WIE of these areas and array them in a descending order. Then, we calculate the differences between adjacent values and array them in an ascending order. Next, we abandon several largest values in sequence 2 and calculate the mean value in sequence 2 as a threshold value. The values in the threshold value region are corresponding to areas with a target. From this result, we can filter out the area without targets. Figure 2 shows the simulation results. The size of all samples is 45 × 45 pixels. The sample is divided into three types, which are the area that includes the up and down, the area that contains the target and the uniform area.
Figure 2

The WIE value with different types of areas.

The above results show that the WIE value of a real SAR image in different areas has a large difference. Based on the difference, we can determine the areas without targets. Therefore, the WIE method can be used to achieve the post-processing of the target detection.

3.3 Processing steps

The flow chart of the visual attention based on the SVD algorithm and WIE method mentioned earlier is shown in Figure 3.
Figure 3

Flow chart of improved visual attention.

Because the algorithm employed in this study is based on the Itti model, the detailed steps of the algorithm will not be presented. In the feature extraction module, we use the features that include the intensity, colour, orientation and consistency. In the centre-surround difference module, the no. 2, 3 and 4 images are selected as the centre image, and the numbers of the surround images are 2, 3 and 4.

4 Simulation and results

4.1 Simulation

To verify the feasibility of the proposed algorithm, several simulations are performed. The simulation data are divided into two types, i.e. type 1 and type 2. Type 1 consists of 20 images, and the size of each image is 384 × 256. These are images in a simple environment and conspicuous target conditions. These targets are tanks in a grassland environment. Type 2 consists of images in a complicated environment and weak targets, and the size of each image is 2,406 × 512. The targets are tanks in grassland and jungle environment.

The simulations are divided into three parts. The first part is the simulation using the Itti model. The second part is the simulation using the proposed algorithm. The last part is the simulation using the two-parameter CFAR. This detection method is a pixel-level target detection method. The reason we choose CFAR for comparison is that the algorithm is not only the most in-depth and most practical but also an extensive method of a class at present. The simulation of each part satisfies two premises, i.e. simple environment and conspicuous target condition and complicated environment and weak-target condition. Figure 4a shows the test image for the first type of data, which are under the simple environment and conspicuous target conditions. Figure 4b,c,d shows the results obtained by the three algorithms. In the figures, red lines and yellow lines show the shifting route and size of FOA, respectively.
Figure 4

The test image and results by different algorithms in a simple environment. (a) The test image of a simple environment. (b) The result of the Itti model algorithm in a simple environment. (c) The result of the two-parameter CFAR method in a simple environment. (d) The result of the improved VA algorithm in a simple environment.

To verify the performance of the improved algorithm under a complicated environment, we perform another simulation using the type 2 data. Figures 5a and 6a show the test images of the type 2 data. They are obtained at different levels of complexity condition. Figures 5b,c,d and 6b,c,d show the results obtained by the three algorithms.
Figure 5

The results of test image 1 by different algorithms in a complicated environment. (a) The test image of a complicated environment. (b) The result of the Itti model algorithm in a complicated environment. (c) The result of the two-parameter CFAR method in a complicated environment. (d) The result of the improved VA algorithm in a complicated environment.

Figure 6

The results of test image 2 by different algorithms in a complicated environment. (a) The test image of a complicated environment. (b) The result of the Itti model algorithm in a complicated environment. (c) The result of the two-parameter CFAR method in a complicated environment. (d) The result of the improved VA algorithm in a complicated environment.

4.2 Results

In the two data types for the simulation, we obtain the following results by comparing the three algorithms. The results are divided into two parts. First, for the data in the simple environment and weak target case, the results are listed in Table 1. The three methods yield four focus of attention (FOA) areas, and the four areas detect the targets well. All three methods do not yield false alarms and undetected targets. We compare the size of the FOA for two kinds of visual attention algorithm. The size of the FOA is measured by the number of pixels in the FOA. Through the simulation, the number of pixels in the FOA in the classic algorithm is 2,000 to 7,000, and it is 1,000 to 1,500 in the proposed algorithm. Therefore, the size of the FOA in the improved method is smaller.
Table 1

Whole performance comparisons among the three algorithms in a simple environment

 

Number of all targets

Number of detected targets

False alarms

Number of undetected targets

Size of FOA (number of pixels)

Classic algorithm

4

4

0

0

2,000 to 7,000 (large)

CFAR

4

4

0

0

-

The proposed algorithm

4

4

0

0

1,000 to 1,500 (small)

From the type 2 data obtained by the simulation, we obtain the results listed in Table 2. Six targets are present in the SAR image. From the classic algorithm, in the six FOA areas obtained by the Itti model, only three areas cover the targets. The others cover shadow and empty areas. The two-parameter CFAR method detected ten targets. Although the six targets are detected, four false alarms occur. On the other hand, all six target areas obtained by the proposed algorithm exactly cover the targets. In the classic algorithm, the number of pixels in the FOA is 2,000 to 9,000, and in the proposed algorithm, it is 1,000 to 1,500. Thus, the FOA size in the improved method is also smaller.
Table 2

Whole performance comparisons among the three algorithms in a complicated environment

 

Number of all targets

Number of detected targets

False alarms

Number of undetected targets

Size of FOA (number of pixels)

Classic algorithm

13

12

6

7

2,000 to 9,000 (large)

CFAR

13

27

14

0

-

The proposed algorithm

13

13

0

0

1,000 to 1,500 (small)

By comparing the detected rates, we determine that the performance of the classic algorithm is significantly lower than that of the other two algorithms and the proposed algorithm performs best. From the comparison of the false alarm rates, the false alarm rate and misses of the classic algorithm are high. Although the CFAR method successfully detects all the targets, the false alarm rate is higher than that of the proposed method. The detected rate is defined as the ratio of the number of correct detected target to real target. The false alarm rate is defined as the ratio of the number of false detected target to real target. By comparing the three methods, the classic method produces too many false alarms which affect the detection accuracy. Itti model methods miss a lot of targets, but the proposed method can make up for the lack. In summary, the proposed algorithm can obtain not only a high detection rate but also a low false alarm rate. The improved visual attention algorithm can adapt to the conditions of a complicated environment and weak target.

After comparing the performance of the three methods, we study the effect of image complexity on the detection performance of the three methods. We use the GLCM features to describe the spatial distribution and whole complexity of an image. The features include the angular second moment, entropy, inverse different moment and contrast. Figure 7 shows the result.
Figure 7

Detection results of the three methods in different complexities. (a) Detection rate. (b) False alarm rate.

Figure 7 shows that the detection rate of the classic method rapidly decreases. Because the complexity of the image increases, the background becomes complicated, and the classic method detects more complicated regions than real targets. The detection rate of the CFAR method decreases slowly. It is based on a large number of false alarms. With the increase in complexity, the detection efficiency rapidly decreases. The proposed method yields better results in both detection and false alarm rates.

5 Conclusions

In this study, we have developed an improved visual attention algorithm adapted to SAR image target detection under complicated environment and weak-target conditions. The method of combining SVD with the pyramid model and using WIE to filter out the false alarms has been introduced in detail. To validate the performance of the method, some simulations were performed. The results show the feasibility of the improved visual attention algorithm to SAR image target detection under complicated environment conditions.

Declarations

Acknowledgements

This work was supported in part by the National Natural Science Foundation of China under Projects 60802065, 61271287 and 61371048.

Authors’ Affiliations

(1)
School of Electronic Engineering, University of Electronic Science and Technology of China, Chengdu, China

References

  1. Walther DB, Koch C: Attention in hierarchical models of object recognition. Prog. Brain Res. 2007, 165: 57-78.View ArticleGoogle Scholar
  2. Itti L, Koch C: Computational modeling of visual attention. Nat. Neurosci. 2001, 2: 194-203. 10.1038/35058500View ArticleGoogle Scholar
  3. Walther D PhD thesis. In Interactions of visual attention and object recognition: computational modeling, algorithms, and psychophysics. Pasadena, CA: California Institute of Technology; 2006.Google Scholar
  4. Tsotsos JK, Liu Y, Martinez-Trujillo J, Pomplun M, Simine E, Zhou K: Attending to visual motion, computer vision and image understanding. Spec. Issue Attention Perform. Comput. Vision 2005, 100: 3-40.Google Scholar
  5. Navalpakkam V, Itti L: An integrated model of top-down and bottom-up attention for optimal object detection. In Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR). New York; June 2006:2049-2056.Google Scholar
  6. Feng J, Cao Z, Pi Y: Multiphase SAR image segmentation with G0 statistical model based active contours. IEEE Trans. GRS 2013, 51(7):4190-4199.Google Scholar
  7. Rybak IA, Gusakova VI, Golovan AV, Podladchikova LN, Shevtsova NA: A model of attention-guided visual perception and recognition. Vis. Res. 1998, 38: 2387-2400. 10.1016/S0042-6989(98)00020-0View ArticleGoogle Scholar
  8. Shah S, Levine MD: Visual information processing in primate cone pathways-part. I. A model, systems, man, and cybernetics. IEEE Trans. 1996, 2: 259-274.Google Scholar
  9. Li Z, Itti L: Saliency and gist features for target detection in satellite images. IEEE Trans. Image Process. 2011, 20(7):2017-2029.MathSciNetView ArticleGoogle Scholar
  10. Desimone R, Duncan J: Neural mechanism of selective visual attention. Annu. Rev. Neurosci. 1995, 18: 193-194. 10.1146/annurev.ne.18.030195.001205View ArticleGoogle Scholar
  11. Lee S, Kim K, Kim JY, Kim M, Yoo HJ: Familiarity based unified visual attention model for fast and robust object recognition. Pattern Recogn. 2010, 43(3):1116-1128. 10.1016/j.patcog.2009.07.014View ArticleMATHGoogle Scholar
  12. Treisman AM, Gelade G: A feature-integration theory of attention. Cogn. Psychol. 1980, 12: 97-136. 10.1016/0010-0285(80)90005-5View ArticleGoogle Scholar
  13. Cater K, Chalmers A, Ward G: Detail to attention: exploiting visual tasks for selective rendering. In Proc. 14th Eurographics Workshop Rendering. Aire-la-Ville: Eurographics Association; 2003:270-280.Google Scholar
  14. Itti L, Koch C, Niebur E: A model of saliency-based visual attention for rapid scene analysis, pattern analysis and machine intelligence. IEEE Trans. 1998, 20: 1254-1259.Google Scholar
  15. Rigas I, Economou G, Fotopoulos S: Low-level visual saliency with application on aerial imagery. Geosci. Remote Sens. Lett. IEEE 2013, 10: 1389-1393.View ArticleGoogle Scholar
  16. Borji A, Itti L: Exploiting local and global patch rarities for saliency detection. In Proc. IEEE Conference on Computer Vision and Pattern Recognition. Providence; 2012:478-485.Google Scholar
  17. Walther D, Rutishauser U, Koch C, Perona P: Selective visual attention enables learning and recognition of multiple objects in cluttered scenes. Comput Vision Image Understanding 2005, 100(1–2):41-63.View ArticleGoogle Scholar
  18. Andrews H, Patterson C: Singular value decomposition (SVD) image coding. IEEE Trans. Commun. 1976, 24(4):425-432. 10.1109/TCOM.1976.1093309View ArticleGoogle Scholar
  19. Arnold B: An investigation into using singular value decomposition as a method of image compression. Department of Mathematics and Statistics, University of Canterbury; 2000.Google Scholar
  20. Yang L, Zhou Y, Yang J: Variance WIE based infrared images processing. Electron. Lett. 2006, 42(15):857-859. 10.1049/el:20060827View ArticleGoogle Scholar
  21. Ruotolo R, Surace C: Using SVD to detect damage in structures with different operational conditions. J. Sound Vib. 1999, 226(3):425-439. 10.1006/jsvi.1999.2305View ArticleGoogle Scholar

Copyright

© Liu and Cao; licensee Springer. 2014

This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly credited.