Scholarly article on topic 'High Range Resolution Profile Automatic Target Recognition Using Sparse Representation'

High Range Resolution Profile Automatic Target Recognition Using Sparse Representation Academic research paper on "Electrical engineering, electronic engineering, information engineering"

CC BY-NC-ND
0
0
Share paper
Keywords
automatic target recognition / high range resolution profile / sparse representation / feature extraction / dictionary generation

Abstract of research paper on Electrical engineering, electronic engineering, information engineering, author of scientific article — Zhou Nuo, Chen Wei

Abstract Sparse representation is a new signal analysis method which is receiving increasing attention in recent years. In this article, a novel scheme solving high range resolution profile automatic target recognition for ground moving targets is proposed. The sparse representation theory is applied to analyzing the components of high range resolution profiles and sparse coefficients are used to describe their features. Numerous experiments with the target type number ranging from 2 to 6 have been implemented. Results show that the proposed scheme not only provides higher recognition preciseness in real time, but also achieves more robust performance as the target type number increases.

Academic research paper on topic "High Range Resolution Profile Automatic Target Recognition Using Sparse Representation"

Available online at www.sciencedirect.com

V ScienceDirect

Chinese Journal ofAeronautics 23(2010) 556-562

Chinese Journal of Aeronautics

www.elsevier.com/locate/cja

High Range Resolution Profile Automatic Target Recognition Using Sparse Representation

Zhou Nuo, Chen Wei*

School of Electronics and Information Engineering, Beijing University ofAeronautics and Astronautics, Beijing 100191, China

Received 29 November 2009; accepted 8 April 2010

Abstract

Sparse representation is a new signal analysis method which is receiving increasing attention in recent years. In this article, a novel scheme solving high range resolution profile automatic target recognition for ground moving targets is proposed. The sparse representation theory is applied to analyzing the components of high range resolution profiles and sparse coefficients are used to describe their features. Numerous experiments with the target type number ranging from 2 to 6 have been implemented. Results show that the proposed scheme not only provides higher recognition preciseness in real time, but also achieves more robust performance as the target type number increases.

Keywords: automatic target recognition; high range resolution profile; sparse representation; feature extraction; dictionary generation

1. Introduction

High range resolution profile (HRRP) is a one-dimensional measurement of target radar reflectivity along radar line-of-sight (LOS). It provides detailed structure information of ground moving targets for identification. Furthermore, compared with two-dimensional synthetic aperture radar (SAR) images, it also has the advantage of easy acquisition, fast processing and less storage space taking. Over the recent years, high range resolution (HRR) mode has been used as a complementary mode to ground moving target indicator (GMTI) and SAR imagery for ground moving target recognition[1-3].

Feature analysis and extraction are significant parts of HRR automatic target recognition (ATR). Their missions are to discover and capture the primary signatures of HRRP as well as to delete plenty of unnecessary information. The obtained features benefit the improvement of recognition preciseness and running speed. Several methods such as HRRP amplitudes[3-5], FFT spectrum analysis[6], S transformation[7], wavelet coefficient fitting[8-9] and nonlinear kernel transformation^0-1^ have been used extensively for HRR ATR

♦Corresponding author.Tel.: +86-10-82317203. E-mail address: chenwei@buaa.edu.cn

1000-9361/$ - see front matter © 2010 Elsevier Ltd. All rights reserved. doi: 10.1016/S1000-9361(09)60254-5

applications. The method based on HRRP amplitudes is more popular due to its clear physical interpretation, convenient acquisition and high recognition precise-ness.

Recently, as a new method of signal analysis, sparse representation has received a great deal of attention. In sparse representation theory, it is considered that signals are mixtures of a wide range of waveforms with different frequency properties, time-scale properties or time-frequency properties. Instead of conventional methods which decompose signals by a single type of waveforms, sparse representation analyzes signals over a redundant dictionary consisting of multiple types of waveforms. The waveforms in the dictionary are called atoms in sparse representation theory. Compared with the conventional methods, local structures within signals could be individually adapted by the atoms which take the most similar properties, and the signal decomposition is more concise. Consequently, sparse representation provides a more flexible and exact approach for signal analysis. To date, sparse representation has been successfully applied in signal analysis, images or videos compression and reconstruction, radar signal formation, and so on[12]. In addition, because of its excellent capability in signal analysis and signal representation, sparse representation has been attempted to be applied to experiments including face recognition[13-14], human movement recognition[15] and cancer diagnosis[16].

In this article, we explore the use of sparse representation into HRRP feature analysis and expect its performance in HRRP recognition.

2. Sparse Representation

2.1. Theory

Sparse representation is originally introduced by S. G. Mallat and Z. F. Zhang in 1993[17], and developed by D. L. Donoho[18-20] later. It is a method to analyze signals under a redundant dictionary.

y = AX = [A A2 ••• AN]

= X a;x

where the measured signal y=[y1 y2 ■■■ yd]T£Rd is a d-dimensional vector, the redundant dictionary A = [A[ A2 - A'n ] e RdxN is a matrix with size of dXN, d<N, the atom A- eA is a d-dimensional vector, i=1,2, — ,N, the coefficient vector X=[x1 x2 ■•• xN]TeRN is a N-dimensional vector obtained by y expanding over A. As d<N, equations y=AX are under determined. There are infinite X satisfying y=AX. The freedom of selection of X facilitates the adaption to various practical applications.

In all the possible solutions, X with fewest nonzero coefficients, X0, is the sparse representation[21]:

X0 = min | |x| |0 subject to y = AX (2)

In fact, X0 is greatly attractive in actual demands since it always means fewer sensors in a radar system or less prerequisites for a machine.

2.2. Dictionary

Dictionary design is crucial for signal component analysis in sparse representation. Generally, atoms in a dictionary should meet two basic requirements: containing a wide range of types of atoms to better adapt to different local structures of signals, and having sufficient number of atoms to make the dictionary redundant.

The available methods for dictionary establishment can be classified into two categories. One category is that the dictionary consists of a collection of g':

f=± g ( ^ )e - *2

where the collection of g' are transformations of a base function g(t)[17], u and ^are dilation, translation and frequency modulation parameters, respectively. g' produced by different 5, u and Stakes various frequency properties, time-scale properties or time-fre-

quency properties. Moreover, because the freedom of the selection of 5, u and £ is unlimited, the number of g' is infinite. That is, the created dictionary can be as redundant as it is needed. However, in practical application, it is sometimes hard to decide the suitable number of g' for a dictionary. Furthermore, as all atoms in the dictionary derive from the same base function, the adaptability in signal component representation is relatively limited.

Another category for dictionary generation is a training sample-based method[14]. The dictionary is combined with training samples from several target types. These training samples are collected under a wide scope of real scenes or their simulations where the measured signals, or also called test samples related to training samples, probably take place. Due to the similar collecting conditions, training samples have similar characteristics with their associated test samples. The dictionary generated by these training samples is able to represent the test samples more efficiently and precisely. Additionally, since the real scenes where signals take place are known or predictable, the atom number in the dictionary can be decided easily.

2.3. Approaches

Finding the sparse representation by the minimum l0 norm in Eq.(2) is complex and extremely hard[18]. Several algorithms have been designed to find suboptimal or approximate solutions instead. There are three approaches widely used in sparse representation over redundant dictionaries. They are matching pursuit (MP), basis pursuit (BP) and least absolute shrinkage and selection operation (LASSO) [17-23].

MP[17-21], proposed by S. G. Mallat, is a greedy pursuit algorithm. It finds the sparse solution by an iterative matching process. Compared with BP and LASSO, MP has the advantage of easy implementing and short running time. However, since the solutions produced by MP are just local optimal, they are sometimes not precise enough.

D. L. Donoho[19] simplified problems minimum l0 norm into problems minimum l1 norm:

X1 = min | |X| [ subject to y = AX (4)

Furthermore, he gave evidence that Xt in Eq.(4) is equal to X0 in Eq.(2) when the sparse representation of y exists. Recently, l1 minimization has been considered as an effective and general method for underdeter-mined equations. Based on the idea in Eq.(4), D. L. Donoho proposed BP[18-21] whose solutions are much sparser and more precise than the solutions MP produces. Nevertheless, computation of BP is still too complicated to be used in applications, especially in the situation that the dimension of y is large.

LASSO is also a method dealing with sparse problem based on Eq.(4). But it tackles l1 minimization

problem by a regression process subject to constraint such as ||X jj!< p.

- AXg subject to jjXj1< p (5)

In the regress process, sparse solution is obtained by optimizing the parameter p value and the coefficients at each chosen value of p. The available literature has demonstrated that solutions obtained by LASSO are sensible approximation to y, or even equal to X0, and the performance of LASSO is fast, efficient and stable no matter the scale of dimension of y is large or small[19-23]. For these merits, LASSO is selected to solve the sparse representation in this article.

3. HRR ATR

3.1. Data description

Due to a lack of real HRRP database obtained from moving ground vehicles, the raw HRRP used in our experiments are formed from a SAR image database collected by an X-band HH-polarization SAR sen-sor.[24] This SAR image database contains several classes of stationary ground vehicles which are no more than 3 m in height, 4 m in width and 8 m in length. Each SAR image in the database is about 128 pixelx128 pixel with 1 foot by 1 foot resolution. Each vehicle corresponds to two groups of SAR images. These two groups of images are obtained under comparable conditions except that one group of SAR images are collected at 15° depression angles over the 360° coverage in aspect while 17° depression angles for the other. At the same depression angle for each vehicle, the interval between two neighboring SAR images is less than 3° aspect angles, and the number of SAR images is more than 120. According to the instruction of the database[24], SAR images from 17° depression angles are used as training samples in a recognition experiment while SAR images from 15° depression angles are used as test samples. The detailed transformation process about the raw HRRP generation from SAR images is explicitly described in Ref. [25].

Due to the formed raw HRRP containing amount of background information which mainly expands in several first and last range bins and is unnecessary for an identification task, only the information in the central 60 range bins of raw HRRP is kept as the HRRP data used in the following experiment.

3.2. Dictionary generation

Considering the superiorities in decomposition accuracy and predictability of the dictionary size, dictionary generation based on training samples is attempted to be used in this experimental study. In the remainder of this section, its feasibility will be discussed from two aspects: the variability of atom types

and the dictionary redundancy.

As radar bandwidth increases over the years, radar resolution in range has been greatly improved. Radar is no longer a blob detector, but a sophisticated system with the abilities to acquire detailed information for target type recognition[26]. Point scattering models are universally adopted to describe target scattering properties which are produced by wideband radars[27].

Assume there are L scattering centers for a given target. Distance of each scattering center to the radar receiver along radar LOS is ri, i=1,2,- ,L. The corresponding intensity and initial phase are ai and w individually. L scattering centers are divided into N contiguous range bins according to ri.

Assume further that there are Ln scattering centers in nth range bin, n=1,2,■■■ ,N, L=L\+L2+-+LN. Their distances to the radar receiver, intensities and initial phases are rni, ani and \yni in sequence, i=1,2,-,Ln. HRRP in nth, hn, is the sum of echoes from Ln scattering centers:

hn = eXP["j(— rm - Vrn )]

where 2 is radar wavelength.

It is seen that there is a close relationship between point scattering models and HRRP. However, point scattering models are strongly dependent on target types and the aspect and depression angles where targets stay ahead. For targets with different types, since the sizes, shapes, materials and scattering properties are distinct, their scattering properties are highly different from each other[28]. Various point scattering models are required to represent them individually. For a given target, a relatively large rotation in aspect or depression, about 10°[29], also results in the changes of target scattering properties. Similarly, a series of point scattering models thus are also needed to describe the scattering properties in continuous angle ranges. In fact, although the point scattering model is constant for a given target, a small rotation in aspect or depression equally produces the changes of scattering center distribution in range bins. It is called migration through resolution cells (MTRC). Assume the range resolution is R, the maximum size of the given target in height and the azimuth plane is no more than Z. If the target rotation satisfies

., R 360

MTRC can be avoided[29-30]. Therefore, it is observed that all of the changes caused by target types, large rotation in aspect or depression and MTRC can cause the changes of L, r, cr„ ^ and Ln, rni, ani, and finally result in the variability of HRRP local structures.

According to the discussion above, it is known that if a dictionary is combined with training samples from multiple target types, the diversity of atom types in the dictionary can be satisfied easily. Furthermore, since the scale of experimental HRRP is 60 and the number

of training samples for each target is more than 120 as described in Section 3.1, the dictionary generated by the concatenation of training samples of various types of targets is sufficiently redundant. Meanwhile, the dictionary size can be correctly predicted by summing up the number of training samples.

In addition, according to Section 3.1, it can also be seen that the maximum size of targets in height and the azimuth plane is no more than 8 m, and thus the rotation avoiding MTRC should be no more than 2.18°. That is, if a rotation in aspect or depression is less than 2.18°, L, ri, cr„ w and Ln, rni, oni, ym■ are relatively stable. The components of HRRP can be viewed as constant. But if the rotation is larger than 2.18°, L, ri, ai, y/i and Ln, rni, ani, y/ni will vary. For the training or test samples of a given target collected at a fixed depression angle, since the angle increment in aspect between contiguous samples is between 1° and 3°, the local structures of samples vary following the change of aspect angles over the 360° aspect coverage. However, the training samples and the test samples gained at the same aspect angle consist of similar local structures in that the difference between them is just 2° in depression. Therefore, atoms combined with training samples of multiple target types not only meet the two basic acquirements of dictionary generation, but also take superiorities to adapt the local structures of individual test samples more flexibly and exactly.

3.3. Feature spar si ty

According to Section 3.2, it is observed that training samples are highly similar to local structures with their corresponding test samples. If a test sample is given from one of the target types, the ideal signal component analysis in theory should approximately locate on the atoms related to this target type, especially the ones obtained at the same aspect angle or around it. Therefore, the decomposition result is sparse. It will be utilized as features in recognition experiments.

3.4. Recognition

Refer to the definition in Eq.(1) and assume that there are c types of targets Ti,T2, — ,Tc. The dictionary is A=[B1 B2 • Bc], B, = A A;,2 Am], i=1,2,-,c,

mi is the number of training samples associated with ith target type. The resolved sparse coefficient vector

is X, = [X X2 ••• Xc]t eRM, X[=[x,1 x,,2 ■

xi m ]t, M=m.\ + m2+~ +mc. y is a test sample from kth

target type, ke[1,cj.

According to the analysis in Section 3.2 and Section 3.3, the nonzero elements should mainly locate on

X'k =[xk,i xk2 • •• xkm ]t , and few locate on X\ =

[x,,i xi2 ••• xim ]t, ie[1,c], i^k. Therefore, there

y = AX = X Bt X Bk X'k=X Ak

i=1 j=1

XAjx,,j << y, i * k j=1

k, jxk, j

The recognition process can be divided into three steps:

1) Sum up the arithmetic product of B{ and X'i

sumI = BiX' = Y A'r,jxr,j, i 6 [1, c] j=1

2) Find the minimum distance d between y and sumj-

min d = min||y - sumj, i e [1, c] (11)

3) Recognize the test sample to the type of targets which the minimum d corresponding to.

4. Results and Discussion

4.1. Sparse decomposition

Numerous test samples have been experimented to validate deduction in Section 3.2 about the feasibility of dictionary generation based on training samples and the assumption in Section 3.3 that the decomposition results are sparse. The results show that the obtained decomposition results are sparse indeed and primary sparse coefficients assemble within the range of the corresponded target types.

Ranges at [1, 135], [136, 283], [284, 421], [422, 658], [659, 795], and [796, 1 032] are related to atoms belonging to 1st, 2nd, 3rd, 4th, 5 th, and 6th target types, respectively. As examples, Fig.1 shows LASSO decomposition results of test samples which belong to 1st, 2nd and 3rd type of targets around 0°, 90°, 180°

0° ¿ispoet angle

S8" aspect angle

136 284 422 659 796 I 032 Dictionary

178" aspect angle

1 136 284 422 659 796 I 032: Dictionary1

272" aspect angle

? 2 w -

136 284 422 659 796 1032 I 136 284 422 659 796 1032 Dictionary Dictionary

(a) I si type oflargels

■1st target lype -----2nd target type---3rd target type

■ 4th target type---5th largel type -------6th target type

Fig.1 Sparse coefficients obtained by sparse analysis.

and 270° aspect angles. It is seen that there are few sparse coefficients appearing in Fig.1. Significant sparse coefficients concentrate on the atoms related to the type of test samples while sparse coefficients of atoms corresponding to other target types are relatively much smaller, or even equal to zero.

4.2. Recognition comparison

Two approaches are considered in recognition comparison. One is the currently widely used template based recognition (TBR) using HRRP magnitudes[3-5]. The other is the sparse representation based recognition (SRBR) employing sparse coefficients proposed in this article. Both approaches are implemented in 5 sorts of recognition experiments. The target type number in these 5 sorts of recognition experiments

ranges from 2 to 6 sequentially.

In order to make the results of SRBR and TBR comparable, test samples used in both approaches are produced as described in Section 3.1; templates employed in TBR and the dictionary for sparse representation are combined by training samples formed in Section 3.1 as well. Once the number of target types in a sort of experiments is fixed, numerous recognition experiments are implemented, and the target types in each experiment are selected randomly from the database; the probability of correct classification (PCC) of each sort of experiments is measured by the statistical mean of the results obtained in the corresponding experiments.

Fig.2 depicts the performance of TBR and SRBR during 2 to 6 recognition experiments. It is seen that, when the target type number is fixed, SRBR keeps higher PCC than TBR does. The improvement in 2 to 6 recognition experiments is 2.371 8%, 3.682 8%, 4.680 1%, 5.368 4% and 5.981 7%, respectively. It can also be observed that as the number of target types increases, both of PCC in SRBR and TBR drop. The rate of PCC declining in SRBR is about 14.809 6%, from 93.508 2% to 78.698 6% during 2 to 6 recognition experiments, while that in TBR is more than 18%, from 91.136 4% to 72.716 9%.

Fig.2 Performance comparison of SRBR and TBR in 2 to 6 recognition experiments.

Experiments are implemented using MATLAB on a PC with 2.8 GHz CPU. The average running time for SRBR and TBR is 4.7 s and 2.7 s, respectively.

4.3. Discussion

From Fig.2, it is known that both of TBR and SRBR provide satisfied recognition results. However, SRBR is superior to TBR throughout 2 to 6 recognition experiments. The improvement in SRBR primarily benefits from the insight of sparse representation into signal decomposition. For example, Fig.3 contains a template matching result that a test sample of 2nd target type at 14° aspect angle. The maximum correlation coefficient appears at the range between 659th template and 795th template which corresponds to 5th target type. Hence, the test sample is judged as the 5th target type accord-

ingly. Obviously, this recognition result is wrong. The main reason of the mistake is that the method based on template matching emphasizes on the rough similarity in general between two samples, but ignores the difference in local structures in detail. The judgment lacks accuracy.

Fig.3 Template matching result of a test sample from 2nd target type at 14° aspect angle.

In contrast, sparse representation, like a microscope, provides an insight into the internal local structure composition. During the signal approximation process, sparse analysis keeps adding atoms adapted to signal local structures, evaluating their significance and removing atoms with less similarity. Fig.4 shows the approximation process of the test sample of 2nd target type at 14° aspect angle in sparse representation. Similarly with template matching, at the beginning, sparse analysis firstly fixes the atom related to 5th target type as the original and rough analysis result, as shown in Fig.4(a). During the successive process, atoms are added or deleted according to their significance at each step of approximation, shown as Figs.4(b)-(g). The final result is a perfect approximation of the test sample with atoms as few as possible (see Fig.4(h)).

Fig.4 Process of sparse analysis for the test sample from 2nd target type at 14° aspect angle.

5. Conclusions

(1) For the experimental data with less dimension but large number of training samples, dictionary generation based on training samples is a simple but efficient means.

(2) Sparse representation provides a more accurate way for HRRP analysis. The decomposition results are sparse and exact.

(3) Compared with TBR, SRBR produces an obvious improvement in recognition precision when the target types range from 2 to 6. Furthermore, performance obtained by SRBR is more robust and stable.

(4) Compared to TBR, the running time of SRBR is little longer than TBR. But the difference is within several seconds, so it is still able to meet the demands in real applications.

References

[1] Westerkamp J J, Williams R L, Palomino A P, et al. Synthetic moving target HRR profile generation using measured and modeled target data. Proceedings of SPIE Algorithms for Synthetic Aperture Radar Imagery VII. 2000; 4053: 631-641.

[2] Blasch E P. Modeling intent for a target tracking and identification scenario. Proceedings of SPIE Signal and Data Processing of Small Targets. 2004; 5428: 260269.

[3] Nguyen D H, Kay J H, Orchard B J, et al. Classification and tracking of moving ground vehicles. Lincoln Laboratory Journal 2002; 13(2): 275-308.

[4] Cong S, Hong L, Blasch E. Performance model for joint tracking and ATR with HRR radar. Proceedings of SPIE Algorithms for Synthetic Aperture Radar Imagery XV. 2008; 6970: 1-12.

[5] Turnbaugh M A, Bauer K W, Oxley M E, et al. HRR

signature classification using syntactic pattern recognition. IEEE Aerospace Conference. 2008: 1-9.

[6] Liu J, Zhang J Y, Zhao F. A new time-shift invariant feature of radar HRRPs. Journal of electronics & Information Technology 2008; 30(8): 1949-1953. [in Chinese]

[7] Wang D W, Li Y J, Zhang K. Time-frequency signature extraction and recognition of target based on S-transform. Acta Aeronautica et Astronautica Sinica 2009; 30(2): 305-309. [in Chinese]

[8] Morris H C, de Pass M M. Wavelet feature extraction of HRR radar profiles using generalized Gaussian distributions for automatic target recognition. Proceedings of SPIE Signal Processing, Sensor Fusion, and Target Recognition XIV. 2005; 5809: 165-175.

[9] Liu H W, Yang Z, He K, et al. Radar high range resolution profiles recognition based on wavelet packet and subband fusion. IEEE International Conference on Acoustics, Speech, and Signal Processing. 2005; 5: 445-448.

[10] Chen B, Liu H W, Bao Z. An efficient kernel optimization method for high range resolution profile recognition. Radar International Conference. 2006: 1-4.

[11] Liu H W, Su H T, Bao Z. Radar high range resolution profiles feature extraction based on kernel PCA and kernel ICA. Advances in Neural Networks 2005; 3496: 913-918.

[12] Wang J Y, Yin Z K, Zhang C M. Sparse analysis and preliminary application in signal and image. Sichuan: Southwest Jiaotong University Press, 2006. [in Chinese]

[13] Wright J, Ganesh A, Zhou Z H, et al. Demo: robust face recognition via sparse representation. IEEE International Conference on Automatic Face and Gesture Recognition. 2008: 1-2.

[14] Wright J, Yang A Y, Ganesh A, et al. Robust face recognition via sparse representation. IEEE Transactions on Pattern Analysis and Machine Intelligence 2009; 31(2): 210-227.

[15] Gkalelis N, Tefas A, Pitas I. Sparse human movement representation and recognition. IEEE 10th Workshop Multimedia Signal Processing. 2008: 165-169.

[16] Hang X Y. Cancer classification by sparse representation using microarray gene expression data. IEEE International Conference on Bioinformatics and Biomeidcine Workshops. 2008: 174-177.

[17] Mallat S G, Zhang Z F. Matching pursuits with time-frequency dictionaries. IEEE Transactions on Signal Processing 1993; 41(12): 3397-3415.

[18] Chen S S, Donoho D L, Saunders M A. Atomic decomposition by basis pursuit. SIAM review 2001; 43(1): 129-159.

[19] Donoho D L, Tsaig Y. Fast solution of l1-norm minimization problems when the solution may be sparse. <http://www-dsp.rice.edu/files/cs/FastL1. pdf >. 2006.

[20] Donoho D L. For most large underdetermined systems of linear equations the minimal 1-norm solution is also the sparsest solution. Communications on Pure and Applied Mathematics 2006; 59(6): 797-829.

[21] Yang J Y, Peng Y G, Xu W L, et al. Ways to sparse representation: an overview. Science in China Series F: Information Sciences 2009; 52(4): 695-703.

[22] Amaldi E, Kann V. On the approximability of minimizing nonzero variables or unsatisfied relations in linear systems. Theoretical Computer Science 1998; 209(1-2): 237-260.

[23] Osborne M R, Presnell B, Turlach B A. A new approach to variable selection in least squares problems. IMA Journal of Numerical Analysis 2000; 20(3): 389-403.

[24] Ross T D, Worrell S W, Velten V J, et al. Standard SAR ATR evaluation experiments using the MSTAR public release data set. Proceedings of SPIE Algorithms for Synthetic Aperture Radar Imagery V. 1998; 3370: 566-573.

[25] Gross D, Oppenheimer M, Kaler B, et al. Preliminary comparison of high range resolution signatures of moving and stationary ground vehicles. Proceedings of SPIE Algorithms for Synthetic Aperture Radar Imagery IX. 2002; 4727: 205-212.

[26] Skolnik M l. Introduction to radar systems. 3rd ed. Beijing: Publishing House of Electronics Industry, 2007.

[27] Huang P K, Yin H C, Xu X J. Characteristics of radar targets. Beijing: Publishing House of Electronics Industry, 2005. [in Chinese]

[28] Shirman Y D. Computer simulation of aerial target radar scattering, recognition, detection, and tracking. Boston: Artech House, 2002.

[29] Du L. Study on radar HRRP target recognition. PhD thesis, Xi'an:Xidian University, 2007. [in Chinese]

[30] Li H J, Yang S H. Using range profiles as feature vectors to identify aerospace objects. IEEE Transactions on Antennas and Propagation 1993; 41(3): 261-268.

Biographies:

Zhou Nuo Born in 1981, she is currently a Ph.D. candidate in Beijing University of Aeronautics and Astronautics. Her research interests include signal processing and automatic target recognition. E-mail: joycesmiling@163.com

Chen Wei Born in 1963, he is an associate professor in Beijing University of Aeronautics and Astronautics. His research interests include signal and information processing, information fusion, digital image processing and target recognition.

E-mail: chenwei@buaa.edu.cn