Automatic Target Acquisition of The DEMo III Program
Automatic Target Acquisition of The DEMo III Program
■-■«•■:=--"-as
Army Research Laboratory
Adelphi,MD 20783-1197
Sandor Der, Alex Chan, Gary Stolovy, Michael Lander, and Matthew Thielke
Sensors and Electron Devices Directorate, ARL
1. Introduction 1
4. Target Recognition 18
4.1 Introduction 18
4.2 The Data 18
4.3 Algorithm Architecture 19
4.3.1 PCA Decomposition/Reconstruction Architecture 20
4.3.2 Linear Weighting of Reconstruction Error 22
4.3.3 Scale and Shift Search Space 22
4.4 Experimental Results 22
List of Figures
Figure 1. ROC curve on Hunter-Liggett April 1992 imagery. The horizontal axis gives the
average number of false alarms per frame, the vertical axis is the target detection
probability '
Figure 2. ROC curve on Yuma July 1992 imagery 7
Figure 3. ROC curve on Greyling August 1992 imagery 8
Figure 4. Easy image from Hunter-Liggett April 1992 data set 8
Figure 5. Results on previous image 9
Figure 6. Moderate image from Hunter-Liggett April 1992 data set 9
Figure 7. Results on previous image 10
Figure 8. ROC curve on 12-bit 2001 Fort Indiantown gap data 11
Figure 9. Histogram of grey levels of 37 Fort Indiantown gap images with no targets 12
Figure 10. Histogram of grey levels of Fort Indiantown gap images with no targets. The y
axis has been magnified 100 times to show tail of distribution 12
Figure 11. Histogram of grey levels of 60 Fort Indiantown gap images with targets 13
Figure 12. Histogram of grey levels of Fort Indiantown gap images with targets. The y axis
has been magnified lOOx to show tail of distribution 13
Figure 13. 100 most dominant PCA eigenvectors extracted from the target chips 15
Figure 14. Performance curves ''
Figure 15. Eigenvectors of HMMWV front side 20
Figure 16. Eigenvectors of HMMWV leftside 20
Figure 17. Eigenvectors of HMMWV back side 20
Figure 18. Eigenvectors of HMMWV right side 20
Figure 19. Eigenvectors of Ml 13 front side 20
Figure 20. Eigenvectors of Ml 13 leftside 20
Figure 21. Eigenvectors of Ml 13 backside 21
Figure 22. Eigenvectors of Ml 13 right side 21
Figure 23. Eigenvectors of target board 1 21
Figure 24. Eigenvectors of target board 2 21
Figure 25. A simple image containing only clutter 23
Figure 26. An image of the left side of an Ml 13 24
Figure 27. Side view of an Ml 13 24
Figure 28. Front view of an Ml 13, on the road near the center of the image 25
Figure 29. View of target board type II 25
List of Tables
in
1. Introduction
This work was performed for the DEMO III Unmanned Ground Vehicle (UGV) program, which
is developing UGVs that will assist U.S. Army scouts. The Electro-Optics Infrared (EOIR)
Image Processing branch (AMSRL-SE-SE) has been tasked with developing algorithms for
acquiring and recognizing targets imaged by the Wescam Forward-Looking Infrared (FLIR)
sensor. These images are sent back to the user upon request or when the automatic target
recognizer (ATR) indicates a location of interest. The user makes the ultimate decision about
whether an object in an image is actually a target. The ATR reduces the bandwidth requirement
of the communication link because the imagery can be sent back at reduced resolution, except
those regions indicated by the ATR as being possible targets. The algorithms consist of a front-
end detector, a clutter rejector, and a recognizer. The next three sections describe these
components.
The algorithm described in this report was designed to address a need for a detection algorithm
with wide applicability which could serve as a prescreener/detector for a number of applications.
While most automatic target detection/recognition (ATD/R) algorithms use much problem-
specific knowledge to improve performance, the result is an algorithm that is tailored to specific
target types and poses. The approximate range to target is often required, with varying amounts
of tolerance. For example, in some scenarios, it is assumed that the range is known to within a
meter from a laser range finder or a digital map. In other scenarios, only the range to the center
of the field-of-view and the depression angle is known so that a flat earth approximation
provides the best estimate. Many algorithms, both model-based and learning-based, either
require accurate range information or compensate for inaccurate information by attempting to
detect targets at a number of different ranges within the tolerance of the range. Because many
such algorithms are quite sensitive to scale, even a modest range tolerance requires that the
algorithm iterate through a large number of closely spaced scales, driving up both the
computational complexity and the false alarm rate. Algorithms have often used statistical
methods [1] or view-based neural networks [2, 3, 4].
The proximate motivation for the development of the scale-insensitive algorithm was to provide
a fast prescreener for a robotic application for which no range information was available. The
algorithm instead attempted to find targets at all ranges between some reasonable minimum,
determined from operational requirements and the maximum effective range of the sensor.
Another motivation was to develop an algorithm that could be applied to a wide variety of image
sets and sensor types, which required it to perform consistently on new data, without the severe
degradation in performance that commonly occurs with learning algorithms, such as neural
networks and principal component analysis (PCA)-based methods, that have been trained on a
limited variety of sensor types, terrain types, and environmental conditions. While we recognize
that with a suitable training set, learning algorithms will often perform better than other methods,
this typically requires a large and expensive training set, which is sometimes not feasible.
where/£,/) is the grey level value of the pixel in the kth row and /th column, N„,(ij) is the
neighborhood of the pixel (ij), defined as a rectangle whose width is the length of the longest
vehicle in the target set and whose height is the height of the tallest vehicle in the target set. For
the applications that we have considered, the width is 7 m and the height is 3 m.
2.2.2 Contrastbox-Feature 1
The contrastbox feature measures the average grey level over a target-sized region and compares
it to the grey level of the local background. It was chosen because many pixels that are not on the
engine or on other particularly hot portions of the target are still somewhat warmer than the
natural background. This feature has been used by a large number of authors. The feature is
defined as
F
h=— n
1 /(*.0-— I /(W). (2)
in (k,l)eNin(ij) "out (kJ)eNM(i,j)
where nout is the number of pixels in N0U,{iJ), nin is the number of pixels in Nin(i,j), Nin(iJ) is the
target-sized neighborhood defined above, and the neighborhood Nout(ij) contains all of the pixels
in a larger rectangle around (ij), except those pixels that are in Nin(ij).
^=7- U
I Gin(ij)-— n
£ Gout{i,j), (3)
in (U)eN,„(ij) out (k,l)SNM(i,j)
where
Gin{i,j) = Gl{i,j) + Gl{i,j), (4)
G*n(i,j) = \f(iJ)-f(i,j + l)\, (5)
Gjn=\f{i,j)-f{i + \,j)\, (6)
and G0lli{ij) is defined similarly.
where
/UU) = - n
I /(W), (9)
in (U)eNin(i.j)
^=i- I ff/nM = —
n
in (U)eN,„{iJ) H
X *UU).
out (k,l)eNm,{i,j)
<10>
where
Hin{UJ) = Ht{i,j) + Hl{i,j), (11)
and G„m{ij) is defined similarly. The parameter / is a function of field-of-view of the system,
target range, and target size. If the sensor or target is significantly tilted, the functions can be
suitably modified to measure edges in other directions, at the cost of more computation and less
discrimination ability.
where
^=77 I Fu>
M
Ü6)
all(U)
and
2
!
^=T7 X (Fu~ßm) ■ (17)
M
all(k,l)
After normalization, the features, each of which is calculated for each pixel, are linearly
combined into a confidence image,
G
iJ = l^nF^N, (18)
m=0
where the feature weights <jm are determined using an algorithm not described here. The
confidence value of each pixel is mapped by a scaling function S:Si -»[0,1], as
S(Gy)= \-eaGl (19)
where a is a constant.
This scaling does not change the relative value of the various pixels, it merely scales them to the
interval [0,1] for convenience. Confidence numbers are often limited to this interval because they
are estimates of the a posteriori probability. While this is not true for our algorithm, using this
interval is convenient for evaluators.
To determine the detection locations from the scaled confidence image, the pixel value with the
maximum confidence value is chosen. Then a target-sized neighborhood around the image is set
to zero so that the search for subsequent detections will not choose a pixel location
corresponding to the same target. The process is then repeated for the fixed number of detections
chosen before the algorithm was run.
*1?*-l
ja
1
Figure 4. Easy image from Hunter-Liggett April
1992 data set.
Figure 5. Results on previous image.
■ "4
lll»*^w
*^*''* %
mm*
The algorithm was also tested on data collected specifically for the DEMO III program to ensure
that performance does not degrade because of the different sensor. The DEMO III sensor is
sensitive in the midwave, 3-5 u region, while the previous data was in the longwave, 8-12 u
region. Figure 8 shows an ROC curve on data collected by the DEMO III sensor at the Fort
Indiantown Gap DEMO III site.
10
Figure 8. ROC curve on 12-bit 2001 Fort Indiantown gap data.
To determine if raw grey level information could be used to locate targets without the use of
shape information, histograms of the 37 Fort Indiantown Gap images were formed. Figure 9
shows the histogram for images that contain no targets, and Figure 10 shows the histogram
magnified lOOx to show the tail of the distribution. The idea is to determine if the tail for images
with targets is higher than for images without targets. Figures 11 and 12 show the corresponding
histograms for images with targets. It appears that the raw grey level information would be a
poor discriminant for target detection.
The algorithm is being used by the DEMO III program to reduce the amount of imagery that
must be trasmitted via radiolink to a human user. It will also be used by the Sensors for UGV
program at Night Vision and Electronic Sensors Directorate (NVESD), to prescreen uncooled
FLIR imagery and indicate potential targets that should be looked at more closely with an active
laser sensor. It has been used by a synthetic image validation tool, by measuring the performance
of the algorithm in comparison to real imagery.
11
400000.0
300000.0
200000.0
100000.0
0.0 JitLb
0.0 20.0 40.0 60.0 80.0
Figure 9. Histogram of grey levels of 37 Fort Indiantown gap images with no targets.
4000.0
3000.0
2000.0
1000.0
0.0
0.0 20.0 40.0 60.0
Figure 10. Histogram of grey levels of Fort Indiantown gap images with no targets.
The y axis has been magnified 100 times to show tail of distribution.
12
1500000.0 n -
1000000.0 - -
500000.0 -
0.0
.
20.0
w hh-^
40.0 60.0 80.0
Figure 11. Histogram of grey levels of 60 Fort Indiantown gap images with targets.
2000.0
1500.0
1000.0
500.0
Figure 12. Histogram of grey levels of Fort Indiantown gap images with targets.
The y axis has been magnified lOOx to show tail of distribution.
13
3. The Clutter Rejection Algorithm
The purpose of the clutter rejection algorithm is to further examine locations indicated by the
detector to determine if targets are present. Because the clutter rejecter doesn't examine the
whole image, the algorithm can be more computationally intensive. The algorithm used here is
based on PCA-based dimensionality reduction, followed by a multilayer perceptron (MLP)
trained to reject clutter and accept targets.
The limited diversity of the training set required that dimensionality reduction be performed
before a neural network is used. This is important because using a learning algorithm prior to
dimensionality reduction requires a large and diverse training set to avoid overtraining, resulting
in a sharp difference between training and testing performance. The architecture of the algorithm
has a front-end PCA dimensionality reduction component, followed by a multilayer perceptron
that uses only the individual PCA components as inputs. The output of the MLP, along with the
feature values from the detector, are combined by a higher level MLP. The following sections
describe the PCA and MLP components and describe experimental results.
3.1 PCA
Also referred to as the Hotelling transform or the discrete Karhunen-Loeve transform, PCA is
based on statistical properties of vector representations. PCA is an important tool for image
processing because it has several useful properties, such as decorrelation of data and compaction
of information (energy). We provide here a summary of the basic theory of PCA.
Assume a population of random vectors of the form
-n
x
2
x= (20)
.r3.
The mean vector and the covariance matrix of the vector population x are defined as
mx = £\-y\, and (21)
Cx=£{(x-mJ(x-mx)r}, (22)
where £arg is the expected value of the argument, and T indicates vector transposition. Because x
is «-dimensional, Cx is a matrix of order nxn. Element c„ of Cx is the variance of x; (the /'th
component of the x vectors in the population), and element Cy of Cx is the covariance between
elements x, Xj of these vectors. The matrix Cx is real and symmetric. If elements x, and Xj are
uncorrelated, their covariance is zero and, therefore, cy = c/( = 0. For N vector samples from a
random population, the mean vector and covariance matrix can be approximated from the
samples by
14
1
m X =— V x P,' and
XT L^ n
(23)
N
1
cx=TrSK<- mm
x x) (24)
p=i
Because Cx is real and symmetric, we can always find a set of« orthonormal eigenvectors for
this covariance matrix. Figure 13 shows the first 100 (out of the 800 possible in this case) most
dominant PCA eigen-targets and eigen-clutters, which were extracted from the target and clutter
chips in the training set, respectively. Having the largest eigenvalues, these eigenvectors capture
the greatest variance or energy as well as the most meaningful features among the training data.
I&Ä..T-* V*^* :-■■* ~"**-- S-pr^w ££»** :-4±..a * * «... ■-,*£ £"fc-£r^
:,
91 ^ ^i;"s^v^1
wÄIlssBä^ls» ü^&iMiM£i «^^^Ws^ssäf äSS^^äS^B ^«SS^F*, Ji ^s*** ^ÄfcKKT <& ^!»(Ä«i««(i &m<&* z^irnm ^ÄtejIaCTär ^^^^^
Figure 13. 100 most dominant PCA eigenvectors extracted from the target chips.
Let ej and A/', / = 1, 2,..., n, be the eigenvectors and the corresponding eigenvalues of Cx, sorted
in a descending order so that kj > A,+i fory = 1,2,..., n-\. Let A be a matrix whose rows are
formed from the eigenvectors of Cx, such that
A= (25)
This A matrix can be used as a linear transformation matrix that maps the xs into vectors,
denoted by ys, as follows:
y = A(x - mx) (26)
15
■""»(fin
Conversely, we may want to reconstruct vector x from vector y. Because the rows of A are
orthonormal vectors, A-1 = AT. Therefore, any vector x can be reconstructed from its
corresponding y by the relation
T.
x = A'y + mx. (27)
Instead of using all the eigenvectors of Cx, we may pick only k eigenvectors corresponding to the
k largest eigenvalues and form a new transformation matrix At of order k x n. In this case, the
resulting y vectors would be ^-dimensional, and the reconstruction given in equation (27) would
no longer be exact. The reconstructed vector using A* is
x = A[y + mx (28)
The mean square error (MSE) between x and x can be computed by the expression
Because the A/s decrease monotonically, equation (29) shows that we can minimize the error by
selecting the k eigenvectors associated with the k largest eigenvalues. Thus, the PCA transform is
optimal in the sense that it minimizes the MSE between vectors x and their approximations x.
As we can see from figure 13, only the first few score of the eigen-targets contain consistent and
structurally significant information pertaining to the training data. These eigentargets exhibit a
reduction in information content as their associated eigenvalues rapidly decrease. For the less
meaningful eigentargets (say, the 50th and all the way up to the 800th) only high-frequency
information is present. In other words, by choosing k = 50 in equation (29) when n = 800, the
resulting distortion error, e, would be small. While the distortion is negligible, there is a 16-fold
reduction in input dimensionality.
3.2 MLP
After projecting an input chip to a chosen set ofk eigen-targets, the resulting k projection values
are fed to an MLP classifier where they are combined nonlinearly. A typical MLP used in our
experiments has k + 1 input nodes (with an extra bias input), several layers of hidden nodes, and
one output node. In addition to full connections between consecutive layers, there are also
shortcut connections directly from one layer to all other layers, which may speed up the learning
process. The MLP classifier is trained to perform a two-class problem, with training output
values of ±1. Its sole task is to decide whether a given input pattern is a target (indicated by a
high-output value of around +1) or clutter (indicated by a low-output value of around -1). The
MLP is trained in batch mode using Qprop [7], a modified backpropagation algorithm, for a
faster but stable learning course.
Alternatively, the eigenspace transformation can be implemented as an additional linear layer
that attaches to the input layer of the simple MLP above. The resulting augmented MLP
classifier, which is collectively referred to as PCAMLP network in this paper, consists of a
transformation layer and a back-end MLP (BMLP). When the weights connecting the new input
nodes to the kth output node of the transformation layer are initialized with the kth PCA
eigenvector, the linear summation at the Mi transformation output node is equivalent to the kth
projection value.
3.3 Experimental Results
The clutter rejection algorithm was trained on data collected at Aberdeen Proving Ground (APG)
and Fort Knox in 1999, and tested on data collected at APG in 1999 and Fort Knox in 2000. The
APG data were divided so that the training and test data were collected on different days, but the
backgrounds are similar because the Perryman test site is rather uniform. ROC curves for the
detection/clutter rejection system are shown in Figure 14. The curves show that the clutter
rejecter that uses the detector features combined with the output of the MLP by a higher level
MLP performs better than the detector alone on the training and test data. Simply using the
output of the MLP of the clutter rejector results in worse performance than the detector alone
because the MLP does a good job of separating targets and clutter but a poor job of estimating
the confidence.
80
5? 60
d3c_pm_i0n05_13tba_„
d3c J5m_il n05_l 3taa__
d3c_pm_t0ri05_13fba._rrr.1_t
d3c_pmJ1n05_13laa_irr
40 d3c_pm_!0n05_131b__ug1 J
d3c..pm..t1n05..13fa_._gt _t
[Link]..t_ri05. .1_fba.ps! .1
d3c._pm_.1n05_.13laa.p_1 .t
._pm._tC-rt05..13f_a. rml _cf2_ _.
:_pm_tl n05_l3(a__rm1 _CI2Q_1I
20
2 3 4 5
False alarm per frame
*%** .o-b'*
- a'
0
" .*■"" •■ ■__*'■'' __-"
iff o' a ,*
■ ■*" ■-*
1;«"'
jfi .* ••*
d3._pm_nnC5.13(„a_rm!..cr2aJes!_..T„_
20
2 3 4 5
False alarm per frame
17
4. Target Recognition
4.1 Introduction
The algorithm described here was designed to address a need for a recognition algorithm that
could be trained with a small amount of data, with poor range and localization information. The
operational scenario is to examine objects that have been detected by another algorithm to
determine if they are one of the objects stored in an existing image library. The detection
location given by the detection algorithm may be poorly centered on the target, and the range to
the target will not be known. The number of training examples for the four different targets
differed radically. This meant that the chosen algorithm must be able to take advantage of a large
training set when it exists but still be able to perform well for smaller data sets.
Many techniques have been applied to the recognition problem [1]. When training sets have been
large, recognition algorithms have typically used complex learning algorithms that use a large
number of features to discriminate between targets. Often the features are either simply the pixel
values, or simple gradient/wavelet features calculated in a dense grid across the target region
[3,4, 8]. The learning algorithms include complex template matching schemes [8] or neural
networks [2-4, 9, 10]. Learning algorithms that are trained on small data sets tend to generalize
poorly, so we chose not to use these algorithms for this work.
Some algorithm designers have used PC A to compress the data prior to recognition [11]. An
advantage of this approach is that it reduces the number of features that a classifier can use, and
thus reduces the size of the required training set. One disadvantage is that the compression
eliminates some of the information that is useful to perform discrimination, and because the PC A
algorithm optimizes the compression of the data without regard for information that is useful for
discrimination, one cannot expect that PCA gives the most discrimination information possible
for a given number of features.
The data set used for our training was lopsided. The algorithm attempts to recognize four targets,
two real (Ml 13 and HMMWV) and two target boards (TB1 and TB2). For the Ml 13 and
HMMWV, we have 1239 and 2080 suitable training samples, whereas for the target boards we
have 14 and 22 suitable training samples. The ramifications of this imbalance will be discussed.
The remainder of this technote is organized as follows: Section 2 describes the data used to train
and test the system. Section 3 describes the architecture of the recognizer. Section 4 gives the
results of experiments performed on a small test set of imagery. Section 5 contains conclusions
and plans for future work.
19
4.3.1 PCA Decomposition/Reconstruction Architecture
The PCA decomposition described above takes a training set of images and turns it into an
ordered set of eigenvectors and corresponding eigenvalues. This decomposition is performed for
each target-pose group of training samples. Because there are two real targets, which are divided
into four pose groups each, and two target board types, which only have one pose, there are a
total of 10 target-pose groups. We have chosen the number of eigenvectors to retain to be 5, for a
total of 50 stored eigenvectors. The eigenvectors are stored at different scales to account for
variability in the range to potential targets. The eigenvectors for each of the 10 target-pose
groups are shown in Figures 15-24.
j0m
lR-: -m&K-
20
Figure 21. Eigenvectors of Ml 13 back side.
The decomposition stage determines the PCA components of a sample being tested. The
components y, for an input target image x are calculated as
n
Thus, the /th PCA component of an input vector x is simply the dot product of x with the fth
eigenvector. The reconstruction using the first k eigenvectors is
!>;*,■
(3D
£
=1L(xJ-xj)2=\\x-x\\ (32)
21
4.3.2 Linear Weighting of Reconstruction Error
To reduce the bias inherent in the PCA decomposition process, the reconstruction errors for each
target-pose group are multiplied by a fixed weight. Thus, the reconstruction error for the /th
target-pose group, £/ is weighted by a weight (»i. The target-pose decision / is given by
I = argmin^co^/). (33)
77
Table 1. Confusion matrix on test set.
HMMWV M113 TB1 TB2
M113 6 41 30 14
TB1 1 11 10 1
TB2 0 6 2 14
Figure 25 shows a sample image that does not contain a target. Figures 26-29 contain targets.
Some of the targets would be difficult for a human to distinguish. The target in Figure 26 is
difficult to distinguish because the shape is not clearly that of an Ml 13, and there is little interior
information because the whole target is hot. Figure 27 is clearly an Ml 13 because the rectangular
plate on the upper front of the target is a distinguishing characteristic. Notice the target is not
level; this makes recognition more difficult because the templates aren't well aligned. The
algorithm doesn't currently tilt the templates to handle such a case. Doing so would make it more
likely to correctly identify tilted targets but would increase the probability of error on level
targets, and would increase computation time. Figure 28 is a good example of a target that is
difficult for an algorithm to detect but easy for a human. The target does not have a clear
boundary, nor is it hotter than its background. The detector and recognizer give correct results
for this image, but that is unusual. Figure 29 shows a target board type II clearly visible in the
left center of the image. While this is clearly a target board, it is not easy to see which type at this
resolution.
23
Figure 26. An image of the left side of an Ml 13.
24
Figure 28. Front view of an Ml 13, on the road
near the center of the image.
There is a great deal of work that can still be done to improve the system described. Future work
on the detector might include a more systematic evaluation of potential features and an improved
classification scheme that allows useful features that appear to be rarely incorporated. In a small
25
minority of FLIR images of targets, a windshield will reflect cold sky, causing a few pixels to be
extremely dark. The current scheme is not set up to incorporate such features because the
weighting would be quite low since the feature is seldom useful. The recognizer would greatly
benefit from a balanced training set, which would allow for a more sophisticated bias reduction
scheme and would enable the formation of a better PCA representation of each target. The
algorithms could benefit from more input information. All of the components would benefit from
more accurate range information, which could be obtained using accurate registration to digital
maps, from structure and motion algorithms, or from a laser range finder. The UGV has a color
TV camera collocated with the FLIR, which could provide additional target screening capability
during the day.
References
1. Bhanu, B. "Automatic Target Recognition: State of the Art Survey." IEEE Transactions on
Aerospace and Electronic Systems, vol. 22, no. 4, pp. 364-379, 1986.
2. Roth, M. W. "Survey of Neural Network Technology for Automatic Target Recognition."
IEEE Transactions on Neural Networks, vol. 1, no. 1, pp. 28-43, 1990.
3. Hecht-Nielsen, R., and Y.-T. Zhou. "VARTAC: A Foveal Active Vision ATR System."
IEEE Transactions on Neural Networks, vol. 8, no. 7, pp. 1309-1321, 1995.
4. Wang, L., S. Der, and N. Nasrabadi. "Modular Neural Network Recognition of Targets in
FLIR Imagery." IEEE Transactions on Image Processing, vol. 7, no. 8, August 1998.
5. Bhanu, B., and T. Jones. "Image Understanding Research for Automatic Target
Recognition." Aerospace Electronic Systems Magazine, pp. 15-22, October 1993.
6. Jolliffee, I. T. Principal Component Analysis, New York: Springer-Verlag, 1986.
7. Fahlman, S. "Faster Learning Variations on Back-Propagation: An Empirical Study."
Proceedings of the 1988 Connectionist Models Summer School, Morgan Kaufmann,
pp. 38-51.
8. Chan, A., and N. Nasrabadi. "Wavelet Based Vector Quantization for Automatic Target
Recognition." International Journal on Artificial Intelligence Tools, vol. 6, no. 2, pp.
165-178, 1997.
9. Neubauer, C. "Evaluation of Convolutional Neural Networks for Visual Recognition." IEEE
Transactions on Neural Networks, vol. 9, no. 4, pp. 685-696, 1998.
10. Chen, C. H., and G. G. Lee. "Multi-Resolution Wavelet Analysis Based Feature Extraction
for Neural Network Classification." Proceedings International Conference Neural Networks
5, pp. 1416-1421,1996.
26
11. Chan L., S. Der, and N. Nasrabadi, "Analysis of Dualband FLIR Imagery for Automatic
Target Detection." Smart Imaging Systems, Bahrain Javidi, (ed.), SPIE Press, 2001.
12. Der, S., C. Dwan, A. Chan, H. Kwon, andN. Nasrabadi. "Scale-Insensitive Detection
Algorithm for FLIR Imagery." ARL-TN-175, U.S. Army Research Laboratory, Adelphi,
MD, February 2000.
13. Kwon, H., S. Der, and N. Nasrabadi. "Multisensor Target Detection Using Adaptive Feature-
Based Fusion." SPIE Aerosense, April 2001.
27
Form Approved
REPORT DOCUMENTATION PAGE OMB No. 0704-0188
Public reporting burden for this collection of information is estimated to average 1 hour per response, including the time for reviewing instructions, searching existing data sources,
gathering and maintaining the data needed, and completing and reviewing the collection of information. Send comments regarding this burden estimate or any other aspect of this
collection of information, including suggestions for reducing this burden, to Washington Headquarters Services, Directorate for Information Operations and Reports, 1215 Jefferson
Davis Highway, Suite 1204, Arlington, VA 22202-4302, and to the Office of Management and Budget, Paperwork Reduction Project (0704-0188), Washington, DC 20503.
1. AGENCY USE ONLY (Leave blank) 2. REPORT DATE 3. REPORT TYPE AND DATES COVERED
17. SECURITY CLASSIFICATION 18. SECURITY CLASSIFICATION 19. SECURITY CLASSIFICATION 20. LIMITATION OF ABSTRACT
OF REPORT OF THIS PAGE OF ABSTRACT
29