8
Feature Fusion and Label Propagation for Textured Object Video Segmentation V. B. Surya Prasath a , Rengarajan Pelapur a , Kannappan Palaniappan a , Gunasekaran Seetharaman b a Department of Computer Science, University of Missouri, Columbia MO 65211 USA b Air Force Research Lab, Rome, NY USA ABSTRACT We study an efficient texture segmentation model for multichannel videos using a local feature fitting based active contour scheme. We propose a flexible motion segmentation approach using fused features computed from texture and intensity components in a globally convex continuous optimization and fusion framework. A fast numerical implementation is demonstrated using an efficient dual minimization formulation. The novel contributions include the fusion of local feature density functions including luminance-chromaticity and local texture in a globally convex active contour variational method, combined with label propagation in scale space using noisy sparse object labels initialized from long term optical flow-based point trajectories. We provide a proof-of-concept demonstration of this novel multi-scale label propagation approach to video object segmentation using synthetic textured video objects embedded in a noisy background and starting with sparse label set trajectories for each object. 1. INTRODUCTION The proliferation of full motion video (FMV) realtime data streams and forensic archives presents a great challenge for automatic processing, exploitation and dissemination. For example, 72 hours of new video is uploaded every minute to the popular video sharing site YouTube. The effective analysis and mining of this exponentially expanding volume of video data can be categorized as a Big Data problem for which a variety of tools from disparate fields such as computer vision, machine learning and multimodal fusion are needed. Object segmentation in videos a critical enabling technology in video understanding and developing content-based video retrieval systems with practical capabilities. Recently there have been efforts to leverage emerging image and video segmentation techniques from computer vision. We can classify these segmentation methods into three broad categories: (a) supervised where the user is required to scribble or select object labels, (b) semi-supervised that uses a small collection of labeled training data in combination with a large corpus of unlabeled data using active learning methods, and (c) unsupervised where the segmentation is completely automatic. In what follows we review some of the main methods which show promise in object segmentation from videos and refer to 1 for other related schemes. Under the category of unsupervised methods: Dense motion segmentation by extending super-pixels (named supervoxels), 2 graph based methods. 3 These methods can lead to over segmentations and leakage during propagation of labels. To avoid over-segmentation one can use supervised or semi-supervised methods. 4, 5 Though these supervised methods obtain better results in general, they usually require labeling from the user which can take considerable effort. In this work we consider a variational approach for label propagation using dense optical flow combined with a multi-scale segmentation model. This unsupervised method uses texture, intensity and color features in a coherent variational formulation for object segmentation from natural videos. For this purpose we utilize the fully unsupervised long distance optical flow to obtain initial labels. 6 Since in inhomogeneous regions there are no reliable structures which can be captured by motion estimation techniques, the object motion vector field is usually sparse. Moreover, computational constraints require reducing the number of motion vectors being analyzed. In this paper we study how to obtain dense segmentations from sparse clusters of object motion vectors. For this purpose we use the recently studied color, texture feature fusion based globally convex active Further author information: (Send correspondence to V. B. S. Prasath).: E-mail: [email protected] Geospatial InfoFusion and Video Analytics IV; and Motion Imagery for ISR and Situational Awareness II, Matthew F. Pellechia, Kannappan Palaniappan, Shiloh L. Dockstader, Peter J. Doucette, Donnie Self, Eds., Proc. of SPIE Vol. 9089, 908904 · © 2014 SPIE · CCC code: 0277-786X/14/$18 · doi: 10.1117/12.2052983 Proc. of SPIE Vol. 9089 908904-1

Feature Fusion and Label Propagation for Textured Object Video …cell.missouri.edu/media/publications/Surya-Videoseg-SPIE-2014.pdf · Feature Fusion and Label Propagation for Textured

  • Upload
    others

  • View
    7

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Feature Fusion and Label Propagation for Textured Object Video …cell.missouri.edu/media/publications/Surya-Videoseg-SPIE-2014.pdf · Feature Fusion and Label Propagation for Textured

Feature Fusion and Label Propagation for TexturedObject Video Segmentation

V. B. Surya Prasatha, Rengarajan Pelapura, Kannappan Palaniappana,Gunasekaran Seetharamanb

a Department of Computer Science, University of Missouri, Columbia MO 65211 USAb Air Force Research Lab, Rome, NY USA

ABSTRACT

We study an efficient texture segmentation model for multichannel videos using a local feature fitting basedactive contour scheme. We propose a flexible motion segmentation approach using fused features computedfrom texture and intensity components in a globally convex continuous optimization and fusion framework. Afast numerical implementation is demonstrated using an efficient dual minimization formulation. The novelcontributions include the fusion of local feature density functions including luminance-chromaticity and localtexture in a globally convex active contour variational method, combined with label propagation in scale spaceusing noisy sparse object labels initialized from long term optical flow-based point trajectories. We provide aproof-of-concept demonstration of this novel multi-scale label propagation approach to video object segmentationusing synthetic textured video objects embedded in a noisy background and starting with sparse label settrajectories for each object.

1. INTRODUCTION

The proliferation of full motion video (FMV) realtime data streams and forensic archives presents a greatchallenge for automatic processing, exploitation and dissemination. For example, 72 hours of new video isuploaded every minute to the popular video sharing site YouTube. The effective analysis and mining of thisexponentially expanding volume of video data can be categorized as a Big Data problem for which a variety oftools from disparate fields such as computer vision, machine learning and multimodal fusion are needed. Objectsegmentation in videos a critical enabling technology in video understanding and developing content-based videoretrieval systems with practical capabilities. Recently there have been efforts to leverage emerging image andvideo segmentation techniques from computer vision. We can classify these segmentation methods into threebroad categories: (a) supervised where the user is required to scribble or select object labels, (b) semi-supervisedthat uses a small collection of labeled training data in combination with a large corpus of unlabeled data usingactive learning methods, and (c) unsupervised where the segmentation is completely automatic.

In what follows we review some of the main methods which show promise in object segmentation fromvideos and refer to1 for other related schemes. Under the category of unsupervised methods: Dense motionsegmentation by extending super-pixels (named supervoxels),2 graph based methods.3 These methods can lead toover segmentations and leakage during propagation of labels. To avoid over-segmentation one can use supervisedor semi-supervised methods.4,5 Though these supervised methods obtain better results in general, they usuallyrequire labeling from the user which can take considerable effort.

In this work we consider a variational approach for label propagation using dense optical flow combined witha multi-scale segmentation model. This unsupervised method uses texture, intensity and color features in acoherent variational formulation for object segmentation from natural videos. For this purpose we utilize thefully unsupervised long distance optical flow to obtain initial labels.6 Since in inhomogeneous regions there areno reliable structures which can be captured by motion estimation techniques, the object motion vector fieldis usually sparse. Moreover, computational constraints require reducing the number of motion vectors beinganalyzed. In this paper we study how to obtain dense segmentations from sparse clusters of object motionvectors. For this purpose we use the recently studied color, texture feature fusion based globally convex active

Further author information: (Send correspondence to V. B. S. Prasath).: E-mail: [email protected]

Geospatial InfoFusion and Video Analytics IV; and Motion Imagery for ISR and Situational Awareness II,Matthew F. Pellechia, Kannappan Palaniappan, Shiloh L. Dockstader, Peter J. Doucette, Donnie Self, Eds.,Proc. of SPIE Vol. 9089, 908904 · © 2014 SPIE · CCC code: 0277-786X/14/$18 · doi: 10.1117/12.2052983

Proc. of SPIE Vol. 9089 908904-1

Page 2: Feature Fusion and Label Propagation for Textured Object Video …cell.missouri.edu/media/publications/Surya-Videoseg-SPIE-2014.pdf · Feature Fusion and Label Propagation for Textured

contour segmentation method which works well for a variety of natural images.7 Experimental results withsynthetic cases illustrating different scenarios are used to show the advantage of using our variational approach.

The rest of the paper is organized as follows. Section 2 introduces the variational segmentation model alongwith the label propagation using point trajectories from dense optical flow and provides some synthetic sequencesand their corresponding segmentations based on our scheme. Finally, Section 3 concludes the paper.

2. SEGMENTATION USING SPATIAL FEATURE FUSION COMBINED WITHMULTISCALE LABEL PROPAGATION

Before embarking upon the detailed explanation of our proposed model for video segmentation we briefly recallthe globally convex version of the active contour method for 2D color image segmentation and refer to7,8 for thedetailed analysis.

2.1 Globally Convex Chan and Vese Spatial Segmentation Model

The convex formulation of the traditional vector valued Chan and Vese model,9 amounts to solving the mini-mization problem:7,8

min0≤φ≤1

µ

∫Ω

|∇φ(x)| dx+

3∑i=3

∫Ω

λiRi(x, ci)φ(x) dx,

(1)

where φ : Ω → [0, 1] is a function of bounded variation, u ∈ BV (Ω) and Ri(x, ci) = (Ii − ciin)2 − (Ii − ciout)2

(i = 1, 2, 3) which is known as the image fitting term, x ∈ Ω represents the pixel x = (x1, x2) and I := (I1, I2, I3) :Ω → R3 represents the input RGB image. The vector ci = (ciin, c

iout) represents the averages (mean values) of

each channel Ii, respectively inside and outside of the segmentation contour. The parameters λi are positivescalars weighting the fitting terms Ri(x, ci) and can be chosen depending upon the imaging modality. The totalvariation term (first term in Eqn.(1)) keeps the segmentation curve regular. It can be shown that for any λi ≥ 0,there exists a minimizer φ of (1), which is a global minimum.8 The energy functional is homogeneous of degreeone in u and has a global minimizer by restricting u such that 0 ≤ u(x) ≤ 1. Moreover, the above model (1)does not require a level set based implementation as in the original formulation9 and can be optimized veryefficiently.10

Since the Chan and Vese model relies only on the mean (piecewise constant) gray values of the input channel Ii

alone (for each channel i), it can neither capture the spatially varying intensity regions nor textured regions withfidelity. In this work, we use the extension of the globally convex Chan-Vese model (1) to support more generalfitting terms and to handle textured regions across multiple channels. For this purpose we use the followingbrightness-chromaticity decomposition based globally convex active contour segmentation model.7 First the inputimage I is decomposed into four components (channels), namely the intensity of brightness I := |I| and threechromaticity components (C1, C2, C3) := (I1/ |I| , I2/ |I| , I3/ |I|). This mapping (I1, I2, I3) → (I, C1, C2, C3)separates the intensity and color information and is found to be useful for image denoising,11 colorization,12 andsegmentation.7 Following the globally convex version of the Chan and Vese model (1), we extend it to the color(RGB) image segmentation as follows.

min0≤φ≤1,ρ

E(φ, ρ) = µ

∫Ω

|∇φ|+∫

Ω

(λT FT + λC FC)φdx (2)

where the multichannel edge-based texture feature fitting term is given by,

FT = (ω × |Gσ ?∇I| − cin)2 − (ω × |Gσ ?∇I| − cout)2 (3)

with ω being the local cumulative intensity histogram weight and the intensity based localized chromaticityfitting term,

FC =

3∑i=1

∣∣Gσ ? Ci − ciin∣∣2 − ∣∣Gσ ? Ci − ciout∣∣2 . (4)

Proc. of SPIE Vol. 9089 908904-2

Page 3: Feature Fusion and Label Propagation for Textured Object Video …cell.missouri.edu/media/publications/Surya-Videoseg-SPIE-2014.pdf · Feature Fusion and Label Propagation for Textured

(a) Input (b) FT (c) FC (d) λT = 0, λC = 1 (e) λT = 1, λC = 0

Figure 1. Globally convex feature fusion based active contour method for color image segmentation.7 (a) Input RGBimage of size 321×321 (b) Multichannel edge-based texture feature fitting term FT , see Eqn. (3) (c) Total intensity basedlocalized chromaticity feature fitting term FC , see Eqn. (4) (d-e) Switching off one of the feature fitting terms gives adifferent object segmentation. The image is taken from the Berkeley Segmentation DataSet (BSDS500).13

These two terms utilize the texture (gradient), and intensity and chromaticity information from the input imagerespectively. The parameter vector is ρ := (cin, cour, c

1in, c

1out, c

2in, c

2out, c

3in, c

3out), the components are the mean

values for image I. Figure 1 shows an example result using the energy minimization function (2) for a naturalimage along with their fitting terms defined in Eqns. (3-4). The energy minimization in Eqn. (2) is solved usinga fast dual minimization scheme7 and will be utilized in our label propagation scheme which we describe next.

2.2 Object Segmentation Based On Label Propagation

We study a variational model that propagates labels across scale space within homogeneous object regionssparsely sampled by clustered point trajectories. We show that we can obtain dense accurate segmentationsfrom sparse initial labels and multiscale label propagation. Our method overcomes the over-segmentation issueassociated with static single frame image segmentation applied to video streams. We first utilize optical flowto extract motion vectors between a pair of frames to be used in our dense segmentation. Due to its efficiencywe use the dense optical flow method studied by Brox and Malik known as Large Displacement Optical Flow(LDOF).6 Clustering analysis of point trajectories is used as a robust tool to initialize (moving) objects andextract object regions from video shots in a fully unsupervised manner. The main question is how to obtaindense or filled segmentations from sparse clusters of point trajectories. For this purpose we utilize the globallyconvex segmentation model in (2) as an additional grouping step with an extension to support label propagationacross scale space to increase robustness and accuracy. We start with the video shot as input, compute pointtrajectories.14 This yields a discrete sparse set of labels (roughly around ∼ 30% pixels are labelled), which areconsistent over the whole shot.

Let u = (u1, u2, . . . , un) : Ω→ 0, 1n, n ∈ N point trajectory labels, i.e.,

ui =

1 if x ∈ Li,0 else,

(5)

where Li is the set of co-ordinates occupied by a trajectory with label i. The objective is to find a functionu = (u1, u2, . . . , un) : Ω → 0, 1n that remains close to the given labels for points in L := ∪ni=1Li. We canminimize,

Edata(u) =1

2

∫Ω

χL

n∑i=1

(ui − ui)2 dx (6)

where χL : Ω → 0, 1 is the label indicator function, a characteristic function with value 1 on L and 0elsewhere. At all other points, the minimizer can take any value, To force other image locations to take specificlabels (interploation or label propagation property), we require a regularizer which can be given in terms of the

Proc. of SPIE Vol. 9089 908904-3

Page 4: Feature Fusion and Label Propagation for Textured Object Video …cell.missouri.edu/media/publications/Surya-Videoseg-SPIE-2014.pdf · Feature Fusion and Label Propagation for Textured

label functions that are relaxed to the continuous domain u : Ω→ [01],

Ereg(u) =

∫Ω

g(|∇I|)ψ

(n∑i=0

|∇ui|2)dx, (7)

where ψ : R+ → R+ is the regularization function imposing the smoothness of the label field; for example,

ψ(s2) =√s2 + ε2, (8)

the regularized total variation with ε = 10−4 is used in our experiments. The diffusivity function g : Ω→ R+ isused to indicate possible edges (jump discontinuities) and here we set it to the traditional edge indicator functionfrom the image segmentation literature,

g(|∇I|2) =1√

|∇I|2 + ε2. (9)

Thus, we minimize the following energy,

minu∈[0,1]

E(u) =1

2

∫Ω

χL

n∑i=1

(ui − ui)2 dx+

∫Ω

g(|∇I|)ψ

(n∑i=0

|∇ui|2)dx. (10)

We further utilize the multi-scale nature of our segmentation model (2) to propagate labels through scale space.Let u = (u0

1, . . . , u0n, u

11, . . . , u

1n, . . . , u

K1 , . . . , u

Kn ) be the label function across the full scale space hierarchy for n

point trajectories (or objects) and K scales. In practice we use K = 3 levels for multiscale segmentation andwe minimize the following energy which takes into account coupled label propagation between scales (or levelhierarchies),

minu∈[0,1]

E(u) =1

2

∫Ω

χL

n∑i=1

(u0i − ui)2 dx+

K∑k=0

∫Ω

gk ψ

(n∑i=0

∣∣∇uki ∣∣2)dx+

K∑k=0

∫Ω

gkl ψ

(n∑i=0

∣∣uki − uk−1i

∣∣2) dx,

(11)

where gkl (Ik, Ik−1) are multi-scale diffusivity functions similar to the single scale case given above and are definedas

gkl (Ik, Ik−1) =1√∣∣Ik − Ik−1

∣∣2 + ε2. (12)

Here

Ik =1

|Ωkm|

∫Ωk

m

I(x) dx (13)

represents mean color of the obtained segmentations (using scheme (2)) and Ωkm are partitioned regions forthe frame at scale or level k, m = 1, . . . ,Mk. The above functional (11) incorporates multi-scale spatial labelpropagation on a frame by frame basis and are summed across channels (not shown for clarity). Initial sparsenoisy labels are extracted based on a separate clustering process of point trajectories.14 The functional providesa regularization framework for propagating labels both across space and across scale, assuming that the sparseclustered point trajectories over time are available and sufficiently accurate for a good initialization but notnecessarily error-free. So in this functional model there is no explicit regularization over time and the multi-scaleregularization is used to produce spatially dense smoothly interpolated labels. Following the success of dualminimization implementation in other domains7,10 we use it here to implement our label propagation acrossscale space; we refer to the work of Chambolle15 for more details.

Proc. of SPIE Vol. 9089 908904-4

Page 5: Feature Fusion and Label Propagation for Textured Object Video …cell.missouri.edu/media/publications/Surya-Videoseg-SPIE-2014.pdf · Feature Fusion and Label Propagation for Textured

(a) Input (b) Ground truth labels (c) Sparse labels (d) Our scheme result

Figure 2. Synthetic video object segmentation using the variational minimization method in (11). We show a scenariowhere two objects moving with same magnitude but opposing direction, and each object has a different texture than thebackground and occlude each other at the end. From top to bottom we show frames #1, 21, 26, 28, 60. (a) Input frame(b) Ground-truth labels, color coded for better visualization (c) Sparse labels obtained by dropping 90% of the real labelsfrom the objects (d) Segmentation using the globally convex variational minimization model combined with multi-scalelabel propagation and optical flow-based label initialization. When the objects overlap then two different shades of bluecolor are used for clarity. See the online version for the color figures.

Proc. of SPIE Vol. 9089 908904-5

Page 6: Feature Fusion and Label Propagation for Textured Object Video …cell.missouri.edu/media/publications/Surya-Videoseg-SPIE-2014.pdf · Feature Fusion and Label Propagation for Textured

(a) Input (b) Ground truth labels (c) Sparse labels (d) Our scheme result

Figure 3. Synthetic video object segmentation using the variational minimization method in (11). We show a scenariowhere two objects moving towards each other at the same velocity and one of the object has the same texture as thebackground (highlighted by a white boundary in the top row). From top to bottom we show frames #1, 28, 52, 55, 60.(a) Input frame (b) Ground-truth labels, color coded for better visualization (c) Sparse labels obtained by dropping 75%of the real labels from the objects (d) Segmentation using the globally convex variational minimization model combinedwith multi-scale label propagation and optical flow-based label initialization. When the objects overlap then two differentshades of blue color are used for clarity. See the online version for the color figures.

Proc. of SPIE Vol. 9089 908904-6

Page 7: Feature Fusion and Label Propagation for Textured Object Video …cell.missouri.edu/media/publications/Surya-Videoseg-SPIE-2014.pdf · Feature Fusion and Label Propagation for Textured

2.3 Preliminary Experimental Results

To test our variational label propagation method we utilize several synthetic cases depicting different scenarios.Figures 2 and 3 illustrate our scheme along with ground truth labels and segmentations for videos (60 framesof size 600× 1024). The ground truth labels (Figure 2(b) and 3(b) columns), and simulated sparse noisy labels(Figure 2(c) and 3(c) columns) are shown after dropping 90% and 75% of the labels from the foreground objectsrespectively. Figure 2 shows sample frames from the first synthetic video which contains two objects havingdifferent texture patterns than the white Gaussian noise background moving with the same speed in oppositedirections. Figure 3 shows another set of sample frames from the second synthetic video where one of the objects(right side) has the same texture as the background (shown as white boundary in Figure 3(a) top row) which isa harder segmentation problem since motion is necessary to distinguish the moving foreground object from thebackground. In both cases our motion-based clustered point trajectory initialized variational scale space labelpropagation method works well and captures moving objects with improved precision.

3. CONCLUSION

In this paper, we studied a variational feature fusion based scale space label propagation for segmenting texturedmoving objects in multichannel video data. By utilizing the large displacement optical flow clustered pointtrajectories as a bottom-up temporal cue we obtain segmentation based on a texture, intensity and color featurefusion using a globally convex active contour formulation. Synthetic examples are given to illustrate the conceptand initial results are promising for obtaining moving object segmentations from videos and currently we areworking on wide area16,17,23 and full motion videos with people19 and objects.20–22

Acknowledgment

This research was partially supported by U.S. Air Force Research Laboratory (AFRL) under agreement AFRLFA8750-14-2-0072. The views and conclusions contained in this document are those of the authors and should notbe interpreted as representing the official policies, either expressed or implied, of AFRL or the U.S. Government.

REFERENCES

[1] Galasso, F., Nagaraja, N. S., Cardenas, T. J., Brox, T., and Schiele, B., “A unified video segmentationbenchmark: Annotation, metrics and analysis,” in [IEEE International Conference on Computer Vision(ICCV) ], (December 2013).

[2] Chang, J., Wei, D., and Fisher III, J. W., “A video representation using temporal superpixels,” in [IEEEConference on Computer Vision and Pattern Recognition (CVPR) ], 2051–2058 (June 2013).

[3] Lezama, J., Alahari, K., Sivic, J., and Laptev, I., “Track to the future: Spatio-temporal video segmentationwith long-range motion cues,” in [IEEE Conference on Computer Vision and Pattern Recognition (CVPR) ],(June 2011).

[4] Vijayanarasimhan, S. and Grauman, K., “Active frame selection for label propagation in videos,” in [Euro-pean Conference on Computer Vision (ECCV) ], 496–509 (October 2012).

[5] Badrinarayanan, V., Budvytis, I., and Cipolla, R., “Semi-supervised video segmentation using tree struc-tured graphical models,” IEEE Transactions on Pattern Analysis and Machine Intelligence 35(11), 2751–2764 (2013).

[6] Brox, T. and Malik, J., “Large displacement optical flow: Descriptor matching in variational motion esti-mation,” IEEE Transactions on Pattern Analysis and Machine Intelligence 33(3), 500–513 (2011).

[7] Prasath, V. B. S., Palaniappan, K., and Seetharaman, G., “Multichannel texture image segmentation usingweighted feature fitting based variational active contours,” in [Eighth Indian Conference on Vision, Graphicsand Image Processing (ICVGIP) ], (December 2012).

[8] Chan, T. F., Esedoglu, S., and Nikolova, M., “Algorithms for finding global minimizers of image segmentationand denoising models,” SIAM Journal on Applied Mathematics 66(5), 1632–1648 (2006).

[9] Chan, T. F. and Vese, L. A., “Active contours without edges,” IEEE Transactions on Image Process-ing 10(2), 266–277 (2001).

Proc. of SPIE Vol. 9089 908904-7

Page 8: Feature Fusion and Label Propagation for Textured Object Video …cell.missouri.edu/media/publications/Surya-Videoseg-SPIE-2014.pdf · Feature Fusion and Label Propagation for Textured

[10] Moreno, J. C., Prasath, V. B. S., Proenca, H., and Palaniappan, K., “Brain MRI segmentation with fast andglobally convex multiphase active contours,” Computer Vision Image Understanding 125, 237–250 (2014).

[11] Prasath, V. B. S., Moreno, J. C., and Palaniappan, K., “Color image denoising by chromatic edges basedvector valued diffusion,” Tech. Rep. 1304.5587, ArXiv (2013).

[12] Kang, S. H. and March, R., “Variational models for image colorization via chromaticity and brightnessdecomposition,” IEEE Transactions on Image Processing 16(9), 2251–2261 (2007).

[13] Arbelaez, P., Maire, M., Fowlkes, C., and Malik, J., “Contour detection and hierarchical image segmenta-tion,” IEEE Transactions on Pattern Analysis and Machine Intelligence 33(5), 898–916 (2011).

[14] Brox, T. and Malik, J., “Object segmentation by long term analysis of point trajectories,” in [EuropeanConference on Computer Vision (ECCV) ], 282–295 (September 2010).

[15] Chambolle, A., “An algorithm for total variation minimization and applications,” Journal of MathematicalImaging and Vision 20(1–2), 89–97 (2004).

[16] Blasch, E., Seetharaman, G., Palaniappan, K., Ling, H., and Chen, G., “Wide-area motion imagery (WAMI)exploitation tools for enhanced situational awareness,” in [Proc. IEEE Applied Imagery Pattern RecognitionWorkshop ], (2012).

[17] Haridas, A., Pelapur, R., Fraser, J., Bunyak, F., and Palaniappan, K., “Visualization of automated andmanual trajectories in wide-area motion imagery,” in [15th Int. Conf. Information Visualization ], 288–293(2011).

[18] Ersoy, I., Palaniappan, K., Seetharaman, G., and Rao, R., “Interactive tracking for persistent wide-areasurveillance,” in [Proc. SPIE Conf. Geospatial InfoFusion II (Defense, Security and Sensing: Sensor Dataand Information Exploitation) ], 8396 (2012).

[19] Kolekar, M. H., Palaniappan, K., and Sengupta, S., “Semantic event detection and classification in cricketvideo sequence,” in [IEEE Indian Conference on Computer Vision, Graphics and Image Processing ], 382–389(2008).

[20] Pelapur, R., Bunyak, F., Seetharaman, G., and Palaniappan, K., “Vehicle detection and orientation esti-mation using the Radon transform,” in [Proc. SPIE Conf. Geospatial InfoFusion III (Defense, Security andSensing: Sensor Data and Information Exploitation) ], 8747, 87470I (2013).

[21] Fraser, J., Haridas, A., Seetharaman, G., Rao, R., and Palaniappan, K., “Kolam: A cross-platform architec-ture for scalable visualization and tracking in wide-area motion imagery,” in [Proc. SPIE Conf. GeospatialInfoFusion III (Defense, Security and Sensing: Sensor Data and Information Exploitation) ], 8747, 87470N(2013).

[22] Ersoy, I., Palaniappan, K., and Seetharaman, G., “Visual tracking with robust target localization,” in [IEEEInt. Conf. Image Processing ], 1365–1368 (2012).

[23] Peng, J., Palaniappan, K., Candemir, S., and Seetharaman, G., “Kullback-Leibler divergence-based datafusion for target tracking,” in [Int. Conf. Pattern Recognition ], (2012).

Proc. of SPIE Vol. 9089 908904-8