15
Superquadric Segmentation in Range Images via Fusion of Region and Boundary Information Dimitrios Katsoulas, Christian Cea Bastidas, and Dimitrios Kosmopoulos Abstract—The high potential of Superquadrics as modeling elements for image segmentation tasks has been pointed out for years in the computer vision community. In this work, we employ superquadrics as modeling elements for multiple object segmentation in range images. Segmentation is executed in two stages: First, a hypothesis about the values of the segmentation parameters is generated. Second, the hypothesis is refined locally. In both stages, object boundary and region information are considered. Boundary information is derived via model-based edge detection in the input range image. Hypothesis generation uses boundary information to isolate image regions that can be accurately described by superquadrics. Within hypothesis refinement, a game-theoretic framework is used to fuse the two information sources by associating an objective function to each information source. Iterative optimization of the two objective functions in succession, outputs a precise description of all image objects. We demonstrate experimentally that this approach substantially improves the most established method in superquadric segmentation in terms of accuracy and computational efficiency. We demonstrate the applicability of our segmentation framework in real-world applications by constructing a novel robotic system for automatic unloading of jumbled box-like objects from platforms. Index Terms—Range data, shape, size and shape, region growing, partitioning, edge and feature detection, surface fitting, applications. Ç 1 INTRODUCTION I MAGE segmentation is admittedly a difficult problem in machine vision, and despite the promising techniques already proposed, for example, in [1], [2], [3], [4], we are still far from solving it in its generality. Considerable improve- ment on the output of the segmentation task regarding all robustness, accuracy, and computational efficiency can be achieved when prior knowledge related to various proper- ties of the objects in the image is utilized. Such knowledge is incorporated in the segmentation task using modeling entities, or simply models, able to accurately express the object properties. This work contributes to segmentation of three-dimensional objects. A variety of 3D entities have already been proposed for their modeling. Among them are generalized cylinders [5], implicit (fourth degree) polyno- mials [6], geons [7], [8], spherical harmonic and Fourier surfaces [9], [10], symmetry seeking models [11], blob models [12], and hyperquadrics [13]. In this work, we employ Superquadrics [14], [15], [16], [17] as modeling entities. The reasons for selecting the particular models are the following: First, superquadrics comprise a small number of parameters with intuitive meaning, which makes their handling straightforward. Second, they have a large expressive power so that they can model a big number of artificial and natural objects. Third, their expressiveness can be easily further enhanced by the addition of global and local shape deformations. Fourth, closed form expressions exist for the models and for special curves on their surface that makes both rendering and usage fast and direct. Finally, fast and robust methods for superquadric fitting to 3D data have been so far developed. This work describes a new framework for segmenting images that contain multiple piled objects, which are modeled by superquadric entities. Input is a range image of the object configuration. The usage of range imagery is advantageous in our case, where 3D objects are dealt with, because object depth information is incorporated in the image. Using intensity images from multiple cameras is an alternative, but in this case, 3D object information must be reconstructed from intensity images, which is known to be strenuous and of questionable robustness. Our strategy improves the state of the art in this domain by an order of magnitude in speed and reduces the model fitting error to a factor of about 3. In the remainder of this section, we describe how superquadric models have been employed up to now for performing segmentation, we give a rough description of our approach and discuss the contributions of this work. 1.1 Related Work on Superquadric Segmentation Existing approaches attempting superquadric segmentation can be roughly divided in those that decouple model parameter estimation from image segmentation and those that interweave them. In the context of the former category (for example, [18], [19], [20], [21], [22], [23]), an initial segmentation of the image is provided. Then, superquadrics are used for representing the already segmented image regions. In other words, given the segmentation the superquadric fitting problem is pri- marily addressed. The models are elaborated with global and local deformations. Hence, these methods demonstrate superior results as far as object representation is concerned: IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL. 30, NO. 5, MAY 2008 781 . D. Katsoulas and D. Kosmopoulos are with the Institute of Informatics and Telecommunications, Computational Intelligence Laboratory, National Center for Scientific Research Demokritos, GR-15310 Aghia Paraskevi, Athens, Greece. E-mail: {dkats, dkosmo}@iit.demokritos.gr. . C.C. Bastidas is with Entel S.A, Avenida Andres Bello 2687, 7550611 Las Condes, Santiago, Chile. E-mail: [email protected]. Manuscript received 1 May 2006; revised 22 Jan. 2007; accepted 3 May 2007; published online 2 July 2007. Recommended for acceptance by K. Kutulakos. For information on obtaining reprints of this article, please send e-mail to: [email protected], and reference IEEECS Log Number TPAMI-0341-0506. Digital Object Identifier no. 10.1109/TPAMI.2007.70736. 0162-8828/08/$25.00 ß 2008 IEEE Published by the IEEE Computer Society

IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE ... › ~afanasye › 10_UniTN_reports › 20_Presentations › … · 2 4 3 5¼ a 1 cosð Þ 1 cosð!Þ 2 a 2 cosð Þ 1 sinð!Þ

  • Upload
    others

  • View
    0

  • Download
    0

Embed Size (px)

Citation preview

Page 1: IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE ... › ~afanasye › 10_UniTN_reports › 20_Presentations › … · 2 4 3 5¼ a 1 cosð Þ 1 cosð!Þ 2 a 2 cosð Þ 1 sinð!Þ

Superquadric Segmentation in Range Imagesvia Fusion of Region and Boundary Information

Dimitrios Katsoulas, Christian Cea Bastidas, and Dimitrios Kosmopoulos

Abstract—The high potential of Superquadrics as modeling elements for image segmentation tasks has been pointed out for years in

the computer vision community. In this work, we employ superquadrics as modeling elements for multiple object segmentation in range

images. Segmentation is executed in two stages: First, a hypothesis about the values of the segmentation parameters is generated.

Second, the hypothesis is refined locally. In both stages, object boundary and region information are considered. Boundary information

is derived via model-based edge detection in the input range image. Hypothesis generation uses boundary information to isolate image

regions that can be accurately described by superquadrics. Within hypothesis refinement, a game-theoretic framework is used to fuse

the two information sources by associating an objective function to each information source. Iterative optimization of the two objective

functions in succession, outputs a precise description of all image objects. We demonstrate experimentally that this approach

substantially improves the most established method in superquadric segmentation in terms of accuracy and computational efficiency.

We demonstrate the applicability of our segmentation framework in real-world applications by constructing a novel robotic system for

automatic unloading of jumbled box-like objects from platforms.

Index Terms—Range data, shape, size and shape, region growing, partitioning, edge and feature detection, surface fitting,

applications.

Ç

1 INTRODUCTION

IMAGE segmentation is admittedly a difficult problem inmachine vision, and despite the promising techniques

already proposed, for example, in [1], [2], [3], [4], we are stillfar from solving it in its generality. Considerable improve-ment on the output of the segmentation task regarding allrobustness, accuracy, and computational efficiency can beachieved when prior knowledge related to various proper-ties of the objects in the image is utilized. Such knowledge isincorporated in the segmentation task using modelingentities, or simply models, able to accurately express theobject properties. This work contributes to segmentation ofthree-dimensional objects. A variety of 3D entities havealready been proposed for their modeling. Among them aregeneralized cylinders [5], implicit (fourth degree) polyno-mials [6], geons [7], [8], spherical harmonic and Fouriersurfaces [9], [10], symmetry seeking models [11], blobmodels [12], and hyperquadrics [13]. In this work, weemploy Superquadrics [14], [15], [16], [17] as modelingentities. The reasons for selecting the particular modelsare the following: First, superquadrics comprise a smallnumber of parameters with intuitive meaning, which makestheir handling straightforward. Second, they have a largeexpressive power so that they can model a big number ofartificial and natural objects. Third, their expressiveness can

be easily further enhanced by the addition of global andlocal shape deformations. Fourth, closed form expressionsexist for the models and for special curves on their surfacethat makes both rendering and usage fast and direct.Finally, fast and robust methods for superquadric fitting to3D data have been so far developed.

This work describes a new framework for segmentingimages that contain multiple piled objects, which aremodeled by superquadric entities. Input is a range image ofthe object configuration. The usage of range imagery isadvantageous in our case, where 3D objects are dealt with,because object depth information is incorporated in theimage. Using intensity images from multiple cameras is analternative, but in this case, 3D object information must bereconstructed from intensity images, which is known to bestrenuous and of questionable robustness. Our strategyimproves the state of the art in this domain by an order ofmagnitude in speed and reduces the model fitting error to afactor of about 3. In the remainder of this section, we describehow superquadric models have been employed up to now forperforming segmentation, we give a rough description of ourapproach and discuss the contributions of this work.

1.1 Related Work on Superquadric Segmentation

Existing approaches attempting superquadric segmentationcan be roughly divided in those that decouple modelparameter estimation from image segmentation and thosethat interweave them.

In the context of the former category (for example, [18],[19], [20], [21], [22], [23]), an initial segmentation of the imageis provided. Then, superquadrics are used for representingthe already segmented image regions. In other words, giventhe segmentation the superquadric fitting problem is pri-marily addressed. The models are elaborated with global andlocal deformations. Hence, these methods demonstratesuperior results as far as object representation is concerned:

IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL. 30, NO. 5, MAY 2008 781

. D. Katsoulas and D. Kosmopoulos are with the Institute of Informatics andTelecommunications, Computational Intelligence Laboratory, NationalCenter for Scientific Research Demokritos, GR-15310 Aghia Paraskevi,Athens, Greece. E-mail: {dkats, dkosmo}@iit.demokritos.gr.

. C.C. Bastidas is with Entel S.A, Avenida Andres Bello 2687, 7550611 LasCondes, Santiago, Chile. E-mail: [email protected].

Manuscript received 1 May 2006; revised 22 Jan. 2007; accepted 3 May 2007;published online 2 July 2007.Recommended for acceptance by K. Kutulakos.For information on obtaining reprints of this article, please send e-mail to:[email protected], and reference IEEECS Log Number TPAMI-0341-0506.Digital Object Identifier no. 10.1109/TPAMI.2007.70736.

0162-8828/08/$25.00 � 2008 IEEE Published by the IEEE Computer Society

Page 2: IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE ... › ~afanasye › 10_UniTN_reports › 20_Presentations › … · 2 4 3 5¼ a 1 cosð Þ 1 cosð!Þ 2 a 2 cosð Þ 1 sinð!Þ

They are able to represent complicated single and multipartobjects with remarkable accuracy. They are referred to assegment-then-fit methods.

Approaches belonging to the latter category, for example,[24], [25], [26], [27], [28], [29], [16], [17], [30] address both thesegmentation and model fitting problem. Because the modelis as well used as a means for segmentation, the ability of thesemethods to represent complicated objects is reduced: theyemploy undeformed superquadrics or at the most super-quadrics enhanced with simple global deformations forobject modeling. Their advantage however, is robustness[17, p.105]. These approaches are referred to as segment-and-fitmethods. They belong to the generic category of model-basedimage segmentation methods using probabilistic techniques:Assuming that the parameters of all models in the inputimage as well as their number comprise the parameter set ofthe segmentation task, they estimate the parameters bymaximizing their posterior probability in the image. Note thatthis is the optimal way to perform segmentation, since itguarantees minimization of the average risk of failing to selectthe actual segmentation parameters [31].

Due to the difficulties in directly optimizing the posterior[32], the segmentation task is usually broken down into aparameter initialization stage and a local optimization stage.The parameter initialization stage is regarded as generation ofa hypothesis about the values of the segmentation parameters.The optimization stage refines the hypothesis by locallyoptimizing the posterior distribution. This is the reason whythis two-stage segmentation framework will be hereinafterreferred to as the hypothesis generation and refinement strategy.

The state of the art for superquadric segmentation, theRecover-and-Select (RAS) paradigm, [33], [16], [17] adheres tothe hypothesis generation and refinement strategy: In thehypothesis generation stage, the segmentation parametersare initialized. The hypothesis refinement stage updates theparameters by means of local optimization. More specifically,the hypothesis generation stage involves a grid-like place-ment of small models in the image. The hypothesis-refine-ment stage updates the parameters of the initialized models,so that they more accurately express the corresponding dataon one hand and rejects models judged to inaccuratelyexpress the data despite parameter updating on the other.RAS is more robust and efficient than other approaches. Itsmain drawback, however, is the misclassification of imagepoints in the neighborhood of object boundaries because thesegmentation process uses the depth values of sets of imagepixels only, that is, region information, for performingsegmentation. Our framework outperforms RAS by fusingboundary and region information within segmentation.

1.2 Our Approach-Contributions

Our approach comprises two processing stages: hypothesisgeneration and refinement. In both stages, in addition toregion information, object boundary information is incorpo-rated, which is acquired via edge detection in the inputrange image. In the hypothesis generation stage, boundaryinformation is used as a guide to the initialization of thesegmentation parameters. In hypothesis refinement, a gametheoretic framework is employed to fuse the informationsources in a balanced way.

Our approach considerably outperforms the state of theart in the segmentation of superquadrics in terms of speedand accuracy. In addition, we show how a game theoreticframework can be used for segmenting multiple objects in

range images. The key to the success of our approach lies inthree aspects: 1) the consideration of model informationfrom the very early steps of all algorithm stages, 2) theadjustment of the data provided by each source in order toexplicitly represent domain specific information, and 3) theadjustment of the model to represent more accurately thedata provided by each information source.

Superquadric segmentation could be applied in a varietyof occasions. Representative examples are object categoriza-tion and recognition, 3D image representation, imageregistration and merging, next viewpoint planning guidance,image compression, and others [17]. In this work, oursegmentation framework is applied in bin picking. Wedeveloped a novel robotic system, which automaticallyunloads piled box-like objects from platforms in real time.This system combines advantages like flexibility, computa-tional efficiency, accuracy, and robustness, the combinationof which cannot be found in existing applications.

1.3 Paper Structure

The remainder of this paper is organized as follows: InSection 2, we present some background information relatedto our modeling elements, which is used by the subsequentsections. In Section 3, we focus on RAS and point out itsproblems. In Section 4, we present our segmentationapproach in detail: we present the components of ourapproach: object boundary information acquisition, the wayin which a hypothesis about the values of the segmentationparameters is initialized, and how the hypothesis is refinedby fusing information sources. In Section 5, we evaluate ourframework experimentally, and we demonstrate its super-iority through its quantitative comparison to the RASframework. In Section 6, we prove the applicability of ourframework in an application of high commercial interest,the robotic unloading of piled objects from platforms.Finally, Section 7 concludes this work.

2 SUPERQUADRICS AND SOME GEOMETRIC

PROPERTIES

The term Superquadrics, was first used in [14] to define afamily of shapes that includes superellipsoids, superhyperbo-loids of one sheet, and superhyperboloids of two sheets, aswell as supertoroids. In the computer vision literature,however, the term superquadrics is used to refer tosuperellipsoid objects. In this work, we will also use theterm superquadrics as a synonym for superellipsoids.

The explicit form of a superquadric model delivers3D points on the model surface and is given by the followingexpression:

Spð�; !Þ ¼xðp; �; !Þyðp; �; !Þzðp; �; !Þ

24

35 ¼ a1 cosð�Þ�1 cosð!Þ�2

a2 cosð�Þ�1 sinð!Þ�2a3 sinð�Þ�1

24

35: ð1Þ

In the preceding equation, � and ! lie in the range½��=2; �=2� and ½��; ��, respectively. The vector p ¼ða1; a2; a3; �1; �2Þ incorporates the superquadric parameters.More specifically, �1 and �2 control the shape of the model, anda1, a2, and a3 express the size of the model along the X, Y, andZ axes of the model coordinate system, respectively. Fig. 1ashows various superquadric models, which are renderedusing (1). Each of these models corresponds to a differentvalue of the shape parameters, within the range [0.1, 2].

782 IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL. 30, NO. 5, MAY 2008

Page 3: IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE ... › ~afanasye › 10_UniTN_reports › 20_Presentations › … · 2 4 3 5¼ a 1 cosð Þ 1 cosð!Þ 2 a 2 cosð Þ 1 sinð!Þ

The implicit superquadric equation can be derived from

(1) by simple algebraic manipulations:

x

a1

� � 2�2

þ y

a2

� � 2�2

!�2�1

þ z

a3

� � 2�1

¼ 1: ð2Þ

Further, the so-called inside-outside function of a super-

quadric model equals the left part of (2)

F ðp; Xðx; y; zÞÞ ¼ x

a1

� � 2�2

þ y

a2

� � 2�2

!�2�1

þ z

a3

� � 2�1

: ð3Þ

In order to express the model translation and rotation,

six additional pose parameters are set in the parameter

vector, which now looks like the following:

p ¼ ða1; a2; a3; �1; �2; px; py; pz; �; �; Þ: ð4Þ

We adopt the standard approach in [15] to fit a model to

the set of 3D points Xi, i ¼ 1 . . .N , according to which the

function Lr, shown in (5), which is proportional to the

likelihood of the data given the parameter vector and

expresses the sum of squares of algebraic distances of the

points to the model, is maximized

LrðpÞ ¼ �Xni¼1

ffiffiffiffiffiffiffiffiffiffiffiffiffiffia1a2a3p ðF�1ðp; XiÞ � 1Þð Þ2; ð5Þ

where the function F is the inside-outside function shown

in (3).In this work, we need to model the boundary of a

Superquadric model. The ideal modeling element for this

purpose is the superquadric RIM, which is defined as the

3D curve on the model surface, which separates its visible

from its invisible part. Fig. 1b shows a model and its RIM

for a particular viewpoint. A detailed discussion about the

definition of the RIM and its properties is presented in [34].

In general, given that the Z axis (or depth axis) of the sensor

coordinate system is the viewing point, 3D points on the

RIM of the model with parameters p are derived as follows:

Rpð!Þ ¼ Spð�ð!Þ; !Þ; ð6Þ

where Sp are points on the superquadric surface obtained

via (2), ! ranges in ½��; ��, and �ð!Þ amounts to

�ð!Þ ¼ arctan � a3

ay

nya1

cosð!Þ2��2 þ oza2

sinð!Þ2��2� � 1

2��1

( ): ð7Þ

In the preceding equation, ay, ny, and oz are elements of

the rotation matrix that depends on ð�; �; Þ (see (4)) and

brings the model to the scene.Note that a uniform (equal-distance) sampling of ! in the

range ½��; �� does not produce uniformly sampled points

along the RIM of a model. As we will see in Section 4.3.1,

uniform point sampling along the RIM is required for

accurate boundary recovery. Our method for RIM sampling

is based on that in [35]: Equal-distance sampling of points

on the RIM is produced using spherical products of proper

equal-distance sampled superellipses.

3 RECOVER-AND-SELECT FOR SUPERQUADRIC

SEGMENTATION

The Recover-and-Select (RAS) paradigm performs the

segmentation task in two stages: First, a rough approxima-

tion, a hypothesis, about the number and the parameters of

the models in the input range image is generated. Second,

the hypothesis is locally refined using image information.

3.1 Hypothesis Generation

This involves an initial placement of the models in the

image, by generating a grid-like subdivision of the input

range image in numerous uniform rectangular windows,

each containing a few range points. The set of points

contained in each window is assumed to be a region of

points corresponding to a superquadric model. Subse-

quently, the model parameters are determined by fitting a

model to each point region. Models with large fitting errors

probably cross object boundaries and thus are rejected from

further consideration. The term seeds is used to refer to the

remaining models, to express both their limited expressive

ability and their small size. Hence, the term seed generation

or seed placement alternatively stands for the hypothesis

generation stage. The Fig. 2c depicts seed placement using

RAS given the range image in Fig. 2b, which was acquired

by a laser sensor from the object configuration in Fig. 2a.

KATSOULAS ET AL.: SUPERQUADRIC SEGMENTATION IN RANGE IMAGES VIA FUSION OF REGION AND BOUNDARY INFORMATION 783

Fig. 1. Properties of Superquadric models. (a) Superquadrics of variousshapes. (b) Given a viewpoint, the continuous black curve shown (RIM)separates the visible from the invisible part of the model.

Page 4: IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE ... › ~afanasye › 10_UniTN_reports › 20_Presentations › … · 2 4 3 5¼ a 1 cosð Þ 1 cosð!Þ 2 a 2 cosð Þ 1 sinð!Þ

3.2 Hypothesis Refinement

The hypothesis refinement stage comprises 1) seed region

growing and 2) model selection.

3.2.1 Region Growing

Due to model rejection by the filtering step of the seedgeneration procedure, image points exist which are notassigned to models. The iterative region growing [36] aims toclassify these points to existing regions and to refine themodel parameters so that the newly classified points aretaken into consideration. Within the region growing process,model parameter estimation through fitting is interwovenwith image point classification to models and, therefore, in[16], [17], it is referred to as the classify-and-fit process.

3.2.2 Model Selection

This process selects the models which represent the under-lying points accurately. The initialization of many seeds andtheir independent growing may cause representation ofpoints by multiple models, which is manifested as modeloverlapping in the image. Model selection rejects redundantmodels by applying the minimum description length (MDL)principle: the model configuration in which the fewestpossible models describe the largest number of imagepoints with low deviation is preferred. This is realized via a

greedy optimization algorithm, which as such is fast butmay deliver suboptimal solutions.

The hypothesis refinement is performed by iterativelyinvoking 1) followed by 2), until a satisfactory data repre-sentation is achieved. Invoking model selection after all seedsare fully grown (recover-then-select approach [17]) is ineffi-cient, owing to the big number of initialized models.Computational costs are reduced by interweaving 1) and 2).Early invocation of model selection may lead to unreliabledata representations, and therefore, in [17], computationalefficiency and reliability are balanced by interrupting theregion growing procedure when at least one growing modelreaches twice its original size or when there are no models leftto grow.

3.3 Discussion

The RAS paradigm is the state of the art in multiplesuperquadric segmentation. The iterative region growingprocess is robust against gross measurement errors, butpoints belonging to neighboring objects are often erroneouslyincluded in the model’s region: Due to its inherent locality,the region growing process has no mechanism to detect thatthe model has reached the boundaries of the object itrepresents (overgrowing). This problem is implicitly addressedwithin RAS by placing numerous seeds in the image and byinvoking the model selection procedure after region growingto get rid of the seeds with large deviation from the data,which probably cross object boundaries.

We have extensively evaluated the RAS paradigm, andwe present here some qualitative results by applying theframework to the range image in Fig. 2b. Fig. 2d shows thefinal result. The region overgrowing problem persists,despite the invocations of the model selection process: therecovered models corresponding to the objects A and E inFig. 2b are overgrown: the greedy model selection processdoes not possess the capability to optimally arrange themodels in a piecewise description. Hence, in many cases,either overgrown models are favored, or an overgrownmodel describing a superset of a data set is selected inaddition to the model optimally describing the particulardata set as is for the object C.

The execution time required for running RAS was about15 minutes in a Pentium 4 PC with a 1.6 Ghz processor:initialization of numerous seeds as a means for addressingthe region overgrowing problem, in combination with thefrequent invocations of the model selection process, rendersthe framework inappropriate for real-time implementation.

To conclude, model overgrowing is the primary cause ofproblems of the RAS strategy. Our approach explicitlyaddresses region overgrowing by incorporating objectboundary information in the segmentation process.

4 INFORMATION FUSION FOR RECOVERING

MULTIPLE SUPERQUADRICS

The mechanism guiding the segmentation task in RAS is theregion growing process, which is perhaps the most famousrepresentative of the so-called region-based approaches forimage segmentation. Fusion of region and boundary-basedapproaches are known to deliver superior segmentationresults ([37 Chapter 5, Section 2]).

Although region-based approaches use the input rangeimage I directly for deriving region information, boundary-based approaches use a boundary image Ib for deriving object

784 IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL. 30, NO. 5, MAY 2008

Fig. 2. A configuration of piled bags used for qualitatively comparing ourapproach with RAS. (a) Intensity image. (b) Range image with objectnumbering. (c) Grid-like seed placement. (d) Output of RAS. (e) Edgeimage. (f) Edge map. (g) Image regions corresponding to the seeds,placed on the edge map. (h) Generated seeds, superimposed to therange image (i) Seed boundaries on the edge map. (j) Recoveredboundaries on the edge map. (k) Recovered models on the range image.

Page 5: IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE ... › ~afanasye › 10_UniTN_reports › 20_Presentations › … · 2 4 3 5¼ a 1 cosð Þ 1 cosð!Þ 2 a 2 cosð Þ 1 sinð!Þ

boundary information. This is usually obtained by applyingan edge detector on I. The problem that arises is how tointegrate the information residing in both I and Ib in asegmentation framework. In general, there is a different set ofsegmentation parameters related to each of the informationsources. Given pr and pb, the segmentation parametersrelated to I and Ib, respectively, segmentation of each sourceseparately within a probabilistic framework for imagesegmentation involves the computation of the likelihoodsP ðIjprÞ and P ðIbjpbÞ [37]. Equal consideration of bothinformation sources requires the computation of the jointlikelihood P ðI; Ibjpr;pbÞ. Since the parameter spaces relatedto each source are not related, this is not an easy task. For thisreason, the majority of methods attempting to integrate regionand boundary-based information [2], [4], [38] form a giantobjective function by constructing a weighted sum of the twolikelihood functions. Subsequently, they simultaneouslyestimate both region and boundary-related parameters bymaximizingthis function. In this way,computationof the jointlikelihood is avoided. Such techniques, named single objectivetechniques have the problem of inadequate analytic andcomputational tractability. Furthermore, they are of ques-tionable robustness, as shown in [37], [39], [40], [41].

The flow diagram of our approach is illustrated in Fig. 3and comprises two stages: First, edge detection is per-formed in the input range image I by means of the scan lineapproximation technique [42]. Output of the edge detectionis the edge map Ib. Second, the main part of thesegmentation is initiated, bounded by the dashed box inFig. 3. Both I and Ib are employed by this stage. Output ofthe hypothesis generation process is a set of superquadricmodels in the range image. If P denotes a vector containingthe parameters of all these models, the hypothesis refine-ment stage updates P to obtain the set of models P�, thefinal output of our segmentation process.

In hypothesis refinement, boundary information is inte-grated in a way inspired by the game theoretic framework in[40], which is here adapted to processing range images:parameter refinement is realized by means of iterativeinvocations of two independent parametric modules insuccession. Each module is related to an information source.A benefit function is associated to each module: the regionmodule fits superquadric models to image regions assumedthey belong to unique objects and associates range pointswith objects. The boundary module fits the RIM of thesuperquadric models to the edge points in the boundaryimage. Hence, each module acts on its own informationdomain, or in other words, dealing with boundary-relatedinformation is decoupled from region information, and viceversa. Information integration is achieved by simply letting

each module pass its corresponding parameters to the other,after its execution. The hypothesis refinement process endswhen no significant change in the model parameters can beperformed.

Performing integration of region and boundary-basedinformation within a game-theoretic framework shows avariety of advantages: First, both information sources areequally taken into consideration by the segmentation process,without computing the joint likelihood function. Second, thebenefit of adopting a decoupled scheme for informationintegration results to computational efficiency, since eachmodule deals with the subset of the entire parameter set,which is related to the corresponding information domain.Finally, robustness is achieved even when the quality of oneof the sources is not high, since each module tends to pull theother away from noise and local maxima on one hand whilepushing itself toward the global optimum on the other.

Our approach is described in detail in the paragraphsthat follow: We first show how boundary detection isperformed. Then, we discuss the hypothesis generation andrefinement stages. Finally, we point out the innovations ofour strategy.

4.1 Edge Detection

We adopt a model-based approach for edge detection in theinput range images, which is inspired by Jiang and Bunke[42]. Edge detection is performed via approximation of therows and the columns of the image with 2D geometricparametric models. In its original form, the techniqueinvolves the following steps: First, an input scan line issplit into smaller segments, which can be approximated bythe modeling elements with a low deviation. A top-downrecursive procedure is employed for this purpose [43].Candidate edge points are the end points of the neighboringsegments produced by the splitting process. The parametersof two neighboring segments are used to compute the edgestrength of each candidate edge point, which is a propor-tional to the probability of a candidate edge point being areal edge point. Since two kinds of edges can be found inrange images, jump edges (discontinuities in depth) andcrease edges (discontinuities in normals), two strengthvalues are associated to each candidate edge point. Thereal edge points are the candidates for which one of the twostrength values exceeds a user-defined threshold.

We have selected this approach because it exhibits manyadvantages: computational efficiency, which is the outcomeof the fast scan line splitting process. Robustness andaccuracy in the localization of the edge points, which is theoutcome of incorporating global shape information in theprocess: the classification of a range point as an edge pointdoes not depend on local information, as is the case for themainstream edge detection approaches for both intensity andrange images but on the parameters of neighboring approx-imating models the estimation of which is influenced by a bignumber of range points. We introduced some additionaloperations to further increase the robustness in [42]: A model-based segment-merging step after the splitting operation, aswell as a procedure for fine localization of the edge points. In[44], we show how this operation considerably improves theoutput of that in [42] in our data set. Our detector employs thetwo-dimensional (2D) analog of the superquadrics, thesuperellipses, for approximating the scan lines of the imagefor our purposes.

KATSOULAS ET AL.: SUPERQUADRIC SEGMENTATION IN RANGE IMAGES VIA FUSION OF REGION AND BOUNDARY INFORMATION 785

Fig. 3. Our approach for multiple superquadric recovery.

Page 6: IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE ... › ~afanasye › 10_UniTN_reports › 20_Presentations › … · 2 4 3 5¼ a 1 cosð Þ 1 cosð!Þ 2 a 2 cosð Þ 1 sinð!Þ

The output of the edge detection operation for the rangeimage in Fig. 2b is the edge image, which comprises 3D edgepoints. This image, superimposed on the range imageFig. 2b, is illustrated in Fig. 2e. Fig. 2f illustrates the edgemap Ib. The edge map is a (2D) binary image, a pixel’s valuex ¼ ðx; yÞ of which is 1, that is, IbðxÞ ¼ 1, only if thecorresponding range point IðxÞ ¼ ðx; y; zÞ is an edge point.In short, the edge map incorporates the orthogonal projec-tions of the boundaries of the objects’ visible surfaces on theimage plane. Thus, the difference in the informationcontained in the edge image and the edge map is that theformer incorporates the depth of the edge points in addition.Since depth information can be derived by the range image,depth information in the edge image is redundant. There-fore, we employ the edge map instead of the edge image forderiving boundary information. In this way, we end updealing with 2D instead of 3D boundary information, which,as we will see, considerably increases the computationalefficiency of the subsequent boundary-related operationswithout hampering their robustness.

4.2 Hypothesis Generation

Target of the hypothesis generation (or seed placement) stageis an initial estimation of the segmentation parameters. Themost widespread strategies for seed placement [45], [46], [47],[48] determine the sought parameters by model fitting toimage regions corresponding to the objects’ interior. Fordetermining these regions, the edge map is employed. Morespecifically, a distance transform is applied to the edge map.The contours in the transformed edge map along which thedistance to the objects’ edges has a constant value (isodistancecontours), enclose image areas in the interior of the objects bydefinition. The main problem that these approaches face ishow to robustly determine the appropriate distance value.This is not easy to do since the objects have differentdimensions, and the edge map may be noisy.

Our approach for seed placement is inspired by Jiang[49], where the regions inside the objects are determined byadaptive closure of the edge map. The strategy in [49] is basedon the observation that any contour in the edge map can beclosed by dilating the edge map. Contour closure isperformed iteratively and involves connected componentlabeling, model fitting to the connected regions, andmorphological dilation of the regions’ boundaries for whicha large fitting error occurred. In [37], we show that usingthis approach, the boundary of each seed region corre-sponds to the isodistance contour of the minimum distancefrom the boundaries in the edge map, whose interior can besatisfactorily modeled by a superquadric entity. Instead ofusing an isodistance contour threshold for determining seedregions, we use a threshold for the maximum acceptablefitting error that is easier to define since it depends onsensor noise and not on the object dimensions in the image.

The seeds created by the application of the adaptivecontour closure approach in the edge map in Fig. 2f arepresented in Figs. 2g and 2h. More specifically, Fig. 2g, showsthe region map of the seeds. Fig. 2h shows the superquadricsfitted to the range points corresponding to each region in themap. A number of conclusions can be drawn by inspectingthese images: Each region corresponds to a different object,and the parameters of the initialized models are sufficientlyclose to the actual object parameters. Furthermore, theboundaries of the regions are very close to the actual object

contours, and their shape is similar to the latter. Our seedplacement procedure manages to produce an accurateinitialization of the segmentation parameters for the parti-cular image.

4.3 Hypothesis Refinement

The hypothesis refinement stage acts on each seedindependently. Inputs are the seed parameter vector prelated to a seed, as well as the region of points associatedto the seed. The seed corresponds to an object in the image,which will be hereinafter referred to as the object of interestof the process. The goal is to accurately segment the objectof interest. This is done iteratively, as shown in the flowdiagram in Fig. 4a.

The process first initializes a 2D image, which partlyincorporates region information by identifying pixels corre-sponding to the object of interest. This image will behereinafter referred to as the region image and will be denotedas Ir. Its dimensions are equal to those of the range image. Apixel of this image has the value 1, if the corresponding pixelof the range image belongs to the object of interest. Initially,the pixels of Ir set to 1 are the pixels comprising the seed’sregion. Subsequently, the core of the process starts. Eachiteration, involves a sequential invocation of two modules:1) The boundary module employs the superquadric RIM tomodel object boundary information. This is allowable, sinceobject boundaries separate the visible from the invisible partof the object, just like the RIM does to the model describing theobject, given a particular viewpoint. The module updates theseed parameter vector by fitting the RIM of the model to theedge map Ib. 2) The region module updates the superquadricparameters by fitting the model to the range image I. Inaddition, it updates the region image Ir. Each modulerecovers the model parameters by interweaving point classi-fication with model parameter estimation. In other words,each of the modules acts as the standard classify-and-fitprocess (see 3) in its own information domain. Further, theoutput of each module is influenced by the output of the other:The output of the boundary module, the fitted RIM, is affectedby the output of the region module, that is, the region image,and vice versa.

786 IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL. 30, NO. 5, MAY 2008

Fig. 4. The iterative hypothesis refinement process, where (b) analyses

the dashed process in (a). (a) Hypothesis refinement. (b) The invocation

of the boundary and the region module in the kth iteration.

Page 7: IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE ... › ~afanasye › 10_UniTN_reports › 20_Presentations › … · 2 4 3 5¼ a 1 cosð Þ 1 cosð!Þ 2 a 2 cosð Þ 1 sinð!Þ

The process stops after its kth iteration, when twoconditions hold simultaneously: The difference between thenewly obtained value of the parameter vector pkþ1 to thevector obtained by the previous iteration pk is smaller than auser-defined threshold �, and the new pixels that the currentiteration assigns to the object of interest, or the difference innumber of pixels set to 1 between the new region image Irkþ1

and the one obtained by the previous iteration Irk is smallerthan a user defined threshold �. In this case, the current modelvector is delivered as the output p� of the process.

After parameter refinement of all seeds in the image hasbeen performed, a postprocessing step is invoked to rejectobjects with large fitting errors to the models, that is,objects that cannot be satisfactorily modeled by super-quadric entities. In the following paragraphs, we come upwith a detailed description of the two modules of theclassify-and-fit process.

4.3.1 Boundary Module: Boundary Finding Influenced by

Region Information

The boundary module refines the model parameter vector ofthe object of interest using 2D boundary information. Assumethat the boundary module is about to be invoked for the kthtime. Inputs of the module are the edge map Ib, and theparameter vector pk, which is the output of the ðk� 1Þthinvocation of the region module and the region image Irk.Output of the kth invocation of the module is the modelparameter vector pkþ (see Fig. 4b).

The refinement of p is implemented via maximization ofits posterior probability, given both Ib and Ir. The way thisis implemented is inspired by Chakraborty et al. [50] andStaib and Duncan [10]. This is described as

pkþ ¼ arg maxpk

Lðpk; Ib; IrkÞ

¼ arg maxpk

�1Lbðpk; IbÞ þ �2Lrðpk; IrkÞ½ �;ð8Þ

where the user defined constants �1 and �2 express therelative significance of the terms Lb and Lr, respectively.

To better illustrate the following, we use the term eRIMto refer to the orthogonal projection of the RIM of theevolving model to the image plane. To clarify, assumingðx; y; zÞ is a 3D point on the RIM obtained by (6) and (7),ðx; yÞ will be the corresponding point of the eRIM on theimage plane. Now, the term Lb of the benefit function L isproportional to the log likelihood of the edge map:maximization of Lb results into fitting the model’s eRIMto the edge pixels of the object of interest in the edge map.Lb amounts to

Lbðpk; IbÞ ¼ logP ðIbjpkÞ

/ �XMi¼1

Ibdðxðpk;!iÞ; yðpk;!iÞÞ2:ð9Þ

In the preceding expression, Ibd is the euclidean distance-transformed image of the edge map, x, y, the image planecoordinates of M 3D points ðx; y; zÞ on the RIM of the modelbeing refined, and !i ¼ 2i�

M , i ¼ 0 . . .M � 1. Note that theneed for a uniform RIM sampling (discussed in Section 2)emerges here: the more uniform the sampling is, the moreaccurately the likelihood function will express the similarityof the model’s eRIM to the object’s eRIM in the edge map,and the more accurate the fitting operation will be.

We now denote as rRIM the boundary of the region inthe region image Irk. If so, we define the second term Lr ofthe benefit function L to be proportional to the loglikelihood of the region image. This is maximized whenfitting the model’s eRIM to rRIM. Its incorporation in thebenefit function integrates region information within theboundary module. In short,

Lrðpk; IrkÞ ¼ logP ðIrkjpkÞ

/ �XMi¼1

Irkbdðxðpk;!iÞ; yðpk;!iÞÞ2:

ð10Þ

In the preceding equation, Irkbd is the euclidean distance-transformed image of the binary image containing the rRIM.

In short, the process shown in (8) fits the eRIM to the edgemap and to the boundary of the region in Ir. Note that addingthe term accounting for region information into L does notrender our approach similar to the single-objective techni-ques discussed in the beginning of this section. The differenceis that both terms involved here deal with 2D-boundaryinformation.

4.3.2 Region Module: Region-Growing Influenced by

Boundary Information

The region module refines the model parameter vectorcorresponding to the object of interest, using 3D rangeinformation. The kth invocation of the region module,succeeds the kth invocation of the boundary module. At thistime point, inputs of the module are the range image I, theregion image Irk obtained in the previous iteration, and theoutput of the kth invocation of the boundary module, that is,the vector pkþ. Outputs of the module are the pkþ1 and Irkþ1

(see Fig. 4b).The region module is realized by means of the standard

region growing process [36]. It performs parameter refine-ment by the iterative succession of a model-fitting step,followed by a pixel classification step. In the model-fittingstep, it is assumed that the image coordinates of range pointscorresponding to the object of interest will be enclosed by theeRIM of the evolving model. In this way, boundary informa-tion is integrated in the region module. If xb is the set of theenclosed pixels, then this step fits the model to all 3D pointswith image coordinates in xb. This is equivalent to an imagelikelihood maximization and is expressed by the followingexpression (see (5)):

pkþ1 ¼ arg maxpkþ

logP ðIjpkþÞ½ �

¼ arg minpkþ

Xx2xb

ffiffiffiffiffiffiffiffiffiffiffiffiffiffia1a2a3p ðF�1ðpkþ; IðxÞÞ � 1Þ2: ð11Þ

The superquadric model with parameters pkþ1 generatedby the fitting step of the region module, and the regionimage Irk are used to generate the updated region image Irkþ1

within the context of the classification step of the regionmodule. More specifically, the pixel classification stepexamines the set of pixels in the neighborhood of the rRIM.The pixels of this set that have a small radial euclideandistance from the model pkþ1 are added in the region of Irk toobtain Irkþ1.

4.4 Discussion

The game-theoretic framework for information integrationwas proposed in [51], and applied to segmentation of

KATSOULAS ET AL.: SUPERQUADRIC SEGMENTATION IN RANGE IMAGES VIA FUSION OF REGION AND BOUNDARY INFORMATION 787

Page 8: IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE ... › ~afanasye › 10_UniTN_reports › 20_Presentations › … · 2 4 3 5¼ a 1 cosð Þ 1 cosð!Þ 2 a 2 cosð Þ 1 sinð!Þ

intensity images in [40], containing a couple of nonoccludedobjects. The contribution of the approach presented here liesin the application of the framework for range imagescontaining multiple piled objects using 3D entities asmodels. This advances significantly the state of the art, aswe will show in Section 5.

More specifically, the novelty of our system lies in theexploitation of decoupling in model recovery, offered by thegame-theoretic framework: 1) Although the target of bothdomains, boundary and region, involved in the recovery isthe estimation of the model’s parameter vector, the model isused only in the region domain for data representation.Using the same model in the boundary domain wouldcomplicate the fitting process since the constraint that datacorrespond to model boundaries should be encoded. Using acurve on the model’s surface (the model’s RIM) for fitting inboundary domain is advantageous since it correspondsexactly to the boundary data and depends exactly on thesame parameters as the superquadric model. 2) Since depthinformation is employed in the region domain for modelrecovery, it does not have again to be taken into considera-tion in the boundary domain calculations. For this reasonfitting in the boundary domain is performed in twodimensions, which considerably accelerates all operationswithout accuracy losses. 3) In cases where additionalknowledge about the shape of the target objects is available(for example, when the objects are box-like), decoupledhandling of boundary and region domains leads into adecomposition in the parameter recovery. This increases theefficiency of our strategy even more. This approach wasapplied in our robotic system for object unloading withexcellent results (see Section 6).

5 EXPERIMENTS

We present three sets of experiments to demonstrate theefficiency of our approach. We used images acquired fromthree different range sensors for this purpose. The goal of thefirst set is to point out the problems of using region andboundary information sources for object segmentationseparately and to show the superiority of the result whenfusing them. We used a commercial range finder, the SICKLMS-200 [52], which is a laser radar based on the time-of-flight measurement technique for acquiring images for thispurpose. The acquired images are relatively noisy: Theadvantages gained by fusing multiple information sourcesare more clearly illustrated when noisy images are dealt with.The second set compares our approach with the RASframework and demonstrates the improvements it intro-duces with respect to computational efficiency and accuracy.We experimented with images acquired using the prototypestructured light range scanner [53], [17]. The same imageswere used in [17] for assessing the RAS framework. Finally,we performed a third set of experiments in order to illustratethe behavior of our approach when segmenting multipartobjects. The images used here were taken from the popularOSU range image database, and the MSU’s Technical Arts100X range scanner was used for their acquisition. A Pentium4 1.6 GHz PC was employed in all experiments.

5.1 Robustness of Information Fusion

We demonstrate the advantages gained when fusing multiplesources of information in image segmentation using the

object configuration shown in Fig. 2a. In particular, we dealtwith recovering one object in the configuration, starting fromthe small seed shown in Figs. 5a and 5b. More specifically,Fig. 5a depicts the seed superimposed on the range image.Fig. 5b shows the edge map along with the eRIM of the seed,as well as the region associated to the seed.

We first tried to recover the object using a region-basedonly approach, namely, the standard region growing process.The result is presented in Figs. 5c and 5d. The former figuredepicts the recovered model in three dimensions, whereas thelatter shows the region of points associated to the model,superimposed on the edge map. Note that the seed has beenplaced inside the object of interest and in a relatively bigdistance from its boundaries. Seed placement in this positionincreases the robustness of the region growing approach [17,p. 157]. Despite this, the result of region segmentation isdiscouraging: Pixels belonging to neighboring points havebeen erroneously included in the region of the evolvingmodel. The model is overgrown. The process needed30 iterations to produce the result shown.

We then attempted object recovery using our approach.In the first case, model recovery is driven by boundaryinformation only. Hence, we used �1 ¼ 1, and �2 ¼ 0 in thebenefit function of the boundary module (see (8)). The resultof boundary-guided recovery is illustrated in Figs. 5e and5f. The problem of using boundary information only forobject segmentation is that noisy edge points on the

788 IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL. 30, NO. 5, MAY 2008

Fig. 5. Recovery results under various fusion modes, as illustrated inSection 5.1. (a) and (b) Seed on a 3D and image plane. (c) and (d) Region-based recovery result. (e) and (f) Boundary-based recovery result. (g) and(h) Result with fusion of region-boundary information sources.

Page 9: IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE ... › ~afanasye › 10_UniTN_reports › 20_Presentations › … · 2 4 3 5¼ a 1 cosð Þ 1 cosð!Þ 2 a 2 cosð Þ 1 sinð!Þ

exposed surface of the object of interest attract the eRIMcurve, so that the model does not grow as it should. Hence,in comparison to the region growing-based segmentation,the boundary-based segmentation suffers from the inverseproblem: model undergrowing. Note that, in this case, thesystem executed five iterations in order to converge.

Finally, we considered that both boundary and region-based information guide the recovery of the object of interest,that is, we set �1 ¼ �2 ¼ 0:5 in (8). The result is illustrated inFigs. 5g and 5h of superior accuracy compared to the previousones, and this is due to the combination of a region force (termLr in (8)), which is responsible for continued model growingalong noisy edge points with the boundary force whichprohibits object growing beyond real boundaries. In this case,the system converged after 12 iterations. The recovery resultsfor all objects in Fig. 2a is shown in Figs. 2j and 2k. There is aconsiderable improvement in terms of robustness andaccuracy with respect to the results delivered by RAS(Fig. 2d). In general, the values of the weights �1 and �2

depend on a priori knowledge on the relative quality of theboundary and region information, respectively. High-qualityboundary information suggests a big value for �1. Similarly,high-quality region information suggests a big value for�2. Inthe experiments with all our data sets, �1 ¼ �2 ¼ 0:5produced satisfactory results.

5.2 Comparison to RAS

We tested our approach on the same images on which RASwas evaluated in [17]. The results are shown in Figs. 6a, 6b,6c, 6d, 6e, 6f, 6g, 6h, and 6i, and Figs. 7a, 7b, 7c, 7d, 7e, 7f, 7g,7h, 7i, 7j, 7k, 7l, 7m, 7n, and 7o.

We performed a comparison of our approach with RASin terms of accuracy. This comparison is based on a ground-truth segmentation of the images: For each range image, wemanually segmented the range points corresponding to eachobject. We denote the set of points belonging to an object i

in the figure as Ri. Now, given a recovered model withparameters pi, which corresponds to the object i, thequantity d3D amounts to the average radial euclideandistance of the points in Ri to the model pi and expressesa measure of recovery accuracy in three dimensions. That is,

d3DiðRi;piÞ ¼1

jRijX

X2Ri

dðX;piÞ: ð12Þ

In addition, we assessed the reconstruction of the objects’boundary. We regarded the image plane coordinates of allpoints in Ri and then extracted the boundary Si of thisregion. This 2D curve corresponds to the ground-trutheRIM of the object i. The metric d2D measures the averagedistance between the eRIM of the recovered object pi(denoted as Cpi ) to the points on Si and expresses theaccuracy in the recovery of the objects boundary, that is,

d2DiðSi;CpiÞ ¼1

jSijXx2Si

dðx;CpiÞ: ð13Þ

The accuracy results of both our approach and RAS areillustrated in Table 1. For each examined image, the average

KATSOULAS ET AL.: SUPERQUADRIC SEGMENTATION IN RANGE IMAGES VIA FUSION OF REGION AND BOUNDARY INFORMATION 789

Fig. 6. Configurations of rounded superquadrics. In (a), (d), and (g) aredisplayed the intensity images. In (b), (e), and (h), the edge map (darkpoints) superimposed to the range image. (c), (f), (i) show results on theimage plane: the recovered eRIMs, the regions corresponding to theseeds (light areas), and the regions added to the seed regions afterobject recovery (dark areas).

Fig. 7. Configurations of various superquadrics. In (a), (d), (g), (j), and(m) are displayed the intensity images. In (b), (e), (h), (k), and (n), theedge map (dark points) superimposed to the range image. (c), (f), (i), (l),and (o) show results on the image plane: the recovered eRIMs, theregions corresponding to the seeds (light areas), and the regions addedto the seed regions after object recovery (dark areas).

Page 10: IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE ... › ~afanasye › 10_UniTN_reports › 20_Presentations › … · 2 4 3 5¼ a 1 cosð Þ 1 cosð!Þ 2 a 2 cosð Þ 1 sinð!Þ

values of d2D and d3D for all the objects in the image aredenoted as �d2D and �d3D, respectively. Given this, the secondand third column in Table 1 show the �d3D and �d2D, when ourapproach is used for segmentation. The fourth and fifthcolumn illustrate the corresponding values when RAS isused. For every image, our approach performs better thanRAS. The last row of the table shows the averaged resultsalong all the rows. The average �d3D for all images examined is0.612 using our approach and 1.196 using RAS. Given thismetric, our approach delivered about two times moreaccurate results. The corresponding results for �d2D is0.747 and 2.422, respectively. As fas as this metric isconcerned, our approach introduces an improvement of afactor around 3. Worth noticing is that, from 79 objectsexamined, RAS delivered better reconstruction accuracyresults in only six cases. The primary reason for the failureof our approach in these cases is the low quality of both regionand boundary-based information.

Moreover, we compared our approach to RAS with respectto computational efficiency. As discussed, our approachrecovers each object independently from the others. In thisway, it offers the potential of parallelization, if within aparallel processing framework, each object is recovered byone processor. Given eight images containing 79 objects intotal (see Figs. 6a, 6b, 6c, 6d, 6e, 6f, 6g, 6h, and 6i, and Figs. 7a,7b, 7c, 7d, 7e, 7f, 7g, 7h, 7i, 7j, 7k, 7l, 7m, 7n, and 7o), wemeasured the average time needed for the recovery of aunique object. This was found to be 25.29 seconds per object.RAS cannot be fully parallelized. The corresponding timeRAS needed for object recovery was about 300 seconds. In thisrespect, our approach introduces an improvement of a factorof about 12 regarding efficiency. Note that our approach wasimplemented in Matlab, whereas RAS in C. An improvementon the efficiency of our method is expected to arise if ourapproach is ported to C.

5.3 Experiments with Multipart Objects

We performed experiments with multipart objects. Tworange images of multipart objects from the OSU (MSU/WSU)range image database where used for this purpose. Theimages along the recovery results are shown in Figs. 8a, 8b, 8c,8d, 8e, and 8f. Not all objects in those images can be wellmodeled by superquadrics. Despite this, in the majority of thecases, our model-based edge detector is able to capture theintersections between the parts and to provide relativelyaccurate boundary information. The adaptive boundaryclosure operation delivers a satisfactory seed placement.Finally, the hypothesis refinement framework delivers in the

majority of these images encouraging segmentation results.Note that the triangular part of the planar object in the firstfigure set cannot be expressed via the model. The same is thecase for the right part of the planar object of the second figureset. Results on the recovery of the objects for these figureshave been as well published in [26]. Despite the problems ofour approach in some cases, the overall results are consider-ably better than in [26].

Furthermore, we performed an experiment with a rangeimage of a wooden doll. The image appears in [17, p. 171].The results appear in Figs. 9a, 9b, 9c, and 9d. All parts of thedoll have been recovered. Again, the result shown here isbetter than the result delivered by RAS: The upper parts ofthe left and right hands of the doll, that is, the partsconnected to the torso are represented in [17] with two

790 IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL. 30, NO. 5, MAY 2008

TABLE 1Comparison of Accuracy of Reconstruction in 3D and 2D

between Our Approach and RAS

Fig. 8. Recovery results on multipart objects. In (a) and (d) are displayedthe intensity images. In (b) and (e) are the recovered RIMs (dark points)superimposed to the edge image (light points). Finally, in (c) and (f), weshow the recovered 3D models. The recovery results are not satisfactoryin some cases. The reason is that some object parts cannot be modeledwith superquadric models. The reader is referred to Section 5.3 fordetails.

Fig. 9. We performedexperiments with the image of a wooden mannequin,originally presented in [17, p. 171]. (a) shows the range image. (b) Therecovered RIMs (dark points) superimposed to the edge image (lightpoints). (c) shows the recovered models in three dimensions. Finally,(d) shows an additional 3D view of the recovered superquadrics.

Page 11: IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE ... › ~afanasye › 10_UniTN_reports › 20_Presentations › … · 2 4 3 5¼ a 1 cosð Þ 1 cosð!Þ 2 a 2 cosð Þ 1 sinð!Þ

superquadric models. In our case, one model expresses eachpart, as it should. Note that, in this case as well, some objectparts are not perfectly modeled with superquadics. Thedifference of these parts to superquadric models is not big.The majority of the boundary points are detected by theedge detector, and the obtained segmentation is reasonable.

5.4 Behavior against Occlusion

We performed a final experiment in order to quantify thebehavior of our algorithm against occlusion. In order toperform viewpoint invariant experiments, we considered aspherical superquadric with radius 30 mm, that is, �1 ¼ �2 ¼ 1and 1 ¼ 2 ¼ 3 ¼ 30. We then created range points on thesurface of the model and embedded the points on itshemisphere in a synthetic range image with an interpixeldistance of 1 mm. The RIM of this model is a circle, on whichwe considered M ¼ 100 points for our experiments. In thisexperiment, we varied independently the amount of bound-ing contour occlusion and the amount of region occlusion.

When the missing percentage of bounding contourinformation is lower than 55 percent, our approach managesto recover the object of interest with an accuracy of 0.5 pixelsfor d2D (see (12)) and 1 mm for d3D (13). For missingboundary information in the range 40-55 percent, the size ofthe seed generated by the hypothesis generation stage issmall. Despite this, the region module gradually increasesthe size of the seed due to the incorporated region growingoperation. Due to the fact that no noisy boundary pointsinhibit the growth of the model, the model manages to reachits actual size. For more than 55 percent missing boundaryinformation, the generated seed is very small, due to the

multiple invocations of the boundary closure operation. Inthis case, less than 10 range points are assigned to theinitialized model. Hence, the superquadric fitting functiondelivers very unreliable results and the system breaks down.

In addition, we performed experiments by varying theamount of region information. For missing region informa-tion, up to 60 percent of our approach delivers a satisfactoryresult. The accuracy in recovering boundary informationexpressed by d2D is up to 3.5 pixels. The accuracy inrecovering region information expressed by d3D is up to5 mm. For missing region information in the range of 70-90 percent, the size of the initialized seed is small. The object’sboundary points are far away from the seed, so the attractionforces they exert to the seed are low. In addition, there is noavailable region information that will be used by the regionmodule to grow the seed. As a result, the output of our systemdeteriorates considerably. Note that, for more than 90 percentmissing region information, the system breaks down, sincethe number of range points are not enough to produce areliable model via the model-fitting operation.

Finally, we measured the reconstruction accuracy of thevisible surface of the model as well. In this case, for missingboundary information up to 70 percent, the recovery accuracyvalues were less than 5 pixels for d2D and 7 mm for d3D. Formissing region information up to 78 percent, the recoveryaccuracy values were less than 4 pixels for d2D and 9 mm ford3D. These results are encouraging, and they demonstrate thatour system is robust against occlusion.

6 APPLICATION: ROBOTIC UNLOADING OF PILED

BOX-LIKE OBJECTS

We applied our segmentation framework to a novel roboticsystem (a description of which appears also in [54], [37],and [55]), for automatic unloading of piled box-like objects,as those in Fig. 2a, as well as in Figs. 6, 7, 8, 9, 10, 11, 12, 13,and 14. The driving force for the implementation of the

KATSOULAS ET AL.: SUPERQUADRIC SEGMENTATION IN RANGE IMAGES VIA FUSION OF REGION AND BOUNDARY INFORMATION 791

Fig. 10. Our robotic system in operation: The start and end positions of

the scanning process are shown in (a) and (b), respectively. (c) shows

the robotic hand while approaching an object. Finally, (d) illustrates the

robot just before grasping the object.

Fig. 11. Bags configuration.

Fig. 12. Pillows configuration.

Fig. 13. Box-like object configuration.

Fig. 14. Cardboard boxes configuration.

Page 12: IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE ... › ~afanasye › 10_UniTN_reports › 20_Presentations › … · 2 4 3 5¼ a 1 cosð Þ 1 cosð!Þ 2 a 2 cosð Þ 1 sinð!Þ

particular system is mainly the industrial requirement forthe reduction of costs in the logistic processes.

In the context of our system, a six degrees of freedomindustrial robotic arm (KUKA KR-15/1) is employed forobject unloading. A laser sensor (SICK LMS-200) ismounted on the hand of the robot for acquiring images ofthe topside of the pallet. Besides, a vacuum gripper is aswell mounted on the hand of the robot for object graspingand placement. Our system operates as follows: First, arange image of the object configuration is acquired. This isdone by a linear movement of the robotic hand across thetwo ends of the pallet. Second, the image is analyzed, and anumber of upper objects on the pile are recovered. Third,the recovered objects are grasped by the robotic hand andplaced at positions defined by the user. This processcontinues iteratively, until no objects lie on the pallet.Fig. 10 illustrates instances of the unloading process.

Given the input range image, our superquadric recoveryframework is applied to determine the parameters of thegraspable objects in the pile. The framework is adapted to theapplication as follows: First, box-like superquadrics are onlyused for modeling. This means that the shape �1, �2 parametersof the models are set in the value 0.2 and are kept constantwithin the recovery process. Second, four additional para-meters expressing global parabolic deformations areintroduced in the parameter vector of (4). These allow formodeling of the slight occurring on some nonrigid objects(for example, bags or pillows). Furthermore, a decompositionin the recovery of the parameter vector of the model isrealized: according to our framework, superquadric para-meter recovery is performed by two modules, each dealingwith its own domain of information. However, given the box-like geometry of the objects, not all the superquadricparameters are related to the domains on which each of thetwo modules acts. Decomposing the recovery of the modelparameter vector results into a search for the optimalparameters in lower dimensional parameter spaces. As aconsequence, both the robustness and computational effi-ciency of the recovery process increase.

Finally, a postprocessing step is included in the frame-work. The process succeeds the hypothesis refinement stageand determines if the refined model corresponds to a graspableobject. A graspable object is by definition an object that is notoccluded by another object. This implies that edge pointsalong the entire RIM of the object are captured by theboundary detection process and are thus present in theboundary image. Therefore, for the graspable objects in theimage, the fitting error residual of the model’s RIM isexpected to be low. In addition, since such objects can bewell modeled by superquadrics, the fitting error residual of asuperquadric model to the range points of the imagebelonging to the object is as well expected to be low. In short,the postprocessing method examines both model fitting errorresiduals to the range image and the boundary image and

decides that the object of interest is graspable if both aresmaller than user-defined thresholds. The model parametersdetermined to correspond to graspable objects are thenforwarded to the robot for performing grasping operations.

We tested the performance of our system on about 40 box-like object configurations. Representative results are shown inFigs. 11, 12, 13, 14, and 15. The first image of each row shows anobject configuration. The second, the range image on whichthe edge image is superimposed. The edge map with therecovered object boundaries (in bold), follow. The last imageshows the recovered models in 3D, superimposed on therange image. Note that, in some occasions, for example, inFigs. 12, 13, and 14, there are objects to which no modelcorresponds. These objects are determined to be nongrasp-able by the postprocessing step, and the correspondingmodels are rejected. The same holds for the recovery outputin Fig. 2k.

Accuracy results are shown in Table 2. The table showsaveraged results along all images for the metrics d2D andd3D, extracted as in Section 5. Note that the recoveryaccuracy in 3D is about 9 mm, that is, in the order of thesensor’s resolution, which is about 10 mm. As well, weachieved almost subpixel reconstruction accuracy in 2D.

We assessed the robustness of our system by comparingthe number of graspable objects in all box-like objectconfigurations against the one determined by a human.Results are shown in Table 2: In the 40 examined images,174 objects exist in total, 159 of which have been successfullyrecovered. That is, the percentage of true positive responseswas more than 91 percent. These statements are summarizedin Table 3. Note, that the system gave false negativeresponses in only six cases. Note as well that using theRAS framework one is not in the position to decide on thegraspability of objects. This feature is offered by ourapproach due to the consideration of boundary information.

Finally, our system is fast: The average object recoverytime was 15.5 seconds (hypothesis refinement) plus 3.13 sec-onds (hypothesis generation), that is, about 18 seconds in aPentium 4 PC, with a 1.6 Ghz processor. The combination ofthe advantages of our system, that is, flexibility, robustness,accuracy, computational efficiency, and low cost cannot befound in any other existing work to the best of ourknowledge [37].

792 IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL. 30, NO. 5, MAY 2008

Fig. 15. Configuration of various types of objects.

TABLE 2Accuracy and Robustness of Robotic Object Unloading System

TABLE 3Efficiency of Robotic Object Unloading System

Page 13: IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE ... › ~afanasye › 10_UniTN_reports › 20_Presentations › … · 2 4 3 5¼ a 1 cosð Þ 1 cosð!Þ 2 a 2 cosð Þ 1 sinð!Þ

7 CONCLUSIONS

We presented a strategy for multiple superquadric segmen-tation in range images. The segmentation task is performedwithin two stages: First, a hypothesis about the number ofobjects in the image and the parameters of the modelsrepresenting them is generated. Second, this hypothesis isrefined by means of iterative fitting of the models to thedata. Within our approach, the two aspects of informationthat a range image can provide are considered: region andobject boundary information. The former is provideddirectly by the input range image. The latter by a 2D edgemap obtained via edge detection on the range image.Hypothesis generation is realized via adaptive closure ofthe edge map, which indicates range point regions belong-ing to the object’s interior, on which seed models are fitted.The hypothesis refinement stage improves the resultdelivered by the latter stage. A game-theoretic frameworkis applied to perform refinement while simultaneouslyfusing the two information sources. This is done viaiterative invocations of two modules in succession: Theregion module refines the model parameters by modelfitting in the range image. The boundary module fits themodel’s boundary to the edge map. This takes place untilthe model parameters do not change significantly.

We demonstrated experimentally that our strategy out-performs the RAS framework [16], [17], the state of the art inmultiple superquadric segmentation, in terms of accuracyand computational efficiency. More specifically, our ap-proach reduced about 2 times the error metric for 3D modelreconstruction and about 3 times the error metric forreconstructing the boundaries of the objects in the 2D imageplane. Additionally, it proved to be more than 12 times fasterthan RAS.

The considerable improvements in accuracy introducedby our approach are due to the consideration of boundaryinformation within segmentation, in addition to regioninformation. In detail, using boundary information in thehypothesis generation stage produces a set of segmentationparameters that are much closer to the actual parametersthan those delivered by the corresponding stage of RAS.Furthermore, consideration of boundary information withinhypothesis refinement solves the region overgrowingproblem, which is the main source of low accuracy withinRAS. Finally, high accuracy is the outcome of incorporationof model information from the early steps and during allstages of the process.

Regarding computational efficiency, consideration ofboundary information in the hypothesis generation processresults into a smaller number of initialized seeds and, thus,in lower overall processing time. Besides, using boundaryinformation for parameter recovery accelerates the growingof the models: except for the standard region growing,models now grow as a result of the attraction forces exertedby the boundary points. In the beginning of the growingoperation, when the model is small, these forces areresponsible for fast growing of the models toward thereal-object boundaries. Note finally that recovery in alldomains is performed using parametric models of closedform, which considerably facilitates the fitting processes.

The novelty of our approach lies in the adjustment of agame-theoretic framework in [51], [40] to the segmentationof range images. More specifically, we exploit the feature ofdecoupling model recovery between the two informationdomains so as to enhance the descriptive ability of the dataand ease the model fitting process. The former is due toletting boundary operations occur in 2D instead of 3D. Thelatter is due to using the model’s boundary curve instead ofthe model itself for model fitting in the boundary domain.

Last but not least, we built a novel robotic system forautomatic piled box-like object unloading. The system isbased on our segmentation framework. A detailed experi-mental evaluation of the system showed that it is accurate,robust, and fast. We showed that the average time for therecovery of an objects was 15.5 seconds on the target platform.Additional knowledge on the shape of the objects in theconfiguration, for example, knowledge that the configurationconsists of rigid boxes only, for example, can further reducethe execution time of our system to 5 seconds per object.Given that the robotic grasping and placement operations donot last more than 4 seconds for an industrial robot in its fullspeed, our system has the potential to perform unloadingoperations faster than human workers. The advantagesexhibited by this system cannot be found in any otherapplication of this kind, up to our knowledge.

ACKNOWLEDGMENTS

This work was partially supported by the German ministry ofeconomy and technology under the PRO INNO ProgramGrant KF 0185401KLF0. The authors would like to thank Ale�sJakli�c and Ale�s Leonardis, University of Ljubljana, Slovenia,for sharing their code and the images they used in the RASevaluation, as well as for very useful discussions. Further-more, they thank all three anonymous reviewers for theirfruitful comments. This work was primarily conducted whenD. Katsoulas and C. Cea were with the Computer ScienceDepartment, Institute for Pattern Recognition and ImageProcessing, University of Freiburg, Germany.

REFERENCES

[1] H. Tek and B. Kimia, “Image Segmentation by Reaction-DiffusionBubbles,” Proc. IEEE Int’l Conf. Computer Vision, pp. 156-162, 1995.

[2] S. Zhu and A. Yuille, “Region Competition: Unifying Snakes,Region Growing, and Bayes/MDL for Multiband Image Segmen-tation,” IEEE Trans. Pattern Analysis and Machine Intelligence,vol. 18, no. 9, pp. 884-900, Sept. 1996.

[3] S. Sclaroff and L. Liu, “Deformable Shape Detection and Descrip-tion via Model-Based Region Grouping,” IEEE Trans. PatternAnalysis and Machine Intelligence, vol. 23, no. 5, pp. 475-489, May2001.

[4] N. Paragios and R. Deriche, “Coupled Geodesic Active Regionsfor Image Segmentation: A Level Set Approach,” Proc. EuropeanConf. Computer Vision, vol. 2, pp. 224-240, 2002.

[5] R. Nevatia and T. Binford, “Description and Recognition ofCurved Objects,” Artificial Intelligence, vol. 8, no. 1, pp. 77-98, 1977.

[6] D. Keren, D. Cooper, and J. Subrahmonia, “Describing Compli-cated Objects by Implicit Polynomials,” IEEE Trans. PatternAnalysis and Machine Intelligence, vol. 16, no. 1, pp. 38-53, Jan. 1994.

[7] I. Biederman, “Human Image Understanding: Recent Researchand a Theory,” Computer Vision Graphics and Image Processing,vol. 32, pp. 29-73, 1985.

[8] K. Wu and M. Levine, “3-D Shape Approximation UsingParametric Geons,” Image Vision Computing, vol. 15, no. 2,pp. 143-158, 1997.

KATSOULAS ET AL.: SUPERQUADRIC SEGMENTATION IN RANGE IMAGES VIA FUSION OF REGION AND BOUNDARY INFORMATION 793

Page 14: IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE ... › ~afanasye › 10_UniTN_reports › 20_Presentations › … · 2 4 3 5¼ a 1 cosð Þ 1 cosð!Þ 2 a 2 cosð Þ 1 sinð!Þ

[9] L. Staib and J. Duncan, “Model Based Deformable Surface Findingfor Medical Images,” IEEE Trans. Medical Imaging, vol. 15, no. 5,pp. 720-731, 1996.

[10] L. Staib and J. Duncan, “Boundary Finding with ParametricallyDeformable Models,” IEEE Trans. Pattern Analysis and MachineIntelligence, vol. 14, no. 11, pp. 1061-1075, Nov. 1992.

[11] D. Terzopoulos, A. Witkin, and M. Kass, “Symmetry-SeekingModels and 3D Object Reconstruction,” Proc. IEEE Int’l Conf.Computer Vision, pp. 269-276, 1987.

[12] S. Muraki, “Volumetric Shape Description of Range Data UsingBlobby Model,” Proc. Int’l Conf. Computer Graphics and InteractiveTechniques (SIGGRAPH ’91), Computer Graphics, vol. 25, no. 4,pp. 227-235, July 1991.

[13] J. Hanson, “Hyperquadrics: Smoothly Deformable Shapes withConvex Polyhedral Bounds,” Computer Vision, Graphics, and ImageProcessing, vol. 44, pp. 191-210, 1988.

[14] A. Barr, “Superquadrics and Angle Preserving Transformations,”IEEE Computer Graphics and Applications, vol. 1, no. 1, pp. 11-23,1981.

[15] F. Solina and R. Bajcsy, “Recovery of Parametric Models fromRange Images: The Case for Superquadrics with Global Deforma-tions,” IEEE Trans. Pattern Analysis and Machine Intelligence, vol. 12,no. 2, pp. 131-147, Feb. 1990.

[16] A. Leonardis, A. Jaklic, and F. Solina, “Superquadrics forSegmenting and Modeling Range Data,” IEEE Trans. PatternAnalysis and Machine Intelligence, vol. 19, no. 11, pp. 1289-1295,Nov. 1997.

[17] A. Jakli�c, A. Leonardis, and F. Solina, “Segmentation andRecovery of Superquadrics,” Computational Imaging and Vision,vol. 20, 2000.

[18] F. Ferrie, J. Lagarde, and P. Whaite, “Darboux Frames, Snakes,and Super-Quadrics: Geometry from the Bottom Up,” IEEE Trans.Pattern Analysis and Machine Intelligence, vol. 15, no. 8, pp. 771-784,Aug. 1993.

[19] D. Metaxas and S. Dickinson, “Integration of Quantitative andQualitative Techniques for Deformable Model Fitting fromOrthographic, Perspective, and Stereo Projections,” Proc. IEEEInt’l Conf. Computer Vision, pp. 641-649, 1993.

[20] D. Terzopoulos and D. Metaxas, “Dynamic 3D Models with Localand Global Deformations: Deformable Superquadrics,” IEEETrans. Pattern Analysis and Machine Intelligence, vol. 13, no. 7,pp. 703-714, July 1991.

[21] N. Raja and A. Jain, “Recognizing Geons from SuperquadricsFitted to Range Data,” Image and Vision Computing, vol. 10, no. 3,pp. 179-190, Apr. 1992.

[22] A. Pentland and S. Sclaroff, “Closed-Form Solutions for PhysicallyBased Shape Modeling and Recognition,” IEEE Trans. PatternAnalysis and Machine Intelligence, vol. 13, no. 7, pp. 715-729, July1991.

[23] A. Gupta, G. Lea, and K. Wohn, “Segmentation, Modeling andClassification of the Compact Objects in a Pile,” Proc. SPIE Symp.Intelligent Robot and Computer Vision VIII: Algorithms and Techni-ques, pp. 98-108, Nov. 1989.

[24] A. Gupta and R. Bajcsy, “Volumetric Segmentation of RangeImages of 3D Objects Using Superquadric Models,” ComputerVision Graphics and Image Processing: Image Understanding, vol. 58,no. 3, pp. 302-326, 1993.

[25] L. Chevalier, F. Jaillet, and A. Baskurt, “Segmentation andSuperquadric Modeling of 3D Objects,” Proc. Int’l Conf. CentralEurope Computer Graphics, Visualization, and Computer Vision, 2003.

[26] S.J. Dickinson, D.N. Metaxas, and A. Pentland, “The Role ofModel-Based Segmentation in the Recovery of Volumetric Partsfrom Range Data,” IEEE Trans. Pattern Analysis and MachineIntelligence, vol. 19, no. 3, pp. 259-267, Mar. 1997.

[27] D. DeCarlo and D. Metaxas, “Shape Evolution with Structural andTopological Changes Using Blending,” IEEE Trans. PatternAnalysis and Machine Intelligence, vol. 20, no. 11, pp. 1186-1205,Nov. 1998.

[28] T. Horikoshi and S. Suzuki, “3D Parts Decomposition from SparseRange Data Using Information Criterion,” Proc. IEEE CS Conf.Computer Vision and Pattern Recognition, pp. 168-173, 1993.

[29] H. Zha, T. Hoshide, and T. Hasegawa, “A Recursive Fitting-and-Splitting Algorithm for 3-D Object Modeling Using Superquad-rics,” Proc. IEEE Int’l Conf. Pattern Recognition, pp. 658-662, 1998.

[30] A. Pentland, “Automatic Extraction of Deformable Part Models,”Int’l J. Computer Vision, vol. 4, no. 2, pp. 107-126, 1990.

[31] R. Duda, P. Hart, and D. Stork, Pattern Classification, second ed.Wiley Interscience, 2001.

[32] J. Schoukens and R. Pintelon, Identification of Linear Systems. APractical Guideline to Accurate Modeling. Pergamon Press, 1991.

[33] A. Leonardis, A. Gupta, and R. Bajcsy, “Segmentation of RangeImages as the Search for Geometric Parametric Models,” Int’l J.Computer Vision, vol. 14, no. 3, pp. 253-277, Apr. 1995.

[34] C. Cea, “Superquadric Recovery in Range Images via RegionGrowing Influenced by Boundary Information,” master’s thesis,Computer Science Dept., Inst. for Pattern Recognition and ImageProcessing, Univ. of Freiburg, 2004.

[35] M. Pilu and R. Fisher, “Equal-Distance Sampling of SuperellipseModels,” Proc. 1995 British Conf. Machine Vision, vol. 1, pp. 257-266, 1995.

[36] P. Besl and R. Jain, “Segmentation through Variable-Order SurfaceFitting,” IEEE Trans. Pattern Analysis and Machine Intelligence,vol. 9, no. 2, pp. 167-192, 1988.

[37] D. Katsoulas, “Robust Recovery of Piled Box-Like Objects inRange Images,” PhD dissertation, Dept. of Computer Science, Inst.for Pattern Recognition and Image Processing, Univ. of Freiburg,2004.

[38] J. Freixenet, X. Munoz, J. Martı, and X. Llado, “Colour TextureSegmentation by Region-Boundary Cooperation,” Proc. EuropeanConf. Computer Vision, vol. 2, pp. 250-261, Sept. 2004.

[39] S. Li and T. Basar, “Distributed Algorithms for the Computationof Non-Cooperative Equilibria,” Automatica, vol. 23, no. 4, pp. 523-533, 1987.

[40] A. Chackraborty and J. Duncan, “Game Theoretic Integration forImage Segmentation,” IEEE Trans. Pattern Analysis and MachineIntelligence, vol. 21, no. 1, pp. 12-30, Jan. 1999.

[41] A. Chakraborty, “Feature and Module Integration for ImageSegmentation,” PhD dissertation, Yale Univ., 1996.

[42] X. Jiang and H. Bunke, “Edge Detection in Range Images Based onScan Line Approximation,” Computer Vision and Image Under-standing, vol. 73, no. 2, pp. 183-199, Feb. 1999.

[43] R. Duda and P. Hart, Pattern Recognition and Scene Analysis. JohnWiley & Sons, 1973.

[44] D. Katsoulas and A. Werber, “Edge Detection in Range Images ofPiled Box-Like Objects,” Proc. 17th Int’l Conf. Pattern Recognition,vol. 2, pp. 80-84, 2004.

[45] X. Munoz, “Image Segmentation Integrating Colour, Texture andBoundary Information,” PhD dissertation, Univ. of Girona, 2003.

[46] J. Benois and D. Barba, “Image Segmentation by Region-ContourCooperation for Image Coding,” Proc. Int’l Conf. Pattern Recogni-tion, vol. C, pp. 331-334, 1992.

[47] D. Sinclair, “Voronoi Seeded Colour Image Segmentation,”Technical Report TR99-04, AT&T Laboratories, Cambridge, May1999.

[48] T. Shimbashi, Y. Kokubo, and N. Shirota, “Region SegmentationUsing Edge Based Circle Growing,” Proc. IEEE Int’l Conf. ImageProcessing, vol. C, pp. 65-68, 1995.

[49] X. Jiang, “An Adaptive Contour Closure Algorithm and ItsExperimental Evaluation,” IEEE Trans. Pattern Analysis andMachine Intelligence, vol. 22, no. 11, pp. 1252-1265, Nov. 2000.

[50] A. Chakraborty, L.H. Staib, and J.S. Duncan, “DeformableBoundary Finding Influenced by Region Homogeneity,” Proc.IEEE Conf. Computer Vision and Pattern Recognition, pp. 624-627,June 1994.

[51] H. Bozma and J. Duncan, “A Game-Theoretic Approach toIntegration of Modules,” IEEE Trans. Pattern Analysis and MachineIntelligence, vol. 16, pp. 1074-1086, 1994.

[52] C. Ye and J. Borenstein, “Characterization of a 2-D Laser Scannerfor Mobile Robot Obstacle Navigation,” Proc. IEEE Int’l Conf.Robotics and Automation, pp. 2512-2518, May 2002.

[53] D. Skocaj and A. Leonardis, “Range Image Acquisition of Objectswith Non-Uniform Albedo Using Structured Light Range Sensor,”Proc. Int’l Conf. Pattern Recognition, pp. 1778-1781, 2000.

[54] D. Katsoulas and D.I. Kosmopoulos, “Box-Like SuperquadricRecovery in Range Images by Fusing Region and BoundaryInformation,” Proc. IEEE Int’l Conf. Pattern Recognition, pp. 719-722, 2006.

[55] D. Katsoulas, “Reliable Recovery of Piled Box-Like Objects viaParabolically Deformable Superquadrics,” Proc. IEEE Int’l Conf.Computer Vision, pp. 931-938, 2003.

794 IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, VOL. 30, NO. 5, MAY 2008

Page 15: IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE ... › ~afanasye › 10_UniTN_reports › 20_Presentations › … · 2 4 3 5¼ a 1 cosð Þ 1 cosð!Þ 2 a 2 cosð Þ 1 sinð!Þ

Dimitrios Katsoulas received the diploma inelectrical engineering from the National Techni-cal University of Athens in 1997 and the PhDdegree from the University of Freiburg, Ger-many, in 2005. He is now with the GreekNational Center for Scientific Research Demok-ritos, Computational Intelligence Laboratory,and with INOS Automation Software. His re-search interests include model-based imagesegmentation, geometric modeling, visual servo-

ing, and multicamera tracking. He is a member of the GermanAssociation for Pattern Recognition and of the Technical Chamber ofGreece.

Christian Cea Bastidas received the diploma incivil engineering from the University of Chile in1997 and the master’s degree in computerscience from the University of Freiburg, Ger-many, in 2004. He worked for a period of fiveyears in the private sector. He joined theComputer Science Department, University ofFreiburg, Germany, in 2002. He is now withEmpresa Nacional De Telecomunicaciones(ENTEL) S.A., Santiago, Chile. His interests

include image processing, computer vision, and optimization theory.

Dimitrios Kosmopoulos received the diplomain electrical engineering and the PhD degreefrom the National Technical University of Athensin 1997 and 2002, respectively. He currentlyworks as a research scientist in the Computa-tional Intelligence Laboratory, Institute of Infor-matics and Telecommunications, Greek NationalCenter for Scientific Research Demokritos. He isalso a lecturer in the Department of ComputerScience and Technology, University of Pelopon-

nese. He has been employed as a researcher and developer for variouscompanies and institutions. His research interests include computer androbotic vision, multimedia processing, and medical imaging. He hasauthored several publications in those fields.

. For more information on this or any other computing topic,please visit our Digital Library at www.computer.org/publications/dlib.

KATSOULAS ET AL.: SUPERQUADRIC SEGMENTATION IN RANGE IMAGES VIA FUSION OF REGION AND BOUNDARY INFORMATION 795