1
Segmentation of dynamic object from low acquisition rate range data Agustin Ortega Juan Andrade-Cetto We address the problem of eliminating moving bodies from low acquisition rate 3d range scans by using high rate image data. Range points are classified according to temporally consistent corresponding images per- pixel class distributions are learnt on-line as Gaussian Mixture. We address the problem of eliminating moving bodies from low acquisition rate 3d range scans by using high rate image data. Range points are classified according to temporally consistent corresponding images per- pixel class distributions are learnt on-line as Gaussian Mixture. Hokuyo UTM- 30LX laser mounted on a slip- ring, with computer-controlled angular position laser resolution has been set to 0.5 degrees in azimuth with 360 degrees omnidirectional field of view, and 0.5 degrees resolution in elevation for a range of 270 degrees. The color camera used is a Pointgray Flea camera with M1214-MP optics and 40.4 degrees field of view Hokuyo UTM- 30LX laser mounted on a slip- ring, with computer-controlled angular position laser resolution has been set to 0.5 degrees in azimuth with 360 degrees omnidirectional field of view, and 0.5 degrees resolution in elevation for a range of 270 degrees. The color camera used is a Pointgray Flea camera with M1214-MP optics and 40.4 degrees field of view Intrinsic camera parameters are computed using Zhang's method [3]. Intrinsic camera parameters are computed using Zhang's method [3]. The probability of each image pixel to be of the static object class is modeled as a weighted mixture of k Gaussians distributions. The probability of each image pixel to be of the static object class is modeled as a weighted mixture of k Gaussians distributions. The mixture is iteratively refined from image data The mixture is iteratively refined from image data We choose as the static object class, the first s<k ordered distributions which add up to a factored weight ws, where We choose as the static object class, the first s<k ordered distributions which add up to a factored weight ws, where For each point in a slice, its corresponding pixel values for the subset of matching synchronized images is measured and compared to the full set of pixel values for the whole sequence by means of the learn mixture distributions. A range point is classified as static if all matching pixels in the synchronized image subset are within the learnt static class For each point in a slice, its corresponding pixel values for the subset of matching synchronized images is measured and compared to the full set of pixel values for the whole sequence by means of the learn mixture distributions. A range point is classified as static if all matching pixels in the synchronized image subset are within the learnt static class The 3D scanner takes about 18 seconds to complete a full scan. Camera frame rate is set to 17 fps. The timestamps between consecutive laser slices, and grabbed images are compared and set to lie within a reasonable threshold Ts in milliseconds. The set of images corresponding to each laser range slice is I. The 3D scanner takes about 18 seconds to complete a full scan. Camera frame rate is set to 17 fps. The timestamps between consecutive laser slices, and grabbed images are compared and set to lie within a reasonable threshold Ts in milliseconds. The set of images corresponding to each laser range slice is I. Submitted to ECMR'11 [1] Submitted to ECMR'11 [1] Problem definition Problem definition Proposal approach Proposal approach Sensor specifications Sensor specifications Sensor synchronization Sensor synchronization Camera calibration Camera calibration Removal of moving objects Removal of moving objects Extrinsic calibration between the 3D laser and the camera is initialized using Hager's method [4] for pose estimation over a human-selected set of point correspondences Extrinsic calibration between the 3D laser and the camera is initialized using Hager's method [4] for pose estimation over a human-selected set of point correspondences Camera-Laser calibration Camera-Laser calibration Camera to laser extrinsic calibration is then refined by minimizing the projection error of line primitives, computed as the weighted sum of both angular and midpoint location errors [2]. Camera to laser extrinsic calibration is then refined by minimizing the projection error of line primitives, computed as the weighted sum of both angular and midpoint location errors [2]. Refining calibration Refining calibration Static class Static class Point classification Point classification Experiments Experiments References References [1] A. Ortega, J. Andrade-Cetto,Segmentation of dynamic objects from laser data, submitted to ECMR'11. . [2]A. Ortega, B. Dias, E. Teniente, A. Bernardino, J. Gaspar, and Juan Andrade-Cetto. Calibrating an outdoor distributed camera network using laser range finder data. IROS'09 [3]Z. Zhang. A flexible new technique for camera calibration. PAMI'00. [4]C.P. Lu, G.D. Hager, and E. Mjolsness. Fast and globally convergent pose estimation from video images.PAMI'00 [1] A. Ortega, J. Andrade-Cetto,Segmentation of dynamic objects from laser data, submitted to ECMR'11. . [2]A. Ortega, B. Dias, E. Teniente, A. Bernardino, J. Gaspar, and Juan Andrade-Cetto. Calibrating an outdoor distributed camera network using laser range finder data. IROS'09 [3]Z. Zhang. A flexible new technique for camera calibration. PAMI'00. [4]C.P. Lu, G.D. Hager, and E. Mjolsness. Fast and globally convergent pose estimation from video images.PAMI'00

Agustin Ortega Juan Andrade-Cetto...Segmentation of dynamic object from low acquisition rate range data Agustin Ortega Juan Andrade-Cetto We address the problem of eliminating moving

  • Upload
    others

  • View
    0

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Agustin Ortega Juan Andrade-Cetto...Segmentation of dynamic object from low acquisition rate range data Agustin Ortega Juan Andrade-Cetto We address the problem of eliminating moving

Segmentation of dynamic object from low acquisition rate range data

Agustin Ortega Juan Andrade-Cetto

We address the problem of eliminating moving bodies from low acquisition rate 3d range scans by using high rate image data.Range points are classified according to temporally consistent corresponding images per- pixel class distributions are learnt on-line as Gaussian Mixture.

We address the problem of eliminating moving bodies from low acquisition rate 3d range scans by using high rate image data.Range points are classified according to temporally consistent corresponding images per- pixel class distributions are learnt on-line as Gaussian Mixture.

Hokuyo UTM- 30LX laser mounted on a slip-ring, with computer-controlled angular position laser resolution has been set to 0.5 degrees in azimuth with 360 degrees omnidirectional field of view, and 0.5 degrees resolution in elevation for a range of 270 degrees. The color camera used is a Pointgray Flea camera with M1214-MP optics and 40.4 degrees field of view

Hokuyo UTM- 30LX laser mounted on a slip-ring, with computer-controlled angular position laser resolution has been set to 0.5 degrees in azimuth with 360 degrees omnidirectional field of view, and 0.5 degrees resolution in elevation for a range of 270 degrees. The color camera used is a Pointgray Flea camera with M1214-MP optics and 40.4 degrees field of view

Intrinsic camera parameters are computed using Zhang's method [3].

Intrinsic camera parameters are computed using Zhang's method [3].

The probability of each image pixel to be of the static object class is modeled as a weighted mixture of k Gaussians distributions.

The probability of each image pixel to be of the static object class is modeled as a weighted mixture of k Gaussians distributions.

The mixture is iteratively refined from image data The mixture is iteratively refined from image data

We choose as the static object class, the first s<k ordered distributions which add up to a factored weight ws, where

We choose as the static object class, the first s<k ordered distributions which add up to a factored weight ws, where

For each point in a slice, its corresponding pixel values for the subset of matching synchronized images is measured and compared to the full set of pixel values for the whole sequence by means of the learn mixture distributions. A range point is classified as static if all matching pixels in the synchronized image subset are within the learnt static class

For each point in a slice, its corresponding pixel values for the subset of matching synchronized images is measured and compared to the full set of pixel values for the whole sequence by means of the learn mixture distributions. A range point is classified as static if all matching pixels in the synchronized image subset are within the learnt static class

The 3D scanner takes about 18 seconds to complete a full scan. Camera frame rate is set to 17 fps. The timestamps between consecutive laser slices, and grabbed images are compared and set to lie within a reasonable threshold Ts in milliseconds. The set of images corresponding to each laser range slice is I.

The 3D scanner takes about 18 seconds to complete a full scan. Camera frame rate is set to 17 fps. The timestamps between consecutive laser slices, and grabbed images are compared and set to lie within a reasonable threshold Ts in milliseconds. The set of images corresponding to each laser range slice is I.

Submitted to ECMR'11 [1]Submitted to ECMR'11 [1]

Problem definitionProblem definition

Proposal approachProposal approach

Sensor specificationsSensor specifications

Sensor synchronizationSensor synchronization

Camera calibrationCamera calibration Removal of moving objectsRemoval of moving objects

Extrinsic calibration between the 3D laser and the camera is initialized using Hager's method [4] for pose estimation over a human-selected set of point correspondences

Extrinsic calibration between the 3D laser and the camera is initialized using Hager's method [4] for pose estimation over a human-selected set of point correspondences

Camera-Laser calibrationCamera-Laser calibration

Camera to laser extrinsic calibration is then refined by minimizing the projection error of line primitives, computed as the weighted sum of both angular and midpoint location errors [2].

Camera to laser extrinsic calibration is then refined by minimizing the projection error of line primitives, computed as the weighted sum of both angular and midpoint location errors [2].

Refining calibrationRefining calibration

Static class Static class

Point classificationPoint classification

ExperimentsExperiments

ReferencesReferences[1] A. Ortega, J. Andrade-Cetto,Segmentation of dynamic objects from laser data, submitted to ECMR'11. .[2]A. Ortega, B. Dias, E. Teniente, A. Bernardino, J. Gaspar, and Juan Andrade-Cetto. Calibrating an outdoor distributed camera network using laser range finder data. IROS'09[3]Z. Zhang. A flexible new technique for camera calibration. PAMI'00.[4]C.P. Lu, G.D. Hager, and E. Mjolsness. Fast and globally convergent pose estimation from video images.PAMI'00

[1] A. Ortega, J. Andrade-Cetto,Segmentation of dynamic objects from laser data, submitted to ECMR'11. .[2]A. Ortega, B. Dias, E. Teniente, A. Bernardino, J. Gaspar, and Juan Andrade-Cetto. Calibrating an outdoor distributed camera network using laser range finder data. IROS'09[3]Z. Zhang. A flexible new technique for camera calibration. PAMI'00.[4]C.P. Lu, G.D. Hager, and E. Mjolsness. Fast and globally convergent pose estimation from video images.PAMI'00