object position and orientation detection for mars science laboratory helicopter test imagery

15
Object Position and Orientation Detection For Mars Science Laboratory Helicopter Test Imagery Michael Johnson EE 368 Stanford University June 5, 2012

Upload: rhonda

Post on 24-Feb-2016

22 views

Category:

Documents


0 download

DESCRIPTION

Object Position and Orientation Detection For Mars Science Laboratory Helicopter Test Imagery. Michael Johnson EE 368 Stanford University June 5, 2012. Background. - PowerPoint PPT Presentation

TRANSCRIPT

Slide 1

Object Position and Orientation DetectionFor Mars Science Laboratory Helicopter Test ImageryMichael JohnsonEE 368Stanford UniversityJune 5, 2012

BackgroundThe Mars Science Laboratory Terminal Descent Sensor, the rovers landing radar, underwent a field test campaign in the Mojave Desert.One purpose of the field test was to gauge how the rovers position during the sky crane descent phase might impact the return of the radar signal during this crucial phase of the mission.To test this, the landing radar was affixed to a helicopter, and a mock up of the rover, with targets attached to the top deck, was attached to a wench underneath the helicopter. The radar was operated while the position of the rover mock up was varied.A camera was positioned near the wench, facing down, to image the rover as the test was underway. Time-tagged images were captured at a rate of 6 frames per second, which will be used to estimate the position of the rover during the test and matched with specific times of radar test data.

Problem StatementAs stated in the background, the problem is to derive estimations of an objects position and orientation in an image.In each image, the object (in this case the rover mock up) will be:Translated along the x and y dimensions.Scaled in size depending on the distance from the camera.Rotated due to helicopter maneuvers, wind, and other factors.Although the mockup may also experience roll and pitch effects, these are assumed to be minimal due to the nature of the test, and will be ignored.Image quality can vary due to:Angle of the sun causing bright reflections on the rover surfaces.Shadows cast by the helicopter over the rover, obscuring the targets.

Proposed SolutionIt is proposed to detect the rovers position and orientation in the images using the following three steps:Perform Sobel edge detection to detect the edges of the circular targets.Calculate the Circular Hough Transform on the filtered images to find the circular targets by detecting the peaks in the Circular Hough Transform. This will be performed over a range of possible radiuses until good peaks are seen. This will give us a set of possible target locations.After possible targets have been identified, a Procrustes method will be performed for matching each possible target with the known actual targets.

Sobel Edge DetectionThe Sobel Operator will be used for edge detection in the images, along both the horizontal and vertical (essential for detecting circles).

The Sobel Operator uses 3x3 kernels convolved with the original image. These kernels are:

The effect of these kernels is to calculate the approximations of the derivatives at each point which can detect abrupt changes from light to dark intensities.

Sobel Edge Detection ResultsAn example of Sobel Edge Detection on the imagery is shown below. The detection does a good job at capturing the edges of the targets while rejecting some of the other edged.It is noted that there is difficulty in detecting the targets obscured by shadows, which is acceptable if enough of the other targets are detected

Circular Hough TransformThe Circular Hough Transform is a series of operations performed on black and white images for circle detection.The general method of the transform is as follows:Create an accumulator matrix the size of the image.For every bright pixel in the image, draw a circle of radius R centered at this pixel in the accumulator matrix. In other words, increment the value of the accumulator matrix by one at every pixel where a circle is drawn.Peaks in the resulting accumulator matrix indicate circles of radius R being present at that pixel in the original image.An illustration:

Circle in source imageRed circles are those drawn around the source circleThe accumulator develops apeak where the circles intersect,which is the center of the circle

Circular Hough Transform Results

BW level image of Sobel Filtered ImageResults of Circular Hough TransformThe centers of the targets have noticeable peaks in them

Circular Hough Transform ResultsDue to irregularities in the shapes and sizes of the target circles, the Circular Hough Transform may not produce a peak at the center of each target, but a small cluster of peaks near the center. In these cases, close peaks (on the order of less than a circle radius away) will be averaged together.

Circular Hough Transform ResultsGenerally, the method is capable of picking up a large fraction of the targets.Some are not detected due to shadowing.Also, clutter around the target (tape round the top-most undetected target) can cause clutter in the detected edges, therefore a poor circle shape.However, further processing will allow for good object tracking even in when some targets are not identified.Method is very good at not erroneously identifying clutter as a target.

Original image with detected targets superimposed.

Procrustes MethodOnce potential targets are identified, the next step is to match them with the actual known targets. A Procrustes Method was employed to do this.The Procrustes Method employed consists of several stages:Pre-process the set of actual set of points by translating them in X and Y such that the mean position of all the targets is zero, and then scale them such that the RMS distance of all points to the origin is one.For the identified points:Translate the identified points in X an Y the same way as the actual points.Scale the identified points the same way as the actual points.Once the points are translated and scaled, finding orientation angle:Iterate over a set of angles and calculate the distance from each identified target to the nearest actual target.For each identified target, choose as its pair the actual target that is closest.Calculate as a cost function the sum of the square of these distances.The rotation angle to use is that which has the smallest value of the cost function.

Procrustes MethodAn orientation angle is now known. However, because the identified targets may be missing some, the translation and scaling first performed may not have been accurate.Therefore, another round of translation and scaling is performed.Translate the identified targets again in X and Y by sweeping over a small range of values, picking the one produces the minimum value of the cost function.Likewise, scale the points again over a small range of values, again picking the scale value that minimizes the cost function.Performing another round of rotation could be done, but generally by this point the identified and actual points will be close enough to not require it.At this point, each identified point can be matched with an actual point by finding the actual point that is closest to each identified point.Now, we have the pixel location in the image for each actual target, and similarly the orientation angle, the scale, and translation.

Results

An example image with each detected target assigned a reference number corresponding to an actual target.A plot of the normalized actual target locations with the locations of the detected targets after translation, scaling, and rotation.

ResultsA subset of 50 images was selected and run through the described method.

An average of 74.4% of targets are identified, plenty to develop a good estimation of the rover position.

14ConclusionPerformance is generally good, although some images are more difficult to track than others due to the reasons previously mentioned: shadowing, sun angle, etc.

The algorithm is efficient, taking a fraction of a second to complete per image with little optimization.

Room for improvements include performance optimizations and more thorough pose estimation based on roll/pitch/yaw characteristics.