3 October 2008 On 3D radar data visualization and merging with camera images
Author Affiliations +
Abstract
The possibilities to support the interpretation of spatial 3D-radar data visually both with and without camera images are studied. Radar measurements and camera pictures of a person are analyzed. First, the received signal amplitudes distributed in three dimensions, spherical range and two angles, are fed to a selection procedure using amplitude and the scene volume of interest. A number of resolution cells will then form images based on a volume representation depending upon the amplitude and location. Projecting the images of all the cells upon an imaging plane then forms the total image. Different images of a radar data set are performed for different projecting planes. The images were studied to find efficient aspect angles to get the target information of most interest. Rotating the target data around a suitable axis may perform such search. In addition, a visualization method for presenting radar data merged with a camera picture has been developed. An aim in this part of the work has been to keep the high information content of the camera image in the merged image. From the 3D-radar measurements the radar data may be projected upon the imaging plane of a camera with an arbitrary viewing center. This possibility is presented in examples with one camera looking at the target scene from the radar location and another camera looking from an aspect angle differing 45° relative to the aspect angle of the radar.
© (2008) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.
J. Kjellgren, "On 3D radar data visualization and merging with camera images", Proc. SPIE 7117, Millimetre Wave and Terahertz Sensors and Technology, 71170G (3 October 2008); doi: 10.1117/12.799811; https://doi.org/10.1117/12.799811
PROCEEDINGS
12 PAGES


SHARE
Back to Top