The Digital Imaging and Remote Sensing Image Generation (DIRSIG) model is an established, first-principles based scene simulation tool that produces synthetic multispectral and hyperspectral images from the visible to long wave infrared (0.4 to 20 microns). Over the last few years, significant enhancements such as spectral polarimetric and active Light Detection and Ranging (LIDAR) models have also been incorporated into the software, providing an extremely powerful tool for algorithm testing and sensor evaluation. However, the extensive time required to create large-scale scenes has limited DIRSIG’s ability to generate scenes “on demand.” To date, scene generation has been a laborious, time-intensive process, as the terrain model, CAD objects and background maps have to be created and attributed manually. To shorten the time required for this process, we are initiating a research effort that aims to reduce the man-in-the-loop requirements for several aspects of synthetic hyperspectral scene construction. Through a fusion of 3D LIDAR data with passive imagery, we are working to semi-automate several of the required tasks in the DIRSIG scene creation process. Additionally, many of the remaining tasks will also realize a shortened implementation time through this application of multi-modal imagery. This paper reports on the progress made thus far in achieving these objectives.
Date of creation, presentation, or exhibit
Department, Program, or Center
Chester F. Carlson Center for Imaging Science (COS)
S. R. Lach, S. D. Brown, J. P. Kerekes, "Semi-automated DIRSIG scene modeling from 3D LIDAR and passive imaging sources", Proc. SPIE 6214, Laser Radar Technology and Applications XI, 62140I (19 May 2006); doi: 10.1117/12.666096; https://doi.org/10.1117/12.666096
RIT – Main Campus