Today, night vision sensor and display systems used in the pilotage or navigation of military helicopters are either long wave IR thermal sensors (8 - 12 microns) or image intensified, visible and near IR (0.6 - 0.9 microns), sensors. The sensor imagery is displayed using a monochrome phosphor on a Cathode Ray Tube or night vision goggle. Currently, there is no fielded capability to combine the best attributes of the emissive radiation sensed by the thermal sensor and the reflected radiation sensed by the image intensified sensor into a single fused image. However, recent advances in signal processing have permitted the real time image fusion and display of multispectral sensors in either monochrome or synthetic chromatic form. The merits of such signal processing is explored. A part task simulation using a desktop computer, video playback unit, and a biocular head mounted display was conducted. Response time and accuracy measures of test subject responses to visual perception tasks were taken. Subjective ratings were collected to determine levels of pilot acceptance. In general, fusion based formats resulted in better subject performance. The benefits of integrating synthetic color to fused imagery, however, is dependent on the color algorithm used, the visual task performed, and scene content.