19 February 2009 From Talbot bands to no mirror terms in Fourier domain OCT: a quest on the fundamental origin of the sensitivity decay with depth
Author Affiliations +
Abstract
We show that the decay of sensitivity with depth in Fourier Domain (FD)-OCT can be explained based on the superposition of wavetrain lengths after dispersion (diffraction). The more coherent the dispersed waves are, the slower the decay of sensitivity. An asymmetry in the decay curve with the optical path difference (OPD) can be introduced via an "intrinsic" delay which has the effect of shifting axially the two wavetrains relative to each other, originating from the object and reference beam of an interferometer. In this way, an equivalent Talbot Bands set-up is implemented, characterized by no mirror terms for the "extrinsic" delay introduced in the interferometer. Such configurations require that the two interfering beams use different parts of the diffraction grating in the interrogating spectrometer. Theory of Talbot Bands is presented and then how this knowledge can be transferred to the field of FD-OCT to achieve A-scans mirror terms free in one step. A theoretical rigorous model and a heuristic model are presented to quantify the Talbot bands.
© (2009) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.
A. Gh. Podoleanu, D. Woods, M. Hughes, "From Talbot bands to no mirror terms in Fourier domain OCT: a quest on the fundamental origin of the sensitivity decay with depth", Proc. SPIE 7168, Optical Coherence Tomography and Coherence Domain Optical Methods in Biomedicine XIII, 71681L (19 February 2009); doi: 10.1117/12.808327; https://doi.org/10.1117/12.808327
PROCEEDINGS
7 PAGES


SHARE
Back to Top