We discuss and demonstrate the dependence of noise on the signal in time-domain optical coherence tomography (TDOCT). We then derive a depth-dependent matched filter to maximize the signal-to-noise ratio at every pixel in a depth scan (A-scan). We use an empirical estimate of the second order statistics of the noise in OCT images of vascular tissue to implement a depth-dependent filter that is matched to these images. The application of our filter results in an average increase of signal-to-noise ratio of about 7 dB compared to a simple averaging operation. Our filter is not specific to time-domain OCT, but it is applicable to other types of OCT systems.
© (2012) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.
Ameneh Boroomand ; Michael S. D. Smith ; Dan P. Popescu ; Michael Sowa and Sherif S. Sherif
Maximization of signal-to-noise ratio in optical coherence tomography using a depth-dependent matched filter
", Proc. SPIE 8412, Photonics North 2012, 841202 (October 23, 2012); doi:10.1117/12.2001691; http://dx.doi.org/10.1117/12.2001691