31 August 2000 Influence of self-absorption on the performance of laser-induced breakdown spectroscopy (LIBS)
Author Affiliations +
Proceedings Volume 4076, Optical Diagnostics for Industrial Applications; (2000) https://doi.org/10.1117/12.397959
Event: Symposium on Applied Photonics, 2000, Glasgow, United Kingdom
Abstract
LIBS is based on atomic emission from plasma formed by laser ablation and excitation. It offers non-contacting and nearly non-destructive elemental analysis, but limited analytical accuracy. An empirical power- law calibration curve is usually required. From our work and from recent work by Gornushkin et al. this arises from self-absorption. Assuming Local Thermal Equilibrium (LTE), irradiance is found from integrals over the Voigt profile, which we compute using the complex error function. Calibration curves show a break between linear and power-law regions, with a square-root dependence at high concentrations. Irradiance depends on width of the Lorentz (pressure broadened) component, and the simple Boltzmann temperature dependence is modified. Gornushkin et al extended calibration curves into the linear region, obtaining the Voigt parameter, but more typically this region is inaccessible. Self- absorption theory should provide improved temperature measurement in the power-law region and, although absolute concentration determination requires the Lorentz width, its known temperature and pressure dependence should reduce the ad-hoc nature of calibration curves.
© (2000) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.
Michael A. Player, John Watson, Jolyon M. O. De Freitas, "Influence of self-absorption on the performance of laser-induced breakdown spectroscopy (LIBS)", Proc. SPIE 4076, Optical Diagnostics for Industrial Applications, (31 August 2000); doi: 10.1117/12.397959; https://doi.org/10.1117/12.397959
PROCEEDINGS
9 PAGES


SHARE
Back to Top