Accuracy of methods for reporting inorganic element concentrations and radioactivity in oil and gas wastewaters from the Appalachian Basin, U.S. based on an inter-laboratory comparison.

Autor: Tasker TL; Department of Civil and Environmental Engineering, The Pennsylvania State University, 212 Sackett Building, University Park, Pennsylvania 16802, USA. ttasker@francis.edu., Burgos WD, Ajemigbitse MA, Lauer NE, Gusa AV, Kuatbek M, May D, Landis JD, Alessi DS, Johnsen AM, Kaste JM, Headrick KL, Wilke FDH, McNeal M, Engle M, Jubb AM, Vidic RD, Vengosh A, Warner NR
Jazyk: angličtina
Zdroj: Environmental science. Processes & impacts [Environ Sci Process Impacts] 2019 Feb 21; Vol. 21 (2), pp. 224-241.
DOI: 10.1039/c8em00359a
Abstrakt: Accurate and precise analyses of oil and gas (O&G) wastewaters and solids (e.g., sediments and sludge) are important for the regulatory monitoring of O&G development and tracing potential O&G contamination in the environment. In this study, 15 laboratories participated in an inter-laboratory comparison on the chemical characterization of three O&G wastewaters from the Appalachian Basin and four solids impacted by O&G development, with the goal of evaluating the quality of data and the accuracy of measurements for various analytes of concern. Using a variety of different methods, analytes in the wastewaters with high concentrations (i.e., >5 mg L-1) were easily detectable with relatively high accuracy, often within ±10% of the most probable value (MPV). In contrast, often less than 7 of the 15 labs were able to report detectable trace metal(loid) concentrations (i.e., Cr, Ni, Cu, Zn, As, and Pb) with accuracies of approximately ±40%. Despite most labs using inductively coupled plasma mass spectrometry (ICP-MS) with low instrument detection capabilities for trace metal analyses, large dilution factors during sample preparation and low trace metal concentrations in the wastewaters limited the number of quantifiable determinations and likely influenced analytical accuracy. In contrast, all the labs measuring Ra in the wastewaters were able to report detectable concentrations using a variety of methods including gamma spectroscopy and wet chemical approaches following Environmental Protection Agency (EPA) standard methods. However, the reported radium activities were often greater than ±30% different to the MPV possibly due to calibration inconsistencies among labs, radon leakage, or failing to correct for self-attenuation. Reported radium activities in solid materials had less variability (±20% from MPV) but accuracy could likely be improved by using certified radium standards and accounting for self-attenuation that results from matrix interferences or a density difference between the calibration standard and the unknown sample. This inter-laboratory comparison illustrates that numerous methods can be used to measure major cation, minor cation, and anion concentrations in O&G wastewaters with relatively high accuracy while trace metal(loid) and radioactivity analyses in liquids may often be over ±20% different from the MPV.
Databáze: MEDLINE