Marco Ferretti
TerraData environmetrics at the Department of Environmental Sciences, University of Siena, Italy
![]() | Marco Ferretti is Technical Director at TerraData environmetrics, a spin-off company of the University of Siena, and lecturer at the same University. He obtained the degree in Forest Sciences at the University of Firenze and received his PhD from the University of Siena. Dr Ferretti has done consultancy work for several Italian and international agencies and has held previous assignments at IUFRO (the International Union of Forest Research Organizations) and UN-ECE ICP-Forests (International Co-operative Programme on Assessment and Monitoring Air Pollution Effects on Forests); for the latter he now serves as chairman of the Quality Assurance Committee. Dr Ferretti has over 200 scientific publications and presentations and has contributed to over 100 national and international reports in the areas of forest and environmental monitoring, and quality assurance. His main scientific interest is on the impact of environmental stressors on forests and to ensure that monitoring is properly designed, implemented and reported. |
While an unprecedented effort in ecological monitoring has been expended since the 1980s, the degree to which data were conclusive in the context of environmental management is controversial.5 Although this may be due to several reasons, a possible explanation needs to consider the actual ability of monitoring results to fit monitoring objectives. Several authors provided convincing evidence that monitoring results can be seriously affected by lack of design and sound statistical concept5,6 and poor data quality.7,8 Based on this concern, the Workshop on “Quality Assurance in environmental data: to what extent environmental monitoring data are reliable?” was organized in Siena, Italy, on 7th March 2008. Although not all possible ecological monitoring fields were covered (for example there were no presentations on issues such as birds, mammals or the marine environment), the workshop was intended to provide concrete examples of possible problems arising with monitoring data, from design to data management, and to suggest a working perspective for improving the value and use of monitoring data. In this special issue, several papers arising from the above Workshop are presented.
(i) sampling errors. They originate from unsuited sampling design that does not represent the population of interest well enough. As reported by Fattorini,10 sampling errors can be estimated and documented. However the control and management of sampling errors depends on the inferential approach (design-based vs. model-based), and/or the sampling concept adopted (probabilistic vs. judgemental).11 When, for example, haphazard sampling is misused in place of probabilistic sampling, not only can sampling error not be controlled, but it also may remain unknown. Yet, probabilistic sampling design is often disregarded in ecological monitoring programmes either because it is perceived as impractical or considered as a sort of exercise of “statistical philosophy”, with little connection with the “real world” of monitoring. This is unfortunate as probabilistic sampling is essential to obtain assumption-free estimates of population parameters (such as totals, mean and variance) and for change detection.3,6,12 The reader is referred to the fundamental textbook by Cochran13 for the basis of sampling. In this special issue, questions related to sampling are presented by Baldaccini et al. for macroinvertabrates monitoring in freshwaters and by Gottardini et al. for pollen quantity and diversity estimates in aerobiological monitoring. In the former, the Authors reported on an exercise based on judgemental sampling and concentrated on sampling efforts in relation to species diversity, abundance and objectivity of the investigation. In the latter, Gottardini et al. evaluated the bias and error arising in pollen counts when the official Italian standard technique is applied. They discussed the assumptions inherent in the standard technique in relation to their impact on the results of pollen counts both in terms of number of grains and in number of species.
(ii) Assessment errors or observer errors. They include measurement and classification errors and are rooted in how Standard Operating Procedures (SOPs) for field and laboratory measurements were prepared and applied, and how personnel were trained and prepared. Measurement errors can be random or one-sided (biased) and can have a serious impact on survey estimators14 and change/trend detection.7 However, they can only be controlled by adequate SOPs, proper training and timely audits, and field checks.15 In this special issue, observer and measurement errors are discussed in relation to a variety of issues: diversity of vascular plants and lichens (Allegrini et al.; Bacaro et al.; Giordani et al.), pollen counting in aerobiology (Berti et al.), forest health assessment and forest inventory (Bussotti et al.; Gasparini et al.), biomonitoring of air pollution (Francini et al.) and chemistry of atmospheric deposition (Marchetto et al.). Despite covering different environmental issues, targets and monitoring techniques (e.g., ground surveys, surveys based on aerial imagery, laboratory analyses), all the above referred papers agreed on some points: (a) SOPs are essential to promote consistency of measurements; (b) continuous training and control is necessary to reduce variability between observers/laboratories; (c) it is important to define data quality objectives in order to have a formal assessment and documentation of the achieved data quality; (d) the use of internal quality tools—e.g., blanks, control charts—promote the achievement of the expected data quality objectives.
(iii) Prediction errors caused by models. These errors occur because (a) measurement errors can propagate in the model output; (b) models are applied for data ranges not covered in the construction of the model; (c) model assumption, that may be attained or not. Models are also necessary when—for example—unsuited sampling renders it difficult to use the data and to assess the value of the information arising from the monitoring. Model-related issues are presented by Gorelli et al. in relation to the validation of the results obtained from an ozone biomonitoring network. They carried out a formal spatial analysis of an ozone biomonitoring dataset based on a preferential selection of sampling sites. The analysis permitted the formal identification of problems in a monitoring network (site distribution over the study area), to quantify and to control the data estimation error in the points not sampled, and therefore to evaluate the strength of the assumption needed to endorse results about spatial and temporal comparisons.
(iv) Non-statistical errors. They include a variety of events: registration errors, errors in data entry, data transfer, errors in calculations are just examples. These errors may occur at every stage, can be serious and cannot be fully controlled. Durrant Houston and Hiederer reported a case-study about data validation and management from an international forest monitoring programme. They concentrated on the various problems data managers have to face (unclear submission rules, unsuited formats, ambiguous relations between tables, missing values, unplausible values, inconsistent values) and on the relevant QA procedures adopted. They emphasised the need for a full integration of data management issues when defining the procedure for field data collection.
Another type of error is represented by definition errors. They are particularly important with large-scale monitoring programmes involving expertise with different education over different countries. Or when data originating from different programmes are combined to derive statistics for the integrated dataset. In these cases, different definitions adopted to identify the same object (and the opposite) may lead to very inconsistent results.9 A particular aspect of problems caused by differences in definition is related to taxonomy, and taxonomic-related problems are discussed by Bacaro et al. in relation to plant species diversity assessment and monitoring. Apart from observer skill, “taxonomic inflation” is also a source of inconsistency between datasets and may pose a severe threat to the long-term and large-scale comparability of biodiversity data.
(i) Identification of the right question the monitoring should answer. This has to be done in close-co-operation with end users of the monitoring results and will facilitate the subsequent phases.
(ii) Definition of unambiguous objectives. Monitoring objectives should be explicit in order to allow conclusive statements about the success of the monitoring and the problem being investigated. If the objective is a population estimate, the required precision level (in terms of width of confidence interval and probability level) should be reported. If the objective is detecting change, the time frame for change detection, the minimum detectable change, the acceptable risk for Type I and Type II errors should be explicit.3,5 Once the “right question” is known, the definition of the objectives can be carried out and graded according to the importance of the question to be answered and the available resources.
(iii) Selection of proper attributes to be measured. Attributes should be selected according to their nature, ease of measurement, known performance and responsiveness to the objective being targeted.18
(iv) Identification of the appropriate sampling strategy. Formal sampling design allows the control and management of sampling errors, thus helping in achievement of the monitoring objectives. It is worth noting that without a probabilistic approach there is no warrant that the results obtained represent the population characteristics. In addition, all the traditional data analysis techniques assume data are originated from probabilistic sampling.3,17
(v) Preparation of adequate Standard Operating Procedures (SOPs). This includes a comprehensive description of data collection activity and covers field as well as laboratory and office activity. Definitions of terms, methods of measurements, range of application, equipment and reporting units, Data Quality Objectives (DQOs—see below), field forms, hardware and software descriptions are typically covered in the SOPs.
(vi) Formal identification of the data quality objectives (DQOs) in order to document the degree to which measurements fit into an explicit acceptable range of variation. In general, DQOs consist of an expressed level of accuracy for each measurement, called Measurement Quality Objectives (MQOs) and of a compliance threshold, termed Data Quality Limits (DQLs).19
(vii) Alongside, identify safe rules for data submission, checks, validation, storage and management. This should be something to be considered across all the design and planning phase.
A formal QA plan covering all the QA activities involved with the acquisition of ecological data (from direct measurements as well as from other sources) is important and very useful in this context.17 At the QA Workshop in Siena there was quite an agreement about the need for a comprehensive QA approach, and an overall QA-based perspective with the relevant QA plan was suggested to be considered for adoption for monitoring programmes in Italy.20 Ideally, such a plan should be required by funding agencies before a grant is assigned to a monitoring programme.
Workshop organization. I am grateful to the other members of the organizing committee: Giorgio Brunialti (TerraData environmetrics, Siena, Italy), Alessandro Chiarucci (University of Siena, Siena, Italy), Paolo Giordani (University of Genova, Genova, Italy), Elena Gottardini (FEM-IASMA, Trento, Italy) and Maurizio Perotti (ISMES Divisione Ambiente e Territorio di CESI S.p.A Unità operativa Atmosfera, Piacenza, Italy). Elisa Baragatti (University of Siena, Siena, Italy), Elisa Santi (University of Siena, Siena, Italy), Arianna Vannini (TerraData environmetrics, Siena, Italy) helped in the organization of the Workshop. I acknowledge also all the colleagues that gave the presentations at the Workshop and submitted manuscripts for this special issue.
Editorials. I am indebted to Tracy Durrant Houston (JRC Ispra, Italy) for the linguistic revision of this paper. Tracy also reviewed other papers of this special issue and I want to thank her also on behalf of the relevant authors. I would also thank Dr Harpal Minhas (editor, the Journal of Environmental Monitoring) for his support and assistance in the preparation of this special issue.
Support. On behalf of the organizing committee of the Workshop, I would like to thank SET S.p.A. Servizi Energetici Teverola and ISMES Divisione Ambiente e Territorio di CESI S.p.A for the financial support. The Italian national Environmental Protection Agency (formerly APAT, now ISPRA) supported the organization of the event.
This journal is © The Royal Society of Chemistry 2009 |