The quality of the data analyzed was also affected strongly by the various problems encountered in taking samples in the field. For example, the surface soil sampling design was quite complex to execute in the field, and it took time for a new sampling crew to learn to take these samples properly. Also, the equipment had a tendency to deteriorate or be altered inadvertently when parts were replaced, so that later data may not have represented the same population as earlier data. The primary result of these and similar field problems was to increase the sample variance, making comparisons among data sets more difficult. Data quality was unavoidably altered to an unknown extent by the engineering operations that were necessary to allow data collection. For instance, if the vegetation were removed totally, as on Janet, the resulting soil disturbance altered the distribution of the TRU activity in the soil. If only aecess lanes were cut, as on other islands, soil disturbance was reduc.Jj but not eliminated. In addition, the data had to be corrected for signal attenuation from the remaining brush, using a subjective estimate of the amount of brush and an empirical brush correction factor. Because of these factors, the general principle used for choosing between alternative statistical approaches was to use the simplest method that would do the job. Certain types of data that were reported by others to the statistician were accepted as accurate because there was no way to verify the information. Examples are the total volumes of soil removed, the nominal depth of soil profile samples taken where the surface was uneven, actual boundaries of soil lifts, brush cover estimates, and similar information. No estimates of variance or reliability could be made for such data, so they were accepted at face value. Cleanup Criteria. The cleanup criteria were stated as averages over specified areas such as 0.25 ha, and specified depth intervals such as 0-3 em. Therefore the statistical methods used had to be appropriate for making estimates of area averages for a given depth interval. Also, the criteria required that the estimation error be considered, so an estimate of the error also had to be made. However, it was not clear at the beginning of the project whether the criteria applied to upper bounds or lower bounds on the estimates. The conservative approach of applying the criteria to the upper bounds was actually used, that is, soil was removed if the estimate plus half its standard deviation exceeded the applicable criterion. The subsurface cleanup criterion was difficult to interpret. Eventually the criterion was restated to reflect the limitations of the subsurface data, so the statistical analysis could aim at locating boundaries of areas to be cleaned rather than estimating subsurface averages. In some instances, though, estimating averages were necessary. For example, the criterion implies that the shallowest 5 em subsurface increment is 2.5 -7.5 em, but this interval was never sampled as such. Therefore, the average in this interval had to be estimated from 0-5 em and 5-10 em data. The method used to estimate the 2.5 - 7.5 em averageis deseribed in Tech Note 19.0. As the cleanup progressed, changes were made in the interpretation of various surface criteria. For more details concerning these changes, see Section 2.2.4. Both the area averaged over and the acceptable average value were altered. This meant that all the statistical analyses had to be flexible enough to allow estimates to be made for different sized areas and compared to various criteria levels. Fortunately, the kriging technique is quite flexible, so the original 50 m data could still be used. In those areas with 25 m data, it was relatively straightforward to compute the arithmetic means for various size areas. 5.3 DATA HANDLING Data handling responsibilities during the Enewetak cleanup project included not only statistical analyses but also data base management, data quality assurance and preservation, and the display of results in clear, useful forms. The types of information involved included not only raw data and final results, but also intermediate results, narrative descriptions of statistical methods, documentation for computer programs, ete. The onsite DRI statistician, assisted by the Navy data technician, had primary responsibility for data responsibility of DRI-Las Vegas. handling on-island. 148 Long-term data preservation was the

Select target paragraph3