Implications of Spatial Correlation in Characterization and Clearance Sampling Design
Abstract not provided.
Abstract not provided.
Abstract not provided.
Abstract not provided.
The Bio-Restoration of Major Transportation Facilities Domestic Demonstration and Application Program (DDAP) is a designed to accelerate the restoration of transportation nodes following an attack with a biological warfare agent. This report documents the technology development work done at SNL for this DDAP, which include development of the BROOM tool, an investigation of surface sample collection efficiency, and a flow cytometry study of chlorine dioxide effects on Bacillus anthracis spore viability.
Abstract not provided.
In February of 2005, a joint exercise involving Sandia National Laboratories (SNL) and the National Institute for Occupational Safety and Health (NIOSH) was conducted in Albuquerque, NM. The SNL participants included the team developing the Building Restoration Operations and Optimization Model (BROOM), a software product developed to expedite sampling and data management activities applicable to facility restoration following a biological contamination event. Integrated data-collection, data-management, and visualization software improve the efficiency of cleanup, minimize facility downtime, and provide a transparent basis for reopening. The exercise was held at an SNL facility, the Coronado Club, a now-closed social club for Sandia employees located on Kirtland Air Force Base. Both NIOSH and SNL had specific objectives for the exercise, and all objectives were met.
Abstract not provided.
Abstract not provided.
Abstract not provided.
Abstract not provided.
Abstract not provided.
Abstract not provided.
Abstract not provided.
Abstract not provided.
Proposed for publication in Engineering Geology.
Abstract not provided.
Proposed for publication in Ground Water.
Abstract not provided.
Environmental and Ecological Statistics
Efficient and reliable unexploded ordnance (UXO) site characterization is needed for decisions regarding future land use. There are several types of data available at UXO sites and geophysical signal maps are one of the most valuable sources of information. Incorporation of such information into site characterization requires a flexible and reliable methodology. Geostatistics allows one to account for exhaustive secondary information (i.e.,, known at every location within the field) in many different ways. Kriging and logistic regression were combined to map the probability of occurrence of at least one geophysical anomaly of interest, such as UXO, from a limited number of indicator data. Logistic regression is used to derive the trend from a geophysical signal map, and kriged residuals are added to the trend to estimate the probabilities of the presence of UXO at unsampled locations (simple kriging with varying local means or SKlm). Each location is identified for further remedial action if the estimated probability is greater than a given threshold. The technique is illustrated using a hypothetical UXO site generated by a UXO simulator, and a corresponding geophysical signal map. Indicator data are collected along two transects located within the site. Classification performances are then assessed by computing proportions of correct classification, false positive, false negative, and Kappa statistics. Two common approaches, one of which does not take any secondary information into account (ordinary indicator kriging) and a variant of common cokriging (collocated cokriging), were used for comparison purposes. Results indicate that accounting for exhaustive secondary information improves the overall characterization of UXO sites if an appropriate methodology, SKlm in this case, is used. © Springer Science+Business Media, Inc. 2005.
Real-time water quality and chemical-specific sensors are becoming more commonplace in water distribution systems. The overall objective of the sensor network is to protect consumers from accidental and malevolent contamination events occurring within the distribution network. This objective can be quantified several different ways including: minimizing the amount of contaminated water consumed, minimizing the extent of the contamination within the network, minimizing the time to detection, etc. We examine the ability of a sensor network to meet these objectives as a function of both the detection limit of the sensors and the number of sensors in the network. A moderately-sized network is used as an example and sensors are placed randomly. The source term is a passive injection into a node and the resulting concentration in the node is a function of the volumetric flow through that node. The concentration of the contaminant at the source node is averaged for all time steps during the injection period. For each combination of a certain number of sensors and a detection limit, the mean values of the different objectives across multiple random sensor placements are evaluated. Results of this analysis allow the tradeoff between the necessary detection limit in a sensor and the number of sensors to be evaluated. Results show that for the example problem examined here, a sensor detection limit of 0.01 of the average source concentration is adequate for maximum protection.
Sandia National Laboratories, under contract to Nuclear Waste Management Organization of Japan (NUMO), is performing research on regional classification of given sites in Japan with respect to potential volcanic disruption using multivariate statistics and geo-statistical interpolation techniques. This report provides results obtained for hierarchical probabilistic regionalization of volcanism for the Sengan region in Japan by applying multivariate statistical techniques and geostatistical interpolation techniques on the geologic data provided by NUMO. A workshop report produced in September 2003 by Sandia National Laboratories (Arnold et al., 2003) on volcanism lists a set of most important geologic variables as well as some secondary information related to volcanism. Geologic data extracted for the Sengan region in Japan from the data provided by NUMO revealed that data are not available at the same locations for all the important geologic variables. In other words, the geologic variable vectors were found to be incomplete spatially. However, it is necessary to have complete geologic variable vectors to perform multivariate statistical analyses. As a first step towards constructing complete geologic variable vectors, the Universal Transverse Mercator (UTM) zone 54 projected coordinate system and a 1 km square regular grid system were selected. The data available for each geologic variable on a geographic coordinate system were transferred to the aforementioned grid system. Also the recorded data on volcanic activity for Sengan region were produced on the same grid system. Each geologic variable map was compared with the recorded volcanic activity map to determine the geologic variables that are most important for volcanism. In the regionalized classification procedure, this step is known as the variable selection step. The following variables were determined as most important for volcanism: geothermal gradient, groundwater temperature, heat discharge, groundwater pH value, presence of volcanic rocks and presence of hydrothermal alteration. Data available for each of these important geologic variables were used to perform directional variogram modeling and kriging to estimate values for each variable at 23949 centers of the chosen 1 km cell grid system that represents the Sengan region. These values formed complete geologic variable vectors at each of the 23,949 one km cell centers.
Multivariate spatial classification schemes such as regionalized classification or principal components analysis combined with kriging rely on all variables being collocated at the sample locations. In these approaches, classification of the multivariate data into a finite number of groups is done prior to the spatial estimation. However, in some cases, the variables may be sampled at different locations with the extreme case being complete heterotopy of the data set. In these situations, it is necessary to adapt existing techniques to work with non-collocated data. Two approaches are considered: (1) kriging of existing data onto a series of 'collection points' where the classification into groups is completed and a measure of the degree of group membership is kriged to all other locations; and (2) independent kriging of all attributes to all locations after which the classification is done at each location. Calculations are conducted using an existing groundwater chemistry data set in the upper Dakota aquifer in Kansas (USA) and previously examined using regionalized classification (Bohling, 1997). This data set has all variables measured at all locations. To test the ability of the first approach for dealing with non-collocated data, each variable is reestimated at each sample location through a cross-validation process and the reestimated values are then used in the regionalized classification. The second approach for non-collocated data requires independent kriging of each attribute across the entire domain prior to classification. Hierarchical and non-hierarchical classification of all vectors is completed and a computationally less burdensome classification approach, 'sequential discrimination', is developed that constrains the classified vectors to be chosen from those with a minimal multivariate kriging variance. Resulting classification and uncertainty maps are compared between all non-collocated approaches as well as to the original collocated approach. The non-collocated approaches lead to significantly different group definitions compared to the collocated case. To some extent, these differences can be explained by the kriging variance of the estimated variables. Sequential discrimination of locations with a minimum multivariate kriging variance constraint produces slightly improved results relative to the collection point and the non-hierarchical classification of the estimated vectors.
Preliminary investigation areas (PIA) for a potential repository of high-level radioactive waste must be evaluated by NUMO with regard to a number of qualifying factors. One of these factors is related to earthquakes and fault activity. This study develops a spatial statistical assessment method that can be applied to the active faults in Japan to perform such screening evaluations. This analysis uses the distribution of seismicity near faults to define the width of the associated process zone. This concept is based on previous observations of aftershock earthquakes clustered near active faults and on the assumption that such seismic activity is indicative of fracturing and associated impacts on bedrock integrity. Preliminary analyses of aggregate data for all of Japan confirmed that the frequency of earthquakes is higher near active faults. Data used in the analysis were obtained from NUMO and consist of three primary sources: (1) active fault attributes compiled in a spreadsheet, (2) earthquake hypocenter data, and (3) active fault locations. Examination of these data revealed several limitations with regard to the ability to associate fault attributes from the spreadsheet to locations of individual fault trace segments. In particular, there was no direct link between attributes of the active faults in the spreadsheet and the active fault locations in the GIS database. In addition, the hypocenter location resolution in the pre-1983 data was less accurate than for later data. These pre-1983 hypocenters were eliminated from further analysis.
Abstract not provided.
Abstract not provided.
This report describes the methodology and results of a project to develop a neural network for the prediction of the measured hydraulic conductivity or transmissivity in a series of boreholes at the Tono, Japan study site. Geophysical measurements were used as the input to EL feed-forward neural network. A simple genetic algorithm was used to evolve the architecture and parameters of the neural network in conjunction with an optimal subset of geophysical measurements for the prediction of hydraulic conductivity. The first attempt was focused on the estimation of the class of the hydraulic conductivity, high, medium or low, from the geophysical logs. This estimation was done while using the genetic algorithm to simultaneously determine which geophysical logs were the most important and optimizing the architecture of the neural network. Initial results showed that certain geophysical logs provided more information than others- most notably the 'short-normal', micro-resistivity, porosity and sonic logs provided the most information on hydraulic conductivity. The neural network produced excellent training results with accuracy of 90 percent or greater, but was unable to produce accurate predictions of the hydraulic conductivity class. The second attempt at prediction was done using a new methodology and a modified data set. The new methodology builds on the results of the first attempts at prediction by limiting the choices of geophysical logs to only those that provide significant information. Additionally, this second attempt uses a modified data set and predicts transmissivity instead of hydraulic conductivity. Results of these simulations indicate that the most informative geophysical measurements for the prediction of transmissivity are depth and sonic log. The long normal resistivity and self potential borehole logs are moderately informative. In addition, it was found that porosity and crack counts (clear, open, or hairline) do not inform predictions of hydraulic transmissivity.
Abstract not provided.
A conceptual model of the MIU site in central Japan, was developed to predict the groundwater system response to pumping. The study area consisted of a fairly large three-dimensional domain, having the size 4.24 x 6 x 3 km{sup 3} with three different geological units, upper and lower fractured zones and a single fault unit. The resulting computational model comprised of 702,204 finite difference cells with variable grid spacing. Both steady-state and transient simulations were completed to evaluate the influence of two different surface boundary conditions: fixed head and no flow. Steady state results were used for particle tracking and also serving as the initial conditions (i.e., starting heads) for the transient simulations. Results of the steady state simulations indicate the significance of the choice of surface (i.e., upper) boundary conditions and its effect on the groundwater flow patterns along the base of the upper fractured zone. Steady state particle tracking results illustrate that all particles exit the top of the model in areas where groundwater discharges to the Hiyoshi and Toki rivers. Particle travel times range from 3.6 x 10{sup 7} sec (i.e., {approx}1.1 years) to 4.4 x 10{sup 10} sec (i.e., {approx}1394 years). For the transient simulations, two pumping zones one above and another one below the fault are considered. For both cases, the pumping period extends for 14 days followed by an additional 36 days of recovery. For the pumping rates used, the maximum drawdown is quite small (ranging from a few centimeters to a few meters) and thus, pumping does not severely impact the groundwater flow system. The range of drawdown values produced by pumping below the fault are generally much less sensitive to the choice of the boundary condition than are the drawdowns resulted from the pumping zone above the fault.
Proposed for publication in the Bulletin of the Oklahoma Geological Survey.
Previous work on sample design has been focused on constructing designs for samples taken at point locations. Significantly less work has been done on sample design for data collected along transects. A review of approaches to point and transect sampling design shows that transects can be considered as a sequential set of point samples. Any two sampling designs can be compared through using each one to predict the value of the quantity being measured on a fixed reference grid. The quality of a design is quantified in two ways: computing either the sum or the product of the eigenvalues of the variance matrix of the prediction error. An important aspect of this analysis is that the reduction of the mean prediction error variance (MPEV) can be calculated for any proposed sample design, including one with straight and/or meandering transects, prior to taking those samples. This reduction in variance can be used as a ''stopping rule'' to determine when enough transect sampling has been completed on the site. Two approaches for the optimization of the transect locations are presented. The first minimizes the sum of the eigenvalues of the predictive error, and the second minimizes the product of these eigenvalues. Simulated annealing is used to identify transect locations that meet either of these objectives. This algorithm is applied to a hypothetical site to determine the optimal locations of two iterations of meandering transects given a previously existing straight transect. The MPEV calculation is also used on both a hypothetical site and on data collected at the Isleta Pueblo to evaluate its potential as a stopping rule. Results show that three or four rounds of systematic sampling with straight parallel transects covering 30 percent or less of the site, can reduce the initial MPEV by as much as 90 percent. The amount of reduction in MPEV can be used as a stopping rule, but the relationship between MPEV and the results of excavation versus no-further-action decisions is site specific and cannot be calculated prior to the sampling. It may be advantageous to use the reduction in MPEV as a stopping rule for systematic sampling across the site that can then be followed by focused sampling in areas identified has having UXO during the systematic sampling. The techniques presented here provide answers to the questions of ''Where to sample?'' and ''When to stop?'' and are capable of running in near real time to support iterative site characterization campaigns.
Groundwater Quality Modeling and Management Under Uncertinity
Abstract not provided.
Variations in water use at short time scales, seconds to minutes, produce variation in transport of solutes through a water supply network. However, the degree to which short term variations in demand influence the solute concentrations at different locations in the network is poorly understood. Here we examine the effect of variability in demand on advective transport of a conservative solute (e.g. chloride) through a water supply network by defining the demand at each node in the model as a stochastic process. The stochastic demands are generated using a Poisson rectangular pulse (PRP) model for the case of a dead-end water line serving 20 homes represented as a single node. The simple dead-end network model is used to examine the variation in Reynolds number, the proportion of time that there is no flow (i.e., stagnant conditions, in the pipe) and the travel time defined as the time for cumulative demand to equal the volume of water in 1000 feet of pipe. Changes in these performance measures are examined as the fine scale demand functions are aggregated over larger and larger time scales. Results are compared to previously developed analytical expressions for the first and second moments of these three performance measures. A new approach to predict the reduction in variance of the performance measures based on perturbation theory is presented and compared to the results of the numerical simulations. The distribution of travel time is relatively consistent across time scales until the time step approaches that of the travel time. However, the proportion of stagnant flow periods decreases rapidly as the simulation time step increases. Both sets of analytical expressions are capable of providing adequate, first-order predictions of the simulation results.
As demonstrated by the anthrax attack through the United States mail, people infected by the biological agent itself will give the first indication of a bioterror attack. Thus, a distributed information system that can rapidly and efficiently gather and analyze public health data would aid epidemiologists in detecting and characterizing emerging diseases, including bioterror attacks. We propose using clusters of adverse health events in space and time to detect possible bioterror attacks. Space-time clusters can indicate exposure to infectious diseases or localized exposure to toxins. Most space-time clustering approaches require individual patient data. To protect the patient's privacy, we have extended these approaches to aggregated data and have embedded this extension in a sequential probability ratio test (SPRT) framework. The real-time and sequential nature of health data makes the SPRT an ideal candidate. The result of space-time clustering gives the statistical significance of a cluster at every location in the surveillance area and can be thought of as a ''health-index'' of the people living in this area. As a surrogate to bioterrorism data, we have experimented with two flu data sets. For both databases, we show that space-time clustering can detect a flu epidemic up to 21 to 28 days earlier than a conventional periodic regression technique. We have also tested using simulated anthrax attack data on top of a respiratory illness diagnostic category. Results show we do very well at detecting an attack as early as the second or third day after infected people start becoming severely symptomatic.
Abstract not provided.
Abstract not provided.
Abstract not provided.
Abstract not provided.
Geostatistical simulation is used to extrapolate data derived from site characterization activities at the MIU site into information describing the three-dimensional distribution of hydraulic conductivity at the site and the uncertainty in the estimates of hydraulic conductivity. This process is demonstrated for six different data sets representing incrementally increasing amounts of characterization data. Short horizontal ranges characterize the spatial variability of both the rock types (facies) and the hydraulic conductivity measurements. For each of the six data sets, 50 geostatistical realizations of the facies and 50 realizations of the hydraulic conductivity are combined to produce 50 final realizations of the hydraulic conductivity distribution. Analysis of these final realizations indicates that the mean hydraulic conductivity value increases with the addition of site characterization data. The average hydraulic conductivity as a function of elevation changes from a uniform profile to a profile showing relatively high hydraulic conductivity values near the top and bottom of the simulation domain. Three-dimensional uncertainty maps show the highest amount of uncertainty in the hydraulic conductivity distribution near the top and bottom of the model. These upper and lower areas of high uncertainty are interpreted to be due to the unconformity at the top of the granitic rocks and the Tsukyoshi fault respectively.
The goal of this project is to predict the drawdown that will be observed in specific piezometers placed in the MIU-2 borehole due to pumping at a single location in the MIU-3 borehole. These predictions will be in the form of distributions obtained through multiple forward runs of a well-test model. Specifically, two distributions will be created for each pumping location--piezometer location pair: (1) the distribution of the times to 1.0 meter of drawdown and (2) the distribution of the drawdown predicted after 12 days of pumping at a discharge rates of 25, 50, 75 and 100 l/hr. Each of the steps in the pumping rate lasts for 3 days (259,200 seconds). This report is based on results that were presented at the Tono Geoscience Center on January 27th, 2000, which was approximately one week prior to the beginning of the interference tests. Hydraulic conductivity (K), specific storage (S{sub s}) and the length of the pathway (L{sub p}) are the input parameters to the well-test analysis model. Specific values of these input parameters are uncertain. This parameter uncertainty is accounted for in the modeling by drawing individual parameter values from distributions defined for each input parameter. For the initial set of runs, the fracture system is assumed to behave as an infinite, homogeneous, isotropic aquifer. These assumptions correspond to conceptualizing the aquifer as having Theis behavior and producing radial flow to the pumping well. A second conceptual model is also used in the drawdown calculations. This conceptual model considers that the fracture system may cause groundwater to move to the pumping well in a more linear (non-radial) manner. The effects of this conceptual model on the drawdown values are examined by casting the flow dimension (F{sub d}) of the fracture pathways as an uncertain variable between 1.0 (purely linear flow) and 2.0 (completely radial flow).
Water Resources Research
We investigated the late-time (asymptotic) behavior of tracer test breakthrough curves (BTCs) with rate-limited mass transfer (e.g., in dual-porosity or multiporosity systems) and found that the late-time concentration c is given by the simple expression c = tad{c0g - [m0(∂g/∂t)]}, for t ≫ tad and tα ≫ tad, where tad is the advection time, c0 is the initial concentration in the medium, m0 is the zeroth moment of the injection pulse, and tα is the mean residence time in the immobile domain (i.e., the characteristic mass transfer time). The function g is proportional to the residence time distribution in the immobile domain; we tabulate g for many geometries, including several distributed (multirate) models of mass transfer. Using this expression, we examine the behavior of late-time concentration for a number of mass transfer models. One key result is that if rate-limited mass transfer causes the BTC to behave as a power law at late time (i.e., c ̃ t-k), then the underlying density function of rate coefficients must also be a power law with the form αk-3 as α → 0. This is true for both density functions of first-order and diffusion rate coefficients. BTCs with k < 3 persisting to the end of the experiment indicate a mean residence time longer than the experiment, and possibly an infinite residence time, and also suggest an effective rate coefficient that is either undefined or changes as a function of observation time. We apply our analysis to breakthrough curves from single-well injection-withdrawal tests at the Waste Isolation Pilot Plant, New Mexico. We investigated the late-time (asymptotic) behavior of tracer test breakthrough curves (BTCs) with rate-limited mass transfer (e.g., in dual-porosity or multiporosity systems) and found that the late-time concentration c is given by the simple expression c = tad{c0g - [m0(∂g/∂t)]}, for t ≫ tad and tα ≫ t ad, where tad is the advection time, c0 is the initial concentration in the medium, m0 is the zeroth moment of the injection pulse, and tα is the mean residence time in the immobile domain (i.e., the characteristic mass transfer time). The function g is proportional to the residence time distribution in the immobile domain; we tabulate g for many geometries, including several distributed (multirate) models of mass transfer. Using this expression, we examine the behavior of late-time concentration for a number of mass transfer models. One key result is that if rate-limited mass transfer causes the BTC to behave as a power law at late time (i.e., c t-k), then the underlying density function of rate coefficients must also be a power law with the form αk-3 as α → 0. This is true for both density functions of first-order and diffusion rate coefficients. BTCs with k < 3 persisting to the end of the experiment indicate a mean residence time longer than the experiment, and possibly an infinite residence time, and also suggest an effective rate coefficient that is either undefined or changes as a function of observation time. We apply our analysis to breakthrough curves from single-well injection-withdrawal tests at the Waste Isolation Pilot Plant, New Mexico.
Multiple techniques have been developed to model the temporal evolution of infectious diseases. Some of these techniques have also been adapted to model the spatial evolution of the disease. This report examines the application of one such technique, the SEIR model, to the spatial and temporal evolution of disease. Applications of the SEIR model are reviewed briefly and an adaptation to the traditional SEIR model is presented. This adaptation allows for modeling the spatial evolution of the disease stages at the individual level. The transmission of the disease between individuals is modeled explicitly through the use of exposure likelihood functions rather than the global transmission rate applied to populations in the traditional implementation of the SEIR model. These adaptations allow for the consideration of spatially variable (heterogeneous) susceptibility and immunity within the population. The adaptations also allow for modeling both contagious and non-contagious diseases. The results of a number of numerical experiments to explore the effect of model parameters on the spread of an example disease are presented.
For the last ten years, the Japanese High-Level Nuclear Waste (HLW) repository program has focused on assessing the feasibility of a basic repository concept, which resulted in the recently published H12 Report. As Japan enters the implementation phase, a new organization must identify, screen and choose potential repository sites. Thus, a rapid mechanism for determining the likelihood of site suitability is critical. The threshold approach, described here, is a simple mechanism for defining the likelihood that a site is suitable given estimates of several critical parameters. We rely on the results of a companion paper, which described a probabilistic performance assessment simulation of the HLW reference case in the H12 report. The most critical two or three input parameters are plotted against each other and treated as spatial variables. Geostatistics is used to interpret the spatial correlation, which in turn is used to simulate multiple realizations of the parameter value maps. By combining an array of realizations, we can look at the probability that a given site, as represented by estimates of this combination of parameters, would be good host for a repository site.
Abstract not provided.