Sample records for estimating large complex

  1. Hybrid estimation of complex systems.

    PubMed

    Hofbaur, Michael W; Williams, Brian C

    2004-10-01

    Modern automated systems evolve both continuously and discretely, and hence require estimation techniques that go well beyond the capability of a typical Kalman Filter. Multiple model (MM) estimation schemes track these system evolutions by applying a bank of filters, one for each discrete system mode. Modern systems, however, are often composed of many interconnected components that exhibit rich behaviors, due to complex, system-wide interactions. Modeling these systems leads to complex stochastic hybrid models that capture the large number of operational and failure modes. This large number of modes makes a typical MM estimation approach infeasible for online estimation. This paper analyzes the shortcomings of MM estimation, and then introduces an alternative hybrid estimation scheme that can efficiently estimate complex systems with large number of modes. It utilizes search techniques from the toolkit of model-based reasoning in order to focus the estimation on the set of most likely modes, without missing symptoms that might be hidden amongst the system noise. In addition, we present a novel approach to hybrid estimation in the presence of unknown behavioral modes. This leads to an overall hybrid estimation scheme for complex systems that robustly copes with unforeseen situations in a degraded, but fail-safe manner.

  2. Low-Complexity Polynomial Channel Estimation in Large-Scale MIMO With Arbitrary Statistics

    NASA Astrophysics Data System (ADS)

    Shariati, Nafiseh; Bjornson, Emil; Bengtsson, Mats; Debbah, Merouane

    2014-10-01

    This paper considers pilot-based channel estimation in large-scale multiple-input multiple-output (MIMO) communication systems, also known as massive MIMO, where there are hundreds of antennas at one side of the link. Motivated by the fact that computational complexity is one of the main challenges in such systems, a set of low-complexity Bayesian channel estimators, coined Polynomial ExpAnsion CHannel (PEACH) estimators, are introduced for arbitrary channel and interference statistics. While the conventional minimum mean square error (MMSE) estimator has cubic complexity in the dimension of the covariance matrices, due to an inversion operation, our proposed estimators significantly reduce this to square complexity by approximating the inverse by a L-degree matrix polynomial. The coefficients of the polynomial are optimized to minimize the mean square error (MSE) of the estimate. We show numerically that near-optimal MSEs are achieved with low polynomial degrees. We also derive the exact computational complexity of the proposed estimators, in terms of the floating-point operations (FLOPs), by which we prove that the proposed estimators outperform the conventional estimators in large-scale MIMO systems of practical dimensions while providing a reasonable MSEs. Moreover, we show that L needs not scale with the system dimensions to maintain a certain normalized MSE. By analyzing different interference scenarios, we observe that the relative MSE loss of using the low-complexity PEACH estimators is smaller in realistic scenarios with pilot contamination. On the other hand, PEACH estimators are not well suited for noise-limited scenarios with high pilot power; therefore, we also introduce the low-complexity diagonalized estimator that performs well in this regime. Finally, we ...

  3. Statistical Analysis of Big Data on Pharmacogenomics

    PubMed Central

    Fan, Jianqing; Liu, Han

    2013-01-01

    This paper discusses statistical methods for estimating complex correlation structure from large pharmacogenomic datasets. We selectively review several prominent statistical methods for estimating large covariance matrix for understanding correlation structure, inverse covariance matrix for network modeling, large-scale simultaneous tests for selecting significantly differently expressed genes and proteins and genetic markers for complex diseases, and high dimensional variable selection for identifying important molecules for understanding molecule mechanisms in pharmacogenomics. Their applications to gene network estimation and biomarker selection are used to illustrate the methodological power. Several new challenges of Big data analysis, including complex data distribution, missing data, measurement error, spurious correlation, endogeneity, and the need for robust statistical methods, are also discussed. PMID:23602905

  4. Accuracy assessment with complex sampling designs

    Treesearch

    Raymond L. Czaplewski

    2010-01-01

    A reliable accuracy assessment of remotely sensed geospatial data requires a sufficiently large probability sample of expensive reference data. Complex sampling designs reduce cost or increase precision, especially with regional, continental and global projects. The General Restriction (GR) Estimator and the Recursive Restriction (RR) Estimator separate a complex...

  5. A Project Management Approach to Using Simulation for Cost Estimation on Large, Complex Software Development Projects

    NASA Technical Reports Server (NTRS)

    Mizell, Carolyn; Malone, Linda

    2007-01-01

    It is very difficult for project managers to develop accurate cost and schedule estimates for large, complex software development projects. None of the approaches or tools available today can estimate the true cost of software with any high degree of accuracy early in a project. This paper provides an approach that utilizes a software development process simulation model that considers and conveys the level of uncertainty that exists when developing an initial estimate. A NASA project will be analyzed using simulation and data from the Software Engineering Laboratory to show the benefits of such an approach.

  6. A channel estimation scheme for MIMO-OFDM systems

    NASA Astrophysics Data System (ADS)

    He, Chunlong; Tian, Chu; Li, Xingquan; Zhang, Ce; Zhang, Shiqi; Liu, Chaowen

    2017-08-01

    In view of the contradiction of the time-domain least squares (LS) channel estimation performance and the practical realization complexity, a reduced complexity channel estimation method for multiple input multiple output-orthogonal frequency division multiplexing (MIMO-OFDM) based on pilot is obtained. This approach can transform the complexity of MIMO-OFDM channel estimation problem into a simple single input single output-orthogonal frequency division multiplexing (SISO-OFDM) channel estimation problem and therefore there is no need for large matrix pseudo-inverse, which greatly reduces the complexity of algorithms. Simulation results show that the bit error rate (BER) performance of the obtained method with time orthogonal training sequences and linear minimum mean square error (LMMSE) criteria is better than that of time-domain LS estimator and nearly optimal performance.

  7. Approximate median regression for complex survey data with skewed response.

    PubMed

    Fraser, Raphael André; Lipsitz, Stuart R; Sinha, Debajyoti; Fitzmaurice, Garrett M; Pan, Yi

    2016-12-01

    The ready availability of public-use data from various large national complex surveys has immense potential for the assessment of population characteristics using regression models. Complex surveys can be used to identify risk factors for important diseases such as cancer. Existing statistical methods based on estimating equations and/or utilizing resampling methods are often not valid with survey data due to complex survey design features. That is, stratification, multistage sampling, and weighting. In this article, we accommodate these design features in the analysis of highly skewed response variables arising from large complex surveys. Specifically, we propose a double-transform-both-sides (DTBS)'based estimating equations approach to estimate the median regression parameters of the highly skewed response; the DTBS approach applies the same Box-Cox type transformation twice to both the outcome and regression function. The usual sandwich variance estimate can be used in our approach, whereas a resampling approach would be needed for a pseudo-likelihood based on minimizing absolute deviations (MAD). Furthermore, the approach is relatively robust to the true underlying distribution, and has much smaller mean square error than a MAD approach. The method is motivated by an analysis of laboratory data on urinary iodine (UI) concentration from the National Health and Nutrition Examination Survey. © 2016, The International Biometric Society.

  8. Approximate Median Regression for Complex Survey Data with Skewed Response

    PubMed Central

    Fraser, Raphael André; Lipsitz, Stuart R.; Sinha, Debajyoti; Fitzmaurice, Garrett M.; Pan, Yi

    2016-01-01

    Summary The ready availability of public-use data from various large national complex surveys has immense potential for the assessment of population characteristics using regression models. Complex surveys can be used to identify risk factors for important diseases such as cancer. Existing statistical methods based on estimating equations and/or utilizing resampling methods are often not valid with survey data due to complex survey design features. That is, stratification, multistage sampling and weighting. In this paper, we accommodate these design features in the analysis of highly skewed response variables arising from large complex surveys. Specifically, we propose a double-transform-both-sides (DTBS) based estimating equations approach to estimate the median regression parameters of the highly skewed response; the DTBS approach applies the same Box-Cox type transformation twice to both the outcome and regression function. The usual sandwich variance estimate can be used in our approach, whereas a resampling approach would be needed for a pseudo-likelihood based on minimizing absolute deviations (MAD). Furthermore, the approach is relatively robust to the true underlying distribution, and has much smaller mean square error than a MAD approach. The method is motivated by an analysis of laboratory data on urinary iodine (UI) concentration from the National Health and Nutrition Examination Survey. PMID:27062562

  9. Tail-scope: Using friends to estimate heavy tails of degree distributions in large-scale complex networks

    NASA Astrophysics Data System (ADS)

    Eom, Young-Ho; Jo, Hang-Hyun

    2015-05-01

    Many complex networks in natural and social phenomena have often been characterized by heavy-tailed degree distributions. However, due to rapidly growing size of network data and concerns on privacy issues about using these data, it becomes more difficult to analyze complete data sets. Thus, it is crucial to devise effective and efficient estimation methods for heavy tails of degree distributions in large-scale networks only using local information of a small fraction of sampled nodes. Here we propose a tail-scope method based on local observational bias of the friendship paradox. We show that the tail-scope method outperforms the uniform node sampling for estimating heavy tails of degree distributions, while the opposite tendency is observed in the range of small degrees. In order to take advantages of both sampling methods, we devise the hybrid method that successfully recovers the whole range of degree distributions. Our tail-scope method shows how structural heterogeneities of large-scale complex networks can be used to effectively reveal the network structure only with limited local information.

  10. Complexity as a Factor of Quality and Cost in Large Scale Software Development.

    DTIC Science & Technology

    1979-12-01

    allocating testing resources." [69 69I V. THE ROLE OF COMPLEXITY IN RESOURCE ESTIMATION AND ALLOCATION A. GENERAL It can be argued that blame for the...and allocation of testing resource by - identifying independent substructures and - identifying heavily used logic paths. 2. Setting a Design Threshold... RESOURCE ESTIMATION -------- 70 1. New Dynamic Field ------------------------- 70 2. Quality and Testing ----------------------- 71 3. Programming Units of

  11. Accounting for Parameter Uncertainty in Complex Atmospheric Models, With an Application to Greenhouse Gas Emissions Evaluation

    NASA Astrophysics Data System (ADS)

    Swallow, B.; Rigby, M. L.; Rougier, J.; Manning, A.; Thomson, D.; Webster, H. N.; Lunt, M. F.; O'Doherty, S.

    2016-12-01

    In order to understand underlying processes governing environmental and physical phenomena, a complex mathematical model is usually required. However, there is an inherent uncertainty related to the parameterisation of unresolved processes in these simulators. Here, we focus on the specific problem of accounting for uncertainty in parameter values in an atmospheric chemical transport model. Systematic errors introduced by failing to account for these uncertainties have the potential to have a large effect on resulting estimates in unknown quantities of interest. One approach that is being increasingly used to address this issue is known as emulation, in which a large number of forward runs of the simulator are carried out, in order to approximate the response of the output to changes in parameters. However, due to the complexity of some models, it is often unfeasible to run large numbers of training runs that is usually required for full statistical emulators of the environmental processes. We therefore present a simplified model reduction method for approximating uncertainties in complex environmental simulators without the need for very large numbers of training runs. We illustrate the method through an application to the Met Office's atmospheric transport model NAME. We show how our parameter estimation framework can be incorporated into a hierarchical Bayesian inversion, and demonstrate the impact on estimates of UK methane emissions, using atmospheric mole fraction data. We conclude that accounting for uncertainties in the parameterisation of complex atmospheric models is vital if systematic errors are to be minimized and all relevant uncertainties accounted for. We also note that investigations of this nature can prove extremely useful in highlighting deficiencies in the simulator that might otherwise be missed.

  12. Human behavioral complexity peaks at age 25

    PubMed Central

    Brugger, Peter

    2017-01-01

    Random Item Generation tasks (RIG) are commonly used to assess high cognitive abilities such as inhibition or sustained attention. They also draw upon our approximate sense of complexity. A detrimental effect of aging on pseudo-random productions has been demonstrated for some tasks, but little is as yet known about the developmental curve of cognitive complexity over the lifespan. We investigate the complexity trajectory across the lifespan of human responses to five common RIG tasks, using a large sample (n = 3429). Our main finding is that the developmental curve of the estimated algorithmic complexity of responses is similar to what may be expected of a measure of higher cognitive abilities, with a performance peak around 25 and a decline starting around 60, suggesting that RIG tasks yield good estimates of such cognitive abilities. Our study illustrates that very short strings of, i.e., 10 items, are sufficient to have their complexity reliably estimated and to allow the documentation of an age-dependent decline in the approximate sense of complexity. PMID:28406953

  13. Robust position estimation of a mobile vehicle

    NASA Astrophysics Data System (ADS)

    Conan, Vania; Boulanger, Pierre; Elgazzar, Shadia

    1994-11-01

    The ability to estimate the position of a mobile vehicle is a key task for navigation over large distances in complex indoor environments such as nuclear power plants. Schematics of the plants are available, but they are incomplete, as real settings contain many objects, such as pipes, cables or furniture, that mask part of the model. The position estimation method described in this paper matches 3-D data with a simple schematic of a plant. It is basically independent of odometry information and viewpoint, robust to noisy data and spurious points and largely insensitive to occlusions. The method is based on a hypothesis/verification paradigm and its complexity is polynomial; it runs in (Omicron) (m4n4), where m represents the number of model patches and n the number of scene patches. Heuristics are presented to speed up the algorithm. Results on real 3-D data show good behavior even when the scene is very occluded.

  14. Airborne lidar-based estimates of tropical forest structure in complex terrain: opportunities and trade-offs for REDD+

    Treesearch

    Veronika Leitold; Michael Keller; Douglas C Morton; Bruce D Cook; Yosio E Shimabukuro

    2015-01-01

    Background: Carbon stocks and fluxes in tropical forests remain large sources of uncertainty in the global carbon budget. Airborne lidar remote sensing is a powerful tool for estimating aboveground biomass, provided that lidar measurements penetrate dense forest vegetation to generate accurate estimates of surface topography and canopy heights. Tropical forest areas...

  15. Comparison of Efficiency of Jackknife and Variance Component Estimators of Standard Errors. Program Statistics Research. Technical Report.

    ERIC Educational Resources Information Center

    Longford, Nicholas T.

    Large scale surveys usually employ a complex sampling design and as a consequence, no standard methods for estimation of the standard errors associated with the estimates of population means are available. Resampling methods, such as jackknife or bootstrap, are often used, with reference to their properties of robustness and reduction of bias. A…

  16. Complex plane integration in the modelling of electromagnetic fields in layered media: part 1. Application to a very large loop

    NASA Astrophysics Data System (ADS)

    Silva, Valdelírio da Silva e.; Régis, Cícero; Howard, Allen Q., Jr.

    2014-02-01

    This paper analyses the details of a procedure for the numerical integration of Hankel transforms in the calculation of the electromagnetic fields generated by a large horizontal loop over a 1D earth. The method performs the integration by deforming the integration path into the complex plane and applying Cauchy's theorem on a modified version of the integrand. The modification is the replacement of the Bessel functions J0 and J1 by the Hankel functions H_0^{(1)} and H_1^{(1)} respectively. The integration in the complex plane takes advantage of the exponentially decaying behaviour of the Hankel functions, allowing calculation on very small segments, instead of the infinite line of the original improper integrals. A crucial point in this problem is the location of the poles. The companion paper shows two methods to estimate the pole locations. We have used this method to calculate the fields of very large loops. Our results show that this method allows the estimation of the integrals with fewer evaluations of the integrand functions than other methods.

  17. Invasion complexity at large spatial scales is an emergent property of interactions among landscape characteristics and invader traits

    USDA-ARS?s Scientific Manuscript database

    Understanding the potential for invasive spread is an important consideration for novel agricultural species that may be translocated or introduced into new regions. However, estimating invasion risks remains a challenging problem, particularly in the context of real, complex landscapes. There is ...

  18. Software Transition Project Retrospectives and the Application of SEL Effort Estimation Model and Boehm's COCOMO to Complex Software Transition Projects

    NASA Technical Reports Server (NTRS)

    McNeill, Justin

    1995-01-01

    The Multimission Image Processing Subsystem (MIPS) at the Jet Propulsion Laboratory (JPL) has managed transitions of application software sets from one operating system and hardware platform to multiple operating systems and hardware platforms. As a part of these transitions, cost estimates were generated from the personal experience of in-house developers and managers to calculate the total effort required for such projects. Productivity measures have been collected for two such transitions, one very large and the other relatively small in terms of source lines of code. These estimates used a cost estimation model similar to the Software Engineering Laboratory (SEL) Effort Estimation Model. Experience in transitioning software within JPL MIPS have uncovered a high incidence of interface complexity. Interfaces, both internal and external to individual software applications, have contributed to software transition project complexity, and thus to scheduling difficulties and larger than anticipated design work on software to be ported.

  19. Combined Parameter and State Estimation Problem in a Complex Domain: RF Hyperthermia Treatment Using Nanoparticles

    NASA Astrophysics Data System (ADS)

    Bermeo Varon, L. A.; Orlande, H. R. B.; Eliçabe, G. E.

    2016-09-01

    The particle filter methods have been widely used to solve inverse problems with sequential Bayesian inference in dynamic models, simultaneously estimating sequential state variables and fixed model parameters. This methods are an approximation of sequences of probability distributions of interest, that using a large set of random samples, with presence uncertainties in the model, measurements and parameters. In this paper the main focus is the solution combined parameters and state estimation in the radiofrequency hyperthermia with nanoparticles in a complex domain. This domain contains different tissues like muscle, pancreas, lungs, small intestine and a tumor which is loaded iron oxide nanoparticles. The results indicated that excellent agreements between estimated and exact value are obtained.

  20. DOE Office of Scientific and Technical Information (OSTI.GOV)

    Bromberger, Seth A.; Klymko, Christine F.; Henderson, Keith A.

    Betweenness centrality is a graph statistic used to nd vertices that are participants in a large number of shortest paths in a graph. This centrality measure is commonly used in path and network interdiction problems and its complete form requires the calculation of all-pairs shortest paths for each vertex. This leads to a time complexity of O(jV jjEj), which is impractical for large graphs. Estimation of betweenness centrality has focused on performing shortest-path calculations on a subset of randomly- selected vertices. This reduces the complexity of the centrality estimation to O(jSjjEj); jSj < jV j, which can be scaled appropriatelymore » based on the computing resources available. An estimation strategy that uses random selection of vertices for seed selection is fast and simple to implement, but may not provide optimal estimation of betweenness centrality when the number of samples is constrained. Our experimentation has identi ed a number of alternate seed-selection strategies that provide lower error than random selection in common scale-free graphs. These strategies are discussed and experimental results are presented.« less

  1. Fast adaptive diamond search algorithm for block-matching motion estimation using spatial correlation

    NASA Astrophysics Data System (ADS)

    Park, Sang-Gon; Jeong, Dong-Seok

    2000-12-01

    In this paper, we propose a fast adaptive diamond search algorithm (FADS) for block matching motion estimation. Many fast motion estimation algorithms reduce the computational complexity by the UESA (Unimodal Error Surface Assumption) where the matching error monotonically increases as the search moves away from the global minimum point. Recently, many fast BMAs (Block Matching Algorithms) make use of the fact that global minimum points in real world video sequences are centered at the position of zero motion. But these BMAs, especially in large motion, are easily trapped into the local minima and result in poor matching accuracy. So, we propose a new motion estimation algorithm using the spatial correlation among the neighboring blocks. We move the search origin according to the motion vectors of the spatially neighboring blocks and their MAEs (Mean Absolute Errors). The computer simulation shows that the proposed algorithm has almost the same computational complexity with DS (Diamond Search), but enhances PSNR. Moreover, the proposed algorithm gives almost the same PSNR as that of FS (Full Search), even for the large motion with half the computational load.

  2. Randomization-Based Inference about Latent Variables from Complex Samples: The Case of Two-Stage Sampling

    ERIC Educational Resources Information Center

    Li, Tiandong

    2012-01-01

    In large-scale assessments, such as the National Assessment of Educational Progress (NAEP), plausible values based on Multiple Imputations (MI) have been used to estimate population characteristics for latent constructs under complex sample designs. Mislevy (1991) derived a closed-form analytic solution for a fixed-effect model in creating…

  3. Data-based discharge extrapolation: estimating annual discharge for a partially gauged large river basin from its small sub-basins

    NASA Astrophysics Data System (ADS)

    Gong, L.

    2013-12-01

    Large-scale hydrological models and land surface models are by far the only tools for accessing future water resources in climate change impact studies. Those models estimate discharge with large uncertainties, due to the complex interaction between climate and hydrology, the limited quality and availability of data, as well as model uncertainties. A new purely data-based scale-extrapolation method is proposed, to estimate water resources for a large basin solely from selected small sub-basins, which are typically two-orders-of-magnitude smaller than the large basin. Those small sub-basins contain sufficient information, not only on climate and land surface, but also on hydrological characteristics for the large basin In the Baltic Sea drainage basin, best discharge estimation for the gauged area was achieved with sub-basins that cover 2-4% of the gauged area. There exist multiple sets of sub-basins that resemble the climate and hydrology of the basin equally well. Those multiple sets estimate annual discharge for gauged area consistently well with 5% average error. The scale-extrapolation method is completely data-based; therefore it does not force any modelling error into the prediction. The multiple predictions are expected to bracket the inherent variations and uncertainties of the climate and hydrology of the basin. The method can be applied in both un-gauged basins and un-gauged periods with uncertainty estimation.

  4. A Large-Scale Multi-Hop Localization Algorithm Based on Regularized Extreme Learning for Wireless Networks.

    PubMed

    Zheng, Wei; Yan, Xiaoyong; Zhao, Wei; Qian, Chengshan

    2017-12-20

    A novel large-scale multi-hop localization algorithm based on regularized extreme learning is proposed in this paper. The large-scale multi-hop localization problem is formulated as a learning problem. Unlike other similar localization algorithms, the proposed algorithm overcomes the shortcoming of the traditional algorithms which are only applicable to an isotropic network, therefore has a strong adaptability to the complex deployment environment. The proposed algorithm is composed of three stages: data acquisition, modeling and location estimation. In data acquisition stage, the training information between nodes of the given network is collected. In modeling stage, the model among the hop-counts and the physical distances between nodes is constructed using regularized extreme learning. In location estimation stage, each node finds its specific location in a distributed manner. Theoretical analysis and several experiments show that the proposed algorithm can adapt to the different topological environments with low computational cost. Furthermore, high accuracy can be achieved by this method without setting complex parameters.

  5. Theoretical prediction of welding distortion in large and complex structures

    NASA Astrophysics Data System (ADS)

    Deng, De-An

    2010-06-01

    Welding technology is widely used to assemble large thin plate structures such as ships, automobiles, and passenger trains because of its high productivity. However, it is impossible to avoid welding-induced distortion during the assembly process. Welding distortion not only reduces the fabrication accuracy of a weldment, but also decreases the productivity due to correction work. If welding distortion can be predicted using a practical method beforehand, the prediction will be useful for taking appropriate measures to control the dimensional accuracy to an acceptable limit. In this study, a two-step computational approach, which is a combination of a thermoelastic-plastic finite element method (FEM) and an elastic finite element with consideration for large deformation, is developed to estimate welding distortion for large and complex welded structures. Welding distortions in several representative large complex structures, which are often used in shipbuilding, are simulated using the proposed method. By comparing the predictions and the measurements, the effectiveness of the two-step computational approach is verified.

  6. Using cumulative diet data and stable isotope analysis to determine trophic position of walleye Sander vitreus in a large, complex system

    USGS Publications Warehouse

    Fincel, Mark J.; James, Daniel A.; Chipps, Steven R.; Davis, Blake A.

    2014-01-01

    Diet studies have traditionally been used to determine prey use and food web dynamics, while stable isotope analysis provides for a time-integrated approach to evaluate food web dynamics and characterize energy flow in aquatic systems. Direct comparison of the two techniques is rare and difficult to conduct in large, species rich systems. We compared changes in walleye Sander vitreus trophic position (TP) derived from paired diet content and stable isotope analysis. Individual diet-derived TP estimates were dissimilar to stable isotope-derived TP estimates. However, cumulative diet-derived TP estimates integrated from May 2001 to May 2002 corresponded to May 2002 isotope-derived estimates of TP. Average walleye TP estimates from the spring season appear representative of feeding throughout the entire previous year.

  7. Fast surface-based travel depth estimation algorithm for macromolecule surface shape description.

    PubMed

    Giard, Joachim; Alface, Patrice Rondao; Gala, Jean-Luc; Macq, Benoît

    2011-01-01

    Travel Depth, introduced by Coleman and Sharp in 2006, is a physical interpretation of molecular depth, a term frequently used to describe the shape of a molecular active site or binding site. Travel Depth can be seen as the physical distance a solvent molecule would have to travel from a point of the surface, i.e., the Solvent-Excluded Surface (SES), to its convex hull. Existing algorithms providing an estimation of the Travel Depth are based on a regular sampling of the molecule volume and the use of the Dijkstra's shortest path algorithm. Since Travel Depth is only defined on the molecular surface, this volume-based approach is characterized by a large computational complexity due to the processing of unnecessary samples lying inside or outside the molecule. In this paper, we propose a surface-based approach that restricts the processing to data defined on the SES. This algorithm significantly reduces the complexity of Travel Depth estimation and makes possible the analysis of large macromolecule surface shape description with high resolution. Experimental results show that compared to existing methods, the proposed algorithm achieves accurate estimations with considerably reduced processing times.

  8. Generalized Likelihood Uncertainty Estimation (GLUE) Using Multi-Optimization Algorithm as Sampling Method

    NASA Astrophysics Data System (ADS)

    Wang, Z.

    2015-12-01

    For decades, distributed and lumped hydrological models have furthered our understanding of hydrological system. The development of hydrological simulation in large scale and high precision elaborated the spatial descriptions and hydrological behaviors. Meanwhile, the new trend is also followed by the increment of model complexity and number of parameters, which brings new challenges of uncertainty quantification. Generalized Likelihood Uncertainty Estimation (GLUE) has been widely used in uncertainty analysis for hydrological models referring to Monte Carlo method coupled with Bayesian estimation. However, the stochastic sampling method of prior parameters adopted by GLUE appears inefficient, especially in high dimensional parameter space. The heuristic optimization algorithms utilizing iterative evolution show better convergence speed and optimality-searching performance. In light of the features of heuristic optimization algorithms, this study adopted genetic algorithm, differential evolution, shuffled complex evolving algorithm to search the parameter space and obtain the parameter sets of large likelihoods. Based on the multi-algorithm sampling, hydrological model uncertainty analysis is conducted by the typical GLUE framework. To demonstrate the superiority of the new method, two hydrological models of different complexity are examined. The results shows the adaptive method tends to be efficient in sampling and effective in uncertainty analysis, providing an alternative path for uncertainty quantilization.

  9. A Powerful Approach to Estimating Annotation-Stratified Genetic Covariance via GWAS Summary Statistics.

    PubMed

    Lu, Qiongshi; Li, Boyang; Ou, Derek; Erlendsdottir, Margret; Powles, Ryan L; Jiang, Tony; Hu, Yiming; Chang, David; Jin, Chentian; Dai, Wei; He, Qidu; Liu, Zefeng; Mukherjee, Shubhabrata; Crane, Paul K; Zhao, Hongyu

    2017-12-07

    Despite the success of large-scale genome-wide association studies (GWASs) on complex traits, our understanding of their genetic architecture is far from complete. Jointly modeling multiple traits' genetic profiles has provided insights into the shared genetic basis of many complex traits. However, large-scale inference sets a high bar for both statistical power and biological interpretability. Here we introduce a principled framework to estimate annotation-stratified genetic covariance between traits using GWAS summary statistics. Through theoretical and numerical analyses, we demonstrate that our method provides accurate covariance estimates, thereby enabling researchers to dissect both the shared and distinct genetic architecture across traits to better understand their etiologies. Among 50 complex traits with publicly accessible GWAS summary statistics (N total ≈ 4.5 million), we identified more than 170 pairs with statistically significant genetic covariance. In particular, we found strong genetic covariance between late-onset Alzheimer disease (LOAD) and amyotrophic lateral sclerosis (ALS), two major neurodegenerative diseases, in single-nucleotide polymorphisms (SNPs) with high minor allele frequencies and in SNPs located in the predicted functional genome. Joint analysis of LOAD, ALS, and other traits highlights LOAD's correlation with cognitive traits and hints at an autoimmune component for ALS. Copyright © 2017 American Society of Human Genetics. Published by Elsevier Inc. All rights reserved.

  10. A review of surface energy balance models for estimating actual evapotranspiration with remote sensing at high spatiotemporal resolution over large extents

    USGS Publications Warehouse

    McShane, Ryan R.; Driscoll, Katelyn P.; Sando, Roy

    2017-09-27

    Many approaches have been developed for measuring or estimating actual evapotranspiration (ETa), and research over many years has led to the development of remote sensing methods that are reliably reproducible and effective in estimating ETa. Several remote sensing methods can be used to estimate ETa at the high spatial resolution of agricultural fields and the large extent of river basins. More complex remote sensing methods apply an analytical approach to ETa estimation using physically based models of varied complexity that require a combination of ground-based and remote sensing data, and are grounded in the theory behind the surface energy balance model. This report, funded through cooperation with the International Joint Commission, provides an overview of selected remote sensing methods used for estimating water consumed through ETa and focuses on Mapping Evapotranspiration at High Resolution with Internalized Calibration (METRIC) and Operational Simplified Surface Energy Balance (SSEBop), two energy balance models for estimating ETa that are currently applied successfully in the United States. The METRIC model can produce maps of ETa at high spatial resolution (30 meters using Landsat data) for specific areas smaller than several hundred square kilometers in extent, an improvement in practice over methods used more generally at larger scales. Many studies validating METRIC estimates of ETa against measurements from lysimeters have shown model accuracies on daily to seasonal time scales ranging from 85 to 95 percent. The METRIC model is accurate, but the greater complexity of METRIC results in greater data requirements, and the internalized calibration of METRIC leads to greater skill required for implementation. In contrast, SSEBop is a simpler model, having reduced data requirements and greater ease of implementation without a substantial loss of accuracy in estimating ETa. The SSEBop model has been used to produce maps of ETa over very large extents (the conterminous United States) using lower spatial resolution (1 kilometer) Moderate Resolution Imaging Spectroradiometer (MODIS) data. Model accuracies ranging from 80 to 95 percent on daily to annual time scales have been shown in numerous studies that validated ETa estimates from SSEBop against eddy covariance measurements. The METRIC and SSEBop models can incorporate low and high spatial resolution data from MODIS and Landsat, but the high spatiotemporal resolution of ETa estimates using Landsat data over large extents takes immense computing power. Cloud computing is providing an opportunity for processing an increasing amount of geospatial “big data” in a decreasing period of time. For example, Google Earth EngineTM has been used to implement METRIC with automated calibration for regional-scale estimates of ETa using Landsat data. The U.S. Geological Survey also is using Google Earth EngineTM to implement SSEBop for estimating ETa in the United States at a continental scale using Landsat data.

  11. Diffusion Limitations in Root Uptake of Cadmium and Zinc, But Not Nickel, and Resulting Bias in the Michaelis Constant1[W][OA

    PubMed Central

    Degryse, Fien; Shahbazi, Afsaneh; Verheyen, Liesbeth; Smolders, Erik

    2012-01-01

    It has long been recognized that diffusive boundary layers affect the determination of active transport parameters, but this has been largely overlooked in plant physiological research. We studied the short-term uptake of cadmium (Cd), zinc (Zn), and nickel (Ni) by spinach (Spinacia oleracea) and tomato (Lycopersicon esculentum) in solutions with or without metal complexes. At same free ion concentration, the presence of complexes, which enhance the diffusion flux, increased the uptake of Cd and Zn, whereas Ni uptake was unaffected. Competition effects of protons on Cd and Zn uptake were observed only at a very large degree of buffering, while competition of magnesium ions on Ni uptake was observed even in unbuffered solutions. These results strongly suggest that uptake of Cd and Zn is limited by diffusion of the free ion to the roots, except at very high degree of solution buffering, whereas Ni uptake is generally internalization limited. All results could be well described by a model that combined a diffusion equation with a competitive Michaelis-Menten equation. Direct uptake of the complex was estimated to be a major contribution only at millimolar concentrations of the complex or at very large ratios of complex to free ion concentration. The true Km for uptake of Cd2+ and Zn2+ was estimated at <5 nm, three orders of magnitude smaller than the Km measured in unbuffered solutions. Published Michaelis constants for plant uptake of Cd and Zn likely strongly overestimate physiological ones and should not be interpreted as an indicator of transporter affinity. PMID:22864584

  12. Spatio-temporal precipitation climatology over complex terrain using a censored additive regression model.

    PubMed

    Stauffer, Reto; Mayr, Georg J; Messner, Jakob W; Umlauf, Nikolaus; Zeileis, Achim

    2017-06-15

    Flexible spatio-temporal models are widely used to create reliable and accurate estimates for precipitation climatologies. Most models are based on square root transformed monthly or annual means, where a normal distribution seems to be appropriate. This assumption becomes invalid on a daily time scale as the observations involve large fractions of zero observations and are limited to non-negative values. We develop a novel spatio-temporal model to estimate the full climatological distribution of precipitation on a daily time scale over complex terrain using a left-censored normal distribution. The results demonstrate that the new method is able to account for the non-normal distribution and the large fraction of zero observations. The new climatology provides the full climatological distribution on a very high spatial and temporal resolution, and is competitive with, or even outperforms existing methods, even for arbitrary locations.

  13. A Feature-based Approach to Big Data Analysis of Medical Images

    PubMed Central

    Toews, Matthew; Wachinger, Christian; Estepar, Raul San Jose; Wells, William M.

    2015-01-01

    This paper proposes an inference method well-suited to large sets of medical images. The method is based upon a framework where distinctive 3D scale-invariant features are indexed efficiently to identify approximate nearest-neighbor (NN) feature matches in O(log N) computational complexity in the number of images N. It thus scales well to large data sets, in contrast to methods based on pair-wise image registration or feature matching requiring O(N) complexity. Our theoretical contribution is a density estimator based on a generative model that generalizes kernel density estimation and K-nearest neighbor (KNN) methods. The estimator can be used for on-the-fly queries, without requiring explicit parametric models or an off-line training phase. The method is validated on a large multi-site data set of 95,000,000 features extracted from 19,000 lung CT scans. Subject-level classification identifies all images of the same subjects across the entire data set despite deformation due to breathing state, including unintentional duplicate scans. State-of-the-art performance is achieved in predicting chronic pulmonary obstructive disorder (COPD) severity across the 5-category GOLD clinical rating, with an accuracy of 89% if both exact and one-off predictions are considered correct. PMID:26221685

  14. A Feature-Based Approach to Big Data Analysis of Medical Images.

    PubMed

    Toews, Matthew; Wachinger, Christian; Estepar, Raul San Jose; Wells, William M

    2015-01-01

    This paper proposes an inference method well-suited to large sets of medical images. The method is based upon a framework where distinctive 3D scale-invariant features are indexed efficiently to identify approximate nearest-neighbor (NN) feature matches-in O (log N) computational complexity in the number of images N. It thus scales well to large data sets, in contrast to methods based on pair-wise image registration or feature matching requiring O(N) complexity. Our theoretical contribution is a density estimator based on a generative model that generalizes kernel density estimation and K-nearest neighbor (KNN) methods.. The estimator can be used for on-the-fly queries, without requiring explicit parametric models or an off-line training phase. The method is validated on a large multi-site data set of 95,000,000 features extracted from 19,000 lung CT scans. Subject-level classification identifies all images of the same subjects across the entire data set despite deformation due to breathing state, including unintentional duplicate scans. State-of-the-art performance is achieved in predicting chronic pulmonary obstructive disorder (COPD) severity across the 5-category GOLD clinical rating, with an accuracy of 89% if both exact and one-off predictions are considered correct.

  15. Evaluation of Penalized and Nonpenalized Methods for Disease Prediction with Large-Scale Genetic Data.

    PubMed

    Won, Sungho; Choi, Hosik; Park, Suyeon; Lee, Juyoung; Park, Changyi; Kwon, Sunghoon

    2015-01-01

    Owing to recent improvement of genotyping technology, large-scale genetic data can be utilized to identify disease susceptibility loci and this successful finding has substantially improved our understanding of complex diseases. However, in spite of these successes, most of the genetic effects for many complex diseases were found to be very small, which have been a big hurdle to build disease prediction model. Recently, many statistical methods based on penalized regressions have been proposed to tackle the so-called "large P and small N" problem. Penalized regressions including least absolute selection and shrinkage operator (LASSO) and ridge regression limit the space of parameters, and this constraint enables the estimation of effects for very large number of SNPs. Various extensions have been suggested, and, in this report, we compare their accuracy by applying them to several complex diseases. Our results show that penalized regressions are usually robust and provide better accuracy than the existing methods for at least diseases under consideration.

  16. High-Molecular-Weight Proanthocyanidins in Foods: Overcoming Analytical Challenges in Pursuit of Novel Dietary Bioactive Components.

    PubMed

    Neilson, Andrew P; O'Keefe, Sean F; Bolling, Bradley W

    2016-01-01

    Proanthocyanidins (PACs) are an abundant but complex class of polyphenols found in foods and botanicals. PACs are polymeric flavanols with a variety of linkages and subunits. Connectivity and degree of polymerization (DP) determine PAC bioavailability and bioactivity. Current quantitative and qualitative methods may ignore a large percentage of dietary PACs. Subsequent correlations between intake and activity are hindered by a lack of understanding of the true PAC complexity in many foods. Additionally, estimates of dietary intakes are likely inaccurate, as nutrient databank values are largely based on standards from cocoa (monomers to decamers) and blueberries (mean DP of 36). Improved analytical methodologies are needed to increase our understanding of the biological roles of these complex compounds.

  17. A large-eddy simulation based power estimation capability for wind farms over complex terrain

    NASA Astrophysics Data System (ADS)

    Senocak, I.; Sandusky, M.; Deleon, R.

    2017-12-01

    There has been an increasing interest in predicting wind fields over complex terrain at the micro-scale for resource assessment, turbine siting, and power forecasting. These capabilities are made possible by advancements in computational speed from a new generation of computing hardware, numerical methods and physics modelling. The micro-scale wind prediction model presented in this work is based on the large-eddy simulation paradigm with surface-stress parameterization. The complex terrain is represented using an immersed-boundary method that takes into account the parameterization of the surface stresses. Governing equations of incompressible fluid flow are solved using a projection method with second-order accurate schemes in space and time. We use actuator disk models with rotation to simulate the influence of turbines on the wind field. Data regarding power production from individual turbines are mostly restricted because of proprietary nature of the wind energy business. Most studies report percentage drop of power relative to power from the first row. There have been different approaches to predict power production. Some studies simply report available wind power in the upstream, some studies estimate power production using power curves available from turbine manufacturers, and some studies estimate power as torque multiplied by rotational speed. In the present work, we propose a black-box approach that considers a control volume around a turbine and estimate the power extracted from the turbine based on the conservation of energy principle. We applied our wind power prediction capability to wind farms over flat terrain such as the wind farm over Mower County, Minnesota and the Horns Rev offshore wind farm in Denmark. The results from these simulations are in good agreement with published data. We also estimate power production from a hypothetical wind farm in complex terrain region and identify potential zones suitable for wind power production.

  18. Estimation of electromagnetic dosimetric values from non-ionizing radiofrequency fields in an indoor commercial airplane environment.

    PubMed

    Aguirre, Erik; Arpón, Javier; Azpilicueta, Leire; López, Peio; de Miguel, Silvia; Ramos, Victoria; Falcone, Francisco

    2014-12-01

    In this article, the impact of topology as well as morphology of a complex indoor environment such as a commercial aircraft in the estimation of dosimetric assessment is presented. By means of an in-house developed deterministic 3D ray-launching code, estimation of electric field amplitude as a function of position for the complete volume of a commercial passenger airplane is obtained. Estimation of electromagnetic field exposure in this environment is challenging, due to the complexity and size of the scenario, as well as to the large metallic content, giving rise to strong multipath components. By performing the calculation with a deterministic technique, the complete scenario can be considered with an optimized balance between accuracy and computational cost. The proposed method can aid in the assessment of electromagnetic dosimetry in the future deployment of embarked wireless systems in commercial aircraft.

  19. Statistical processing of large image sequences.

    PubMed

    Khellah, F; Fieguth, P; Murray, M J; Allen, M

    2005-01-01

    The dynamic estimation of large-scale stochastic image sequences, as frequently encountered in remote sensing, is important in a variety of scientific applications. However, the size of such images makes conventional dynamic estimation methods, for example, the Kalman and related filters, impractical. In this paper, we present an approach that emulates the Kalman filter, but with considerably reduced computational and storage requirements. Our approach is illustrated in the context of a 512 x 512 image sequence of ocean surface temperature. The static estimation step, the primary contribution here, uses a mixture of stationary models to accurately mimic the effect of a nonstationary prior, simplifying both computational complexity and modeling. Our approach provides an efficient, stable, positive-definite model which is consistent with the given correlation structure. Thus, the methods of this paper may find application in modeling and single-frame estimation.

  20. Characterize kinematic rupture history of large earthquakes with Multiple Haskell sources

    NASA Astrophysics Data System (ADS)

    Jia, Z.; Zhan, Z.

    2017-12-01

    Earthquakes are often regarded as continuous rupture along a single fault, but the occurrence of complex large events involving multiple faults and dynamic triggering challenges this view. Such rupture complexities cause difficulties in existing finite fault inversion algorithms, because they rely on specific parameterizations and regularizations to obtain physically meaningful solutions. Furthermore, it is difficult to assess reliability and uncertainty of obtained rupture models. Here we develop a Multi-Haskell Source (MHS) method to estimate rupture process of large earthquakes as a series of sub-events of varying location, timing and directivity. Each sub-event is characterized by a Haskell rupture model with uniform dislocation and constant unilateral rupture velocity. This flexible yet simple source parameterization allows us to constrain first-order rupture complexity of large earthquakes robustly. Additionally, relatively few parameters in the inverse problem yields improved uncertainty analysis based on Markov chain Monte Carlo sampling in a Bayesian framework. Synthetic tests and application of MHS method on real earthquakes show that our method can capture major features of large earthquake rupture process, and provide information for more detailed rupture history analysis.

  1. Filtering Gene Ontology semantic similarity for identifying protein complexes in large protein interaction networks.

    PubMed

    Wang, Jian; Xie, Dong; Lin, Hongfei; Yang, Zhihao; Zhang, Yijia

    2012-06-21

    Many biological processes recognize in particular the importance of protein complexes, and various computational approaches have been developed to identify complexes from protein-protein interaction (PPI) networks. However, high false-positive rate of PPIs leads to challenging identification. A protein semantic similarity measure is proposed in this study, based on the ontology structure of Gene Ontology (GO) terms and GO annotations to estimate the reliability of interactions in PPI networks. Interaction pairs with low GO semantic similarity are removed from the network as unreliable interactions. Then, a cluster-expanding algorithm is used to detect complexes with core-attachment structure on filtered network. Our method is applied to three different yeast PPI networks. The effectiveness of our method is examined on two benchmark complex datasets. Experimental results show that our method performed better than other state-of-the-art approaches in most evaluation metrics. The method detects protein complexes from large scale PPI networks by filtering GO semantic similarity. Removing interactions with low GO similarity significantly improves the performance of complex identification. The expanding strategy is also effective to identify attachment proteins of complexes.

  2. Investigations of (Delta)14C, (delta)13C, and (delta)15N in vertebrae of white shark (Carcharodon carcharias) from the eastern North Pacific Ocean

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Kerr, L A; Andrews, A H; Cailliet, G M

    The white shark (Carcharodon carcharias) has a complex life history that is characterized by large scale movements and a highly variable diet. Estimates of age and growth for the white shark from the eastern North Pacific Ocean indicate they have a slow growth rate and a relatively high longevity. Age, growth, and longevity estimates useful for stock assessment and fishery models, however, require some form of validation. By counting vertebral growth band pairs, ages can be estimated, but because not all sharks deposit annual growth bands and many are not easily discernable, it is necessary to validate growth band periodicitymore » with an independent method. Radiocarbon ({sup 14}C) age validation uses the discrete {sup 14}C signal produced from thermonuclear testing in the 1950s and 1960s that is retained in skeletal structures as a time-specific marker. Growth band pairs in vertebrae, estimated as annual and spanning the 1930s to 1990s, were analyzed for {Delta}{sup 14}C and stable carbon and nitrogen isotopes ({delta}{sup 13}C and {delta}{sup 15}N). The aim of this study was to evaluate the utility of {sup 14}C age validation for a wide-ranging species with a complex life history and to use stable isotope measurements in vertebrae as a means of resolving complexity introduced into the {sup 14}C chronology by ontogenetic shifts in diet and habitat. Stable isotopes provided useful trophic position information; however, validation of age estimates was confounded by what may have been some combination of the dietary source of carbon to the vertebrae, large-scale movement patterns, and steep {sup 14}C gradients with depth in the eastern North Pacific Ocean.« less

  3. A Reduced Dimension Static, Linearized Kalman Filter and Smoother

    NASA Technical Reports Server (NTRS)

    Fukumori, I.

    1995-01-01

    An approximate Kalman filter and smoother, based on approximations of the state estimation error covariance matrix, is described. Approximations include a reduction of the effective state dimension, use of a static asymptotic error limit, and a time-invariant linearization of the dynamic model for error integration. The approximations lead to dramatic computational savings in applying estimation theory to large complex systems. Examples of use come from TOPEX/POSEIDON.

  4. Spacecraft structural system identification by modal test

    NASA Technical Reports Server (NTRS)

    Chen, J.-C.; Peretti, L. F.; Garba, J. A.

    1984-01-01

    A structural parameter estimation procedure using the measured natural frequencies and kinetic energy distribution as observers is proposed. The theoretical derivation of the estimation procedure is described and its constraints and limitations are explained. This procedure is applied to a large complex spacecraft structural system to identify the inertia matrix using modal test results. The inertia matrix is chosen after the stiffness matrix has been updated by the static test results.

  5. Systematic Underestimation of Earthquake Magnitudes from Large Intracontinental Reverse Faults: Historical Ruptures Break Across Segment Boundaries

    NASA Technical Reports Server (NTRS)

    Rubin, C. M.

    1996-01-01

    Because most large-magnitude earthquakes along reverse faults have such irregular and complicated rupture patterns, reverse-fault segments defined on the basis of geometry alone may not be very useful for estimating sizes of future seismic sources. Most modern large ruptures of historical earthquakes generated by intracontinental reverse faults have involved geometrically complex rupture patterns. Ruptures across surficial discontinuities and complexities such as stepovers and cross-faults are common. Specifically, segment boundaries defined on the basis of discontinuities in surficial fault traces, pronounced changes in the geomorphology along strike, or the intersection of active faults commonly have not proven to be major impediments to rupture. Assuming that the seismic rupture will initiate and terminate at adjacent major geometric irregularities will commonly lead to underestimation of magnitudes of future large earthquakes.

  6. Comparison of different statistical methods for estimation of extreme sea levels with wave set-up contribution

    NASA Astrophysics Data System (ADS)

    Kergadallan, Xavier; Bernardara, Pietro; Benoit, Michel; Andreewsky, Marc; Weiss, Jérôme

    2013-04-01

    Estimating the probability of occurrence of extreme sea levels is a central issue for the protection of the coast. Return periods of sea level with wave set-up contribution are estimated here in one site : Cherbourg in France in the English Channel. The methodology follows two steps : the first one is computation of joint probability of simultaneous wave height and still sea level, the second one is interpretation of that joint probabilities to assess a sea level for a given return period. Two different approaches were evaluated to compute joint probability of simultaneous wave height and still sea level : the first one is multivariate extreme values distributions of logistic type in which all components of the variables become large simultaneously, the second one is conditional approach for multivariate extreme values in which only one component of the variables have to be large. Two different methods were applied to estimate sea level with wave set-up contribution for a given return period : Monte-Carlo simulation in which estimation is more accurate but needs higher calculation time and classical ocean engineering design contours of type inverse-FORM in which the method is simpler and allows more complex estimation of wave setup part (wave propagation to the coast for example). We compare results from the two different approaches with the two different methods. To be able to use both Monte-Carlo simulation and design contours methods, wave setup is estimated with an simple empirical formula. We show advantages of the conditional approach compared to the multivariate extreme values approach when extreme sea-level occurs when either surge or wave height is large. We discuss the validity of the ocean engineering design contours method which is an alternative when computation of sea levels is too complex to use Monte-Carlo simulation method.

  7. Complexation of Cd, Ni, and Zn by DOC in polluted groundwater: A comparison of approaches using resin exchange, aquifer material sorption, and computer speciation models (WHAM and MINTEQA2)

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Christensen, J.B.; Christensen, T.H.

    1999-11-01

    Complexation of cadmium (Cd), nickel (Ni), and zinc (Zn) by dissolved organic carbon (DOC) in leachate-polluted groundwater was measured using a resin equilibrium method and an aquifer material sorption technique. The first method is commonly used in complexation studies, while the second method better represents aquifer conditions. The two approaches gave similar results. Metal-DOC complexation was measured over a range of DOC concentrations using the resin equilibrium method, and the results were compared to simulations made by two speciation models containing default databases on metal-DOC complexes (WHAM and MINTEQA2). The WHAM model gave reasonable estimates of Cd and Ni complexationmore » by DOC for both leachate-polluted groundwater samples. The estimated effect of complexation differed less than 50% from the experimental values corresponding to a deviation on the activity of the free metal ion of a factor of 2.5. The effect of DOC complexation for Zn was largely overestimated by the WHAM model, and it was found that using a binding constant of 1.7 instead of the default value of 1.3 would improve the fit between the simulations and experimental data. The MINTEQA2 model gave reasonable predictions of the complexation of Cd and Zn by DOC, whereas deviations in the estimated activity of the free Ni{sup 2+} ion as compared to experimental results are up to a factor of 5.« less

  8. Stoichiometry for binding and transport by the twin arginine translocation system.

    PubMed

    Celedon, Jose M; Cline, Kenneth

    2012-05-14

    Twin arginine translocation (Tat) systems transport large folded proteins across sealed membranes. Tat systems accomplish this feat with three membrane components organized in two complexes. In thylakoid membranes, cpTatC and Hcf106 comprise a large receptor complex containing an estimated eight cpTatC-Hcf106 pairs. Protein transport occurs when Tha4 joins the receptor complex as an oligomer of uncertain size that is thought to form the protein-conducting structure. Here, binding analyses with intact membranes or purified complexes indicate that each receptor complex could bind eight precursor proteins. Kinetic analysis of translocation showed that each precursor-bound site was independently functional for transport, and, with sufficient Tha4, all sites were concurrently active for transport. Tha4 titration determined that ∼26 Tha4 protomers were required for transport of each OE17 (oxygen-evolving complex subunit of 17 kD) precursor protein. Our results suggest that, when fully saturated with precursor proteins and Tha4, the Tat translocase is an ∼2.2-megadalton complex that can individually transport eight precursor proteins or cooperatively transport multimeric precursors.

  9. Flux Calculation Using CARIBIC DOAS Aircraft Measurements: SO2 Emission of Norilsk

    NASA Technical Reports Server (NTRS)

    Walter, D.; Heue, K.-P.; Rauthe-Schoech, A.; Brenninkmeijer, C. A. M.; Lamsal, L. N.; Krotkov, N. A.; Platt, U.

    2012-01-01

    Based on a case-study of the nickel smelter in Norilsk (Siberia), the retrieval of trace gas fluxes using airborne remote sensing is discussed. A DOAS system onboard an Airbus 340 detected large amounts of SO2 and NO2 near Norilsk during a regular passenger flight within the CARIBIC project. The remote sensing data were combined with ECMWF wind data to estimate the SO2 output of the Norilsk industrial complex to be around 1 Mt per year, which is in agreement with independent estimates. This value is compared to results using data from satellite remote sensing (GOME, OMI). The validity of the assumptions underlying our estimate is discussed, including the adaptation of this method to other gases and sources like the NO2 emissions of large industries or cities.

  10. Depth inpainting by tensor voting.

    PubMed

    Kulkarni, Mandar; Rajagopalan, Ambasamudram N

    2013-06-01

    Depth maps captured by range scanning devices or by using optical cameras often suffer from missing regions due to occlusions, reflectivity, limited scanning area, sensor imperfections, etc. In this paper, we propose a fast and reliable algorithm for depth map inpainting using the tensor voting (TV) framework. For less complex missing regions, local edge and depth information is utilized for synthesizing missing values. The depth variations are modeled by local planes using 3D TV, and missing values are estimated using plane equations. For large and complex missing regions, we collect and evaluate depth estimates from self-similar (training) datasets. We align the depth maps of the training set with the target (defective) depth map and evaluate the goodness of depth estimates among candidate values using 3D TV. We demonstrate the effectiveness of the proposed approaches on real as well as synthetic data.

  11. The space-time structure of oil and gas field growth in a complex depositional system

    USGS Publications Warehouse

    Drew, L.J.; Mast, R.F.; Schuenemeyer, J.H.

    1994-01-01

    Shortly after the discovery of an oil and gas field, an initial estimate is usually made of the ultimate recovery of the field. With the passage of time, this initial estimate is almost always revised upward. The phenomenon of the growth of the expected ultimate recovery of a field, which is known as "field growth," is important to resource assessment analysts for several reasons. First, field growth is the source of a large part of future additions to the inventory of proved reserves of crude oil and natural gas in most petroliferous areas of the world. Second, field growth introduces a large negative bias in the forecast of the future rates of discovery of oil and gas fields made by discovery process models. In this study, the growth in estimated ultimate recovery of oil and gas in fields made up of sandstone reservoirs formed in a complex depositional environment (Frio strand plain exploration play) is examined. The results presented here show how the growth of oil and gas fields is tied directly to the architectural element of the shoreline processes and tectonics that caused the deposition of the individual sand bodies hosting the producible hydrocarbon. ?? 1994 Oxford University Press.

  12. Tomographic inversion of P-wave velocity and Q structures beneath the Kirishima volcanic complex, Southern Japan, based on finite difference calculations of complex traveltimes

    USGS Publications Warehouse

    Tomatsu, T.; Kumagai, H.; Dawson, P.B.

    2001-01-01

    We estimate the P-wave velocity and attenuation structures beneath the Kirishima volcanic complex, southern Japan, by inverting the complex traveltimes (arrival times and pulse widths) of waveform data obtained during an active seismic experiment conducted in 1994. In this experiment, six 200-250 kg shots were recorded at 163 temporary seismic stations deployed on the volcanic complex. We use first-arrival times for the shots, which were hand-measured interactively. The waveform data are Fourier transformed into the frequency domain and analysed using a new method based on autoregressive modelling of complex decaying oscillations in the frequency domain to determine pulse widths for the first-arrival phases. A non-linear inversion method is used to invert 893 first-arrival times and 325 pulse widths to estimate the velocity and attenuation structures of the volcanic complex. Wavefronts for the inversion are calculated with a finite difference method based on the Eikonal equation, which is well suited to estimating the complex traveltimes for the structures of the Kirishima volcano complex, where large structural heterogeneities are expected. The attenuation structure is derived using ray paths derived from the velocity structure. We obtain 3-D velocity and attenuation structures down to 1.5 and 0.5 km below sea level, respectively. High-velocity pipe-like structures with correspondingly low attenuation are found under the summit craters. These pipe-like structures are interpreted as remnant conduits of solidified magma. No evidence of a shallow magma chamber is visible in the tomographic images.

  13. The Limitations of Model-Based Experimental Design and Parameter Estimation in Sloppy Systems.

    PubMed

    White, Andrew; Tolman, Malachi; Thames, Howard D; Withers, Hubert Rodney; Mason, Kathy A; Transtrum, Mark K

    2016-12-01

    We explore the relationship among experimental design, parameter estimation, and systematic error in sloppy models. We show that the approximate nature of mathematical models poses challenges for experimental design in sloppy models. In many models of complex biological processes it is unknown what are the relevant physical mechanisms that must be included to explain system behaviors. As a consequence, models are often overly complex, with many practically unidentifiable parameters. Furthermore, which mechanisms are relevant/irrelevant vary among experiments. By selecting complementary experiments, experimental design may inadvertently make details that were ommitted from the model become relevant. When this occurs, the model will have a large systematic error and fail to give a good fit to the data. We use a simple hyper-model of model error to quantify a model's discrepancy and apply it to two models of complex biological processes (EGFR signaling and DNA repair) with optimally selected experiments. We find that although parameters may be accurately estimated, the discrepancy in the model renders it less predictive than it was in the sloppy regime where systematic error is small. We introduce the concept of a sloppy system-a sequence of models of increasing complexity that become sloppy in the limit of microscopic accuracy. We explore the limits of accurate parameter estimation in sloppy systems and argue that identifying underlying mechanisms controlling system behavior is better approached by considering a hierarchy of models of varying detail rather than focusing on parameter estimation in a single model.

  14. The Limitations of Model-Based Experimental Design and Parameter Estimation in Sloppy Systems

    PubMed Central

    Tolman, Malachi; Thames, Howard D.; Mason, Kathy A.

    2016-01-01

    We explore the relationship among experimental design, parameter estimation, and systematic error in sloppy models. We show that the approximate nature of mathematical models poses challenges for experimental design in sloppy models. In many models of complex biological processes it is unknown what are the relevant physical mechanisms that must be included to explain system behaviors. As a consequence, models are often overly complex, with many practically unidentifiable parameters. Furthermore, which mechanisms are relevant/irrelevant vary among experiments. By selecting complementary experiments, experimental design may inadvertently make details that were ommitted from the model become relevant. When this occurs, the model will have a large systematic error and fail to give a good fit to the data. We use a simple hyper-model of model error to quantify a model’s discrepancy and apply it to two models of complex biological processes (EGFR signaling and DNA repair) with optimally selected experiments. We find that although parameters may be accurately estimated, the discrepancy in the model renders it less predictive than it was in the sloppy regime where systematic error is small. We introduce the concept of a sloppy system–a sequence of models of increasing complexity that become sloppy in the limit of microscopic accuracy. We explore the limits of accurate parameter estimation in sloppy systems and argue that identifying underlying mechanisms controlling system behavior is better approached by considering a hierarchy of models of varying detail rather than focusing on parameter estimation in a single model. PMID:27923060

  15. Harnessing quantitative genetics and genomics for understanding and improving complex traits in crops

    USDA-ARS?s Scientific Manuscript database

    Classical quantitative genetics aids crop improvement by providing the means to estimate heritability, genetic correlations, and predicted responses to various selection schemes. Genomics has the potential to aid quantitative genetics and applied crop improvement programs via large-scale, high-thro...

  16. Quantifying complexity of financial short-term time series by composite multiscale entropy measure

    NASA Astrophysics Data System (ADS)

    Niu, Hongli; Wang, Jun

    2015-05-01

    It is significant to study the complexity of financial time series since the financial market is a complex evolved dynamic system. Multiscale entropy is a prevailing method used to quantify the complexity of a time series. Due to its less reliability of entropy estimation for short-term time series at large time scales, a modification method, the composite multiscale entropy, is applied to the financial market. To qualify its effectiveness, its applications in the synthetic white noise and 1 / f noise with different data lengths are reproduced first in the present paper. Then it is introduced for the first time to make a reliability test with two Chinese stock indices. After conducting on short-time return series, the CMSE method shows the advantages in reducing deviations of entropy estimation and demonstrates more stable and reliable results when compared with the conventional MSE algorithm. Finally, the composite multiscale entropy of six important stock indices from the world financial markets is investigated, and some useful and interesting empirical results are obtained.

  17. Metamorphic P-T conditions across the Chugach Metamorphic Complex (Alaska)—A record of focussed exhumation during transpression

    NASA Astrophysics Data System (ADS)

    Bruand, Emilie; Gasser, Deta; Stüwe, Kurt

    2014-03-01

    The Chugach Metamorphic Complex (CMC) is a large high-grade metamorphic complex that developed in the Eocene within the Chugach accretionary complex along the margin of Alaska where subduction is still ongoing. The CMC has a conspicuous asymmetric structure with a migmatitic zone flanked in the north and west by amphibolite facies schists and in the south by a metabasite belt. To the north and south, major, crustal-scale fault zones juxtapose the Chugach terrane against much lower-grade and less-deformed sequences belonging to different terranes. Curiously these crustal-scale structures are known to have largely strike slip motion posing the question as to the nature of the exhumation of the high-grade complex between them. However, P-T conditions which would allow an estimation of the amount of exhumation were lacking for large parts of the complex. This paper presents petrographic descriptions, biotite-garnet thermometry, RSCM thermometry, average P-T calculations and pseudosection modelling from three major across-strike transects covering the complex from west to south-east. Our results reveal that, both temperature and pressure vary substantially across the complex. More specifically, peak metamorphic conditions evolve from 4-7 kbar and ~ 550-650 °C in the northern schist zone to 5-11 kbar and ~ 650-750 °C in the migmatite zone in the south of the complex. The higher pressure estimates in the south of the complex indicate that focussed exhumation must have occurred in this area and was probably initiated by the subduction of a high topographic relief (intra-oceanic arc or ridge subduction) and the accretion of the metabasite belt in the south. Exhumation of the CMC occurred in an overall transpressive strain regime, with strike-slip deformation concentrated along the northern Border Range fault zone and thrusting and exhumation focussed within the southern migmatite zone and splay faults of the Contact fault zone. The T/P ratios in the southern migmatite zone indicate that the thermal perturbation of the migmatites is less than previously inferred. These new results, associated with the structural data and the accretion of a metabasite belt in the south of the complex, seem incompatible with the existing ridge-subduction models.

  18. Predicting Fish Densities in Lotic Systems: a Simple Modeling Approach

    EPA Science Inventory

    Fish density models are essential tools for fish ecologists and fisheries managers. However, applying these models can be difficult because of high levels of model complexity and the large number of parameters that must be estimated. We designed a simple fish density model and te...

  19. UHF (Ultra-High-Frequency) Propagation in Vegetative Media.

    DTIC Science & Technology

    1980-04-01

    Y V /ik) where k = 2A/X is the wave number and the asterisk indicates complex conjugate. In order to obtain useful results for average values that are...easy to make an accurate estimation of the expected effects under one set of conditions on the basis of experimental observa- tions carried out under... systems propagating horizontally through vegetation. The large quantity A-13 of measured data demonstrates the complex effects upon path loss of irregu

  20. Interaction of a supersonic particle with a three-dimensional complex plasma

    NASA Astrophysics Data System (ADS)

    Zaehringer, E.; Schwabe, M.; Zhdanov, S.; Mohr, D. P.; Knapek, C. A.; Huber, P.; Semenov, I. L.; Thomas, H. M.

    2018-03-01

    The influence of a supersonic projectile on a three-dimensional complex plasma is studied. Micron sized particles in a low-temperature plasma formed a large undisturbed system in the new "Zyflex" chamber during microgravity conditions. A supersonic probe particle excited a Mach cone with Mach number M ≈ 1.5-2 and double Mach cone structure in the large weakly damped particle cloud. The speed of sound is measured with different methods and particle charge estimations are compared to the calculations from standard theories. The high image resolution enables the study of Mach cones in microgravity on the single particle level of a three-dimensional complex plasma and gives insight to the dynamics. A heating of the microparticles is discovered behind the supersonic projectile but not in the flanks of the Mach cone.

  1. Complex Population Dynamics and the Coalescent Under Neutrality

    PubMed Central

    Volz, Erik M.

    2012-01-01

    Estimates of the coalescent effective population size Ne can be poorly correlated with the true population size. The relationship between Ne and the population size is sensitive to the way in which birth and death rates vary over time. The problem of inference is exacerbated when the mechanisms underlying population dynamics are complex and depend on many parameters. In instances where nonparametric estimators of Ne such as the skyline struggle to reproduce the correct demographic history, model-based estimators that can draw on prior information about population size and growth rates may be more efficient. A coalescent model is developed for a large class of populations such that the demographic history is described by a deterministic nonlinear dynamical system of arbitrary dimension. This class of demographic model differs from those typically used in population genetics. Birth and death rates are not fixed, and no assumptions are made regarding the fraction of the population sampled. Furthermore, the population may be structured in such a way that gene copies reproduce both within and across demes. For this large class of models, it is shown how to derive the rate of coalescence, as well as the likelihood of a gene genealogy with heterochronous sampling and labeled taxa, and how to simulate a coalescent tree conditional on a complex demographic history. This theoretical framework encapsulates many of the models used by ecologists and epidemiologists and should facilitate the integration of population genetics with the study of mathematical population dynamics. PMID:22042576

  2. Statistical Field Estimation and Scale Estimation for Complex Coastal Regions and Archipelagos

    DTIC Science & Technology

    2009-05-01

    instruments applied to mode-73. Deep-Sea Research, 23:559–582. Brown , R. G. and Hwang , P. Y. C. (1997). Introduction to Random Signals and Applied Kalman ...the covariance matrix becomes neg- ative due to numerical issues ( Brown and Hwang , 1997). Some useful techniques to counter these divergence problems...equations ( Brown and Hwang , 1997). If the number of observations is large, divergence problems can arise under certain con- ditions due to truncation errors

  3. Structure in the 3D Galaxy Distribution. III. Fourier Transforming the Universe: Phase and Power Spectra

    NASA Technical Reports Server (NTRS)

    Scargle, Jeffrey D.; Way, M. J.; Gazis, P. G.

    2017-01-01

    We demonstrate the effectiveness of a relatively straightforward analysis of the complex 3D Fourier transform of galaxy coordinates derived from redshift surveys. Numerical demonstrations of this approach are carried out on a volume-limited sample of the Sloan Digital Sky Survey redshift survey. The direct unbinned transform yields a complex 3D data cube quite similar to that from the Fast Fourier Transform of finely binned galaxy positions. In both cases, deconvolution of the sampling window function yields estimates of the true transform. Simple power spectrum estimates from these transforms are roughly consistent with those using more elaborate methods. The complex Fourier transform characterizes spatial distributional properties beyond the power spectrum in a manner different from (and we argue is more easily interpreted than) the conventional multipoint hierarchy. We identify some threads of modern large-scale inference methodology that will presumably yield detections in new wider and deeper surveys.

  4. POWER ANALYSIS FOR COMPLEX MEDIATIONAL DESIGNS USING MONTE CARLO METHODS

    PubMed Central

    Thoemmes, Felix; MacKinnon, David P.; Reiser, Mark R.

    2013-01-01

    Applied researchers often include mediation effects in applications of advanced methods such as latent variable models and linear growth curve models. Guidance on how to estimate statistical power to detect mediation for these models has not yet been addressed in the literature. We describe a general framework for power analyses for complex mediational models. The approach is based on the well known technique of generating a large number of samples in a Monte Carlo study, and estimating power as the percentage of cases in which an estimate of interest is significantly different from zero. Examples of power calculation for commonly used mediational models are provided. Power analyses for the single mediator, multiple mediators, three-path mediation, mediation with latent variables, moderated mediation, and mediation in longitudinal designs are described. Annotated sample syntax for Mplus is appended and tabled values of required sample sizes are shown for some models. PMID:23935262

  5. VLA Zeeman Observations of the NGC 6334 Complex

    NASA Astrophysics Data System (ADS)

    Mayo, E. A.; Sarma, A. P.; Troland, T. H.

    2004-05-01

    We present OH 1665 and 1667 MHz observations of the NGC 6334 complex taken with the Very Large Array in the BnA configuration. We have combined our data with the lower resolution CnB data of Sarma et al (1999), in order to perform a detailed study of Source A, a compact continuum source in the SW region of the complex. Our observations reveal magnetic fields with peak values of the order of 700μ G toward Source A. Virial estimates presented indicate the significance of the magnetic field in the support of the molecular cloud against gravitational collapse.

  6. SAS procedures for designing and analyzing sample surveys

    USGS Publications Warehouse

    Stafford, Joshua D.; Reinecke, Kenneth J.; Kaminski, Richard M.

    2003-01-01

    Complex surveys often are necessary to estimate occurrence (or distribution), density, and abundance of plants and animals for purposes of re-search and conservation. Most scientists are familiar with simple random sampling, where sample units are selected from a population of interest (sampling frame) with equal probability. However, the goal of ecological surveys often is to make inferences about populations over large or complex spatial areas where organisms are not homogeneously distributed or sampling frames are in-convenient or impossible to construct. Candidate sampling strategies for such complex surveys include stratified,multistage, and adaptive sampling (Thompson 1992, Buckland 1994).

  7. Interaction Network Estimation: Predicting Problem-Solving Diversity in Interactive Environments

    ERIC Educational Resources Information Center

    Eagle, Michael; Hicks, Drew; Barnes, Tiffany

    2015-01-01

    Intelligent tutoring systems and computer aided learning environments aimed at developing problem solving produce large amounts of transactional data which make it a challenge for both researchers and educators to understand how students work within the environment. Researchers have modeled student-tutor interactions using complex networks in…

  8. Estimating the Local Size and Coverage of Interaction Network Regions

    ERIC Educational Resources Information Center

    Eagle, Michael; Barnes, Tiffany

    2015-01-01

    Interactive problem solving environments, such as intelligent tutoring systems and educational video games, produce large amounts of transactional data which make it a challenge for both researchers and educators to understand how students work within the environment. Researchers have modeled the student-tutor interactions using complex network…

  9. Large scale affinity calculations of cyclodextrin host-guest complexes: Understanding the role of reorganization in the molecular recognition process

    PubMed Central

    Wickstrom, Lauren; He, Peng; Gallicchio, Emilio; Levy, Ronald M.

    2013-01-01

    Host-guest inclusion complexes are useful models for understanding the structural and energetic aspects of molecular recognition. Due to their small size relative to much larger protein-ligand complexes, converged results can be obtained rapidly for these systems thus offering the opportunity to more reliably study fundamental aspects of the thermodynamics of binding. In this work, we have performed a large scale binding affinity survey of 57 β-cyclodextrin (CD) host guest systems using the binding energy distribution analysis method (BEDAM) with implicit solvation (OPLS-AA/AGBNP2). Converged estimates of the standard binding free energies are obtained for these systems by employing techniques such as parallel Hamitionian replica exchange molecular dynamics, conformational reservoirs and multistate free energy estimators. Good agreement with experimental measurements is obtained in terms of both numerical accuracy and affinity rankings. Overall, average effective binding energies reproduce affinity rank ordering better than the calculated binding affinities, even though calculated binding free energies, which account for effects such as conformational strain and entropy loss upon binding, provide lower root mean square errors when compared to measurements. Interestingly, we find that binding free energies are superior rank order predictors for a large subset containing the most flexible guests. The results indicate that, while challenging, accurate modeling of reorganization effects can lead to ligand design models of superior predictive power for rank ordering relative to models based only on ligand-receptor interaction energies. PMID:25147485

  10. A Review of Global Precipitation Data Sets: Data Sources, Estimation, and Intercomparisons

    NASA Astrophysics Data System (ADS)

    Sun, Qiaohong; Miao, Chiyuan; Duan, Qingyun; Ashouri, Hamed; Sorooshian, Soroosh; Hsu, Kuo-Lin

    2018-03-01

    In this paper, we present a comprehensive review of the data sources and estimation methods of 30 currently available global precipitation data sets, including gauge-based, satellite-related, and reanalysis data sets. We analyzed the discrepancies between the data sets from daily to annual timescales and found large differences in both the magnitude and the variability of precipitation estimates. The magnitude of annual precipitation estimates over global land deviated by as much as 300 mm/yr among the products. Reanalysis data sets had a larger degree of variability than the other types of data sets. The degree of variability in precipitation estimates also varied by region. Large differences in annual and seasonal estimates were found in tropical oceans, complex mountain areas, northern Africa, and some high-latitude regions. Overall, the variability associated with extreme precipitation estimates was slightly greater at lower latitudes than at higher latitudes. The reliability of precipitation data sets is mainly limited by the number and spatial coverage of surface stations, the satellite algorithms, and the data assimilation models. The inconsistencies described limit the capability of the products for climate monitoring, attribution, and model validation.

  11. Terrestrial laser scanning to quantify above-ground biomass of structurally complex coastal wetland vegetation

    NASA Astrophysics Data System (ADS)

    Owers, Christopher J.; Rogers, Kerrylee; Woodroffe, Colin D.

    2018-05-01

    Above-ground biomass represents a small yet significant contributor to carbon storage in coastal wetlands. Despite this, above-ground biomass is often poorly quantified, particularly in areas where vegetation structure is complex. Traditional methods for providing accurate estimates involve harvesting vegetation to develop mangrove allometric equations and quantify saltmarsh biomass in quadrats. However broad scale application of these methods may not capture structural variability in vegetation resulting in a loss of detail and estimates with considerable uncertainty. Terrestrial laser scanning (TLS) collects high resolution three-dimensional point clouds capable of providing detailed structural morphology of vegetation. This study demonstrates that TLS is a suitable non-destructive method for estimating biomass of structurally complex coastal wetland vegetation. We compare volumetric models, 3-D surface reconstruction and rasterised volume, and point cloud elevation histogram modelling techniques to estimate biomass. Our results show that current volumetric modelling approaches for estimating TLS-derived biomass are comparable to traditional mangrove allometrics and saltmarsh harvesting. However, volumetric modelling approaches oversimplify vegetation structure by under-utilising the large amount of structural information provided by the point cloud. The point cloud elevation histogram model presented in this study, as an alternative to volumetric modelling, utilises all of the information within the point cloud, as opposed to sub-sampling based on specific criteria. This method is simple but highly effective for both mangrove (r2 = 0.95) and saltmarsh (r2 > 0.92) vegetation. Our results provide evidence that application of TLS in coastal wetlands is an effective non-destructive method to accurately quantify biomass for structurally complex vegetation.

  12. Multi-agent based control of large-scale complex systems employing distributed dynamic inference engine

    NASA Astrophysics Data System (ADS)

    Zhang, Daili

    Increasing societal demand for automation has led to considerable efforts to control large-scale complex systems, especially in the area of autonomous intelligent control methods. The control system of a large-scale complex system needs to satisfy four system level requirements: robustness, flexibility, reusability, and scalability. Corresponding to the four system level requirements, there arise four major challenges. First, it is difficult to get accurate and complete information. Second, the system may be physically highly distributed. Third, the system evolves very quickly. Fourth, emergent global behaviors of the system can be caused by small disturbances at the component level. The Multi-Agent Based Control (MABC) method as an implementation of distributed intelligent control has been the focus of research since the 1970s, in an effort to solve the above-mentioned problems in controlling large-scale complex systems. However, to the author's best knowledge, all MABC systems for large-scale complex systems with significant uncertainties are problem-specific and thus difficult to extend to other domains or larger systems. This situation is partly due to the control architecture of multiple agents being determined by agent to agent coupling and interaction mechanisms. Therefore, the research objective of this dissertation is to develop a comprehensive, generalized framework for the control system design of general large-scale complex systems with significant uncertainties, with the focus on distributed control architecture design and distributed inference engine design. A Hybrid Multi-Agent Based Control (HyMABC) architecture is proposed by combining hierarchical control architecture and module control architecture with logical replication rings. First, it decomposes a complex system hierarchically; second, it combines the components in the same level as a module, and then designs common interfaces for all of the components in the same module; third, replications are made for critical agents and are organized into logical rings. This architecture maintains clear guidelines for complexity decomposition and also increases the robustness of the whole system. Multiple Sectioned Dynamic Bayesian Networks (MSDBNs) as a distributed dynamic probabilistic inference engine, can be embedded into the control architecture to handle uncertainties of general large-scale complex systems. MSDBNs decomposes a large knowledge-based system into many agents. Each agent holds its partial perspective of a large problem domain by representing its knowledge as a Dynamic Bayesian Network (DBN). Each agent accesses local evidence from its corresponding local sensors and communicates with other agents through finite message passing. If the distributed agents can be organized into a tree structure, satisfying the running intersection property and d-sep set requirements, globally consistent inferences are achievable in a distributed way. By using different frequencies for local DBN agent belief updating and global system belief updating, it balances the communication cost with the global consistency of inferences. In this dissertation, a fully factorized Boyen-Koller (BK) approximation algorithm is used for local DBN agent belief updating, and the static Junction Forest Linkage Tree (JFLT) algorithm is used for global system belief updating. MSDBNs assume a static structure and a stable communication network for the whole system. However, for a real system, sub-Bayesian networks as nodes could be lost, and the communication network could be shut down due to partial damage in the system. Therefore, on-line and automatic MSDBNs structure formation is necessary for making robust state estimations and increasing survivability of the whole system. A Distributed Spanning Tree Optimization (DSTO) algorithm, a Distributed D-Sep Set Satisfaction (DDSSS) algorithm, and a Distributed Running Intersection Satisfaction (DRIS) algorithm are proposed in this dissertation. Combining these three distributed algorithms and a Distributed Belief Propagation (DBP) algorithm in MSDBNs makes state estimations robust to partial damage in the whole system. Combining the distributed control architecture design and the distributed inference engine design leads to a process of control system design for a general large-scale complex system. As applications of the proposed methodology, the control system design of a simplified ship chilled water system and a notional ship chilled water system have been demonstrated step by step. Simulation results not only show that the proposed methodology gives a clear guideline for control system design for general large-scale complex systems with dynamic and uncertain environment, but also indicate that the combination of MSDBNs and HyMABC can provide excellent performance for controlling general large-scale complex systems.

  13. Temporal and spectral characteristics of dynamic functional connectivity between resting-state networks reveal information beyond static connectivity

    PubMed Central

    Yeh, Hsiang J.; Guindani, Michele; Vannucci, Marina; Haneef, Zulfi; Stern, John M.

    2018-01-01

    Estimation of functional connectivity (FC) has become an increasingly powerful tool for investigating healthy and abnormal brain function. Static connectivity, in particular, has played a large part in guiding conclusions from the majority of resting-state functional MRI studies. However, accumulating evidence points to the presence of temporal fluctuations in FC, leading to increasing interest in estimating FC as a dynamic quantity. One central issue that has arisen in this new view of connectivity is the dramatic increase in complexity caused by dynamic functional connectivity (dFC) estimation. To computationally handle this increased complexity, a limited set of dFC properties, primarily the mean and variance, have generally been considered. Additionally, it remains unclear how to integrate the increased information from dFC into pattern recognition techniques for subject-level prediction. In this study, we propose an approach to address these two issues based on a large number of previously unexplored temporal and spectral features of dynamic functional connectivity. A Generalized Autoregressive Conditional Heteroskedasticity (GARCH) model is used to estimate time-varying patterns of functional connectivity between resting-state networks. Time-frequency analysis is then performed on dFC estimates, and a large number of previously unexplored temporal and spectral features drawn from signal processing literature are extracted for dFC estimates. We apply the investigated features to two neurologic populations of interest, healthy controls and patients with temporal lobe epilepsy, and show that the proposed approach leads to substantial increases in predictive performance compared to both traditional estimates of static connectivity as well as current approaches to dFC. Variable importance is assessed and shows that there are several quantities that can be extracted from dFC signal which are more informative than the traditional mean or variance of dFC. This work illuminates many previously unexplored facets of the dynamic properties of functional connectivity between resting-state networks, and provides a platform for dynamic functional connectivity analysis that facilitates its usage as an investigative measure for healthy as well as abnormal brain function. PMID:29320526

  14. Advanced Kalman Filter for Real-Time Responsiveness in Complex Systems

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Welch, Gregory Francis; Zhang, Jinghe

    2014-06-10

    Complex engineering systems pose fundamental challenges in real-time operations and control because they are highly dynamic systems consisting of a large number of elements with severe nonlinearities and discontinuities. Today’s tools for real-time complex system operations are mostly based on steady state models, unable to capture the dynamic nature and too slow to prevent system failures. We developed advanced Kalman filtering techniques and the formulation of dynamic state estimation using Kalman filtering techniques to capture complex system dynamics in aiding real-time operations and control. In this work, we looked at complex system issues including severe nonlinearity of system equations, discontinuitiesmore » caused by system controls and network switches, sparse measurements in space and time, and real-time requirements of power grid operations. We sought to bridge the disciplinary boundaries between Computer Science and Power Systems Engineering, by introducing methods that leverage both existing and new techniques. While our methods were developed in the context of electrical power systems, they should generalize to other large-scale scientific and engineering applications.« less

  15. Enabling Predictive Simulation and UQ of Complex Multiphysics PDE Systems by the Development of Goal-Oriented Variational Sensitivity Analysis and a-Posteriori Error Estimation Methods

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Estep, Donald

    2015-11-30

    This project addressed the challenge of predictive computational analysis of strongly coupled, highly nonlinear multiphysics systems characterized by multiple physical phenomena that span a large range of length- and time-scales. Specifically, the project was focused on computational estimation of numerical error and sensitivity analysis of computational solutions with respect to variations in parameters and data. In addition, the project investigated the use of accurate computational estimates to guide efficient adaptive discretization. The project developed, analyzed and evaluated new variational adjoint-based techniques for integration, model, and data error estimation/control and sensitivity analysis, in evolutionary multiphysics multiscale simulations.

  16. Relation of the lunar volcano complexes lying on the identical linear gravity anomaly

    NASA Astrophysics Data System (ADS)

    Yamamoto, K.; Haruyama, J.; Ohtake, M.; Iwata, T.; Ishihara, Y.

    2015-12-01

    There are several large-scale volcanic complexes, e.g., Marius Hills, Aristarchus Plateau, Rumker Hills, and Flamsteed area in western Oceanus Procellarum of the lunar nearside. For better understanding of the lunar thermal history, it is important to study these areas intensively. The magmatisms and volcanic eruption mechanisms of these volcanic complexes have been discussed from geophysical and geochemical perspectives using data sets acquired by lunar explorers. In these data sets, precise gravity field data obtained by Gravity Recovery and Interior Laboratory (GRAIL) gives information on mass anomalies below the lunar surface, and useful to estimate location and mass of the embedded magmas. Using GRAIL data, Andrews-Hanna et al. (2014) prepared gravity gradient map of the Moon. They discussed the origin of the quasi-rectangular pattern of narrow linear gravity gradient anomalies located along the border of Oceanus Procellarum and suggested that the underlying dikes played important roles in magma plumbing system. In the gravity gradient map, we found that there are also several small linear gravity gradient anomaly patterns in the inside of the large quasi-rectangular pattern, and that one of the linear anomalies runs through multiple gravity anomalies in the vicinity of Aristarchus, Marius and Flamstead volcano complexes. Our concern is whether the volcanisms of these complexes are caused by common factors or not. To clarify this, we firstly estimated the mass and depth of the embedded magmas as well as the directions of the linear gravity anomalies. The results were interpreted by comparing with the chronological and KREEP distribution maps on the lunar surface. We suggested providing mechanisms of the magma to these regions and finally discussed whether the volcanisms of these multiple volcano complex regions are related with each other or not.

  17. Neighborhood Contributions to Racial and Ethnic Disparities in Obesity Among New York City Adults.

    PubMed

    Lim, Sungwoo; Harris, Tiffany G

    2015-01-01

    Objectives. We assessed neighborhood confounding on racial/ethnic obesity disparities among adults in New York City after accounting for complex sampling, and how much neighborhood factors (walkability, percentage Black or Hispanic, poverty) contributed to this effect. Methods. We combined New York City Community Health Survey 2002-2004 data with Census 2000 zip code-level data. We estimated odds ratios (ORs) for obesity with 2 sets of regression analyses. First, we used the method incorporating the conditional pseudolikelihood into complex sample adjustment. Second, we compared ORs for race/ethnicity from a conventional multilevel model for each neighborhood factor with those from a hybrid fixed-effect model. Results. The weighted estimate for obesity for Blacks versus Whites (OR = 1.8; 95% confidence interval = 1.6, 2.0) was attenuated when we controlled neighborhood confounding (OR = 1.4; 95% confidence interval = 1.2, 1.6; first analysis). Percentage of Blacks in the neighborhood made a large contribution whereas the walkability contribution was minimal (second analysis). Conclusions. Percentage of Blacks in New York City neighborhoods explained a large portion of the disparity in obesity between Blacks and Whites. The study highlights the importance of estimating valid neighborhood effects for public health surveillance and intervention.

  18. Navigating complex sample analysis using national survey data.

    PubMed

    Saylor, Jennifer; Friedmann, Erika; Lee, Hyeon Joo

    2012-01-01

    The National Center for Health Statistics conducts the National Health and Nutrition Examination Survey and other national surveys with probability-based complex sample designs. Goals of national surveys are to provide valid data for the population of the United States. Analyses of data from population surveys present unique challenges in the research process but are valuable avenues to study the health of the United States population. The aim of this study was to demonstrate the importance of using complex data analysis techniques for data obtained with complex multistage sampling design and provide an example of analysis using the SPSS Complex Samples procedure. Illustration of challenges and solutions specific to secondary data analysis of national databases are described using the National Health and Nutrition Examination Survey as the exemplar. Oversampling of small or sensitive groups provides necessary estimates of variability within small groups. Use of weights without complex samples accurately estimates population means and frequency from the sample after accounting for over- or undersampling of specific groups. Weighting alone leads to inappropriate population estimates of variability, because they are computed as if the measures were from the entire population rather than a sample in the data set. The SPSS Complex Samples procedure allows inclusion of all sampling design elements, stratification, clusters, and weights. Use of national data sets allows use of extensive, expensive, and well-documented survey data for exploratory questions but limits analysis to those variables included in the data set. The large sample permits examination of multiple predictors and interactive relationships. Merging data files, availability of data in several waves of surveys, and complex sampling are techniques used to provide a representative sample but present unique challenges. In sophisticated data analysis techniques, use of these data is optimized.

  19. The Effect of Black Peers on Black Test Scores

    ERIC Educational Resources Information Center

    Armor, David J.; Duck, Stephanie

    2007-01-01

    Recent studies have used increasingly complex methodologies to estimate the effect of peer characteristics--race, poverty, and ability--on student achievement. A paper by Hanushek, Kain, and Rivkin using Texas state testing data has received particularly wide attention because it found a large negative effect of school percent black on black math…

  20. Balancing Computer Resources with Institutional Needs. AIR Forum Paper 1978.

    ERIC Educational Resources Information Center

    McLaughlin, Gerald W.; And Others

    To estimate computer needs at a higher education institution, the major types of users and their future needs should be determined. In a large or complex university, three major groups of users are typically instructional, research, and administrative. After collecting information on the needs of these users, the needs can be translated into…

  1. DOE Office of Scientific and Technical Information (OSTI.GOV)

    Eken Tuna, Kevin Mayeda, Abraham Hofstetter, Rengin Gok, Gonca Orgulu, Niyazi Turkelli

    A recently developed coda magnitude methodology was applied to selected broadband stations in Turkey for the purpose of testing the coda method in a large, laterally complex region. As found in other, albeit smaller regions, coda envelope amplitude measurements are significantly less variable than distance-corrected direct wave measurements (i.e., L{sub g} and surface waves) by roughly a factor 3-to-4. Despite strong lateral crustal heterogeneity in Turkey, they found that the region could be adequately modeled assuming a simple 1-D, radially symmetric path correction. After calibrating the stations ISP, ISKB and MALT for local and regional distances, single-station moment-magnitude estimates (M{submore » W}) derived from the coda spectra were in excellent agreement with those determined from multistation waveform modeling inversions, exhibiting a data standard deviation of 0.17. Though the calibration was validated using large events, the results of the calibration will extend M{sub W} estimates to significantly smaller events which could not otherwise be waveform modeled. The successful application of the method is remarkable considering the significant lateral complexity in Turkey and the simple assumptions used in the coda method.« less

  2. Lexical decision as an endophenotype for reading comprehension: An exploration of an association

    PubMed Central

    NAPLES, ADAM; KATZ, LEN; GRIGORENKO, ELENA L.

    2012-01-01

    Based on numerous suggestions in the literature, we evaluated lexical decision (LD) as a putative endophenotype for reading comprehension by investigating heritability estimates and segregation analyses parameter estimates for both of these phenotypes. Specifically, in a segregation analysis of a large sample of families, we established that there is little to no overlap between genes contributing to LD and reading comprehension and that the genetic mechanism behind LD derived from this analysis appears to be more complex than that for reading comprehension. We conclude that in our sample, LD is not a good candidate as an endophenotype for reading comprehension, despite previous suggestions from the literature. Based on this conclusion, we discuss the role and benefit of the endophenotype approach in studies of complex human cognitive functions. PMID:23062302

  3. Using constraints and their value for optimization of large ODE systems

    PubMed Central

    Domijan, Mirela; Rand, David A.

    2015-01-01

    We provide analytical tools to facilitate a rigorous assessment of the quality and value of the fit of a complex model to data. We use this to provide approaches to model fitting, parameter estimation, the design of optimization functions and experimental optimization. This is in the context where multiple constraints are used to select or optimize a large model defined by differential equations. We illustrate the approach using models of circadian clocks and the NF-κB signalling system. PMID:25673300

  4. A simple method for estimating the size of nuclei on fractal surfaces

    NASA Astrophysics Data System (ADS)

    Zeng, Qiang

    2017-10-01

    Determining the size of nuclei on complex surfaces remains a big challenge in aspects of biological, material and chemical engineering. Here the author reported a simple method to estimate the size of the nuclei in contact with complex (fractal) surfaces. The established approach was based on the assumptions of contact area proportionality for determining nucleation density and the scaling congruence between nuclei and surfaces for identifying contact regimes. It showed three different regimes governing the equations for estimating the nucleation site density. Nuclei in the size large enough could eliminate the effect of fractal structure. Nuclei in the size small enough could lead to the independence of nucleation site density on fractal parameters. Only when nuclei match the fractal scales, the nucleation site density is associated with the fractal parameters and the size of the nuclei in a coupling pattern. The method was validated by the experimental data reported in the literature. The method may provide an effective way to estimate the size of nuclei on fractal surfaces, through which a number of promising applications in relative fields can be envisioned.

  5. STRUCTURE IN THE 3D GALAXY DISTRIBUTION: III. FOURIER TRANSFORMING THE UNIVERSE: PHASE AND POWER SPECTRA.

    PubMed

    Scargle, Jeffrey D; Way, M J; Gazis, P R

    2017-04-10

    We demonstrate the effectiveness of a relatively straightforward analysis of the complex 3D Fourier transform of galaxy coordinates derived from redshift surveys. Numerical demonstrations of this approach are carried out on a volume-limited sample of the Sloan Digital Sky Survey redshift survey. The direct unbinned transform yields a complex 3D data cube quite similar to that from the Fast Fourier Transform (FFT) of finely binned galaxy positions. In both cases deconvolution of the sampling window function yields estimates of the true transform. Simple power spectrum estimates from these transforms are roughly consistent with those using more elaborate methods. The complex Fourier transform characterizes spatial distributional properties beyond the power spectrum in a manner different from (and we argue is more easily interpreted than) the conventional multi-point hierarchy. We identify some threads of modern large scale inference methodology that will presumably yield detections in new wider and deeper surveys.

  6. STRUCTURE IN THE 3D GALAXY DISTRIBUTION: III. FOURIER TRANSFORMING THE UNIVERSE: PHASE AND POWER SPECTRA

    PubMed Central

    Scargle, Jeffrey D.; Way, M. J.; Gazis, P. R.

    2017-01-01

    We demonstrate the effectiveness of a relatively straightforward analysis of the complex 3D Fourier transform of galaxy coordinates derived from redshift surveys. Numerical demonstrations of this approach are carried out on a volume-limited sample of the Sloan Digital Sky Survey redshift survey. The direct unbinned transform yields a complex 3D data cube quite similar to that from the Fast Fourier Transform (FFT) of finely binned galaxy positions. In both cases deconvolution of the sampling window function yields estimates of the true transform. Simple power spectrum estimates from these transforms are roughly consistent with those using more elaborate methods. The complex Fourier transform characterizes spatial distributional properties beyond the power spectrum in a manner different from (and we argue is more easily interpreted than) the conventional multi-point hierarchy. We identify some threads of modern large scale inference methodology that will presumably yield detections in new wider and deeper surveys. PMID:29628519

  7. Structure in the 3D Galaxy Distribution. III. Fourier Transforming the Universe: Phase and Power Spectra

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Scargle, Jeffrey D.; Way, M. J.; Gazis, P. R., E-mail: Jeffrey.D.Scargle@nasa.gov, E-mail: Michael.J.Way@nasa.gov, E-mail: PGazis@sbcglobal.net

    We demonstrate the effectiveness of a relatively straightforward analysis of the complex 3D Fourier transform of galaxy coordinates derived from redshift surveys. Numerical demonstrations of this approach are carried out on a volume-limited sample of the Sloan Digital Sky Survey redshift survey. The direct unbinned transform yields a complex 3D data cube quite similar to that from the Fast Fourier Transform of finely binned galaxy positions. In both cases, deconvolution of the sampling window function yields estimates of the true transform. Simple power spectrum estimates from these transforms are roughly consistent with those using more elaborate methods. The complex Fouriermore » transform characterizes spatial distributional properties beyond the power spectrum in a manner different from (and we argue is more easily interpreted than) the conventional multipoint hierarchy. We identify some threads of modern large-scale inference methodology that will presumably yield detections in new wider and deeper surveys.« less

  8. Structure in the 3D Galaxy Distribution: III. Fourier Transforming the Universe: Phase and Power Spectra

    NASA Technical Reports Server (NTRS)

    Scargle, Jeffrey D.; Way, M. J.; Gazis, P. R.

    2017-01-01

    We demonstrate the effectiveness of a relatively straightforward analysis of the complex 3D Fourier transform of galaxy coordinates derived from redshift surveys. Numerical demonstrations of this approach are carried out on a volume-limited sample of the Sloan Digital Sky Survey redshift survey. The direct unbinned transform yields a complex 3D data cube quite similar to that from the Fast Fourier Transform (FFT) of finely binned galaxy positions. In both cases deconvolution of the sampling window function yields estimates of the true transform. Simple power spectrum estimates from these transforms are roughly consistent with those using more elaborate methods. The complex Fourier transform characterizes spatial distributional properties beyond the power spectrum in a manner different from (and we argue is more easily interpreted than) the conventional multi-point hierarchy. We identify some threads of modern large scale inference methodology that will presumably yield detections in new wider and deeper surveys.

  9. Estimating Lion Abundance using N-mixture Models for Social Species

    PubMed Central

    Belant, Jerrold L.; Bled, Florent; Wilton, Clay M.; Fyumagwa, Robert; Mwampeta, Stanslaus B.; Beyer, Dean E.

    2016-01-01

    Declining populations of large carnivores worldwide, and the complexities of managing human-carnivore conflicts, require accurate population estimates of large carnivores to promote their long-term persistence through well-informed management We used N-mixture models to estimate lion (Panthera leo) abundance from call-in and track surveys in southeastern Serengeti National Park, Tanzania. Because of potential habituation to broadcasted calls and social behavior, we developed a hierarchical observation process within the N-mixture model conditioning lion detectability on their group response to call-ins and individual detection probabilities. We estimated 270 lions (95% credible interval = 170–551) using call-ins but were unable to estimate lion abundance from track data. We found a weak negative relationship between predicted track density and predicted lion abundance from the call-in surveys. Luminosity was negatively correlated with individual detection probability during call-in surveys. Lion abundance and track density were influenced by landcover, but direction of the corresponding effects were undetermined. N-mixture models allowed us to incorporate multiple parameters (e.g., landcover, luminosity, observer effect) influencing lion abundance and probability of detection directly into abundance estimates. We suggest that N-mixture models employing a hierarchical observation process can be used to estimate abundance of other social, herding, and grouping species. PMID:27786283

  10. Estimating Lion Abundance using N-mixture Models for Social Species.

    PubMed

    Belant, Jerrold L; Bled, Florent; Wilton, Clay M; Fyumagwa, Robert; Mwampeta, Stanslaus B; Beyer, Dean E

    2016-10-27

    Declining populations of large carnivores worldwide, and the complexities of managing human-carnivore conflicts, require accurate population estimates of large carnivores to promote their long-term persistence through well-informed management We used N-mixture models to estimate lion (Panthera leo) abundance from call-in and track surveys in southeastern Serengeti National Park, Tanzania. Because of potential habituation to broadcasted calls and social behavior, we developed a hierarchical observation process within the N-mixture model conditioning lion detectability on their group response to call-ins and individual detection probabilities. We estimated 270 lions (95% credible interval = 170-551) using call-ins but were unable to estimate lion abundance from track data. We found a weak negative relationship between predicted track density and predicted lion abundance from the call-in surveys. Luminosity was negatively correlated with individual detection probability during call-in surveys. Lion abundance and track density were influenced by landcover, but direction of the corresponding effects were undetermined. N-mixture models allowed us to incorporate multiple parameters (e.g., landcover, luminosity, observer effect) influencing lion abundance and probability of detection directly into abundance estimates. We suggest that N-mixture models employing a hierarchical observation process can be used to estimate abundance of other social, herding, and grouping species.

  11. [Review of estimation on oceanic primary productivity by using remote sensing methods.

    PubMed

    Xu, Hong Yun; Zhou, Wei Feng; Ji, Shi Jian

    2016-09-01

    Accuracy estimation of oceanic primary productivity is of great significance in the assessment and management of fisheries resources, marine ecology systems, global change and other fields. The traditional measurement and estimation of oceanic primary productivity has to rely on in situ sample data by vessels. Satellite remote sensing has advantages of providing dynamic and eco-environmental parameters of ocean surface at large scale in real time. Thus, satellite remote sensing has increasingly become an important means for oceanic primary productivity estimation on large spatio-temporal scale. Combining with the development of ocean color sensors, the models to estimate the oceanic primary productivity by satellite remote sensing have been developed that could be mainly summarized as chlorophyll-based, carbon-based and phytoplankton absorption-based approach. The flexibility and complexity of the three kinds of models were presented in the paper. On this basis, the current research status for global estimation of oceanic primary productivity was analyzed and evaluated. In view of these, four research fields needed to be strengthened in further stu-dy: 1) Global oceanic primary productivity estimation should be segmented and studied, 2) to dee-pen the research on absorption coefficient of phytoplankton, 3) to enhance the technology of ocea-nic remote sensing, 4) to improve the in situ measurement of primary productivity.

  12. [Complex estimation of proliferative activity of epithelial cells of the large intestine damaged by polyps and cancer].

    PubMed

    Nalieskina, L A; Zabarko, L B; Polishchuk, L Z; Oliĭnichenko, G P; Zakhartseva, L M; Koshel', K V

    2001-01-01

    Peculiarities of mitotic regime and expression of proliferating cell nuclear antigen were investigated in 18 polyps and 35 cases of colorectal cancer. Direct relationship between spectrum and degree of manifestation of proliferative activity, level of morphological malignant tumors and accumulation of oncopathology in the patient pedigrees was established.

  13. Estimating aboveground net primary productivity in forest-dominated ecosystems

    Treesearch

    Brian D. Kloeppel; Mark E. Harmon; Timothy J. Fahey

    2007-01-01

    The measurement of net primary productivity (NPP) in forest ecosystems presents a variety of challenges because of the large and complex dimensions of trees and the difficulties of quantifying several components of NPP. As summarized by Clark et al. (2001a), these methodological challenges can be overcome, and more reliable spatial and temporal comparisons can be...

  14. Natural Allelic Variations in Highly Polyploidy Saccharum Complex

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Song, Jian; Yang, Xiping; Resende, Jr., Marcio F. R.

    Sugarcane ( Saccharum spp.) is an important sugar and biofuel crop with high polyploid and complex genomes. The Saccharum complex, comprised of Saccharum genus and a few related genera, are important genetic resources for sugarcane breeding. A large amount of natural variation exists within the Saccharum complex. Though understanding their allelic variation has been challenging, it is critical to dissect allelic structure and to identify the alleles controlling important traits in sugarcane. To characterize natural variations in Saccharum complex, a target enrichment sequencing approach was used to assay 12 representative germplasm accessions. In total, 55,946 highly efficient probes were designedmore » based on the sorghum genome and sugarcane unigene set targeting a total of 6 Mb of the sugarcane genome. A pipeline specifically tailored for polyploid sequence variants and genotype calling was established. BWAmem and sorghum genome approved to be an acceptable aligner and reference for sugarcane target enrichment sequence analysis, respectively. Genetic variations including 1,166,066 non-redundant SNPs, 150,421 InDels, 919 gene copy number variations, and 1,257 gene presence/absence variations were detected. SNPs from three different callers (Samtools, Freebayes, and GATK) were compared and the validation rates were nearly 90%. Based on the SNP loci of each accession and their ploidy levels, 999,258 single dosage SNPs were identified and most loci were estimated as largely homozygotes. An average of 34,397 haplotype blocks for each accession was inferred. The highest divergence time among the Saccharum spp. was estimated as 1.2 million years ago (MYA). Saccharum spp. diverged from Erianthus and Sorghum approximately 5 and 6 MYA, respectively. Furthermore, the target enrichment sequencing approach provided an effective way to discover and catalog natural allelic variation in highly polyploid or heterozygous genomes.« less

  15. Natural Allelic Variations in Highly Polyploidy Saccharum Complex

    DOE PAGES

    Song, Jian; Yang, Xiping; Resende, Jr., Marcio F. R.; ...

    2016-06-08

    Sugarcane ( Saccharum spp.) is an important sugar and biofuel crop with high polyploid and complex genomes. The Saccharum complex, comprised of Saccharum genus and a few related genera, are important genetic resources for sugarcane breeding. A large amount of natural variation exists within the Saccharum complex. Though understanding their allelic variation has been challenging, it is critical to dissect allelic structure and to identify the alleles controlling important traits in sugarcane. To characterize natural variations in Saccharum complex, a target enrichment sequencing approach was used to assay 12 representative germplasm accessions. In total, 55,946 highly efficient probes were designedmore » based on the sorghum genome and sugarcane unigene set targeting a total of 6 Mb of the sugarcane genome. A pipeline specifically tailored for polyploid sequence variants and genotype calling was established. BWAmem and sorghum genome approved to be an acceptable aligner and reference for sugarcane target enrichment sequence analysis, respectively. Genetic variations including 1,166,066 non-redundant SNPs, 150,421 InDels, 919 gene copy number variations, and 1,257 gene presence/absence variations were detected. SNPs from three different callers (Samtools, Freebayes, and GATK) were compared and the validation rates were nearly 90%. Based on the SNP loci of each accession and their ploidy levels, 999,258 single dosage SNPs were identified and most loci were estimated as largely homozygotes. An average of 34,397 haplotype blocks for each accession was inferred. The highest divergence time among the Saccharum spp. was estimated as 1.2 million years ago (MYA). Saccharum spp. diverged from Erianthus and Sorghum approximately 5 and 6 MYA, respectively. Furthermore, the target enrichment sequencing approach provided an effective way to discover and catalog natural allelic variation in highly polyploid or heterozygous genomes.« less

  16. Continued investigation of solid propulsion economics. Task 1B: Large solid rocket motor case fabrication methods - Supplement process complexity factor cost technique

    NASA Technical Reports Server (NTRS)

    Baird, J.

    1967-01-01

    This supplement to Task lB-Large Solid Rocket Motor Case Fabrication Methods supplies additional supporting cost data and discusses in detail the methodology that was applied to the task. For the case elements studied, the cost was found to be directly proportional to the Process Complexity Factor (PCF). The PCF was obtained for each element by identifying unit processes that are common to the elements and their alternative manufacturing routes, by assigning a weight to each unit process, and by summing the weighted counts. In three instances of actual manufacture, the actual cost per pound equaled the cost estimate based on PCF per pound, but this supplement, recognizes that the methodology is of limited, rather than general, application.

  17. Effective concentration as a tool for quantitatively addressing preorganization in multicomponent assemblies: application to the selective complexation of lanthanide cations.

    PubMed

    Canard, Gabriel; Koeller, Sylvain; Bernardinelli, Gérald; Piguet, Claude

    2008-01-23

    The beneficial entropic effect, which may be expected from the connection of three tridentate binding units to a strain-free covalent tripod for complexing nine-coordinate cations (Mz+ = Ca2+, La3+, Eu3+, Lu3+), is quantitatively analyzed by using a simple thermodynamic additive model. The switch from pure intermolecular binding processes, characterizing the formation of the triple-helical complexes [M(L2)3]z+, to a combination of inter- and intramolecular complexation events in [M(L8)]z+ shows that the ideal structural fit observed in [M(L8)]z+ indeed masks large energetic constraints. This limitation is evidenced by the faint effective concentrations, ceff, which control the intramolecular ring-closing reactions operating in [M(L8)]z+. This predominence of the thermodynamic approach over the usual structural analysis agrees with the hierarchical relationships linking energetics and structures. Its simple estimation by using a single microscopic parameter, ceff, opens novel perspectives for the molecular tuning of specific receptors for the recognition of large cations, a crucial point for the programming of heterometallic f-f complexes under thermodynamic control.

  18. On the Origins of Suboptimality in Human Probabilistic Inference

    PubMed Central

    Acerbi, Luigi; Vijayakumar, Sethu; Wolpert, Daniel M.

    2014-01-01

    Humans have been shown to combine noisy sensory information with previous experience (priors), in qualitative and sometimes quantitative agreement with the statistically-optimal predictions of Bayesian integration. However, when the prior distribution becomes more complex than a simple Gaussian, such as skewed or bimodal, training takes much longer and performance appears suboptimal. It is unclear whether such suboptimality arises from an imprecise internal representation of the complex prior, or from additional constraints in performing probabilistic computations on complex distributions, even when accurately represented. Here we probe the sources of suboptimality in probabilistic inference using a novel estimation task in which subjects are exposed to an explicitly provided distribution, thereby removing the need to remember the prior. Subjects had to estimate the location of a target given a noisy cue and a visual representation of the prior probability density over locations, which changed on each trial. Different classes of priors were examined (Gaussian, unimodal, bimodal). Subjects' performance was in qualitative agreement with the predictions of Bayesian Decision Theory although generally suboptimal. The degree of suboptimality was modulated by statistical features of the priors but was largely independent of the class of the prior and level of noise in the cue, suggesting that suboptimality in dealing with complex statistical features, such as bimodality, may be due to a problem of acquiring the priors rather than computing with them. We performed a factorial model comparison across a large set of Bayesian observer models to identify additional sources of noise and suboptimality. Our analysis rejects several models of stochastic behavior, including probability matching and sample-averaging strategies. Instead we show that subjects' response variability was mainly driven by a combination of a noisy estimation of the parameters of the priors, and by variability in the decision process, which we represent as a noisy or stochastic posterior. PMID:24945142

  19. Localization Algorithm Based on a Spring Model (LASM) for Large Scale Wireless Sensor Networks.

    PubMed

    Chen, Wanming; Mei, Tao; Meng, Max Q-H; Liang, Huawei; Liu, Yumei; Li, Yangming; Li, Shuai

    2008-03-15

    A navigation method for a lunar rover based on large scale wireless sensornetworks is proposed. To obtain high navigation accuracy and large exploration area, highnode localization accuracy and large network scale are required. However, thecomputational and communication complexity and time consumption are greatly increasedwith the increase of the network scales. A localization algorithm based on a spring model(LASM) method is proposed to reduce the computational complexity, while maintainingthe localization accuracy in large scale sensor networks. The algorithm simulates thedynamics of physical spring system to estimate the positions of nodes. The sensor nodesare set as particles with masses and connected with neighbor nodes by virtual springs. Thevirtual springs will force the particles move to the original positions, the node positionscorrespondingly, from the randomly set positions. Therefore, a blind node position can bedetermined from the LASM algorithm by calculating the related forces with the neighbornodes. The computational and communication complexity are O(1) for each node, since thenumber of the neighbor nodes does not increase proportionally with the network scale size.Three patches are proposed to avoid local optimization, kick out bad nodes and deal withnode variation. Simulation results show that the computational and communicationcomplexity are almost constant despite of the increase of the network scale size. The time consumption has also been proven to remain almost constant since the calculation steps arealmost unrelated with the network scale size.

  20. Initiating Molecular Growth in the Interstellar Medium via Dimeric Complexes of Observed Ions and Molecules

    NASA Technical Reports Server (NTRS)

    Bera, Partha P.; Head-Gordon, Martin; Lee, Timothy J.

    2011-01-01

    A feasible initiation step for particle growth in the interstellar medium (ISM) is simulated by means of ab quantum chemistry methods. The systems studied are dimer ions formed by pairing nitrogen containing small molecules known to exist in the ISM with ions of unsaturated hydrocarbons or vice versa. Complexation energies, structures of ensuing complexes and electronic excitation spectra of the encounter complexes are estimated using various quantum chemistry methods. Moller-Plesset perturbation theory (MP2, Z-averaged perturbation theory (ZAP2), coupled cluster singles and doubles with perturbative triples corrections (CCSD(T)), and density functional theory (DFT) methods (B3LYP) were employed along with the correlation consistent cc-pVTZ and aug-cc-pVTZ basis sets. Two types of complexes are predicted. One type of complex has electrostatic binding with moderate (7-20 kcal per mol) binding energies, that are nonetheless significantly stronger than typical van der Waals interactions between molecules of this size. The other type of complex develops strong covalent bonds between the fragments. Cyclic isomers of the nitrogen containing complexes are produced very easily by ion-molecule reactions. Some of these complexes show intense ultraviolet visible spectra for electronic transitions with large oscillator strengths at the B3LYP, omegaB97, and equations of motion coupled cluster (EOM-CCSD) levels. The open shell nitrogen containing carbonaceous complexes especially exhibit a large oscillator strength electronic transition in the visible region of the electromagnetic spectrum.

  1. Computationally efficient confidence intervals for cross-validated area under the ROC curve estimates.

    PubMed

    LeDell, Erin; Petersen, Maya; van der Laan, Mark

    In binary classification problems, the area under the ROC curve (AUC) is commonly used to evaluate the performance of a prediction model. Often, it is combined with cross-validation in order to assess how the results will generalize to an independent data set. In order to evaluate the quality of an estimate for cross-validated AUC, we obtain an estimate of its variance. For massive data sets, the process of generating a single performance estimate can be computationally expensive. Additionally, when using a complex prediction method, the process of cross-validating a predictive model on even a relatively small data set can still require a large amount of computation time. Thus, in many practical settings, the bootstrap is a computationally intractable approach to variance estimation. As an alternative to the bootstrap, we demonstrate a computationally efficient influence curve based approach to obtaining a variance estimate for cross-validated AUC.

  2. Computationally efficient confidence intervals for cross-validated area under the ROC curve estimates

    PubMed Central

    Petersen, Maya; van der Laan, Mark

    2015-01-01

    In binary classification problems, the area under the ROC curve (AUC) is commonly used to evaluate the performance of a prediction model. Often, it is combined with cross-validation in order to assess how the results will generalize to an independent data set. In order to evaluate the quality of an estimate for cross-validated AUC, we obtain an estimate of its variance. For massive data sets, the process of generating a single performance estimate can be computationally expensive. Additionally, when using a complex prediction method, the process of cross-validating a predictive model on even a relatively small data set can still require a large amount of computation time. Thus, in many practical settings, the bootstrap is a computationally intractable approach to variance estimation. As an alternative to the bootstrap, we demonstrate a computationally efficient influence curve based approach to obtaining a variance estimate for cross-validated AUC. PMID:26279737

  3. Space Weather Studies Using Ground-based Experimental Complex in Kazakhstan

    NASA Astrophysics Data System (ADS)

    Kryakunova, O.; Yakovets, A.; Monstein, C.; Nikolayevskiy, N.; Zhumabayev, B.; Gordienko, G.; Andreyev, A.; Malimbayev, A.; Levin, Yu.; Salikhov, N.; Sokolova, O.; Tsepakina, I.

    2015-12-01

    Kazakhstan ground-based experimental complex for space weather study is situated near Almaty. Results of space environment monitoring are accessible via Internet on the web-site of the Institute of Ionosphere (http://www.ionos.kz/?q=en/node/21) in real time. There is a complex database with hourly data of cosmic ray intensity, geomagnetic field intensity, and solar radio flux at 10.7 cm and 27.8 cm wavelengths. Several studies using those data are reported. They are an estimation of speed of a coronal mass ejection, a study of large scale traveling distrubances, an analysis of geomagnetically induced currents using the geomagnetic field data, and a solar energetic proton event on 27 January 2012.

  4. Complexity is simple!

    NASA Astrophysics Data System (ADS)

    Cottrell, William; Montero, Miguel

    2018-02-01

    In this note we investigate the role of Lloyd's computational bound in holographic complexity. Our goal is to translate the assumptions behind Lloyd's proof into the bulk language. In particular, we discuss the distinction between orthogonalizing and `simple' gates and argue that these notions are useful for diagnosing holographic complexity. We show that large black holes constructed from series circuits necessarily employ simple gates, and thus do not satisfy Lloyd's assumptions. We also estimate the degree of parallel processing required in this case for elementary gates to orthogonalize. Finally, we show that for small black holes at fixed chemical potential, the orthogonalization condition is satisfied near the phase transition, supporting a possible argument for the Weak Gravity Conjecture first advocated in [1].

  5. Fast and Epsilon-Optimal Discretized Pursuit Learning Automata.

    PubMed

    Zhang, JunQi; Wang, Cheng; Zhou, MengChu

    2015-10-01

    Learning automata (LA) are powerful tools for reinforcement learning. A discretized pursuit LA is the most popular one among them. During an iteration its operation consists of three basic phases: 1) selecting the next action; 2) finding the optimal estimated action; and 3) updating the state probability. However, when the number of actions is large, the learning becomes extremely slow because there are too many updates to be made at each iteration. The increased updates are mostly from phases 1 and 3. A new fast discretized pursuit LA with assured ε -optimality is proposed to perform both phases 1 and 3 with the computational complexity independent of the number of actions. Apart from its low computational complexity, it achieves faster convergence speed than the classical one when operating in stationary environments. This paper can promote the applications of LA toward the large-scale-action oriented area that requires efficient reinforcement learning tools with assured ε -optimality, fast convergence speed, and low computational complexity for each iteration.

  6. Automated Determination of Magnitude and Source Length of Large Earthquakes

    NASA Astrophysics Data System (ADS)

    Wang, D.; Kawakatsu, H.; Zhuang, J.; Mori, J. J.; Maeda, T.; Tsuruoka, H.; Zhao, X.

    2017-12-01

    Rapid determination of earthquake magnitude is of importance for estimating shaking damages, and tsunami hazards. However, due to the complexity of source process, accurately estimating magnitude for great earthquakes in minutes after origin time is still a challenge. Mw is an accurate estimate for large earthquakes. However, calculating Mw requires the whole wave trains including P, S, and surface phases, which takes tens of minutes to reach stations at tele-seismic distances. To speed up the calculation, methods using W phase and body wave are developed for fast estimating earthquake sizes. Besides these methods that involve Green's Functions and inversions, there are other approaches that use empirically simulated relations to estimate earthquake magnitudes, usually for large earthquakes. The nature of simple implementation and straightforward calculation made these approaches widely applied at many institutions such as the Pacific Tsunami Warning Center, the Japan Meteorological Agency, and the USGS. Here we developed an approach that was originated from Hara [2007], estimating magnitude by considering P-wave displacement and source duration. We introduced a back-projection technique [Wang et al., 2016] instead to estimate source duration using array data from a high-sensitive seismograph network (Hi-net). The introduction of back-projection improves the method in two ways. Firstly, the source duration could be accurately determined by seismic array. Secondly, the results can be more rapidly calculated, and data derived from farther stations are not required. We purpose to develop an automated system for determining fast and reliable source information of large shallow seismic events based on real time data of a dense regional array and global data, for earthquakes that occur at distance of roughly 30°- 85° from the array center. This system can offer fast and robust estimates of magnitudes and rupture extensions of large earthquakes in 6 to 13 min (plus source duration time) depending on the epicenter distances. It may be a promising aid for disaster mitigation right after a damaging earthquake, especially when dealing with the tsunami evacuation and emergency rescue.

  7. Automated Determination of Magnitude and Source Extent of Large Earthquakes

    NASA Astrophysics Data System (ADS)

    Wang, Dun

    2017-04-01

    Rapid determination of earthquake magnitude is of importance for estimating shaking damages, and tsunami hazards. However, due to the complexity of source process, accurately estimating magnitude for great earthquakes in minutes after origin time is still a challenge. Mw is an accurate estimate for large earthquakes. However, calculating Mw requires the whole wave trains including P, S, and surface phases, which takes tens of minutes to reach stations at tele-seismic distances. To speed up the calculation, methods using W phase and body wave are developed for fast estimating earthquake sizes. Besides these methods that involve Green's Functions and inversions, there are other approaches that use empirically simulated relations to estimate earthquake magnitudes, usually for large earthquakes. The nature of simple implementation and straightforward calculation made these approaches widely applied at many institutions such as the Pacific Tsunami Warning Center, the Japan Meteorological Agency, and the USGS. Here we developed an approach that was originated from Hara [2007], estimating magnitude by considering P-wave displacement and source duration. We introduced a back-projection technique [Wang et al., 2016] instead to estimate source duration using array data from a high-sensitive seismograph network (Hi-net). The introduction of back-projection improves the method in two ways. Firstly, the source duration could be accurately determined by seismic array. Secondly, the results can be more rapidly calculated, and data derived from farther stations are not required. We purpose to develop an automated system for determining fast and reliable source information of large shallow seismic events based on real time data of a dense regional array and global data, for earthquakes that occur at distance of roughly 30°- 85° from the array center. This system can offer fast and robust estimates of magnitudes and rupture extensions of large earthquakes in 6 to 13 min (plus source duration time) depending on the epicenter distances. It may be a promising aid for disaster mitigation right after a damaging earthquake, especially when dealing with the tsunami evacuation and emergency rescue.

  8. The effects of numerical-model complexity and observation type on estimated porosity values

    USGS Publications Warehouse

    Starn, Jeffrey; Bagtzoglou, Amvrossios C.; Green, Christopher T.

    2015-01-01

    The relative merits of model complexity and types of observations employed in model calibration are compared. An existing groundwater flow model coupled with an advective transport simulation of the Salt Lake Valley, Utah (USA), is adapted for advective transport, and effective porosity is adjusted until simulated tritium concentrations match concentrations in samples from wells. Two calibration approaches are used: a “complex” highly parameterized porosity field and a “simple” parsimonious model of porosity distribution. The use of an atmospheric tracer (tritium in this case) and apparent ages (from tritium/helium) in model calibration also are discussed. Of the models tested, the complex model (with tritium concentrations and tritium/helium apparent ages) performs best. Although tritium breakthrough curves simulated by complex and simple models are very generally similar, and there is value in the simple model, the complex model is supported by a more realistic porosity distribution and a greater number of estimable parameters. Culling the best quality data did not lead to better calibration, possibly because of processes and aquifer characteristics that are not simulated. Despite many factors that contribute to shortcomings of both the models and the data, useful information is obtained from all the models evaluated. Although any particular prediction of tritium breakthrough may have large errors, overall, the models mimic observed trends.

  9. Methodology used in Cuba for estimating economic losses caused by forest fires

    Treesearch

    Marcos Pedro Ramos Rodríguez; Raúl González Rodríguez

    2013-01-01

    Assessment of economic losses caused by forest fires is a highly complex but important activity. It is complicated first by the large number of effects, in different periods, brought about in the social, economic and environmental fields. Secondly, the difficulty of assigning a market value to resources such as biodiversity or endangered species should be mentioned. It...

  10. Going to the Source: A Practical Way to Simplify the FAFSA

    ERIC Educational Resources Information Center

    Asher, Lauren

    2007-01-01

    There is widespread agreement that the complexity of the current Free Application for Federal Student Aid (FAFSA) is a barrier to college access and success. One indication is the large and growing number of lower income college students who do not apply for aid, even though they are likely eligible for a Pell grant: an estimated 1.5 million in…

  11. Modeling the Psychometric Properties of Complex Performance Assessment Tasks Using Confirmatory Factor Analysis: A Multistage Model for Calibrating Tasks

    ERIC Educational Resources Information Center

    Kahraman, Nilufer; De Champlain, Andre; Raymond, Mark

    2012-01-01

    Item-level information, such as difficulty and discrimination are invaluable to the test assembly, equating, and scoring practices. Estimating these parameters within the context of large-scale performance assessments is often hindered by the use of unbalanced designs for assigning examinees to tasks and raters because such designs result in very…

  12. Spatial allocation of market and nonmarket values in wildland fire management: A case study

    Treesearch

    John W. Benoit; Armando González-Cabán; Francis M. Fujioka; Shyh-Chin Chen; José J. Sanchez

    2013-01-01

    We developed a methodology to evaluate the efficacy of fuel treatments by estimating their costs and potential costs/losses with and without treatments in the San Jacinto Ranger District of the San Bernardino National Forest, California. This district is a typical southern California forest complex containing a large amount of high-valued real estate. We chose four...

  13. A measurement system for large, complex software programs

    NASA Technical Reports Server (NTRS)

    Rone, Kyle Y.; Olson, Kitty M.; Davis, Nathan E.

    1994-01-01

    This paper describes measurement systems required to forecast, measure, and control activities for large, complex software development and support programs. Initial software cost and quality analysis provides the foundation for meaningful management decisions as a project evolves. In modeling the cost and quality of software systems, the relationship between the functionality, quality, cost, and schedule of the product must be considered. This explicit relationship is dictated by the criticality of the software being developed. This balance between cost and quality is a viable software engineering trade-off throughout the life cycle. Therefore, the ability to accurately estimate the cost and quality of software systems is essential to providing reliable software on time and within budget. Software cost models relate the product error rate to the percent of the project labor that is required for independent verification and validation. The criticality of the software determines which cost model is used to estimate the labor required to develop the software. Software quality models yield an expected error discovery rate based on the software size, criticality, software development environment, and the level of competence of the project and developers with respect to the processes being employed.

  14. Flattening of Caribbean coral reefs: region-wide declines in architectural complexity

    PubMed Central

    Alvarez-Filip, Lorenzo; Dulvy, Nicholas K.; Gill, Jennifer A.; Côté, Isabelle M.; Watkinson, Andrew R.

    2009-01-01

    Coral reefs are rich in biodiversity, in large part because their highly complex architecture provides shelter and resources for a wide range of organisms. Recent rapid declines in hard coral cover have occurred across the Caribbean region, but the concomitant consequences for reef architecture have not been quantified on a large scale to date. We provide, to our knowledge, the first region-wide analysis of changes in reef architectural complexity, using nearly 500 surveys across 200 reefs, between 1969 and 2008. The architectural complexity of Caribbean reefs has declined nonlinearly with the near disappearance of the most complex reefs over the last 40 years. The flattening of Caribbean reefs was apparent by the early 1980s, followed by a period of stasis between 1985 and 1998 and then a resumption of the decline in complexity to the present. Rates of loss are similar on shallow (<6 m), mid-water (6–20 m) and deep (>20 m) reefs and are consistent across all five subregions. The temporal pattern of declining architecture coincides with key events in recent Caribbean ecological history: the loss of structurally complex Acropora corals, the mass mortality of the grazing urchin Diadema antillarum and the 1998 El Nino Southern Oscillation-induced worldwide coral bleaching event. The consistently low estimates of current architectural complexity suggest regional-scale degradation and homogenization of reef structure. The widespread loss of architectural complexity is likely to have serious consequences for reef biodiversity, ecosystem functioning and associated environmental services. PMID:19515663

  15. Empirical Bayes Estimation of Semi-parametric Hierarchical Mixture Models for Unbiased Characterization of Polygenic Disease Architectures

    PubMed Central

    Nishino, Jo; Kochi, Yuta; Shigemizu, Daichi; Kato, Mamoru; Ikari, Katsunori; Ochi, Hidenori; Noma, Hisashi; Matsui, Kota; Morizono, Takashi; Boroevich, Keith A.; Tsunoda, Tatsuhiko; Matsui, Shigeyuki

    2018-01-01

    Genome-wide association studies (GWAS) suggest that the genetic architecture of complex diseases consists of unexpectedly numerous variants with small effect sizes. However, the polygenic architectures of many diseases have not been well characterized due to lack of simple and fast methods for unbiased estimation of the underlying proportion of disease-associated variants and their effect-size distribution. Applying empirical Bayes estimation of semi-parametric hierarchical mixture models to GWAS summary statistics, we confirmed that schizophrenia was extremely polygenic [~40% of independent genome-wide SNPs are risk variants, most within odds ratio (OR = 1.03)], whereas rheumatoid arthritis was less polygenic (~4 to 8% risk variants, significant portion reaching OR = 1.05 to 1.1). For rheumatoid arthritis, stratified estimations revealed that expression quantitative loci in blood explained large genetic variance, and low- and high-frequency derived alleles were prone to be risk and protective, respectively, suggesting a predominance of deleterious-risk and advantageous-protective mutations. Despite genetic correlation, effect-size distributions for schizophrenia and bipolar disorder differed across allele frequency. These analyses distinguished disease polygenic architectures and provided clues for etiological differences in complex diseases. PMID:29740473

  16. Molecular counting by photobleaching in protein complexes with many subunits: best practices and application to the cellulose synthesis complex

    PubMed Central

    Chen, Yalei; Deffenbaugh, Nathan C.; Anderson, Charles T.; Hancock, William O.

    2014-01-01

    The constituents of large, multisubunit protein complexes dictate their functions in cells, but determining their precise molecular makeup in vivo is challenging. One example of such a complex is the cellulose synthesis complex (CSC), which in plants synthesizes cellulose, the most abundant biopolymer on Earth. In growing plant cells, CSCs exist in the plasma membrane as six-lobed rosettes that contain at least three different cellulose synthase (CESA) isoforms, but the number and stoichiometry of CESAs in each CSC are unknown. To begin to address this question, we performed quantitative photobleaching of GFP-tagged AtCESA3-containing particles in living Arabidopsis thaliana cells using variable-angle epifluorescence microscopy and developed a set of information-based step detection procedures to estimate the number of GFP molecules in each particle. The step detection algorithms account for changes in signal variance due to changing numbers of fluorophores, and the subsequent analysis avoids common problems associated with fitting multiple Gaussian functions to binned histogram data. The analysis indicates that at least 10 GFP-AtCESA3 molecules can exist in each particle. These procedures can be applied to photobleaching data for any protein complex with large numbers of fluorescently tagged subunits, providing a new analytical tool with which to probe complex composition and stoichiometry. PMID:25232006

  17. Molecular counting by photobleaching in protein complexes with many subunits: best practices and application to the cellulose synthesis complex

    DOE PAGES

    Chen, Yalei; Deffenbaugh, Nathan C.; Anderson, Charles T.; ...

    2014-09-17

    The constituents of large, multisubunit protein complexes dictate their functions in cells, but determining their precise molecular makeup in vivo is challenging. One example of such a complex is the cellulose synthesis complex (CSC), which in plants synthesizes cellulose, the most abundant biopolymer on Earth. In growing plant cells, CSCs exist in the plasma membrane as six-lobed rosettes that contain at least three different cellulose synthase (CESA) isoforms, but the number and stoichiometry of CESAs in each CSC are unknown. To begin to address this question, we performed quantitative photobleaching of GFP-tagged AtCESA3-containing particles in living Arabidopsis thaliana cells usingmore » variable-angle epifluorescence microscopy and developed a set of information-based step detection procedures to estimate the number of GFP molecules in each particle. The step detection algorithms account for changes in signal variance due to changing numbers of fluorophores, and the subsequent analysis avoids common problems associated with fitting multiple Gaussian functions to binned histogram data. The analysis indicates that at least 10 GFP-AtCESA3 molecules can exist in each particle. In conclusion, these procedures can be applied to photobleaching data for any protein complex with large numbers of fluorescently tagged subunits, providing a new analytical tool with which to probe complex composition and stoichiometry.« less

  18. Scalable Parameter Estimation for Genome-Scale Biochemical Reaction Networks

    PubMed Central

    Kaltenbacher, Barbara; Hasenauer, Jan

    2017-01-01

    Mechanistic mathematical modeling of biochemical reaction networks using ordinary differential equation (ODE) models has improved our understanding of small- and medium-scale biological processes. While the same should in principle hold for large- and genome-scale processes, the computational methods for the analysis of ODE models which describe hundreds or thousands of biochemical species and reactions are missing so far. While individual simulations are feasible, the inference of the model parameters from experimental data is computationally too intensive. In this manuscript, we evaluate adjoint sensitivity analysis for parameter estimation in large scale biochemical reaction networks. We present the approach for time-discrete measurement and compare it to state-of-the-art methods used in systems and computational biology. Our comparison reveals a significantly improved computational efficiency and a superior scalability of adjoint sensitivity analysis. The computational complexity is effectively independent of the number of parameters, enabling the analysis of large- and genome-scale models. Our study of a comprehensive kinetic model of ErbB signaling shows that parameter estimation using adjoint sensitivity analysis requires a fraction of the computation time of established methods. The proposed method will facilitate mechanistic modeling of genome-scale cellular processes, as required in the age of omics. PMID:28114351

  19. Applications of species accumulation curves in large-scale biological data analysis.

    PubMed

    Deng, Chao; Daley, Timothy; Smith, Andrew D

    2015-09-01

    The species accumulation curve, or collector's curve, of a population gives the expected number of observed species or distinct classes as a function of sampling effort. Species accumulation curves allow researchers to assess and compare diversity across populations or to evaluate the benefits of additional sampling. Traditional applications have focused on ecological populations but emerging large-scale applications, for example in DNA sequencing, are orders of magnitude larger and present new challenges. We developed a method to estimate accumulation curves for predicting the complexity of DNA sequencing libraries. This method uses rational function approximations to a classical non-parametric empirical Bayes estimator due to Good and Toulmin [Biometrika, 1956, 43, 45-63]. Here we demonstrate how the same approach can be highly effective in other large-scale applications involving biological data sets. These include estimating microbial species richness, immune repertoire size, and k -mer diversity for genome assembly applications. We show how the method can be modified to address populations containing an effectively infinite number of species where saturation cannot practically be attained. We also introduce a flexible suite of tools implemented as an R package that make these methods broadly accessible.

  20. Applications of species accumulation curves in large-scale biological data analysis

    PubMed Central

    Deng, Chao; Daley, Timothy; Smith, Andrew D

    2016-01-01

    The species accumulation curve, or collector’s curve, of a population gives the expected number of observed species or distinct classes as a function of sampling effort. Species accumulation curves allow researchers to assess and compare diversity across populations or to evaluate the benefits of additional sampling. Traditional applications have focused on ecological populations but emerging large-scale applications, for example in DNA sequencing, are orders of magnitude larger and present new challenges. We developed a method to estimate accumulation curves for predicting the complexity of DNA sequencing libraries. This method uses rational function approximations to a classical non-parametric empirical Bayes estimator due to Good and Toulmin [Biometrika, 1956, 43, 45–63]. Here we demonstrate how the same approach can be highly effective in other large-scale applications involving biological data sets. These include estimating microbial species richness, immune repertoire size, and k-mer diversity for genome assembly applications. We show how the method can be modified to address populations containing an effectively infinite number of species where saturation cannot practically be attained. We also introduce a flexible suite of tools implemented as an R package that make these methods broadly accessible. PMID:27252899

  1. Geometric modeling of subcellular structures, organelles, and multiprotein complexes

    PubMed Central

    Feng, Xin; Xia, Kelin; Tong, Yiying; Wei, Guo-Wei

    2013-01-01

    SUMMARY Recently, the structure, function, stability, and dynamics of subcellular structures, organelles, and multi-protein complexes have emerged as a leading interest in structural biology. Geometric modeling not only provides visualizations of shapes for large biomolecular complexes but also fills the gap between structural information and theoretical modeling, and enables the understanding of function, stability, and dynamics. This paper introduces a suite of computational tools for volumetric data processing, information extraction, surface mesh rendering, geometric measurement, and curvature estimation of biomolecular complexes. Particular emphasis is given to the modeling of cryo-electron microscopy data. Lagrangian-triangle meshes are employed for the surface presentation. On the basis of this representation, algorithms are developed for surface area and surface-enclosed volume calculation, and curvature estimation. Methods for volumetric meshing have also been presented. Because the technological development in computer science and mathematics has led to multiple choices at each stage of the geometric modeling, we discuss the rationales in the design and selection of various algorithms. Analytical models are designed to test the computational accuracy and convergence of proposed algorithms. Finally, we select a set of six cryo-electron microscopy data representing typical subcellular complexes to demonstrate the efficacy of the proposed algorithms in handling biomolecular surfaces and explore their capability of geometric characterization of binding targets. This paper offers a comprehensive protocol for the geometric modeling of subcellular structures, organelles, and multiprotein complexes. PMID:23212797

  2. Metal adsorption onto bacterial surfaces: development of a predictive approach

    NASA Astrophysics Data System (ADS)

    Fein, Jeremy B.; Martin, Aaron M.; Wightman, Peter G.

    2001-12-01

    Aqueous metal cation adsorption onto bacterial surfaces can be successfully modeled by means of a surface complexation approach. However, relatively few stability constants for metal-bacterial surface complexes have been measured. In order to determine the bacterial adsorption behavior of cations that have not been studied in the laboratory, predictive techniques are required that enable estimation of the stability constants of bacterial surface complexes. In this study, we use a linear free-energy approach to compare previously measured stability constants for Bacillus subtilis metal-carboxyl surface complexes with aqueous metal-organic acid anion stability constants. The organic acids that we consider are acetic, oxalic, citric, and tiron. We add to this limited data set by conducting metal adsorption experiments onto Bacillus subtilis, determining bacterial surface stability constants for Co, Nd, Ni, Sr, and Zn. The adsorption behavior of each of the metals studied here was described well by considering metal-carboxyl bacterial surface complexation only, except for the Zn adsorption behavior, which required carboxyl and phosphoryl complexation to obtain a suitable fit to the data. The best correlation between bacterial carboxyl surface complexes and aqueous organic acid anion stability constants was obtained by means of metal-acetate aqueous complexes, with a linear correlation coefficient of 0.97. This correlation applies only to unhydrolyzed aqueous cations and only to carboxyl binding of those cations, and it does not predict the binding behavior under conditions where metal binding to other bacterial surface site types occurs. However, the relationship derived in this study permits estimation of the carboxyl site adsorption behavior of a wide range of aqueous metal cations for which there is an absence of experimental data. This technique, coupled with the observation of similar adsorption behaviors across bacterial species (Yee and Fein, 2001), enables estimation of the effects of bacterial adsorption on metal mobilities for a large number of environmental and geologic applications.

  3. Applying the 15 Public Health Emergency Preparedness Capabilities to Support Large-Scale Tuberculosis Investigations in Complex Congregate Settings

    PubMed Central

    Toren, Katelynne Gardner; Elsenboss, Carina; Narita, Masahiro

    2017-01-01

    Public Health—Seattle and King County, a metropolitan health department in western Washington, experiences rates of tuberculosis (TB) that are 1.6 times higher than are state and national averages. The department’s TB Control Program uses public health emergency management tools and capabilities sustained with Centers for Disease Control and Prevention grant funding to manage large-scale complex case investigations. We have described 3 contact investigations in large congregate settings that the TB Control Program conducted in 2015 and 2016. The program managed the investigations using public health emergency management tools, with support from the Preparedness Program. The 3 investigations encompassed medical evaluation of more than 1600 people, used more than 100 workers, identified nearly 30 individuals with latent TB infection, and prevented an estimated 3 cases of active disease. These incidents exemplify how investments in public health emergency preparedness can enhance health outcomes in traditional areas of public health. PMID:28892445

  4. Partitioning heritability by functional annotation using genome-wide association summary statistics.

    PubMed

    Finucane, Hilary K; Bulik-Sullivan, Brendan; Gusev, Alexander; Trynka, Gosia; Reshef, Yakir; Loh, Po-Ru; Anttila, Verneri; Xu, Han; Zang, Chongzhi; Farh, Kyle; Ripke, Stephan; Day, Felix R; Purcell, Shaun; Stahl, Eli; Lindstrom, Sara; Perry, John R B; Okada, Yukinori; Raychaudhuri, Soumya; Daly, Mark J; Patterson, Nick; Neale, Benjamin M; Price, Alkes L

    2015-11-01

    Recent work has demonstrated that some functional categories of the genome contribute disproportionately to the heritability of complex diseases. Here we analyze a broad set of functional elements, including cell type-specific elements, to estimate their polygenic contributions to heritability in genome-wide association studies (GWAS) of 17 complex diseases and traits with an average sample size of 73,599. To enable this analysis, we introduce a new method, stratified LD score regression, for partitioning heritability from GWAS summary statistics while accounting for linked markers. This new method is computationally tractable at very large sample sizes and leverages genome-wide information. Our findings include a large enrichment of heritability in conserved regions across many traits, a very large immunological disease-specific enrichment of heritability in FANTOM5 enhancers and many cell type-specific enrichments, including significant enrichment of central nervous system cell types in the heritability of body mass index, age at menarche, educational attainment and smoking behavior.

  5. In-memory integration of existing software components for parallel adaptive unstructured mesh workflows

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Smith, Cameron W.; Granzow, Brian; Diamond, Gerrett

    Unstructured mesh methods, like finite elements and finite volumes, support the effective analysis of complex physical behaviors modeled by partial differential equations over general threedimensional domains. The most reliable and efficient methods apply adaptive procedures with a-posteriori error estimators that indicate where and how the mesh is to be modified. Although adaptive meshes can have two to three orders of magnitude fewer elements than a more uniform mesh for the same level of accuracy, there are many complex simulations where the meshes required are so large that they can only be solved on massively parallel systems.

  6. In-memory integration of existing software components for parallel adaptive unstructured mesh workflows

    DOE PAGES

    Smith, Cameron W.; Granzow, Brian; Diamond, Gerrett; ...

    2017-01-01

    Unstructured mesh methods, like finite elements and finite volumes, support the effective analysis of complex physical behaviors modeled by partial differential equations over general threedimensional domains. The most reliable and efficient methods apply adaptive procedures with a-posteriori error estimators that indicate where and how the mesh is to be modified. Although adaptive meshes can have two to three orders of magnitude fewer elements than a more uniform mesh for the same level of accuracy, there are many complex simulations where the meshes required are so large that they can only be solved on massively parallel systems.

  7. Incorporation of Solar-Induced Chlorophyll Fluorescence into the Breathing Earth System Simulator (BESS)

    NASA Astrophysics Data System (ADS)

    Dechant, B.; Ryu, Y.; Jiang, C.; Yang, K.

    2017-12-01

    Solar-induced chlorophyll fluorescence (SIF) is rapidly becoming an important tool to remotely estimate terrestrial gross primary productivity (GPP) at large spatial scales. Many findings, however, are based on empirical relationships between SIF and GPP that have been found to be dependent on plant functional types. Therefore, combining model-based analysis with observations is crucial to improve our understanding of SIF-GPP relationships. So far, most model-based results were based on SCOPE, a complex ecophysiological model with explicit description of canopy layers and a large number of parameters that may not be easily obtained reliably on large scales. Here, we report on our efforts to incorporate SIF into a two-big leaf (sun and shade) process-based model that is suitable for obtaining its inputs entirely from satellite products. We examine if the SIF-GPP relationships are consistent with the findings from SCOPE simulations and investigate if incorporation of the SIF signal into BESS can help improve GPP estimation. A case study in a rice paddy is presented.

  8. Does reef architectural complexity influence resource availability for a large reef-dwelling invertebrate?

    NASA Astrophysics Data System (ADS)

    Lozano-Álvarez, Enrique; Luviano-Aparicio, Nelia; Negrete-Soto, Fernando; Barradas-Ortiz, Cecilia; Aguíñiga-García, Sergio; Morillo-Velarde, Piedad S.; Álvarez-Filip, Lorenzo; Briones-Fourzán, Patricia

    2017-10-01

    In coral reefs, loss of architectural complexity and its associated habitat degradation is expected to affect reef specialists in particular due to changes in resource availability. We explored whether these features could potentially affect populations of a large invertebrate, the spotted spiny lobster Panulirus guttatus, which is an obligate Caribbean coral reef-dweller with a limited home range. We selected two separate large coral reef patches in Puerto Morelos (Mexico) that differed significantly in structural complexity and level of degradation, as assessed via the rugosity index, habitat assessment score, and percent cover of various benthic components. On each reef, we estimated density of P. guttatus and sampled lobsters to analyze their stomach contents, three different condition indices, and stable isotopes (δ15N and δ13C) in muscle. Lobster density did not vary with reef, suggesting that available crevices in the less complex patch still provided adequate refuge to these lobsters. Lobsters consumed many food types, dominated by mollusks and crustaceans, but proportionally more crustaceans (herbivore crabs) in the less complex patch, which had more calcareous macroalgae and algal turf. Lobsters from both reefs had a similar condition (all three indices) and mean δ15N, suggesting a similar quality of diet between reefs related to their opportunistic feeding, but differed in mean δ13C values, reflecting the different carbon sources between reefs and providing indirect evidence of individuals of P. guttatus foraging exclusively over their home reef. Overall, we found no apparent effects of architectural complexity, at least to the degree observed in our less complex patch, on density, condition, or trophic level of P. guttatus.

  9. Confidence range estimate of extended source imagery acquisition algorithms via computer simulations. [in optical communication systems

    NASA Technical Reports Server (NTRS)

    Chen, CHIEN-C.; Hui, Elliot; Okamoto, Garret

    1992-01-01

    Spatial acquisition using the sun-lit Earth as a beacon source provides several advantages over active beacon-based systems for deep-space optical communication systems. However, since the angular extend of the Earth image is large compared to the laser beam divergence, the acquisition subsystem must be capable of resolving the image to derive the proper pointing orientation. The algorithms used must be capable of deducing the receiver location given the blurring introduced by the imaging optics and the large Earth albedo fluctuation. Furthermore, because of the complexity of modelling the Earth and the tracking algorithms, an accurate estimate of the algorithm accuracy can only be made via simulation using realistic Earth images. An image simulator was constructed for this purpose, and the results of the simulation runs are reported.

  10. Equilibrium expert: an add-in to Microsoft Excel for multiple binding equilibrium simulations and parameter estimations.

    PubMed

    Raguin, Olivier; Gruaz-Guyon, Anne; Barbet, Jacques

    2002-11-01

    An add-in to Microsoft Excel was developed to simulate multiple binding equilibriums. A partition function, readily written even when the equilibrium is complex, describes the experimental system. It involves the concentrations of the different free molecular species and of the different complexes present in the experiment. As a result, the software is not restricted to a series of predefined experimental setups but can handle a large variety of problems involving up to nine independent molecular species. Binding parameters are estimated by nonlinear least-square fitting of experimental measurements as supplied by the user. The fitting process allows user-defined weighting of the experimental data. The flexibility of the software and the way it may be used to describe common experimental situations and to deal with usual problems such as tracer reactivity or nonspecific binding is demonstrated by a few examples. The software is available free of charge upon request.

  11. Visual analysis of geocoded twin data puts nature and nurture on the map.

    PubMed

    Davis, O S P; Haworth, C M A; Lewis, C M; Plomin, R

    2012-09-01

    Twin studies allow us to estimate the relative contributions of nature and nurture to human phenotypes by comparing the resemblance of identical and fraternal twins. Variation in complex traits is a balance of genetic and environmental influences; these influences are typically estimated at a population level. However, what if the balance of nature and nurture varies depending on where we grow up? Here we use statistical and visual analysis of geocoded data from over 6700 families to show that genetic and environmental contributions to 45 childhood cognitive and behavioral phenotypes vary geographically in the United Kingdom. This has implications for detecting environmental exposures that may interact with the genetic influences on complex traits, and for the statistical power of samples recruited for genetic association studies. More broadly, our experience demonstrates the potential for collaborative exploratory visualization to act as a lingua franca for large-scale interdisciplinary research.

  12. Practical characterization of quantum devices without tomography

    NASA Astrophysics Data System (ADS)

    Landon-Cardinal, Olivier; Flammia, Steven; Silva, Marcus; Liu, Yi-Kai; Poulin, David

    2012-02-01

    Quantum tomography is the main method used to assess the quality of quantum information processing devices, but its complexity presents a major obstacle for the characterization of even moderately large systems. Part of the reason for this complexity is that tomography generates much more information than is usually sought. Taking a more targeted approach, we develop schemes that enable (i) estimating the ?delity of an experiment to a theoretical ideal description, (ii) learning which description within a reduced subset best matches the experimental data. Both these approaches yield a signi?cant reduction in resources compared to tomography. In particular, we show how to estimate the ?delity between a predicted pure state and an arbitrary experimental state using only a constant number of Pauli expectation values selected at random according to an importance-weighting rule. In addition, we propose methods for certifying quantum circuits and learning continuous-time quantum dynamics that are described by local Hamiltonians or Lindbladians.

  13. Comprehensive inventory of protein complexes in the Protein Data Bank from consistent classification of interfaces.

    PubMed

    Bordner, Andrew J; Gorin, Andrey A

    2008-05-12

    Protein-protein interactions are ubiquitous and essential for all cellular processes. High-resolution X-ray crystallographic structures of protein complexes can reveal the details of their function and provide a basis for many computational and experimental approaches. Differentiation between biological and non-biological contacts and reconstruction of the intact complex is a challenging computational problem. A successful solution can provide additional insights into the fundamental principles of biological recognition and reduce errors in many algorithms and databases utilizing interaction information extracted from the Protein Data Bank (PDB). We have developed a method for identifying protein complexes in the PDB X-ray structures by a four step procedure: (1) comprehensively collecting all protein-protein interfaces; (2) clustering similar protein-protein interfaces together; (3) estimating the probability that each cluster is relevant based on a diverse set of properties; and (4) combining these scores for each PDB entry in order to predict the complex structure. The resulting clusters of biologically relevant interfaces provide a reliable catalog of evolutionary conserved protein-protein interactions. These interfaces, as well as the predicted protein complexes, are available from the Protein Interface Server (PInS) website (see Availability and requirements section). Our method demonstrates an almost two-fold reduction of the annotation error rate as evaluated on a large benchmark set of complexes validated from the literature. We also estimate relative contributions of each interface property to the accurate discrimination of biologically relevant interfaces and discuss possible directions for further improving the prediction method.

  14. Evaluating habitat for black-footed ferrets: Revision of an existing model

    USGS Publications Warehouse

    Biggins, Dean E.; Lockhart, J. Michael; Godbey, Jerry L.

    2006-01-01

    Black-footed ferrets (Mustela nigripes) are highly dependent on prairie dogs (Cynomys spp.) as prey, and prairie dog colonies are the only known habitats that sustain black-footed ferret populations. An existing model used extensively for evaluating black-footed ferret reintroduction habitat defined complexes by interconnecting colonies with 7-km line segments. Although the 7-km complex remains a useful construct, we propose additional, smaller-scale evaluations that consider 1.5-km subcomplexes. The original model estimated the carrying capacity of complexes based on energy requirements of ferrets and density estimates of their prairie dog prey. Recent data have supported earlier contentions of intraspecific competition and intrasexual territorial behavior in ferrets. We suggest a revised model that retains the fixed linear relationship of the existing model when prairie dog densities are <18/ha and uses a curvilinear relationship that reflects increasing effects of ferret territoriality when there are 18–42 prairie dogs per hectare. We discuss possible effects of colony size and shape, interacting with territoriality, as justification for the exclusion of territorial influences if a prairie dog colony supports only a single female ferret. We also present data to support continued use of active prairie dog burrow densities as indices suitable for broad-scale estimates of prairie dog density. Calculation of percent of complexes that are occupied by prairie dog colonies was recommended as part of the original habitat evaluation process. That attribute has been largely ignored, resulting in rating anomalies.

  15. A genetic-algorithm approach for assessing the liquefaction potential of sandy soils

    NASA Astrophysics Data System (ADS)

    Sen, G.; Akyol, E.

    2010-04-01

    The determination of liquefaction potential is required to take into account a large number of parameters, which creates a complex nonlinear structure of the liquefaction phenomenon. The conventional methods rely on simple statistical and empirical relations or charts. However, they cannot characterise these complexities. Genetic algorithms are suited to solve these types of problems. A genetic algorithm-based model has been developed to determine the liquefaction potential by confirming Cone Penetration Test datasets derived from case studies of sandy soils. Software has been developed that uses genetic algorithms for the parameter selection and assessment of liquefaction potential. Then several estimation functions for the assessment of a Liquefaction Index have been generated from the dataset. The generated Liquefaction Index estimation functions were evaluated by assessing the training and test data. The suggested formulation estimates the liquefaction occurrence with significant accuracy. Besides, the parametric study on the liquefaction index curves shows a good relation with the physical behaviour. The total number of misestimated cases was only 7.8% for the proposed method, which is quite low when compared to another commonly used method.

  16. Gray: a ray tracing-based Monte Carlo simulator for PET

    NASA Astrophysics Data System (ADS)

    Freese, David L.; Olcott, Peter D.; Buss, Samuel R.; Levin, Craig S.

    2018-05-01

    Monte Carlo simulation software plays a critical role in PET system design. Performing complex, repeated Monte Carlo simulations can be computationally prohibitive, as even a single simulation can require a large amount of time and a computing cluster to complete. Here we introduce Gray, a Monte Carlo simulation software for PET systems. Gray exploits ray tracing methods used in the computer graphics community to greatly accelerate simulations of PET systems with complex geometries. We demonstrate the implementation of models for positron range, annihilation acolinearity, photoelectric absorption, Compton scatter, and Rayleigh scatter. For validation, we simulate the GATE PET benchmark, and compare energy, distribution of hits, coincidences, and run time. We show a speedup using Gray, compared to GATE for the same simulation, while demonstrating nearly identical results. We additionally simulate the Siemens Biograph mCT system with both the NEMA NU-2 scatter phantom and sensitivity phantom. We estimate the total sensitivity within % when accounting for differences in peak NECR. We also estimate the peak NECR to be kcps, or within % of published experimental data. The activity concentration of the peak is also estimated within 1.3%.

  17. Re-assessment of the mass balance of the Abbot and Getz sectors of West Antarctica

    NASA Astrophysics Data System (ADS)

    Chuter, S.; Bamber, J. L.

    2016-12-01

    Large discrepancies exist in mass balance estimates for the Getz and Abbot drainage basins, primarily due to previous poor knowledge of ice thickness at the grounding line, poor coverage by previous altimetry missions and signal leakage issues for GRACE. Large errors arise when using ice thickness measurements derived from ERS-1 and/or ICESat altimetry data due to poor track spacing, `loss of lock' issues near the grounding line and the complex morphology of these shelves, requiring fine resolution to derive robust and accurate elevations close to the grounding line. However, the advent of CryoSat-2 with its unique orbit and SARIn mode of operation has overcome these issues and enabled the determination of ice shelf thickness at a much higher accuracy than possible from previous satellites, particularly within the grounding zone. Here we present a contemporary estimate of ice sheet mass balance for the both the Getz and Abbot drainage basins. This is achieved through the use of contemporary velocity data derived from Landsat feature tracking and the use of CryoSat-2 derived ice thickness measurements. Additionally, we use this new ice thickness dataset to reassess mass balance estimates from 2008/2009, where there were large disparities between results from radar altimetry and Input-Output methodologies over the Abbot region in particular. These contemporary results are compared with other present day estimates from gravimetry and altimetry elevation changes.

  18. The ribosomes of Drosophila. II. Studies on intraspecific variation.

    PubMed

    Berger, E M; Weber, L

    1974-12-01

    Electrophoretic comparisons of 40S and 55S ribosomal subunit proteins from 18 strains of Drosophila melanogaster revealed the virtual absence of allelic variation. More detailed two-dimensional studies on the large subunit proteins in 6 of the strains demonstrated additional complexity but still no interstrain variation. The significance of these results is discussed with respect to present estimates of genic heterozygosity in natural populations.

  19. Testing for independence in J×K contingency tables with complex sample survey data.

    PubMed

    Lipsitz, Stuart R; Fitzmaurice, Garrett M; Sinha, Debajyoti; Hevelone, Nathanael; Giovannucci, Edward; Hu, Jim C

    2015-09-01

    The test of independence of row and column variables in a (J×K) contingency table is a widely used statistical test in many areas of application. For complex survey samples, use of the standard Pearson chi-squared test is inappropriate due to correlation among units within the same cluster. Rao and Scott (1981, Journal of the American Statistical Association 76, 221-230) proposed an approach in which the standard Pearson chi-squared statistic is multiplied by a design effect to adjust for the complex survey design. Unfortunately, this test fails to exist when one of the observed cell counts equals zero. Even with the large samples typical of many complex surveys, zero cell counts can occur for rare events, small domains, or contingency tables with a large number of cells. Here, we propose Wald and score test statistics for independence based on weighted least squares estimating equations. In contrast to the Rao-Scott test statistic, the proposed Wald and score test statistics always exist. In simulations, the score test is found to perform best with respect to type I error. The proposed method is motivated by, and applied to, post surgical complications data from the United States' Nationwide Inpatient Sample (NIS) complex survey of hospitals in 2008. © 2015, The International Biometric Society.

  20. Geometric k-nearest neighbor estimation of entropy and mutual information

    NASA Astrophysics Data System (ADS)

    Lord, Warren M.; Sun, Jie; Bollt, Erik M.

    2018-03-01

    Nonparametric estimation of mutual information is used in a wide range of scientific problems to quantify dependence between variables. The k-nearest neighbor (knn) methods are consistent, and therefore expected to work well for a large sample size. These methods use geometrically regular local volume elements. This practice allows maximum localization of the volume elements, but can also induce a bias due to a poor description of the local geometry of the underlying probability measure. We introduce a new class of knn estimators that we call geometric knn estimators (g-knn), which use more complex local volume elements to better model the local geometry of the probability measures. As an example of this class of estimators, we develop a g-knn estimator of entropy and mutual information based on elliptical volume elements, capturing the local stretching and compression common to a wide range of dynamical system attractors. A series of numerical examples in which the thickness of the underlying distribution and the sample sizes are varied suggest that local geometry is a source of problems for knn methods such as the Kraskov-Stögbauer-Grassberger estimator when local geometric effects cannot be removed by global preprocessing of the data. The g-knn method performs well despite the manipulation of the local geometry. In addition, the examples suggest that the g-knn estimators can be of particular relevance to applications in which the system is large, but the data size is limited.

  1. On estimating the phase of periodic waveform in additive Gaussian noise, part 2

    NASA Astrophysics Data System (ADS)

    Rauch, L. L.

    1984-11-01

    Motivated by advances in signal processing technology that support more complex algorithms, a new look is taken at the problem of estimating the phase and other parameters of a periodic waveform in additive Gaussian noise. The general problem was introduced and the maximum a posteriori probability criterion with signal space interpretation was used to obtain the structures of optimum and some suboptimum phase estimators for known constant frequency and unknown constant phase with an a priori distribution. Optimal algorithms are obtained for some cases where the frequency is a parameterized function of time with the unknown parameters and phase having a joint a priori distribution. In the last section, the intrinsic and extrinsic geometry of hypersurfaces is introduced to provide insight to the estimation problem for the small noise and large noise cases.

  2. On Estimating the Phase of Periodic Waveform in Additive Gaussian Noise, Part 2

    NASA Technical Reports Server (NTRS)

    Rauch, L. L.

    1984-01-01

    Motivated by advances in signal processing technology that support more complex algorithms, a new look is taken at the problem of estimating the phase and other parameters of a periodic waveform in additive Gaussian noise. The general problem was introduced and the maximum a posteriori probability criterion with signal space interpretation was used to obtain the structures of optimum and some suboptimum phase estimators for known constant frequency and unknown constant phase with an a priori distribution. Optimal algorithms are obtained for some cases where the frequency is a parameterized function of time with the unknown parameters and phase having a joint a priori distribution. In the last section, the intrinsic and extrinsic geometry of hypersurfaces is introduced to provide insight to the estimation problem for the small noise and large noise cases.

  3. Platform options for the Space Station program

    NASA Technical Reports Server (NTRS)

    Mangano, M. J.; Rowley, R. W.

    1986-01-01

    Platforms for polar and 28.5 deg orbits were studied to determine the platform requirements and characteristics necessary to support the science objectives. Large platforms supporting the Earth-Observing System (EOS) were initially studied. Co-orbiting platforms were derived from these designs. Because cost estimates indicated that the large platform approach was likely to be too expensive, require several launches, and generally be excessively complex, studies of small platforms were undertaken. Results of these studies show the small platform approach to be technically feasible at lower overall cost. All designs maximized hardware inheritance from the Space Station program to reduce costs. Science objectives as defined at the time of these studies are largely achievable.

  4. Hapl-o-Mat: open-source software for HLA haplotype frequency estimation from ambiguous and heterogeneous data.

    PubMed

    Schäfer, Christian; Schmidt, Alexander H; Sauter, Jürgen

    2017-05-30

    Knowledge of HLA haplotypes is helpful in many settings as disease association studies, population genetics, or hematopoietic stem cell transplantation. Regarding the recruitment of unrelated hematopoietic stem cell donors, HLA haplotype frequencies of specific populations are used to optimize both donor searches for individual patients and strategic donor registry planning. However, the estimation of haplotype frequencies from HLA genotyping data is challenged by the large amount of genotype data, the complex HLA nomenclature, and the heterogeneous and ambiguous nature of typing records. To meet these challenges, we have developed the open-source software Hapl-o-Mat. It estimates haplotype frequencies from population data including an arbitrary number of loci using an expectation-maximization algorithm. Its key features are the processing of different HLA typing resolutions within a given population sample and the handling of ambiguities recorded via multiple allele codes or genotype list strings. Implemented in C++, Hapl-o-Mat facilitates efficient haplotype frequency estimation from large amounts of genotype data. We demonstrate its accuracy and performance on the basis of artificial and real genotype data. Hapl-o-Mat is a versatile and efficient software for HLA haplotype frequency estimation. Its capability of processing various forms of HLA genotype data allows for a straightforward haplotype frequency estimation from typing records usually found in stem cell donor registries.

  5. Fast Component Pursuit for Large-Scale Inverse Covariance Estimation.

    PubMed

    Han, Lei; Zhang, Yu; Zhang, Tong

    2016-08-01

    The maximum likelihood estimation (MLE) for the Gaussian graphical model, which is also known as the inverse covariance estimation problem, has gained increasing interest recently. Most existing works assume that inverse covariance estimators contain sparse structure and then construct models with the ℓ 1 regularization. In this paper, different from existing works, we study the inverse covariance estimation problem from another perspective by efficiently modeling the low-rank structure in the inverse covariance, which is assumed to be a combination of a low-rank part and a diagonal matrix. One motivation for this assumption is that the low-rank structure is common in many applications including the climate and financial analysis, and another one is that such assumption can reduce the computational complexity when computing its inverse. Specifically, we propose an efficient COmponent Pursuit (COP) method to obtain the low-rank part, where each component can be sparse. For optimization, the COP method greedily learns a rank-one component in each iteration by maximizing the log-likelihood. Moreover, the COP algorithm enjoys several appealing properties including the existence of an efficient solution in each iteration and the theoretical guarantee on the convergence of this greedy approach. Experiments on large-scale synthetic and real-world datasets including thousands of millions variables show that the COP method is faster than the state-of-the-art techniques for the inverse covariance estimation problem when achieving comparable log-likelihood on test data.

  6. Cross-slope Movement Patterns in Landslides

    NASA Astrophysics Data System (ADS)

    Petley, D.; Murphy, W.; Bulmer, M. H.; Keefer, D.

    2002-12-01

    There is growing evidence that there is a significant element of cross-slope movement in many large landslide systems. These movements may result in changing states of stress between landslide blocks that can establish complex displacement patterns. Such motions, which are not considered in traditional two-dimensional limit-equilibrium analyses, are important in the investigation of a variety of landslide types, such as those triggered by earthquakes. In addition, these movements may introduce considerable errors into the interpretation of strain patterns as derived from InSAR studies. Finally, even traditional interpretation techniques may lead to the amount of total displacement being underestimated. These observations suggest that a three dimensional form of analysis may be more appropriate for large landslide complexes. The significance of such cross-slope movements are being investigated using a detailed investigation of the Lishan landslide complex in Central Taiwan. This landslide system, which was reactivated in 1990 related to the construction of a hotel. The total recorded movements have been approximately 1.5 m over an area of sliding that is estimated to be 450 m wide and 200 m long. Extensive damage has been caused to roads and buildings within the town. Remediation work has resulted largely in the stabilization of the landslide complex. Detailed geomorphological mapping has revealed that the landslide complex is composed of two main components. The first, immediately upslope of the hotel construction site, is a relatively shallow earthflow. The second, which has formed a large headscarp upslope from the main road in the centre of the town, is a deeper translational slide. Both appear to have been reactivations of previous failures. While the displacement patterns of the earthflow indicate a relatively simple downslope movement, the vectors derived from kinematic analysis of surface features have indicated that the movement of the deeper-seated landslide was more complex. Though the dominant movement vector is downslope, there is evidence to suggest that there has been a cross-slope component of motion that corresponds to the bedding orientation.

  7. Hardware design and implementation of fast DOA estimation method based on multicore DSP

    NASA Astrophysics Data System (ADS)

    Guo, Rui; Zhao, Yingxiao; Zhang, Yue; Lin, Qianqiang; Chen, Zengping

    2016-10-01

    In this paper, we present a high-speed real-time signal processing hardware platform based on multicore digital signal processor (DSP). The real-time signal processing platform shows several excellent characteristics including high performance computing, low power consumption, large-capacity data storage and high speed data transmission, which make it able to meet the constraint of real-time direction of arrival (DOA) estimation. To reduce the high computational complexity of DOA estimation algorithm, a novel real-valued MUSIC estimator is used. The algorithm is decomposed into several independent steps and the time consumption of each step is counted. Based on the statistics of the time consumption, we present a new parallel processing strategy to distribute the task of DOA estimation to different cores of the real-time signal processing hardware platform. Experimental results demonstrate that the high processing capability of the signal processing platform meets the constraint of real-time direction of arrival (DOA) estimation.

  8. Improved biovolume estimation of Microcystis aeruginosa colonies: A statistical approach.

    PubMed

    Alcántara, I; Piccini, C; Segura, A M; Deus, S; González, C; Martínez de la Escalera, G; Kruk, C

    2018-05-27

    The Microcystis aeruginosa complex (MAC) clusters many of the most common freshwater and brackish bloom-forming cyanobacteria. In monitoring protocols, biovolume estimation is a common approach to determine MAC colonies biomass and useful for prediction purposes. Biovolume (μm 3 mL -1 ) is calculated multiplying organism abundance (orgL -1 ) by colonial volume (μm 3 org -1 ). Colonial volume is estimated based on geometric shapes and requires accurate measurements of dimensions using optical microscopy. A trade-off between easy-to-measure but low-accuracy simple shapes (e.g. sphere) and time costly but high-accuracy complex shapes (e.g. ellipsoid) volume estimation is posed. Overestimations effects in ecological studies and management decisions associated to harmful blooms are significant due to the large sizes of MAC colonies. In this work, we aimed to increase the precision of MAC biovolume estimations by developing a statistical model based on two easy-to-measure dimensions. We analyzed field data from a wide environmental gradient (800 km) spanning freshwater to estuarine and seawater. We measured length, width and depth from ca. 5700 colonies under an inverted microscope and estimated colonial volume using three different recommended geometrical shapes (sphere, prolate spheroid and ellipsoid). Because of the non-spherical shape of MAC the ellipsoid resulted in the most accurate approximation, whereas the sphere overestimated colonial volume (3-80) especially for large colonies (MLD higher than 300 μm). Ellipsoid requires measuring three dimensions and is time-consuming. Therefore, we constructed different statistical models to predict organisms depth based on length and width. Splitting the data into training (2/3) and test (1/3) sets, all models resulted in low training (1.41-1.44%) and testing average error (1.3-2.0%). The models were also evaluated using three other independent datasets. The multiple linear model was finally selected to calculate MAC volume as an ellipsoid based on length and width. This work contributes to achieve a better estimation of MAC volume applicable to monitoring programs as well as to ecological research. Copyright © 2017. Published by Elsevier B.V.

  9. A heuristic simulation model of Lake Ontario circulation and mass balance transport

    USGS Publications Warehouse

    McKenna, J.E.; Chalupnicki, M.A.

    2011-01-01

    The redistribution of suspended organisms and materials by large-scale currents is part of natural ecological processes in large aquatic systems but can contribute to ecosystem disruption when exotic elements are introduced into the system. Toxic compounds and planktonic organisms spend various lengths of time in suspension before settling to the bottom or otherwise being removed. We constructed a simple physical simulation model, including the influence of major tributaries, to qualitatively examine circulation patterns in Lake Ontario. We used a simple mass balance approach to estimate the relative water input to and export from each of 10 depth regime-specific compartments (nearshore vs. offshore) comprising Lake Ontario. Despite its simplicity, our model produced circulation patterns similar to those reported by more complex studies in the literature. A three-gyre pattern, with the classic large counterclockwise central lake circulation, and a simpler two-gyre system were both observed. These qualitative simulations indicate little offshore transport along the south shore, except near the mouths of the Niagara River and Oswego River. Complex flow structure was evident, particularly near the Niagara River mouth and in offshore waters of the eastern basin. Average Lake Ontario residence time is 8 years, but the fastest model pathway indicated potential transport of plankton through the lake in as little as 60 days. This simulation illustrates potential invasion pathways and provides rough estimates of planktonic larval dispersal or chemical transport among nearshore and offshore areas of Lake Ontario. ?? 2011 Taylor & Francis.

  10. Voltage collapse in complex power grids

    PubMed Central

    Simpson-Porco, John W.; Dörfler, Florian; Bullo, Francesco

    2016-01-01

    A large-scale power grid's ability to transfer energy from producers to consumers is constrained by both the network structure and the nonlinear physics of power flow. Violations of these constraints have been observed to result in voltage collapse blackouts, where nodal voltages slowly decline before precipitously falling. However, methods to test for voltage collapse are dominantly simulation-based, offering little theoretical insight into how grid structure influences stability margins. For a simplified power flow model, here we derive a closed-form condition under which a power network is safe from voltage collapse. The condition combines the complex structure of the network with the reactive power demands of loads to produce a node-by-node measure of grid stress, a prediction of the largest nodal voltage deviation, and an estimate of the distance to collapse. We extensively test our predictions on large-scale systems, highlighting how our condition can be leveraged to increase grid stability margins. PMID:26887284

  11. Monte Carlo estimation of total variation distance of Markov chains on large spaces, with application to phylogenetics.

    PubMed

    Herbei, Radu; Kubatko, Laura

    2013-03-26

    Markov chains are widely used for modeling in many areas of molecular biology and genetics. As the complexity of such models advances, it becomes increasingly important to assess the rate at which a Markov chain converges to its stationary distribution in order to carry out accurate inference. A common measure of convergence to the stationary distribution is the total variation distance, but this measure can be difficult to compute when the state space of the chain is large. We propose a Monte Carlo method to estimate the total variation distance that can be applied in this situation, and we demonstrate how the method can be efficiently implemented by taking advantage of GPU computing techniques. We apply the method to two Markov chains on the space of phylogenetic trees, and discuss the implications of our findings for the development of algorithms for phylogenetic inference.

  12. Leveraging population admixture to explain missing heritability of complex traits

    PubMed Central

    Zaitlen, Noah; Pasaniuc, Bogdan; Sankararaman, Sriram; Bhatia, Gaurav; Zhang, Jianqi; Gusev, Alexander; Young, Taylor; Tandon, Arti; Pollack, Samuela; Vilhjálmsson, Bjarni J.; Assimes, Themistocles L.; Berndt, Sonja I.; Blot, William J.; Chanock, Stephen; Franceschini, Nora; Goodman, Phyllis G.; He, Jing; Hennis, Anselm JM; Hsing, Ann; Ingles, Sue A.; Isaacs, William; Kittles, Rick A.; Klein, Eric A.; Lange, Leslie A.; Nemesure, Barbara; Patterson, Nick; Reich, David; Rybicki, Benjamin A.; Stanford, Janet L.; Stevens, Victoria L; Strom, Sara S.; Whitsel, Eric A; Witte, John S.; Xu, Jianfeng; Haiman, Christopher; Wilson, James G.; Kooperberg, Charles; Stram, Daniel; Reiner, Alex P.; Tang, Hua; Price, Alkes L.

    2014-01-01

    Despite recent progress on estimating the heritability explained by genotyped SNPs (hg2), a large gap between hg2 and estimates of total narrow-sense heritability (h2) remains. Explanations for this gap include rare variants, or upward bias in family-based estimates of h2 due to shared environment or epistasis. We estimate h2 from unrelated individuals in admixed populations by first estimating the heritability explained by local ancestry (hγ2). We show that hγ2 = 2FSTCθ(1−θ)h2, where FSTC measures frequency differences between populations at causal loci and θ is the genome-wide ancestry proportion. Our approach is not susceptible to biases caused by epistasis or shared environment. We examined 21,497 African Americans from three cohorts, analyzing 13 phenotypes. For height and BMI, we obtained h2 estimates of 0.55 ± 0.09 and 0.23 ± 0.06, respectively, which are larger than estimates of hg2 in these and other data, but smaller than family-based estimates of h2. PMID:25383972

  13. Leveraging population admixture to characterize the heritability of complex traits.

    PubMed

    Zaitlen, Noah; Pasaniuc, Bogdan; Sankararaman, Sriram; Bhatia, Gaurav; Zhang, Jianqi; Gusev, Alexander; Young, Taylor; Tandon, Arti; Pollack, Samuela; Vilhjálmsson, Bjarni J; Assimes, Themistocles L; Berndt, Sonja I; Blot, William J; Chanock, Stephen; Franceschini, Nora; Goodman, Phyllis G; He, Jing; Hennis, Anselm J M; Hsing, Ann; Ingles, Sue A; Isaacs, William; Kittles, Rick A; Klein, Eric A; Lange, Leslie A; Nemesure, Barbara; Patterson, Nick; Reich, David; Rybicki, Benjamin A; Stanford, Janet L; Stevens, Victoria L; Strom, Sara S; Whitsel, Eric A; Witte, John S; Xu, Jianfeng; Haiman, Christopher; Wilson, James G; Kooperberg, Charles; Stram, Daniel; Reiner, Alex P; Tang, Hua; Price, Alkes L

    2014-12-01

    Despite recent progress on estimating the heritability explained by genotyped SNPs (h(2)g), a large gap between h(2)g and estimates of total narrow-sense heritability (h(2)) remains. Explanations for this gap include rare variants or upward bias in family-based estimates of h(2) due to shared environment or epistasis. We estimate h(2) from unrelated individuals in admixed populations by first estimating the heritability explained by local ancestry (h(2)γ). We show that h(2)γ = 2FSTCθ(1 - θ)h(2), where FSTC measures frequency differences between populations at causal loci and θ is the genome-wide ancestry proportion. Our approach is not susceptible to biases caused by epistasis or shared environment. We applied this approach to the analysis of 13 phenotypes in 21,497 African-American individuals from 3 cohorts. For height and body mass index (BMI), we obtained h(2) estimates of 0.55 ± 0.09 and 0.23 ± 0.06, respectively, which are larger than estimates of h(2)g in these and other data but smaller than family-based estimates of h(2).

  14. Mechanosensitive channels in bacteria as membrane tension reporters

    NASA Technical Reports Server (NTRS)

    Sukharev, S.

    1999-01-01

    The purpose of this short review is to discuss recent data on the molecular structure and mechanism of gating of MscL, a mechanosensitive channel of large conductance from Escherichia coli. MscL is the first isolated molecule shown to convert mechanical stress of the membrane into a simple response, the opening of a large aqueous pore. The functional complex appears to be a stable homo-pentamer of 15-kDa subunits, the gating transitions in which are driven by stretch forces conveyed through the lipid bilayer. We have measured the open probability of MscL and the kinetics of transitions as a function of membrane tension. The parameters extracted from the single-channel current recordings and dose-response curves such as the energy difference between the closed, open, and intermediate conducting states, and the transition-related changes in protein dimensions suggest a large conformational rearrangement of the channel complex. The estimations show that in native conditions MscL openings could be driven primarily by forces of osmotic nature. The thermodynamic and spatial parameters reasonably correlate with the available data on the structure of a single MscL subunit and multimeric organization of the complex. Combined with the functional analysis of mutations, these data give grounds to hypotheses on the nature of the channel mechanosensitivity.

  15. Analysis of Hydrogen Atom Abstraction from Ethylbenzene by an FeVO(TAML) Complex.

    PubMed

    Shen, Longzhu Q; Kundu, Soumen; Collins, Terrence J; Bominaar, Emile L

    2017-04-17

    It was shown previously (Chem. Eur. J. 2015, 21, 1803) that the rate of hydrogen atom abstraction, k, from ethylbenzene (EB) by TAML complex [Fe V (O)B*] - (1) in acetonitrile exhibits a large kinetic isotope effect (KIE ∼ 26) in the experimental range 233-243 K. The extrapolated tangents of ln(k/T) vs T -1 plots for EB-d 10 and EB gave a large, negative intercept difference, Int(EB) - Int(EB-d 10 ) = -34.5 J mol -1 K -1 for T -1 → 0, which is shown to be exclusively due to an isotopic mass effect on tunneling. A decomposition of the apparent activation barrier in terms of electronic, ZPE, thermal enthalpic, tunneling, and entropic contributions is presented. Tunneling corrections to ΔH ⧧ and ΔS ⧧ are estimated to be large. The DFT prediction, using functional B3LYP and basis set 6-311G, for the electronic contribution is significantly smaller than suggested by experiment. However, the agreement improves after correction for the basis set superposition error in the interaction between EB and 1. The kinetic model employed has been used to predict rate constants outside the experimental temperature range, which enabled us to compare the reactivity of 1 with those of other hydrogen abstracting complexes.

  16. ARK: Aggregation of Reads by K-Means for Estimation of Bacterial Community Composition.

    PubMed

    Koslicki, David; Chatterjee, Saikat; Shahrivar, Damon; Walker, Alan W; Francis, Suzanna C; Fraser, Louise J; Vehkaperä, Mikko; Lan, Yueheng; Corander, Jukka

    2015-01-01

    Estimation of bacterial community composition from high-throughput sequenced 16S rRNA gene amplicons is a key task in microbial ecology. Since the sequence data from each sample typically consist of a large number of reads and are adversely impacted by different levels of biological and technical noise, accurate analysis of such large datasets is challenging. There has been a recent surge of interest in using compressed sensing inspired and convex-optimization based methods to solve the estimation problem for bacterial community composition. These methods typically rely on summarizing the sequence data by frequencies of low-order k-mers and matching this information statistically with a taxonomically structured database. Here we show that the accuracy of the resulting community composition estimates can be substantially improved by aggregating the reads from a sample with an unsupervised machine learning approach prior to the estimation phase. The aggregation of reads is a pre-processing approach where we use a standard K-means clustering algorithm that partitions a large set of reads into subsets with reasonable computational cost to provide several vectors of first order statistics instead of only single statistical summarization in terms of k-mer frequencies. The output of the clustering is then processed further to obtain the final estimate for each sample. The resulting method is called Aggregation of Reads by K-means (ARK), and it is based on a statistical argument via mixture density formulation. ARK is found to improve the fidelity and robustness of several recently introduced methods, with only a modest increase in computational complexity. An open source, platform-independent implementation of the method in the Julia programming language is freely available at https://github.com/dkoslicki/ARK. A Matlab implementation is available at http://www.ee.kth.se/ctsoftware.

  17. Integrating Gene Expression with Summary Association Statistics to Identify Genes Associated with 30 Complex Traits.

    PubMed

    Mancuso, Nicholas; Shi, Huwenbo; Goddard, Pagé; Kichaev, Gleb; Gusev, Alexander; Pasaniuc, Bogdan

    2017-03-02

    Although genome-wide association studies (GWASs) have identified thousands of risk loci for many complex traits and diseases, the causal variants and genes at these loci remain largely unknown. Here, we introduce a method for estimating the local genetic correlation between gene expression and a complex trait and utilize it to estimate the genetic correlation due to predicted expression between pairs of traits. We integrated gene expression measurements from 45 expression panels with summary GWAS data to perform 30 multi-tissue transcriptome-wide association studies (TWASs). We identified 1,196 genes whose expression is associated with these traits; of these, 168 reside more than 0.5 Mb away from any previously reported GWAS significant variant. We then used our approach to find 43 pairs of traits with significant genetic correlation at the level of predicted expression; of these, eight were not found through genetic correlation at the SNP level. Finally, we used bi-directional regression to find evidence that BMI causally influences triglyceride levels and that triglyceride levels causally influence low-density lipoprotein. Together, our results provide insight into the role of gene expression in the susceptibility of complex traits and diseases. Copyright © 2017 American Society of Human Genetics. Published by Elsevier Inc. All rights reserved.

  18. Benthic Crustacea from tropical and temperate reef locations: differences in assemblages and their relationship with habitat structure

    NASA Astrophysics Data System (ADS)

    Kramer, Michael J.; Bellwood, David R.; Taylor, Richard B.; Bellwood, Orpha

    2017-09-01

    Tropical and temperate marine habitats have long been recognised as fundamentally different system, yet comparative studies are rare, particularly for small organisms such as Crustacea. This study investigates the ecological attributes (abundance, biomass and estimated productivity) of benthic Crustacea in selected microhabitats from a tropical and a temperate location, revealing marked differences in the crustacean assemblages. In general, microhabitats from the tropical location (dead coral, the epilithic algal matrix [algal turfs] and sand) supported high abundances of small individuals (mean length = 0.53 mm vs. 0.96 mm in temperate microhabitats), while temperate microhabitats (the brown seaweed Carpophyllum sp., coralline turf and sand) had substantially greater biomasses of crustaceans and higher estimated productivity rates. In both locations, the most important microhabitats for crustaceans (per unit area) were complex structures: tropical dead coral and temperate Carpophyllum sp. It appears that the differences between microhabitats are largely driven by the size and relative abundance of key crustacean groups. Temperate microhabitats have a higher proportion of relatively large Peracarida (Amphipoda and Isopoda), whereas tropical microhabitats are dominated by small detrital- and microalgal-feeding crustaceans (harpacticoid copepods and ostracods). These differences highlight the vulnerability of tropical and temperate systems to the loss of complex benthic structures and their associated crustacean assemblages.

  19. Mapping of the extinction in giant molecular clouds using optical star counts

    NASA Astrophysics Data System (ADS)

    Cambrésy, L.

    1999-05-01

    This paper presents large scale extinction maps of most nearby Giant Molecular Clouds of the Galaxy (Lupus, rho Ophiuchus, Scorpius, Coalsack, Taurus, Chamaeleon, Musca, Corona Australis, Serpens, IC 5146, Vela, Orion, Monoceros R1 and R2, Rosette, Carina) derived from a star count method using an adaptive grid and a wavelet decomposition applied to the optical data provided by the USNO-Precision Measuring Machine. The distribution of the extinction in the clouds leads to estimate their total individual masses M and their maximum of extinction. I show that the relation between the mass contained within an iso-extinction contour and the extinction is similar from cloud to cloud and allows the extrapolation of the maximum of extinction in the range 5.7 to 25.5 magnitudes. I found that about half of the mass is contained in regions where the visual extinction is smaller than 1 magnitude. The star count method used on large scale ( ~ 250 square degrees) is a powerful and relatively straightforward method to estimate the mass of molecular complexes. A systematic study of the all sky would lead to discover new clouds as I did in the Lupus complex for which I found a sixth cloud of about 10(4) M_⊙.

  20. Housing accessibility for senior citizens in Sweden: Estimation of the effects of targeted elimination of environmental barriers.

    PubMed

    Pettersson, Cecilia; Slaug, Björn; Granbom, Marianne; Kylberg, Marianne; Iwarsson, Susanne

    2017-01-24

    To estimate the effects of targeted elimination of environmental barriers (EB) in the ordinary housing stock in Sweden, and to explore the estimated effects on accessibility at a population level in relation to (a) residents with different functional profiles, (b) different housing types and (c) building periods. Data on dwellings from existing Swedish research databases were utilized. EB and accessibility were assessed by means of the Housing Enabler instrument. In simulations of EB removal, five items that correspond to the most common housing adaptations were selected. The simulations were applied to four functional profiles of different complexity. EB known to be commonly removed by housing adaptations exist in large proportions of the existing ordinary housing stock. Estimated targeted elimination of selected barriers would have the largest accessibility effects for the more complex functional profiles. The effects would be consistently larger for one-family houses, and for all types of dwellings built before 1960. The elimination of the EB most commonly addressed by housing adaptations could result in a reduction of the housing accessibility problems that community-living older people are facing. For society to solve the housing situation for the ageing population well-informed and efficient upgrading of ordinary housing is imperative.

  1. Reducing uncertainties in energy dissipation measurements in atomic force spectroscopy of molecular networks and cell-adhesion studies.

    PubMed

    Biswas, Soma; Leitao, Samuel; Theillaud, Quentin; Erickson, Blake W; Fantner, Georg E

    2018-06-20

    Atomic force microscope (AFM) based single molecule force spectroscopy (SMFS) is a valuable tool in biophysics to investigate the ligand-receptor interactions, cell adhesion and cell mechanics. However, the force spectroscopy data analysis needs to be done carefully to extract the required quantitative parameters correctly. Especially the large number of molecules, commonly involved in complex networks formation; leads to very complicated force spectroscopy curves. One therefore, generally characterizes the total dissipated energy over a whole pulling cycle, as it is difficult to decompose the complex force curves into individual single molecule events. However, calculating the energy dissipation directly from the transformed force spectroscopy curves can lead to a significant over-estimation of the dissipated energy during a pulling experiment. The over-estimation of dissipated energy arises from the finite stiffness of the cantilever used for AFM based SMFS. Although this error can be significant, it is generally not compensated for. This can lead to significant misinterpretation of the energy dissipation (up to the order of 30%). In this paper, we show how in complex SMFS the excess dissipated energy caused by the stiffness of the cantilever can be identified and corrected using a high throughput algorithm. This algorithm is then applied to experimental results from molecular networks and cell-adhesion measurements to quantify the improvement in the estimation of the total energy dissipation.

  2. The quality estimation of exterior wall’s and window filling’s construction design

    NASA Astrophysics Data System (ADS)

    Saltykov, Ivan; Bovsunovskaya, Maria

    2017-10-01

    The article reveals the term of “artificial envelope” in dwelling building. Authors offer a complex multifactorial approach to the design quality estimation of external fencing structures, which is based on various parameters impact. These referred parameters are: functional, exploitation, cost, and also, the environmental index is among them. The quality design index Qк is inputting for the complex characteristic of observed above parameters. The mathematical relation of this index from these parameters is the target function for the quality design estimation. For instance, the article shows the search of optimal variant for wall and window designs in small, middle and large square dwelling premises of economic class buildings. The graphs of target function single parameters are expressed for the three types of residual chamber’s dimensions. As a result of the showing example, there is a choice of window opening’s dimensions, which make the wall’s and window’s constructions properly correspondent to the producible complex requirements. The authors reveal the comparison of recommended window filling’s square in accordance with the building standards, and the square, due to the finding of the optimal variant of the design quality index. The multifactorial approach for optimal design searching, which is mentioned in this article, can be used in consideration of various construction elements of dwelling buildings in accounting of suitable climate, social and economic construction area features.

  3. On the predictivity of pore-scale simulations: Estimating uncertainties with multilevel Monte Carlo

    NASA Astrophysics Data System (ADS)

    Icardi, Matteo; Boccardo, Gianluca; Tempone, Raúl

    2016-09-01

    A fast method with tunable accuracy is proposed to estimate errors and uncertainties in pore-scale and Digital Rock Physics (DRP) problems. The overall predictivity of these studies can be, in fact, hindered by many factors including sample heterogeneity, computational and imaging limitations, model inadequacy and not perfectly known physical parameters. The typical objective of pore-scale studies is the estimation of macroscopic effective parameters such as permeability, effective diffusivity and hydrodynamic dispersion. However, these are often non-deterministic quantities (i.e., results obtained for specific pore-scale sample and setup are not totally reproducible by another ;equivalent; sample and setup). The stochastic nature can arise due to the multi-scale heterogeneity, the computational and experimental limitations in considering large samples, and the complexity of the physical models. These approximations, in fact, introduce an error that, being dependent on a large number of complex factors, can be modeled as random. We propose a general simulation tool, based on multilevel Monte Carlo, that can reduce drastically the computational cost needed for computing accurate statistics of effective parameters and other quantities of interest, under any of these random errors. This is, to our knowledge, the first attempt to include Uncertainty Quantification (UQ) in pore-scale physics and simulation. The method can also provide estimates of the discretization error and it is tested on three-dimensional transport problems in heterogeneous materials, where the sampling procedure is done by generation algorithms able to reproduce realistic consolidated and unconsolidated random sphere and ellipsoid packings and arrangements. A totally automatic workflow is developed in an open-source code [1], that include rigid body physics and random packing algorithms, unstructured mesh discretization, finite volume solvers, extrapolation and post-processing techniques. The proposed method can be efficiently used in many porous media applications for problems such as stochastic homogenization/upscaling, propagation of uncertainty from microscopic fluid and rock properties to macro-scale parameters, robust estimation of Representative Elementary Volume size for arbitrary physics.

  4. Parameter Estimation for Viscoplastic Material Modeling

    NASA Technical Reports Server (NTRS)

    Saleeb, Atef F.; Gendy, Atef S.; Wilt, Thomas E.

    1997-01-01

    A key ingredient in the design of engineering components and structures under general thermomechanical loading is the use of mathematical constitutive models (e.g. in finite element analysis) capable of accurate representation of short and long term stress/deformation responses. In addition to the ever-increasing complexity of recent viscoplastic models of this type, they often also require a large number of material constants to describe a host of (anticipated) physical phenomena and complicated deformation mechanisms. In turn, the experimental characterization of these material parameters constitutes the major factor in the successful and effective utilization of any given constitutive model; i.e., the problem of constitutive parameter estimation from experimental measurements.

  5. A Backward-Lagrangian-Stochastic Footprint Model for the Urban Environment

    NASA Astrophysics Data System (ADS)

    Wang, Chenghao; Wang, Zhi-Hua; Yang, Jiachuan; Li, Qi

    2018-02-01

    Built terrains, with their complexity in morphology, high heterogeneity, and anthropogenic impact, impose substantial challenges in Earth-system modelling. In particular, estimation of the source areas and footprints of atmospheric measurements in cities requires realistic representation of the landscape characteristics and flow physics in urban areas, but has hitherto been heavily reliant on large-eddy simulations. In this study, we developed physical parametrization schemes for estimating urban footprints based on the backward-Lagrangian-stochastic algorithm, with the built environment represented by street canyons. The vertical profile of mean streamwise velocity is parametrized for the urban canopy and boundary layer. Flux footprints estimated by the proposed model show reasonable agreement with analytical predictions over flat surfaces without roughness elements, and with experimental observations over sparse plant canopies. Furthermore, comparisons of canyon flow and turbulence profiles and the subsequent footprints were made between the proposed model and large-eddy simulation data. The results suggest that the parametrized canyon wind and turbulence statistics, based on the simple similarity theory used, need to be further improved to yield more realistic urban footprint modelling.

  6. Building continental-scale 3D subsurface layers in the Digital Crust project: constrained interpolation and uncertainty estimation.

    NASA Astrophysics Data System (ADS)

    Yulaeva, E.; Fan, Y.; Moosdorf, N.; Richard, S. M.; Bristol, S.; Peters, S. E.; Zaslavsky, I.; Ingebritsen, S.

    2015-12-01

    The Digital Crust EarthCube building block creates a framework for integrating disparate 3D/4D information from multiple sources into a comprehensive model of the structure and composition of the Earth's upper crust, and to demonstrate the utility of this model in several research scenarios. One of such scenarios is estimation of various crustal properties related to fluid dynamics (e.g. permeability and porosity) at each node of any arbitrary unstructured 3D grid to support continental-scale numerical models of fluid flow and transport. Starting from Macrostrat, an existing 4D database of 33,903 chronostratigraphic units, and employing GeoDeepDive, a software system for extracting structured information from unstructured documents, we construct 3D gridded fields of sediment/rock porosity, permeability and geochemistry for large sedimentary basins of North America, which will be used to improve our understanding of large-scale fluid flow, chemical weathering rates, and geochemical fluxes into the ocean. In this talk, we discuss the methods, data gaps (particularly in geologically complex terrain), and various physical and geological constraints on interpolation and uncertainty estimation.

  7. Multicategory Composite Least Squares Classifiers

    PubMed Central

    Park, Seo Young; Liu, Yufeng; Liu, Dacheng; Scholl, Paul

    2010-01-01

    Classification is a very useful statistical tool for information extraction. In particular, multicategory classification is commonly seen in various applications. Although binary classification problems are heavily studied, extensions to the multicategory case are much less so. In view of the increased complexity and volume of modern statistical problems, it is desirable to have multicategory classifiers that are able to handle problems with high dimensions and with a large number of classes. Moreover, it is necessary to have sound theoretical properties for the multicategory classifiers. In the literature, there exist several different versions of simultaneous multicategory Support Vector Machines (SVMs). However, the computation of the SVM can be difficult for large scale problems, especially for problems with large number of classes. Furthermore, the SVM cannot produce class probability estimation directly. In this article, we propose a novel efficient multicategory composite least squares classifier (CLS classifier), which utilizes a new composite squared loss function. The proposed CLS classifier has several important merits: efficient computation for problems with large number of classes, asymptotic consistency, ability to handle high dimensional data, and simple conditional class probability estimation. Our simulated and real examples demonstrate competitive performance of the proposed approach. PMID:21218128

  8. A stepwise regression tree for nonlinear approximation: applications to estimating subpixel land cover

    USGS Publications Warehouse

    Huang, C.; Townshend, J.R.G.

    2003-01-01

    A stepwise regression tree (SRT) algorithm was developed for approximating complex nonlinear relationships. Based on the regression tree of Breiman et al . (BRT) and a stepwise linear regression (SLR) method, this algorithm represents an improvement over SLR in that it can approximate nonlinear relationships and over BRT in that it gives more realistic predictions. The applicability of this method to estimating subpixel forest was demonstrated using three test data sets, on all of which it gave more accurate predictions than SLR and BRT. SRT also generated more compact trees and performed better than or at least as well as BRT at all 10 equal forest proportion interval ranging from 0 to 100%. This method is appealing to estimating subpixel land cover over large areas.

  9. Improving breeding efficiency in potato using molecular and quantitative genetics.

    PubMed

    Slater, Anthony T; Cogan, Noel O I; Hayes, Benjamin J; Schultz, Lee; Dale, M Finlay B; Bryan, Glenn J; Forster, John W

    2014-11-01

    Potatoes are highly heterozygous and the conventional breeding of superior germplasm is challenging, but use of a combination of MAS and EBVs can accelerate genetic gain. Cultivated potatoes are highly heterozygous due to their outbreeding nature, and suffer acute inbreeding depression. Modern potato cultivars also exhibit tetrasomic inheritance. Due to this genetic heterogeneity, the large number of target traits and the specific requirements of commercial cultivars, potato breeding is challenging. A conventional breeding strategy applies phenotypic recurrent selection over a number of generations, a process which can take over 10 years. Recently, major advances in genetics and molecular biology have provided breeders with molecular tools to accelerate gains for some traits. Marker-assisted selection (MAS) can be effectively used for the identification of major genes and quantitative trait loci that exhibit large effects. There are also a number of complex traits of interest, such as yield, that are influenced by a large number of genes of individual small effect where MAS will be difficult to deploy. Progeny testing and the use of pedigree in the analysis can provide effective identification of the superior genetic factors that underpin these complex traits. Recently, it has been shown that estimated breeding values (EBVs) can be developed for complex potato traits. Using a combination of MAS and EBVs for simple and complex traits can lead to a significant reduction in the length of the breeding cycle for the identification of superior germplasm.

  10. Probability based remaining capacity estimation using data-driven and neural network model

    NASA Astrophysics Data System (ADS)

    Wang, Yujie; Yang, Duo; Zhang, Xu; Chen, Zonghai

    2016-05-01

    Since large numbers of lithium-ion batteries are composed in pack and the batteries are complex electrochemical devices, their monitoring and safety concerns are key issues for the applications of battery technology. An accurate estimation of battery remaining capacity is crucial for optimization of the vehicle control, preventing battery from over-charging and over-discharging and ensuring the safety during its service life. The remaining capacity estimation of a battery includes the estimation of state-of-charge (SOC) and state-of-energy (SOE). In this work, a probability based adaptive estimator is presented to obtain accurate and reliable estimation results for both SOC and SOE. For the SOC estimation, an n ordered RC equivalent circuit model is employed by combining an electrochemical model to obtain more accurate voltage prediction results. For the SOE estimation, a sliding window neural network model is proposed to investigate the relationship between the terminal voltage and the model inputs. To verify the accuracy and robustness of the proposed model and estimation algorithm, experiments under different dynamic operation current profiles are performed on the commercial 1665130-type lithium-ion batteries. The results illustrate that accurate and robust estimation can be obtained by the proposed method.

  11. The Hidden Fortress: structure and substructure of the complex strong lensing cluster SDSS J1029+2623

    NASA Astrophysics Data System (ADS)

    Oguri, Masamune; Schrabback, Tim; Jullo, Eric; Ota, Naomi; Kochanek, Christopher S.; Dai, Xinyu; Ofek, Eran O.; Richards, Gordon T.; Blandford, Roger D.; Falco, Emilio E.; Fohlmeister, Janine

    2013-02-01

    We present Hubble Space Telescope (HST) Advanced Camera for Surveys (ACS) and Wide Field Camera 3 (WFC3) observations of SDSS J1029+2623, a three-image quasar lens system produced by a foreground cluster at z = 0.584. Our strong lensing analysis reveals six additional multiply imaged galaxies in addition to the multiply imaged quasar. We confirm the complex nature of the mass distribution of the lensing cluster, with a bimodal dark matter distribution which deviates from the Chandra X-ray surface brightness distribution. The Einstein radius of the lensing cluster is estimated to be θE = 15.2 ± 0.5 arcsec for the quasar redshift of z = 2.197. We derive a radial mass distribution from the combination of strong lensing, HST/ACS weak lensing and Subaru/Suprime-cam weak lensing analysis results, finding a best-fitting virial mass of Mvir = 1.55+ 0.40- 0.35 × 1014 h- 1 M⊙ and a concentration parameter of cvir = 25.7+ 14.1- 7.5. The lensing mass estimate at the outer radius is smaller than the X-ray mass estimate by a factor of ˜2. We ascribe this large mass discrepancy to shock heating of the intracluster gas during a merger, which is also suggested by the complex mass and gas distributions and the high value of the concentration parameter. In the HST image, we also identify a probable galaxy, GX, in the vicinity of the faintest quasar image C. In strong lens models, the inclusion of GX explains the anomalous flux ratios between the quasar images. The morphology of the highly elongated quasar host galaxy is also well reproduced. The best-fitting model suggests large total magnifications of 30 for the quasar and 35 for the quasar host galaxy, and has an AB time delay consistent with the measured value.

  12. Analysis of Power Laws, Shape Collapses, and Neural Complexity: New Techniques and MATLAB Support via the NCC Toolbox.

    PubMed

    Marshall, Najja; Timme, Nicholas M; Bennett, Nicholas; Ripp, Monica; Lautzenhiser, Edward; Beggs, John M

    2016-01-01

    Neural systems include interactions that occur across many scales. Two divergent methods for characterizing such interactions have drawn on the physical analysis of critical phenomena and the mathematical study of information. Inferring criticality in neural systems has traditionally rested on fitting power laws to the property distributions of "neural avalanches" (contiguous bursts of activity), but the fractal nature of avalanche shapes has recently emerged as another signature of criticality. On the other hand, neural complexity, an information theoretic measure, has been used to capture the interplay between the functional localization of brain regions and their integration for higher cognitive functions. Unfortunately, treatments of all three methods-power-law fitting, avalanche shape collapse, and neural complexity-have suffered from shortcomings. Empirical data often contain biases that introduce deviations from true power law in the tail and head of the distribution, but deviations in the tail have often been unconsidered; avalanche shape collapse has required manual parameter tuning; and the estimation of neural complexity has relied on small data sets or statistical assumptions for the sake of computational efficiency. In this paper we present technical advancements in the analysis of criticality and complexity in neural systems. We use maximum-likelihood estimation to automatically fit power laws with left and right cutoffs, present the first automated shape collapse algorithm, and describe new techniques to account for large numbers of neural variables and small data sets in the calculation of neural complexity. In order to facilitate future research in criticality and complexity, we have made the software utilized in this analysis freely available online in the MATLAB NCC (Neural Complexity and Criticality) Toolbox.

  13. Contribution of Large Region Joint Associations to Complex Traits Genetics

    PubMed Central

    Paré, Guillaume; Asma, Senay; Deng, Wei Q.

    2015-01-01

    A polygenic model of inheritance, whereby hundreds or thousands of weakly associated variants contribute to a trait’s heritability, has been proposed to underlie the genetic architecture of complex traits. However, relatively few genetic variants have been positively identified so far and they collectively explain only a small fraction of the predicted heritability. We hypothesized that joint association of multiple weakly associated variants over large chromosomal regions contributes to complex traits variance. Confirmation of such regional associations can help identify new loci and lead to a better understanding of known ones. To test this hypothesis, we first characterized the ability of commonly used genetic association models to identify large region joint associations. Through theoretical derivation and simulation, we showed that multivariate linear models where multiple SNPs are included as independent predictors have the most favorable association profile. Based on these results, we tested for large region association with height in 3,740 European participants from the Health and Retirement Study (HRS) study. Adjusting for SNPs with known association with height, we demonstrated clustering of weak associations (p = 2x10-4) in regions extending up to 433.0 Kb from known height loci. The contribution of regional associations to phenotypic variance was estimated at 0.172 (95% CI 0.063-0.279; p < 0.001), which compared favorably to 0.129 explained by known height variants. Conversely, we showed that suggestively associated regions are enriched for known height loci. To extend our findings to other traits, we also tested BMI, HDLc and CRP for large region associations, with consistent results for CRP. Our results demonstrate the presence of large region joint associations and suggest these can be used to pinpoint weakly associated SNPs. PMID:25856144

  14. Optimal space-time attacks on system state estimation under a sparsity constraint

    NASA Astrophysics Data System (ADS)

    Lu, Jingyang; Niu, Ruixin; Han, Puxiao

    2016-05-01

    System state estimation in the presence of an adversary that injects false information into sensor readings has attracted much attention in wide application areas, such as target tracking with compromised sensors, secure monitoring of dynamic electric power systems, secure driverless cars, and radar tracking and detection in the presence of jammers. From a malicious adversary's perspective, the optimal strategy for attacking a multi-sensor dynamic system over sensors and over time is investigated. It is assumed that the system defender can perfectly detect the attacks and identify and remove sensor data once they are corrupted by false information injected by the adversary. With this in mind, the adversary's goal is to maximize the covariance matrix of the system state estimate by the end of attack period under a sparse attack constraint such that the adversary can only attack the system a few times over time and over sensors. The sparsity assumption is due to the adversary's limited resources and his/her intention to reduce the chance of being detected by the system defender. This becomes an integer programming problem and its optimal solution, the exhaustive search, is intractable with a prohibitive complexity, especially for a system with a large number of sensors and over a large number of time steps. Several suboptimal solutions, such as those based on greedy search and dynamic programming are proposed to find the attack strategies. Examples and numerical results are provided in order to illustrate the effectiveness and the reduced computational complexities of the proposed attack strategies.

  15. Adaptive Elastic Net for Generalized Methods of Moments.

    PubMed

    Caner, Mehmet; Zhang, Hao Helen

    2014-01-30

    Model selection and estimation are crucial parts of econometrics. This paper introduces a new technique that can simultaneously estimate and select the model in generalized method of moments (GMM) context. The GMM is particularly powerful for analyzing complex data sets such as longitudinal and panel data, and it has wide applications in econometrics. This paper extends the least squares based adaptive elastic net estimator of Zou and Zhang (2009) to nonlinear equation systems with endogenous variables. The extension is not trivial and involves a new proof technique due to estimators lack of closed form solutions. Compared to Bridge-GMM of Caner (2009), we allow for the number of parameters to diverge to infinity as well as collinearity among a large number of variables, also the redundant parameters set to zero via a data dependent technique. This method has the oracle property, meaning that we can estimate nonzero parameters with their standard limit and the redundant parameters are dropped from the equations simultaneously. Numerical examples are used to illustrate the performance of the new method.

  16. Using geostatistical methods to estimate snow water equivalence distribution in a mountain watershed

    USGS Publications Warehouse

    Balk, B.; Elder, K.; Baron, Jill S.

    1998-01-01

    Knowledge of the spatial distribution of snow water equivalence (SWE) is necessary to adequately forecast the volume and timing of snowmelt runoff.  In April 1997, peak accumulation snow depth and density measurements were independently taken in the Loch Vale watershed (6.6 km2), Rocky Mountain National Park, Colorado.  Geostatistics and classical statistics were used to estimate SWE distribution across the watershed.  Snow depths were spatially distributed across the watershed through kriging interpolation methods which provide unbiased estimates that have minimum variances.  Snow densities were spatially modeled through regression analysis.  Combining the modeled depth and density with snow-covered area (SCA produced an estimate of the spatial distribution of SWE.  The kriged estimates of snow depth explained 37-68% of the observed variance in the measured depths.  Steep slopes, variably strong winds, and complex energy balance in the watershed contribute to a large degree of heterogeneity in snow depth.

  17. Parameter estimation procedure for complex non-linear systems: calibration of ASM No. 1 for N-removal in a full-scale oxidation ditch.

    PubMed

    Abusam, A; Keesman, K J; van Straten, G; Spanjers, H; Meinema, K

    2001-01-01

    When applied to large simulation models, the process of parameter estimation is also called calibration. Calibration of complex non-linear systems, such as activated sludge plants, is often not an easy task. On the one hand, manual calibration of such complex systems is usually time-consuming, and its results are often not reproducible. On the other hand, conventional automatic calibration methods are not always straightforward and often hampered by local minima problems. In this paper a new straightforward and automatic procedure, which is based on the response surface method (RSM) for selecting the best identifiable parameters, is proposed. In RSM, the process response (output) is related to the levels of the input variables in terms of a first- or second-order regression model. Usually, RSM is used to relate measured process output quantities to process conditions. However, in this paper RSM is used for selecting the dominant parameters, by evaluating parameters sensitivity in a predefined region. Good results obtained in calibration of ASM No. 1 for N-removal in a full-scale oxidation ditch proved that the proposed procedure is successful and reliable.

  18. Horvitz-Thompson survey sample methods for estimating large-scale animal abundance

    USGS Publications Warehouse

    Samuel, M.D.; Garton, E.O.

    1994-01-01

    Large-scale surveys to estimate animal abundance can be useful for monitoring population status and trends, for measuring responses to management or environmental alterations, and for testing ecological hypotheses about abundance. However, large-scale surveys may be expensive and logistically complex. To ensure resources are not wasted on unattainable targets, the goals and uses of each survey should be specified carefully and alternative methods for addressing these objectives always should be considered. During survey design, the impoflance of each survey error component (spatial design, propofiion of detected animals, precision in detection) should be considered carefully to produce a complete statistically based survey. Failure to address these three survey components may produce population estimates that are inaccurate (biased low), have unrealistic precision (too precise) and do not satisfactorily meet the survey objectives. Optimum survey design requires trade-offs in these sources of error relative to the costs of sampling plots and detecting animals on plots, considerations that are specific to the spatial logistics and survey methods. The Horvitz-Thompson estimators provide a comprehensive framework for considering all three survey components during the design and analysis of large-scale wildlife surveys. Problems of spatial and temporal (especially survey to survey) heterogeneity in detection probabilities have received little consideration, but failure to account for heterogeneity produces biased population estimates. The goal of producing unbiased population estimates is in conflict with the increased variation from heterogeneous detection in the population estimate. One solution to this conflict is to use an MSE-based approach to achieve a balance between bias reduction and increased variation. Further research is needed to develop methods that address spatial heterogeneity in detection, evaluate the effects of temporal heterogeneity on survey objectives and optimize decisions related to survey bias and variance. Finally, managers and researchers involved in the survey design process must realize that obtaining the best survey results requires an interactive and recursive process of survey design, execution, analysis and redesign. Survey refinements will be possible as further knowledge is gained on the actual abundance and distribution of the population and on the most efficient techniques for detection animals.

  19. Gray: a ray tracing-based Monte Carlo simulator for PET.

    PubMed

    Freese, David L; Olcott, Peter D; Buss, Samuel R; Levin, Craig S

    2018-05-21

    Monte Carlo simulation software plays a critical role in PET system design. Performing complex, repeated Monte Carlo simulations can be computationally prohibitive, as even a single simulation can require a large amount of time and a computing cluster to complete. Here we introduce Gray, a Monte Carlo simulation software for PET systems. Gray exploits ray tracing methods used in the computer graphics community to greatly accelerate simulations of PET systems with complex geometries. We demonstrate the implementation of models for positron range, annihilation acolinearity, photoelectric absorption, Compton scatter, and Rayleigh scatter. For validation, we simulate the GATE PET benchmark, and compare energy, distribution of hits, coincidences, and run time. We show a [Formula: see text] speedup using Gray, compared to GATE for the same simulation, while demonstrating nearly identical results. We additionally simulate the Siemens Biograph mCT system with both the NEMA NU-2 scatter phantom and sensitivity phantom. We estimate the total sensitivity within [Formula: see text]% when accounting for differences in peak NECR. We also estimate the peak NECR to be [Formula: see text] kcps, or within [Formula: see text]% of published experimental data. The activity concentration of the peak is also estimated within 1.3%.

  20. A near-optimal low complexity sensor fusion technique for accurate indoor localization based on ultrasound time of arrival measurements from low-quality sensors

    NASA Astrophysics Data System (ADS)

    Mitilineos, Stelios A.; Argyreas, Nick D.; Thomopoulos, Stelios C. A.

    2009-05-01

    A fusion-based localization technique for location-based services in indoor environments is introduced herein, based on ultrasound time-of-arrival measurements from multiple off-the-shelf range estimating sensors which are used in a market-available localization system. In-situ field measurements results indicated that the respective off-the-shelf system was unable to estimate position in most of the cases, while the underlying sensors are of low-quality and yield highly inaccurate range and position estimates. An extensive analysis is performed and a model of the sensor-performance characteristics is established. A low-complexity but accurate sensor fusion and localization technique is then developed, which consists inof evaluating multiple sensor measurements and selecting the one that is considered most-accurate based on the underlying sensor model. Optimality, in the sense of a genie selecting the optimum sensor, is subsequently evaluated and compared to the proposed technique. The experimental results indicate that the proposed fusion method exhibits near-optimal performance and, albeit being theoretically suboptimal, it largely overcomes most flaws of the underlying single-sensor system resulting in a localization system of increased accuracy, robustness and availability.

  1. Quantum algorithm for association rules mining

    NASA Astrophysics Data System (ADS)

    Yu, Chao-Hua; Gao, Fei; Wang, Qing-Le; Wen, Qiao-Yan

    2016-10-01

    Association rules mining (ARM) is one of the most important problems in knowledge discovery and data mining. Given a transaction database that has a large number of transactions and items, the task of ARM is to acquire consumption habits of customers by discovering the relationships between itemsets (sets of items). In this paper, we address ARM in the quantum settings and propose a quantum algorithm for the key part of ARM, finding frequent itemsets from the candidate itemsets and acquiring their supports. Specifically, for the case in which there are Mf(k ) frequent k -itemsets in the Mc(k ) candidate k -itemsets (Mf(k )≤Mc(k ) ), our algorithm can efficiently mine these frequent k -itemsets and estimate their supports by using parallel amplitude estimation and amplitude amplification with complexity O (k/√{Mc(k )Mf(k ) } ɛ ) , where ɛ is the error for estimating the supports. Compared with the classical counterpart, i.e., the classical sampling-based algorithm, whose complexity is O (k/Mc(k ) ɛ2) , our quantum algorithm quadratically improves the dependence on both ɛ and Mc(k ) in the best case when Mf(k )≪Mc(k ) and on ɛ alone in the worst case when Mf(k )≈Mc(k ) .

  2. Department of Defense Fiscal Year (FY) 2005 Budget Estimates. Research, Development, Test and Evaluation, Defense-Wide. Volume 1 - Defense Advanced Research Projects Agency

    DTIC Science & Technology

    2004-02-01

    UNCLASSIFIED − Conducted experiments to determine the usability of general-purpose anomaly detection algorithms to monitor a large, complex military...reaction and detection modules to perform tailored analysis sequences to monitor environmental conditions, health hazards and physiological states...scalability of lab proven anomaly detection techniques for intrusion detection in real world high volume environments. Narrative Title FY 2003

  3. Identifying Optimal Measurement Subspace for the Ensemble Kalman Filter

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Zhou, Ning; Huang, Zhenyu; Welch, Greg

    2012-05-24

    To reduce the computational load of the ensemble Kalman filter while maintaining its efficacy, an optimization algorithm based on the generalized eigenvalue decomposition method is proposed for identifying the most informative measurement subspace. When the number of measurements is large, the proposed algorithm can be used to make an effective tradeoff between computational complexity and estimation accuracy. This algorithm also can be extended to other Kalman filters for measurement subspace selection.

  4. A Spectral Method for Spatial Downscaling

    PubMed Central

    Reich, Brian J.; Chang, Howard H.; Foley, Kristen M.

    2014-01-01

    Summary Complex computer models play a crucial role in air quality research. These models are used to evaluate potential regulatory impacts of emission control strategies and to estimate air quality in areas without monitoring data. For both of these purposes, it is important to calibrate model output with monitoring data to adjust for model biases and improve spatial prediction. In this article, we propose a new spectral method to study and exploit complex relationships between model output and monitoring data. Spectral methods allow us to estimate the relationship between model output and monitoring data separately at different spatial scales, and to use model output for prediction only at the appropriate scales. The proposed method is computationally efficient and can be implemented using standard software. We apply the method to compare Community Multiscale Air Quality (CMAQ) model output with ozone measurements in the United States in July 2005. We find that CMAQ captures large-scale spatial trends, but has low correlation with the monitoring data at small spatial scales. PMID:24965037

  5. Rule-based modeling and simulations of the inner kinetochore structure.

    PubMed

    Tschernyschkow, Sergej; Herda, Sabine; Gruenert, Gerd; Döring, Volker; Görlich, Dennis; Hofmeister, Antje; Hoischen, Christian; Dittrich, Peter; Diekmann, Stephan; Ibrahim, Bashar

    2013-09-01

    Combinatorial complexity is a central problem when modeling biochemical reaction networks, since the association of a few components can give rise to a large variation of protein complexes. Available classical modeling approaches are often insufficient for the analysis of very large and complex networks in detail. Recently, we developed a new rule-based modeling approach that facilitates the analysis of spatial and combinatorially complex problems. Here, we explore for the first time how this approach can be applied to a specific biological system, the human kinetochore, which is a multi-protein complex involving over 100 proteins. Applying our freely available SRSim software to a large data set on kinetochore proteins in human cells, we construct a spatial rule-based simulation model of the human inner kinetochore. The model generates an estimation of the probability distribution of the inner kinetochore 3D architecture and we show how to analyze this distribution using information theory. In our model, the formation of a bridge between CenpA and an H3 containing nucleosome only occurs efficiently for higher protein concentration realized during S-phase but may be not in G1. Above a certain nucleosome distance the protein bridge barely formed pointing towards the importance of chromatin structure for kinetochore complex formation. We define a metric for the distance between structures that allow us to identify structural clusters. Using this modeling technique, we explore different hypothetical chromatin layouts. Applying a rule-based network analysis to the spatial kinetochore complex geometry allowed us to integrate experimental data on kinetochore proteins, suggesting a 3D model of the human inner kinetochore architecture that is governed by a combinatorial algebraic reaction network. This reaction network can serve as bridge between multiple scales of modeling. Our approach can be applied to other systems beyond kinetochores. Copyright © 2013 Elsevier Ltd. All rights reserved.

  6. Markov state models from short non-equilibrium simulations—Analysis and correction of estimation bias

    NASA Astrophysics Data System (ADS)

    Nüske, Feliks; Wu, Hao; Prinz, Jan-Hendrik; Wehmeyer, Christoph; Clementi, Cecilia; Noé, Frank

    2017-03-01

    Many state-of-the-art methods for the thermodynamic and kinetic characterization of large and complex biomolecular systems by simulation rely on ensemble approaches, where data from large numbers of relatively short trajectories are integrated. In this context, Markov state models (MSMs) are extremely popular because they can be used to compute stationary quantities and long-time kinetics from ensembles of short simulations, provided that these short simulations are in "local equilibrium" within the MSM states. However, over the last 15 years since the inception of MSMs, it has been controversially discussed and not yet been answered how deviations from local equilibrium can be detected, whether these deviations induce a practical bias in MSM estimation, and how to correct for them. In this paper, we address these issues: We systematically analyze the estimation of MSMs from short non-equilibrium simulations, and we provide an expression for the error between unbiased transition probabilities and the expected estimate from many short simulations. We show that the unbiased MSM estimate can be obtained even from relatively short non-equilibrium simulations in the limit of long lag times and good discretization. Further, we exploit observable operator model (OOM) theory to derive an unbiased estimator for the MSM transition matrix that corrects for the effect of starting out of equilibrium, even when short lag times are used. Finally, we show how the OOM framework can be used to estimate the exact eigenvalues or relaxation time scales of the system without estimating an MSM transition matrix, which allows us to practically assess the discretization quality of the MSM. Applications to model systems and molecular dynamics simulation data of alanine dipeptide are included for illustration. The improved MSM estimator is implemented in PyEMMA of version 2.3.

  7. Estimation of Soil Moisture from Optical and Thermal Remote Sensing: A Review

    PubMed Central

    Zhang, Dianjun; Zhou, Guoqing

    2016-01-01

    As an important parameter in recent and numerous environmental studies, soil moisture (SM) influences the exchange of water and energy at the interface between the land surface and atmosphere. Accurate estimate of the spatio-temporal variations of SM is critical for numerous large-scale terrestrial studies. Although microwave remote sensing provides many algorithms to obtain SM at large scale, such as SMOS and SMAP etc., resulting in many data products, they are almost low resolution and not applicable in small catchment or field scale. Estimations of SM from optical and thermal remote sensing have been studied for many years and significant progress has been made. In contrast to previous reviews, this paper presents a new, comprehensive and systematic review of using optical and thermal remote sensing for estimating SM. The physical basis and status of the estimation methods are analyzed and summarized in detail. The most important and latest advances in soil moisture estimation using temporal information have been shown in this paper. SM estimation from optical and thermal remote sensing mainly depends on the relationship between SM and the surface reflectance or vegetation index. The thermal infrared remote sensing methods uses the relationship between SM and the surface temperature or variations of surface temperature/vegetation index. These approaches often have complex derivation processes and many approximations. Therefore, combinations of optical and thermal infrared remotely sensed data can provide more valuable information for SM estimation. Moreover, the advantages and weaknesses of different approaches are compared and applicable conditions as well as key issues in current soil moisture estimation algorithms are discussed. Finally, key problems and suggested solutions are proposed for future research. PMID:27548168

  8. Estimation of Soil Moisture from Optical and Thermal Remote Sensing: A Review.

    PubMed

    Zhang, Dianjun; Zhou, Guoqing

    2016-08-17

    As an important parameter in recent and numerous environmental studies, soil moisture (SM) influences the exchange of water and energy at the interface between the land surface and atmosphere. Accurate estimate of the spatio-temporal variations of SM is critical for numerous large-scale terrestrial studies. Although microwave remote sensing provides many algorithms to obtain SM at large scale, such as SMOS and SMAP etc., resulting in many data products, they are almost low resolution and not applicable in small catchment or field scale. Estimations of SM from optical and thermal remote sensing have been studied for many years and significant progress has been made. In contrast to previous reviews, this paper presents a new, comprehensive and systematic review of using optical and thermal remote sensing for estimating SM. The physical basis and status of the estimation methods are analyzed and summarized in detail. The most important and latest advances in soil moisture estimation using temporal information have been shown in this paper. SM estimation from optical and thermal remote sensing mainly depends on the relationship between SM and the surface reflectance or vegetation index. The thermal infrared remote sensing methods uses the relationship between SM and the surface temperature or variations of surface temperature/vegetation index. These approaches often have complex derivation processes and many approximations. Therefore, combinations of optical and thermal infrared remotely sensed data can provide more valuable information for SM estimation. Moreover, the advantages and weaknesses of different approaches are compared and applicable conditions as well as key issues in current soil moisture estimation algorithms are discussed. Finally, key problems and suggested solutions are proposed for future research.

  9. The Transition from Hydrogen Bonding to Ionization in (HCI)n(NH3)n and (HCI)n(H2O)n Clusters: Consequences for Anharmonic Vibrational Spectroscopy

    NASA Technical Reports Server (NTRS)

    Chaban, Galina M.; Gerber, R. Benny; Janda, Kenneth C.; Kwak, Dochan (Technical Monitor)

    2001-01-01

    Anharmonic vibrational frequencies and intensities are calculated for 1:1 and 2:2 (HCl)(sub n)(NH3)(sub n) and (HCl)(sub n)(H2O)(sub n) complexes, employing the correlation-corrected vibrational self-consistent field method with ab initio potential surfaces at the MP2/TZP computational level. In this method, the anharmonic coupling between all vibrational modes is included, which is found to be important for the systems studied. For the 4:4 (HCl)(sub n)(H2O)(sub n) complex, the vibrational spectra are calculated at the harmonic level, and anharmonic effects are estimated. Just as the (HCl)(sub n)(NH3)(sub n) structure switches from hydrogen-bonded to ionic for n=2, the (HCl)(sub n)(H2O)(sub n) switches to ionic structure for n=4. For (HCl)2(H2O)2, the lowest energy structure corresponds to the hydrogen-bonded form. However, configurations of the ionic form are separated from this minimum by a barrier of less than an O-H stretching quantum. This suggests the possibility of experiments on ionization dynamics using infrared excitation of the hydrogen-bonded form. The strong cooperative effects on the hydrogen bonding, and concomitant transition to ionic bonding, makes an accurate estimate of the large anharmonicity crucial for understanding the infrared spectra of these systems. The anharmonicity is typically of the order of several hundred wave numbers for the proton stretching motions involved in hydrogen or ionic bonding, and can also be quite large for the intramolecular modes. In addition, the large cooperative effects in the 2:2 and higher order (HCl(sub n)(H2O)(sub n) complexes may have interesting implications for solvation of hydrogen halides at ice surfaces.

  10. Reduced order models for prediction of groundwater quality impacts from CO₂ and brine leakage

    DOE PAGES

    Zheng, Liange; Carroll, Susan; Bianchi, Marco; ...

    2014-12-31

    A careful assessment of the risk associated with geologic CO₂ storage is critical to the deployment of large-scale storage projects. A potential risk is the deterioration of groundwater quality caused by the leakage of CO₂ and brine leakage from deep subsurface reservoirs. In probabilistic risk assessment studies, numerical modeling is the primary tool employed to assess risk. However, the application of traditional numerical models to fully evaluate the impact of CO₂ leakage on groundwater can be computationally complex, demanding large processing times and resources, and involving large uncertainties. As an alternative, reduced order models (ROMs) can be used as highlymore » efficient surrogates for the complex process-based numerical models. In this study, we represent the complex hydrogeological and geochemical conditions in a heterogeneous aquifer and subsequent risk by developing and using two separate ROMs. The first ROM is derived from a model that accounts for the heterogeneous flow and transport conditions in the presence of complex leakage functions for CO₂ and brine. The second ROM is obtained from models that feature similar, but simplified flow and transport conditions, and allow for a more complex representation of all relevant geochemical reactions. To quantify possible impacts to groundwater aquifers, the basic risk metric is taken as the aquifer volume in which the water quality of the aquifer may be affected by an underlying CO₂ storage project. The integration of the two ROMs provides an estimate of the impacted aquifer volume taking into account uncertainties in flow, transport and chemical conditions. These two ROMs can be linked in a comprehensive system level model for quantitative risk assessment of the deep storage reservoir, wellbore leakage, and shallow aquifer impacts to assess the collective risk of CO₂ storage projects.« less

  11. Using Full Genomic Information to Predict Disease: Breaking Down the Barriers Between Complex and Mendelian Diseases.

    PubMed

    Jordan, Daniel M; Do, Ron

    2018-04-11

    While sequence-based genetic tests have long been available for specific loci, especially for Mendelian disease, the rapidly falling costs of genome-wide genotyping arrays, whole-exome sequencing, and whole-genome sequencing are moving us toward a future where full genomic information might inform the prognosis and treatment of a variety of diseases, including complex disease. Similarly, the availability of large populations with full genomic information has enabled new insights about the etiology and genetic architecture of complex disease. Insights from the latest generation of genomic studies suggest that our categorization of diseases as complex may conceal a wide spectrum of genetic architectures and causal mechanisms that ranges from Mendelian forms of complex disease to complex regulatory structures underlying Mendelian disease. Here, we review these insights, along with advances in the prediction of disease risk and outcomes from full genomic information. Expected final online publication date for the Annual Review of Genomics and Human Genetics Volume 19 is August 31, 2018. Please see http://www.annualreviews.org/page/journal/pubdates for revised estimates.

  12. Bayesian dose-response analysis for epidemiological studies with complex uncertainty in dose estimation.

    PubMed

    Kwon, Deukwoo; Hoffman, F Owen; Moroz, Brian E; Simon, Steven L

    2016-02-10

    Most conventional risk analysis methods rely on a single best estimate of exposure per person, which does not allow for adjustment for exposure-related uncertainty. Here, we propose a Bayesian model averaging method to properly quantify the relationship between radiation dose and disease outcomes by accounting for shared and unshared uncertainty in estimated dose. Our Bayesian risk analysis method utilizes multiple realizations of sets (vectors) of doses generated by a two-dimensional Monte Carlo simulation method that properly separates shared and unshared errors in dose estimation. The exposure model used in this work is taken from a study of the risk of thyroid nodules among a cohort of 2376 subjects who were exposed to fallout from nuclear testing in Kazakhstan. We assessed the performance of our method through an extensive series of simulations and comparisons against conventional regression risk analysis methods. When the estimated doses contain relatively small amounts of uncertainty, the Bayesian method using multiple a priori plausible draws of dose vectors gave similar results to the conventional regression-based methods of dose-response analysis. However, when large and complex mixtures of shared and unshared uncertainties are present, the Bayesian method using multiple dose vectors had significantly lower relative bias than conventional regression-based risk analysis methods and better coverage, that is, a markedly increased capability to include the true risk coefficient within the 95% credible interval of the Bayesian-based risk estimate. An evaluation of the dose-response using our method is presented for an epidemiological study of thyroid disease following radiation exposure. Copyright © 2015 John Wiley & Sons, Ltd.

  13. A Low Cost Matching Motion Estimation Sensor Based on the NIOS II Microprocessor

    PubMed Central

    González, Diego; Botella, Guillermo; Meyer-Baese, Uwe; García, Carlos; Sanz, Concepción; Prieto-Matías, Manuel; Tirado, Francisco

    2012-01-01

    This work presents the implementation of a matching-based motion estimation sensor on a Field Programmable Gate Array (FPGA) and NIOS II microprocessor applying a C to Hardware (C2H) acceleration paradigm. The design, which involves several matching algorithms, is mapped using Very Large Scale Integration (VLSI) technology. These algorithms, as well as the hardware implementation, are presented here together with an extensive analysis of the resources needed and the throughput obtained. The developed low-cost system is practical for real-time throughput and reduced power consumption and is useful in robotic applications, such as tracking, navigation using an unmanned vehicle, or as part of a more complex system. PMID:23201989

  14. Analysis of Power Laws, Shape Collapses, and Neural Complexity: New Techniques and MATLAB Support via the NCC Toolbox

    PubMed Central

    Marshall, Najja; Timme, Nicholas M.; Bennett, Nicholas; Ripp, Monica; Lautzenhiser, Edward; Beggs, John M.

    2016-01-01

    Neural systems include interactions that occur across many scales. Two divergent methods for characterizing such interactions have drawn on the physical analysis of critical phenomena and the mathematical study of information. Inferring criticality in neural systems has traditionally rested on fitting power laws to the property distributions of “neural avalanches” (contiguous bursts of activity), but the fractal nature of avalanche shapes has recently emerged as another signature of criticality. On the other hand, neural complexity, an information theoretic measure, has been used to capture the interplay between the functional localization of brain regions and their integration for higher cognitive functions. Unfortunately, treatments of all three methods—power-law fitting, avalanche shape collapse, and neural complexity—have suffered from shortcomings. Empirical data often contain biases that introduce deviations from true power law in the tail and head of the distribution, but deviations in the tail have often been unconsidered; avalanche shape collapse has required manual parameter tuning; and the estimation of neural complexity has relied on small data sets or statistical assumptions for the sake of computational efficiency. In this paper we present technical advancements in the analysis of criticality and complexity in neural systems. We use maximum-likelihood estimation to automatically fit power laws with left and right cutoffs, present the first automated shape collapse algorithm, and describe new techniques to account for large numbers of neural variables and small data sets in the calculation of neural complexity. In order to facilitate future research in criticality and complexity, we have made the software utilized in this analysis freely available online in the MATLAB NCC (Neural Complexity and Criticality) Toolbox. PMID:27445842

  15. CosmoSIS: Modular cosmological parameter estimation

    DOE PAGES

    Zuntz, J.; Paterno, M.; Jennings, E.; ...

    2015-06-09

    Cosmological parameter estimation is entering a new era. Large collaborations need to coordinate high-stakes analyses using multiple methods; furthermore such analyses have grown in complexity due to sophisticated models of cosmology and systematic uncertainties. In this paper we argue that modularity is the key to addressing these challenges: calculations should be broken up into interchangeable modular units with inputs and outputs clearly defined. Here we present a new framework for cosmological parameter estimation, CosmoSIS, designed to connect together, share, and advance development of inference tools across the community. We describe the modules already available in CosmoSIS, including CAMB, Planck, cosmicmore » shear calculations, and a suite of samplers. Lastly, we illustrate it using demonstration code that you can run out-of-the-box with the installer available at http://bitbucket.org/joezuntz/cosmosis« less

  16. NASA Software Cost Estimation Model: An Analogy Based Estimation Model

    NASA Technical Reports Server (NTRS)

    Hihn, Jairus; Juster, Leora; Menzies, Tim; Mathew, George; Johnson, James

    2015-01-01

    The cost estimation of software development activities is increasingly critical for large scale integrated projects such as those at DOD and NASA especially as the software systems become larger and more complex. As an example MSL (Mars Scientific Laboratory) developed at the Jet Propulsion Laboratory launched with over 2 million lines of code making it the largest robotic spacecraft ever flown (Based on the size of the software). Software development activities are also notorious for their cost growth, with NASA flight software averaging over 50% cost growth. All across the agency, estimators and analysts are increasingly being tasked to develop reliable cost estimates in support of program planning and execution. While there has been extensive work on improving parametric methods there is very little focus on the use of models based on analogy and clustering algorithms. In this paper we summarize our findings on effort/cost model estimation and model development based on ten years of software effort estimation research using data mining and machine learning methods to develop estimation models based on analogy and clustering. The NASA Software Cost Model performance is evaluated by comparing it to COCOMO II, linear regression, and K-­ nearest neighbor prediction model performance on the same data set.

  17. Volumes and bulk densities of forty asteroids from ADAM shape modeling

    NASA Astrophysics Data System (ADS)

    Hanuš, J.; Viikinkoski, M.; Marchis, F.; Ďurech, J.; Kaasalainen, M.; Delbo', M.; Herald, D.; Frappa, E.; Hayamizu, T.; Kerr, S.; Preston, S.; Timerson, B.; Dunham, D.; Talbot, J.

    2017-05-01

    Context. Disk-integrated photometric data of asteroids do not contain accurate information on shape details or size scale. Additional data such as disk-resolved images or stellar occultation measurements further constrain asteroid shapes and allow size estimates. Aims: We aim to use all the available disk-resolved images of approximately forty asteroids obtained by the Near-InfraRed Camera (Nirc2) mounted on the W.M. Keck II telescope together with the disk-integrated photometry and stellar occultation measurements to determine their volumes. We can then use the volume, in combination with the known mass, to derive the bulk density. Methods: We downloaded and processed all the asteroid disk-resolved images obtained by the Nirc2 that are available in the Keck Observatory Archive (KOA). We combined optical disk-integrated data and stellar occultation profiles with the disk-resolved images and use the All-Data Asteroid Modeling (ADAM) algorithm for the shape and size modeling. Our approach provides constraints on the expected uncertainty in the volume and size as well. Results: We present shape models and volume for 41 asteroids. For 35 of these asteroids, the knowledge of their mass estimates from the literature allowed us to derive their bulk densities. We see a clear trend of lower bulk densities for primitive objects (C-complex) and higher bulk densities for S-complex asteroids. The range of densities in the X-complex is large, suggesting various compositions. We also identified a few objects with rather peculiar bulk densities, which is likely a hint of their poor mass estimates. Asteroid masses determined from the Gaia astrometric observations should further refine most of the density estimates.

  18. An Exploratory Study of Cost Engineering in Axiomatic Design: Creation of the Cost Model Based on an FR-DP Map

    NASA Technical Reports Server (NTRS)

    Lee, Taesik; Jeziorek, Peter

    2004-01-01

    Large complex projects cost large sums of money throughout their life cycle for a variety of reasons and causes. For such large programs, the credible estimation of the project cost, a quick assessment of the cost of making changes, and the management of the project budget with effective cost reduction determine the viability of the project. Cost engineering that deals with these issues requires a rigorous method and systematic processes. This paper introduces a logical framework to a&e effective cost engineering. The framework is built upon Axiomatic Design process. The structure in the Axiomatic Design process provides a good foundation to closely tie engineering design and cost information together. The cost framework presented in this paper is a systematic link between the functional domain (FRs), physical domain (DPs), cost domain (CUs), and a task/process-based model. The FR-DP map relates a system s functional requirements to design solutions across all levels and branches of the decomposition hierarchy. DPs are mapped into CUs, which provides a means to estimate the cost of design solutions - DPs - from the cost of the physical entities in the system - CUs. The task/process model describes the iterative process ot-developing each of the CUs, and is used to estimate the cost of CUs. By linking the four domains, this framework provides a superior traceability from requirements to cost information.

  19. DOE Office of Scientific and Technical Information (OSTI.GOV)

    Eken, T; Mayeda, K; Hofstetter, A

    A recently developed coda magnitude methodology was applied to selected broadband stations in Turkey for the purpose of testing the coda method in a large, laterally complex region. As found in other, albeit smaller regions, coda envelope amplitude measurements are significantly less variable than distance-corrected direct wave measurements (i.e., L{sub g} and surface waves) by roughly a factor 3-to-4. Despite strong lateral crustal heterogeneity in Turkey, we found that the region could be adequately modeled assuming a simple 1-D, radially symmetric path correction for 10 narrow frequency bands ranging between 0.02 to 2.0 Hz. For higher frequencies however, 2-D pathmore » corrections will be necessary and will be the subject of a future study. After calibrating the stations ISP, ISKB, and MALT for local and regional distances, single-station moment-magnitude estimates (M{sub w}) derived from the coda spectra were in excellent agreement with those determined from multi-station waveform modeling inversions of long-period data, exhibiting a data standard deviation of 0.17. Though the calibration was validated using large events, the results of the calibration will extend M{sub w} estimates to significantly smaller events which could not otherwise be waveform modeled due to poor signal-to-noise ratio at long periods and sparse station coverage. The successful application of the method is remarkable considering the significant lateral complexity in Turkey and the simple assumptions used in the coda method.« less

  20. Methodological Considerations in Estimation of Phenotype Heritability Using Genome-Wide SNP Data, Illustrated by an Analysis of the Heritability of Height in a Large Sample of African Ancestry Adults

    PubMed Central

    Chen, Fang; He, Jing; Zhang, Jianqi; Chen, Gary K.; Thomas, Venetta; Ambrosone, Christine B.; Bandera, Elisa V.; Berndt, Sonja I.; Bernstein, Leslie; Blot, William J.; Cai, Qiuyin; Carpten, John; Casey, Graham; Chanock, Stephen J.; Cheng, Iona; Chu, Lisa; Deming, Sandra L.; Driver, W. Ryan; Goodman, Phyllis; Hayes, Richard B.; Hennis, Anselm J. M.; Hsing, Ann W.; Hu, Jennifer J.; Ingles, Sue A.; John, Esther M.; Kittles, Rick A.; Kolb, Suzanne; Leske, M. Cristina; Monroe, Kristine R.; Murphy, Adam; Nemesure, Barbara; Neslund-Dudas, Christine; Nyante, Sarah; Ostrander, Elaine A; Press, Michael F.; Rodriguez-Gil, Jorge L.; Rybicki, Ben A.; Schumacher, Fredrick; Stanford, Janet L.; Signorello, Lisa B.; Strom, Sara S.; Stevens, Victoria; Van Den Berg, David; Wang, Zhaoming; Witte, John S.; Wu, Suh-Yuh; Yamamura, Yuko; Zheng, Wei; Ziegler, Regina G.; Stram, Alexander H.; Kolonel, Laurence N.; Marchand, Loïc Le; Henderson, Brian E.; Haiman, Christopher A.; Stram, Daniel O.

    2015-01-01

    Height has an extremely polygenic pattern of inheritance. Genome-wide association studies (GWAS) have revealed hundreds of common variants that are associated with human height at genome-wide levels of significance. However, only a small fraction of phenotypic variation can be explained by the aggregate of these common variants. In a large study of African-American men and women (n = 14,419), we genotyped and analyzed 966,578 autosomal SNPs across the entire genome using a linear mixed model variance components approach implemented in the program GCTA (Yang et al Nat Genet 2010), and estimated an additive heritability of 44.7% (se: 3.7%) for this phenotype in a sample of evidently unrelated individuals. While this estimated value is similar to that given by Yang et al in their analyses, we remain concerned about two related issues: (1) whether in the complete absence of hidden relatedness, variance components methods have adequate power to estimate heritability when a very large number of SNPs are used in the analysis; and (2) whether estimation of heritability may be biased, in real studies, by low levels of residual hidden relatedness. We addressed the first question in a semi-analytic fashion by directly simulating the distribution of the score statistic for a test of zero heritability with and without low levels of relatedness. The second question was addressed by a very careful comparison of the behavior of estimated heritability for both observed (self-reported) height and simulated phenotypes compared to imputation R2 as a function of the number of SNPs used in the analysis. These simulations help to address the important question about whether today's GWAS SNPs will remain useful for imputing causal variants that are discovered using very large sample sizes in future studies of height, or whether the causal variants themselves will need to be genotyped de novo in order to build a prediction model that ultimately captures a large fraction of the variability of height, and by implication other complex phenotypes. Our overall conclusions are that when study sizes are quite large (5,000 or so) the additive heritability estimate for height is not apparently biased upwards using the linear mixed model; however there is evidence in our simulation that a very large number of causal variants (many thousands) each with very small effect on phenotypic variance will need to be discovered to fill the gap between the heritability explained by known versus unknown causal variants. We conclude that today's GWAS data will remain useful in the future for causal variant prediction, but that finding the causal variants that need to be predicted may be extremely laborious. PMID:26125186

  1. Methodological Considerations in Estimation of Phenotype Heritability Using Genome-Wide SNP Data, Illustrated by an Analysis of the Heritability of Height in a Large Sample of African Ancestry Adults.

    PubMed

    Chen, Fang; He, Jing; Zhang, Jianqi; Chen, Gary K; Thomas, Venetta; Ambrosone, Christine B; Bandera, Elisa V; Berndt, Sonja I; Bernstein, Leslie; Blot, William J; Cai, Qiuyin; Carpten, John; Casey, Graham; Chanock, Stephen J; Cheng, Iona; Chu, Lisa; Deming, Sandra L; Driver, W Ryan; Goodman, Phyllis; Hayes, Richard B; Hennis, Anselm J M; Hsing, Ann W; Hu, Jennifer J; Ingles, Sue A; John, Esther M; Kittles, Rick A; Kolb, Suzanne; Leske, M Cristina; Millikan, Robert C; Monroe, Kristine R; Murphy, Adam; Nemesure, Barbara; Neslund-Dudas, Christine; Nyante, Sarah; Ostrander, Elaine A; Press, Michael F; Rodriguez-Gil, Jorge L; Rybicki, Ben A; Schumacher, Fredrick; Stanford, Janet L; Signorello, Lisa B; Strom, Sara S; Stevens, Victoria; Van Den Berg, David; Wang, Zhaoming; Witte, John S; Wu, Suh-Yuh; Yamamura, Yuko; Zheng, Wei; Ziegler, Regina G; Stram, Alexander H; Kolonel, Laurence N; Le Marchand, Loïc; Henderson, Brian E; Haiman, Christopher A; Stram, Daniel O

    2015-01-01

    Height has an extremely polygenic pattern of inheritance. Genome-wide association studies (GWAS) have revealed hundreds of common variants that are associated with human height at genome-wide levels of significance. However, only a small fraction of phenotypic variation can be explained by the aggregate of these common variants. In a large study of African-American men and women (n = 14,419), we genotyped and analyzed 966,578 autosomal SNPs across the entire genome using a linear mixed model variance components approach implemented in the program GCTA (Yang et al Nat Genet 2010), and estimated an additive heritability of 44.7% (se: 3.7%) for this phenotype in a sample of evidently unrelated individuals. While this estimated value is similar to that given by Yang et al in their analyses, we remain concerned about two related issues: (1) whether in the complete absence of hidden relatedness, variance components methods have adequate power to estimate heritability when a very large number of SNPs are used in the analysis; and (2) whether estimation of heritability may be biased, in real studies, by low levels of residual hidden relatedness. We addressed the first question in a semi-analytic fashion by directly simulating the distribution of the score statistic for a test of zero heritability with and without low levels of relatedness. The second question was addressed by a very careful comparison of the behavior of estimated heritability for both observed (self-reported) height and simulated phenotypes compared to imputation R2 as a function of the number of SNPs used in the analysis. These simulations help to address the important question about whether today's GWAS SNPs will remain useful for imputing causal variants that are discovered using very large sample sizes in future studies of height, or whether the causal variants themselves will need to be genotyped de novo in order to build a prediction model that ultimately captures a large fraction of the variability of height, and by implication other complex phenotypes. Our overall conclusions are that when study sizes are quite large (5,000 or so) the additive heritability estimate for height is not apparently biased upwards using the linear mixed model; however there is evidence in our simulation that a very large number of causal variants (many thousands) each with very small effect on phenotypic variance will need to be discovered to fill the gap between the heritability explained by known versus unknown causal variants. We conclude that today's GWAS data will remain useful in the future for causal variant prediction, but that finding the causal variants that need to be predicted may be extremely laborious.

  2. Fast Poisson noise removal by biorthogonal Haar domain hypothesis testing

    NASA Astrophysics Data System (ADS)

    Zhang, B.; Fadili, M. J.; Starck, J.-L.; Digel, S. W.

    2008-07-01

    Methods based on hypothesis tests (HTs) in the Haar domain are widely used to denoise Poisson count data. Facing large datasets or real-time applications, Haar-based denoisers have to use the decimated transform to meet limited-memory or computation-time constraints. Unfortunately, for regular underlying intensities, decimation yields discontinuous estimates and strong “staircase” artifacts. In this paper, we propose to combine the HT framework with the decimated biorthogonal Haar (Bi-Haar) transform instead of the classical Haar. The Bi-Haar filter bank is normalized such that the p-values of Bi-Haar coefficients (p) provide good approximation to those of Haar (pH) for high-intensity settings or large scales; for low-intensity settings and small scales, we show that p are essentially upper-bounded by pH. Thus, we may apply the Haar-based HTs to Bi-Haar coefficients to control a prefixed false positive rate. By doing so, we benefit from the regular Bi-Haar filter bank to gain a smooth estimate while always maintaining a low computational complexity. A Fisher-approximation-based threshold implementing the HTs is also established. The efficiency of this method is illustrated on an example of hyperspectral-source-flux estimation.

  3. An approximate Kalman filter for ocean data assimilation: An example with an idealized Gulf Stream model

    NASA Technical Reports Server (NTRS)

    Fukumori, Ichiro; Malanotte-Rizzoli, Paola

    1995-01-01

    A practical method of data assimilation for use with large, nonlinear, ocean general circulation models is explored. A Kalman filter based on approximation of the state error covariance matrix is presented, employing a reduction of the effective model dimension, the error's asymptotic steady state limit, and a time-invariant linearization of the dynamic model for the error integration. The approximations lead to dramatic computational savings in applying estimation theory to large complex systems. We examine the utility of the approximate filter in assimilating different measurement types using a twin experiment of an idealized Gulf Stream. A nonlinear primitive equation model of an unstable east-west jet is studied with a state dimension exceeding 170,000 elements. Assimilation of various pseudomeasurements are examined, including velocity, density, and volume transport at localized arrays and realistic distributions of satellite altimetry and acoustic tomography observations. Results are compared in terms of their effects on the accuracies of the estimation. The approximate filter is shown to outperform an empirical nudging scheme used in a previous study. The examples demonstrate that useful approximate estimation errors can be computed in a practical manner for general circulation models.

  4. An approximate Kalman filter for ocean data assimilation: An example with an idealized Gulf Stream model

    NASA Astrophysics Data System (ADS)

    Fukumori, Ichiro; Malanotte-Rizzoli, Paola

    1995-04-01

    A practical method of data assimilation for use with large, nonlinear, ocean general circulation models is explored. A Kaiman filter based on approximations of the state error covariance matrix is presented, employing a reduction of the effective model dimension, the error's asymptotic steady state limit, and a time-invariant linearization of the dynamic model for the error integration. The approximations lead to dramatic computational savings in applying estimation theory to large complex systems. We examine the utility of the approximate filter in assimilating different measurement types using a twin experiment of an idealized Gulf Stream. A nonlinear primitive equation model of an unstable east-west jet is studied with a state dimension exceeding 170,000 elements. Assimilation of various pseudomeasurements are examined, including velocity, density, and volume transport at localized arrays and realistic distributions of satellite altimetry and acoustic tomography observations. Results are compared in terms of their effects on the accuracies of the estimation. The approximate filter is shown to outperform an empirical nudging scheme used in a previous study. The examples demonstrate that useful approximate estimation errors can be computed in a practical manner for general circulation models.

  5. Computational tools for exact conditional logistic regression.

    PubMed

    Corcoran, C; Mehta, C; Patel, N; Senchaudhuri, P

    Logistic regression analyses are often challenged by the inability of unconditional likelihood-based approximations to yield consistent, valid estimates and p-values for model parameters. This can be due to sparseness or separability in the data. Conditional logistic regression, though useful in such situations, can also be computationally unfeasible when the sample size or number of explanatory covariates is large. We review recent developments that allow efficient approximate conditional inference, including Monte Carlo sampling and saddlepoint approximations. We demonstrate through real examples that these methods enable the analysis of significantly larger and more complex data sets. We find in this investigation that for these moderately large data sets Monte Carlo seems a better alternative, as it provides unbiased estimates of the exact results and can be executed in less CPU time than can the single saddlepoint approximation. Moreover, the double saddlepoint approximation, while computationally the easiest to obtain, offers little practical advantage. It produces unreliable results and cannot be computed when a maximum likelihood solution does not exist. Copyright 2001 John Wiley & Sons, Ltd.

  6. An Approach to Experimental Design for the Computer Analysis of Complex Phenomenon

    NASA Technical Reports Server (NTRS)

    Rutherford, Brian

    2000-01-01

    The ability to make credible system assessments, predictions and design decisions related to engineered systems and other complex phenomenon is key to a successful program for many large-scale investigations in government and industry. Recently, many of these large-scale analyses have turned to computational simulation to provide much of the required information. Addressing specific goals in the computer analysis of these complex phenomenon is often accomplished through the use of performance measures that are based on system response models. The response models are constructed using computer-generated responses together with physical test results where possible. They are often based on probabilistically defined inputs and generally require estimation of a set of response modeling parameters. As a consequence, the performance measures are themselves distributed quantities reflecting these variabilities and uncertainties. Uncertainty in the values of the performance measures leads to uncertainties in predicted performance and can cloud the decisions required of the analysis. A specific goal of this research has been to develop methodology that will reduce this uncertainty in an analysis environment where limited resources and system complexity together restrict the number of simulations that can be performed. An approach has been developed that is based on evaluation of the potential information provided for each "intelligently selected" candidate set of computer runs. Each candidate is evaluated by partitioning the performance measure uncertainty into two components - one component that could be explained through the additional computational simulation runs and a second that would remain uncertain. The portion explained is estimated using a probabilistic evaluation of likely results for the additional computational analyses based on what is currently known about the system. The set of runs indicating the largest potential reduction in uncertainty is then selected and the computational simulations are performed. Examples are provided to demonstrate this approach on small scale problems. These examples give encouraging results. Directions for further research are indicated.

  7. Processing and comparison of two weighing lysimeters at the Rietholzbach catchment

    NASA Astrophysics Data System (ADS)

    Ruth, Conall; Michel, Dominik; Hirschi, Martin; Seneviratne, Sonia I.

    2017-04-01

    Weighing lysimeters are a well-established means of accurately obtaining local-scale estimates of actual evapotranspiration and seepage within soils. Current state-of-the-art devices have very high temporal resolutions and weighing precisions, and can also be used to estimate precipitation. These, however, require complex filtering to first remove noise (e.g. resulting from wind influence) from the mass measurements. At the Rietholzbach research catchment in northeastern Switzerland, two weighing lysimeters are in operation. One is a recently-installed state-of-the-art mini-lysimeter with a pump-controlled lower boundary; the other is a large free-drainage lysimeter in operation since 1976. To determine the optimal processing approach for the mini-lysimeter, a number of reported approaches were applied, with the resulting evapotranspiration and precipitation records being compared to those of the large lysimeter and a tipping bucket, respectively. Out of those examined, we found the Adaptive-Window and Adaptive-Threshold (AWAT) filter and a similar, non-adaptive approach, to perform best. Using the AWAT-filtered mini-lysimeter data as a reference, additional, retrospectively-applicable processing steps for the large lysimeter were then investigated. Those found to be most beneficial were the application of a three-point (10-min) moving mean to the mass measurements, and the setting-to-zero of estimated evapotranspiration and condensation in hours with greater-than-zero reference tipping bucket precipitation recordings. A comparison of lysimeter mass increases associated with precipitation revealed that the large lysimeter experiences a previously unknown under-catch of 11.1% (for liquid precipitation). Daily seepage measurements were found to be generally greater from the mini-lysimeter, probably reflecting the reduced input of water to the large lysimeter due to this under-catch.

  8. Optical Remote Sensing Algorithm Validation using High-Frequency Underway Biogeochemical Measurements in Three Large Global River Systems

    NASA Astrophysics Data System (ADS)

    Kuhn, C.; Richey, J. E.; Striegl, R. G.; Ward, N.; Sawakuchi, H. O.; Crawford, J.; Loken, L. C.; Stadler, P.; Dornblaser, M.; Butman, D. E.

    2017-12-01

    More than 93% of the world's river-water volume occurs in basins impacted by large dams and about 43% of river water discharge is impacted by flow regulation. Human land use also alters nutrient and carbon cycling and the emission of carbon dioxide from inland reservoirs. Increased water residence times and warmer temperatures in reservoirs fundamentally alter the physical settings for biogeochemical processing in large rivers, yet river biogeochemistry for many large systems remains undersampled. Satellite remote sensing holds promise as a methodology for responsive regional and global water resources management. Decades of ocean optics research has laid the foundation for the use of remote sensing reflectance in optical wavelengths (400 - 700 nm) to produce satellite-derived, near-surface estimates of phytoplankton chlorophyll concentration. Significant improvements between successive generations of ocean color sensors have enabled the scientific community to document changes in global ocean productivity (NPP) and estimate ocean biomass with increasing accuracy. Despite large advances in ocean optics, application of optical methods to inland waters has been limited to date due to their optical complexity and small spatial scale. To test this frontier, we present a study evaluating the accuracy and suitability of empirical inversion approaches for estimating chlorophyll-a, turbidity and temperature for the Amazon, Columbia and Mississippi rivers using satellite remote sensing. We demonstrate how riverine biogeochemical measurements collected at high frequencies from underway vessels can be used as in situ matchups to evaluate remotely-sensed, near-surface temperature, turbidity, chlorophyll-a derived from the Landsat 8 (NASA) and Sentinel 2 (ESA) satellites. We investigate the use of remote sensing water reflectance to infer trophic status as well as tributary influences on the optical characteristics of the Amazon, Mississippi and Columbia rivers.

  9. Host-guest complexes between cucurbit[n]urils and acetanilides having aminopropyl units.

    PubMed

    Buaki-Sogo, Mireia; Montes-Navajas, Pedro; Alvaro, Mercedes; Garcia, Hermenegildo

    2013-06-01

    2-(Propylamino)acetamide of aniline (1a), and bis-2-(propylamino)acetamide of ortho- (1b) and para-(1c) phenylenediamine form host-guest complexes with CB[6], CB[7] and CB[8] as evidenced by the variations in the (1)H NMR spectroscopy chemical shifts and observation in MALDI-TOF-MS and ESI-MS of ions at the corresponding mass. Binding constants for the 1:1 complexes were estimated from fluorescence titrations and were in the range 10(5)-10(6)M(-1). Models based on molecular mechanics for these supramolecular complexes are provided. In spite of the different geometries arising from the ortho- or para-substitution, phenylenediamides form complexes of similar strength in which the hydrophobic alkyl chains are accommodated inside the host cavity. Formation of these host-guest complexes in the solid state was also achieved by modifying an aminopropyl silica with chloroacetanilides and preparing three silica having analogues of compounds 1a-c anchored to the solid particles. Titrations showed, however, that these solids can adsorb a large percentage of CBs by unselective interactions that are not related to the formation of inclusion complexes. Copyright © 2013 Elsevier Inc. All rights reserved.

  10. Assessment of actual evapotranspiration over a semiarid heterogeneous land surface by means of coupled low-resolution remote sensing data with an energy balance model: comparison to extra-large aperture scintillometer measurements

    NASA Astrophysics Data System (ADS)

    Saadi, Sameh; Boulet, Gilles; Bahir, Malik; Brut, Aurore; Delogu, Émilie; Fanise, Pascal; Mougenot, Bernard; Simonneaux, Vincent; Lili Chabaane, Zohra

    2018-04-01

    In semiarid areas, agricultural production is restricted by water availability; hence, efficient agricultural water management is a major issue. The design of tools providing regional estimates of evapotranspiration (ET), one of the most relevant water balance fluxes, may help the sustainable management of water resources. Remote sensing provides periodic data about actual vegetation temporal dynamics (through the normalized difference vegetation index, NDVI) and water availability under water stress (through the surface temperature Tsurf), which are crucial factors controlling ET. In this study, spatially distributed estimates of ET (or its energy equivalent, the latent heat flux LE) in the Kairouan plain (central Tunisia) were computed by applying the Soil Plant Atmosphere and Remote Sensing Evapotranspiration (SPARSE) model fed by low-resolution remote sensing data (Terra and Aqua MODIS). The work's goal was to assess the operational use of the SPARSE model and the accuracy of the modeled (i) sensible heat flux (H) and (ii) daily ET over a heterogeneous semiarid landscape with complex land cover (i.e., trees, winter cereals, summer vegetables). SPARSE was run to compute instantaneous estimates of H and LE fluxes at the satellite overpass times. The good correspondence (R2 = 0.60 and 0.63 and RMSE = 57.89 and 53.85 W m-2 for Terra and Aqua, respectively) between instantaneous H estimates and large aperture scintillometer (XLAS) H measurements along a path length of 4 km over the study area showed that the SPARSE model presents satisfactory accuracy. Results showed that, despite the fairly large scatter, the instantaneous LE can be suitably estimated at large scales (RMSE = 47.20 and 43.20 W m-2 for Terra and Aqua, respectively, and R2 = 0.55 for both satellites). Additionally, water stress was investigated by comparing modeled (SPARSE) and observed (XLAS) water stress values; we found that most points were located within a 0.2 confidence interval, thus the general tendencies are well reproduced. Even though extrapolation of instantaneous latent heat flux values to daily totals was less obvious, daily ET estimates are deemed acceptable.

  11. State estimation for autopilot control of small unmanned aerial vehicles in windy conditions

    NASA Astrophysics Data System (ADS)

    Poorman, David Paul

    The use of small unmanned aerial vehicles (UAVs) both in the military and civil realms is growing. This is largely due to the proliferation of inexpensive sensors and the increase in capability of small computers that has stemmed from the personal electronic device market. Methods for performing accurate state estimation for large scale aircraft have been well known and understood for decades, which usually involve a complex array of expensive high accuracy sensors. Performing accurate state estimation for small unmanned aircraft is a newer area of study and often involves adapting known state estimation methods to small UAVs. State estimation for small UAVs can be more difficult than state estimation for larger UAVs due to small UAVs employing limited sensor suites due to cost, and the fact that small UAVs are more susceptible to wind than large aircraft. The purpose of this research is to evaluate the ability of existing methods of state estimation for small UAVs to accurately capture the states of the aircraft that are necessary for autopilot control of the aircraft in a Dryden wind field. The research begins by showing which aircraft states are necessary for autopilot control in Dryden wind. Then two state estimation methods that employ only accelerometer, gyro, and GPS measurements are introduced. The first method uses assumptions on aircraft motion to directly solve for attitude information and smooth GPS data, while the second method integrates sensor data to propagate estimates between GPS measurements and then corrects those estimates with GPS information. The performance of both methods is analyzed with and without Dryden wind, in straight and level flight, in a coordinated turn, and in a wings level ascent. It is shown that in zero wind, the first method produces significant steady state attitude errors in both a coordinated turn and in a wings level ascent. In Dryden wind, it produces large noise on the estimates for its attitude states, and has a non-zero mean error that increases when gyro bias is increased. The second method is shown to not exhibit any steady state error in the tested scenarios that is inherent to its design. The second method can correct for attitude errors that arise from both integration error and gyro bias states, but it suffers from lack of attitude error observability. The attitude errors are shown to be more observable in wind, but increased integration error in wind outweighs the increase in attitude corrections that such increased observability brings, resulting in larger attitude errors in wind. Overall, this work highlights many technical deficiencies of both of these methods of state estimation that could be improved upon in the future to enhance state estimation for small UAVs in windy conditions.

  12. Site Transfer Functions of Three-Component Ground Motion in Western Turkey

    NASA Astrophysics Data System (ADS)

    Ozgur Kurtulmus, Tevfik; Akyol, Nihal; Camyildiz, Murat; Gungor, Talip

    2015-04-01

    Because of high seismicity accommodating crustal deformation and deep graben structures, on which have, urbanized and industrialized large cities in western Turkey, the importance of site-specific seismic hazard assessments becomes more crucial. Characterizing source, site and path effects is important for both assessing the seismic hazard in a specific region and generation of the building codes/or renewing previous ones. In this study, we evaluated three-component recordings for micro- and moderate-size earthquakes with local magnitudes ranging between 2.0 and 5.6. This dataset is used for site transfer function estimations, utilizing two different spectral ratio approaches 'Standard Spectral Ratio-(SSR)' and 'Horizontal to Vertical Spectral Ratio-(HVSR)' and a 'Generalized Inversion Technique-(GIT)' to highlight site-specific seismic hazard potential of deep basin structures of the region. Obtained transfer functions revealed that the sites located near the basin edges are characterized by broader HVSR curves. Broad HVSR peaks could be attributed to the complexity of wave propagation related to significant 2D/3D velocity variations at the sediment-bedrock interface near the basin edges. Comparison of HVSR and SSR estimates for the sites located on the grabens showed that SSR estimates give larger values at lower frequencies which could be attributed to lateral variations in regional velocity and attenuation values caused by basin geometry and edge effects. However, large amplitude values of vertical component GIT site transfer functions were observed at varying frequency ranges for some of the stations. These results imply that vertical component of ground motion is not amplification free. Contamination of HVSR site transfer function estimates at different frequency bands could be related to complexities in the wave field caused by deep or shallow heterogeneities in the region such as differences in the basin geometries, fracturing and fluid saturation along different propagation paths. The results also show that, even if the site is located on a horst, the presence of weathered zones near the surface could cause moderate frequency dependent site effects.

  13. Predicting the effect of deep-rooted hybrid poplars on the groundwater flow system at a large-scale phytoremediation site.

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Quinn, J. J.; Negri, M. C.; Hinchman, R. R.

    2001-03-01

    Estimating the effect of phreatophytes on the groundwater flow field is critical in the design or evaluation of a phytoremediation system. Complex hydrogeological conditions and the transient water use rates of trees require the application of numerical modeling to address such issues as hydraulic containment, seasonality, and system design. In 1999, 809 hybrid poplars and willows were planted to phytoremediate the 317 and 319 Areas of Argonne National Laboratory near Chicago, Illinois. Contaminants of concern are volatile organic compounds and tritium. The site hydrogeology is a complex framework of glacial tills interlaced with sands, gravels, and silts of varying character,more » thickness, and lateral extent. A total of 420 poplars were installed using a technology to direct the roots through a 25-ft (8-m)-thick till to a contaminated aquifer. Numerical modeling was used to simulate the effect of the deep-rooted poplars on this aquifer of concern. Initially, the best estimates of input parameters and boundary conditions were determined to provide a suitable match to historical transient ground-water flow conditions. The model was applied to calculate the future effect of the developing deep-rooted poplars over a 6 year period. The first 3 years represent the development period of the trees. In the fourth year, canopy closure is expected to occur; modeling continues through the first 3 years of the mature plantation. Monthly estimates of water use by the trees are incorporated. The modeling suggested that the mature trees in the plantation design will provide a large degree of containment of groundwater from the upgradient source areas, despite the seasonal nature of the trees' water consumption. The results indicate the likely areas where seasonal dewatering of the aquifer may limit the availability of water for the trees. The modeling also provided estimates of the residence time of groundwater in the geochemically altered rhizosphere of the plantation.« less

  14. Spatial Investigation of Columnar AOD and Near-Surface PM2.5 Concentrations During the 2013 American and Yosemite Rim Fires

    NASA Astrophysics Data System (ADS)

    Loria Salazar, S. M.; Holmes, H.; Arnott, W. P.; Moosmuller, H.; Liming, A.; Echevarria, B.

    2014-12-01

    The study of aerosol pollution transport and optical properties in the western U.S. is a challenge due to the complex terrain, bright surfaces, presence of anthropogenic and biogenic emissions, secondary organic aerosol formation, and smoke from wild fires. In addition, the complex terrain influences transport phenomena by recirculating mountain air from California to Nevada, where air pollution from the Sierra Nevada Mountains (SNM) is mixed with urban air from the Central Valley in California. Previous studies in Reno hypothesize that elevated aerosol concentrations aloft, above the convective boundary layer height, make air quality monitoring in Reno challenging with MODIS products. Here, we analyze data from August 2013 as a case study for wildfire smoke plumes in California and Nevada. During this time period, northern California was impacted by large wild fires known as the American and Yosemite Rim fires. Thousands of acres burned, generating large quantities of aerosol pollutants that were transported downwind. The aim of the present work is to investigate the fire plume behavior and transport phenomena using ground level PM2.5 concentrations from routine monitoring networks and aerosol optical properties from AERONET, both at multiple locations in California and Nevada. In addition, the accuracy of MODIS (Collection 6) and VIIRS aerosol satellite products will be evaluated. The multispectral photoacoustic instruments and reciprocal nephelometers located in Reno support the estimation of approximated aerosol height. The objectives are to investigate the impact of the vertical distribution of PM concentrations on satellite aerosol optical depth (AOD) retrievals; assess the ability to estimate ground level PM2.5 mass concentrations for wildfire smoke plumes from satellite remote sensing; and investigate the influence of complex terrain on the transport of pollutants, convective boundary layer depth, and aerosol optical height.

  15. The scaling of complex craters

    NASA Technical Reports Server (NTRS)

    Croft, S. K.

    1985-01-01

    The empirical relation between the transient crater diameter (Dg) and final crater diameter (Dr) of complex craters and basins is estimated using cumulative terrace widths, central uplift diameters, continuous ejecta radii, and transient crater reconstructions determined from lunar and terrestrial impact structures. The ratio Dg/Dr is a power law function of Dr, decreasing uniformly from unity at the diameter of the simple-complex crater morphology transition to about 0.5 for large multiring basins like Imbrium on the moon. The empirical constants in the Dg/Dr relation are interpreted physically to mean that the position of the final rim relative to the transient crater, and hence the extent of collapse, is controlled or greatly influenced by the properties of the zone of dissociated material produced by the impact shock. The continuity of the Dg/Dr relation over the entire spectrum of morphologic types from complex craters to multiring basins implies that the rims of all these structures form in the same tectonic environment despite morphologic differences.

  16. Development of improved wildfire smoke exposure estimates for health studies in the western U.S.

    NASA Astrophysics Data System (ADS)

    Ivey, C.; Holmes, H.; Loria Salazar, S. M.; Pierce, A.; Liu, C.

    2016-12-01

    Wildfire smoke exposure is a significant health concern in the western U.S. because large wildfires have increased in size and frequency over the past four years due to drought conditions. The transport phenomena in complex terrain and timing of the wildfire emissions make the smoke plumes difficult to simulate using conventional air quality models. Monitoring data can be used to estimate exposure metrics, but in rural areas the monitoring networks are too sparse to calculate wildfire exposure metrics for the entire population in a region. Satellite retrievals provide global, spatiotemporal air quality information and are used to track pollution plumes, estimate human exposures, model emissions, and determine sources (i.e., natural versus anthropogenic) in regulatory applications. Particulate matter (PM) exposures can be estimated using columnar aerosol optical depth (AOD), where satellite AOD retrievals serve as a spatial surrogate to simulate surface PM gradients. These exposure models have been successfully used in health effects studies in the eastern U.S. where complex mountainous terrain and surface reflectance do not limit AOD retrival from satellites. Using results from a chemical transport model (CTM) is another effective method to determine spatial gradients of pollutants. However, the CTM does not adequately capture the temporal and spatial distribution of wildfire smoke plumes. By combining the spatiotemporal pollutant fields from both satellite retrievals and CTM results with ground based pollutant observations the spatial wildfire smoke exposure model can be improved. This work will address the challenge of understanding the spatiotemporal distributions of pollutant concentrations to model human exposures of wildfire smoke in regions with complex terrain, where meteorological conditions as well as emission sources significantly influence the spatial distribution of pollutants. The focus will be on developing models to enhance exposure estimates of elevated PM and ozone concentrations from wildfire smoke plumes in the western U.S.

  17. State-space models’ dirty little secrets: even simple linear Gaussian models can have estimation problems

    NASA Astrophysics Data System (ADS)

    Auger-Méthé, Marie; Field, Chris; Albertsen, Christoffer M.; Derocher, Andrew E.; Lewis, Mark A.; Jonsen, Ian D.; Mills Flemming, Joanna

    2016-05-01

    State-space models (SSMs) are increasingly used in ecology to model time-series such as animal movement paths and population dynamics. This type of hierarchical model is often structured to account for two levels of variability: biological stochasticity and measurement error. SSMs are flexible. They can model linear and nonlinear processes using a variety of statistical distributions. Recent ecological SSMs are often complex, with a large number of parameters to estimate. Through a simulation study, we show that even simple linear Gaussian SSMs can suffer from parameter- and state-estimation problems. We demonstrate that these problems occur primarily when measurement error is larger than biological stochasticity, the condition that often drives ecologists to use SSMs. Using an animal movement example, we show how these estimation problems can affect ecological inference. Biased parameter estimates of a SSM describing the movement of polar bears (Ursus maritimus) result in overestimating their energy expenditure. We suggest potential solutions, but show that it often remains difficult to estimate parameters. While SSMs are powerful tools, they can give misleading results and we urge ecologists to assess whether the parameters can be estimated accurately before drawing ecological conclusions from their results.

  18. Myocardial motion estimation of tagged cardiac magnetic resonance images using tag motion constraints and multi-level b-splines interpolation.

    PubMed

    Liu, Hong; Yan, Meng; Song, Enmin; Wang, Jie; Wang, Qian; Jin, Renchao; Jin, Lianghai; Hung, Chih-Cheng

    2016-05-01

    Myocardial motion estimation of tagged cardiac magnetic resonance (TCMR) images is of great significance in clinical diagnosis and the treatment of heart disease. Currently, the harmonic phase analysis method (HARP) and the local sine-wave modeling method (SinMod) have been proven as two state-of-the-art motion estimation methods for TCMR images, since they can directly obtain the inter-frame motion displacement vector field (MDVF) with high accuracy and fast speed. By comparison, SinMod has better performance over HARP in terms of displacement detection, noise and artifacts reduction. However, the SinMod method has some drawbacks: 1) it is unable to estimate local displacements larger than half of the tag spacing; 2) it has observable errors in tracking of tag motion; and 3) the estimated MDVF usually has large local errors. To overcome these problems, we present a novel motion estimation method in this study. The proposed method tracks the motion of tags and then estimates the dense MDVF by using the interpolation. In this new method, a parameter estimation procedure for global motion is applied to match tag intersections between different frames, ensuring specific kinds of large displacements being correctly estimated. In addition, a strategy of tag motion constraints is applied to eliminate most of errors produced by inter-frame tracking of tags and the multi-level b-splines approximation algorithm is utilized, so as to enhance the local continuity and accuracy of the final MDVF. In the estimation of the motion displacement, our proposed method can obtain a more accurate MDVF compared with the SinMod method and our method can overcome the drawbacks of the SinMod method. However, the motion estimation accuracy of our method depends on the accuracy of tag lines detection and our method has a higher time complexity. Copyright © 2015 Elsevier Inc. All rights reserved.

  19. Habitat Complexity Metrics to Guide Restoration of Large Rivers

    NASA Astrophysics Data System (ADS)

    Jacobson, R. B.; McElroy, B. J.; Elliott, C.; DeLonay, A.

    2011-12-01

    Restoration strategies on large, channelized rivers typically strive to recover lost habitat complexity, based on the assumption complexity and biophysical capacity are directly related. Although definition of links between complexity and biotic responses can be tenuous, complexity metrics have appeal because of their potential utility in quantifying habitat quality, defining reference conditions and design criteria, and measuring restoration progress. Hydroacoustic instruments provide many ways to measure complexity on large rivers, yet substantive questions remain about variables and scale of complexity that are meaningful to biota, and how complexity can be measured and monitored cost effectively. We explore these issues on the Missouri River, using the example of channel re-engineering projects that are intended to aid in recovery of the pallid sturgeon, an endangered benthic fish. We are refining understanding of what habitat complexity means for adult fish by combining hydroacoustic habitat assessments with acoustic telemetry to map locations during reproductive migrations and spawning. These data indicate that migrating sturgeon select points with relatively low velocity but adjacent to areas of high velocity (that is, with high velocity gradients); the integration of points defines pathways which minimize energy expenditures during upstream migrations of 10's to 100's of km. Complexity metrics that efficiently quantify migration potential at the reach scale are therefore directly relevant to channel restoration strategies. We are also exploring complexity as it relates to larval sturgeon dispersal. Larvae may drift for as many as 17 days (100's of km at mean velocities) before using up their yolk sac, after which they "settle" into habitats where they initiate feeding. An assumption underlying channel re-engineering is that additional channel complexity, specifically increased shallow, slow water, is necessary for early feeding and refugia. Development of complexity metrics is complicated by the fact that characteristics of channel morphology may increase complexity scores without necessarily increasing biophysical capacity for target species. For example, a cross section that samples depths and velocities across the thalweg (navigation channel) and into lentic habitat may score high on most measures of hydraulic or geomorphic complexity, but does not necessarily provide habitats beneficial to native species. Complexity measures need to be bounded by best estimates of native species requirements. In the absence of specific information, creation of habitat complexity for the sake of complexity may lead to unintended consequences, for example, lentic habitats that increase a complexity score but support invasive species. An additional practical constraint on complexity measures is the need to develop metrics that are can be deployed cost-effectively in an operational monitoring program. Design of a monitoring program requires informed choices of measurement variables, definition of reference sites, and design of sampling effort to capture spatial and temporal variability.

  20. Top-down Estimates of Greenhouse Gas Intensities and Emissions for Individual Oil Sands Facilities in Alberta Canada

    NASA Astrophysics Data System (ADS)

    Liggio, J.; Li, S. M.; Staebler, R. M.; Hayden, K. L.; Mittermeier, R. L.; McLaren, R.; Baray, S.; Darlington, A.; Worthy, D.; O'Brien, J.

    2017-12-01

    The oil sands (OS) region of Alberta contributes approximately 10% to Canada's overall anthropogenic greenhouse gas (GHG) emissions. Such emissions have traditionally been estimated through "bottom-up" methods which seek to account for all individual sources of GHGs within a given facility. However, it is recognized that bottom-up approaches for complex industrial facilities can be subject to uncertainties associated with incomplete or inaccurate emission factor and/or activity data. In order to quantify air pollutant emissions from oil sands activities an aircraft-based measurement campaign was performed in the summer of 2013. The aircraft measurements could also be used to quantify GHG emissions for comparison to the bottom up emissions estimates. Utilizing specific flight patterns, together with an emissions estimation algorithm and measurements of CO2 and methane, a "top-down" estimate of GHG intensities for several large surface mining operations was obtained. The results demonstrate that there is a wide variation in emissions intensities (≈80 - 220 kg CO2/barrel oil) across OS facilities, which in some cases agree with calculated intensities, and in other cases are larger than that estimated using industry reported GHG emission and oil production data. When translated to annual GHG emissions, the "top-down" approach results in a CO2 emission of approximately 41 Mega Tonnes (MT) CO2/year for the 4 OS facilities investigated, in contrast to the ≈26 MT CO2/year reported by industry. The results presented here highlight the importance of using "top-down" approaches as a complimentary method in evaluating GHG emissions from large industrial sources.

  1. Estimating psychiatric manpower requirements based on patients' needs.

    PubMed

    Faulkner, L R; Goldman, C R

    1997-05-01

    To provide a better understanding of the complexities of estimating psychiatric manpower requirements, the authors describe several approaches to estimation and present a method based on patients' needs. A five-step method for psychiatric manpower estimation is used, with estimates of data pertinent to each step, to calculate the total psychiatric manpower requirements for the United States. The method is also used to estimate the hours of psychiatric service per patient per year that might be available under current psychiatric practice and under a managed care scenario. Depending on assumptions about data at each step in the method, the total psychiatric manpower requirements for the U.S. population range from 2,989 to 358,696 full-time-equivalent psychiatrists. The number of available hours of psychiatric service per patient per year is 14.1 hours under current psychiatric practice and 2.8 hours under the managed care scenario. The key to psychiatric manpower estimation lies in clarifying the assumptions that underlie the specific method used. Even small differences in assumptions mean large differences in estimates. Any credible manpower estimation process must include discussions and negotiations between psychiatrists, other clinicians, administrators, and patients and families to clarify the treatment needs of patients and the roles, responsibilities, and job description of psychiatrists.

  2. A Corticothalamic Circuit Model for Sound Identification in Complex Scenes

    PubMed Central

    Otazu, Gonzalo H.; Leibold, Christian

    2011-01-01

    The identification of the sound sources present in the environment is essential for the survival of many animals. However, these sounds are not presented in isolation, as natural scenes consist of a superposition of sounds originating from multiple sources. The identification of a source under these circumstances is a complex computational problem that is readily solved by most animals. We present a model of the thalamocortical circuit that performs level-invariant recognition of auditory objects in complex auditory scenes. The circuit identifies the objects present from a large dictionary of possible elements and operates reliably for real sound signals with multiple concurrently active sources. The key model assumption is that the activities of some cortical neurons encode the difference between the observed signal and an internal estimate. Reanalysis of awake auditory cortex recordings revealed neurons with patterns of activity corresponding to such an error signal. PMID:21931668

  3. A Sparse Reconstruction Approach for Identifying Gene Regulatory Networks Using Steady-State Experiment Data

    PubMed Central

    Zhang, Wanhong; Zhou, Tong

    2015-01-01

    Motivation Identifying gene regulatory networks (GRNs) which consist of a large number of interacting units has become a problem of paramount importance in systems biology. Situations exist extensively in which causal interacting relationships among these units are required to be reconstructed from measured expression data and other a priori information. Though numerous classical methods have been developed to unravel the interactions of GRNs, these methods either have higher computing complexities or have lower estimation accuracies. Note that great similarities exist between identification of genes that directly regulate a specific gene and a sparse vector reconstruction, which often relates to the determination of the number, location and magnitude of nonzero entries of an unknown vector by solving an underdetermined system of linear equations y = Φx. Based on these similarities, we propose a novel framework of sparse reconstruction to identify the structure of a GRN, so as to increase accuracy of causal regulation estimations, as well as to reduce their computational complexity. Results In this paper, a sparse reconstruction framework is proposed on basis of steady-state experiment data to identify GRN structure. Different from traditional methods, this approach is adopted which is well suitable for a large-scale underdetermined problem in inferring a sparse vector. We investigate how to combine the noisy steady-state experiment data and a sparse reconstruction algorithm to identify causal relationships. Efficiency of this method is tested by an artificial linear network, a mitogen-activated protein kinase (MAPK) pathway network and the in silico networks of the DREAM challenges. The performance of the suggested approach is compared with two state-of-the-art algorithms, the widely adopted total least-squares (TLS) method and those available results on the DREAM project. Actual results show that, with a lower computational cost, the proposed method can significantly enhance estimation accuracy and greatly reduce false positive and negative errors. Furthermore, numerical calculations demonstrate that the proposed algorithm may have faster convergence speed and smaller fluctuation than other methods when either estimate error or estimate bias is considered. PMID:26207991

  4. A hierarchical estimator development for estimation of tire-road friction coefficient

    PubMed Central

    Zhang, Xudong; Göhlich, Dietmar

    2017-01-01

    The effect of vehicle active safety systems is subject to the friction force arising from the contact of tires and the road surface. Therefore, an adequate knowledge of the tire-road friction coefficient is of great importance to achieve a good performance of these control systems. This paper presents a tire-road friction coefficient estimation method for an advanced vehicle configuration, four-motorized-wheel electric vehicles, in which the longitudinal tire force is easily obtained. A hierarchical structure is adopted for the proposed estimation design. An upper estimator is developed based on unscented Kalman filter to estimate vehicle state information, while a hybrid estimation method is applied as the lower estimator to identify the tire-road friction coefficient using general regression neural network (GRNN) and Bayes' theorem. GRNN aims at detecting road friction coefficient under small excitations, which are the most common situations in daily driving. GRNN is able to accurately create a mapping from input parameters to the friction coefficient, avoiding storing an entire complex tire model. As for large excitations, the estimation algorithm is based on Bayes' theorem and a simplified “magic formula” tire model. The integrated estimation method is established by the combination of the above-mentioned estimators. Finally, the simulations based on a high-fidelity CarSim vehicle model are carried out on different road surfaces and driving maneuvers to verify the effectiveness of the proposed estimation method. PMID:28178332

  5. A hierarchical estimator development for estimation of tire-road friction coefficient.

    PubMed

    Zhang, Xudong; Göhlich, Dietmar

    2017-01-01

    The effect of vehicle active safety systems is subject to the friction force arising from the contact of tires and the road surface. Therefore, an adequate knowledge of the tire-road friction coefficient is of great importance to achieve a good performance of these control systems. This paper presents a tire-road friction coefficient estimation method for an advanced vehicle configuration, four-motorized-wheel electric vehicles, in which the longitudinal tire force is easily obtained. A hierarchical structure is adopted for the proposed estimation design. An upper estimator is developed based on unscented Kalman filter to estimate vehicle state information, while a hybrid estimation method is applied as the lower estimator to identify the tire-road friction coefficient using general regression neural network (GRNN) and Bayes' theorem. GRNN aims at detecting road friction coefficient under small excitations, which are the most common situations in daily driving. GRNN is able to accurately create a mapping from input parameters to the friction coefficient, avoiding storing an entire complex tire model. As for large excitations, the estimation algorithm is based on Bayes' theorem and a simplified "magic formula" tire model. The integrated estimation method is established by the combination of the above-mentioned estimators. Finally, the simulations based on a high-fidelity CarSim vehicle model are carried out on different road surfaces and driving maneuvers to verify the effectiveness of the proposed estimation method.

  6. Bias and robustness of uncertainty components estimates in transient climate projections

    NASA Astrophysics Data System (ADS)

    Hingray, Benoit; Blanchet, Juliette; Jean-Philippe, Vidal

    2016-04-01

    A critical issue in climate change studies is the estimation of uncertainties in projections along with the contribution of the different uncertainty sources, including scenario uncertainty, the different components of model uncertainty and internal variability. Quantifying the different uncertainty sources faces actually different problems. For instance and for the sake of simplicity, an estimate of model uncertainty is classically obtained from the empirical variance of the climate responses obtained for the different modeling chains. These estimates are however biased. Another difficulty arises from the limited number of members that are classically available for most modeling chains. In this case, the climate response of one given chain and the effect of its internal variability may be actually difficult if not impossible to separate. The estimate of scenario uncertainty, model uncertainty and internal variability components are thus likely to be not really robust. We explore the importance of the bias and the robustness of the estimates for two classical Analysis of Variance (ANOVA) approaches: a Single Time approach (STANOVA), based on the only data available for the considered projection lead time and a time series based approach (QEANOVA), which assumes quasi-ergodicity of climate outputs over the whole available climate simulation period (Hingray and Saïd, 2014). We explore both issues for a simple but classical configuration where uncertainties in projections are composed of two single sources: model uncertainty and internal climate variability. The bias in model uncertainty estimates is explored from theoretical expressions of unbiased estimators developed for both ANOVA approaches. The robustness of uncertainty estimates is explored for multiple synthetic ensembles of time series projections generated with MonteCarlo simulations. For both ANOVA approaches, when the empirical variance of climate responses is used to estimate model uncertainty, the bias is always positive. It can be especially high with STANOVA. In the most critical configurations, when the number of members available for each modeling chain is small (< 3) and when internal variability explains most of total uncertainty variance (75% or more), the overestimation is higher than 100% of the true model uncertainty variance. The bias can be considerably reduced with a time series ANOVA approach, owing to the multiple time steps accounted for. The longer the transient time period used for the analysis, the larger the reduction. When a quasi-ergodic ANOVA approach is applied to decadal data for the whole 1980-2100 period, the bias is reduced by a factor 2.5 to 20 depending on the projection lead time. In all cases, the bias is likely to be not negligible for a large number of climate impact studies resulting in a likely large overestimation of the contribution of model uncertainty to total variance. For both approaches, the robustness of all uncertainty estimates is higher when more members are available, when internal variability is smaller and/or the response-to-uncertainty ratio is higher. QEANOVA estimates are much more robust than STANOVA ones: QEANOVA simulated confidence intervals are roughly 3 to 5 times smaller than STANOVA ones. Excepted for STANOVA when less than 3 members is available, the robustness is rather high for total uncertainty and moderate for internal variability estimates. For model uncertainty or response-to-uncertainty ratio estimates, the robustness is conversely low for QEANOVA to very low for STANOVA. In the most critical configurations (small number of member, large internal variability), large over- or underestimation of uncertainty components is very thus likely. To propose relevant uncertainty analyses and avoid misleading interpretations, estimates of uncertainty components should be therefore bias corrected and ideally come with estimates of their robustness. This work is part of the COMPLEX Project (European Collaborative Project FP7-ENV-2012 number: 308601; http://www.complex.ac.uk/). Hingray, B., Saïd, M., 2014. Partitioning internal variability and model uncertainty components in a multimodel multireplicate ensemble of climate projections. J.Climate. doi:10.1175/JCLI-D-13-00629.1 Hingray, B., Blanchet, J. (revision) Unbiased estimators for uncertainty components in transient climate projections. J. Climate Hingray, B., Blanchet, J., Vidal, J.P. (revision) Robustness of uncertainty components estimates in climate projections. J.Climate

  7. Ariadne's Thread: A Robust Software Solution Leading to Automated Absolute and Relative Quantification of SRM Data.

    PubMed

    Nasso, Sara; Goetze, Sandra; Martens, Lennart

    2015-09-04

    Selected reaction monitoring (SRM) MS is a highly selective and sensitive technique to quantify protein abundances in complex biological samples. To enhance the pace of SRM large studies, a validated, robust method to fully automate absolute quantification and to substitute for interactive evaluation would be valuable. To address this demand, we present Ariadne, a Matlab software. To quantify monitored targets, Ariadne exploits metadata imported from the transition lists, and targets can be filtered according to mProphet output. Signal processing and statistical learning approaches are combined to compute peptide quantifications. To robustly estimate absolute abundances, the external calibration curve method is applied, ensuring linearity over the measured dynamic range. Ariadne was benchmarked against mProphet and Skyline by comparing its quantification performance on three different dilution series, featuring either noisy/smooth traces without background or smooth traces with complex background. Results, evaluated as efficiency, linearity, accuracy, and precision of quantification, showed that Ariadne's performance is independent of data smoothness and complex background presence and that Ariadne outperforms mProphet on the noisier data set and improved 2-fold Skyline's accuracy and precision for the lowest abundant dilution with complex background. Remarkably, Ariadne could statistically distinguish from each other all different abundances, discriminating dilutions as low as 0.1 and 0.2 fmol. These results suggest that Ariadne offers reliable and automated analysis of large-scale SRM differential expression studies.

  8. Computationally efficient algorithm for high sampling-frequency operation of active noise control

    NASA Astrophysics Data System (ADS)

    Rout, Nirmal Kumar; Das, Debi Prasad; Panda, Ganapati

    2015-05-01

    In high sampling-frequency operation of active noise control (ANC) system the length of the secondary path estimate and the ANC filter are very long. This increases the computational complexity of the conventional filtered-x least mean square (FXLMS) algorithm. To reduce the computational complexity of long order ANC system using FXLMS algorithm, frequency domain block ANC algorithms have been proposed in past. These full block frequency domain ANC algorithms are associated with some disadvantages such as large block delay, quantization error due to computation of large size transforms and implementation difficulties in existing low-end DSP hardware. To overcome these shortcomings, the partitioned block ANC algorithm is newly proposed where the long length filters in ANC are divided into a number of equal partitions and suitably assembled to perform the FXLMS algorithm in the frequency domain. The complexity of this proposed frequency domain partitioned block FXLMS (FPBFXLMS) algorithm is quite reduced compared to the conventional FXLMS algorithm. It is further reduced by merging one fast Fourier transform (FFT)-inverse fast Fourier transform (IFFT) combination to derive the reduced structure FPBFXLMS (RFPBFXLMS) algorithm. Computational complexity analysis for different orders of filter and partition size are presented. Systematic computer simulations are carried out for both the proposed partitioned block ANC algorithms to show its accuracy compared to the time domain FXLMS algorithm.

  9. Bayesian deconvolution and quantification of metabolites in complex 1D NMR spectra using BATMAN.

    PubMed

    Hao, Jie; Liebeke, Manuel; Astle, William; De Iorio, Maria; Bundy, Jacob G; Ebbels, Timothy M D

    2014-01-01

    Data processing for 1D NMR spectra is a key bottleneck for metabolomic and other complex-mixture studies, particularly where quantitative data on individual metabolites are required. We present a protocol for automated metabolite deconvolution and quantification from complex NMR spectra by using the Bayesian automated metabolite analyzer for NMR (BATMAN) R package. BATMAN models resonances on the basis of a user-controllable set of templates, each of which specifies the chemical shifts, J-couplings and relative peak intensities for a single metabolite. Peaks are allowed to shift position slightly between spectra, and peak widths are allowed to vary by user-specified amounts. NMR signals not captured by the templates are modeled non-parametrically by using wavelets. The protocol covers setting up user template libraries, optimizing algorithmic input parameters, improving prior information on peak positions, quality control and evaluation of outputs. The outputs include relative concentration estimates for named metabolites together with associated Bayesian uncertainty estimates, as well as the fit of the remainder of the spectrum using wavelets. Graphical diagnostics allow the user to examine the quality of the fit for multiple spectra simultaneously. This approach offers a workflow to analyze large numbers of spectra and is expected to be useful in a wide range of metabolomics studies.

  10. A Novel Hybrid Dimension Reduction Technique for Undersized High Dimensional Gene Expression Data Sets Using Information Complexity Criterion for Cancer Classification

    PubMed Central

    Pamukçu, Esra; Bozdogan, Hamparsum; Çalık, Sinan

    2015-01-01

    Gene expression data typically are large, complex, and highly noisy. Their dimension is high with several thousand genes (i.e., features) but with only a limited number of observations (i.e., samples). Although the classical principal component analysis (PCA) method is widely used as a first standard step in dimension reduction and in supervised and unsupervised classification, it suffers from several shortcomings in the case of data sets involving undersized samples, since the sample covariance matrix degenerates and becomes singular. In this paper we address these limitations within the context of probabilistic PCA (PPCA) by introducing and developing a new and novel approach using maximum entropy covariance matrix and its hybridized smoothed covariance estimators. To reduce the dimensionality of the data and to choose the number of probabilistic PCs (PPCs) to be retained, we further introduce and develop celebrated Akaike's information criterion (AIC), consistent Akaike's information criterion (CAIC), and the information theoretic measure of complexity (ICOMP) criterion of Bozdogan. Six publicly available undersized benchmark data sets were analyzed to show the utility, flexibility, and versatility of our approach with hybridized smoothed covariance matrix estimators, which do not degenerate to perform the PPCA to reduce the dimension and to carry out supervised classification of cancer groups in high dimensions. PMID:25838836

  11. Economic analysis of recycling contaminated concrete

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Stephen, A.; Ayers, K.W.; Boren, J.K.

    1997-02-01

    Decontamination and Decommissioning activities in the DOE complex generate large volumes of radioactively contaminated and uncontaminated concrete. Currently, this concrete is usually decontaminated, the contaminated waste is disposed of in a LLW facility and the decontaminated concrete is placed in C&D landfills. A number of alternatives to this practice are available including recycling of the concrete. Cost estimates for six alternatives were developed using a spreadsheet model. The results of this analysis show that recycling alternatives are at least as economical as current practice.

  12. CPTC and NIST-sponsored Yeast Reference Material Now Publicly Available | Office of Cancer Clinical Proteomics Research

    Cancer.gov

    The yeast protein extract (RM8323) developed by National Institute of Standards and Technology (NIST) under the auspices of NCI's CPTC initiative is currently available to the public at https://www-s.nist.gov/srmors/view_detail.cfm?srm=8323. The yeast proteome offers researchers a unique biological reference material. RM8323 is the most extensively characterized complex biological proteome and the only one associated with several large-scale studies to estimate protein abundance across a wide concentration range.

  13. DOE Office of Scientific and Technical Information (OSTI.GOV)

    Ehlen, Mark A.; Sun, Amy C.; Pepple, Mark A.

    The potential impacts of man-made and natural disasters on chemical plants, complexes, and supply chains are of great importance to homeland security. To be able to estimate these impacts, we developed an agent-based chemical supply chain model that includes: chemical plants with enterprise operations such as purchasing, production scheduling, and inventories; merchant chemical markets, and multi-modal chemical shipments. Large-scale simulations of chemical-plant activities and supply chain interactions, running on desktop computers, are used to estimate the scope and duration of disruptive-event impacts, and overall system resilience, based on the extent to which individual chemical plants can adjust their internal operationsmore » (e.g., production mixes and levels) versus their external interactions (market sales and purchases, and transportation routes and modes). As a result, to illustrate how the model estimates the impacts of a hurricane disruption, a simple example model centered on 1,4-butanediol is presented.« less

  14. Chemical supply chain modeling for analysis of homeland security events

    DOE PAGES

    Ehlen, Mark A.; Sun, Amy C.; Pepple, Mark A.; ...

    2013-09-06

    The potential impacts of man-made and natural disasters on chemical plants, complexes, and supply chains are of great importance to homeland security. To be able to estimate these impacts, we developed an agent-based chemical supply chain model that includes: chemical plants with enterprise operations such as purchasing, production scheduling, and inventories; merchant chemical markets, and multi-modal chemical shipments. Large-scale simulations of chemical-plant activities and supply chain interactions, running on desktop computers, are used to estimate the scope and duration of disruptive-event impacts, and overall system resilience, based on the extent to which individual chemical plants can adjust their internal operationsmore » (e.g., production mixes and levels) versus their external interactions (market sales and purchases, and transportation routes and modes). As a result, to illustrate how the model estimates the impacts of a hurricane disruption, a simple example model centered on 1,4-butanediol is presented.« less

  15. A Satellite Mortality Study to Support Space Systems Lifetime Prediction

    NASA Technical Reports Server (NTRS)

    Fox, George; Salazar, Ronald; Habib-Agahi, Hamid; Dubos, Gregory

    2013-01-01

    Estimating the operational lifetime of satellites and spacecraft is a complex process. Operational lifetime can differ from mission design lifetime for a variety of reasons. Unexpected mortality can occur due to human errors in design and fabrication, to human errors in launch and operations, to random anomalies of hardware and software or even satellite function degradation or technology change, leading to unrealized economic or mission return. This study focuses on data collection of public information using, for the first time, a large, publically available dataset, and preliminary analysis of satellite lifetimes, both operational lifetime and design lifetime. The objective of this study is the illustration of the relationship of design life to actual lifetime for some representative classes of satellites and spacecraft. First, a Weibull and Exponential lifetime analysis comparison is performed on the ratio of mission operating lifetime to design life, accounting for terminated and ongoing missions. Next a Kaplan-Meier survivor function, standard practice for clinical trials analysis, is estimated from operating lifetime. Bootstrap resampling is used to provide uncertainty estimates of selected survival probabilities. This study highlights the need for more detailed databases and engineering reliability models of satellite lifetime that include satellite systems and subsystems, operations procedures and environmental characteristics to support the design of complex, multi-generation, long-lived space systems in Earth orbit.

  16. Forest height Mapping using the fusion of Lidar and MULTI-ANGLE spectral data

    NASA Astrophysics Data System (ADS)

    Pang, Y.; Li, Z.

    2016-12-01

    Characterizing the complexity of forest ecosystem over large area is highly complex. Light detection and Ranging (LIDAR) approaches have demonstrated a high capacity to accurately estimate forest structural parameters. A number of satellite mission concepts have been proposed to fuse LiDAR with other optical imagery allowing Multi-angle spectral observations to be captured using the Bidirectional Reflectance Distribution Function (BRDF) characteristics of forests. China is developing the concept of Chinese Terrestrial Carbon Mapping Satellite. A multi-beam waveform Lidar is the main sensor. A multi-angle imagery system is considered as the spatial mapping sensor. In this study, we explore the fusion potential of Lidar and multi-angle spectral data to estimate forest height across different scales. We flew intensive airborne Lidar and Multi-angle hyperspectral data in Genhe Forest Ecological Research Station, Northeast China. Then extended the spatial scale with some long transect flights to cover more forest structures. Forest height data derived from airborne lidar data was used as reference data and the multi-angle hyperspectral data was used as model inputs. Our results demonstrate that the multi-angle spectral data can be used to estimate forest height with the RMSE of 1.1 m with an R2 approximately 0.8.

  17. The application of sensitivity analysis to models of large scale physiological systems

    NASA Technical Reports Server (NTRS)

    Leonard, J. I.

    1974-01-01

    A survey of the literature of sensitivity analysis as it applies to biological systems is reported as well as a brief development of sensitivity theory. A simple population model and a more complex thermoregulatory model illustrate the investigatory techniques and interpretation of parameter sensitivity analysis. The role of sensitivity analysis in validating and verifying models, and in identifying relative parameter influence in estimating errors in model behavior due to uncertainty in input data is presented. This analysis is valuable to the simulationist and the experimentalist in allocating resources for data collection. A method for reducing highly complex, nonlinear models to simple linear algebraic models that could be useful for making rapid, first order calculations of system behavior is presented.

  18. A Model of Compound Heterozygous, Loss-of-Function Alleles Is Broadly Consistent with Observations from Complex-Disease GWAS Datasets

    PubMed Central

    Sanjak, Jaleal S.; Long, Anthony D.; Thornton, Kevin R.

    2017-01-01

    The genetic component of complex disease risk in humans remains largely unexplained. A corollary is that the allelic spectrum of genetic variants contributing to complex disease risk is unknown. Theoretical models that relate population genetic processes to the maintenance of genetic variation for quantitative traits may suggest profitable avenues for future experimental design. Here we use forward simulation to model a genomic region evolving under a balance between recurrent deleterious mutation and Gaussian stabilizing selection. We consider multiple genetic and demographic models, and several different methods for identifying genomic regions harboring variants associated with complex disease risk. We demonstrate that the model of gene action, relating genotype to phenotype, has a qualitative effect on several relevant aspects of the population genetic architecture of a complex trait. In particular, the genetic model impacts genetic variance component partitioning across the allele frequency spectrum and the power of statistical tests. Models with partial recessivity closely match the minor allele frequency distribution of significant hits from empirical genome-wide association studies without requiring homozygous effect sizes to be small. We highlight a particular gene-based model of incomplete recessivity that is appealing from first principles. Under that model, deleterious mutations in a genomic region partially fail to complement one another. This model of gene-based recessivity predicts the empirically observed inconsistency between twin and SNP based estimated of dominance heritability. Furthermore, this model predicts considerable levels of unexplained variance associated with intralocus epistasis. Our results suggest a need for improved statistical tools for region based genetic association and heritability estimation. PMID:28103232

  19. Incorporation of MRI-AIF Information For Improved Kinetic Modelling of Dynamic PET Data

    NASA Astrophysics Data System (ADS)

    Sari, Hasan; Erlandsson, Kjell; Thielemans, Kris; Atkinson, David; Ourselin, Sebastien; Arridge, Simon; Hutton, Brian F.

    2015-06-01

    In the analysis of dynamic PET data, compartmental kinetic analysis methods require an accurate knowledge of the arterial input function (AIF). Although arterial blood sampling is the gold standard of the methods used to measure the AIF, it is usually not preferred as it is an invasive method. An alternative method is the simultaneous estimation method (SIME), where physiological parameters and the AIF are estimated together, using information from different anatomical regions. Due to the large number of parameters to estimate in its optimisation, SIME is a computationally complex method and may sometimes fail to give accurate estimates. In this work, we try to improve SIME by utilising an input function derived from a simultaneously obtained DSC-MRI scan. With the assumption that the true value of one of the six parameter PET-AIF model can be derived from an MRI-AIF, the method is tested using simulated data. The results indicate that SIME can yield more robust results when the MRI information is included with a significant reduction in absolute bias of Ki estimates.

  20. Multiple Two-Way Time Message Exchange (TTME) Time Synchronization for Bridge Monitoring Wireless Sensor Networks

    PubMed Central

    Shi, Fanrong; Tuo, Xianguo; Yang, Simon X.; Li, Huailiang; Shi, Rui

    2017-01-01

    Wireless sensor networks (WSNs) have been widely used to collect valuable information in Structural Health Monitoring (SHM) of bridges, using various sensors, such as temperature, vibration and strain sensors. Since multiple sensors are distributed on the bridge, accurate time synchronization is very important for multi-sensor data fusion and information processing. Based on shape of the bridge, a spanning tree is employed to build linear topology WSNs and achieve time synchronization in this paper. Two-way time message exchange (TTME) and maximum likelihood estimation (MLE) are employed for clock offset estimation. Multiple TTMEs are proposed to obtain a subset of TTME observations. The time out restriction and retry mechanism are employed to avoid the estimation errors that are caused by continuous clock offset and software latencies. The simulation results show that the proposed algorithm could avoid the estimation errors caused by clock drift and minimize the estimation error due to the large random variable delay jitter. The proposed algorithm is an accurate and low complexity time synchronization algorithm for bridge health monitoring. PMID:28471418

  1. Multiple Two-Way Time Message Exchange (TTME) Time Synchronization for Bridge Monitoring Wireless Sensor Networks.

    PubMed

    Shi, Fanrong; Tuo, Xianguo; Yang, Simon X; Li, Huailiang; Shi, Rui

    2017-05-04

    Wireless sensor networks (WSNs) have been widely used to collect valuable information in Structural Health Monitoring (SHM) of bridges, using various sensors, such as temperature, vibration and strain sensors. Since multiple sensors are distributed on the bridge, accurate time synchronization is very important for multi-sensor data fusion and information processing. Based on shape of the bridge, a spanning tree is employed to build linear topology WSNs and achieve time synchronization in this paper. Two-way time message exchange (TTME) and maximum likelihood estimation (MLE) are employed for clock offset estimation. Multiple TTMEs are proposed to obtain a subset of TTME observations. The time out restriction and retry mechanism are employed to avoid the estimation errors that are caused by continuous clock offset and software latencies. The simulation results show that the proposed algorithm could avoid the estimation errors caused by clock drift and minimize the estimation error due to the large random variable delay jitter. The proposed algorithm is an accurate and low complexity time synchronization algorithm for bridge health monitoring.

  2. Evaluating analytical approaches for estimating pelagic fish biomass using simulated fish communities

    USGS Publications Warehouse

    Yule, Daniel L.; Adams, Jean V.; Warner, David M.; Hrabik, Thomas R.; Kocovsky, Patrick M.; Weidel, Brian C.; Rudstam, Lars G.; Sullivan, Patrick J.

    2013-01-01

    Pelagic fish assessments often combine large amounts of acoustic-based fish density data and limited midwater trawl information to estimate species-specific biomass density. We compared the accuracy of five apportionment methods for estimating pelagic fish biomass density using simulated communities with known fish numbers that mimic Lakes Superior, Michigan, and Ontario, representing a range of fish community complexities. Across all apportionment methods, the error in the estimated biomass generally declined with increasing effort, but methods that accounted for community composition changes with water column depth performed best. Correlations between trawl catch and the true species composition were highest when more fish were caught, highlighting the benefits of targeted trawling in locations of high fish density. Pelagic fish surveys should incorporate geographic and water column depth stratification in the survey design, use apportionment methods that account for species-specific depth differences, target midwater trawling effort in areas of high fish density, and include at least 15 midwater trawls. With relatively basic biological information, simulations of fish communities and sampling programs can optimize effort allocation and reduce error in biomass estimates.

  3. Online Estimation of Allan Variance Coefficients Based on a Neural-Extended Kalman Filter

    PubMed Central

    Miao, Zhiyong; Shen, Feng; Xu, Dingjie; He, Kunpeng; Tian, Chunmiao

    2015-01-01

    As a noise analysis method for inertial sensors, the traditional Allan variance method requires the storage of a large amount of data and manual analysis for an Allan variance graph. Although the existing online estimation methods avoid the storage of data and the painful procedure of drawing slope lines for estimation, they require complex transformations and even cause errors during the modeling of dynamic Allan variance. To solve these problems, first, a new state-space model that directly models the stochastic errors to obtain a nonlinear state-space model was established for inertial sensors. Then, a neural-extended Kalman filter algorithm was used to estimate the Allan variance coefficients. The real noises of an ADIS16405 IMU and fiber optic gyro-sensors were analyzed by the proposed method and traditional methods. The experimental results show that the proposed method is more suitable to estimate the Allan variance coefficients than the traditional methods. Moreover, the proposed method effectively avoids the storage of data and can be easily implemented using an online processor. PMID:25625903

  4. Density functional theory study of the interaction of vinyl radical, ethyne, and ethene with benzene, aimed to define an affordable computational level to investigate stability trends in large van der Waals complexes

    NASA Astrophysics Data System (ADS)

    Maranzana, Andrea; Giordana, Anna; Indarto, Antonius; Tonachini, Glauco; Barone, Vincenzo; Causà, Mauro; Pavone, Michele

    2013-12-01

    Our purpose is to identify a computational level sufficiently dependable and affordable to assess trends in the interaction of a variety of radical or closed shell unsaturated hydro-carbons A adsorbed on soot platelet models B. These systems, of environmental interest, would unavoidably have rather large sizes, thus prompting to explore in this paper the performances of relatively low-level computational methods and compare them with higher-level reference results. To this end, the interaction of three complexes between non-polar species, vinyl radical, ethyne, or ethene (A) with benzene (B) is studied, since these species, involved themselves in growth processes of polycyclic aromatic hydrocarbons (PAHs) and soot particles, are small enough to allow high-level reference calculations of the interaction energy ΔEAB. Counterpoise-corrected interaction energies ΔEAB are used at all stages. (1) Density Functional Theory (DFT) unconstrained optimizations of the A-B complexes are carried out, using the B3LYP-D, ωB97X-D, and M06-2X functionals, with six basis sets: 6-31G(d), 6-311 (2d,p), and 6-311++G(3df,3pd); aug-cc-pVDZ and aug-cc-pVTZ; N07T. (2) Then, unconstrained optimizations by Møller-Plesset second order Perturbation Theory (MP2), with each basis set, allow subsequent single point Coupled Cluster Singles Doubles and perturbative estimate of the Triples energy computations with the same basis sets [CCSD(T)//MP2]. (3) Based on an additivity assumption of (i) the estimated MP2 energy at the complete basis set limit [EMP2/CBS] and (ii) the higher-order correlation energy effects in passing from MP2 to CCSD(T) at the aug-cc-pVTZ basis set, ΔECC-MP, a CCSD(T)/CBS estimate is obtained and taken as a computational energy reference. At DFT, variations in ΔEAB with basis set are not large for the title molecules, and the three functionals perform rather satisfactorily even with rather small basis sets [6-31G(d) and N07T], exhibiting deviation from the computational reference of less than 1 kcal mol-1. The zero-point vibrational energy corrected estimates Δ(EAB+ZPE), obtained with the three functionals and the 6-31G(d) and N07T basis sets, are compared with experimental D0 measures, when available. In particular, this comparison is finally extended to the naphthalene and coronene dimers and to three π-π associations of different PAHs (R, made by 10, 16, or 24 C atoms) and P (80 C atoms).

  5. Density functional theory study of the interaction of vinyl radical, ethyne, and ethene with benzene, aimed to define an affordable computational level to investigate stability trends in large van der Waals complexes.

    PubMed

    Maranzana, Andrea; Giordana, Anna; Indarto, Antonius; Tonachini, Glauco; Barone, Vincenzo; Causà, Mauro; Pavone, Michele

    2013-12-28

    Our purpose is to identify a computational level sufficiently dependable and affordable to assess trends in the interaction of a variety of radical or closed shell unsaturated hydro-carbons A adsorbed on soot platelet models B. These systems, of environmental interest, would unavoidably have rather large sizes, thus prompting to explore in this paper the performances of relatively low-level computational methods and compare them with higher-level reference results. To this end, the interaction of three complexes between non-polar species, vinyl radical, ethyne, or ethene (A) with benzene (B) is studied, since these species, involved themselves in growth processes of polycyclic aromatic hydrocarbons (PAHs) and soot particles, are small enough to allow high-level reference calculations of the interaction energy ΔEAB. Counterpoise-corrected interaction energies ΔEAB are used at all stages. (1) Density Functional Theory (DFT) unconstrained optimizations of the A-B complexes are carried out, using the B3LYP-D, ωB97X-D, and M06-2X functionals, with six basis sets: 6-31G(d), 6-311 (2d,p), and 6-311++G(3df,3pd); aug-cc-pVDZ and aug-cc-pVTZ; N07T. (2) Then, unconstrained optimizations by Møller-Plesset second order Perturbation Theory (MP2), with each basis set, allow subsequent single point Coupled Cluster Singles Doubles and perturbative estimate of the Triples energy computations with the same basis sets [CCSD(T)//MP2]. (3) Based on an additivity assumption of (i) the estimated MP2 energy at the complete basis set limit [EMP2/CBS] and (ii) the higher-order correlation energy effects in passing from MP2 to CCSD(T) at the aug-cc-pVTZ basis set, ΔECC-MP, a CCSD(T)/CBS estimate is obtained and taken as a computational energy reference. At DFT, variations in ΔEAB with basis set are not large for the title molecules, and the three functionals perform rather satisfactorily even with rather small basis sets [6-31G(d) and N07T], exhibiting deviation from the computational reference of less than 1 kcal mol(-1). The zero-point vibrational energy corrected estimates Δ(EAB+ZPE), obtained with the three functionals and the 6-31G(d) and N07T basis sets, are compared with experimental D0 measures, when available. In particular, this comparison is finally extended to the naphthalene and coronene dimers and to three π-π associations of different PAHs (R, made by 10, 16, or 24 C atoms) and P (80 C atoms).

  6. Improved argument-FFT frequency offset estimation for QPSK coherent optical Systems

    NASA Astrophysics Data System (ADS)

    Han, Jilong; Li, Wei; Yuan, Zhilin; Li, Haitao; Huang, Liyan; Hu, Qianggao

    2016-02-01

    A frequency offset estimation (FOE) algorithm based on fast Fourier transform (FFT) of the signal's argument is investigated, which does not require removing the modulated data phase. In this paper, we analyze the flaw of the argument-FFT algorithm and propose a combined FOE algorithm, in which the absolute of frequency offset (FO) is accurately calculated by argument-FFT algorithm with a relatively large number of samples and the sign of FO is determined by FFT-based interpolation discrete Fourier transformation (DFT) algorithm with a relatively small number of samples. Compared with the previous algorithms based on argument-FFT, the proposed one has low complexity and can still effectively work with a relatively less number of samples.

  7. Active marks structure optimization for optical-electronic systems of spatial position control of industrial objects

    NASA Astrophysics Data System (ADS)

    Sycheva, Elena A.; Vasilev, Aleksandr S.; Lashmanov, Oleg U.; Korotaev, Valery V.

    2017-06-01

    The article is devoted to the optimization of optoelectronic systems of the spatial position of objects. Probabilistic characteristics of the detection of an active structured mark on a random noisy background are investigated. The developed computer model and the results of the study allow us to estimate the probabilistic characteristics of detection of a complex structured mark on a random gradient background, and estimate the error of spatial coordinates. The results of the study make it possible to improve the accuracy of measuring the coordinates of the object. Based on the research recommendations are given on the choice of parameters of the optimal mark structure for use in opticalelectronic systems for monitoring the spatial position of large-sized structures.

  8. Trans-dimensional Bayesian inversion of airborne electromagnetic data for 2D conductivity profiles

    NASA Astrophysics Data System (ADS)

    Hawkins, Rhys; Brodie, Ross C.; Sambridge, Malcolm

    2018-02-01

    This paper presents the application of a novel trans-dimensional sampling approach to a time domain airborne electromagnetic (AEM) inverse problem to solve for plausible conductivities of the subsurface. Geophysical inverse field problems, such as time domain AEM, are well known to have a large degree of non-uniqueness. Common least-squares optimisation approaches fail to take this into account and provide a single solution with linearised estimates of uncertainty that can result in overly optimistic appraisal of the conductivity of the subsurface. In this new non-linear approach, the spatial complexity of a 2D profile is controlled directly by the data. By examining an ensemble of proposed conductivity profiles it accommodates non-uniqueness and provides more robust estimates of uncertainties.

  9. Global Ocean Vertical Velocity From a Dynamically Consistent Ocean State Estimate

    NASA Astrophysics Data System (ADS)

    Liang, Xinfeng; Spall, Michael; Wunsch, Carl

    2017-10-01

    Estimates of the global ocean vertical velocities (Eulerian, eddy-induced, and residual) from a dynamically consistent and data-constrained ocean state estimate are presented and analyzed. Conventional patterns of vertical velocity, Ekman pumping, appear in the upper ocean, with topographic dominance at depth. Intense and vertically coherent upwelling and downwelling occur in the Southern Ocean, which are likely due to the interaction of the Antarctic Circumpolar Current and large-scale topographic features and are generally canceled out in the conventional zonally averaged results. These "elevators" at high latitudes connect the upper to the deep and abyssal oceans and working together with isopycnal mixing are likely a mechanism, in addition to the formation of deep and abyssal waters, for fast responses of the deep and abyssal oceans to the changing climate. Also, Eulerian and parameterized eddy-induced components are of opposite signs in numerous regions around the global ocean, particularly in the ocean interior away from surface and bottom. Nevertheless, residual vertical velocity is primarily determined by the Eulerian component, and related to winds and large-scale topographic features. The current estimates of vertical velocities can serve as a useful reference for investigating the vertical exchange of ocean properties and tracers, and its complex spatial structure ultimately permits regional tests of basic oceanographic concepts such as Sverdrup balance and coastal upwelling/downwelling.

  10. Estimation of health effects of prenatal methylmercury exposure using structural equation models.

    PubMed

    Budtz-Jørgensen, Esben; Keiding, Niels; Grandjean, Philippe; Weihe, Pal

    2002-10-14

    Observational studies in epidemiology always involve concerns regarding validity, especially measurement error, confounding, missing data, and other problems that may affect the study outcomes. Widely used standard statistical techniques, such as multiple regression analysis, may to some extent adjust for these shortcomings. However, structural equations may incorporate most of these considerations, thereby providing overall adjusted estimations of associations. This approach was used in a large epidemiological data set from a prospective study of developmental methyl-mercury toxicity. Structural equation models were developed for assessment of the association between biomarkers of prenatal mercury exposure and neuropsychological test scores in 7 year old children. Eleven neurobehavioral outcomes were grouped into motor function and verbally mediated function. Adjustment for local dependence and item bias was necessary for a satisfactory fit of the model, but had little impact on the estimated mercury effects. The mercury effect on the two latent neurobehavioral functions was similar to the strongest effects seen for individual test scores of motor function and verbal skills. Adjustment for contaminant exposure to poly chlorinated biphenyls (PCBs) changed the estimates only marginally, but the mercury effect could be reduced to non-significance by assuming a large measurement error for the PCB biomarker. The structural equation analysis allows correction for measurement error in exposure variables, incorporation of multiple outcomes and incomplete cases. This approach therefore deserves to be applied more frequently in the analysis of complex epidemiological data sets.

  11. A Modular Localization System as a Positioning Service for Road Transport

    PubMed Central

    Brida, Peter; Machaj, Juraj; Benikovsky, Jozef

    2014-01-01

    In recent times smart devices have attracted a large number of users. Since many of these devices allow position estimation using Global Navigation Satellite Systems (GNSS) signals, a large number of location-based applications and services have emerged, especially in transport systems. However GNSS signals are affected by the environment and are not always present, especially in dense urban environment or indoors. In this work firstly a Modular Localization Algorithm is proposed to allow seamless switching between different positioning modules. This helps us develop a positioning system that is able to provide position estimates in both indoor and outdoor environments without any user interaction. Since the proposed system can run as a service on any smart device, it could allow users to navigate not only in outdoor environments, but also indoors, e.g., underground garages, tunnels etc. Secondly we present the proposal of a 2-phase map reduction algorithm which allows one to significantly reduce the complexity of position estimation processes in case that positioning is performed using a fingerprinting framework. The proposed 2-phase map reduction algorithm can also improve the accuracy of the position estimates by filtering out reference points that are far from the mobile device. Both algorithms were implemented into a positioning system and tested in real world conditions in both indoor and outdoor environments. PMID:25353979

  12. Local Ancestry Inference in a Large US-Based Hispanic/Latino Study: Hispanic Community Health Study/Study of Latinos (HCHS/SOL)

    PubMed Central

    Browning, Sharon R.; Grinde, Kelsey; Plantinga, Anna; Gogarten, Stephanie M.; Stilp, Adrienne M.; Kaplan, Robert C.; Avilés-Santa, M. Larissa; Browning, Brian L.; Laurie, Cathy C.

    2016-01-01

    We estimated local ancestry on the autosomes and X chromosome in a large US-based study of 12,793 Hispanic/Latino individuals using the RFMix method, and we compared different reference panels and approaches to local ancestry estimation on the X chromosome by means of Mendelian inconsistency rates as a proxy for accuracy. We developed a novel and straightforward approach to performing ancestry-specific PCA after finding artifactual behavior in the results from an existing approach. Using the ancestry-specific PCA, we found significant population structure within African, European, and Amerindian ancestries in the Hispanic/Latino individuals in our study. In the African ancestral component of the admixed individuals, individuals whose grandparents were from Central America clustered separately from individuals whose grandparents were from the Caribbean, and also from reference Yoruba and Mandenka West African individuals. In the European component, individuals whose grandparents were from Puerto Rico diverged partially from other background groups. In the Amerindian ancestral component, individuals clustered into multiple different groups depending on the grandparental country of origin. Therefore, local ancestry estimation provides further insight into the complex genetic structure of US Hispanic/Latino populations, which must be properly accounted for in genotype-phenotype association studies. It also provides a basis for admixture mapping and ancestry-specific allele frequency estimation, which are useful in the identification of risk factors for disease. PMID:27172203

  13. pplacer: linear time maximum-likelihood and Bayesian phylogenetic placement of sequences onto a fixed reference tree

    PubMed Central

    2010-01-01

    Background Likelihood-based phylogenetic inference is generally considered to be the most reliable classification method for unknown sequences. However, traditional likelihood-based phylogenetic methods cannot be applied to large volumes of short reads from next-generation sequencing due to computational complexity issues and lack of phylogenetic signal. "Phylogenetic placement," where a reference tree is fixed and the unknown query sequences are placed onto the tree via a reference alignment, is a way to bring the inferential power offered by likelihood-based approaches to large data sets. Results This paper introduces pplacer, a software package for phylogenetic placement and subsequent visualization. The algorithm can place twenty thousand short reads on a reference tree of one thousand taxa per hour per processor, has essentially linear time and memory complexity in the number of reference taxa, and is easy to run in parallel. Pplacer features calculation of the posterior probability of a placement on an edge, which is a statistically rigorous way of quantifying uncertainty on an edge-by-edge basis. It also can inform the user of the positional uncertainty for query sequences by calculating expected distance between placement locations, which is crucial in the estimation of uncertainty with a well-sampled reference tree. The software provides visualizations using branch thickness and color to represent number of placements and their uncertainty. A simulation study using reads generated from 631 COG alignments shows a high level of accuracy for phylogenetic placement over a wide range of alignment diversity, and the power of edge uncertainty estimates to measure placement confidence. Conclusions Pplacer enables efficient phylogenetic placement and subsequent visualization, making likelihood-based phylogenetics methodology practical for large collections of reads; it is freely available as source code, binaries, and a web service. PMID:21034504

  14. Practical problems in aggregating expert opinions

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Booker, J.M.; Picard, R.R.; Meyer, M.A.

    1993-11-01

    Expert opinion is data given by a qualified person in response to a technical question. In these analyses, expert opinion provides information where other data are either sparse or non-existent. Improvements in forecasting result from the advantageous addition of expert opinion to observed data in many areas, such as meteorology and econometrics. More generally, analyses of large, complex systems often involve experts on various components of the system supplying input to a decision process; applications include such wide-ranging areas as nuclear reactor safety, management science, and seismology. For large or complex applications, no single expert may be knowledgeable enough aboutmore » the entire application. In other problems, decision makers may find it comforting that a consensus or aggregation of opinions is usually better than a single opinion. Many risk and reliability studies require a single estimate for modeling, analysis, reporting, and decision making purposes. For problems with large uncertainties, the strategy of combining as diverse a set of experts as possible hedges against underestimation of that uncertainty. Decision makers are frequently faced with the task of selecting the experts and combining their opinions. However, the aggregation is often the responsibility of an analyst. Whether the decision maker or the analyst does the aggregation, the input for it, such as providing weights for experts or estimating other parameters, is imperfect owing to a lack of omniscience. Aggregation methods for expert opinions have existed for over thirty years; yet many of the difficulties with their use remain unresolved. The bulk of these problem areas are summarized in the sections that follow: sensitivities of results to assumptions, weights for experts, correlation of experts, and handling uncertainties. The purpose of this paper is to discuss the sources of these problems and describe their effects on aggregation.« less

  15. Delineating parameter unidentifiabilities in complex models

    NASA Astrophysics Data System (ADS)

    Raman, Dhruva V.; Anderson, James; Papachristodoulou, Antonis

    2017-03-01

    Scientists use mathematical modeling as a tool for understanding and predicting the properties of complex physical systems. In highly parametrized models there often exist relationships between parameters over which model predictions are identical, or nearly identical. These are known as structural or practical unidentifiabilities, respectively. They are hard to diagnose and make reliable parameter estimation from data impossible. They furthermore imply the existence of an underlying model simplification. We describe a scalable method for detecting unidentifiabilities, as well as the functional relations defining them, for generic models. This allows for model simplification, and appreciation of which parameters (or functions thereof) cannot be estimated from data. Our algorithm can identify features such as redundant mechanisms and fast time-scale subsystems, as well as the regimes in parameter space over which such approximations are valid. We base our algorithm on a quantification of regional parametric sensitivity that we call `multiscale sloppiness'. Traditionally, the link between parametric sensitivity and the conditioning of the parameter estimation problem is made locally, through the Fisher information matrix. This is valid in the regime of infinitesimal measurement uncertainty. We demonstrate the duality between multiscale sloppiness and the geometry of confidence regions surrounding parameter estimates made where measurement uncertainty is non-negligible. Further theoretical relationships are provided linking multiscale sloppiness to the likelihood-ratio test. From this, we show that a local sensitivity analysis (as typically done) is insufficient for determining the reliability of parameter estimation, even with simple (non)linear systems. Our algorithm can provide a tractable alternative. We finally apply our methods to a large-scale, benchmark systems biology model of necrosis factor (NF)-κ B , uncovering unidentifiabilities.

  16. Airborne Lidar-Based Estimates of Tropical Forest Structure in Complex Terrain: Opportunities and Trade-Offs for REDD+

    NASA Technical Reports Server (NTRS)

    Leitold, Veronika; Keller, Michael; Morton, Douglas C.; Cook, Bruce D.; Shimabukuro, Yosio E.

    2015-01-01

    Background: Carbon stocks and fluxes in tropical forests remain large sources of uncertainty in the global carbon budget. Airborne lidar remote sensing is a powerful tool for estimating aboveground biomass, provided that lidar measurements penetrate dense forest vegetation to generate accurate estimates of surface topography and canopy heights. Tropical forest areas with complex topography present a challenge for lidar remote sensing. Results: We compared digital terrain models (DTM) derived from airborne lidar data from a mountainous region of the Atlantic Forest in Brazil to 35 ground control points measured with survey grade GNSS receivers. The terrain model generated from full-density (approx. 20 returns/sq m) data was highly accurate (mean signed error of 0.19 +/-0.97 m), while those derived from reduced-density datasets (8/sq m, 4/sq m, 2/sq m and 1/sq m) were increasingly less accurate. Canopy heights calculated from reduced-density lidar data declined as data density decreased due to the inability to accurately model the terrain surface. For lidar return densities below 4/sq m, the bias in height estimates translated into errors of 80-125 Mg/ha in predicted aboveground biomass. Conclusions: Given the growing emphasis on the use of airborne lidar for forest management, carbon monitoring, and conservation efforts, the results of this study highlight the importance of careful survey planning and consistent sampling for accurate quantification of aboveground biomass stocks and dynamics. Approaches that rely primarily on canopy height to estimate aboveground biomass are sensitive to DTM errors from variability in lidar sampling density.

  17. Airborne lidar-based estimates of tropical forest structure in complex terrain: opportunities and trade-offs for REDD+

    PubMed

    Leitold, Veronika; Keller, Michael; Morton, Douglas C; Cook, Bruce D; Shimabukuro, Yosio E

    2015-12-01

    Carbon stocks and fluxes in tropical forests remain large sources of uncertainty in the global carbon budget. Airborne lidar remote sensing is a powerful tool for estimating aboveground biomass, provided that lidar measurements penetrate dense forest vegetation to generate accurate estimates of surface topography and canopy heights. Tropical forest areas with complex topography present a challenge for lidar remote sensing. We compared digital terrain models (DTM) derived from airborne lidar data from a mountainous region of the Atlantic Forest in Brazil to 35 ground control points measured with survey grade GNSS receivers. The terrain model generated from full-density (~20 returns m -2 ) data was highly accurate (mean signed error of 0.19 ± 0.97 m), while those derived from reduced-density datasets (8 m -2 , 4 m -2 , 2 m -2 and 1 m -2 ) were increasingly less accurate. Canopy heights calculated from reduced-density lidar data declined as data density decreased due to the inability to accurately model the terrain surface. For lidar return densities below 4 m -2 , the bias in height estimates translated into errors of 80-125 Mg ha -1 in predicted aboveground biomass. Given the growing emphasis on the use of airborne lidar for forest management, carbon monitoring, and conservation efforts, the results of this study highlight the importance of careful survey planning and consistent sampling for accurate quantification of aboveground biomass stocks and dynamics. Approaches that rely primarily on canopy height to estimate aboveground biomass are sensitive to DTM errors from variability in lidar sampling density.

  18. Estimating under-five mortality in space and time in a developing world context.

    PubMed

    Wakefield, Jon; Fuglstad, Geir-Arne; Riebler, Andrea; Godwin, Jessica; Wilson, Katie; Clark, Samuel J

    2018-01-01

    Accurate estimates of the under-five mortality rate in a developing world context are a key barometer of the health of a nation. This paper describes a new model to analyze survey data on mortality in this context. We are interested in both spatial and temporal description, that is wishing to estimate under-five mortality rate across regions and years and to investigate the association between the under-five mortality rate and spatially varying covariate surfaces. We illustrate the methodology by producing yearly estimates for subnational areas in Kenya over the period 1980-2014 using data from the Demographic and Health Surveys, which use stratified cluster sampling. We use a binomial likelihood with fixed effects for the urban/rural strata and random effects for the clustering to account for the complex survey design. Smoothing is carried out using Bayesian hierarchical models with continuous spatial and temporally discrete components. A key component of the model is an offset to adjust for bias due to the effects of HIV epidemics. Substantively, there has been a sharp decline in Kenya in the under-five mortality rate in the period 1980-2014, but large variability in estimated subnational rates remains. A priority for future research is understanding this variability. In exploratory work, we examine whether a variety of spatial covariate surfaces can explain the variability in under-five mortality rate. Temperature, precipitation, a measure of malaria infection prevalence, and a measure of nearness to cities were candidates for inclusion in the covariate model, but the interplay between space, time, and covariates is complex.

  19. Understanding and estimating effective population size for practical application in marine species management.

    PubMed

    Hare, Matthew P; Nunney, Leonard; Schwartz, Michael K; Ruzzante, Daniel E; Burford, Martha; Waples, Robin S; Ruegg, Kristen; Palstra, Friso

    2011-06-01

    Effective population size (N(e)) determines the strength of genetic drift in a population and has long been recognized as an important parameter for evaluating conservation status and threats to genetic health of populations. Specifically, an estimate of N(e) is crucial to management because it integrates genetic effects with the life history of the species, allowing for predictions of a population's current and future viability. Nevertheless, compared with ecological and demographic parameters, N(e) has had limited influence on species management, beyond its application in very small populations. Recent developments have substantially improved N(e) estimation; however, some obstacles remain for the practical application of N(e) estimates. For example, the need to define the spatial and temporal scale of measurement makes the concept complex and sometimes difficult to interpret. We reviewed approaches to estimation of N(e) over both long-term and contemporary time frames, clarifying their interpretations with respect to local populations and the global metapopulation. We describe multiple experimental factors affecting robustness of contemporary N(e) estimates and suggest that different sampling designs can be combined to compare largely independent measures of N(e) for improved confidence in the result. Large populations with moderate gene flow pose the greatest challenges to robust estimation of contemporary N(e) and require careful consideration of sampling and analysis to minimize estimator bias. We emphasize the practical utility of estimating N(e) by highlighting its relevance to the adaptive potential of a population and describing applications in management of marine populations, where the focus is not always on critically endangered populations. Two cases discussed include the mechanisms generating N(e) estimates many orders of magnitude lower than census N in harvested marine fishes and the predicted reduction in N(e) from hatchery-based population supplementation. ©2011 Society for Conservation Biology.

  20. Improving a regional model using reduced complexity and parameter estimation

    USGS Publications Warehouse

    Kelson, Victor A.; Hunt, Randall J.; Haitjema, Henk M.

    2002-01-01

    The availability of powerful desktop computers and graphical user interfaces for ground water flow models makes possible the construction of ever more complex models. A proposed copper-zinc sulfide mine in northern Wisconsin offers a unique case in which the same hydrologic system has been modeled using a variety of techniques covering a wide range of sophistication and complexity. Early in the permitting process, simple numerical models were used to evaluate the necessary amount of water to be pumped from the mine, reductions in streamflow, and the drawdowns in the regional aquifer. More complex models have subsequently been used in an attempt to refine the predictions. Even after so much modeling effort, questions regarding the accuracy and reliability of the predictions remain. We have performed a new analysis of the proposed mine using the two-dimensional analytic element code GFLOW coupled with the nonlinear parameter estimation code UCODE. The new model is parsimonious, containing fewer than 10 parameters, and covers a region several times larger in areal extent than any of the previous models. The model demonstrates the suitability of analytic element codes for use with parameter estimation codes. The simplified model results are similar to the more complex models; predicted mine inflows and UCODE-derived 95% confidence intervals are consistent with the previous predictions. More important, the large areal extent of the model allowed us to examine hydrological features not included in the previous models, resulting in new insights about the effects that far-field boundary conditions can have on near-field model calibration and parameterization. In this case, the addition of surface water runoff into a lake in the headwaters of a stream while holding recharge constant moved a regional ground watershed divide and resulted in some of the added water being captured by the adjoining basin. Finally, a simple analytical solution was used to clarify the GFLOW model's prediction that, for a model that is properly calibrated for heads, regional drawdowns are relatively unaffected by the choice of aquifer properties, but that mine inflows are strongly affected. Paradoxically, by reducing model complexity, we have increased the understanding gained from the modeling effort.

  1. Improving a regional model using reduced complexity and parameter estimation.

    PubMed

    Kelson, Victor A; Hunt, Randall J; Haitjema, Henk M

    2002-01-01

    The availability of powerful desktop computers and graphical user interfaces for ground water flow models makes possible the construction of ever more complex models. A proposed copper-zinc sulfide mine in northern Wisconsin offers a unique case in which the same hydrologic system has been modeled using a variety of techniques covering a wide range of sophistication and complexity. Early in the permitting process, simple numerical models were used to evaluate the necessary amount of water to be pumped from the mine, reductions in streamflow, and the drawdowns in the regional aquifer. More complex models have subsequently been used in an attempt to refine the predictions. Even after so much modeling effort, questions regarding the accuracy and reliability of the predictions remain. We have performed a new analysis of the proposed mine using the two-dimensional analytic element code GFLOW coupled with the nonlinear parameter estimation code UCODE. The new model is parsimonious, containing fewer than 10 parameters, and covers a region several times larger in areal extent than any of the previous models. The model demonstrates the suitability of analytic element codes for use with parameter estimation codes. The simplified model results are similar to the more complex models; predicted mine inflows and UCODE-derived 95% confidence intervals are consistent with the previous predictions. More important, the large areal extent of the model allowed us to examine hydrological features not included in the previous models, resulting in new insights about the effects that far-field boundary conditions can have on near-field model calibration and parameterization. In this case, the addition of surface water runoff into a lake in the headwaters of a stream while holding recharge constant moved a regional ground watershed divide and resulted in some of the added water being captured by the adjoining basin. Finally, a simple analytical solution was used to clarify the GFLOW model's prediction that, for a model that is properly calibrated for heads, regional drawdowns are relatively unaffected by the choice of aquifer properties, but that mine inflows are strongly affected. Paradoxically, by reducing model complexity, we have increased the understanding gained from the modeling effort.

  2. Estimating variation in stomatal frequency at intra-individual, intra-site, and inter-taxonomic levels in populations of the Leonardoxa africana (Fabaceae) complex over environmental gradients in Cameroon

    NASA Astrophysics Data System (ADS)

    Finsinger, Walter; Dos Santos, Thibaut; McKey, Doyle

    2013-07-01

    Variation of stomatal frequency (stomatal density and stomatal index) includes genetically-based, potentially-adaptive variation, and variation due to phenotypic plasticity, the degree of which may be fundamental to the ability to maintain high water-use efficiency and thus to deal with environmental change. We analysed stomatal frequency and morphology (pore length, pore width) in leaves from several individuals from nine populations of four sub-species of the Leonardoxa africana complex. The dataset represents a hierarchical sampling wherein factors are nested within each level (leaves in individuals, individuals in sites, etc.), allowing estimation of the contribution of different levels to overall variation, using variance-component analysis. SI showed significant variation among sites ("site" is largely confounded with "sub-species"), being highest in the sub-species localized in the highest-elevation site. However, most of the observed variance was accounted for at intra-site and intra-individual levels. This variance could reflect great phenotypic plasticity, presumably in response to highly local variation in micro-environmental conditions.

  3. Limited plasticity in the phenotypic variance-covariance matrix for male advertisement calls in the black field cricket, Teleogryllus commodus

    PubMed Central

    Pitchers, W. R.; Brooks, R.; Jennions, M. D.; Tregenza, T.; Dworkin, I.; Hunt, J.

    2013-01-01

    Phenotypic integration and plasticity are central to our understanding of how complex phenotypic traits evolve. Evolutionary change in complex quantitative traits can be predicted using the multivariate breeders’ equation, but such predictions are only accurate if the matrices involved are stable over evolutionary time. Recent work, however, suggests that these matrices are temporally plastic, spatially variable and themselves evolvable. The data available on phenotypic variance-covariance matrix (P) stability is sparse, and largely focused on morphological traits. Here we compared P for the structure of the complex sexual advertisement call of six divergent allopatric populations of the Australian black field cricket, Teleogryllus commodus. We measured a subset of calls from wild-caught crickets from each of the populations and then a second subset after rearing crickets under common-garden conditions for three generations. In a second experiment, crickets from each population were reared in the laboratory on high- and low-nutrient diets and their calls recorded. In both experiments, we estimated P for call traits and used multiple methods to compare them statistically (Flury hierarchy, geometric subspace comparisons and random skewers). Despite considerable variation in means and variances of individual call traits, the structure of P was largely conserved among populations, across generations and between our rearing diets. Our finding that P remains largely stable, among populations and between environmental conditions, suggests that selection has preserved the structure of call traits in order that they can function as an integrated unit. PMID:23530814

  4. An analysis of species boundaries and biogeographic patterns in a cryptic species complex: the rotifer--Brachionus plicatilis.

    PubMed

    Suatoni, Elizabeth; Vicario, Saverio; Rice, Sean; Snell, Terry; Caccone, Adalgisa

    2006-10-01

    Since the advent of molecular phylogenetics, there is increasing evidence that many small aquatic and marine invertebrates--once believed to be single, cosmopolitan species--are in fact cryptic species complexes. Although the application of the biological species concept is central to the identification of species boundaries in these cryptic complexes, tests of reproductive isolation do not frequently accompany phylogenetic studies. Because different species concepts generally identify different boundaries in cryptic complexes, studies that apply multiple species concepts are needed to gain a more detailed understanding of patterns of diversification in these taxa. Here we explore different methods of empirically delimiting species boundaries in the salt water rotifer Brachionus plicatilis by comparing reproductive data (i.e., the traditional biological species concept) to phylogenetic data (the genealogical species concept). Based on a high degree of molecular sequence divergence and largely concordant genetic patterns in COI and ITS1, the genealogical species hypothesis indicates the existence of at least 14 species--the highest estimate for the group thus far. A test of the genealogical species concept with biological crosses shows a fairly high level of concordance, depending on the degree of reproductive success used to draw boundaries. The convergence of species concepts in this group suggests that many of the species within the group may be old. Although the diversity of the group is higher than previously understood, geographic distributions remain broad. Efficient passive dispersal has resulted in global distributions for many species with some evidence of isolation by distance over large geographic scales. These patterns concur with expectations that micro-meiofauna (0.1-1mm) have biogeographies intermediate to microbial organisms and large vertebrates. Sympatry of genetically distant strains is common.

  5. Towards the understanding of network information processing in biology

    NASA Astrophysics Data System (ADS)

    Singh, Vijay

    Living organisms perform incredibly well in detecting a signal present in the environment. This information processing is achieved near optimally and quite reliably, even though the sources of signals are highly variable and complex. The work in the last few decades has given us a fair understanding of how individual signal processing units like neurons and cell receptors process signals, but the principles of collective information processing on biological networks are far from clear. Information processing in biological networks, like the brain, metabolic circuits, cellular-signaling circuits, etc., involves complex interactions among a large number of units (neurons, receptors). The combinatorially large number of states such a system can exist in makes it impossible to study these systems from the first principles, starting from the interactions between the basic units. The principles of collective information processing on such complex networks can be identified using coarse graining approaches. This could provide insights into the organization and function of complex biological networks. Here I study models of biological networks using continuum dynamics, renormalization, maximum likelihood estimation and information theory. Such coarse graining approaches identify features that are essential for certain processes performed by underlying biological networks. We find that long-range connections in the brain allow for global scale feature detection in a signal. These also suppress the noise and remove any gaps present in the signal. Hierarchical organization with long-range connections leads to large-scale connectivity at low synapse numbers. Time delays can be utilized to separate a mixture of signals with temporal scales. Our observations indicate that the rules in multivariate signal processing are quite different from traditional single unit signal processing.

  6. The Second Galactic Center Black Hole? A Possible Detection of Ionized Gas Orbiting around an IMBH Embedded in the Galactic Center IRS13E Complex

    NASA Astrophysics Data System (ADS)

    Tsuboi, Masato; Kitamura, Yoshimi; Tsutsumi, Takahiro; Uehara, Kenta; Miyoshi, Makoto; Miyawaki, Ryosuke; Miyazaki, Atsushi

    2017-11-01

    The Galactic Center is the nuclear region of the nearest spiral galaxy, the Milky Way, and contains the supermassive black hole with M˜ 4× {10}6 {M}⊙ , Sagittarius A* (Sgr A*). One of the basic questions about the Galactic Center is whether or not Sgr A* is the only “massive” black hole in the region. The IRS13E complex is a very intriguing infrared (IR) object that contains a large dark mass comparable to the mass of an intermediate mass black hole (IMBH) from the proper motions of the main member stars. However, the existence of the IMBH remains controversial. There are some objections to accepting the existence of the IMBH. In this study, we detected ionized gas with a very large velocity width ({{Δ }}{v}{FWZI}˜ 650 km s-1) and a very compact size (r˜ 400 au) in the complex using the Atacama Large Millimeter/submillimeter Array (ALMA). We also found an extended component connecting with the compact ionized gas. The properties suggest that this is an ionized gas flow on the Keplerian orbit with high eccentricity. The enclosed mass is estimated to be {10}4 {M}⊙ by the analysis of the orbit. The mass does not conflict with the upper limit mass of the IMBH around Sgr A*, which is derived by the long-term astrometry with the Very Long Baseline Array (VLBA). In addition, the object probably has an X-ray counterpart. Consequently, a very fascinating possibility is that the detected ionized gas is rotating around an IMBH embedded in the IRS13E complex.

  7. Robust Vision-Based Pose Estimation Algorithm for AN Uav with Known Gravity Vector

    NASA Astrophysics Data System (ADS)

    Kniaz, V. V.

    2016-06-01

    Accurate estimation of camera external orientation with respect to a known object is one of the central problems in photogrammetry and computer vision. In recent years this problem is gaining an increasing attention in the field of UAV autonomous flight. Such application requires a real-time performance and robustness of the external orientation estimation algorithm. The accuracy of the solution is strongly dependent on the number of reference points visible on the given image. The problem only has an analytical solution if 3 or more reference points are visible. However, in limited visibility conditions it is often needed to perform external orientation with only 2 visible reference points. In such case the solution could be found if the gravity vector direction in the camera coordinate system is known. A number of algorithms for external orientation estimation for the case of 2 known reference points and a gravity vector were developed to date. Most of these algorithms provide analytical solution in the form of polynomial equation that is subject to large errors in the case of complex reference points configurations. This paper is focused on the development of a new computationally effective and robust algorithm for external orientation based on positions of 2 known reference points and a gravity vector. The algorithm implementation for guidance of a Parrot AR.Drone 2.0 micro-UAV is discussed. The experimental evaluation of the algorithm proved its computational efficiency and robustness against errors in reference points positions and complex configurations.

  8. PENDISC: a simple method for constructing a mathematical model from time-series data of metabolite concentrations.

    PubMed

    Sriyudthsak, Kansuporn; Iwata, Michio; Hirai, Masami Yokota; Shiraishi, Fumihide

    2014-06-01

    The availability of large-scale datasets has led to more effort being made to understand characteristics of metabolic reaction networks. However, because the large-scale data are semi-quantitative, and may contain biological variations and/or analytical errors, it remains a challenge to construct a mathematical model with precise parameters using only these data. The present work proposes a simple method, referred to as PENDISC (Parameter Estimation in a N on- DImensionalized S-system with Constraints), to assist the complex process of parameter estimation in the construction of a mathematical model for a given metabolic reaction system. The PENDISC method was evaluated using two simple mathematical models: a linear metabolic pathway model with inhibition and a branched metabolic pathway model with inhibition and activation. The results indicate that a smaller number of data points and rate constant parameters enhances the agreement between calculated values and time-series data of metabolite concentrations, and leads to faster convergence when the same initial estimates are used for the fitting. This method is also shown to be applicable to noisy time-series data and to unmeasurable metabolite concentrations in a network, and to have a potential to handle metabolome data of a relatively large-scale metabolic reaction system. Furthermore, it was applied to aspartate-derived amino acid biosynthesis in Arabidopsis thaliana plant. The result provides confirmation that the mathematical model constructed satisfactorily agrees with the time-series datasets of seven metabolite concentrations.

  9. Jackknife Estimation of Sampling Variance of Ratio Estimators in Complex Samples: Bias and the Coefficient of Variation. Research Report. ETS RR-06-19

    ERIC Educational Resources Information Center

    Oranje, Andreas

    2006-01-01

    A multitude of methods has been proposed to estimate the sampling variance of ratio estimates in complex samples (Wolter, 1985). Hansen and Tepping (1985) studied some of those variance estimators and found that a high coefficient of variation (CV) of the denominator of a ratio estimate is indicative of a biased estimate of the standard error of a…

  10. Detectability of change in winter precipitation within mountain landscapes: Spatial patterns and uncertainty

    NASA Astrophysics Data System (ADS)

    Silverman, N. L.; Maneta, M. P.

    2016-06-01

    Detecting long-term change in seasonal precipitation using ground observations is dependent on the representativity of the point measurement to the surrounding landscape. In mountainous regions, representativity can be poor and lead to large uncertainties in precipitation estimates at high elevations or in areas where observations are sparse. If the uncertainty in the estimate is large compared to the long-term shifts in precipitation, then the change will likely go undetected. In this analysis, we examine the minimum detectable change across mountainous terrain in western Montana, USA. We ask the question: What is the minimum amount of change that is necessary to be detected using our best estimates of precipitation in complex terrain? We evaluate the spatial uncertainty in the precipitation estimates by conditioning historic regional climate model simulations to ground observations using Bayesian inference. By using this uncertainty as a null hypothesis, we test for detectability across the study region. To provide context for the detectability calculations, we look at a range of future scenarios from the Coupled Model Intercomparison Project 5 (CMIP5) multimodel ensemble downscaled to 4 km resolution using the MACAv2-METDATA data set. When using the ensemble averages we find that approximately 65% of the significant increases in winter precipitation go undetected at midelevations. At high elevation, approximately 75% of significant increases in winter precipitation are undetectable. Areas where change can be detected are largely controlled by topographic features. Elevation and aspect are key characteristics that determine whether or not changes in winter precipitation can be detected. Furthermore, we find that undetected increases in winter precipitation at high elevation will likely remain as snow under climate change scenarios. Therefore, there is potential for these areas to offset snowpack loss at lower elevations and confound the effects of climate change on water resources.

  11. Catchment Tomography - Joint Estimation of Surface Roughness and Hydraulic Conductivity with the EnKF

    NASA Astrophysics Data System (ADS)

    Baatz, D.; Kurtz, W.; Hendricks Franssen, H. J.; Vereecken, H.; Kollet, S. J.

    2017-12-01

    Parameter estimation for physically based, distributed hydrological models becomes increasingly challenging with increasing model complexity. The number of parameters is usually large and the number of observations relatively small, which results in large uncertainties. A moving transmitter - receiver concept to estimate spatially distributed hydrological parameters is presented by catchment tomography. In this concept, precipitation, highly variable in time and space, serves as a moving transmitter. As response to precipitation, runoff and stream discharge are generated along different paths and time scales, depending on surface and subsurface flow properties. Stream water levels are thus an integrated signal of upstream parameters, measured by stream gauges which serve as the receivers. These stream water level observations are assimilated into a distributed hydrological model, which is forced with high resolution, radar based precipitation estimates. Applying a joint state-parameter update with the Ensemble Kalman Filter, the spatially distributed Manning's roughness coefficient and saturated hydraulic conductivity are estimated jointly. The sequential data assimilation continuously integrates new information into the parameter estimation problem, especially during precipitation events. Every precipitation event constrains the possible parameter space. In the approach, forward simulations are performed with ParFlow, a variable saturated subsurface and overland flow model. ParFlow is coupled to the Parallel Data Assimilation Framework for the data assimilation and the joint state-parameter update. In synthetic, 3-dimensional experiments including surface and subsurface flow, hydraulic conductivity and the Manning's coefficient are efficiently estimated with the catchment tomography approach. A joint update of the Manning's coefficient and hydraulic conductivity tends to improve the parameter estimation compared to a single parameter update, especially in cases of biased initial parameter ensembles. The computational experiments additionally show to which degree of spatial heterogeneity and to which degree of uncertainty of subsurface flow parameters the Manning's coefficient and hydraulic conductivity can be estimated efficiently.

  12. State estimation and prediction using clustered particle filters.

    PubMed

    Lee, Yoonsang; Majda, Andrew J

    2016-12-20

    Particle filtering is an essential tool to improve uncertain model predictions by incorporating noisy observational data from complex systems including non-Gaussian features. A class of particle filters, clustered particle filters, is introduced for high-dimensional nonlinear systems, which uses relatively few particles compared with the standard particle filter. The clustered particle filter captures non-Gaussian features of the true signal, which are typical in complex nonlinear dynamical systems such as geophysical systems. The method is also robust in the difficult regime of high-quality sparse and infrequent observations. The key features of the clustered particle filtering are coarse-grained localization through the clustering of the state variables and particle adjustment to stabilize the method; each observation affects only neighbor state variables through clustering and particles are adjusted to prevent particle collapse due to high-quality observations. The clustered particle filter is tested for the 40-dimensional Lorenz 96 model with several dynamical regimes including strongly non-Gaussian statistics. The clustered particle filter shows robust skill in both achieving accurate filter results and capturing non-Gaussian statistics of the true signal. It is further extended to multiscale data assimilation, which provides the large-scale estimation by combining a cheap reduced-order forecast model and mixed observations of the large- and small-scale variables. This approach enables the use of a larger number of particles due to the computational savings in the forecast model. The multiscale clustered particle filter is tested for one-dimensional dispersive wave turbulence using a forecast model with model errors.

  13. Nutritional correlates and mate acquisition role of multiple sexual traits in male collared flycatchers

    NASA Astrophysics Data System (ADS)

    Hegyi, Gergely; Szöllősi, Eszter; Jenni-Eiermann, Susanne; Török, János; Eens, Marcel; Garamszegi, László Zsolt

    2010-06-01

    The information content of a sexual signal may predict its importance in a multiple signal system. Many studies have correlated sexual signal expression with the absolute levels of nutrient reserves. In contrast, the changes of nutrient reserves associated with signal expression are largely unknown in the wild due to technical limitations although they are important determinants of signal information content. We compared two visual and eight acoustic sexual traits in male collared flycatchers to see whether the nutritional correlates of expression predict the role of the signal in sexual selection. We used single point assays of plasma lipid metabolites to estimate short-term changes in nutritional state in relation to sexual trait expression during courtship. As a measure of sexual selection, we estimated the relationship with pairing latency after arrival in a 4-year dataset. Males which found a mate rapidly were characterized by large wing and forehead patches, but small song strophe complexity and small figure repertoire size. Traits more strongly related to pairing latency were also more closely related to changes in nutrient reserves. This indicates a link between signal role and information content. Small wing patches and, surprisingly, complex songs seemed to indicate poor phenotypic quality and were apparently disfavoured at mate acquisition in our population. Future studies of the information content of sexual traits, especially dynamic traits such as song, may benefit from the use of plasma metabolite profiles as non-invasive indicators of short-term changes in body condition.

  14. State estimation and prediction using clustered particle filters

    PubMed Central

    Lee, Yoonsang; Majda, Andrew J.

    2016-01-01

    Particle filtering is an essential tool to improve uncertain model predictions by incorporating noisy observational data from complex systems including non-Gaussian features. A class of particle filters, clustered particle filters, is introduced for high-dimensional nonlinear systems, which uses relatively few particles compared with the standard particle filter. The clustered particle filter captures non-Gaussian features of the true signal, which are typical in complex nonlinear dynamical systems such as geophysical systems. The method is also robust in the difficult regime of high-quality sparse and infrequent observations. The key features of the clustered particle filtering are coarse-grained localization through the clustering of the state variables and particle adjustment to stabilize the method; each observation affects only neighbor state variables through clustering and particles are adjusted to prevent particle collapse due to high-quality observations. The clustered particle filter is tested for the 40-dimensional Lorenz 96 model with several dynamical regimes including strongly non-Gaussian statistics. The clustered particle filter shows robust skill in both achieving accurate filter results and capturing non-Gaussian statistics of the true signal. It is further extended to multiscale data assimilation, which provides the large-scale estimation by combining a cheap reduced-order forecast model and mixed observations of the large- and small-scale variables. This approach enables the use of a larger number of particles due to the computational savings in the forecast model. The multiscale clustered particle filter is tested for one-dimensional dispersive wave turbulence using a forecast model with model errors. PMID:27930332

  15. Lost in transportation: Information measures and cognitive limits in multilayer navigation.

    PubMed

    Gallotti, Riccardo; Porter, Mason A; Barthelemy, Marc

    2016-02-01

    Cities and their transportation systems become increasingly complex and multimodal as they grow, and it is natural to wonder whether it is possible to quantitatively characterize our difficulty navigating in them and whether such navigation exceeds our cognitive limits. A transition between different search strategies for navigating in metropolitan maps has been observed for large, complex metropolitan networks. This evidence suggests the existence of a limit associated with cognitive overload and caused by a large amount of information that needs to be processed. In this light, we analyzed the world's 15 largest metropolitan networks and estimated the information limit for determining a trip in a transportation system to be on the order of 8 bits. Similar to the "Dunbar number," which represents a limit to the size of an individual's friendship circle, our cognitive limit suggests that maps should not consist of more than 250 connection points to be easily readable. We also show that including connections with other transportation modes dramatically increases the information needed to navigate in multilayer transportation networks. In large cities such as New York, Paris, and Tokyo, more than 80% of the trips are above the 8-bit limit. Multimodal transportation systems in large cities have thus already exceeded human cognitive limits and, consequently, the traditional view of navigation in cities has to be revised substantially.

  16. Lost in transportation: Information measures and cognitive limits in multilayer navigation

    PubMed Central

    Gallotti, Riccardo; Porter, Mason A.; Barthelemy, Marc

    2016-01-01

    Cities and their transportation systems become increasingly complex and multimodal as they grow, and it is natural to wonder whether it is possible to quantitatively characterize our difficulty navigating in them and whether such navigation exceeds our cognitive limits. A transition between different search strategies for navigating in metropolitan maps has been observed for large, complex metropolitan networks. This evidence suggests the existence of a limit associated with cognitive overload and caused by a large amount of information that needs to be processed. In this light, we analyzed the world’s 15 largest metropolitan networks and estimated the information limit for determining a trip in a transportation system to be on the order of 8 bits. Similar to the “Dunbar number,” which represents a limit to the size of an individual’s friendship circle, our cognitive limit suggests that maps should not consist of more than 250 connection points to be easily readable. We also show that including connections with other transportation modes dramatically increases the information needed to navigate in multilayer transportation networks. In large cities such as New York, Paris, and Tokyo, more than 80% of the trips are above the 8-bit limit. Multimodal transportation systems in large cities have thus already exceeded human cognitive limits and, consequently, the traditional view of navigation in cities has to be revised substantially. PMID:26989769

  17. Extraordinary Structured Noncoding RNAs Revealed by Bacterial Metagenome Analysis

    PubMed Central

    Weinberg, Zasha; Perreault, Jonathan; Meyer, Michelle M.; Breaker, Ronald R.

    2012-01-01

    Estimates of the total number of bacterial species1-3 suggest that existing DNA sequence databases carry only a tiny fraction of the total amount of DNA sequence space represented by this division of life. Indeed, environmental DNA samples have been shown to encode many previously unknown classes of proteins4 and RNAs5. Bioinformatics searches6-10 of genomic DNA from bacteria commonly identify novel noncoding RNAs (ncRNAs)10-12 such as riboswitches13,14. In rare instances, RNAs that exhibit more extensive sequence and structural conservation across a wide range of bacteria are encountered15,16. Given that large structured RNAs are known to carry out complex biochemical functions such as protein synthesis and RNA processing reactions, identifying more RNAs of great size and intricate structure is likely to reveal additional biochemical functions that can be achieved by RNA. We applied an updated computational pipeline17 to discover ncRNAs that rival the known large ribozymes in size and structural complexity or that are among the most abundant RNAs in bacteria that encode them. These RNAs would have been difficult or impossible to detect without examining environmental DNA sequences, suggesting that numerous RNAs with extraordinary size, structural complexity, or other exceptional characteristics remain to be discovered in unexplored sequence space. PMID:19956260

  18. Investigation of natural gas plume dispersion using mobile observations and large eddy simulations

    NASA Astrophysics Data System (ADS)

    Caulton, Dana R.; Li, Qi; Golston, Levi; Pan, Da; Bou-Zeid, Elie; Fitts, Jeff; Lane, Haley; Lu, Jessica; Zondlo, Mark A.

    2016-04-01

    Recent work suggests the distribution of methane emissions from fracking operations is skewed with a small percentage of emitters contributing a large proportion of the total emissions. These sites are known as 'super-emitters.' The Marcellus shale, the most productive natural gas shale field in the United States, has received less intense focus for well-level emissions and is here used as a test site for targeted analysis between current standard trace-gas advection practices and possible improvements via advanced modeling techniques. The Marcellus shale is topographically complex, making traditional techniques difficult to implement and evaluate. For many ground based mobile studies, the inverse Gaussian plume method (IGM) is used to produce emission rates. This method is best applied to well-mixed plumes from strong point sources and may not currently be well-suited for use with disperse weak sources, short-time frame measurements or data collected in complex terrain. To assess the quality of IGM results and to improve source-strength estimations, a robust study that combines observational data with a hierarchy of models of increasing complexity will be presented. The field test sites were sampled with multiple passes using a mobile lab as well as a stationary tower. This mobile lab includes a Garmin GPS unit, Vaisala weather station (WTX520), LICOR 7700 CH4 open path sensor and LICOR 7500 CO2/H2O open path sensor. The sampling tower was constructed consisting of a Metek uSonic-3 Class A sonic anemometer, and an additional LICOR 7700 and 7500. Data were recorded for at least one hour at these sites. The modeling will focus on large eddy simulations (LES) of the wind and CH4 concentration fields for these test sites. The LES model used 2 m horizontal and 1 m vertical resolution and was integrated in time for 45 min for various test sites under stable, neutral and unstable conditions. It is here considered as the reference to which various IGM approaches can be compared. Preliminary results show large variability in this region which, under the observed meteorological conditions, is determined to be a factor of 2 for IGM results. While this level of uncertainty appears adequate to identify super-emitters under most circumstances, there is large uncertainty on individual measurements. LES can provide insights into the expected variability and its sources and into sampling patterns that will allow more robust error estimates.

  19. A comparison of three approaches to compute the effective Reynolds number of the implicit large-eddy simulations

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Zhou, Ye; Thornber, Ben

    2016-04-12

    Here, the implicit large-eddy simulation (ILES) has been utilized as an effective approach for calculating many complex flows at high Reynolds number flows. Richtmyer–Meshkov instability (RMI) induced flow can be viewed as a homogeneous decaying turbulence (HDT) after the passage of the shock. In this article, a critical evaluation of three methods for estimating the effective Reynolds number and the effective kinematic viscosity is undertaken utilizing high-resolution ILES data. Effective Reynolds numbers based on the vorticity and dissipation rate, or the integral and inner-viscous length scales, are found to be the most self-consistent when compared to the expected phenomenology andmore » wind tunnel experiments.« less

  20. Scope Complexity Options Risks Excursions (SCORE) Factor Mathematical Description.

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Gearhart, Jared Lee; Samberson, Jonell Nicole; Shettigar, Subhasini

    The purpose of the Scope, Complexity, Options, Risks, Excursions (SCORE) model is to estimate the relative complexity of design variants of future warhead options, resulting in scores. SCORE factors extend this capability by providing estimates of complexity relative to a base system (i.e., all design options are normalized to one weapon system). First, a clearly defined set of scope elements for a warhead option is established. The complexity of each scope element is estimated by Subject Matter Experts (SMEs), including a level of uncertainty, relative to a specific reference system. When determining factors, complexity estimates for a scope element canmore » be directly tied to the base system or chained together via comparable scope elements in a string of reference systems that ends with the base system. The SCORE analysis process is a growing multi-organizational Nuclear Security Enterprise (NSE) effort, under the management of the NA-12 led Enterprise Modeling and Analysis Consortium (EMAC). Historically, it has provided the data elicitation, integration, and computation needed to support the out-year Life Extension Program (LEP) cost estimates included in the Stockpile Stewardship Management Plan (SSMP).« less

  1. Regional Evaluation of Groundwater Age Distributions Using Lumped Parameter Models with Large, Sparse Datasets: Example from the Central Valley, California, USA

    NASA Astrophysics Data System (ADS)

    Jurgens, B. C.; Bohlke, J. K.; Voss, S.; Fram, M. S.; Esser, B.

    2015-12-01

    Tracer-based, lumped parameter models (LPMs) are an appealing way to estimate the distribution of age for groundwater because the cost of sampling wells is often less than building numerical groundwater flow models sufficiently complex to provide groundwater age distributions. In practice, however, tracer datasets are often incomplete because of anthropogenic or terrigenic contamination of tracers, or analytical limitations. While age interpretations using such datsets can have large uncertainties, it may still be possible to identify key parts of the age distribution if LPMs are carefully chosen to match hydrogeologic conceptualization and the degree of age mixing is reasonably estimated. We developed a systematic approach for evaluating groundwater age distributions using LPMs with a large but incomplete set of tracer data (3H, 3Hetrit, 14C, and CFCs) from 535 wells, mostly used for public supply, in the Central Valley, California, USA that were sampled by the USGS for the California State Water Resources Control Board Groundwater Ambient Monitoring and Assessment or the USGS National Water Quality Assessment Programs. In addition to mean ages, LPMs gave estimates of unsaturated zone travel times, recharge rates for pre- and post-development groundwater, the degree of age mixing in wells, proportion of young water (<60 yrs), and the depth of the boundary between post-development and predevelopment groundwater throughout the Central Valley. Age interpretations were evaluated by comparing past nitrate trends with LPM predicted trends, and whether the presence or absence of anthropogenic organic compounds was consistent with model results. This study illustrates a practical approach for assessing groundwater age information at a large scale to reveal important characteristics about the age structure of a major aquifer, and of the water supplies being derived from it.

  2. Titan dune heights retrieval by using Cassini Radar Altimeter

    NASA Astrophysics Data System (ADS)

    Mastrogiuseppe, M.; Poggiali, V.; Seu, R.; Martufi, R.; Notarnicola, C.

    2014-02-01

    The Cassini Radar is a Ku band multimode instrument capable of providing topographic and mapping information. During several of the 93 Titan fly-bys performed by Cassini, the radar collected a large amount of data observing many dune fields in multiple modes such as SAR, Altimeter, Scatterometer and Radiometer. Understanding dune characteristics, such as shape and height, will reveal important clues on Titan's climatic and geological history providing a better understanding of aeolian processes on Earth. Dunes are believed to be sculpted by the action of the wind, weak at the surface but still able to activate the process of sand-sized particle transport. This work aims to estimate dunes height by modeling the shape of the real Cassini Radar Altimeter echoes. Joint processing of SAR/Altimeter data has been adopted to localize the altimeter footprints overlapping dune fields excluding non-dune features. The height of the dunes was estimated by applying Maximum Likelihood Estimation along with a non-coherent electromagnetic (EM) echo model, thus comparing the real averaged waveform with the theoretical curves. Such analysis has been performed over the Fensal dune field observed during the T30 flyby (May 2007). As a result we found that the estimated dunes' peak to trough heights difference was in the order of 60-120 m. Estimation accuracy and robustness of the MLE for different complex scenarios was assessed via radar simulations and Monte-Carlo approach. We simulated dunes-interdunes different composition and roughness for a large set of values verifying that, in the range of possible Titan environment conditions, these two surface parameters have weak effects on our estimates of standard dune heights deviation. Results presented here are the first part of a study that will cover all Titan's sand seas.

  3. Estimation of the interference coupling into cables within electrically large multiroom structures

    NASA Astrophysics Data System (ADS)

    Keghie, J.; Kanyou Nana, R.; Schetelig, B.; Potthast, S.; Dickmann, S.

    2010-10-01

    Communication cables are used to transfer data between components of a system. As a part of the EMC analysis of complex systems, it is necessary to determine which level of interference can be expected at the input of connected devices due to the coupling into the irradiated cable. For electrically large systems consisting of several rooms with cables connecting components located in different rooms, an estimation of the coupled disturbances inside cables using commercial field computation software is often not feasible without several restrictions. In many cases, this is related to the non-availability of computing memory and processing power needed for the computation. In this paper, we are going to show that, starting from a topological analysis of the entire system, weak coupling paths within the system can be can be identified. By neglecting these coupling paths and using the transmission line approach, the original system will be simplified so that a simpler estimation is possible. Using the example of a system which is composed of two rooms, multiple apertures, and a network cable located in both chambers, it is shown that an estimation of the coupled disturbances due to external electromagnetic sources is feasible with this approach. Starting from an incident electromagnetic field, we determine transfer functions describing the coupling means (apertures, cables). Using these transfer functions and the knowledge of the weak coupling paths above, a decision is taken regarding the means for paths that can be neglected during the estimation. The estimation of the coupling into the cable is then made while taking only paths with strong coupling into account. The remaining part of the wiring harness in areas with weak coupling is represented by its input impedance. A comparison with the original network shows a good agreement.

  4. Performance of Random Effects Model Estimators under Complex Sampling Designs

    ERIC Educational Resources Information Center

    Jia, Yue; Stokes, Lynne; Harris, Ian; Wang, Yan

    2011-01-01

    In this article, we consider estimation of parameters of random effects models from samples collected via complex multistage designs. Incorporation of sampling weights is one way to reduce estimation bias due to unequal probabilities of selection. Several weighting methods have been proposed in the literature for estimating the parameters of…

  5. Estimating Children's Soil/Dust Ingestion Rates through ...

    EPA Pesticide Factsheets

    Background: Soil/dust ingestion rates are important variables in assessing children’s health risks in contaminated environments. Current estimates are based largely on soil tracer methodology, which is limited by analytical uncertainty, small sample size, and short study duration. Objectives: The objective was to estimate site-specific soil/dust ingestion rates through reevaluation of the lead absorption dose–response relationship using new bioavailability data from the Bunker Hill Mining and Metallurgical Complex Superfund Site (BHSS) in Idaho, USA. Methods: The U.S. Environmental Protection Agency (EPA) in vitro bioavailability methodology was applied to archived BHSS soil and dust samples. Using age-specific biokinetic slope factors, we related bioavailable lead from these sources to children’s blood lead levels (BLLs) monitored during cleanup from 1988 through 2002. Quantitative regression analyses and exposure assessment guidance were used to develop candidate soil/dust source partition scenarios estimating lead intake, allowing estimation of age-specific soil/dust ingestion rates. These ingestion rate and bioavailability estimates were simultaneously applied to the U.S. EPA Integrated Exposure Uptake Biokinetic Model for Lead in Children to determine those combinations best approximating observed BLLs. Results: Absolute soil and house dust bioavailability averaged 33% (SD ± 4%) and 28% (SD ± 6%), respectively. Estimated BHSS age-specific soil/du

  6. Implicit Particle Filter for Power System State Estimation with Large Scale Renewable Power Integration.

    NASA Astrophysics Data System (ADS)

    Uzunoglu, B.; Hussaini, Y.

    2017-12-01

    Implicit Particle Filter is a sequential Monte Carlo method for data assimilation that guides the particles to the high-probability by an implicit step . It optimizes a nonlinear cost function which can be inherited from legacy assimilation routines . Dynamic state estimation for almost real-time applications in power systems are becomingly increasingly more important with integration of variable wind and solar power generation. New advanced state estimation tools that will replace the old generation state estimation in addition to having a general framework of complexities should be able to address the legacy software and able to integrate the old software in a mathematical framework while allowing the power industry need for a cautious and evolutionary change in comparison to a complete revolutionary approach while addressing nonlinearity and non-normal behaviour. This work implements implicit particle filter as a state estimation tool for the estimation of the states of a power system and presents the first implicit particle filter application study on a power system state estimation. The implicit particle filter is introduced into power systems and the simulations are presented for a three-node benchmark power system . The performance of the filter on the presented problem is analyzed and the results are presented.

  7. Photochemical grid model performance with varying horizontal grid resolution and sub-grid plume treatment for the Martins Creek near-field SO2 study

    NASA Astrophysics Data System (ADS)

    Baker, Kirk R.; Hawkins, Andy; Kelly, James T.

    2014-12-01

    Near source modeling is needed to assess primary and secondary pollutant impacts from single sources and single source complexes. Source-receptor relationships need to be resolved from tens of meters to tens of kilometers. Dispersion models are typically applied for near-source primary pollutant impacts but lack complex photochemistry. Photochemical models provide a realistic chemical environment but are typically applied using grid cell sizes that may be larger than the distance between sources and receptors. It is important to understand the impacts of grid resolution and sub-grid plume treatments on photochemical modeling of near-source primary pollution gradients. Here, the CAMx photochemical grid model is applied using multiple grid resolutions and sub-grid plume treatment for SO2 and compared with a receptor mesonet largely impacted by nearby sources approximately 3-17 km away in a complex terrain environment. Measurements are compared with model estimates of SO2 at 4- and 1-km resolution, both with and without sub-grid plume treatment and inclusion of finer two-way grid nests. Annual average estimated SO2 mixing ratios are highest nearest the sources and decrease as distance from the sources increase. In general, CAMx estimates of SO2 do not compare well with the near-source observations when paired in space and time. Given the proximity of these sources and receptors, accuracy in wind vector estimation is critical for applications that pair pollutant predictions and observations in time and space. In typical permit applications, predictions and observations are not paired in time and space and the entire distributions of each are directly compared. Using this approach, model estimates using 1-km grid resolution best match the distribution of observations and are most comparable to similar studies that used dispersion and Lagrangian modeling systems. Model-estimated SO2 increases as grid cell size decreases from 4 km to 250 m. However, it is notable that the 1-km model estimates using 1-km meteorological model input are higher than the 1-km model simulation that used interpolated 4-km meteorology. The inclusion of sub-grid plume treatment did not improve model skill in predicting SO2 in time and space and generally acts to keep emitted mass aloft.

  8. Numerical Demons in Monte Carlo Estimation of Bayesian Model Evidence with Application to Soil Respiration Models

    NASA Astrophysics Data System (ADS)

    Elshall, A. S.; Ye, M.; Niu, G. Y.; Barron-Gafford, G.

    2016-12-01

    Bayesian multimodel inference is increasingly being used in hydrology. Estimating Bayesian model evidence (BME) is of central importance in many Bayesian multimodel analysis such as Bayesian model averaging and model selection. BME is the overall probability of the model in reproducing the data, accounting for the trade-off between the goodness-of-fit and the model complexity. Yet estimating BME is challenging, especially for high dimensional problems with complex sampling space. Estimating BME using the Monte Carlo numerical methods is preferred, as the methods yield higher accuracy than semi-analytical solutions (e.g. Laplace approximations, BIC, KIC, etc.). However, numerical methods are prone the numerical demons arising from underflow of round off errors. Although few studies alluded to this issue, to our knowledge this is the first study that illustrates these numerical demons. We show that the precision arithmetic can become a threshold on likelihood values and Metropolis acceptance ratio, which results in trimming parameter regions (when likelihood function is less than the smallest floating point number that a computer can represent) and corrupting of the empirical measures of the random states of the MCMC sampler (when using log-likelihood function). We consider two of the most powerful numerical estimators of BME that are the path sampling method of thermodynamic integration (TI) and the importance sampling method of steppingstone sampling (SS). We also consider the two most widely used numerical estimators, which are the prior sampling arithmetic mean (AS) and posterior sampling harmonic mean (HM). We investigate the vulnerability of these four estimators to the numerical demons. Interesting, the most biased estimator, namely the HM, turned out to be the least vulnerable. While it is generally assumed that AM is a bias-free estimator that will always approximate the true BME by investing in computational effort, we show that arithmetic underflow can hamper AM resulting in severe underestimation of BME. TI turned out to be the most vulnerable, resulting in BME overestimation. Finally, we show how SS can be largely invariant to rounding errors, yielding the most accurate and computational efficient results. These research results are useful for MC simulations to estimate Bayesian model evidence.

  9. Charge Redistribution in the β-NAPHTHOL-WATER Complex as Measured by High Resolution Stark Spectroscopy in the Gas Phase.

    NASA Astrophysics Data System (ADS)

    Fleisher, Adam J.; Pratt, David W.; Cembran, Alessandro; Gao, Jiali

    2010-06-01

    The extensively studied photoacid β-naphthol exhibits a large decrease in pKa upon irradiation with ultraviolet light, in the condensed phase. β-naphthol is almost 10 million times more acidic in the excited electronic state, compared to the ground state. Motivated by this fact, we report here the measurement of the electronic dipole moments of the β-naphthol-water complex in both electronic states, from which estimates of the charge transfer from solute to solvent in both states will be made. Comparisons to ab initio and density functional theory calculations will also be reported. N. Mataga and T. Kubota, Molecular Interactions and Electronic Spectra (Marcel Dekker, New York, 1970). Y. Mo, J. Gao, S.D. Peyerimhoff, J. Chem. Phys. 112, 5530 (2000).

  10. Fast optimization algorithms and the cosmological constant

    NASA Astrophysics Data System (ADS)

    Bao, Ning; Bousso, Raphael; Jordan, Stephen; Lackey, Brad

    2017-11-01

    Denef and Douglas have observed that in certain landscape models the problem of finding small values of the cosmological constant is a large instance of a problem that is hard for the complexity class NP (Nondeterministic Polynomial-time). The number of elementary operations (quantum gates) needed to solve this problem by brute force search exceeds the estimated computational capacity of the observable Universe. Here we describe a way out of this puzzling circumstance: despite being NP-hard, the problem of finding a small cosmological constant can be attacked by more sophisticated algorithms whose performance vastly exceeds brute force search. In fact, in some parameter regimes the average-case complexity is polynomial. We demonstrate this by explicitly finding a cosmological constant of order 10-120 in a randomly generated 1 09-dimensional Arkani-Hamed-Dimopoulos-Kachru landscape.

  11. Bayesian WLS/GLS regression for regional skewness analysis for regions with large crest stage gage networks

    USGS Publications Warehouse

    Veilleux, Andrea G.; Stedinger, Jery R.; Eash, David A.

    2012-01-01

    This paper summarizes methodological advances in regional log-space skewness analyses that support flood-frequency analysis with the log Pearson Type III (LP3) distribution. A Bayesian Weighted Least Squares/Generalized Least Squares (B-WLS/B-GLS) methodology that relates observed skewness coefficient estimators to basin characteristics in conjunction with diagnostic statistics represents an extension of the previously developed B-GLS methodology. B-WLS/B-GLS has been shown to be effective in two California studies. B-WLS/B-GLS uses B-WLS to generate stable estimators of model parameters and B-GLS to estimate the precision of those B-WLS regression parameters, as well as the precision of the model. The study described here employs this methodology to develop a regional skewness model for the State of Iowa. To provide cost effective peak-flow data for smaller drainage basins in Iowa, the U.S. Geological Survey operates a large network of crest stage gages (CSGs) that only record flow values above an identified recording threshold (thus producing a censored data record). CSGs are different from continuous-record gages, which record almost all flow values and have been used in previous B-GLS and B-WLS/B-GLS regional skewness studies. The complexity of analyzing a large CSG network is addressed by using the B-WLS/B-GLS framework along with the Expected Moments Algorithm (EMA). Because EMA allows for the censoring of low outliers, as well as the use of estimated interval discharges for missing, censored, and historic data, it complicates the calculations of effective record length (and effective concurrent record length) used to describe the precision of sample estimators because the peak discharges are no longer solely represented by single values. Thus new record length calculations were developed. The regional skewness analysis for the State of Iowa illustrates the value of the new B-WLS/BGLS methodology with these new extensions.

  12. Large Scale Screening of Low Cost Ferritic Steel Designs For Advanced Ultra Supercritical Boiler Using First Principles Methods

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Ouyang, Lizhi

    Advanced Ultra Supercritical Boiler (AUSC) requires materials that can operate in corrosive environment at temperature and pressure as high as 760°C (or 1400°F) and 5000psi, respectively, while at the same time maintain good ductility at low temperature. We develop automated simulation software tools to enable fast large scale screening studies of candidate designs. While direct evaluation of creep rupture strength and ductility are currently not feasible, properties such as energy, elastic constants, surface energy, interface energy, and stack fault energy can be used to assess their relative ductility and creeping strength. We implemented software to automate the complex calculations tomore » minimize human inputs in the tedious screening studies which involve model structures generation, settings for first principles calculations, results analysis and reporting. The software developed in the project and library of computed mechanical properties of phases found in ferritic steels, many are complex solid solutions estimated for the first time, will certainly help the development of low cost ferritic steel for AUSC.« less

  13. Thiocyanato Chromium (III) Complexes: Separation by Paper Electrophoresis and Estimate of Stability Constants

    ERIC Educational Resources Information Center

    Larsen, Erik; Eriksen, J.

    1975-01-01

    Describes an experiment wherein the student can demonstrate the existence of all the thiocyanato chromium complexes, estimate the stepwise formation constants, demonstrate the robustness of chromium III complexes, and show the principles of paper electrophoresis. (GS)

  14. Self-optimized construction of transition rate matrices from accelerated atomistic simulations with Bayesian uncertainty quantification

    NASA Astrophysics Data System (ADS)

    Swinburne, Thomas D.; Perez, Danny

    2018-05-01

    A massively parallel method to build large transition rate matrices from temperature-accelerated molecular dynamics trajectories is presented. Bayesian Markov model analysis is used to estimate the expected residence time in the known state space, providing crucial uncertainty quantification for higher-scale simulation schemes such as kinetic Monte Carlo or cluster dynamics. The estimators are additionally used to optimize where exploration is performed and the degree of temperature acceleration on the fly, giving an autonomous, optimal procedure to explore the state space of complex systems. The method is tested against exactly solvable models and used to explore the dynamics of C15 interstitial defects in iron. Our uncertainty quantification scheme allows for accurate modeling of the evolution of these defects over timescales of several seconds.

  15. Integrating Efficiency of Industry Processes and Practices Alongside Technology Effectiveness in Space Transportation Cost Modeling and Analysis

    NASA Technical Reports Server (NTRS)

    Zapata, Edgar

    2012-01-01

    This paper presents past and current work in dealing with indirect industry and NASA costs when providing cost estimation or analysis for NASA projects and programs. Indirect costs, when defined as those costs in a project removed from the actual hardware or software hands-on labor; makes up most of the costs of today's complex large scale NASA space/industry projects. This appears to be the case across phases from research into development into production and into the operation of the system. Space transportation is the case of interest here. Modeling and cost estimation as a process rather than a product will be emphasized. Analysis as a series of belief systems in play among decision makers and decision factors will also be emphasized to provide context.

  16. Angular-domain scattering interferometry.

    PubMed

    Shipp, Dustin W; Qian, Ruobing; Berger, Andrew J

    2013-11-15

    We present an angular-scattering optical method that is capable of measuring the mean size of scatterers in static ensembles within a field of view less than 20 μm in diameter. Using interferometry, the method overcomes the inability of intensity-based models to tolerate the large speckle grains associated with such small illumination areas. By first estimating each scatterer's location, the method can model between-scatterer interference as well as traditional single-particle Mie scattering. Direct angular-domain measurements provide finer angular resolution than digitally transformed image-plane recordings. This increases sensitivity to size-dependent scattering features, enabling more robust size estimates. The sensitivity of these angular-scattering measurements to various sizes of polystyrene beads is demonstrated. Interferometry also allows recovery of the full complex scattered field, including a size-dependent phase profile in the angular-scattering pattern.

  17. An Efficient VLSI Architecture of the Enhanced Three Step Search Algorithm

    NASA Astrophysics Data System (ADS)

    Biswas, Baishik; Mukherjee, Rohan; Saha, Priyabrata; Chakrabarti, Indrajit

    2016-09-01

    The intense computational complexity of any video codec is largely due to the motion estimation unit. The Enhanced Three Step Search is a popular technique that can be adopted for fast motion estimation. This paper proposes a novel VLSI architecture for the implementation of the Enhanced Three Step Search Technique. A new addressing mechanism has been introduced which enhances the speed of operation and reduces the area requirements. The proposed architecture when implemented in Verilog HDL on Virtex-5 Technology and synthesized using Xilinx ISE Design Suite 14.1 achieves a critical path delay of 4.8 ns while the area comes out to be 2.9K gate equivalent. It can be incorporated in commercial devices like smart-phones, camcorders, video conferencing systems etc.

  18. Multivariate Meta-Analysis of Genetic Association Studies: A Simulation Study

    PubMed Central

    Neupane, Binod; Beyene, Joseph

    2015-01-01

    In a meta-analysis with multiple end points of interests that are correlated between or within studies, multivariate approach to meta-analysis has a potential to produce more precise estimates of effects by exploiting the correlation structure between end points. However, under random-effects assumption the multivariate estimation is more complex (as it involves estimation of more parameters simultaneously) than univariate estimation, and sometimes can produce unrealistic parameter estimates. Usefulness of multivariate approach to meta-analysis of the effects of a genetic variant on two or more correlated traits is not well understood in the area of genetic association studies. In such studies, genetic variants are expected to roughly maintain Hardy-Weinberg equilibrium within studies, and also their effects on complex traits are generally very small to modest and could be heterogeneous across studies for genuine reasons. We carried out extensive simulation to explore the comparative performance of multivariate approach with most commonly used univariate inverse-variance weighted approach under random-effects assumption in various realistic meta-analytic scenarios of genetic association studies of correlated end points. We evaluated the performance with respect to relative mean bias percentage, and root mean square error (RMSE) of the estimate and coverage probability of corresponding 95% confidence interval of the effect for each end point. Our simulation results suggest that multivariate approach performs similarly or better than univariate method when correlations between end points within or between studies are at least moderate and between-study variation is similar or larger than average within-study variation for meta-analyses of 10 or more genetic studies. Multivariate approach produces estimates with smaller bias and RMSE especially for the end point that has randomly or informatively missing summary data in some individual studies, when the missing data in the endpoint are imputed with null effects and quite large variance. PMID:26196398

  19. Multivariate Meta-Analysis of Genetic Association Studies: A Simulation Study.

    PubMed

    Neupane, Binod; Beyene, Joseph

    2015-01-01

    In a meta-analysis with multiple end points of interests that are correlated between or within studies, multivariate approach to meta-analysis has a potential to produce more precise estimates of effects by exploiting the correlation structure between end points. However, under random-effects assumption the multivariate estimation is more complex (as it involves estimation of more parameters simultaneously) than univariate estimation, and sometimes can produce unrealistic parameter estimates. Usefulness of multivariate approach to meta-analysis of the effects of a genetic variant on two or more correlated traits is not well understood in the area of genetic association studies. In such studies, genetic variants are expected to roughly maintain Hardy-Weinberg equilibrium within studies, and also their effects on complex traits are generally very small to modest and could be heterogeneous across studies for genuine reasons. We carried out extensive simulation to explore the comparative performance of multivariate approach with most commonly used univariate inverse-variance weighted approach under random-effects assumption in various realistic meta-analytic scenarios of genetic association studies of correlated end points. We evaluated the performance with respect to relative mean bias percentage, and root mean square error (RMSE) of the estimate and coverage probability of corresponding 95% confidence interval of the effect for each end point. Our simulation results suggest that multivariate approach performs similarly or better than univariate method when correlations between end points within or between studies are at least moderate and between-study variation is similar or larger than average within-study variation for meta-analyses of 10 or more genetic studies. Multivariate approach produces estimates with smaller bias and RMSE especially for the end point that has randomly or informatively missing summary data in some individual studies, when the missing data in the endpoint are imputed with null effects and quite large variance.

  20. Applying transfer matrix method to the estimation of the modal characteristics of the NASA Mini-Mass Truss

    NASA Technical Reports Server (NTRS)

    Shen, Ji-Yao; Taylor, Lawrence W., Jr.

    1994-01-01

    It is beneficial to use a distributed parameter model for large space structures because the approach minimizes the number of model parameters. Holzer's transfer matrix method provides a useful means to simplify and standardize the procedure for solving the system of partial differential equations. Any large space structures can be broken down into sub-structures with simple elastic and dynamical properties. For each single element, such as beam, tether, or rigid body, we can derive the corresponding transfer matrix. Combining these elements' matrices enables the solution of the global system equations. The characteristics equation can then be formed by satisfying the appropriate boundary conditions. Then natural frequencies and mode shapes can be determined by searching the roots of the characteristic equation at frequencies within the range of interest. This paper applies this methodology, and the maximum likelihood estimation method, to refine the modal characteristics of the NASA Mini-Mast Truss by successively matching the theoretical response to the test data of the truss. The method is being applied to more complex configurations.

  1. Design of DNA pooling to allow incorporation of covariates in rare variants analysis.

    PubMed

    Guan, Weihua; Li, Chun

    2014-01-01

    Rapid advances in next-generation sequencing technologies facilitate genetic association studies of an increasingly wide array of rare variants. To capture the rare or less common variants, a large number of individuals will be needed. However, the cost of a large scale study using whole genome or exome sequencing is still high. DNA pooling can serve as a cost-effective approach, but with a potential limitation that the identity of individual genomes would be lost and therefore individual characteristics and environmental factors could not be adjusted in association analysis, which may result in power loss and a biased estimate of genetic effect. For case-control studies, we propose a design strategy for pool creation and an analysis strategy that allows covariate adjustment, using multiple imputation technique. Simulations show that our approach can obtain reasonable estimate for genotypic effect with only slight loss of power compared to the much more expensive approach of sequencing individual genomes. Our design and analysis strategies enable more powerful and cost-effective sequencing studies of complex diseases, while allowing incorporation of covariate adjustment.

  2. Long-ranged contributions to solvation free energies from theory and short-ranged models

    PubMed Central

    Remsing, Richard C.; Liu, Shule; Weeks, John D.

    2016-01-01

    Long-standing problems associated with long-ranged electrostatic interactions have plagued theory and simulation alike. Traditional lattice sum (Ewald-like) treatments of Coulomb interactions add significant overhead to computer simulations and can produce artifacts from spurious interactions between simulation cell images. These subtle issues become particularly apparent when estimating thermodynamic quantities, such as free energies of solvation in charged and polar systems, to which long-ranged Coulomb interactions typically make a large contribution. In this paper, we develop a framework for determining very accurate solvation free energies of systems with long-ranged interactions from models that interact with purely short-ranged potentials. Our approach is generally applicable and can be combined with existing computational and theoretical techniques for estimating solvation thermodynamics. We demonstrate the utility of our approach by examining the hydration thermodynamics of hydrophobic and ionic solutes and the solvation of a large, highly charged colloid that exhibits overcharging, a complex nonlinear electrostatic phenomenon whereby counterions from the solvent effectively overscreen and locally invert the integrated charge of the solvated object. PMID:26929375

  3. Heritability maps of human face morphology through large-scale automated three-dimensional phenotyping

    NASA Astrophysics Data System (ADS)

    Tsagkrasoulis, Dimosthenis; Hysi, Pirro; Spector, Tim; Montana, Giovanni

    2017-04-01

    The human face is a complex trait under strong genetic control, as evidenced by the striking visual similarity between twins. Nevertheless, heritability estimates of facial traits have often been surprisingly low or difficult to replicate. Furthermore, the construction of facial phenotypes that correspond to naturally perceived facial features remains largely a mystery. We present here a large-scale heritability study of face geometry that aims to address these issues. High-resolution, three-dimensional facial models have been acquired on a cohort of 952 twins recruited from the TwinsUK registry, and processed through a novel landmarking workflow, GESSA (Geodesic Ensemble Surface Sampling Algorithm). The algorithm places thousands of landmarks throughout the facial surface and automatically establishes point-wise correspondence across faces. These landmarks enabled us to intuitively characterize facial geometry at a fine level of detail through curvature measurements, yielding accurate heritability maps of the human face (www.heritabilitymaps.info).

  4. Multilocus phylogeny and coalescent species delimitation in Kotschy's gecko, Mediodactylus kotschyi: Hidden diversity and cryptic species.

    PubMed

    Kotsakiozi, Panayiota; Jablonski, Daniel; Ilgaz, Çetin; Kumlutaş, Yusuf; Avcı, Aziz; Meiri, Shai; Itescu, Yuval; Kukushkin, Oleg; Gvoždík, Václav; Scillitani, Giovanni; Roussos, Stephanos A; Jandzik, David; Kasapidis, Panagiotis; Lymberakis, Petros; Poulakakis, Nikos

    2018-08-01

    Kotschy's Gecko, Mediodactylus kotschyi, is a small gecko native to southeastern Europe and the Levant. It displays great morphological variation with a large number of morphologically recognized subspecies. However, it has been suggested that it constitutes a species complex of several yet unrecognized species. In this study, we used multilocus sequence data (three mitochondrial and three nuclear gene fragments) to estimate the phylogenetic relationships of 174 specimens from 129 sampling localities, covering a substantial part of the distribution range of the species. Our results revealed high genetic diversity of M. kotschyi populations and contributed to our knowledge about the phylogenetic relationships and the estimation of the divergence times between them. Diversification within M. kotschyi began approximately 15 million years ago (Mya) in the Middle Miocene, whereas the diversification within most of the major clades have been occurred in the last 5 Mya. Species delimitation analysis suggests there exists five species within the complex, and we propose to tentatively recognize the following taxa as full species: M. kotschyi (mainland Balkans, most of Aegean islands, and Italy), M. orientalis (Levant, Cyprus, southern Anatolia, and south-eastern Aegean islands), M. danilewskii (Black Sea region and south-western Anatolia), M. bartoni (Crete), and M. oertzeni (southern Dodecanese Islands). This newly recognized diversity underlines the complex biogeographical history of the Eastern Mediterranean region. Copyright © 2018 Elsevier Inc. All rights reserved.

  5. Prediction of TF target sites based on atomistic models of protein-DNA complexes

    PubMed Central

    Angarica, Vladimir Espinosa; Pérez, Abel González; Vasconcelos, Ana T; Collado-Vides, Julio; Contreras-Moreira, Bruno

    2008-01-01

    Background The specific recognition of genomic cis-regulatory elements by transcription factors (TFs) plays an essential role in the regulation of coordinated gene expression. Studying the mechanisms determining binding specificity in protein-DNA interactions is thus an important goal. Most current approaches for modeling TF specific recognition rely on the knowledge of large sets of cognate target sites and consider only the information contained in their primary sequence. Results Here we describe a structure-based methodology for predicting sequence motifs starting from the coordinates of a TF-DNA complex. Our algorithm combines information regarding the direct and indirect readout of DNA into an atomistic statistical model, which is used to estimate the interaction potential. We first measure the ability of our method to correctly estimate the binding specificities of eight prokaryotic and eukaryotic TFs that belong to different structural superfamilies. Secondly, the method is applied to two homology models, finding that sampling of interface side-chain rotamers remarkably improves the results. Thirdly, the algorithm is compared with a reference structural method based on contact counts, obtaining comparable predictions for the experimental complexes and more accurate sequence motifs for the homology models. Conclusion Our results demonstrate that atomic-detail structural information can be feasibly used to predict TF binding sites. The computational method presented here is universal and might be applied to other systems involving protein-DNA recognition. PMID:18922190

  6. Modeling Mediterranean forest structure using airborne laser scanning data

    NASA Astrophysics Data System (ADS)

    Bottalico, Francesca; Chirici, Gherardo; Giannini, Raffaello; Mele, Salvatore; Mura, Matteo; Puxeddu, Michele; McRoberts, Ronald E.; Valbuena, Ruben; Travaglini, Davide

    2017-05-01

    The conservation of biological diversity is recognized as a fundamental component of sustainable development, and forests contribute greatly to its preservation. Structural complexity increases the potential biological diversity of a forest by creating multiple niches that can host a wide variety of species. To facilitate greater understanding of the contributions of forest structure to forest biological diversity, we modeled relationships between 14 forest structure variables and airborne laser scanning (ALS) data for two Italian study areas representing two common Mediterranean forests, conifer plantations and coppice oaks subjected to irregular intervals of unplanned and non-standard silvicultural interventions. The objectives were twofold: (i) to compare model prediction accuracies when using two types of ALS metrics, echo-based metrics and canopy height model (CHM)-based metrics, and (ii) to construct inferences in the form of confidence intervals for large area structural complexity parameters. Our results showed that the effects of the two study areas on accuracies were greater than the effects of the two types of ALS metrics. In particular, accuracies were less for the more complex study area in terms of species composition and forest structure. However, accuracies achieved using the echo-based metrics were only slightly greater than when using the CHM-based metrics, thus demonstrating that both options yield reliable and comparable results. Accuracies were greatest for dominant height (Hd) (R2 = 0.91; RMSE% = 8.2%) and mean height weighted by basal area (R2 = 0.83; RMSE% = 10.5%) when using the echo-based metrics, 99th percentile of the echo height distribution and interquantile distance. For the forested area, the generalized regression (GREG) estimate of mean Hd was similar to the simple random sampling (SRS) estimate, 15.5 m for GREG and 16.2 m SRS. Further, the GREG estimator with standard error of 0.10 m was considerable more precise than the SRS estimator with standard error of 0.69 m.

  7. Why do we differ in number sense? Evidence from a genetically sensitive investigation☆

    PubMed Central

    Tosto, M.G.; Petrill, S.A.; Halberda, J.; Trzaskowski, M.; Tikhomirova, T.N.; Bogdanova, O.Y.; Ly, R.; Wilmer, J.B.; Naiman, D.Q.; Germine, L.; Plomin, R.; Kovas, Y.

    2014-01-01

    Basic intellectual abilities of quantity and numerosity estimation have been detected across animal species. Such abilities are referred to as ‘number sense’. For human species, individual differences in number sense are detectable early in life, persist in later development, and relate to general intelligence. The origins of these individual differences are unknown. To address this question, we conducted the first large-scale genetically sensitive investigation of number sense, assessing numerosity discrimination abilities in 837 pairs of monozygotic and 1422 pairs of dizygotic 16-year-old twin pairs. Univariate genetic analysis of the twin data revealed that number sense is modestly heritable (32%), with individual differences being largely explained by non-shared environmental influences (68%) and no contribution from shared environmental factors. Sex-Limitation model fitting revealed no differences between males and females in the etiology of individual differences in number sense abilities. We also carried out Genome-wide Complex Trait Analysis (GCTA) that estimates the population variance explained by additive effects of DNA differences among unrelated individuals. For 1118 unrelated individuals in our sample with genotyping information on 1.7 million DNA markers, GCTA estimated zero heritability for number sense, unlike other cognitive abilities in the same twin study where the GCTA heritability estimates were about 25%. The low heritability of number sense, observed in this study, is consistent with the directional selection explanation whereby additive genetic variance for evolutionary important traits is reduced. PMID:24696527

  8. Local Ancestry Inference in a Large US-Based Hispanic/Latino Study: Hispanic Community Health Study/Study of Latinos (HCHS/SOL).

    PubMed

    Browning, Sharon R; Grinde, Kelsey; Plantinga, Anna; Gogarten, Stephanie M; Stilp, Adrienne M; Kaplan, Robert C; Avilés-Santa, M Larissa; Browning, Brian L; Laurie, Cathy C

    2016-06-01

    We estimated local ancestry on the autosomes and X chromosome in a large US-based study of 12,793 Hispanic/Latino individuals using the RFMix method, and we compared different reference panels and approaches to local ancestry estimation on the X chromosome by means of Mendelian inconsistency rates as a proxy for accuracy. We developed a novel and straightforward approach to performing ancestry-specific PCA after finding artifactual behavior in the results from an existing approach. Using the ancestry-specific PCA, we found significant population structure within African, European, and Amerindian ancestries in the Hispanic/Latino individuals in our study. In the African ancestral component of the admixed individuals, individuals whose grandparents were from Central America clustered separately from individuals whose grandparents were from the Caribbean, and also from reference Yoruba and Mandenka West African individuals. In the European component, individuals whose grandparents were from Puerto Rico diverged partially from other background groups. In the Amerindian ancestral component, individuals clustered into multiple different groups depending on the grandparental country of origin. Therefore, local ancestry estimation provides further insight into the complex genetic structure of US Hispanic/Latino populations, which must be properly accounted for in genotype-phenotype association studies. It also provides a basis for admixture mapping and ancestry-specific allele frequency estimation, which are useful in the identification of risk factors for disease. Copyright © 2016 Browning et al.

  9. BAYESIAN LARGE-SCALE MULTIPLE REGRESSION WITH SUMMARY STATISTICS FROM GENOME-WIDE ASSOCIATION STUDIES1

    PubMed Central

    Zhu, Xiang; Stephens, Matthew

    2017-01-01

    Bayesian methods for large-scale multiple regression provide attractive approaches to the analysis of genome-wide association studies (GWAS). For example, they can estimate heritability of complex traits, allowing for both polygenic and sparse models; and by incorporating external genomic data into the priors, they can increase power and yield new biological insights. However, these methods require access to individual genotypes and phenotypes, which are often not easily available. Here we provide a framework for performing these analyses without individual-level data. Specifically, we introduce a “Regression with Summary Statistics” (RSS) likelihood, which relates the multiple regression coefficients to univariate regression results that are often easily available. The RSS likelihood requires estimates of correlations among covariates (SNPs), which also can be obtained from public databases. We perform Bayesian multiple regression analysis by combining the RSS likelihood with previously proposed prior distributions, sampling posteriors by Markov chain Monte Carlo. In a wide range of simulations RSS performs similarly to analyses using the individual data, both for estimating heritability and detecting associations. We apply RSS to a GWAS of human height that contains 253,288 individuals typed at 1.06 million SNPs, for which analyses of individual-level data are practically impossible. Estimates of heritability (52%) are consistent with, but more precise, than previous results using subsets of these data. We also identify many previously unreported loci that show evidence for association with height in our analyses. Software is available at https://github.com/stephenslab/rss. PMID:29399241

  10. The Determination of Molecular Quantities from Measurements on Macroscopic Systems.V. Existence and Properties of 1:1 and 2:1-Electron-Donor-Acceptor Complexes of Hexamethylbenzene with Tetracyanoethylene

    NASA Astrophysics Data System (ADS)

    Liptay, Wolfgang; Rehm, Torsten; Wehning, Detlev; Schanne, Lothar; Baumann, Wolfram; Lang, Werner

    1982-12-01

    The formation of electron-donor-acceptor complexes of hexamethylbenzene (HMB) with tetracyanoethylene (TCNE) was investigated by measurements of the optical absorptions, the densities, the permittivities and the electro-optical absorptions of solutions in CCl4. The careful evaluation of data based on some previously reported models, has shown that the assumption of the formation of the 1: 1 and the 2 : 1 complex agrees with all experimental data, but that the assumption of the formation of only the 1: 1 complex is contradictory to experimental facts even if the activity effects on the equilibrium constant and of the solvent dependences of observed molar quantities are taken into account. The evaluation leads to the molar optical absorption coefficients and the molar volumes of both complexes and to their electric dipole moments in the electronic ground state and the considered excited state. According to these results the complexes are of the sandwich type HMB-TCNE and HMB-TCNE-HMB. In spite of the fact that the 2: 1 complex owns a center of symmetry, at least approximately, there is a rather large electric dipole moment in its excited state. Furthermore, values for the equilibrium constants and for the standard reaction enthalpies of both complex formation reactions are estimated from experimental data.

  11. Estimation of river pollution index in a tidal stream using kriging analysis.

    PubMed

    Chen, Yen-Chang; Yeh, Hui-Chung; Wei, Chiang

    2012-08-29

    Tidal streams are complex watercourses that represent a transitional zone between riverine and marine systems; they occur where fresh and marine waters converge. Because tidal circulation processes cause substantial turbulence in these highly dynamic zones, tidal streams are the most productive of water bodies. Their rich biological diversity, combined with the convenience of land and water transports, provide sites for concentrated populations that evolve into large cities. Domestic wastewater is generally discharged directly into tidal streams in Taiwan, necessitating regular evaluation of the water quality of these streams. Given the complex flow dynamics of tidal streams, only a few models can effectively evaluate and identify pollution levels. This study evaluates the river pollution index (RPI) in tidal streams by using kriging analysis. This is a geostatistical method for interpolating random spatial variation to estimate linear grid points in two or three dimensions. A kriging-based method is developed to evaluate RPI in tidal streams, which is typically considered as 1D in hydraulic engineering. The proposed method efficiently evaluates RPI in tidal streams with the minimum amount of water quality data. Data of the Tanshui River downstream reach available from an estuarine area validate the accuracy and reliability of the proposed method. Results of this study demonstrate that this simple yet reliable method can effectively estimate RPI in tidal streams.

  12. High performance computing enabling exhaustive analysis of higher order single nucleotide polymorphism interaction in Genome Wide Association Studies.

    PubMed

    Goudey, Benjamin; Abedini, Mani; Hopper, John L; Inouye, Michael; Makalic, Enes; Schmidt, Daniel F; Wagner, John; Zhou, Zeyu; Zobel, Justin; Reumann, Matthias

    2015-01-01

    Genome-wide association studies (GWAS) are a common approach for systematic discovery of single nucleotide polymorphisms (SNPs) which are associated with a given disease. Univariate analysis approaches commonly employed may miss important SNP associations that only appear through multivariate analysis in complex diseases. However, multivariate SNP analysis is currently limited by its inherent computational complexity. In this work, we present a computational framework that harnesses supercomputers. Based on our results, we estimate a three-way interaction analysis on 1.1 million SNP GWAS data requiring over 5.8 years on the full "Avoca" IBM Blue Gene/Q installation at the Victorian Life Sciences Computation Initiative. This is hundreds of times faster than estimates for other CPU based methods and four times faster than runtimes estimated for GPU methods, indicating how the improvement in the level of hardware applied to interaction analysis may alter the types of analysis that can be performed. Furthermore, the same analysis would take under 3 months on the currently largest IBM Blue Gene/Q supercomputer "Sequoia" at the Lawrence Livermore National Laboratory assuming linear scaling is maintained as our results suggest. Given that the implementation used in this study can be further optimised, this runtime means it is becoming feasible to carry out exhaustive analysis of higher order interaction studies on large modern GWAS.

  13. Metal-organic complexes in geochemical processes: Estimation of standard partial molal thermodynamic properties of aqueous complexes between metal cations and monovalent organic acid ligands at high pressures and temperatures

    NASA Astrophysics Data System (ADS)

    Shock, Everetr L.; Koretsky, Carla M.

    1995-04-01

    Regression of standard state equilibrium constants with the revised Helgeson-Kirkham-Flowers (HKF) equation of state allows evaluation of standard partial molal entropies ( overlineSo) of aqueous metal-organic complexes involving monovalent organic acid ligands. These values of overlineSo provide the basis for correlations that can be used, together with correlation algorithms among standard partial molal properties of aqueous complexes and equation-of-state parameters, to estimate thermodynamic properties including equilibrium constants for complexes between aqueous metals and several monovalent organic acid ligands at the elevated pressures and temperatures of many geochemical processes which involve aqueous solutions. Data, parameters, and estimates are given for 270 formate, propanoate, n-butanoate, n-pentanoate, glycolate, lactate, glycinate, and alanate complexes, and a consistent algorithm is provided for making other estimates. Standard partial molal entropies of association ( Δ -Sro) for metal-monovalent organic acid ligand complexes fall into at least two groups dependent upon the type of functional groups present in the ligand. It is shown that isothermal correlations among equilibrium constants for complex formation are consistent with one another and with similar correlations for inorganic metal-ligand complexes. Additional correlations allow estimates of standard partial molal Gibbs free energies of association at 25°C and 1 bar which can be used in cases where no experimentally derived values are available.

  14. Geomorphology and Ice Content of Glacier - Rock Glacier &ndash; Moraine Complexes in Ak-Shiirak Range (Inner Tien Shan, Kyrgyzstan)

    NASA Astrophysics Data System (ADS)

    Bolch, Tobias; Kutuzov, Stanislav; Rohrbach, Nico; Fischer, Andrea; Osmonov, Azamat

    2015-04-01

    Meltwater originating from the Tien Shan is of high importance for the runoff to the arid and semi-arid region of Central Asia. Previous studies estimate a glaciers' contribution of about 40% for the Aksu-Tarim Catchment, a transboundary watershed between Kyrgyzstan and China. Large parts of the Ak-Shiirak Range drain into this watershed. Glaciers in Central and Inner Tien Shan are typically polythermal or even cold and surrounded by permafrost. Several glaciers terminate into large moraine complexes which show geomorphological indicators of ice content such as thermo-karst like depressions, and further downvalley signs of creep such as ridges and furrows and a fresh, steep rock front which are typical indicators for permafrost creep ("rock glacier"). Hence, glaciers and permafrost co-exist in this region and their interactions are important to consider, e.g. for the understanding of glacial and periglacial processes. It can also be assumed that the ice stored in these relatively large dead-ice/moraine-complexes is a significant amount of the total ice storage. However, no detailed investigations exist so far. In an initial study, we investigated the structure and ice content of two typical glacier-moraine complexes in the Ak-Shiirak-Range using different ground penetrating radar (GPR) devices. In addition, the geomorphology was mapped using high resolution satellite imagery. The structure of the moraine-rock glacier complex is in general heterogeneous. Several dead ice bodies with different thicknesses and moraine-derived rock glaciers with different stages of activities could be identified. Few parts of these "rock glaciers" contain also massive ice but the largest parts are likely characterised by rock-ice layers of different thickness and ice contents. In one glacier forefield, the thickness of the rock-ice mixture is partly more than 300 m. This is only slightly lower than the maximum thickness of the glacier ice. Our measurements revealed that up to 20% of the total ice of the entire glacier-rock glacier-moraine-complex could be stored in the moraine-rock glacier parts.

  15. Measuring vulnerability to disaster displacement

    NASA Astrophysics Data System (ADS)

    Brink, Susan A.; Khazai, Bijan; Power, Christopher; Wenzel, Friedemann

    2015-04-01

    Large scale disasters can cause devastating impacts in terms of population displacement. Between 2008 and 2013, on average 27 million people were displaced annually by disasters (Yonetani 2014). After large events such as hurricane Katrina or the Port-au-Prince earthquake, images of inadequate public shelter and concerns about large scale and often inequitable migration have been broadcast around the world. Population displacement can often be one of the most devastating and visible impacts of a natural disaster. Despite the importance of population displacement in disaster events, measures to understand the socio-economic vulnerability of a community often use broad metrics to estimate the total socio-economic risk of an event rather than focusing on the specific impacts that a community faces in a disaster. Population displacement is complex and multi-causal with the physical impact of a disaster interacting with vulnerability arising from the response, environmental issues (e.g., weather), cultural concerns (e.g., expectations of adequate shelter), and many individual factors (e.g., mobility, risk perception). In addition to the complexity of the causes, population displacement is difficult to measure because of the wide variety of different terms and definitions and its multi-dimensional nature. When we speak of severe population displacement, we may refer to a large number of displaced people, an extended length of displacement or associated difficulties such as poor shelter quality, risk of violence and crime in shelter communities, discrimination in aid, a lack of access to employment or other difficulties that can be associated with large scale population displacement. We have completed a thorough review of the literature on disaster population displacement. Research has been conducted on historic events to understand the types of negative impacts associated with population displacement and also the vulnerability of different groups to these impacts. We aggregate these ideas into a framework of disaster displacement vulnerability that distinguishes between three main aspects of disaster displacement. Disaster displacement can be considered in terms of the number of displaced people and the length of that displacement. However, the literature emphasizes that the severity of disaster displacement can not be measured completely in quantitative terms. Thus, we include a measure representing people who are trapped and unable to leave their homes due to mobility, resources or for other reasons. Finally the third main aspect considers the difficulties that are associated with displacement and reflects the difference between the experiences of those who are displaced into safe and supportive environments as compared to those whose only alternate shelter is dangerous and inadequate for their needs. Finally, we apply the framework to demonstrate a methodology to estimate vulnerability to disaster displacement. Using data from the Global Earthquake Model (GEM) Social and Economic Vulnerability sub-National Database, we generate an index to measure the vulnerability of Japanese prefectures to the dimensions of displacement included in the framework. References Yonitani, M. (2014). Global Estimates 2014: People displaced by disasters. http://www.internal-displacement.org/publications/2014/global-estimates-2014-people-displaced-by-disasters/

  16. Scope Complexity Options Risks Excursions (SCORE) Version 3.0 Mathematical Description.

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Gearhart, Jared Lee; Samberson, Jonell Nicole; Shettigar, Subhasini

    The purpose of the Scope, Complexity, Options, Risks, Excursions (SCORE) model is to estimate the relative complexity of design variants of future warhead options. The results of this model allow those considering these options to understand the complexity tradeoffs between proposed warhead options. The core idea of SCORE is to divide a warhead option into a well- defined set of scope elements and then estimate the complexity of each scope element against a well understood reference system. The uncertainty associated with estimates can also be captured. A weighted summation of the relative complexity of each scope element is used tomore » determine the total complexity of the proposed warhead option or portions of the warhead option (i.e., a National Work Breakdown Structure code). The SCORE analysis process is a growing multi-organizational Nuclear Security Enterprise (NSE) effort, under the management of the NA- 12 led Enterprise Modeling and Analysis Consortium (EMAC), that has provided the data elicitation, integration and computation needed to support the out-year Life Extension Program (LEP) cost estimates included in the Stockpile Stewardship Management Plan (SSMP).« less

  17. A Subspace Pursuit–based Iterative Greedy Hierarchical Solution to the Neuromagnetic Inverse Problem

    PubMed Central

    Babadi, Behtash; Obregon-Henao, Gabriel; Lamus, Camilo; Hämäläinen, Matti S.; Brown, Emery N.; Purdon, Patrick L.

    2013-01-01

    Magnetoencephalography (MEG) is an important non-invasive method for studying activity within the human brain. Source localization methods can be used to estimate spatiotemporal activity from MEG measurements with high temporal resolution, but the spatial resolution of these estimates is poor due to the ill-posed nature of the MEG inverse problem. Recent developments in source localization methodology have emphasized temporal as well as spatial constraints to improve source localization accuracy, but these methods can be computationally intense. Solutions emphasizing spatial sparsity hold tremendous promise, since the underlying neurophysiological processes generating MEG signals are often sparse in nature, whether in the form of focal sources, or distributed sources representing large-scale functional networks. Recent developments in the theory of compressed sensing (CS) provide a rigorous framework to estimate signals with sparse structure. In particular, a class of CS algorithms referred to as greedy pursuit algorithms can provide both high recovery accuracy and low computational complexity. Greedy pursuit algorithms are difficult to apply directly to the MEG inverse problem because of the high-dimensional structure of the MEG source space and the high spatial correlation in MEG measurements. In this paper, we develop a novel greedy pursuit algorithm for sparse MEG source localization that overcomes these fundamental problems. This algorithm, which we refer to as the Subspace Pursuit-based Iterative Greedy Hierarchical (SPIGH) inverse solution, exhibits very low computational complexity while achieving very high localization accuracy. We evaluate the performance of the proposed algorithm using comprehensive simulations, as well as the analysis of human MEG data during spontaneous brain activity and somatosensory stimuli. These studies reveal substantial performance gains provided by the SPIGH algorithm in terms of computational complexity, localization accuracy, and robustness. PMID:24055554

  18. Lagged kernel machine regression for identifying time windows of susceptibility to exposures of complex mixtures.

    PubMed

    Liu, Shelley H; Bobb, Jennifer F; Lee, Kyu Ha; Gennings, Chris; Claus Henn, Birgit; Bellinger, David; Austin, Christine; Schnaas, Lourdes; Tellez-Rojo, Martha M; Hu, Howard; Wright, Robert O; Arora, Manish; Coull, Brent A

    2018-07-01

    The impact of neurotoxic chemical mixtures on children's health is a critical public health concern. It is well known that during early life, toxic exposures may impact cognitive function during critical time intervals of increased vulnerability, known as windows of susceptibility. Knowledge on time windows of susceptibility can help inform treatment and prevention strategies, as chemical mixtures may affect a developmental process that is operating at a specific life phase. There are several statistical challenges in estimating the health effects of time-varying exposures to multi-pollutant mixtures, such as: multi-collinearity among the exposures both within time points and across time points, and complex exposure-response relationships. To address these concerns, we develop a flexible statistical method, called lagged kernel machine regression (LKMR). LKMR identifies critical exposure windows of chemical mixtures, and accounts for complex non-linear and non-additive effects of the mixture at any given exposure window. Specifically, LKMR estimates how the effects of a mixture of exposures change with the exposure time window using a Bayesian formulation of a grouped, fused lasso penalty within a kernel machine regression (KMR) framework. A simulation study demonstrates the performance of LKMR under realistic exposure-response scenarios, and demonstrates large gains over approaches that consider each time window separately, particularly when serial correlation among the time-varying exposures is high. Furthermore, LKMR demonstrates gains over another approach that inputs all time-specific chemical concentrations together into a single KMR. We apply LKMR to estimate associations between neurodevelopment and metal mixtures in Early Life Exposures in Mexico and Neurotoxicology, a prospective cohort study of child health in Mexico City.

  19. Adaptive Distributed Video Coding with Correlation Estimation using Expectation Propagation

    PubMed Central

    Cui, Lijuan; Wang, Shuang; Jiang, Xiaoqian; Cheng, Samuel

    2013-01-01

    Distributed video coding (DVC) is rapidly increasing in popularity by the way of shifting the complexity from encoder to decoder, whereas no compression performance degrades, at least in theory. In contrast with conventional video codecs, the inter-frame correlation in DVC is explored at decoder based on the received syndromes of Wyner-Ziv (WZ) frame and side information (SI) frame generated from other frames available only at decoder. However, the ultimate decoding performances of DVC are based on the assumption that the perfect knowledge of correlation statistic between WZ and SI frames should be available at decoder. Therefore, the ability of obtaining a good statistical correlation estimate is becoming increasingly important in practical DVC implementations. Generally, the existing correlation estimation methods in DVC can be classified into two main types: pre-estimation where estimation starts before decoding and on-the-fly (OTF) estimation where estimation can be refined iteratively during decoding. As potential changes between frames might be unpredictable or dynamical, OTF estimation methods usually outperforms pre-estimation techniques with the cost of increased decoding complexity (e.g., sampling methods). In this paper, we propose a low complexity adaptive DVC scheme using expectation propagation (EP), where correlation estimation is performed OTF as it is carried out jointly with decoding of the factor graph-based DVC code. Among different approximate inference methods, EP generally offers better tradeoff between accuracy and complexity. Experimental results show that our proposed scheme outperforms the benchmark state-of-the-art DISCOVER codec and other cases without correlation tracking, and achieves comparable decoding performance but with significantly low complexity comparing with sampling method. PMID:23750314

  20. Effects of model complexity and priors on estimation using sequential importance sampling/resampling for species conservation

    USGS Publications Warehouse

    Dunham, Kylee; Grand, James B.

    2016-01-01

    We examined the effects of complexity and priors on the accuracy of models used to estimate ecological and observational processes, and to make predictions regarding population size and structure. State-space models are useful for estimating complex, unobservable population processes and making predictions about future populations based on limited data. To better understand the utility of state space models in evaluating population dynamics, we used them in a Bayesian framework and compared the accuracy of models with differing complexity, with and without informative priors using sequential importance sampling/resampling (SISR). Count data were simulated for 25 years using known parameters and observation process for each model. We used kernel smoothing to reduce the effect of particle depletion, which is common when estimating both states and parameters with SISR. Models using informative priors estimated parameter values and population size with greater accuracy than their non-informative counterparts. While the estimates of population size and trend did not suffer greatly in models using non-informative priors, the algorithm was unable to accurately estimate demographic parameters. This model framework provides reasonable estimates of population size when little to no information is available; however, when information on some vital rates is available, SISR can be used to obtain more precise estimates of population size and process. Incorporating model complexity such as that required by structured populations with stage-specific vital rates affects precision and accuracy when estimating latent population variables and predicting population dynamics. These results are important to consider when designing monitoring programs and conservation efforts requiring management of specific population segments.

  1. Adaptive distributed video coding with correlation estimation using expectation propagation

    NASA Astrophysics Data System (ADS)

    Cui, Lijuan; Wang, Shuang; Jiang, Xiaoqian; Cheng, Samuel

    2012-10-01

    Distributed video coding (DVC) is rapidly increasing in popularity by the way of shifting the complexity from encoder to decoder, whereas no compression performance degrades, at least in theory. In contrast with conventional video codecs, the inter-frame correlation in DVC is explored at decoder based on the received syndromes of Wyner-Ziv (WZ) frame and side information (SI) frame generated from other frames available only at decoder. However, the ultimate decoding performances of DVC are based on the assumption that the perfect knowledge of correlation statistic between WZ and SI frames should be available at decoder. Therefore, the ability of obtaining a good statistical correlation estimate is becoming increasingly important in practical DVC implementations. Generally, the existing correlation estimation methods in DVC can be classified into two main types: pre-estimation where estimation starts before decoding and on-the-fly (OTF) estimation where estimation can be refined iteratively during decoding. As potential changes between frames might be unpredictable or dynamical, OTF estimation methods usually outperforms pre-estimation techniques with the cost of increased decoding complexity (e.g., sampling methods). In this paper, we propose a low complexity adaptive DVC scheme using expectation propagation (EP), where correlation estimation is performed OTF as it is carried out jointly with decoding of the factor graph-based DVC code. Among different approximate inference methods, EP generally offers better tradeoff between accuracy and complexity. Experimental results show that our proposed scheme outperforms the benchmark state-of-the-art DISCOVER codec and other cases without correlation tracking, and achieves comparable decoding performance but with significantly low complexity comparing with sampling method.

  2. Adaptive Distributed Video Coding with Correlation Estimation using Expectation Propagation.

    PubMed

    Cui, Lijuan; Wang, Shuang; Jiang, Xiaoqian; Cheng, Samuel

    2012-10-15

    Distributed video coding (DVC) is rapidly increasing in popularity by the way of shifting the complexity from encoder to decoder, whereas no compression performance degrades, at least in theory. In contrast with conventional video codecs, the inter-frame correlation in DVC is explored at decoder based on the received syndromes of Wyner-Ziv (WZ) frame and side information (SI) frame generated from other frames available only at decoder. However, the ultimate decoding performances of DVC are based on the assumption that the perfect knowledge of correlation statistic between WZ and SI frames should be available at decoder. Therefore, the ability of obtaining a good statistical correlation estimate is becoming increasingly important in practical DVC implementations. Generally, the existing correlation estimation methods in DVC can be classified into two main types: pre-estimation where estimation starts before decoding and on-the-fly (OTF) estimation where estimation can be refined iteratively during decoding. As potential changes between frames might be unpredictable or dynamical, OTF estimation methods usually outperforms pre-estimation techniques with the cost of increased decoding complexity (e.g., sampling methods). In this paper, we propose a low complexity adaptive DVC scheme using expectation propagation (EP), where correlation estimation is performed OTF as it is carried out jointly with decoding of the factor graph-based DVC code. Among different approximate inference methods, EP generally offers better tradeoff between accuracy and complexity. Experimental results show that our proposed scheme outperforms the benchmark state-of-the-art DISCOVER codec and other cases without correlation tracking, and achieves comparable decoding performance but with significantly low complexity comparing with sampling method.

  3. Geographic and ecologic distributions of the Anopheles gambiae complex predicted using a genetic algorithm.

    PubMed

    Levine, Rebecca S; Peterson, A Townsend; Benedict, Mark Q

    2004-02-01

    The distribution of the Anopheles gambiae complex of malaria vectors in Africa is uncertain due to under-sampling of vast regions. We use ecologic niche modeling to predict the potential distribution of three members of the complex (A. gambiae, A. arabiensis, and A. quadriannulatus) and demonstrate the statistical significance of the models. Predictions correspond well to previous estimates, but provide detail regarding spatial discontinuities in the distribution of A. gambiae s.s. that are consistent with population genetic studies. Our predictions also identify large areas of Africa where the presence of A. arabiensis is predicted, but few specimens have been obtained, suggesting under-sampling of the species. Finally, we project models developed from African distribution data for the late 1900s into the past and to South America to determine retrospectively whether the deadly 1929 introduction of A. gambiae sensu lato into Brazil was more likely that of A. gambiae sensu stricto or A. arabiensis.

  4. Global daily reference evapotranspiration modeling and evaluation

    USGS Publications Warehouse

    Senay, G.B.; Verdin, J.P.; Lietzow, R.; Melesse, Assefa M.

    2008-01-01

    Accurate and reliable evapotranspiration (ET) datasets are crucial in regional water and energy balance studies. Due to the complex instrumentation requirements, actual ET values are generally estimated from reference ET values by adjustment factors using coefficients for water stress and vegetation conditions, commonly referred to as crop coefficients. Until recently, the modeling of reference ET has been solely based on important weather variables collected from weather stations that are generally located in selected agro-climatic locations. Since 2001, the National Oceanic and Atmospheric Administration’s Global Data Assimilation System (GDAS) has been producing six-hourly climate parameter datasets that are used to calculate daily reference ET for the whole globe at 1-degree spatial resolution. The U.S. Geological Survey Center for Earth Resources Observation and Science has been producing daily reference ET (ETo) since 2001, and it has been used on a variety of operational hydrological models for drought and streamflow monitoring all over the world. With the increasing availability of local station-based reference ET estimates, we evaluated the GDAS-based reference ET estimates using data from the California Irrigation Management Information System (CIMIS). Daily CIMIS reference ET estimates from 85 stations were compared with GDAS-based reference ET at different spatial and temporal scales using five-year daily data from 2002 through 2006. Despite the large difference in spatial scale (point vs. ∼100 km grid cell) between the two datasets, the correlations between station-based ET and GDAS-ET were very high, exceeding 0.97 on a daily basis to more than 0.99 on time scales of more than 10 days. Both the temporal and spatial correspondences in trend/pattern and magnitudes between the two datasets were satisfactory, suggesting the reliability of using GDAS parameter-based reference ET for regional water and energy balance studies in many parts of the world. While the study revealed the potential of GDAS ETo for large-scale hydrological applications, site-specific use of GDAS ETo in complex hydro-climatic regions such as coastal areas and rugged terrain may require the application of bias correction and/or disaggregation of the GDAS ETo using downscaling techniques.

  5. Mapping the universe in three dimensions

    PubMed Central

    Haynes, Martha P.

    1996-01-01

    The determination of the three-dimensional layout of galaxies is critical to our understanding of the evolution of galaxies and the structures in which they lie, to our determination of the fundamental parameters of cosmology, and to our understanding of both the past and future histories of the universe at large. The mapping of the large scale structure in the universe via the determination of galaxy red shifts (Doppler shifts) is a rapidly growing industry thanks to technological developments in detectors and spectrometers at radio and optical wavelengths. First-order application of the red shift-distance relation (Hubble’s law) allows the analysis of the large-scale distribution of galaxies on scales of hundreds of megaparsecs. Locally, the large-scale structure is very complex but the overall topology is not yet clear. Comparison of the observed red shifts with ones expected on the basis of other distance estimates allows mapping of the gravitational field and the underlying total density distribution. The next decade holds great promise for our understanding of the character of large-scale structure and its origin. PMID:11607714

  6. Mapping the universe in three dimensions.

    PubMed

    Haynes, M P

    1996-12-10

    The determination of the three-dimensional layout of galaxies is critical to our understanding of the evolution of galaxies and the structures in which they lie, to our determination of the fundamental parameters of cosmology, and to our understanding of both the past and future histories of the universe at large. The mapping of the large scale structure in the universe via the determination of galaxy red shifts (Doppler shifts) is a rapidly growing industry thanks to technological developments in detectors and spectrometers at radio and optical wavelengths. First-order application of the red shift-distance relation (Hubble's law) allows the analysis of the large-scale distribution of galaxies on scales of hundreds of megaparsecs. Locally, the large-scale structure is very complex but the overall topology is not yet clear. Comparison of the observed red shifts with ones expected on the basis of other distance estimates allows mapping of the gravitational field and the underlying total density distribution. The next decade holds great promise for our understanding of the character of large-scale structure and its origin.

  7. Estimating sediment budgets at the interface between rivers and estuaries with application to the Sacramento-San Joaquin River Delta

    USGS Publications Warehouse

    Wright, S.A.; Schoellhamer, D.H.

    2005-01-01

    [1] Where rivers encounter estuaries, a transition zone develops where riverine and tidal processes both affect sediment transport processes. One such transition zone is the Sacramento-San Joaquin River Delta, a large, complex system where several rivers meet to form an estuary (San Francisco Bay). Herein we present the results of a detailed sediment budget for this river/estuary transitional system. The primary regional goal of the study was to measure sediment transport rates and pathways in the delta in support of ecosystem restoration efforts. In addition to achieving this regional goal, the study has produced general methods to collect, edit, and analyze (including error analysis) sediment transport data at the interface of rivers and estuaries. Estimating sediment budgets for these systems is difficult because of the mixed nature of riverine versus tidal transport processes, the different timescales of transport in fluvial and tidal environments, and the sheer complexity and size of systems such as the Sacramento-San Joaquin River Delta. Sediment budgets also require error estimates in order to assess whether differences in inflows and outflows, which could be small compared to overall fluxes, are indeed distinguishable from zero. Over the 4 year period of this study, water years 1999-2002, 6.6 ?? 0.9 Mt of sediment entered the delta and 2.2 ?? 0.7 Mt exited, resulting in 4.4 ?? 1.1 Mt (67 ?? 17%) of deposition. The estimated deposition rate corresponding to this mass of sediment compares favorably with measured inorganic sediment accumulation on vegetated wetlands in the delta.

  8. Error Estimates for Approximate Solutions of the Riccati Equation with Real or Complex Potentials

    NASA Astrophysics Data System (ADS)

    Finster, Felix; Smoller, Joel

    2010-09-01

    A method is presented for obtaining rigorous error estimates for approximate solutions of the Riccati equation, with real or complex potentials. Our main tool is to derive invariant region estimates for complex solutions of the Riccati equation. We explain the general strategy for applying these estimates and illustrate the method in typical examples, where the approximate solutions are obtained by gluing together WKB and Airy solutions of corresponding one-dimensional Schrödinger equations. Our method is motivated by, and has applications to, the analysis of linear wave equations in the geometry of a rotating black hole.

  9. A global food demand model for the assessment of complex human-earth systems

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    EDMONDS, JAMES A.; LINK, ROBERT; WALDHOFF, STEPHANIE T.

    Demand for agricultural products is an important problem in climate change economics. Food consumption will shape and shaped by climate change and emissions mitigation policies through interactions with bioenergy and afforestation, two critical issues in meeting international climate goals such as two-degrees. We develop a model of food demand for staple and nonstaple commodities that evolves with changing incomes and prices. The model addresses a long-standing issue in estimating food demands, the evolution of demand relationships across large changes in income and prices. We discuss the model, some of its properties and limitations. We estimate parameter values using pooled cross-sectional-time-seriesmore » observations and the Metropolis Monte Carlo method and cross-validate the model by estimating parameters using a subset of the observations and test its ability to project into the unused observations. Finally, we apply bias correction techniques borrowed from the climate-modeling community and report results.« less

  10. Model synthesis in frequency analysis of Missouri floods

    USGS Publications Warehouse

    Hauth, Leland D.

    1974-01-01

    Synthetic flood records for 43 small-stream sites aided in definition of techniques for estimating the magnitude and frequency of floods in Missouri. The long-term synthetic flood records were generated by use of a digital computer model of the rainfall-runoff process. A relatively short period of concurrent rainfall and runoff data observed at each of the 43 sites was used to calibrate the model, and rainfall records covering from 66 to 78 years for four Missouri sites and pan-evaporation data were used to generate the synthetic records. Flood magnitude and frequency characteristics of both the synthetic records and observed long-term flood records available for 109 large-stream sites were used in a multiple-regression analysis to define relations for estimating future flood characteristics at ungaged sites. That analysis indicated that drainage basin size and slope were the most useful estimating variables. It also indicated that a more complex regression model than the commonly used log-linear one was needed for the range of drainage basin sizes available in this study.

  11. [Perception by teenagers and adults of the changed by amplitude sound sequences used in models of movement of the sound source].

    PubMed

    Andreeva, I G; Vartanian, I A

    2012-01-01

    The ability to evaluate direction of amplitude changes of sound stimuli was studied in adults and in the 11-12- and 15-16-year old teenagers. The stimuli representing sequences of fragments of the tone of 1 kHz, whose amplitude is changing with time, are used as model of approach and withdrawal of the sound sources. The 11-12-year old teenagers at estimation of direction of amplitude changes were shown to make the significantly higher number of errors as compared with two other examined groups, including those in repeated experiments. The structure of errors - the ratio of the portion of errors at estimation of increasing and decreasing by amplitude stimulus - turned out to be different in teenagers and in adults. The question is discussed about the effect of unspecific activation of the large hemisphere cortex in teenagers on processes if taking solution about the complex sound stimulus, including a possibility estimation of approach and withdrawal of the sound source.

  12. Target Tracking Using SePDAF under Ambiguous Angles for Distributed Array Radar.

    PubMed

    Long, Teng; Zhang, Honggang; Zeng, Tao; Chen, Xinliang; Liu, Quanhua; Zheng, Le

    2016-09-09

    Distributed array radar can improve radar detection capability and measurement accuracy. However, it will suffer cyclic ambiguity in its angle estimates according to the spatial Nyquist sampling theorem since the large sparse array is undersampling. Consequently, the state estimation accuracy and track validity probability degrades when the ambiguous angles are directly used for target tracking. This paper proposes a second probability data association filter (SePDAF)-based tracking method for distributed array radar. Firstly, the target motion model and radar measurement model is built. Secondly, the fusion result of each radar's estimation is employed to the extended Kalman filter (EKF) to finish the first filtering. Thirdly, taking this result as prior knowledge, and associating with the array-processed ambiguous angles, the SePDAF is applied to accomplish the second filtering, and then achieving a high accuracy and stable trajectory with relatively low computational complexity. Moreover, the azimuth filtering accuracy will be promoted dramatically and the position filtering accuracy will also improve. Finally, simulations illustrate the effectiveness of the proposed method.

  13. Age, extent and carbon storage of the central Congo Basin peatland complex.

    PubMed

    Dargie, Greta C; Lewis, Simon L; Lawson, Ian T; Mitchard, Edward T A; Page, Susan E; Bocko, Yannick E; Ifo, Suspense A

    2017-02-02

    Peatlands are carbon-rich ecosystems that cover just three per cent of Earth's land surface, but store one-third of soil carbon. Peat soils are formed by the build-up of partially decomposed organic matter under waterlogged anoxic conditions. Most peat is found in cool climatic regions where unimpeded decomposition is slower, but deposits are also found under some tropical swamp forests. Here we present field measurements from one of the world's most extensive regions of swamp forest, the Cuvette Centrale depression in the central Congo Basin. We find extensive peat deposits beneath the swamp forest vegetation (peat defined as material with an organic matter content of at least 65 per cent to a depth of at least 0.3 metres). Radiocarbon dates indicate that peat began accumulating from about 10,600 years ago, coincident with the onset of more humid conditions in central Africa at the beginning of the Holocene. The peatlands occupy large interfluvial basins, and seem to be largely rain-fed and ombrotrophic-like (of low nutrient status) systems. Although the peat layer is relatively shallow (with a maximum depth of 5.9 metres and a median depth of 2.0 metres), by combining in situ and remotely sensed data, we estimate the area of peat to be approximately 145,500 square kilometres (95 per cent confidence interval of 131,900-156,400 square kilometres), making the Cuvette Centrale the most extensive peatland complex in the tropics. This area is more than five times the maximum possible area reported for the Congo Basin in a recent synthesis of pantropical peat extent. We estimate that the peatlands store approximately 30.6 petagrams (30.6 × 10 15  grams) of carbon belowground (95 per cent confidence interval of 6.3-46.8 petagrams of carbon)-a quantity that is similar to the above-ground carbon stocks of the tropical forests of the entire Congo Basin. Our result for the Cuvette Centrale increases the best estimate of global tropical peatland carbon stocks by 36 per cent, to 104.7 petagrams of carbon (minimum estimate of 69.6 petagrams of carbon; maximum estimate of 129.8 petagrams of carbon). This stored carbon is vulnerable to land-use change and any future reduction in precipitation.

  14. Adaptive neuro fuzzy inference system-based power estimation method for CMOS VLSI circuits

    NASA Astrophysics Data System (ADS)

    Vellingiri, Govindaraj; Jayabalan, Ramesh

    2018-03-01

    Recent advancements in very large scale integration (VLSI) technologies have made it feasible to integrate millions of transistors on a single chip. This greatly increases the circuit complexity and hence there is a growing need for less-tedious and low-cost power estimation techniques. The proposed work employs Back-Propagation Neural Network (BPNN) and Adaptive Neuro Fuzzy Inference System (ANFIS), which are capable of estimating the power precisely for the complementary metal oxide semiconductor (CMOS) VLSI circuits, without requiring any knowledge on circuit structure and interconnections. The ANFIS to power estimation application is relatively new. Power estimation using ANFIS is carried out by creating initial FIS modes using hybrid optimisation and back-propagation (BP) techniques employing constant and linear methods. It is inferred that ANFIS with the hybrid optimisation technique employing the linear method produces better results in terms of testing error that varies from 0% to 0.86% when compared to BPNN as it takes the initial fuzzy model and tunes it by means of a hybrid technique combining gradient descent BP and mean least-squares optimisation algorithms. ANFIS is the best suited for power estimation application with a low RMSE of 0.0002075 and a high coefficient of determination (R) of 0.99961.

  15. Risks of Large Portfolios

    PubMed Central

    Fan, Jianqing; Liao, Yuan; Shi, Xiaofeng

    2014-01-01

    The risk of a large portfolio is often estimated by substituting a good estimator of the volatility matrix. However, the accuracy of such a risk estimator is largely unknown. We study factor-based risk estimators under a large amount of assets, and introduce a high-confidence level upper bound (H-CLUB) to assess the estimation. The H-CLUB is constructed using the confidence interval of risk estimators with either known or unknown factors. We derive the limiting distribution of the estimated risks in high dimensionality. We find that when the dimension is large, the factor-based risk estimators have the same asymptotic variance no matter whether the factors are known or not, which is slightly smaller than that of the sample covariance-based estimator. Numerically, H-CLUB outperforms the traditional crude bounds, and provides an insightful risk assessment. In addition, our simulated results quantify the relative error in the risk estimation, which is usually negligible using 3-month daily data. PMID:26195851

  16. Bayesian estimation of extreme flood quantiles using a rainfall-runoff model and a stochastic daily rainfall generator

    NASA Astrophysics Data System (ADS)

    Costa, Veber; Fernandes, Wilson

    2017-11-01

    Extreme flood estimation has been a key research topic in hydrological sciences. Reliable estimates of such events are necessary as structures for flood conveyance are continuously evolving in size and complexity and, as a result, their failure-associated hazards become more and more pronounced. Due to this fact, several estimation techniques intended to improve flood frequency analysis and reducing uncertainty in extreme quantile estimation have been addressed in the literature in the last decades. In this paper, we develop a Bayesian framework for the indirect estimation of extreme flood quantiles from rainfall-runoff models. In the proposed approach, an ensemble of long daily rainfall series is simulated with a stochastic generator, which models extreme rainfall amounts with an upper-bounded distribution function, namely, the 4-parameter lognormal model. The rationale behind the generation model is that physical limits for rainfall amounts, and consequently for floods, exist and, by imposing an appropriate upper bound for the probabilistic model, more plausible estimates can be obtained for those rainfall quantiles with very low exceedance probabilities. Daily rainfall time series are converted into streamflows by routing each realization of the synthetic ensemble through a conceptual hydrologic model, the Rio Grande rainfall-runoff model. Calibration of parameters is performed through a nonlinear regression model, by means of the specification of a statistical model for the residuals that is able to accommodate autocorrelation, heteroscedasticity and nonnormality. By combining the outlined steps in a Bayesian structure of analysis, one is able to properly summarize the resulting uncertainty and estimating more accurate credible intervals for a set of flood quantiles of interest. The method for extreme flood indirect estimation was applied to the American river catchment, at the Folsom dam, in the state of California, USA. Results show that most floods, including exceptionally large non-systematic events, were reasonably estimated with the proposed approach. In addition, by accounting for uncertainties in each modeling step, one is able to obtain a better understanding of the influential factors in large flood formation dynamics.

  17. Statistical analysis of donation--transfusion data with complex correlation.

    PubMed

    Reilly, Marie; Szulkin, Robert

    2007-12-30

    Blood-borne transmission of disease is estimated from linked data records from blood donors and transfusion recipients. However, such data are characterized by complex correlation due to donors typically contributing many donations and recipients being transfused with multiple units of blood product. In this paper, we present a method for analysing such data, by using a modification of a nested case-control design. For recipients who develop the disease of interest (cases) and their matched controls, all donors who contributed blood to these individuals define clusters or 'families' of related individuals. Using a Cox regression model for the hazard of the individuals within clusters of donors, we estimate the risk of transmission, and a bootstrap step provides valid standard errors provided the clusters are independent. As an illustration, we apply the method to the analysis of a large database of Swedish donor and recipient records linked to the population cancer register. We investigate whether there is an increased risk of cancer in recipients transfused with blood from donors who develop cancer after donating. Our method provides a powerful alternative to the small 'look-back' studies typical of transfusion medicine and can make an important contribution to haemovigilance efforts. Copyright (c) 2007 John Wiley & Sons, Ltd.

  18. Stress anisotropy analysis and its effect on unconventional resource development in Montney play, Kakwa, Canada

    NASA Astrophysics Data System (ADS)

    Tak, Heewon; Choi, Jaewon; Jo, Sohyun; Hwang, Sukyeon

    2017-04-01

    Stress anisotropy analysis is important for estimating both stress regime and fracture geometry for the efficient development of unconventional resources. Despite being within the same play, different areas can have different stress regimes, which can affect drilling decisions. The Montney play is located in Canada between British Columbia and Alberta. In British Columbia it is known for its ductile shale and high horizontal stress anisotropy because of the Rocky Mountains; however, in Alberta, it has different geological characteristics with some studies finding weak horizontal stress anisotropy. Therefore, we studied the horizontal stress anisotropy using full azimuth seismic and well data in the Kakwa area in order to establish a drilling plan. Minimal horizontal anisotropy was discovered within the area and the direction of maximum horizontal anisotropy corresponded with the regional scale (i.e., NE-SW). The induced fractures were assumed to have a normal stress regime because of the large depth (> 3000 m). Additionally, because of the very high brittleness (Young's modulus > 9) and relatively weak horizontal stress anisotropy, the fracture geometry in the Kakwa area was estimated as complex or complex planar, as opposed to simply planar.

  19. POD evaluation using simulation: A phased array UT case on a complex geometry part

    NASA Astrophysics Data System (ADS)

    Dominguez, Nicolas; Reverdy, Frederic; Jenson, Frederic

    2014-02-01

    The use of Probability of Detection (POD) for NDT performances demonstration is a key link in products lifecycle management. The POD approach is to apply the given NDT procedure on a series of known flaws to estimate the probability to detect with respect to the flaw size. A POD is relevant if and only if NDT operations are carried out within the range of variability authorized by the procedure. Such experimental campaigns require collection of large enough datasets to cover the range of variability with sufficient occurrences to build a reliable POD statistics, leading to expensive costs to get POD curves. In the last decade research activities have been led in the USA with the MAPOD group and later in Europe with the SISTAE and PICASSO projects based on the idea to use models and simulation tools to feed POD estimations. This paper proposes an example of application of POD using simulation on the inspection procedure of a complex -full 3D- geometry part using phased arrays ultrasonic testing. It illustrates the methodology and the associated tools developed in the CIVA software. The paper finally provides elements of further progress in the domain.

  20. Grid-based lattice summation of electrostatic potentials by assembled rank-structured tensor approximation

    NASA Astrophysics Data System (ADS)

    Khoromskaia, Venera; Khoromskij, Boris N.

    2014-12-01

    Our recent method for low-rank tensor representation of sums of the arbitrarily positioned electrostatic potentials discretized on a 3D Cartesian grid reduces the 3D tensor summation to operations involving only 1D vectors however retaining the linear complexity scaling in the number of potentials. Here, we introduce and study a novel tensor approach for fast and accurate assembled summation of a large number of lattice-allocated potentials represented on 3D N × N × N grid with the computational requirements only weakly dependent on the number of summed potentials. It is based on the assembled low-rank canonical tensor representations of the collected potentials using pointwise sums of shifted canonical vectors representing the single generating function, say the Newton kernel. For a sum of electrostatic potentials over L × L × L lattice embedded in a box the required storage scales linearly in the 1D grid-size, O(N) , while the numerical cost is estimated by O(NL) . For periodic boundary conditions, the storage demand remains proportional to the 1D grid-size of a unit cell, n = N / L, while the numerical cost reduces to O(N) , that outperforms the FFT-based Ewald-type summation algorithms of complexity O(N3 log N) . The complexity in the grid parameter N can be reduced even to the logarithmic scale O(log N) by using data-sparse representation of canonical N-vectors via the quantics tensor approximation. For justification, we prove an upper bound on the quantics ranks for the canonical vectors in the overall lattice sum. The presented approach is beneficial in applications which require further functional calculus with the lattice potential, say, scalar product with a function, integration or differentiation, which can be performed easily in tensor arithmetics on large 3D grids with 1D cost. Numerical tests illustrate the performance of the tensor summation method and confirm the estimated bounds on the tensor ranks.

  1. Gaussian polarizable-ion tight binding.

    PubMed

    Boleininger, Max; Guilbert, Anne Ay; Horsfield, Andrew P

    2016-10-14

    To interpret ultrafast dynamics experiments on large molecules, computer simulation is required due to the complex response to the laser field. We present a method capable of efficiently computing the static electronic response of large systems to external electric fields. This is achieved by extending the density-functional tight binding method to include larger basis sets and by multipole expansion of the charge density into electrostatically interacting Gaussian distributions. Polarizabilities for a range of hydrocarbon molecules are computed for a multipole expansion up to quadrupole order, giving excellent agreement with experimental values, with average errors similar to those from density functional theory, but at a small fraction of the cost. We apply the model in conjunction with the polarizable-point-dipoles model to estimate the internal fields in amorphous poly(3-hexylthiophene-2,5-diyl).

  2. Gaussian polarizable-ion tight binding

    NASA Astrophysics Data System (ADS)

    Boleininger, Max; Guilbert, Anne AY; Horsfield, Andrew P.

    2016-10-01

    To interpret ultrafast dynamics experiments on large molecules, computer simulation is required due to the complex response to the laser field. We present a method capable of efficiently computing the static electronic response of large systems to external electric fields. This is achieved by extending the density-functional tight binding method to include larger basis sets and by multipole expansion of the charge density into electrostatically interacting Gaussian distributions. Polarizabilities for a range of hydrocarbon molecules are computed for a multipole expansion up to quadrupole order, giving excellent agreement with experimental values, with average errors similar to those from density functional theory, but at a small fraction of the cost. We apply the model in conjunction with the polarizable-point-dipoles model to estimate the internal fields in amorphous poly(3-hexylthiophene-2,5-diyl).

  3. Organic Carbon Deposits of Soils Overlying the Ice Complex in the Lena River Delta

    NASA Astrophysics Data System (ADS)

    Zubrzycki, Sebastian; Pfeiffer, Eva-Maria; Kutzbach, Lars; Desiatkin, Aleksei

    2017-04-01

    The Lena River Delta (LRD) is located in northeast Siberia and extends over a soil covered area of around 21,500 km2. LRD likely stores more than half of the entire soil organic carbon (SOC) mass stored in deltas affected by permafrost. LRD consists of several geomorphic units. Recent studies showed that the spatially dominating Holocene units of the LRD (61 % of the area) store around 240 Tg of SOC and 12 Tg of nitrogen (N) within the first meter of ground. These units are a river terrace dominated by wet sedge polygons and the active floodplains. About 50 % of these reported storages are located in the perennially frozen ground below 50 cm depth and are excluded from intense biogeochemical exchange with the atmosphere today. However, these storages are likely to be mineralized in near future due to the projected temperature increases in this region. A substantial part of the LRD (1,712 km2) belongs to the so-called Ice Complex (Yedoma) Region, which formed during the Late Pleistocene. This oldest unit of the LRD is characterized by extensive plains incised by thermo-erosional valleys and large thermokarst depressions. Such depressions are called Alases and cover around 20 % of the area. Ice Complex deposits in the LDR are known to store high amounts of SOC. However, within the LRD no detailed spatial studies on SOC and N in the soils overlying Ice Complex and thermokarst depressions were carried out so far. We present here our "investigation in progress" on soils in these landscape units of the LRD. Our first estimates, based on 69 pedons sampled in 2008, show that the mean SOC stocks for the upper 30 cm of soils on both units were estimated at 13.0 kg m2 ± 4.8 kg m2 on the Ice Complex surfaces and at 13.1 kg m2 ± 3.8 kg m2 in the Alases. The stocks of N were estimated at 0.69 kg m2 ± 0.25 kg m2 and at 0.70 kg m2 ± 0.18 kg m2 on the Ice Complex surfaces and in the Alases, respectively. The estimated SOC and N pools for the depth of 30 cm within the investigated part of the LRD add to 20.9 Tg and 1.1 Tg, respectively. The Ice Complex surfaces (1,313 km2) store 17.1 ± 6.3 Tg SOC and 0.9 ± 0.3 Tg N, whereas the Alases (287 km2) store 3.8 ± 1.1 Tg SOC and 0.2 ± 0.05 Tg N within the investigated depth of 30 cm. Further analyses of the soil core material collected in 2015 will provide SOC and N pool estimates for a depth of 100 cm including both, the seasonally active layer and the perennially frozen ground. With continuing advanced analyses of an available digital elevation model, slopes will be designated with their extents and inclinations since the planar extents of slopes derived from satellite imagery do not correspond to the actual slope soil surface area, which is vital for spatial SOC and N storage calculations as well as trace gas release estimates. The actual soil surface area of slopes will be calculated prior to result extrapolations.

  4. Geophysical study of the San Juan Mountains batholith complex, southwestern Colorado

    USGS Publications Warehouse

    Drenth, Benjamin J.; Keller, G. Randy; Thompson, Ren A.

    2012-01-01

    One of the largest and most pronounced gravity lows over North America is over the rugged San Juan Mountains of southwestern Colorado (USA). The mountain range is coincident with the San Juan volcanic field (SJVF), the largest erosional remnant of a widespread mid-Cenozoic volcanic field that spanned much of the southern Rocky Mountains. A buried, low-density silicic batholith complex related to the volcanic field has been the accepted interpretation of the source of the gravity low since the 1970s. However, this interpretation was based on gravity data processed with standard techniques that are problematic in the SJVF region. The combination of high-relief topography, topography with low densities, and the use of a common reduction density of 2670 kg/m3produces spurious large-amplitude gravity lows that may distort the geophysical signature of deeper features such as a batholith complex. We applied an unconventional processing procedure that uses geologically appropriate densities for the uppermost crust and digital topography to mostly remove the effect of the low-density units that underlie the topography associated with the SJVF. This approach resulted in a gravity map that provides an improved representation of deeper sources, including reducing the amplitude of the anomaly attributed to a batholith complex. We also reinterpreted vintage seismic refraction data that indicate the presence of low-velocity zones under the SJVF. Assuming that the source of the gravity low on the improved gravity anomaly map is the same as the source of the low seismic velocities, integrated modeling corroborates the interpretation of a batholith complex and then defines the dimensions and overall density contrast of the complex. Models show that the thickness of the batholith complex varies laterally to a significant degree, with the greatest thickness (∼20 km) under the western SJVF, and lesser thicknesses (<10 km) under the eastern SJVF. The largest group of nested calderas on the surface of the SJVF, the central caldera cluster, is not correlated with the thickest part of the batholith complex. This result is consistent with petrologic interpretations from recent studies that the batholith complex continued to be modified after cessation of volcanism and therefore is not necessarily representative of synvolcanic magma chambers. The total volume of the batholith complex is estimated to be 82,000–130,000 km3. The formation of such a large felsic batholith complex would inevitably involve production of a considerably greater volume of residuum, which could be present in the lower crust or uppermost mantle. The interpreted vertically averaged density contrast (–60 to –110 kg/m3), density (2590–2640 kg/m3), and seismic expression of the batholith complex are consistent with results of geophysical studies of other large batholiths in the western United States.

  5. Evaluation of Different Phenological Information to Map Crop Rotation in Complex Irrigated Indus Basin

    NASA Astrophysics Data System (ADS)

    Ismaeel, A.; Zhou, Q.

    2018-04-01

    Accurate information of crop rotation in large basin is essential for policy decisions on land, water and nutrient resources around the world. Crop area estimation using low spatial resolution remote sensing data is challenging in a large heterogeneous basin having more than one cropping seasons. This study aims to evaluate the accuracy of two phenological datasets individually and in combined form to map crop rotations in complex irrigated Indus basin without image segmentation. Phenology information derived from Normalized Difference Vegetation Index (NDVI) and Leaf Area Index (LAI) of Moderate Resolution Imaging Spectroradiometer (MODIS) sensor, having 8-day temporal and 1000 m spatial resolution, was used in the analysis. An unsupervised (temporal space clustering) to supervised (area knowledge and phenology behavior) classification approach was adopted to identify 13 crop rotations. Estimated crop area was compared with reported area collected by field census. Results reveal that combined dataset (NDVI*LAI) performs better in mapping wheat-rice, wheat-cotton and wheat-fodder rotation by attaining root mean square error (RMSE) of 34.55, 16.84, 20.58 and mean absolute percentage error (MAPE) of 24.56 %, 36.82 %, 30.21 % for wheat, rice and cotton crop respectively. For sugarcane crop mapping, LAI produce good results by achieving RMSE of 8.60 and MAPE of 34.58 %, as compared to NDVI (10.08, 40.53 %) and NDVI*LAI (10.83, 39.45 %). The availability of major crop rotation statistics provides insight to develop better strategies for land, water and nutrient accounting frameworks to improve agriculture productivity.

  6. Towards an automatic wind speed and direction profiler for Wide Field adaptive optics systems

    NASA Astrophysics Data System (ADS)

    Sivo, G.; Turchi, A.; Masciadri, E.; Guesalaga, A.; Neichel, B.

    2018-05-01

    Wide Field Adaptive Optics (WFAO) systems are among the most sophisticated adaptive optics (AO) systems available today on large telescopes. Knowledge of the vertical spatio-temporal distribution of wind speed (WS) and direction (WD) is fundamental to optimize the performance of such systems. Previous studies already proved that the Gemini Multi-Conjugated AO system (GeMS) is able to retrieve measurements of the WS and WD stratification using the SLOpe Detection And Ranging (SLODAR) technique and to store measurements in the telemetry data. In order to assess the reliability of these estimates and of the SLODAR technique applied to such complex AO systems, in this study we compared WS and WD values retrieved from GeMS with those obtained with the atmospheric model Meso-NH on a rich statistical sample of nights. It has previously been proved that the latter technique provided excellent agreement with a large sample of radiosoundings, both in statistical terms and on individual flights. It can be considered, therefore, as an independent reference. The excellent agreement between GeMS measurements and the model that we find in this study proves the robustness of the SLODAR approach. To bypass the complex procedures necessary to achieve automatic measurements of the wind with GeMS, we propose a simple automatic method to monitor nightly WS and WD using Meso-NH model estimates. Such a method can be applied to whatever present or new-generation facilities are supported by WFAO systems. The interest of this study is, therefore, well beyond the optimization of GeMS performance.

  7. Optimal Model-Based Fault Estimation and Correction for Particle Accelerators and Industrial Plants Using Combined Support Vector Machines and First Principles Models

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Sayyar-Rodsari, Bijan; Schweiger, Carl; /SLAC /Pavilion Technologies, Inc., Austin, TX

    2010-08-25

    Timely estimation of deviations from optimal performance in complex systems and the ability to identify corrective measures in response to the estimated parameter deviations has been the subject of extensive research over the past four decades. The implications in terms of lost revenue from costly industrial processes, operation of large-scale public works projects and the volume of the published literature on this topic clearly indicates the significance of the problem. Applications range from manufacturing industries (integrated circuits, automotive, etc.), to large-scale chemical plants, pharmaceutical production, power distribution grids, and avionics. In this project we investigated a new framework for buildingmore » parsimonious models that are suited for diagnosis and fault estimation of complex technical systems. We used Support Vector Machines (SVMs) to model potentially time-varying parameters of a First-Principles (FP) description of the process. The combined SVM & FP model was built (i.e. model parameters were trained) using constrained optimization techniques. We used the trained models to estimate faults affecting simulated beam lifetime. In the case where a large number of process inputs are required for model-based fault estimation, the proposed framework performs an optimal nonlinear principal component analysis of the large-scale input space, and creates a lower dimension feature space in which fault estimation results can be effectively presented to the operation personnel. To fulfill the main technical objectives of the Phase I research, our Phase I efforts have focused on: (1) SVM Training in a Combined Model Structure - We developed the software for the constrained training of the SVMs in a combined model structure, and successfully modeled the parameters of a first-principles model for beam lifetime with support vectors. (2) Higher-order Fidelity of the Combined Model - We used constrained training to ensure that the output of the SVM (i.e. the parameters of the beam lifetime model) are physically meaningful. (3) Numerical Efficiency of the Training - We investigated the numerical efficiency of the SVM training. More specifically, for the primal formulation of the training, we have developed a problem formulation that avoids the linear increase in the number of the constraints as a function of the number of data points. (4) Flexibility of Software Architecture - The software framework for the training of the support vector machines was designed to enable experimentation with different solvers. We experimented with two commonly used nonlinear solvers for our simulations. The primary application of interest for this project has been the sustained optimal operation of particle accelerators at the Stanford Linear Accelerator Center (SLAC). Particle storage rings are used for a variety of applications ranging from 'colliding beam' systems for high-energy physics research to highly collimated x-ray generators for synchrotron radiation science. Linear accelerators are also used for collider research such as International Linear Collider (ILC), as well as for free electron lasers, such as the Linear Coherent Light Source (LCLS) at SLAC. One common theme in the operation of storage rings and linear accelerators is the need to precisely control the particle beams over long periods of time with minimum beam loss and stable, yet challenging, beam parameters. We strongly believe that beyond applications in particle accelerators, the high fidelity and cost benefits of a combined model-based fault estimation/correction system will attract customers from a wide variety of commercial and scientific industries. Even though the acquisition of Pavilion Technologies, Inc. by Rockwell Automation Inc. in 2007 has altered the small business status of the Pavilion and it no longer qualifies for a Phase II funding, our findings in the course of the Phase I research have convinced us that further research will render a workable model-based fault estimation and correction for particle accelerators and industrial plants feasible.« less

  8. How are flood risk estimates affected by the choice of return-periods?

    NASA Astrophysics Data System (ADS)

    Ward, P. J.; de Moel, H.; Aerts, J. C. J. H.

    2011-12-01

    Flood management is more and more adopting a risk based approach, whereby flood risk is the product of the probability and consequences of flooding. One of the most common approaches in flood risk assessment is to estimate the damage that would occur for floods of several exceedance probabilities (or return periods), to plot these on an exceedance probability-loss curve (risk curve) and to estimate risk as the area under the curve. However, there is little insight into how the selection of the return-periods (which ones and how many) used to calculate risk actually affects the final risk calculation. To gain such insights, we developed and validated an inundation model capable of rapidly simulating inundation extent and depth, and dynamically coupled this to an existing damage model. The method was applied to a section of the River Meuse in the southeast of the Netherlands. Firstly, we estimated risk based on a risk curve using yearly return periods from 2 to 10 000 yr (€ 34 million p.a.). We found that the overall risk is greatly affected by the number of return periods used to construct the risk curve, with over-estimations of annual risk between 33% and 100% when only three return periods are used. In addition, binary assumptions on dike failure can have a large effect (a factor two difference) on risk estimates. Also, the minimum and maximum return period considered in the curve affects the risk estimate considerably. The results suggest that more research is needed to develop relatively simple inundation models that can be used to produce large numbers of inundation maps, complementary to more complex 2-D-3-D hydrodynamic models. It also suggests that research into flood risk could benefit by paying more attention to the damage caused by relatively high probability floods.

  9. [Atmospheric parameter estimation for LAMOST/GUOSHOUJING spectra].

    PubMed

    Lu, Yu; Li, Xiang-Ru; Yang, Tan

    2014-11-01

    It is a key task to estimate the atmospheric parameters from the observed stellar spectra in exploring the nature of stars and universe. With our Large Sky Area Multi-Object Fiber Spectroscopy Telescope (LAMOST) which begun its formal Sky Survey in September 2012, we are obtaining a mass of stellar spectra in an unprecedented speed. It has brought a new opportunity and a challenge for the research of galaxies. Due to the complexity of the observing system, the noise in the spectrum is relatively large. At the same time, the preprocessing procedures of spectrum are also not ideal, such as the wavelength calibration and the flow calibration. Therefore, there is a slight distortion of the spectrum. They result in the high difficulty of estimating the atmospheric parameters for the measured stellar spectra. It is one of the important issues to estimate the atmospheric parameters for the massive stellar spectra of LAMOST. The key of this study is how to eliminate noise and improve the accuracy and robustness of estimating the atmospheric parameters for the measured stellar spectra. We propose a regression model for estimating the atmospheric parameters of LAMOST stellar(SVM(lasso)). The basic idea of this model is: First, we use the Haar wavelet to filter spectrum, suppress the adverse effects of the spectral noise and retain the most discrimination information of spectrum. Secondly, We use the lasso algorithm for feature selection and extract the features of strongly correlating with the atmospheric parameters. Finally, the features are input to the support vector regression model for estimating the parameters. Because the model has better tolerance to the slight distortion and the noise of the spectrum, the accuracy of the measurement is improved. To evaluate the feasibility of the above scheme, we conduct experiments extensively on the 33 963 pilot surveys spectrums by LAMOST. The accuracy of three atmospheric parameters is log Teff: 0.006 8 dex, log g: 0.155 1 dex, [Fe/H]: 0.104 0 dex.

  10. Modelling the evolution of complex conductivity during calcite precipitation on glass beads

    NASA Astrophysics Data System (ADS)

    Leroy, Philippe; Li, Shuai; Jougnot, Damien; Revil, André; Wu, Yuxin

    2017-04-01

    When pH and alkalinity increase, calcite frequently precipitates and hence modifies the petrophysical properties of porous media. The complex conductivity method can be used to directly monitor calcite precipitation in porous media because it is sensitive to the evolution of the mineralogy, pore structure and its connectivity. We have developed a mechanistic grain polarization model considering the electrochemical polarization of the Stern and diffuse layers surrounding calcite particles. Our complex conductivity model depends on the surface charge density of the Stern layer and on the electrical potential at the onset of the diffuse layer, which are computed using a basic Stern model of the calcite/water interface. The complex conductivity measurements of Wu et al. on a column packed with glass beads where calcite precipitation occurs are reproduced by our surface complexation and complex conductivity models. The evolution of the size and shape of calcite particles during the calcite precipitation experiment is estimated by our complex conductivity model. At the early stage of the calcite precipitation experiment, modelled particles sizes increase and calcite particles flatten with time because calcite crystals nucleate at the surface of glass beads and grow into larger calcite grains. At the later stage of the calcite precipitation experiment, modelled sizes and cementation exponents of calcite particles decrease with time because large calcite grains aggregate over multiple glass beads and only small calcite crystals polarize.

  11. Adjoint-Based Algorithms for Adaptation and Design Optimizations on Unstructured Grids

    NASA Technical Reports Server (NTRS)

    Nielsen, Eric J.

    2006-01-01

    Schemes based on discrete adjoint algorithms present several exciting opportunities for significantly advancing the current state of the art in computational fluid dynamics. Such methods provide an extremely efficient means for obtaining discretely consistent sensitivity information for hundreds of design variables, opening the door to rigorous, automated design optimization of complex aerospace configuration using the Navier-Stokes equation. Moreover, the discrete adjoint formulation provides a mathematically rigorous foundation for mesh adaptation and systematic reduction of spatial discretization error. Error estimates are also an inherent by-product of an adjoint-based approach, valuable information that is virtually non-existent in today's large-scale CFD simulations. An overview of the adjoint-based algorithm work at NASA Langley Research Center is presented, with examples demonstrating the potential impact on complex computational problems related to design optimization as well as mesh adaptation.

  12. Large-Scale SRM Screen of Urothelial Bladder Cancer Candidate Biomarkers in Urine.

    PubMed

    Duriez, Elodie; Masselon, Christophe D; Mesmin, Cédric; Court, Magali; Demeure, Kevin; Allory, Yves; Malats, Núria; Matondo, Mariette; Radvanyi, François; Garin, Jérôme; Domon, Bruno

    2017-04-07

    Urothelial bladder cancer is a condition associated with high recurrence and substantial morbidity and mortality. Noninvasive urinary tests that would detect bladder cancer and tumor recurrence are required to significantly improve patient care. Over the past decade, numerous bladder cancer candidate biomarkers have been identified in the context of extensive proteomics or transcriptomics studies. To translate these findings in clinically useful biomarkers, the systematic evaluation of these candidates remains the bottleneck. Such evaluation involves large-scale quantitative LC-SRM (liquid chromatography-selected reaction monitoring) measurements, targeting hundreds of signature peptides by monitoring thousands of transitions in a single analysis. The design of highly multiplexed SRM analyses is driven by several factors: throughput, robustness, selectivity and sensitivity. Because of the complexity of the samples to be analyzed, some measurements (transitions) can be interfered by coeluting isobaric species resulting in biased or inconsistent estimated peptide/protein levels. Thus the assessment of the quality of SRM data is critical to allow flagging these inconsistent data. We describe an efficient and robust method to process large SRM data sets, including the processing of the raw data, the detection of low-quality measurements, the normalization of the signals for each protein, and the estimation of protein levels. Using this methodology, a variety of proteins previously associated with bladder cancer have been assessed through the analysis of urine samples from a large cohort of cancer patients and corresponding controls in an effort to establish a priority list of most promising candidates to guide subsequent clinical validation studies.

  13. Oxalate metal complexes in aerosol particles: implications for the hygroscopicity of oxalate-containing particles

    NASA Astrophysics Data System (ADS)

    Furukawa, T.; Takahashi, Y.

    2011-05-01

    Atmospheric aerosols have both a direct and an indirect cooling effect that influences the radiative balance at the Earth's surface. It has been estimated that the degree of cooling is large enough to weaken the warming effect of carbon dioxide. Among the cooling factors, secondary organic aerosols (SOA) play an important role in the solar radiation balance in the troposphere as SOA can act as cloud condensation nuclei (CCN) and extend the lifespan of clouds because of their high hygroscopic and water soluble nature. Oxalic acid is an important component of SOA, and is produced via several formation pathways in the atmosphere. However, it is not certain whether oxalic acid exists as free oxalic acid or as metal oxalate complexes in aerosols, although there is a marked difference in their solubility in water and their hygroscopicity. We employed X-ray absorption fine structure spectroscopy to characterize the calcium (Ca) and zinc (Zn) in aerosols collected at Tsukuba in Japan. Size-fractionated aerosol samples were collected for this purpose using an impactor aerosol sampler. It was shown that 10-60% and 20-100% of the total Ca and Zn in the finer particles (<2.1 μm) were present as Ca and Zn oxalate complexes, respectively. Oxalic acid is hygroscopic and can thus increase the CCN activity of aerosol particles, while complexes with various polyvalent metal ions such as Ca and Zn are not hygroscopic, which cannot contribute to the increase of the CCN activity of aerosols. Based on the concentrations of noncomplexed and metal-complexed oxalate species, we found that most of the oxalic acid is present as metal oxalate complexes in the aerosols, suggesting that oxalic acid does not always increase the hygroscopicity of aerosols in the atmosphere. Similar results are expected for other dicarboxylic acids, such as malonic and succinic acids. Thus, it is advisable that the cooling effect of organic aerosols should be estimated by including the information on metal oxalate complexes and metal complexes with other dicarboxylic acids in aerosols.

  14. A proposed framework for consensus-based lung tumour volume auto-segmentation in 4D computed tomography imaging

    NASA Astrophysics Data System (ADS)

    Martin, Spencer; Brophy, Mark; Palma, David; Louie, Alexander V.; Yu, Edward; Yaremko, Brian; Ahmad, Belal; Barron, John L.; Beauchemin, Steven S.; Rodrigues, George; Gaede, Stewart

    2015-02-01

    This work aims to propose and validate a framework for tumour volume auto-segmentation based on ground-truth estimates derived from multi-physician input contours to expedite 4D-CT based lung tumour volume delineation. 4D-CT datasets of ten non-small cell lung cancer (NSCLC) patients were manually segmented by 6 physicians. Multi-expert ground truth (GT) estimates were constructed using the STAPLE algorithm for the gross tumour volume (GTV) on all respiratory phases. Next, using a deformable model-based method, multi-expert GT on each individual phase of the 4D-CT dataset was propagated to all other phases providing auto-segmented GTVs and motion encompassing internal gross target volumes (IGTVs) based on GT estimates (STAPLE) from each respiratory phase of the 4D-CT dataset. Accuracy assessment of auto-segmentation employed graph cuts for 3D-shape reconstruction and point-set registration-based analysis yielding volumetric and distance-based measures. STAPLE-based auto-segmented GTV accuracy ranged from (81.51  ±  1.92) to (97.27  ±  0.28)% volumetric overlap of the estimated ground truth. IGTV auto-segmentation showed significantly improved accuracies with reduced variance for all patients ranging from 90.87 to 98.57% volumetric overlap of the ground truth volume. Additional metrics supported these observations with statistical significance. Accuracy of auto-segmentation was shown to be largely independent of selection of the initial propagation phase. IGTV construction based on auto-segmented GTVs within the 4D-CT dataset provided accurate and reliable target volumes compared to manual segmentation-based GT estimates. While inter-/intra-observer effects were largely mitigated, the proposed segmentation workflow is more complex than that of current clinical practice and requires further development.

  15. A proposed framework for consensus-based lung tumour volume auto-segmentation in 4D computed tomography imaging.

    PubMed

    Martin, Spencer; Brophy, Mark; Palma, David; Louie, Alexander V; Yu, Edward; Yaremko, Brian; Ahmad, Belal; Barron, John L; Beauchemin, Steven S; Rodrigues, George; Gaede, Stewart

    2015-02-21

    This work aims to propose and validate a framework for tumour volume auto-segmentation based on ground-truth estimates derived from multi-physician input contours to expedite 4D-CT based lung tumour volume delineation. 4D-CT datasets of ten non-small cell lung cancer (NSCLC) patients were manually segmented by 6 physicians. Multi-expert ground truth (GT) estimates were constructed using the STAPLE algorithm for the gross tumour volume (GTV) on all respiratory phases. Next, using a deformable model-based method, multi-expert GT on each individual phase of the 4D-CT dataset was propagated to all other phases providing auto-segmented GTVs and motion encompassing internal gross target volumes (IGTVs) based on GT estimates (STAPLE) from each respiratory phase of the 4D-CT dataset. Accuracy assessment of auto-segmentation employed graph cuts for 3D-shape reconstruction and point-set registration-based analysis yielding volumetric and distance-based measures. STAPLE-based auto-segmented GTV accuracy ranged from (81.51  ±  1.92) to (97.27  ±  0.28)% volumetric overlap of the estimated ground truth. IGTV auto-segmentation showed significantly improved accuracies with reduced variance for all patients ranging from 90.87 to 98.57% volumetric overlap of the ground truth volume. Additional metrics supported these observations with statistical significance. Accuracy of auto-segmentation was shown to be largely independent of selection of the initial propagation phase. IGTV construction based on auto-segmented GTVs within the 4D-CT dataset provided accurate and reliable target volumes compared to manual segmentation-based GT estimates. While inter-/intra-observer effects were largely mitigated, the proposed segmentation workflow is more complex than that of current clinical practice and requires further development.

  16. Heavy-Tailed Fluctuations in the Spiking Output Intensity of Semiconductor Lasers with Optical Feedback

    PubMed Central

    2016-01-01

    Although heavy-tailed fluctuations are ubiquitous in complex systems, a good understanding of the mechanisms that generate them is still lacking. Optical complex systems are ideal candidates for investigating heavy-tailed fluctuations, as they allow recording large datasets under controllable experimental conditions. A dynamical regime that has attracted a lot of attention over the years is the so-called low-frequency fluctuations (LFFs) of semiconductor lasers with optical feedback. In this regime, the laser output intensity is characterized by abrupt and apparently random dropouts. The statistical analysis of the inter-dropout-intervals (IDIs) has provided many useful insights into the underlying dynamics. However, the presence of large temporal fluctuations in the IDI sequence has not yet been investigated. Here, by applying fluctuation analysis we show that the experimental distribution of IDI fluctuations is heavy-tailed, and specifically, is well-modeled by a non-Gaussian stable distribution. We find a good qualitative agreement with simulations of the Lang-Kobayashi model. Moreover, we uncover a transition from a less-heavy-tailed state at low pump current to a more-heavy-tailed state at higher pump current. Our results indicate that fluctuation analysis can be a useful tool for investigating the output signals of complex optical systems; it can be used for detecting underlying regime shifts, for model validation and parameter estimation. PMID:26901346

  17. Investigation of mesoscale precipitation processes in the Carolinas using a radar-based climatology

    NASA Astrophysics Data System (ADS)

    Boyles, Ryan Patrick

    The complex topography, shoreline, soils, and land use patterns makes the Carolinas a unique location to study mesoscale processes. Using gage-calibrated radar estimates and a series of numerical model simulations, warm season mesoscale precipitation patterns are analyzed over the Carolinas. Gage-calibrated radar precipitation estimates are compared with surface gage observations. Stage IV estimates generally compared better than Stage II estimates, but some Stage II and Stage IV estimates have gross errors during autumn, winter, and spring seasons. Analysis of days when sea breeze is observed suggests that sea breeze induced precipitation occurs on nearly 40% of days in June, July, and August, but only 18% in May and 6% of days in April. Precipitation on days with sea breeze convection can contribute to over 50% of seasonal precipitation. Rainfall associated with sea breeze is generally maximized along east-facing shores 10-20 km inland, and minimized along south-facing shores in North Carolina. The shape of the shoreline along Cape Fear is associated with a local precipitation maximum that may be caused by the convergence of two sea breeze fronts from the south and east shores. Differential heating associated with contrasting soils along the Carolina Sandhills is suggested as a mechanism for enhancement in local precipitation. A high-resolution summer precipitation climatology suggests that precipitation is enhanced along the Sandhills region in both wet and dry years. Analysis of four numerical simulations suggests that contrasts in soils over the Carolinas Sandhills dominates over vegetation contrasts to produce heat flux gradients and a convergence zone along the sand-to-clay transition. Orographically induced precipitation is consistently observed in the summer, and appears to be isolated along windward slopes at 20km--40km from the ridge line. Amounts over external ridges are generally 50-100% higher than amounts observed over the foothills. Precipitation amounts over interior ridges and valleys are lower than observed on exterior ridges and are similar to values observed over the foothills. When compared with Stage IV estimates, the PRISM (Precipitation-elevation Regressions on Independent Slopes Model) method for estimating precipitation in complex terrain appears to largely over-estimate precipitation amounts over the interior ridges.

  18. Realizing actual feedback control of complex network

    NASA Astrophysics Data System (ADS)

    Tu, Chengyi; Cheng, Yuhua

    2014-06-01

    In this paper, we present the concept of feedbackability and how to identify the Minimum Feedbackability Set of an arbitrary complex directed network. Furthermore, we design an estimator and a feedback controller accessing one MFS to realize actual feedback control, i.e. control the system to our desired state according to the estimated system internal state from the output of estimator. Last but not least, we perform numerical simulations of a small linear time-invariant dynamics network and a real simple food network to verify the theoretical results. The framework presented here could make an arbitrary complex directed network realize actual feedback control and deepen our understanding of complex systems.

  19. Parameter Estimation in Epidemiology: from Simple to Complex Dynamics

    NASA Astrophysics Data System (ADS)

    Aguiar, Maíra; Ballesteros, Sebastién; Boto, João Pedro; Kooi, Bob W.; Mateus, Luís; Stollenwerk, Nico

    2011-09-01

    We revisit the parameter estimation framework for population biological dynamical systems, and apply it to calibrate various models in epidemiology with empirical time series, namely influenza and dengue fever. When it comes to more complex models like multi-strain dynamics to describe the virus-host interaction in dengue fever, even most recently developed parameter estimation techniques, like maximum likelihood iterated filtering, come to their computational limits. However, the first results of parameter estimation with data on dengue fever from Thailand indicate a subtle interplay between stochasticity and deterministic skeleton. The deterministic system on its own already displays complex dynamics up to deterministic chaos and coexistence of multiple attractors.

  20. Networks consolidation program: Maintenance and Operations (M&O) staffing estimates

    NASA Technical Reports Server (NTRS)

    Goodwin, J. P.

    1981-01-01

    The Mark IV-A consolidate deep space and high elliptical Earth orbiter (HEEO) missions tracking and implements centralized control and monitoring at the deep space communications complexes (DSCC). One of the objectives of the network design is to reduce maintenance and operations (M&O) costs. To determine if the system design meets this objective an M&O staffing model for Goldstone was developed which was used to estimate the staffing levels required to support the Mark IV-A configuration. The study was performed for the Goldstone complex and the program office translated these estimates for the overseas complexes to derive the network estimates.

  1. Complex sample survey estimation in static state-space

    Treesearch

    Raymond L. Czaplewski

    2010-01-01

    Increased use of remotely sensed data is a key strategy adopted by the Forest Inventory and Analysis Program. However, multiple sensor technologies require complex sampling units and sampling designs. The Recursive Restriction Estimator (RRE) accommodates this complexity. It is a design-consistent Empirical Best Linear Unbiased Prediction for the state-vector, which...

  2. A new analytical method for characterizing nonlinear visual processes with stimuli of arbitrary distribution: Theory and applications.

    PubMed

    Hayashi, Ryusuke; Watanabe, Osamu; Yokoyama, Hiroki; Nishida, Shin'ya

    2017-06-01

    Characterization of the functional relationship between sensory inputs and neuronal or observers' perceptual responses is one of the fundamental goals of systems neuroscience and psychophysics. Conventional methods, such as reverse correlation and spike-triggered data analyses are limited in their ability to resolve complex and inherently nonlinear neuronal/perceptual processes because these methods require input stimuli to be Gaussian with a zero mean. Recent studies have shown that analyses based on a generalized linear model (GLM) do not require such specific input characteristics and have advantages over conventional methods. GLM, however, relies on iterative optimization algorithms and its calculation costs become very expensive when estimating the nonlinear parameters of a large-scale system using large volumes of data. In this paper, we introduce a new analytical method for identifying a nonlinear system without relying on iterative calculations and yet also not requiring any specific stimulus distribution. We demonstrate the results of numerical simulations, showing that our noniterative method is as accurate as GLM in estimating nonlinear parameters in many cases and outperforms conventional, spike-triggered data analyses. As an example of the application of our method to actual psychophysical data, we investigated how different spatiotemporal frequency channels interact in assessments of motion direction. The nonlinear interaction estimated by our method was consistent with findings from previous vision studies and supports the validity of our method for nonlinear system identification.

  3. Modeling complex aquifer systems: a case study in Baton Rouge, Louisiana (USA)

    NASA Astrophysics Data System (ADS)

    Pham, Hai V.; Tsai, Frank T.-C.

    2017-05-01

    This study targets two challenges in groundwater model development: grid generation and model calibration for aquifer systems that are fluvial in origin. Realistic hydrostratigraphy can be developed using a large quantity of well log data to capture the complexity of an aquifer system. However, generating valid groundwater model grids to be consistent with the complex hydrostratigraphy is non-trivial. Model calibration can also become intractable for groundwater models that intend to match the complex hydrostratigraphy. This study uses the Baton Rouge aquifer system, Louisiana (USA), to illustrate a technical need to cope with grid generation and model calibration issues. A grid generation technique is introduced based on indicator kriging to interpolate 583 wireline well logs in the Baton Rouge area to derive a hydrostratigraphic architecture with fine vertical discretization. Then, an upscaling procedure is developed to determine a groundwater model structure with 162 layers that captures facies geometry in the hydrostratigraphic architecture. To handle model calibration for such a large model, this study utilizes a derivative-free optimization method in parallel computing to complete parameter estimation in a few months. The constructed hydrostratigraphy indicates the Baton Rouge aquifer system is fluvial in origin. The calibration result indicates hydraulic conductivity for Miocene sands is higher than that for Pliocene to Holocene sands and indicates the Baton Rouge fault and the Denham Springs-Scotlandville fault to be low-permeability leaky aquifers. The modeling result shows significantly low groundwater level in the "2,000-foot" sand due to heavy pumping, indicating potential groundwater upward flow from the "2,400-foot" sand.

  4. Molecular Gas toward the Gemini OB1 Molecular Cloud Complex. II. CO Outflow Candidates with Possible WISE Associations

    NASA Astrophysics Data System (ADS)

    Li, Yingjie; Li, Fa-Cheng; Xu, Ye; Wang, Chen; Du, Xin-Yu; Yang, Wenjin; Yang, Ji

    2018-03-01

    We present a large-scale survey of CO outflows in the Gem OB1 molecular cloud complex and its surroundings, using the Purple Mountain Observatory Delingha 13.7 m telescope. A total of 198 outflow candidates were identified over a large area (∼58.5 square degrees), of which 193 are newly detected. Approximately 68% (134/198) are associated with the Gem OB1 molecular cloud complex, including clouds GGMC 1, GGMC 2, BFS 52, GGMC 3, and GGMC 4. Other regions studied are: the Local arm (Local Lynds, West Front), Swallow, Horn, and Remote cloud. Outflow candidates in GGMC 1, BFS 52, and Swallow are mainly located at ring-like or filamentary structures. To avoid excessive uncertainty in distant regions (≳3.8 kpc), we only estimated the physical parameters for clouds in the Gem OB1 molecular cloud complex and in the Local arm. In those clouds, the total kinetic energy and the energy injection rate of the identified outflow candidates are ≲1% and ≲3% of the turbulent energy and the turbulent dissipation rate of each cloud, indicating that the identified outflow candidates cannot provide enough energy to balance turbulence of their host cloud at the scale of the entire cloud (several to dozens of parsecs). The gravitational binding energy of each cloud is ≳135 times the total kinetic energy of the identified outflow candidates within the corresponding cloud, indicating that the identified outflow candidates cannot cause major disruptions to the integrity of their host cloud at the scale of the entire cloud.

  5. Evaluating a complex system-wide intervention using the difference in differences method: the Delivering Choice Programme

    PubMed Central

    Round, Jeff; Drake, Robyn; Kendall, Edward; Addicott, Rachael; Agelopoulos, Nicky; Jones, Louise

    2015-01-01

    Objectives We report the use of difference in differences (DiD) methodology to evaluate a complex, system-wide healthcare intervention. We use the worked example of evaluating the Marie Curie Delivering Choice Programme (DCP) for advanced illness in a large urban healthcare economy. Methods DiD was selected because a randomised controlled trial was not feasible. The method allows for before and after comparison of changes that occur in an intervention site with a matched control site. This enables analysts to control for the effect of the intervention in the absence of a local control. Any policy, seasonal or other confounding effects over the test period are assumed to have occurred in a balanced way at both sites. Data were obtained from primary care trusts. Outcomes were place of death, inpatient admissions, length of stay and costs. Results Small changes were identified between pre- and post-DCP outputs in the intervention site. The proportion of home deaths and median cost increased slightly, while the number of admissions per patient and the average length of stay per admission decreased slightly. None of these changes was statistically significant. Conclusions Effects estimates were limited by small numbers accessing new services and selection bias in sample population and comparator site. In evaluating the effect of a complex healthcare intervention, the choice of analysis method and output measures is crucial. Alternatives to randomised controlled trials may be required for evaluating large scale complex interventions and the DiD approach is suitable, subject to careful selection of measured outputs and control population. PMID:24644163

  6. Dynamic ruptures on faults of complex geometry: insights from numerical simulations, from large-scale curvature to small-scale fractal roughness

    NASA Astrophysics Data System (ADS)

    Ulrich, T.; Gabriel, A. A.

    2016-12-01

    The geometry of faults is subject to a large degree of uncertainty. As buried structures being not directly observable, their complex shapes may only be inferred from surface traces, if available, or through geophysical methods, such as reflection seismology. As a consequence, most studies aiming at assessing the potential hazard of faults rely on idealized fault models, based on observable large-scale features. Yet, real faults are known to be wavy at all scales, their geometric features presenting similar statistical properties from the micro to the regional scale. The influence of roughness on the earthquake rupture process is currently a driving topic in the computational seismology community. From the numerical point of view, rough faults problems are challenging problems that require optimized codes able to run efficiently on high-performance computing infrastructure and simultaneously handle complex geometries. Physically, simulated ruptures hosted by rough faults appear to be much closer to source models inverted from observation in terms of complexity. Incorporating fault geometry on all scales may thus be crucial to model realistic earthquake source processes and to estimate more accurately seismic hazard. In this study, we use the software package SeisSol, based on an ADER-Discontinuous Galerkin scheme, to run our numerical simulations. SeisSol allows solving the spontaneous dynamic earthquake rupture problem and the wave propagation problem with high-order accuracy in space and time efficiently on large-scale machines. In this study, the influence of fault roughness on dynamic rupture style (e.g. onset of supershear transition, rupture front coherence, propagation of self-healing pulses, etc) at different length scales is investigated by analyzing ruptures on faults of varying roughness spectral content. In particular, we investigate the existence of a minimum roughness length scale in terms of rupture inherent length scales below which the rupture ceases to be sensible. Finally, the effect of fault geometry on ground-motions, in the near-field, is considered. Our simulations feature a classical linear slip weakening on the fault and a viscoplastic constitutive model off the fault. The benefits of using a more elaborate fast velocity-weakening friction law will also be considered.

  7. Streamflow model of the six-country transboundary Ganges-Bhramaputra and Meghna river basin

    NASA Astrophysics Data System (ADS)

    Rahman, K.; Lehmann, A.; Dennedy-Frank, P. J.; Gorelick, S.

    2014-12-01

    Extremely large-scale river basin modelling remains a challenge for water resources planning in the developing world. Such planning is particularly difficult in the developing world because of the lack of data on both natural (climatological, hydrological) processes and complex anthropological influences. We simulate three enormous river basins located in south Asia. The Ganges-Bhramaputra and Meghna (GBM) River Basins cover an area of 1.75 million km2 associated with 6 different countries, including the Bengal delta, which is the most densely populated delta in the world with ~600 million people. We target this developing region to better understand the hydrological system and improve water management planning in these transboundary watersheds. This effort uses the Soil and Water Assessment Tool (SWAT) to simulate streamflow in the GBM River Basins and assess the use of global climatological datasets for such large scale river modeling. We evaluate the utility of three global rainfall datasets to reproduce measured river discharge: the Tropical Rainfall Measuring Mission (TRMM) from NASA, the National Centers for Environmental Prediction (NCEP) reanalysis, and the World Metrological Organization (WMO) reanalysis. We use global datasets for spatial information as well: 90m DEM from the Shuttle Radar Topographic Mission, 300m GlobCover land use maps, and 1000 km FAO soil map. We find that SWAT discharge estimates match the observed streamflow well (NSE=0.40-0.66, R2=0.60-0.70) when using meteorological estimates from the NCEP reanalysis. However, SWAT estimates diverge from observed discharge when using meteorological estimates from TRMM and the WMO reanalysis.

  8. Spatio-temporal optimization of sampling for bluetongue vectors (Culicoides) near grazing livestock

    PubMed Central

    2013-01-01

    Background Estimating the abundance of Culicoides using light traps is influenced by a large variation in abundance in time and place. This study investigates the optimal trapping strategy to estimate the abundance or presence/absence of Culicoides on a field with grazing animals. We used 45 light traps to sample specimens from the Culicoides obsoletus species complex on a 14 hectare field during 16 nights in 2009. Findings The large number of traps and catch nights enabled us to simulate a series of samples consisting of different numbers of traps (1-15) on each night. We also varied the number of catch nights when simulating the sampling, and sampled with increasing minimum distances between traps. We used resampling to generate a distribution of different mean and median abundance in each sample. Finally, we used the hypergeometric distribution to estimate the probability of falsely detecting absence of vectors on the field. The variation in the estimated abundance decreased steeply when using up to six traps, and was less pronounced when using more traps, although no clear cutoff was found. Conclusions Despite spatial clustering in vector abundance, we found no effect of increasing the distance between traps. We found that 18 traps were generally required to reach 90% probability of a true positive catch when sampling just one night. But when sampling over two nights the same probability level was obtained with just three traps per night. The results are useful for the design of vector monitoring programmes on fields with grazing animals. PMID:23705770

  9. Use of n-butyl cyanoacrylate to reduce left to right shunting of an abdominal arteriovenous malformation in a dog.

    PubMed

    Eason, B D; Hogan, D F; Lim, C; Hogan, M J

    2017-08-01

    A 9-month old castrated male Labradoodle presented to the cardiology service at Purdue University for evaluation of a low-grade murmur. Physical examination, thoracic radiography, and echocardiography were strongly supportive of an extracardiac left-to-right shunt. Subsequent evaluation with nuclear scintigraphy and computed tomography angiography revealed a large, complex arteriovenous malformation within the cranial abdomen. Staged interventional attenuation of the shunt was performed using n-butyl cyanoacrylate that resulted in a reduction in echocardiographic and nuclear scintigraphy derived shunt estimation. Copyright © 2017 Elsevier B.V. All rights reserved.

  10. Sobol Sensitivity Analysis: A Tool to Guide the Development and Evaluation of Systems Pharmacology Models

    PubMed Central

    Trame, MN; Lesko, LJ

    2015-01-01

    A systems pharmacology model typically integrates pharmacokinetic, biochemical network, and systems biology concepts into a unifying approach. It typically consists of a large number of parameters and reaction species that are interlinked based upon the underlying (patho)physiology and the mechanism of drug action. The more complex these models are, the greater the challenge of reliably identifying and estimating respective model parameters. Global sensitivity analysis provides an innovative tool that can meet this challenge. CPT Pharmacometrics Syst. Pharmacol. (2015) 4, 69–79; doi:10.1002/psp4.6; published online 25 February 2015 PMID:27548289

  11. Mars: Fretted and chaotic terrains

    NASA Technical Reports Server (NTRS)

    Sharp, R. P.

    1973-01-01

    Fretted Martian terrain is characterized by smooth, flat, lowland areas separated from a cratered upland by abrupt escarpments of complex planimetric configuration and a maximum estimated height approaching 1 to 2 km. It is the product of some unusual erosive or abstractive process that has created steep escarpments. Chaotic terrain differs from fretted terrain in having a rough floor topography featuring a haphazard jumble of large angular blocks, and by arc-shaped slump blocks on its bounding escarpments. Its existence has now been confirmed by Mariner 9 pictures, and the characteristics, location, and areal extent of chaotic terrain have been more accurately and completely defined.

  12. Combined analysis of modeled and monitored SO2 concentrations at a complex smelting facility.

    PubMed

    Rehbein, Peter J G; Kennedy, Michael G; Cotsman, David J; Campeau, Madonna A; Greenfield, Monika M; Annett, Melissa A; Lepage, Mike F

    2014-03-01

    Vale Canada Limited owns and operates a large nickel smelting facility located in Sudbury, Ontario. This is a complex facility with many sources of SO2 emissions, including a mix of source types ranging from passive building roof vents to North America's tallest stack. In addition, as this facility performs batch operations, there is significant variability in the emission rates depending on the operations that are occurring. Although SO2 emission rates for many of the sources have been measured by source testing, the reliability of these emission rates has not been tested from a dispersion modeling perspective. This facility is a significant source of SO2 in the local region, making it critical that when modeling the emissions from this facility for regulatory or other purposes, that the resulting concentrations are representative of what would actually be measured or otherwise observed. To assess the accuracy of the modeling, a detailed analysis of modeled and monitored data for SO2 at the facility was performed. A mobile SO2 monitor sampled at five locations downwind of different source groups for different wind directions resulting in a total of 168 hr of valid data that could be used for the modeled to monitored results comparison. The facility was modeled in AERMOD (American Meteorological Society/U.S. Environmental Protection Agency Regulatory Model) using site-specific meteorological data such that the modeled periods coincided with the same times as the monitored events. In addition, great effort was invested into estimating the actual SO2 emission rates that would likely be occurring during each of the monitoring events. SO2 concentrations were modeled for receptors around each monitoring location so that the modeled data could be directly compared with the monitored data. The modeled and monitored concentrations were compared and showed that there were no systematic biases in the modeled concentrations. This paper is a case study of a Combined Analysis of Modelled and Monitored Data (CAMM), which is an approach promulgated within air quality regulations in the Province of Ontario, Canada. Although combining dispersion models and monitoring data to estimate or refine estimates of source emission rates is not a new technique, this study shows how, with a high degree of rigor in the design of the monitoring and filtering of the data, it can be applied to a large industrial facility, with a variety of emission sources. The comparison of modeled and monitored SO2 concentrations in this case study also provides an illustration of the AERMOD model performance for a large industrial complex with many sources, at short time scales in comparison with monitored data. Overall, this analysis demonstrated that the AERMOD model performed well.

  13. A Comparison of Multisensor Precipitation Estimation Methods in Complex Terrain for Flash Flood Warning and Mitigation

    NASA Astrophysics Data System (ADS)

    Cifelli, R.; Chen, H.; Chandrasekar, C. V.; Willie, D.; Reynolds, D.; Campbell, C.; Zhang, Y.; Sukovich, E.

    2012-12-01

    Investigating the uncertainties and improving the accuracy of quantitative precipitation estimation (QPE) is a critical mission of the National Oceanic and Atmospheric Administration (NOAA). QPE is extremely challenging in regions of complex terrain like the western U.S. because of the sparse coverage of ground-based radar, complex orographic precipitation processes, and the effects of beam blockages (e.g., Westrick et al. 1999). In addition, the rain gauge density in complex terrain is often inadequate to capture spatial variability in the precipitation patterns. The NOAA Hydrometeorology Testbed (HMT) conducts research on precipitation and weather conditions that can lead to flooding, and fosters transition of scientific advances and new tools into forecasting operations (see hmt.noaa.gov). The HMT program consists of a series of demonstration projects in different geographical regions to enhance understanding of region specific processes related to precipitation, including QPE. There are a number of QPE systems that are widely used across NOAA for precipitation estimation (e.g., Cifelli et al. 2011; Chandrasekar et al. 2012). Two of these systems have been installed at the NOAA Earth System Research Laboratory: Multisensor Precipitation Estimator (MPE) and National Mosaic and Multi-sensor QPE (NMQ) developed by NWS and NSSL, respectively. Both provide gridded QPE products that include radar-only, gauge-only and gauge-radar-merged, etc; however, these systems often provide large differences in QPE (in terms of amounts and spatial patterns) due to differences in Z-R selection, vertical profile of reflectivity correction, and gauge interpolation procedures. Determining the appropriate QPE product and quantification of QPE uncertainty is critical for operational applications, including water management decisions and flood warnings. For example, hourly QPE is used to correct radar based rain rates used by the Flash Flood Monitoring and Prediction (FFMP) package in the NWS forecast offices for issuance of flash flood warnings. This study will evaluate the performance of MPE and NMQ QPE products using independent gauges, object identification techniques for spatial verification and impact on surface runoff using a distributed hydrologic model. The effort will consist of baseline evaluations of these QPE systems to determine which combination of algorithm features is appropriate as well as investigate new methods for combining the gage and radar data. The Russian River Basin in California is used to demonstrate the comparison methodology with data collected from several rainfall events in March 2012.

  14. Removal of waterborne microorganisms by filtration using clay-polymer complexes.

    PubMed

    Undabeytia, Tomas; Posada, Rosa; Nir, Shlomo; Galindo, Irene; Laiz, Leonila; Saiz-Jimenez, Cesareo; Morillo, Esmeralda

    2014-08-30

    Clay-polymer composites were designed for use in filtration processes for disinfection during the course of water purification. The composites were formed by sorption of polymers based on starch modified with quaternary ammonium ethers onto the negatively charged clay mineral bentonite. The performance of the clay-polymer complexes in removal of bacteria was strongly dependent on the conformation adopted by the polycation on the clay surface, the charge density of the polycation itself and the ratio between the concentrations of clay and polymer used during the sorption process. The antimicrobial effect exerted by the clay-polymer system was due to the cationic monomers adsorbed on the clay surface, which resulted in a positive surface potential of the complexes and charge reversal. Clay-polymer complexes were more toxic to bacteria than the polymers alone. Filtration employing our optimal clay-polymer composite yielded 100% removal of bacteria after the passage of 3L, whereas an equivalent filter with granular activated carbon (GAC) hardly yielded removal of bacteria after 0.5L. Regeneration of clay-polymer complexes saturated with bacteria was demonstrated. Modeling of the filtration processes permitted to optimize the design of filters and estimation of experimental conditions for purifying large water volumes in short periods. Copyright © 2014 Elsevier B.V. All rights reserved.

  15. Affinity and specificity of serine endopeptidase-protein inhibitor interactions. Empirical free energy calculations based on X-ray crystallographic structures.

    PubMed

    Krystek, S; Stouch, T; Novotny, J

    1993-12-05

    An empirical function was used to calculate free energy change (delta G) of complex formation between the following inhibitors and enzymes: Kunitz inhibitor (BPTI) with trypsin, trypsinogen and kallikrein; turkey ovomucoid 3rd domain (OMTKY3) with alpha-chymotrypsin and the Streptomyces griseus protease B; the potato chymotrypsin inhibitor with the protease B; and the barely chymotrypsin inhibitor and eglin-c with subtilisin and thermitase. Using X-ray coordinates of the nine complexes, we estimated the contributions that hydrophobic effect, electrostatic interactions and side-chain conformational entropy make towards the stability of the complexes. The calculated delta G values showed good agreement with the experimentally measured ones, the only exception being the kallikrein/BPTI complex whose X-ray structure was solved at an exceptionally low pH. In complexes with different enzymes, the same inhibitor residues contributed identically towards complex formation (delta G(residue) Spearman rank correlation coefficient 0.7 to 1.0). The most productive enzyme-contacting residues in OMTKY3, eglin-c, and the chymotrypsin inhibitors were found in analogous positions on their respective binding loops; thus, our calculations identified a functional (energetic) motif that parallels the well-known structural similarity of the binding loops. The delta G values calculated for BPTI complexed with trypsin (-21.7 kcal) and trypsinogen (-23.4 kcal) were similar and close to the experimental delta G value of the trypsin/BPTI complex (-18.1 kcal), lending support to the suggestion that the 10(7) difference in the observed stabilities (KA) of these two complexes reflects the energetic cost of conformational changes induced in trypsinogen during the pre-equilibrium stages of complex formation. In almost all of the complexes studied, the stabilization free energy contributed by the inhibitors was larger than that donated by the enzymes. In the trypsin-BPTI complex, the calculated delta G contribution of the amino group from the BPTI residue Lys15 (9.7 kcal) was somewhat higher than that arrived at in experiments with semisynthetic inhibitor analogs (7.5 kcal). In OMTKY3, different binding loop residues are known to affect differently the binding (delta delta G) to alpha-chymotrypsin and protease B; a good qualitative agreement was found between the calculated delta G(residue) estimates and the experimental delta delta G data (correlation coefficient 0.7). Large variations were observed in local surface complementarity and related interfacial volume in the two OMTKY3 complexes (by 20 to 60% for some side-chains).(ABSTRACT TRUNCATED AT 400 WORDS)

  16. Exogenous factors matter when interpreting the results of an impact evaluation: a case study of rainfall and child health programme intervention in Rwanda.

    PubMed

    Mukabutera, Assumpta; Thomson, Dana R; Hedt-Gauthier, Bethany L; Atwood, Sidney; Basinga, Paulin; Nyirazinyoye, Laetitia; Savage, Kevin P; Habimana, Marcellin; Murray, Megan

    2017-12-01

    Public health interventions are often implemented at large scale, and their evaluation seems to be difficult because they are usually multiple and their pathways to effect are complex and subject to modification by contextual factors. We assessed whether controlling for rainfall-related variables altered estimates of the efficacy of a health programme in rural Rwanda and have a quantifiable effect on an intervention evaluation outcomes. We conducted a retrospective quasi-experimental study using previously collected cross-sectional data from the 2005 and 2010 Rwanda Demographic and Health Surveys (DHS), 2010 DHS oversampled data, monthly rainfall data collected from meteorological stations over the same period, and modelled output of long-term rainfall averages, soil moisture, and rain water run-off. Difference-in-difference models were used. Rainfall factors confounded the PIH intervention impact evaluation. When we adjusted our estimates of programme effect by controlling for a variety of rainfall variables, several effectiveness estimates changed by 10% or more. The analyses that did not adjust for rainfall-related variables underestimated the intervention effect on the prevalence of ARI by 14.3%, fever by 52.4% and stunting by 10.2%. Conversely, the unadjusted analysis overestimated the intervention's effect on diarrhoea by 56.5% and wasting by 80%. Rainfall-related patterns have a quantifiable effect on programme evaluation results and highlighted the importance and complexity of controlling for contextual factors in quasi-experimental design evaluations. © 2017 John Wiley & Sons Ltd.

  17. A place-based model of local activity spaces: individual place exposure and characteristics

    NASA Astrophysics Data System (ADS)

    Hasanzadeh, Kamyar; Laatikainen, Tiina; Kyttä, Marketta

    2018-01-01

    Researchers for long have hypothesized relationships between mobility, urban context, and health. Despite the ample amount of discussions, the empirical findings corroborating such associations remain to be marginal in the literature. It is growingly believed that the weakness of the observed associations can be largely explained by the common misspecification of the geographical context. Researchers coming from different fields have developed a wide range of methods for estimating the extents of these geographical contexts. In this article, we argue that no single approach yet has sufficiently been capable of capturing the complexity of human mobility patterns. Subsequently, we discuss that reaching a better understanding of individual activity spaces can be possible through a spatially sensitive estimation of place exposure. Following this discussion, we take an integrative person and place-based approach to create an individualized residential exposure model (IREM) to estimate the local activity spaces (LAS) of the individuals. This model is created using data collected through public participation GIS. Following a brief comparison of IREM with other commonly used LAS models, the article continues by presenting an empirical study of aging citizens in Helsinki area to demonstrate the usability of the proposed framework. In this study, we identify the main dimensions of LASs and seek their associations with socio-demographic characteristics of individuals and their location in the region. The promising results from comparisons and the interesting findings from the empirical part suggest both a methodological and conceptual improvement in capturing the complexity of local activity spaces.

  18. Multi-factorial analysis of class prediction error: estimating optimal number of biomarkers for various classification rules.

    PubMed

    Khondoker, Mizanur R; Bachmann, Till T; Mewissen, Muriel; Dickinson, Paul; Dobrzelecki, Bartosz; Campbell, Colin J; Mount, Andrew R; Walton, Anthony J; Crain, Jason; Schulze, Holger; Giraud, Gerard; Ross, Alan J; Ciani, Ilenia; Ember, Stuart W J; Tlili, Chaker; Terry, Jonathan G; Grant, Eilidh; McDonnell, Nicola; Ghazal, Peter

    2010-12-01

    Machine learning and statistical model based classifiers have increasingly been used with more complex and high dimensional biological data obtained from high-throughput technologies. Understanding the impact of various factors associated with large and complex microarray datasets on the predictive performance of classifiers is computationally intensive, under investigated, yet vital in determining the optimal number of biomarkers for various classification purposes aimed towards improved detection, diagnosis, and therapeutic monitoring of diseases. We investigate the impact of microarray based data characteristics on the predictive performance for various classification rules using simulation studies. Our investigation using Random Forest, Support Vector Machines, Linear Discriminant Analysis and k-Nearest Neighbour shows that the predictive performance of classifiers is strongly influenced by training set size, biological and technical variability, replication, fold change and correlation between biomarkers. Optimal number of biomarkers for a classification problem should therefore be estimated taking account of the impact of all these factors. A database of average generalization errors is built for various combinations of these factors. The database of generalization errors can be used for estimating the optimal number of biomarkers for given levels of predictive accuracy as a function of these factors. Examples show that curves from actual biological data resemble that of simulated data with corresponding levels of data characteristics. An R package optBiomarker implementing the method is freely available for academic use from the Comprehensive R Archive Network (http://www.cran.r-project.org/web/packages/optBiomarker/).

  19. Estimation of Global Network Statistics from Incomplete Data

    PubMed Central

    Bliss, Catherine A.; Danforth, Christopher M.; Dodds, Peter Sheridan

    2014-01-01

    Complex networks underlie an enormous variety of social, biological, physical, and virtual systems. A profound complication for the science of complex networks is that in most cases, observing all nodes and all network interactions is impossible. Previous work addressing the impacts of partial network data is surprisingly limited, focuses primarily on missing nodes, and suggests that network statistics derived from subsampled data are not suitable estimators for the same network statistics describing the overall network topology. We generate scaling methods to predict true network statistics, including the degree distribution, from only partial knowledge of nodes, links, or weights. Our methods are transparent and do not assume a known generating process for the network, thus enabling prediction of network statistics for a wide variety of applications. We validate analytical results on four simulated network classes and empirical data sets of various sizes. We perform subsampling experiments by varying proportions of sampled data and demonstrate that our scaling methods can provide very good estimates of true network statistics while acknowledging limits. Lastly, we apply our techniques to a set of rich and evolving large-scale social networks, Twitter reply networks. Based on 100 million tweets, we use our scaling techniques to propose a statistical characterization of the Twitter Interactome from September 2008 to November 2008. Our treatment allows us to find support for Dunbar's hypothesis in detecting an upper threshold for the number of active social contacts that individuals maintain over the course of one week. PMID:25338183

  20. DnaSAM: Software to perform neutrality testing for large datasets with complex null models.

    PubMed

    Eckert, Andrew J; Liechty, John D; Tearse, Brandon R; Pande, Barnaly; Neale, David B

    2010-05-01

    Patterns of DNA sequence polymorphisms can be used to understand the processes of demography and adaptation within natural populations. High-throughput generation of DNA sequence data has historically been the bottleneck with respect to data processing and experimental inference. Advances in marker technologies have largely solved this problem. Currently, the limiting step is computational, with most molecular population genetic software allowing a gene-by-gene analysis through a graphical user interface. An easy-to-use analysis program that allows both high-throughput processing of multiple sequence alignments along with the flexibility to simulate data under complex demographic scenarios is currently lacking. We introduce a new program, named DnaSAM, which allows high-throughput estimation of DNA sequence diversity and neutrality statistics from experimental data along with the ability to test those statistics via Monte Carlo coalescent simulations. These simulations are conducted using the ms program, which is able to incorporate several genetic parameters (e.g. recombination) and demographic scenarios (e.g. population bottlenecks). The output is a set of diversity and neutrality statistics with associated probability values under a user-specified null model that are stored in easy to manipulate text file. © 2009 Blackwell Publishing Ltd.

  1. Relation between energy radiation ratio and rupture speed in numerically simulated earthquakes

    NASA Astrophysics Data System (ADS)

    Noda, H.; Lapusta, N.; Kanamori, H.

    2011-12-01

    One of the prominent questions in seismology is energy partitioning during an earthquake. Venkataraman and Kanamori [2004] discussed radiation ratio η_R, the ratio of radiated energy E_R to partial strain energy change ΔW_0 which is the total released strain energy minus the energy that would have been dissipated if a fault had slipped at the final stress. They found positive correlation between η_R and rupture speed in large earthquakes, and compared these data with theoretical estimates from simplified models. The relation between η_R and rupture speed is of great interest since both quantities can be estimated independently although there are large uncertainties. We conduct numerical simulations of dynamic ruptures and study the obtained energy partitioning (and η_R) and averaged rupture speeds V_r. So far, we have considered problems based on TPV103 from the SCEC/USGS Spontaneous Rupture Code Verification Project [Harris et al., 2009, http://scecdata.usc.edu/cvws/], which is a 3-D problem with the possibility of remarkable rate weakening at coseismic slip rates caused by flash heating of microscopic asperities [Rice, 1999]. We study the effect of background shear stress level τ_b and the manner in which rupture is arrested, either in rate-strengthening or unbreakable areas of the fault. Note that rupture speed at each fault point is defined when the rupture is still in progress, while η_R is defined after all dynamic processes such as propagation of a rupture front, healing fronts, and seismic waves have been completed. Those complexities may cause a difference from the theoretical estimates based on simple models, an issue we explore in this study. Overall, our simulations produce the relation between η_R and V_r broadly consistent with the study of Venkataraman and Kanamori (2004) for natural earthquakes and the corresponding theoretical estimates. The model by Mott [1948] agrees best with the cases studied so far, although it is not rigorously correct [Freund, 1990]. For example, a case which is similar to TPV103 except in the nucleation procedure yields a pulse-like rupture with a spatially averaged rupture speed V_r = 0.59 c_s and η_R = 0.32, while the theoretical estimates [Fossum and Freund, 1975 for mode II and Kostrov, 1966; Ehselby, 1969 for mode III] predict η_R of about 0.5 for this rupture speed. This difference is not significant compared with the large observational error. As τ_b increases, V_r increases monotonically, while η_R exhibits more complex behavior: it increases with τ_b for pulse-like ruptures, decreases by about 0.1 at the transition to crack-like ruptures, and then increases again. Frictional dissipation is significant when a rupture front reaches a rate-strengthening region. If the barrier is changed to an unbreakable region, η_R decreases and V_r/c_s increases at most by 0.3 and 0.1, respectively. Although sharper arrest of rupture causes larger E_R per seismic moment due to the stopping phases, ΔW_0 per seismic moment increases more remarkably due to large wavenumber components in final slip distribution.

  2. The effects of riverine physical complexity on anadromy and genetic diversity in steelhead or rainbow trout Oncorhynchus mykiss around the Pacific Rim.

    PubMed

    McPhee, M V; Whited, D C; Kuzishchin, K V; Stanford, J A

    2014-07-01

    This study explored the relationship between riverine physical complexity, as determined from remotely sensed metrics, and anadromy and genetic diversity in steelhead or rainbow trout Oncorhynchus mykiss. The proportion of anadromy (estimated fraction of individuals within a drainage that are anadromous) was correlated with riverine complexity, but this correlation appeared to be driven largely by a confounding negative relationship between drainage area and the proportion of anadromy. Genetic diversity decreased with latitude, was lower in rivers with only non-anadromous individuals and also decreased with an increasing ratio of floodplain area to total drainage area. Anadromy may be less frequent in larger drainages due to the higher cost of migration associated with reaches farther from the ocean, and the negative relationship between genetic diversity and floodplain area may be due to lower effective population size resulting from greater population fluctuations associated with higher rates of habitat turnover. Ultimately, the relationships between riverine physical complexity and migratory life history or genetic diversity probably depend on the spatial scale of analysis. © 2014 The Fisheries Society of the British Isles.

  3. Valuing Insect Pollination Services with Cost of Replacement

    PubMed Central

    Allsopp, Mike H.; de Lange, Willem J.; Veldtman, Ruan

    2008-01-01

    Value estimates of ecosystem goods and services are useful to justify the allocation of resources towards conservation, but inconclusive estimates risk unsustainable resource allocations. Here we present replacement costs as a more accurate value estimate of insect pollination as an ecosystem service, although this method could also be applied to other services. The importance of insect pollination to agriculture is unequivocal. However, whether this service is largely provided by wild pollinators (genuine ecosystem service) or managed pollinators (commercial service), and which of these requires immediate action amidst reports of pollinator decline, remains contested. If crop pollination is used to argue for biodiversity conservation, clear distinction should be made between values of managed- and wild pollination services. Current methods either under-estimate or over-estimate the pollination service value, and make use of criticised general insect and managed pollinator dependence factors. We apply the theoretical concept of ascribing a value to a service by calculating the cost to replace it, as a novel way of valuing wild and managed pollination services. Adjusted insect and managed pollinator dependence factors were used to estimate the cost of replacing insect- and managed pollination services for the Western Cape deciduous fruit industry of South Africa. Using pollen dusting and hand pollination as suitable replacements, we value pollination services significantly higher than current market prices for commercial pollination, although lower than traditional proportional estimates. The complexity associated with inclusive value estimation of pollination services required several defendable assumptions, but made estimates more inclusive than previous attempts. Consequently this study provides the basis for continued improvement in context specific pollination service value estimates. PMID:18781196

  4. Learning multivariate distributions by competitive assembly of marginals.

    PubMed

    Sánchez-Vega, Francisco; Younes, Laurent; Geman, Donald

    2013-02-01

    We present a new framework for learning high-dimensional multivariate probability distributions from estimated marginals. The approach is motivated by compositional models and Bayesian networks, and designed to adapt to small sample sizes. We start with a large, overlapping set of elementary statistical building blocks, or "primitives," which are low-dimensional marginal distributions learned from data. Each variable may appear in many primitives. Subsets of primitives are combined in a Lego-like fashion to construct a probabilistic graphical model; only a small fraction of the primitives will participate in any valid construction. Since primitives can be precomputed, parameter estimation and structure search are separated. Model complexity is controlled by strong biases; we adapt the primitives to the amount of training data and impose rules which restrict the merging of them into allowable compositions. The likelihood of the data decomposes into a sum of local gains, one for each primitive in the final structure. We focus on a specific subclass of networks which are binary forests. Structure optimization corresponds to an integer linear program and the maximizing composition can be computed for reasonably large numbers of variables. Performance is evaluated using both synthetic data and real datasets from natural language processing and computational biology.

  5. A simplified method for power-law modelling of metabolic pathways from time-course data and steady-state flux profiles.

    PubMed

    Kitayama, Tomoya; Kinoshita, Ayako; Sugimoto, Masahiro; Nakayama, Yoichi; Tomita, Masaru

    2006-07-17

    In order to improve understanding of metabolic systems there have been attempts to construct S-system models from time courses. Conventionally, non-linear curve-fitting algorithms have been used for modelling, because of the non-linear properties of parameter estimation from time series. However, the huge iterative calculations required have hindered the development of large-scale metabolic pathway models. To solve this problem we propose a novel method involving power-law modelling of metabolic pathways from the Jacobian of the targeted system and the steady-state flux profiles by linearization of S-systems. The results of two case studies modelling a straight and a branched pathway, respectively, showed that our method reduced the number of unknown parameters needing to be estimated. The time-courses simulated by conventional kinetic models and those described by our method behaved similarly under a wide range of perturbations of metabolite concentrations. The proposed method reduces calculation complexity and facilitates the construction of large-scale S-system models of metabolic pathways, realizing a practical application of reverse engineering of dynamic simulation models from the Jacobian of the targeted system and steady-state flux profiles.

  6. A large coaxial reflection cell for broadband dielectric characterization of coarse-grained materials

    NASA Astrophysics Data System (ADS)

    Bore, Thierry; Bhuyan, Habibullah; Bittner, Tilman; Murgan, Vignesh; Wagner, Norman; Scheuermann, Alexander

    2018-01-01

    Knowledge of the frequency-dependent electromagnetic properties of coarse-grained materials is imperative for the successful application of high frequency electromagnetic measurement techniques for near and subsurface monitoring. This paper reports the design, calibration and application of a novel one-port large coaxial cell for broadband complex permittivity measurements of civil engineering materials. It was designed to allow the characterization of heterogeneous material with large aggregate dimensions (up to 28 mm) over a frequency range from 1 MHz-860 MHz. In the first step, the system parameters were calibrated using the measured scattering function in a perfectly known dielectric material in an optimization scheme. In the second step, the method was validated with measurements made on standard liquids. Then the performance of the cell was evaluated on a compacted coarse-grained soil. The dielectric spectra were obtained by means of fitting the measured scattering function using a transverse electromagnetic mode propagation model considering the frequency-dependent complex permittivity. Two scenarios were systematically analyzed and compared. The first scenario consisted of a broadband generalized dielectric relaxation model with two Cole-Cole type relaxation processes related to the interaction of the aqueous phase and the solid phase, a constant high frequency contribution as well as an apparent direct current conductivity term. The second scenario relied on a three-phase theoretical mixture equation which was used in a forward approach in order to calibrate the model. Both scenarios provide almost identical results for the broadband effective complex relative permittivity. The combination of both scenarios suggests the simultaneous estimation of water content, density, bulk and pore water conductivity for road base materials for in situ applications.

  7. Investigation of the complexity of streamflow fluctuations in a large heterogeneous lake catchment in China

    NASA Astrophysics Data System (ADS)

    Ye, Xuchun; Xu, Chong-Yu; Li, Xianghu; Zhang, Qi

    2018-05-01

    The occurrence of flood and drought frequency is highly correlated with the temporal fluctuations of streamflow series; understanding of these fluctuations is essential for the improved modeling and statistical prediction of extreme changes in river basins. In this study, the complexity of daily streamflow fluctuations was investigated by using multifractal detrended fluctuation analysis (MF-DFA) in a large heterogeneous lake basin, the Poyang Lake basin in China, and the potential impacts of human activities were also explored. Major results indicate that the multifractality of streamflow fluctuations shows significant regional characteristics. In the study catchment, all the daily streamflow series present a strong long-range correlation with Hurst exponents bigger than 0.8. The q-order Hurst exponent h( q) of all the hydrostations can be characterized well by only two parameters: a (0.354 ≤ a ≤ 0.384) and b (0.627 ≤ b ≤ 0.677), with no pronounced differences. Singularity spectrum analysis pointed out that small fluctuations play a dominant role in all daily streamflow series. Our research also revealed that both the correlation properties and the broad probability density function (PDF) of hydrological series can be responsible for the multifractality of streamflow series that depends on watershed areas. In addition, we emphasized the relationship between watershed area and the estimated multifractal parameters, such as the Hurst exponent and fitted parameters a and b from the q-order Hurst exponent h( q). However, the relationship between the width of the singularity spectrum (Δ α) and watershed area is not clear. Further investigation revealed that increasing forest coverage and reservoir storage can effectively enhance the persistence of daily streamflow, decrease the hydrological complexity of large fluctuations, and increase the small fluctuations.

  8. Forecasting production in Liquid Rich Shale plays

    NASA Astrophysics Data System (ADS)

    Nikfarman, Hanieh

    Production from Liquid Rich Shale (LRS) reservoirs is taking center stage in the exploration and production of unconventional reservoirs. Production from the low and ultra-low permeability LRS plays is possible only through multi-fractured horizontal wells (MFHW's). There is no existing workflow that is applicable to forecasting multi-phase production from MFHW's in LRS plays. This project presents a practical and rigorous workflow for forecasting multiphase production from MFHW's in LRS reservoirs. There has been much effort in developing workflows and methodology for forecasting in tight/shale plays in recent years. The existing workflows, however, are applicable only to single phase flow, and are primarily used in shale gas plays. These methodologies do not apply to the multi-phase flow that is inevitable in LRS plays. To account for complexities of multiphase flow in MFHW's the only available technique is dynamic modeling in compositional numerical simulators. These are time consuming and not practical when it comes to forecasting production and estimating reserves for a large number of producers. A workflow was developed, and validated by compositional numerical simulation. The workflow honors physics of flow, and is sufficiently accurate while practical so that an analyst can readily apply it to forecast production and estimate reserves in a large number of producers in a short period of time. To simplify the complex multiphase flow in MFHW, the workflow divides production periods into an initial period where large production and pressure declines are expected, and the subsequent period where production decline may converge into a common trend for a number of producers across an area of interest in the field. Initial period assumes the production is dominated by single-phase flow of oil and uses the tri-linear flow model of Erdal Ozkan to estimate the production history. Commercial software readily available can simulate flow and forecast production in this period. In the subsequent Period, dimensionless rate and dimensionless time functions are introduced that help identify transition from initial period into subsequent period. The production trends in terms of the dimensionless parameters converge for a range of rock permeability and stimulation intensity. This helps forecast production beyond transition to the end of life of well. This workflow is applicable to single fluid system.

  9. A self-sensing active magnetic bearing based on a direct current measurement approach.

    PubMed

    Niemann, Andries C; van Schoor, George; du Rand, Carel P

    2013-09-11

    Active magnetic bearings (AMBs) have become a key technology in various industrial applications. Self-sensing AMBs provide an integrated sensorless solution for position estimation, consolidating the sensing and actuating functions into a single electromagnetic transducer. The approach aims to reduce possible hardware failure points, production costs, and system complexity. Despite these advantages, self-sensing methods must address various technical challenges to maximize the performance thereof. This paper presents the direct current measurement (DCM) approach for self-sensing AMBs, denoting the direct measurement of the current ripple component. In AMB systems, switching power amplifiers (PAs) modulate the rotor position information onto the current waveform. Demodulation self-sensing techniques then use bandpass and lowpass filters to estimate the rotor position from the voltage and current signals. However, the additional phase-shift introduced by these filters results in lower stability margins. The DCM approach utilizes a novel PA switching method that directly measures the current ripple to obtain duty-cycle invariant position estimates. Demodulation filters are largely excluded to minimize additional phase-shift in the position estimates. Basic functionality and performance of the proposed self-sensing approach are demonstrated via a transient simulation model as well as a high current (10 A) experimental system. A digital implementation of amplitude modulation self-sensing serves as a comparative estimator.

  10. Reduction of Topographic Effect for Curve Number Estimated from Remotely Sensed Imagery

    NASA Astrophysics Data System (ADS)

    Zhang, Wen-Yan; Lin, Chao-Yuan

    2016-04-01

    The Soil Conservation Service Curve Number (SCS-CN) method is commonly used in hydrology to estimate direct runoff volume. The CN is the empirical parameter which corresponding to land use/land cover, hydrologic soil group and antecedent soil moisture condition. In large watersheds with complex topography, satellite remote sensing is the appropriate approach to acquire the land use change information. However, the topographic effect have been usually found in the remotely sensed imageries and resulted in land use classification. This research selected summer and winter scenes of Landsat-5 TM during 2008 to classified land use in Chen-You-Lan Watershed, Taiwan. The b-correction, the empirical topographic correction method, was applied to Landsat-5 TM data. Land use were categorized using K-mean classification into 4 groups i.e. forest, grassland, agriculture and river. Accuracy assessment of image classification was performed with national land use map. The results showed that after topographic correction, the overall accuracy of classification was increased from 68.0% to 74.5%. The average CN estimated from remotely sensed imagery decreased from 48.69 to 45.35 where the average CN estimated from national LULC map was 44.11. Therefore, the topographic correction method was recommended to normalize the topographic effect from the satellite remote sensing data before estimating the CN.

  11. Estimating survival probabilities by exposure levels: utilizing vital statistics and complex survey data with mortality follow-up.

    PubMed

    Landsman, V; Lou, W Y W; Graubard, B I

    2015-05-20

    We present a two-step approach for estimating hazard rates and, consequently, survival probabilities, by levels of general categorical exposure. The resulting estimator utilizes three sources of data: vital statistics data and census data are used at the first step to estimate the overall hazard rate for a given combination of gender and age group, and cohort data constructed from a nationally representative complex survey with linked mortality records, are used at the second step to divide the overall hazard rate by exposure levels. We present an explicit expression for the resulting estimator and consider two methods for variance estimation that account for complex multistage sample design: (1) the leaving-one-out jackknife method, and (2) the Taylor linearization method, which provides an analytic formula for the variance estimator. The methods are illustrated with smoking and all-cause mortality data from the US National Health Interview Survey Linked Mortality Files, and the proposed estimator is compared with a previously studied crude hazard rate estimator that uses survey data only. The advantages of a two-step approach and possible extensions of the proposed estimator are discussed. Copyright © 2015 John Wiley & Sons, Ltd.

  12. DEM Based Modeling: Grid or TIN? The Answer Depends

    NASA Astrophysics Data System (ADS)

    Ogden, F. L.; Moreno, H. A.

    2015-12-01

    The availability of petascale supercomputing power has enabled process-based hydrological simulations on large watersheds and two-way coupling with mesoscale atmospheric models. Of course with increasing watershed scale come corresponding increases in watershed complexity, including wide ranging water management infrastructure and objectives, and ever increasing demands for forcing data. Simulations of large watersheds using grid-based models apply a fixed resolution over the entire watershed. In large watersheds, this means an enormous number of grids, or coarsening of the grid resolution to reduce memory requirements. One alternative to grid-based methods is the triangular irregular network (TIN) approach. TINs provide the flexibility of variable resolution, which allows optimization of computational resources by providing high resolution where necessary and low resolution elsewhere. TINs also increase required effort in model setup, parameter estimation, and coupling with forcing data which are often gridded. This presentation discusses the costs and benefits of the use of TINs compared to grid-based methods, in the context of large watershed simulations within the traditional gridded WRF-HYDRO framework and the new TIN-based ADHydro high performance computing watershed simulator.

  13. MATE: Machine Learning for Adaptive Calibration Template Detection

    PubMed Central

    Donné, Simon; De Vylder, Jonas; Goossens, Bart; Philips, Wilfried

    2016-01-01

    The problem of camera calibration is two-fold. On the one hand, the parameters are estimated from known correspondences between the captured image and the real world. On the other, these correspondences themselves—typically in the form of chessboard corners—need to be found. Many distinct approaches for this feature template extraction are available, often of large computational and/or implementational complexity. We exploit the generalized nature of deep learning networks to detect checkerboard corners: our proposed method is a convolutional neural network (CNN) trained on a large set of example chessboard images, which generalizes several existing solutions. The network is trained explicitly against noisy inputs, as well as inputs with large degrees of lens distortion. The trained network that we evaluate is as accurate as existing techniques while offering improved execution time and increased adaptability to specific situations with little effort. The proposed method is not only robust against the types of degradation present in the training set (lens distortions, and large amounts of sensor noise), but also to perspective deformations, e.g., resulting from multi-camera set-ups. PMID:27827920

  14. OLYMPEX Data Workshop: GPM View

    NASA Technical Reports Server (NTRS)

    Petersen, W.

    2017-01-01

    OLYMPEX Primary Objectives: Datasets to enable: (1) Direct validation over complex terrain at multiple scales, liquid and frozen precip types, (a) Do we capture terrain and synoptic regime transitions, orographic enhancements/structure, full range of precipitation intensity (e.g., very light to heavy) and types, spatial variability? (b) How well can we estimate space/time-accumulated precipitation over terrain (liquid + frozen)? (2) Physical validation of algorithms in mid-latitude cold season frontal systems over ocean and complex terrain, (a) What are the column properties of frozen, melting, liquid hydrometeors-their relative contributions to estimated surface precipitation, transition under the influence of terrain gradients, and systematic variability as a function of synoptic regime? (3) Integrated hydrologic validation in complex terrain, (a) Can satellite estimates be combined with modeling over complex topography to drive improved products (assimilation, downscaling) [Level IV products] (b) What are capabilities and limitations for use of satellite-based precipitation estimates in stream/river flow forecasting?

  15. Estimation of absolute solvent and solvation shell entropies via permutation reduction

    NASA Astrophysics Data System (ADS)

    Reinhard, Friedemann; Grubmüller, Helmut

    2007-01-01

    Despite its prominent contribution to the free energy of solvated macromolecules such as proteins or DNA, and although principally contained within molecular dynamics simulations, the entropy of the solvation shell is inaccessible to straightforward application of established entropy estimation methods. The complication is twofold. First, the configurational space density of such systems is too complex for a sufficiently accurate fit. Second, and in contrast to the internal macromolecular dynamics, the configurational space volume explored by the diffusive motion of the solvent molecules is too large to be exhaustively sampled by current simulation techniques. Here, we develop a method to overcome the second problem and to significantly alleviate the first one. We propose to exploit the permutation symmetry of the solvent by transforming the trajectory in a way that renders established estimation methods applicable, such as the quasiharmonic approximation or principal component analysis. Our permutation-reduced approach involves a combinatorial problem, which is solved through its equivalence with the linear assignment problem, for which O(N3) methods exist. From test simulations of dense Lennard-Jones gases, enhanced convergence and improved entropy estimates are obtained. Moreover, our approach renders diffusive systems accessible to improved fit functions.

  16. Age synthesis and estimation via faces: a survey.

    PubMed

    Fu, Yun; Guo, Guodong; Huang, Thomas S

    2010-11-01

    Human age, as an important personal trait, can be directly inferred by distinct patterns emerging from the facial appearance. Derived from rapid advances in computer graphics and machine vision, computer-based age synthesis and estimation via faces have become particularly prevalent topics recently because of their explosively emerging real-world applications, such as forensic art, electronic customer relationship management, security control and surveillance monitoring, biometrics, entertainment, and cosmetology. Age synthesis is defined to rerender a face image aesthetically with natural aging and rejuvenating effects on the individual face. Age estimation is defined to label a face image automatically with the exact age (year) or the age group (year range) of the individual face. Because of their particularity and complexity, both problems are attractive yet challenging to computer-based application system designers. Large efforts from both academia and industry have been devoted in the last a few decades. In this paper, we survey the complete state-of-the-art techniques in the face image-based age synthesis and estimation topics. Existing models, popular algorithms, system performances, technical difficulties, popular face aging databases, evaluation protocols, and promising future directions are also provided with systematic discussions.

  17. Spaceborne Applications of P Band Imaging Radars for Measuring Forest Biomass

    NASA Technical Reports Server (NTRS)

    Rignot, Eric J.; Zimmermann, Reiner; vanZyl, Jakob J.

    1995-01-01

    In three sites of boreal and temperate forests, P band HH, HV, and VV polarization data combined estimate total aboveground dry woody biomass within 12 to 27% of the values derived from allometric equations, depending on forest complexity. Biomass estimates derived from HV-polarization data only are 2 to 14% less accurate. When the radar operates at circular polarization, the errors exceed 100% over flooded forests, wet or damaged trees and sparse open tall forests because double-bounce reflections of the radar signals yield radar signatures similar to that of tall and massive forests. Circular polarizations, which minimize the effect of Faraday rotation in spaceborne applications, are therefore of limited use for measuring forest biomass. In the tropical rain forest of Manu, in Peru, where forest biomass ranges from 4 kg/sq m in young forest succession up to 50 kg/sq m in old, undisturbed floodplain stands, the P band horizontal and vertical polarization data combined separate biomass classes in good agreement with forest inventory estimates. The worldwide need for large scale, updated, biomass estimates, achieved with a uniformly applied method, justifies a more in-depth exploration of multi-polarization long wavelength imaging radar applications for tropical forests inventories.

  18. In vivo estimation of normal amygdala volume from structural MRI scans with anatomical-based segmentation.

    PubMed

    Siozopoulos, Achilleas; Thomaidis, Vasilios; Prassopoulos, Panos; Fiska, Aliki

    2018-02-01

    Literature includes a number of studies using structural MRI (sMRI) to determine the volume of the amygdala, which is modified in various pathologic conditions. The reported values vary widely mainly because of different anatomical approaches to the complex. This study aims at estimating of the normal amygdala volume from sMRI scans using a recent anatomical definition described in a study based on post-mortem material. The amygdala volume has been calculated in 106 healthy subjects, using sMRI and anatomical-based segmentation. The resulting volumes have been analyzed for differences related to hemisphere, sex, and age. The mean amygdalar volume was estimated at 1.42 cm 3 . The mean right amygdala volume has been found larger than the left, but the difference for the raw values was within the limits of the method error. No intersexual differences or age-related alterations have been observed. The study provides a method for determining the boundaries of the amygdala in sMRI scans based on recent anatomical considerations and an estimation of the mean normal amygdala volume from a quite large number of scans for future use in comparative studies.

  19. Repeat participation in annual cross-sectional surveys of drug users and its implications for analysis.

    PubMed

    Agius, P A; Aitken, C K; Breen, C; Dietze, P M

    2018-06-04

    We sought to establish the extent of repeat participation in a large annual cross-sectional survey of people who inject drugs and assess its implications for analysis. We used "porn star names" (the name of each participant's first pet followed by the name of the first street in which they lived) to identify repeat participation in three Australian Illicit Drug Reporting System surveys. Over 2013-2015, 2468 porn star names (96.2%) appeared only once, 88 (3.4%) twice, and nine (0.4%) in all 3 years. We measured design effects, based on the between-cluster variability for selected estimates, of 1.01-1.07 for seven key variables. These values indicate that the complex sample is (e.g.) 7% less efficient in estimating prevalence of heroin use (ever) than a simple random sample, and 1% less efficient in estimating number of heroin overdoses (ever). Porn star names are a useful means of tracking research participants longitudinally while maintaining their anonymity. Repeat participation in the Australian Illicit Drug Reporting System is low (less than 5% per annum), meaning point-prevalence and effect estimation without correction for the lack of independence in observations is unlikely to seriously affect population inference.

  20. Improving Lidar-based Aboveground Biomass Estimation with Site Productivity for Central Hardwood Forests, USA

    NASA Astrophysics Data System (ADS)

    Shao, G.; Gallion, J.; Fei, S.

    2016-12-01

    Sound forest aboveground biomass estimation is required to monitor diverse forest ecosystems and their impacts on the changing climate. Lidar-based regression models provided promised biomass estimations in most forest ecosystems. However, considerable uncertainties of biomass estimations have been reported in the temperate hardwood and hardwood-dominated mixed forests. Varied site productivities in temperate hardwood forests largely diversified height and diameter growth rates, which significantly reduced the correlation between tree height and diameter at breast height (DBH) in mature and complex forests. It is, therefore, difficult to utilize height-based lidar metrics to predict DBH-based field-measured biomass through a simple regression model regardless the variation of site productivity. In this study, we established a multi-dimension nonlinear regression model incorporating lidar metrics and site productivity classes derived from soil features. In the regression model, lidar metrics provided horizontal and vertical structural information and productivity classes differentiated good and poor forest sites. The selection and combination of lidar metrics were discussed. Multiple regression models were employed and compared. Uncertainty analysis was applied to the best fit model. The effects of site productivity on the lidar-based biomass model were addressed.

  1. LSST Astroinformatics And Astrostatistics: Data-oriented Astronomical Research

    NASA Astrophysics Data System (ADS)

    Borne, Kirk D.; Stassun, K.; Brunner, R. J.; Djorgovski, S. G.; Graham, M.; Hakkila, J.; Mahabal, A.; Paegert, M.; Pesenson, M.; Ptak, A.; Scargle, J.; Informatics, LSST; Statistics Team

    2011-01-01

    The LSST Informatics and Statistics Science Collaboration (ISSC) focuses on research and scientific discovery challenges posed by the very large and complex data collection that LSST will generate. Application areas include astroinformatics, machine learning, data mining, astrostatistics, visualization, scientific data semantics, time series analysis, and advanced signal processing. Research problems to be addressed with these methodologies include transient event characterization and classification, rare class discovery, correlation mining, outlier/anomaly/surprise detection, improved estimators (e.g., for photometric redshift or early onset supernova classification), exploration of highly dimensional (multivariate) data catalogs, and more. We present sample science results from these data-oriented approaches to large-data astronomical research. We present results from LSST ISSC team members, including the EB (Eclipsing Binary) Factory, the environmental variations in the fundamental plane of elliptical galaxies, and outlier detection in multivariate catalogs.

  2. An abundance of rare functional variants in 202 drug target genes sequenced in 14,002 people

    PubMed Central

    Nelson, Matthew R.; Wegmann, Daniel; Ehm, Margaret G.; Kessner, Darren; St. Jean, Pamela; Verzilli, Claudio; Shen, Judong; Tang, Zhengzheng; Bacanu, Silviu-Alin; Fraser, Dana; Warren, Liling; Aponte, Jennifer; Zawistowski, Matthew; Liu, Xiao; Zhang, Hao; Zhang, Yong; Li, Jun; Li, Yun; Li, Li; Woollard, Peter; Topp, Simon; Hall, Matthew D.; Nangle, Keith; Wang, Jun; Abecasis, Gonçalo; Cardon, Lon R.; Zöllner, Sebastian; Whittaker, John C.; Chissoe, Stephanie L.; Novembre, John; Mooser, Vincent

    2015-01-01

    Rare genetic variants contribute to complex disease risk; however, the abundance of rare variants in human populations remains unknown. We explored this spectrum of variation by sequencing 202 genes encoding drug targets in 14,002 individuals. We find rare variants are abundant (one every 17 bases) and geographically localized, such that even with large sample sizes, rare variant catalogs will be largely incomplete. We used the observed patterns of variation to estimate population growth parameters, the proportion of variants in a given frequency class that are putatively deleterious, and mutation rates for each gene. Overall we conclude that, due to rapid population growth and weak purifying selection, human populations harbor an abundance of rare variants, many of which are deleterious and have relevance to understanding disease risk. PMID:22604722

  3. Estimating the effect of multiple environmental stressors on coral bleaching and mortality.

    PubMed

    Welle, Paul D; Small, Mitchell J; Doney, Scott C; Azevedo, Inês L

    2017-01-01

    Coral cover has been declining in recent decades due to increased temperatures and environmental stressors. However, the extent to which different stressors contribute both individually and in concert to bleaching and mortality is still very uncertain. We develop and use a novel regression approach, using non-linear parametric models that control for unobserved time invariant effects to estimate the effects on coral bleaching and mortality due to temperature, solar radiation, depth, hurricanes and anthropogenic stressors using historical data from a large bleaching event in 2005 across the Caribbean. Two separate models are created, one to predict coral bleaching, and the other to predict near-term mortality. A large ensemble of supporting data is assembled to control for omitted variable bias and improve fit, and a significant improvement in fit is observed from univariate linear regression based on temperature alone. The results suggest that climate stressors (temperature and radiation) far outweighed direct anthropogenic stressors (using distance from shore and nearby human population density as a proxy for such stressors) in driving coral health outcomes during the 2005 event. Indeed, temperature was found to play a role ~4 times greater in both the bleaching and mortality response than population density across their observed ranges. The empirical models tested in this study have large advantages over ordinary-least squares-they offer unbiased estimates for censored data, correct for spatial correlation, and are capable of handling more complex relationships between dependent and independent variables. The models offer a framework for preparing for future warming events and climate change; guiding monitoring and attribution of other bleaching and mortality events regionally and around the globe; and informing adaptive management and conservation efforts.

  4. Estimating the effect of multiple environmental stressors on coral bleaching and mortality

    PubMed Central

    Welle, Paul D.; Small, Mitchell J.; Doney, Scott C.; Azevedo, Inês L.

    2017-01-01

    Coral cover has been declining in recent decades due to increased temperatures and environmental stressors. However, the extent to which different stressors contribute both individually and in concert to bleaching and mortality is still very uncertain. We develop and use a novel regression approach, using non-linear parametric models that control for unobserved time invariant effects to estimate the effects on coral bleaching and mortality due to temperature, solar radiation, depth, hurricanes and anthropogenic stressors using historical data from a large bleaching event in 2005 across the Caribbean. Two separate models are created, one to predict coral bleaching, and the other to predict near-term mortality. A large ensemble of supporting data is assembled to control for omitted variable bias and improve fit, and a significant improvement in fit is observed from univariate linear regression based on temperature alone. The results suggest that climate stressors (temperature and radiation) far outweighed direct anthropogenic stressors (using distance from shore and nearby human population density as a proxy for such stressors) in driving coral health outcomes during the 2005 event. Indeed, temperature was found to play a role ~4 times greater in both the bleaching and mortality response than population density across their observed ranges. The empirical models tested in this study have large advantages over ordinary-least squares–they offer unbiased estimates for censored data, correct for spatial correlation, and are capable of handling more complex relationships between dependent and independent variables. The models offer a framework for preparing for future warming events and climate change; guiding monitoring and attribution of other bleaching and mortality events regionally and around the globe; and informing adaptive management and conservation efforts. PMID:28472031

  5. Estimating fish exploitation and aquatic habitat loss across diffuse inland recreational fisheries.

    PubMed

    de Kerckhove, Derrick Tupper; Minns, Charles Kenneth; Chu, Cindy

    2015-01-01

    The current state of many freshwater fish stocks worldwide is largely unknown but suspected to be vulnerable to exploitation from recreational fisheries and habitat degradation. Both these factors, combined with complex ecological dynamics and the diffuse nature of inland fisheries could lead to an invisible collapse: the drastic decline in fish stocks without great public or management awareness. In this study we provide a method to address the pervasive knowledge gaps in regional rates of exploitation and habitat degradation, and demonstrate its use in one of North America's largest and most diffuse recreational freshwater fisheries (Ontario, Canada). We estimated that (1) fish stocks were highly exploited and in apparent danger of collapse in management zones close to large population centres, and (2) fish habitat was under a low but constant threat of degradation at rates comparable to deforestation in Ontario and throughout Canada. These findings confirm some commonly held, but difficult to quantify, beliefs in inland fisheries management but also provide some further insights including (1) large anthropogenic projects greater than one hectare could contribute much more to fish habitat loss on an area basis than the cumulative effect of smaller projects within one year, (2) hooking mortality from catch-and-release fisheries is likely a greater source of mortality than the harvest itself, and (3) in most northern management zones over 50% of the fisheries resources are not yet accessible to anglers. While this model primarily provides a framework to prioritize management decisions and further targeted stock assessments, we note that our regional estimates of fisheries productivity and exploitation were similar to broadscale monitoring efforts by the Province of Ontario. We discuss the policy implications from our results and extending the model to other jurisdictions and countries.

  6. Estimating Fish Exploitation and Aquatic Habitat Loss across Diffuse Inland Recreational Fisheries

    PubMed Central

    de Kerckhove, Derrick Tupper; Minns, Charles Kenneth; Chu, Cindy

    2015-01-01

    The current state of many freshwater fish stocks worldwide is largely unknown but suspected to be vulnerable to exploitation from recreational fisheries and habitat degradation. Both these factors, combined with complex ecological dynamics and the diffuse nature of inland fisheries could lead to an invisible collapse: the drastic decline in fish stocks without great public or management awareness. In this study we provide a method to address the pervasive knowledge gaps in regional rates of exploitation and habitat degradation, and demonstrate its use in one of North America’s largest and most diffuse recreational freshwater fisheries (Ontario, Canada). We estimated that 1) fish stocks were highly exploited and in apparent danger of collapse in management zones close to large population centres, and 2) fish habitat was under a low but constant threat of degradation at rates comparable to deforestation in Ontario and throughout Canada. These findings confirm some commonly held, but difficult to quantify, beliefs in inland fisheries management but also provide some further insights including 1) large anthropogenic projects greater than one hectare could contribute much more to fish habitat loss on an area basis than the cumulative effect of smaller projects within one year, 2) hooking mortality from catch-and-release fisheries is likely a greater source of mortality than the harvest itself, and 3) in most northern management zones over 50% of the fisheries resources are not yet accessible to anglers. While this model primarily provides a framework to prioritize management decisions and further targeted stock assessments, we note that our regional estimates of fisheries productivity and exploitation were similar to broadscale monitoring efforts by the Province of Ontario. We discuss the policy implications from our results and extending the model to other jurisdictions and countries. PMID:25875790

  7. Use of an exchange method to estimate the association and dissociation rate constants of cadmium complexes formed with low-molecular-weight organic acids commonly exuded by plant roots.

    PubMed

    Schneider, André; Nguyen, Christophe

    2011-01-01

    Organic acids released from plant roots can form complexes with cadmium (Cd) in the soil solution and influence metal bioavailability not only due to the nature and concentration of the complexes but also due to their lability. The lability of a complex influences its ability to buffer changes in the concentration of free ions (Cd); it depends on the association (, m mol s) and dissociation (, s) rate constants. A resin exchange method was used to estimate and (m mol s), which is the conditional estimate of depending on the calcium (Ca) concentration in solution. The constants were estimated for oxalate, citrate, and malate, three low-molecular-weight organic acids commonly exuded by plant roots and expected to strongly influence Cd uptake by plants. For all three organic acids, the and estimates were around 2.5 10 m mol s and 1.3 × 10 s, respectively. Based on the literature, these values indicate that the Cd- low-molecular-weight organic acids complexes formed between Cd and low-molecular-weight organic acids may be less labile than complexes formed with soil soluble organic matter but more labile than those formed with aminopolycarboxylic chelates. Copyright © by the American Society of Agronomy, Crop Science Society of America, and Soil Science Society of America, Inc.

  8. Inference of the genetic architecture underlying BMI and height with the use of 20,240 sibling pairs.

    PubMed

    Hemani, Gibran; Yang, Jian; Vinkhuyzen, Anna; Powell, Joseph E; Willemsen, Gonneke; Hottenga, Jouke-Jan; Abdellaoui, Abdel; Mangino, Massimo; Valdes, Ana M; Medland, Sarah E; Madden, Pamela A; Heath, Andrew C; Henders, Anjali K; Nyholt, Dale R; de Geus, Eco J C; Magnusson, Patrik K E; Ingelsson, Erik; Montgomery, Grant W; Spector, Timothy D; Boomsma, Dorret I; Pedersen, Nancy L; Martin, Nicholas G; Visscher, Peter M

    2013-11-07

    Evidence that complex traits are highly polygenic has been presented by population-based genome-wide association studies (GWASs) through the identification of many significant variants, as well as by family-based de novo sequencing studies indicating that several traits have a large mutational target size. Here, using a third study design, we show results consistent with extreme polygenicity for body mass index (BMI) and height. On a sample of 20,240 siblings (from 9,570 nuclear families), we used a within-family method to obtain narrow-sense heritability estimates of 0.42 (SE = 0.17, p = 0.01) and 0.69 (SE = 0.14, p = 6 × 10(-)(7)) for BMI and height, respectively, after adjusting for covariates. The genomic inflation factors from locus-specific linkage analysis were 1.69 (SE = 0.21, p = 0.04) for BMI and 2.18 (SE = 0.21, p = 2 × 10(-10)) for height. This inflation is free of confounding and congruent with polygenicity, consistent with observations of ever-increasing genomic-inflation factors from GWASs with large sample sizes, implying that those signals are due to true genetic signals across the genome rather than population stratification. We also demonstrate that the distribution of the observed test statistics is consistent with both rare and common variants underlying a polygenic architecture and that previous reports of linkage signals in complex traits are probably a consequence of polygenic architecture rather than the segregation of variants with large effects. The convergent empirical evidence from GWASs, de novo studies, and within-family segregation implies that family-based sequencing studies for complex traits require very large sample sizes because the effects of causal variants are small on average. Copyright © 2013 The American Society of Human Genetics. Published by Elsevier Inc. All rights reserved.

  9. Practicality of magnetic compression for plasma density control

    DOE PAGES

    Gueroult, Renaud; Fisch, Nathaniel J.

    2016-03-16

    Here, plasma densification through magnetic compression has been suggested for time-resolved control of the wave properties in plasma-based accelerators [P. F. Schmit and N. J. Fisch, Phys. Rev. Lett. 109, 255003 (2012)]. Using particle in cell simulations with real mass ratio, the practicality of large magnetic compression on timescales shorter than the ion gyro-period is investigated. For compression times shorter than the transit time of a compressional Alfven wave across the plasma slab, results show the formation of two counter-propagating shock waves, leading to a highly non-uniform plasma density profile. Furthermore, the plasma slab displays large hydromagnetic like oscillations aftermore » the driving field has reached steady state. Peak compression is obtained when the two shocks collide in the mid-plane. At this instant, very large plasma heating is observed, and the plasmaβ is estimated to be about 1. Although these results point out a densification mechanism quite different and more complex than initially envisioned, these features still might be advantageous in particle accelerators.« less

  10. Comparison of actual and seismologically inferred stress drops in dynamic models of microseismicity

    NASA Astrophysics Data System (ADS)

    Lin, Y. Y.; Lapusta, N.

    2017-12-01

    Estimating source parameters for small earthquakes is commonly based on either Brune or Madariaga source models. These models assume circular rupture that starts from the center of a fault and spreads axisymmetrically with a constant rupture speed. The resulting stress drops are moment-independent, with large scatter. However, more complex source behaviors are commonly discovered by finite-fault inversions for both large and small earthquakes, including directivity, heterogeneous slip, and non-circular shapes. Recent studies (Noda, Lapusta, and Kanamori, GJI, 2013; Kaneko and Shearer, GJI, 2014; JGR, 2015) have shown that slip heterogeneity and directivity can result in large discrepancies between the actual and estimated stress drops. We explore the relation between the actual and seismologically estimated stress drops for several types of numerically produced microearthquakes. For example, an asperity-type circular fault patch with increasing normal stress towards the middle of the patch, surrounded by a creeping region, is a potentially common microseismicity source. In such models, a number of events rupture the portion of the patch near its circumference, producing ring-like ruptures, before a patch-spanning event occurs. We calculate the far-field synthetic waveforms for our simulated sources and estimate their spectral properties. The distribution of corner frequencies over the focal sphere is markedly different for the ring-like sources compared to the Madariaga model. Furthermore, most waveforms for the ring-like sources are better fitted by a high-frequency fall-off rate different from the commonly assumed value of 2 (from the so-called omega-squared model), with the average value over the focal sphere being 1.5. The application of Brune- or Madariaga-type analysis to these sources results in the stress drops estimates different from the actual stress drops by a factor of up to 125 in the models we considered. We will report on our current studies of other types of seismic sources, such as repeating earthquakes and foreshock-like events, and whether the potentially realistic and common sources different from the standard Brune and Madariaga models can be identified from their focal spectral signatures and studied using a more tailored seismological analysis.

  11. Efficient Ensemble State-Parameters Estimation Techniques in Ocean Ecosystem Models: Application to the North Atlantic

    NASA Astrophysics Data System (ADS)

    El Gharamti, M.; Bethke, I.; Tjiputra, J.; Bertino, L.

    2016-02-01

    Given the recent strong international focus on developing new data assimilation systems for biological models, we present in this comparative study the application of newly developed state-parameters estimation tools to an ocean ecosystem model. It is quite known that the available physical models are still too simple compared to the complexity of the ocean biology. Furthermore, various biological parameters remain poorly unknown and hence wrong specifications of such parameters can lead to large model errors. Standard joint state-parameters augmentation technique using the ensemble Kalman filter (Stochastic EnKF) has been extensively tested in many geophysical applications. Some of these assimilation studies reported that jointly updating the state and the parameters might introduce significant inconsistency especially for strongly nonlinear models. This is usually the case for ecosystem models particularly during the period of the spring bloom. A better handling of the estimation problem is often carried out by separating the update of the state and the parameters using the so-called Dual EnKF. The dual filter is computationally more expensive than the Joint EnKF but is expected to perform more accurately. Using a similar separation strategy, we propose a new EnKF estimation algorithm in which we apply a one-step-ahead smoothing to the state. The new state-parameters estimation scheme is derived in a consistent Bayesian filtering framework and results in separate update steps for the state and the parameters. Unlike the classical filtering path, the new scheme starts with an update step and later a model propagation step is performed. We test the performance of the new smoothing-based schemes against the standard EnKF in a one-dimensional configuration of the Norwegian Earth System Model (NorESM) in the North Atlantic. We use nutrients profile (up to 2000 m deep) data and surface partial CO2 measurements from Mike weather station (66o N, 2o E) to estimate different biological parameters of phytoplanktons and zooplanktons. We analyze the performance of the filters in terms of complexity and accuracy of the state and parameters estimates.

  12. Testing earthquake source inversion methodologies

    USGS Publications Warehouse

    Page, M.; Mai, P.M.; Schorlemmer, D.

    2011-01-01

    Source Inversion Validation Workshop; Palm Springs, California, 11-12 September 2010; Nowadays earthquake source inversions are routinely performed after large earthquakes and represent a key connection between recorded seismic and geodetic data and the complex rupture process at depth. The resulting earthquake source models quantify the spatiotemporal evolution of ruptures. They are also used to provide a rapid assessment of the severity of an earthquake and to estimate losses. However, because of uncertainties in the data, assumed fault geometry and velocity structure, and chosen rupture parameterization, it is not clear which features of these source models are robust. Improved understanding of the uncertainty and reliability of earthquake source inversions will allow the scientific community to use the robust features of kinematic inversions to more thoroughly investigate the complexity of the rupture process and to better constrain other earthquakerelated computations, such as ground motion simulations and static stress change calculations.

  13. Stochastic competitive learning in complex networks.

    PubMed

    Silva, Thiago Christiano; Zhao, Liang

    2012-03-01

    Competitive learning is an important machine learning approach which is widely employed in artificial neural networks. In this paper, we present a rigorous definition of a new type of competitive learning scheme realized on large-scale networks. The model consists of several particles walking within the network and competing with each other to occupy as many nodes as possible, while attempting to reject intruder particles. The particle's walking rule is composed of a stochastic combination of random and preferential movements. The model has been applied to solve community detection and data clustering problems. Computer simulations reveal that the proposed technique presents high precision of community and cluster detections, as well as low computational complexity. Moreover, we have developed an efficient method for estimating the most likely number of clusters by using an evaluator index that monitors the information generated by the competition process itself. We hope this paper will provide an alternative way to the study of competitive learning..

  14. Polypeptide Translocation Through the Mitochondrial TOM Channel: Temperature-Dependent Rates at the Single-Molecule Level.

    PubMed

    Mahendran, Kozhinjampara R; Lamichhane, Usha; Romero-Ruiz, Mercedes; Nussberger, Stephan; Winterhalter, Mathias

    2013-01-03

    The TOM protein complex facilitates the transfer of nearly all mitochondrial preproteins across outer mitochondrial membranes. Here we characterized the effect of temperature on facilitated translocation of a mitochondrial presequence peptide pF1β. Ion current fluctuations analysis through single TOM channels revealed thermodynamic and kinetic parameters of substrate binding and allowed determining the energy profile of peptide translocation. The activation energy for the on-rate and off-rate of the presequence peptide into the TOM complex was symmetric with respect to the electric field and estimated to be about 15 and 22 kT per peptide. These values are above that expected for free diffusion of ions in water (6 kT) and reflect the stronger interaction in the channel. Both values are in the range for typical enzyme kinetics and suggest one process without involving large conformational changes within the channel protein.

  15. Modeling, Analysis, and Interpretation of Photoelectron Energy Spectra at Enceladus Observed by Cassini

    NASA Astrophysics Data System (ADS)

    Taylor, S. A.; Coates, A. J.; Jones, G. H.; Wellbrock, A.; Fazakerley, A. N.; Desai, R. T.; Caro-Carretero, R.; Michiko, M. W.; Schippers, P.; Waite, J. H.

    2018-01-01

    The Electron Spectrometer (ELS) of the Cassini Plasma Spectrometer has observed photoelectrons produced in the plume of Enceladus. These photoelectrons are observed during Enceladus encounters in the energetic particle shadow where the spacecraft is largely shielded from penetrating radiation by the moon. We present a complex electron spectrum at Enceladus including evidence of two previously unidentified electron populations at 6-10 eV and 10-16 eV. We estimate that the proportion of "hot" (>15 eV) to "cold" (<15 eV) electrons during the Enceladus flybys is ≈ 0.1-0.5%. We have constructed a model of photoelectron production in the plume and compared it with ELS Enceladus flyby data by scaling and energy shifting according to spacecraft potential. We suggest that the complex structure of the electron spectrum observed can be explained entirely by photoelectron production in the plume ionosphere.

  16. Using mixture tuned match filtering to measure changes in subpixel vegetation area in Las Vegas, Nevada

    NASA Astrophysics Data System (ADS)

    Brelsford, Christa; Shepherd, Doug

    2013-09-01

    In desert cities, securing sufficient water supply to meet the needs of both existing population and future growth is a complex problem with few easy solutions. Grass lawns are a major driver of water consumption and accurate measurements of vegetation area are necessary to understand drivers of changes in household water consumption. Measuring vegetation change in a heterogeneous urban environment requires sub-pixel estimation of vegetation area. Mixture Tuned Match Filtering has been successfully applied to target detection for materials that only cover small portions of a satellite image pixel. There have been few successful applications of MTMF to fractional area estimation, despite theory that suggests feasibility. We use a ground truth dataset over ten times larger than that available for any previous MTMF application to estimate the bias between ground truth data and matched filter results. We find that the MTMF algorithm underestimates the fractional area of vegetation by 5-10%, and calculate that averaging over 20 to 30 pixels is necessary to correct this bias. We conclude that with a large ground truth dataset, using MTMF for fractional area estimation is possible when results can be estimated at a lower spatial resolution than the base image. When this method is applied to estimating vegetation area in Las Vegas, NV spatial and temporal trends are consistent with expectations from known population growth and policy goals.

  17. Evaporation estimates from the Dead Sea and their implications on its water balance

    NASA Astrophysics Data System (ADS)

    Oroud, Ibrahim M.

    2011-12-01

    The Dead Sea (DS) is a terminal hypersaline water body situated in the deepest part of the Jordan Valley. There is a growing interest in linking the DS to the open seas due to severe water shortages in the area and the serious geological and environmental hazards to its vicinity caused by the rapid level drop of the DS. A key issue in linking the DS with the open seas would be an accurate determination of evaporation rates. There exist large uncertainties of evaporation estimates from the DS due to the complex feedback mechanisms between meteorological forcings and thermophysical properties of hypersaline solutions. Numerous methods have been used to estimate current and historical (pre-1960) evaporation rates, with estimates differing by ˜100%. Evaporation from the DS is usually deduced indirectly using energy, water balance, or pan methods with uncertainty in many parameters. Accumulated errors resulting from these uncertainties are usually pooled into the estimates of evaporation rates. In this paper, a physically based method with minimum empirical parameters is used to evaluate historical and current evaporation estimates from the DS. The more likely figures for historical and current evaporation rates from the DS were 1,500-1,600 and 1,200-1,250 mm per annum, respectively. Results obtained are congruent with field observations and with more elaborate procedures.

  18. Volcanism and erosion during the past 930 k.y. at the Tatara-San Pedro complex, Chilean Andes

    USGS Publications Warehouse

    Singer, B.S.; Thompson, R.A.; Dungan, M.A.; Feeley, T.C.; Nelson, S.T.; Pickens, J.C.; Brown, L.L.; Wulff, A.W.; Davidson, J.P.; Metzger, J.

    1997-01-01

    Geologic mapping, together with 73 new K-Ar and 40Ar/39Ar age determinations of 45 samples from 17 different volcanic units, plus paleomagnetic orientations, geochemical compositions, and terrestrial photogrammetry are used to define the chronostratigraphy of the Tatara-San Pedro complex, an eruptive center at 36??S on the volcanic front of the Andean southern volcanic zone. The Tatara-San Pedro complex preserves ???55 km3 of lavas that erupted from at least three central vent regions. Remnant, unconformity-bound sequences of lavas are separated by lacunae that include significant periods of erosion. Quaternary volcanism commenced ca. 930 ka with eruption of voluminous dacitic magma, followed 100 k.y. later by the only major rhyolitic eruption. From 780 ka onward, more than 80% of the preserved volume is basaltic andesite (52%-57% SiO2), but petrographically and geochemically diverse dacitic magmas (63%-69% SiO2) erupted sporadically throughout this younger, dominantly mafic phase of activity. A few basaltic lavas (49%-52% SiO2) are present, mainly in portions of the complex older than 230 ka. The number of vents, the petrologic and geochemical diversity, and the temporal distribution of mafic and silicic lavas are consistent with emplacement of many separate batches of made magma into the shallow crust beneath the Tatara-San Pedro complex over the past million years. Nearly two-thirds of the preserved volume of the Tatara-San Pedro complex comprises the two youngest volcanoes, which were active between ca. 188-83 ka and 90-19 ka. Repeated advances of mountain glaciers punctuated growth of the complex with major erosional episodes that removed much of the pre-200 ka volcanic record, particularly on the south flank of the complex. Dating the inception of a glaciation on the basis of preserved material is difficult, but the age of the oldest lava above a lacuna may be used to estimate the timing of deglaciation. On this basis, the argon ages of basal lavas of multiple sequences indicate minimum upper limits of lacunae at ca. 830, 790, 610, 400, 330, 230, 110, and 17 ka. These are broadly consistent with global ice-volume peaks predicted by the oxygen isotope-based astronomical time scale and with age brackets on North American glacial advances. Estimated growth rates for the two young volcanoes are 0.2 to 0.3 km3/k.y.; these are three to five times greater than a growth rate estimated from all preserved lavas in the complex (0.06 km3/k.y.). Removal of up to 50%-95% of the material erupted between 930 and 200 ka by repeated glacial advances largely explains this discrepancy, and it raises the possibility that episodic erosion of midlatitude frontal arc complexes may be extensive and common.

  19. Turbulent and Laminar Flow in Karst Conduits Under Unsteady Flow Conditions: Interpretation of Pumping Tests by Discrete Conduit-Continuum Modeling

    NASA Astrophysics Data System (ADS)

    Giese, M.; Reimann, T.; Bailly-Comte, V.; Maréchal, J.-C.; Sauter, M.; Geyer, T.

    2018-03-01

    Due to the duality in terms of (1) the groundwater flow field and (2) the discharge conditions, flow patterns of karst aquifer systems are complex. Estimated aquifer parameters may differ by several orders of magnitude from local (borehole) to regional (catchment) scale because of the large contrast in hydraulic parameters between matrix and conduit, their heterogeneity and anisotropy. One approach to deal with the scale effect problem in the estimation of hydraulic parameters of karst aquifers is the application of large-scale experiments such as long-term high-abstraction conduit pumping tests, stimulating measurable groundwater drawdown in both, the karst conduit system as well as the fractured matrix. The numerical discrete conduit-continuum modeling approach MODFLOW-2005 Conduit Flow Process Mode 1 (CFPM1) is employed to simulate laminar and nonlaminar conduit flow, induced by large-scale experiments, in combination with Darcian matrix flow. Effects of large-scale experiments were simulated for idealized settings. Subsequently, diagnostic plots and analyses of different fluxes are applied to interpret differences in the simulated conduit drawdown and general flow patterns. The main focus is set on the question to which extent different conduit flow regimes will affect the drawdown in conduit and matrix depending on the hydraulic properties of the conduit system, i.e., conduit diameter and relative roughness. In this context, CFPM1 is applied to investigate the importance of considering turbulent conditions for the simulation of karst conduit flow. This work quantifies the relative error that results from assuming laminar conduit flow for the interpretation of a synthetic large-scale pumping test in karst.

  20. Estimating micro area behavioural risk factor prevalence from large population-based surveys: a full Bayesian approach.

    PubMed

    Seliske, L; Norwood, T A; McLaughlin, J R; Wang, S; Palleschi, C; Holowaty, E

    2016-06-07

    An important public health goal is to decrease the prevalence of key behavioural risk factors, such as tobacco use and obesity. Survey information is often available at the regional level, but heterogeneity within large geographic regions cannot be assessed. Advanced spatial analysis techniques are demonstrated to produce sensible micro area estimates of behavioural risk factors that enable identification of areas with high prevalence. A spatial Bayesian hierarchical model was used to estimate the micro area prevalence of current smoking and excess bodyweight for the Erie-St. Clair region in southwestern Ontario. Estimates were mapped for male and female respondents of five cycles of the Canadian Community Health Survey (CCHS). The micro areas were 2006 Census Dissemination Areas, with an average population of 400-700 people. Two individual-level models were specified: one controlled for survey cycle and age group (model 1), and one controlled for survey cycle, age group and micro area median household income (model 2). Post-stratification was used to derive micro area behavioural risk factor estimates weighted to the population structure. SaTScan analyses were conducted on the granular, postal-code level CCHS data to corroborate findings of elevated prevalence. Current smoking was elevated in two urban areas for both sexes (Sarnia and Windsor), and an additional small community (Chatham) for males only. Areas of excess bodyweight were prevalent in an urban core (Windsor) among males, but not females. Precision of the posterior post-stratified current smoking estimates was improved in model 2, as indicated by narrower credible intervals and a lower coefficient of variation. For excess bodyweight, both models had similar precision. Aggregation of the micro area estimates to CCHS design-based estimates validated the findings. This is among the first studies to apply a full Bayesian model to complex sample survey data to identify micro areas with variation in risk factor prevalence, accounting for spatial correlation and other covariates. Application of micro area analysis techniques helps define areas for public health planning, and may be informative to surveillance and research modeling of relevant chronic disease outcomes.

  1. Getting Astrophysical Information from LISA Data

    NASA Technical Reports Server (NTRS)

    Stebbins, R. T.; Bender, P. L.; Folkner, W. M.

    1997-01-01

    Gravitational wave signals from a large number of astrophysical sources will be present in the LISA data. Information about as many sources as possible must be estimated from time series of strain measurements. Several types of signals are expected to be present: simple periodic signals from relatively stable binary systems, chirped signals from coalescing binary systems, complex waveforms from highly relativistic binary systems, stochastic backgrounds from galactic and extragalactic binary systems and possibly stochastic backgrounds from the early Universe. The orbital motion of the LISA antenna will modulate the phase and amplitude of all these signals, except the isotropic backgrounds and thereby give information on the directions of sources. Here we describe a candidate process for disentangling the gravitational wave signals and estimating the relevant astrophysical parameters from one year of LISA data. Nearly all of the sources will be identified by searching with templates based on source parameters and directions.

  2. Impact Lithologies and Post-Impact Hydrothermal Alteration Exposed by the Chicxulub Scientific Drilling Project, Yaxcopoil, Mexico

    NASA Technical Reports Server (NTRS)

    Kring, David A.; Zurcher, Lukas; Horz, Friedrich

    2003-01-01

    The Chicxulub Scientific Drilling Project recovered a continuous core from the Yaxcopoil-1 (YAX-1) borehole, which is approx.60-65 km from the center of the Chicxulub structure, approx.15 km beyond the limit of the estimated approx.50 km radius transient crater (excavation cavity), but within the rim of the estimated approx.90 km radius final crater. Approximately approx.100 m of melt-bearing impactites were recoverd from a depth of 794 to 895 m, above approx.600 m of underlying megablocks of Cretaceous target sediments, before bottoming at 1511 m. Compared to lithologies at impact craters like the Ries, the YAX-1 impactite sequence is incredibly rich in impact melts of unusual textural variety and complexity. The impactite sequence has also been altered by hydrothermal activity that may have largely been produced by the impact event.

  3. Ionization of Interstellar Hydrogen Beyond the Termination Shock

    NASA Astrophysics Data System (ADS)

    Gruntman, Mike

    2016-11-01

    Models of solar wind interaction with the surrounding interstellar medium usually disregard ionization of interstellar hydrogen atoms beyond the solar wind termination shock. If and when included, the effects of ionization in the heliospheric interface region are often obscured by complexities of the interaction. This work assesses the importance of interstellar hydrogen ionization in the heliosheath. Photoionization could be accounted for in a straightforward way. In contrast, electron impact ionization is largely unknown because of poorly understood energy transfer to electrons at the termination shock and beyond. We first estimate the effect of photoionization and then use it as a yardstick to assess the role of electron impact ionization. The physical estimates show that ionization of interstellar hydrogen may lead to significant mass loading in the inner heliosheath which would slow down plasma flowing toward the heliotail and deplete populations of nonthermal protons, with the corresponding effect on heliospheric fluxes of energetic neutral atoms.

  4. Morphometricity as a measure of the neuroanatomical signature of a trait.

    PubMed

    Sabuncu, Mert R; Ge, Tian; Holmes, Avram J; Smoller, Jordan W; Buckner, Randy L; Fischl, Bruce

    2016-09-27

    Complex physiological and behavioral traits, including neurological and psychiatric disorders, often associate with distributed anatomical variation. This paper introduces a global metric, called morphometricity, as a measure of the anatomical signature of different traits. Morphometricity is defined as the proportion of phenotypic variation that can be explained by macroscopic brain morphology. We estimate morphometricity via a linear mixed-effects model that uses an anatomical similarity matrix computed based on measurements derived from structural brain MRI scans. We examined over 3,800 unique MRI scans from nine large-scale studies to estimate the morphometricity of a range of phenotypes, including clinical diagnoses such as Alzheimer's disease, and nonclinical traits such as measures of cognition. Our results demonstrate that morphometricity can provide novel insights about the neuroanatomical correlates of a diverse set of traits, revealing associations that might not be detectable through traditional statistical techniques.

  5. Evidence for a Low Bulk Crustal Density for Mars from Gravity and Topography.

    PubMed

    Goossens, Sander; Sabaka, Terence J; Genova, Antonio; Mazarico, Erwan; Nicholas, Joseph B; Neumann, Gregory A

    2017-08-16

    Knowledge of the average density of the crust of a planet is important in determining its interior structure. The combination of high-resolution gravity and topography data has yielded a low density for the Moon's crust, yet for other terrestrial planets the resolution of the gravity field models has hampered reasonable estimates. By using well-chosen constraints derived from topography during gravity field model determination using satellite tracking data, we show that we can robustly and independently determine the average bulk crustal density directly from the tracking data, using the admittance between topography and imperfect gravity. We find a low average bulk crustal density for Mars, 2582 ± 209 kg m -3 . This bulk crustal density is lower than that assumed until now. Densities for volcanic complexes are higher, consistent with earlier estimates, implying large lateral variations in crustal density. In addition, we find indications that the crustal density increases with depth.

  6. Life Outside the Golden Window: Statistical Angles on the Signal-to-Noise Problem

    NASA Astrophysics Data System (ADS)

    Wagman, Michael

    2018-03-01

    Lattice QCD simulations of multi-baryon correlation functions can predict the structure and reactions of nuclei without encountering the baryon chemical potential sign problem. However, they suffer from a signal-to-noise problem where Monte Carlo estimates of observables have quantum fluctuations that are exponentially larger than their average values. Recent lattice QCD results demonstrate that the complex phase of baryon correlations functions relates the baryon signal-to-noise problem to a sign problem and exhibits unexpected statistical behavior resembling a heavy-tailed random walk on the unit circle. Estimators based on differences of correlation function phases evaluated at different Euclidean times are discussed that avoid the usual signal-to-noise problem, instead facing a signal-to-noise problem as the time interval associated with the phase difference is increased, and allow hadronic observables to be determined from arbitrarily large-time correlation functions.

  7. Morphometricity as a measure of the neuroanatomical signature of a trait

    PubMed Central

    Sabuncu, Mert R.; Ge, Tian; Holmes, Avram J.; Smoller, Jordan W.; Buckner, Randy L.; Fischl, Bruce

    2016-01-01

    Complex physiological and behavioral traits, including neurological and psychiatric disorders, often associate with distributed anatomical variation. This paper introduces a global metric, called morphometricity, as a measure of the anatomical signature of different traits. Morphometricity is defined as the proportion of phenotypic variation that can be explained by macroscopic brain morphology. We estimate morphometricity via a linear mixed-effects model that uses an anatomical similarity matrix computed based on measurements derived from structural brain MRI scans. We examined over 3,800 unique MRI scans from nine large-scale studies to estimate the morphometricity of a range of phenotypes, including clinical diagnoses such as Alzheimer’s disease, and nonclinical traits such as measures of cognition. Our results demonstrate that morphometricity can provide novel insights about the neuroanatomical correlates of a diverse set of traits, revealing associations that might not be detectable through traditional statistical techniques. PMID:27613854

  8. UNCERTAINTY ON RADIATION DOSES ESTIMATED BY BIOLOGICAL AND RETROSPECTIVE PHYSICAL METHODS.

    PubMed

    Ainsbury, Elizabeth A; Samaga, Daniel; Della Monaca, Sara; Marrale, Maurizio; Bassinet, Celine; Burbidge, Christopher I; Correcher, Virgilio; Discher, Michael; Eakins, Jon; Fattibene, Paola; Güçlü, Inci; Higueras, Manuel; Lund, Eva; Maltar-Strmecki, Nadica; McKeever, Stephen; Rääf, Christopher L; Sholom, Sergey; Veronese, Ivan; Wieser, Albrecht; Woda, Clemens; Trompier, Francois

    2018-03-01

    Biological and physical retrospective dosimetry are recognised as key techniques to provide individual estimates of dose following unplanned exposures to ionising radiation. Whilst there has been a relatively large amount of recent development in the biological and physical procedures, development of statistical analysis techniques has failed to keep pace. The aim of this paper is to review the current state of the art in uncertainty analysis techniques across the 'EURADOS Working Group 10-Retrospective dosimetry' members, to give concrete examples of implementation of the techniques recommended in the international standards, and to further promote the use of Monte Carlo techniques to support characterisation of uncertainties. It is concluded that sufficient techniques are available and in use by most laboratories for acute, whole body exposures to highly penetrating radiation, but further work will be required to ensure that statistical analysis is always wholly sufficient for the more complex exposure scenarios.

  9. Laser radar cross-section estimation from high-resolution image data.

    PubMed

    Osche, G R; Seeber, K N; Lok, Y F; Young, D S

    1992-05-10

    A methodology for the estimation of ladar cross sections from high-resolution image data of geometrically complex targets is presented. Coherent CO(2) laser radar was used to generate high-resolution amplitude imagery of a UC-8 Buffalo test aircraft at a range of 1.3 km at nine different aspect angles. The average target ladar cross section was synthesized from these data and calculated to be sigma(T) = 15.4 dBsm, which is similar to the expected microwave radar cross sections. The aspect angle dependence of the cross section shows pronounced peaks at nose on and broadside, which are also in agreement with radar results. Strong variations in both the mean amplitude and the statistical distributions of amplitude with the aspect angle have also been observed. The relative mix of diffuse and specular returns causes significant deviations from a simple Lambertian or Swerling II target, especially at broadside where large normal surfaces are present.

  10. Network structure from rich but noisy data

    NASA Astrophysics Data System (ADS)

    Newman, M. E. J.

    2018-06-01

    Driven by growing interest across the sciences, a large number of empirical studies have been conducted in recent years of the structure of networks ranging from the Internet and the World Wide Web to biological networks and social networks. The data produced by these experiments are often rich and multimodal, yet at the same time they may contain substantial measurement error1-7. Accurate analysis and understanding of networked systems requires a way of estimating the true structure of networks from such rich but noisy data8-15. Here we describe a technique that allows us to make optimal estimates of network structure from complex data in arbitrary formats, including cases where there may be measurements of many different types, repeated observations, contradictory observations, annotations or metadata, or missing data. We give example applications to two different social networks, one derived from face-to-face interactions and one from self-reported friendships.

  11. Comparative Population Genomics Analysis of the Mammalian Fungal Pathogen Pneumocystis

    PubMed Central

    Ma, Liang; Wei Huang, Da; Khil, Pavel P.; Dekker, John P.; Kutty, Geetha; Bishop, Lisa; Liu, Yueqin; Deng, Xilong; Pagni, Marco; Hirsch, Vanessa; Lempicki, Richard A.

    2018-01-01

    ABSTRACT Pneumocystis species are opportunistic mammalian pathogens that cause severe pneumonia in immunocompromised individuals. These fungi are highly host specific and uncultivable in vitro. Human Pneumocystis infections present major challenges because of a limited therapeutic arsenal and the rise of drug resistance. To investigate the diversity and demographic history of natural populations of Pneumocystis infecting humans, rats, and mice, we performed whole-genome and large-scale multilocus sequencing of infected tissues collected in various geographic locations. Here, we detected reduced levels of recombination and variations in historical demography, which shape the global population structures. We report estimates of evolutionary rates, levels of genetic diversity, and population sizes. Molecular clock estimates indicate that Pneumocystis species diverged before their hosts, while the asynchronous timing of population declines suggests host shifts. Our results have uncovered complex patterns of genetic variation influenced by multiple factors that shaped the adaptation of Pneumocystis populations during their spread across mammals. PMID:29739910

  12. Endogenous Price Bubbles in a Multi-Agent System of the Housing Market

    PubMed Central

    2015-01-01

    Economic history shows a large number of boom-bust cycles, with the U.S. real estate market as one of the latest examples. Classical economic models have not been able to provide a full explanation for this type of market dynamics. Therefore, we analyze home prices in the U.S. using an alternative approach, a multi-agent complex system. Instead of the classical assumptions of agent rationality and market efficiency, agents in the model are heterogeneous, adaptive, and boundedly rational. We estimate the multi-agent system with historical house prices for the U.S. market. The model fits the data well and a deterministic version of the model can endogenously produce boom-and-bust cycles on the basis of the estimated coefficients. This implies that trading between agents themselves can create major price swings in absence of fundamental news. PMID:26107740

  13. Depth to the bottom of magnetic sources (DBMS) from aeromagnetic data of Central India using modified centroid method for fractal distribution of sources

    NASA Astrophysics Data System (ADS)

    Bansal, A. R.; Anand, S. P.; Rajaram, Mita; Rao, V. K.; Dimri, V. P.

    2013-09-01

    The depth to the bottom of the magnetic sources (DBMS) has been estimated from the aeromagnetic data of Central India. The conventional centroid method of DBMS estimation assumes random uniform uncorrelated distribution of sources and to overcome this limitation a modified centroid method based on scaling distribution has been proposed. Shallower values of the DBMS are found for the south western region. The DBMS values are found as low as 22 km in the south west Deccan trap covered regions and as deep as 43 km in the Chhattisgarh Basin. In most of the places DBMS are much shallower than the Moho depth, earlier found from the seismic study and may be representing the thermal/compositional/petrological boundaries. The large variation in the DBMS indicates the complex nature of the Indian crust.

  14. Leucosome distribution in migmatitic paragneisses and orthogneisses: A record of self-organized melt migration and entrapment in a heterogeneous partially-molten crust

    NASA Astrophysics Data System (ADS)

    Yakymchuk, C.; Brown, M.; Ivanic, T. J.; Korhonen, F. J.

    2013-09-01

    The depth to the bottom of the magnetic sources (DBMS) has been estimated from the aeromagnetic data of Central India. The conventional centroid method of DBMS estimation assumes random uniform uncorrelated distribution of sources and to overcome this limitation a modified centroid method based on scaling distribution has been proposed. Shallower values of the DBMS are found for the south western region. The DBMS values are found as low as 22 km in the south west Deccan trap covered regions and as deep as 43 km in the Chhattisgarh Basin. In most of the places DBMS are much shallower than the Moho depth, earlier found from the seismic study and may be representing the thermal/compositional/petrological boundaries. The large variation in the DBMS indicates the complex nature of the Indian crust.

  15. High cell density media for Escherichia coli are generally designed for aerobic cultivations – consequences for large-scale bioprocesses and shake flask cultures

    PubMed Central

    Soini, Jaakko; Ukkonen, Kaisa; Neubauer, Peter

    2008-01-01

    Background For the cultivation of Escherichia coli in bioreactors trace element solutions are generally designed for optimal growth under aerobic conditions. They do normally not contain selenium and nickel. Molybdenum is only contained in few of them. These elements are part of the formate hydrogen lyase (FHL) complex which is induced under anaerobic conditions. As it is generally known that oxygen limitation appears in shake flask cultures and locally in large-scale bioreactors, function of the FHL complex may influence the process behaviour. Formate has been described to accumulate in large-scale cultures and may have toxic effects on E. coli. Although the anaerobic metabolism of E. coli is well studied, reference data which estimate the impact of the FHL complex on bioprocesses of E. coli with oxygen limitation have so far not been published, but are important for a better process understanding. Results Two sets of fed-batch cultures with conditions triggering oxygen limitation and formate accumulation were performed. Permanent oxygen limitation which is typical for shake flask cultures was caused in a bioreactor by reduction of the agitation rate. Transient oxygen limitation, which has been described to eventually occur in the feed-zone of large-scale bioreactors, was mimicked in a two-compartment scale-down bioreactor consisting of a stirred tank reactor and a plug flow reactor (PFR) with continuous glucose feeding into the PFR. In both models formate accumulated up to about 20 mM in the culture medium without addition of selenium, molybdenum and nickel. By addition of these trace elements the formate accumulation decreased below the level observed in well-mixed laboratory-scale cultures. Interestingly, addition of the extra trace elements caused accumulation of large amounts of lactate and reduced biomass yield in the simulator with permanent oxygen limitation, but not in the scale-down two-compartment bioreactor. Conclusion The accumulation of formate in oxygen limited cultivations of E. coli can be fully prevented by addition of the trace elements selenium, nickel and molybdenum, necessary for the function of FHL complex. For large-scale cultivations, if glucose gradients are likely, the results from the two-compartment scale-down bioreactor indicate that the addition of the extra trace elements is beneficial. No negative effects on the biomass yield or on any other bioprocess parameters could be observed in cultures with the extra trace elements if the cells were repeatedly exposed to transient oxygen limitation. PMID:18687130

  16. Complexity: an internet resource for analysis of DNA sequence complexity

    PubMed Central

    Orlov, Y. L.; Potapov, V. N.

    2004-01-01

    The search for DNA regions with low complexity is one of the pivotal tasks of modern structural analysis of complete genomes. The low complexity may be preconditioned by strong inequality in nucleotide content (biased composition), by tandem or dispersed repeats or by palindrome-hairpin structures, as well as by a combination of all these factors. Several numerical measures of textual complexity, including combinatorial and linguistic ones, together with complexity estimation using a modified Lempel–Ziv algorithm, have been implemented in a software tool called ‘Complexity’ (http://wwwmgs.bionet.nsc.ru/mgs/programs/low_complexity/). The software enables a user to search for low-complexity regions in long sequences, e.g. complete bacterial genomes or eukaryotic chromosomes. In addition, it estimates the complexity of groups of aligned sequences. PMID:15215465

  17. General Aviation Aircraft Reliability Study

    NASA Technical Reports Server (NTRS)

    Pettit, Duane; Turnbull, Andrew; Roelant, Henk A. (Technical Monitor)

    2001-01-01

    This reliability study was performed in order to provide the aviation community with an estimate of Complex General Aviation (GA) Aircraft System reliability. To successfully improve the safety and reliability for the next generation of GA aircraft, a study of current GA aircraft attributes was prudent. This was accomplished by benchmarking the reliability of operational Complex GA Aircraft Systems. Specifically, Complex GA Aircraft System reliability was estimated using data obtained from the logbooks of a random sample of the Complex GA Aircraft population.

  18. Continuous stroke volume estimation from aortic pressure using zero dimensional cardiovascular model: proof of concept study from porcine experiments.

    PubMed

    Kamoi, Shun; Pretty, Christopher; Docherty, Paul; Squire, Dougie; Revie, James; Chiew, Yeong Shiong; Desaive, Thomas; Shaw, Geoffrey M; Chase, J Geoffrey

    2014-01-01

    Accurate, continuous, left ventricular stroke volume (SV) measurements can convey large amounts of information about patient hemodynamic status and response to therapy. However, direct measurements are highly invasive in clinical practice, and current procedures for estimating SV require specialized devices and significant approximation. This study investigates the accuracy of a three element Windkessel model combined with an aortic pressure waveform to estimate SV. Aortic pressure is separated into two components capturing; 1) resistance and compliance, 2) characteristic impedance. This separation provides model-element relationships enabling SV to be estimated while requiring only one of the three element values to be known or estimated. Beat-to-beat SV estimation was performed using population-representative optimal values for each model element. This method was validated using measured SV data from porcine experiments (N = 3 female Pietrain pigs, 29-37 kg) in which both ventricular volume and aortic pressure waveforms were measured simultaneously. The median difference between measured SV from left ventricle (LV) output and estimated SV was 0.6 ml with a 90% range (5th-95th percentile) -12.4 ml-14.3 ml. During periods when changes in SV were induced, cross correlations in between estimated and measured SV were above R = 0.65 for all cases. The method presented demonstrates that the magnitude and trends of SV can be accurately estimated from pressure waveforms alone, without the need for identification of complex physiological metrics where strength of correlations may vary significantly from patient to patient.

  19. New Insights into the Estimation of Extreme Geomagnetic Storm Occurrences

    NASA Astrophysics Data System (ADS)

    Ruffenach, Alexis; Winter, Hugo; Lavraud, Benoit; Bernardara, Pietro

    2017-04-01

    Space weather events such as intense geomagnetic storms are major disturbances of the near-Earth environment that may lead to serious impacts on our modern society. As such, it is of great importance to estimate their probability, and in particular that of extreme events. One approach largely used in statistical sciences for extreme events probability estimates is Extreme Value Analysis (EVA). Using this rigorous statistical framework, estimations of the occurrence of extreme geomagnetic storms are performed here based on the most relevant global parameters related to geomagnetic storms, such as ground parameters (e.g. geomagnetic Dst and aa indexes), and space parameters related to the characteristics of Coronal Mass Ejections (CME) (velocity, southward magnetic field component, electric field). Using our fitted model, we estimate the annual probability of a Carrington-type event (Dst = -850nT) to be on the order of 10-3, with a lower limit of the uncertainties on the return period of ˜500 years. Our estimate is significantly higher than that of most past studies, which typically had a return period of a few 100 years at maximum. Thus precautions are required when extrapolating intense values. Currently, the complexity of the processes and the length of available data inevitably leads to significant uncertainties in return period estimates for the occurrence of extreme geomagnetic storms. However, our application of extreme value models for extrapolating into the tail of the distribution provides a mathematically justified framework for the estimation of extreme return periods, thereby enabling the determination of more accurate estimates and reduced associated uncertainties.

  20. Methane bubbling from northern lakes: present and future contributions to the global methane budget.

    PubMed

    Walter, Katey M; Smith, Laurence C; Chapin, F Stuart

    2007-07-15

    Large uncertainties in the budget of atmospheric methane (CH4) limit the accuracy of climate change projections. Here we describe and quantify an important source of CH4 -- point-source ebullition (bubbling) from northern lakes -- that has not been incorporated in previous regional or global methane budgets. Employing a method recently introduced to measure ebullition more accurately by taking into account its spatial patchiness in lakes, we estimate point-source ebullition for 16 lakes in Alaska and Siberia that represent several common northern lake types: glacial, alluvial floodplain, peatland and thermokarst (thaw) lakes. Extrapolation of measured fluxes from these 16 sites to all lakes north of 45 degrees N using circumpolar databases of lake and permafrost distributions suggests that northern lakes are a globally significant source of atmospheric CH4, emitting approximately 24.2+/-10.5Tg CH4yr(-1). Thermokarst lakes have particularly high emissions because they release CH4 produced from organic matter previously sequestered in permafrost. A carbon mass balance calculation of CH4 release from thermokarst lakes on the Siberian yedoma ice complex suggests that these lakes alone would emit as much as approximately 49000Tg CH4 if this ice complex was to thaw completely. Using a space-for-time substitution based on the current lake distributions in permafrost-dominated and permafrost-free terrains, we estimate that lake emissions would be reduced by approximately 12% in a more probable transitional permafrost scenario and by approximately 53% in a 'permafrost-free' Northern Hemisphere. Long-term decline in CH4 ebullition from lakes due to lake area loss and permafrost thaw would occur only after the large release of CH4 associated thermokarst lake development in the zone of continuous permafrost.

  1. CAUSAL INFERENCE WITH A GRAPHICAL HIERARCHY OF INTERVENTIONS

    PubMed Central

    Shpitser, Ilya; Tchetgen, Eric Tchetgen

    2017-01-01

    Identifying causal parameters from observational data is fraught with subtleties due to the issues of selection bias and confounding. In addition, more complex questions of interest, such as effects of treatment on the treated and mediated effects may not always be identified even in data where treatment assignment is known and under investigator control, or may be identified under one causal model but not another. Increasingly complex effects of interest, coupled with a diversity of causal models in use resulted in a fragmented view of identification. This fragmentation makes it unnecessarily difficult to determine if a given parameter is identified (and in what model), and what assumptions must hold for this to be the case. This, in turn, complicates the development of estimation theory and sensitivity analysis procedures. In this paper, we give a unifying view of a large class of causal effects of interest, including novel effects not previously considered, in terms of a hierarchy of interventions, and show that identification theory for this large class reduces to an identification theory of random variables under interventions from this hierarchy. Moreover, we show that one type of intervention in the hierarchy is naturally associated with queries identified under the Finest Fully Randomized Causally Interpretable Structure Tree Graph (FFRCISTG) model of Robins (via the extended g-formula), and another is naturally associated with queries identified under the Non-Parametric Structural Equation Model with Independent Errors (NPSEM-IE) of Pearl, via a more general functional we call the edge g-formula. Our results motivate the study of estimation theory for the edge g-formula, since we show it arises both in mediation analysis, and in settings where treatment assignment has unobserved causes, such as models associated with Pearl’s front-door criterion. PMID:28919652

  2. Dynamics of the Wulong Landslide Revealed by Broadband Seismic Records

    NASA Astrophysics Data System (ADS)

    Huang, X.; Dan, Y.

    2016-12-01

    Long-period seismic signals are frequently used to trace the dynamic process of large scale landslides. The catastrophic WuLong landslide occurred at 14:51 on 5 June 2009 (Beijing time, UTC+8) in Wulong Prefecture, Southwest China. The topography in landslide area varies dramatically, enhancing the complexity in its movement characteristics. The mass started sliding northward on the upper part of the cliff located upon the west slope of the Tiejianggou gully, and shifted its movement direction to northeastward after being blocked by stable bedrock in front, leaving a scratch zone. The sliding mass then moved downward along the west slope of the gully until it collided with the east slope, and broke up into small pieces after the collision, forming a debris flow along the gully. We use long-period seismic signals extracted from eight broadband seismic stations within 250 km of the landslide to estimate its source time functions. Combining with topographic surveys done before and after the event, we can also resolve kinematic parameters of sliding mass, i.e. velocities, displacements and trajectories, perfectly characterizing its movement features. The runout trajectory deduced from source time functions is consistent with the sliding path, including two direction changing processes, corresponding to scratching the western bedrock and collision with the east slope respectively. Topographic variations can be reflected from estimated velocities. The maximum velocity of the sliding mass reaches 35 m/s before the collision with the east slope of the Tiejianggou gully, resulting from the height difference between the source zone and the deposition zone. What is important is that dynamics of scratching and collision can be characterized by source time functions. Our results confirm that long-period seismic signals are sufficient to characterize dynamics and kinematics of large scale landslides which occur in a region with complex topography.

  3. A Fast Multiple-Kernel Method With Applications to Detect Gene-Environment Interaction.

    PubMed

    Marceau, Rachel; Lu, Wenbin; Holloway, Shannon; Sale, Michèle M; Worrall, Bradford B; Williams, Stephen R; Hsu, Fang-Chi; Tzeng, Jung-Ying

    2015-09-01

    Kernel machine (KM) models are a powerful tool for exploring associations between sets of genetic variants and complex traits. Although most KM methods use a single kernel function to assess the marginal effect of a variable set, KM analyses involving multiple kernels have become increasingly popular. Multikernel analysis allows researchers to study more complex problems, such as assessing gene-gene or gene-environment interactions, incorporating variance-component based methods for population substructure into rare-variant association testing, and assessing the conditional effects of a variable set adjusting for other variable sets. The KM framework is robust, powerful, and provides efficient dimension reduction for multifactor analyses, but requires the estimation of high dimensional nuisance parameters. Traditional estimation techniques, including regularization and the "expectation-maximization (EM)" algorithm, have a large computational cost and are not scalable to large sample sizes needed for rare variant analysis. Therefore, under the context of gene-environment interaction, we propose a computationally efficient and statistically rigorous "fastKM" algorithm for multikernel analysis that is based on a low-rank approximation to the nuisance effect kernel matrices. Our algorithm is applicable to various trait types (e.g., continuous, binary, and survival traits) and can be implemented using any existing single-kernel analysis software. Through extensive simulation studies, we show that our algorithm has similar performance to an EM-based KM approach for quantitative traits while running much faster. We also apply our method to the Vitamin Intervention for Stroke Prevention (VISP) clinical trial, examining gene-by-vitamin effects on recurrent stroke risk and gene-by-age effects on change in homocysteine level. © 2015 WILEY PERIODICALS, INC.

  4. CAUSAL INFERENCE WITH A GRAPHICAL HIERARCHY OF INTERVENTIONS.

    PubMed

    Shpitser, Ilya; Tchetgen, Eric Tchetgen

    2016-12-01

    Identifying causal parameters from observational data is fraught with subtleties due to the issues of selection bias and confounding. In addition, more complex questions of interest, such as effects of treatment on the treated and mediated effects may not always be identified even in data where treatment assignment is known and under investigator control, or may be identified under one causal model but not another. Increasingly complex effects of interest, coupled with a diversity of causal models in use resulted in a fragmented view of identification. This fragmentation makes it unnecessarily difficult to determine if a given parameter is identified (and in what model), and what assumptions must hold for this to be the case. This, in turn, complicates the development of estimation theory and sensitivity analysis procedures. In this paper, we give a unifying view of a large class of causal effects of interest, including novel effects not previously considered, in terms of a hierarchy of interventions, and show that identification theory for this large class reduces to an identification theory of random variables under interventions from this hierarchy. Moreover, we show that one type of intervention in the hierarchy is naturally associated with queries identified under the Finest Fully Randomized Causally Interpretable Structure Tree Graph (FFRCISTG) model of Robins (via the extended g-formula), and another is naturally associated with queries identified under the Non-Parametric Structural Equation Model with Independent Errors (NPSEM-IE) of Pearl, via a more general functional we call the edge g-formula. Our results motivate the study of estimation theory for the edge g-formula, since we show it arises both in mediation analysis, and in settings where treatment assignment has unobserved causes, such as models associated with Pearl's front-door criterion.

  5. Decoding 2D-PAGE complex maps: relevance to proteomics.

    PubMed

    Pietrogrande, Maria Chiara; Marchetti, Nicola; Dondi, Francesco; Righetti, Pier Giorgio

    2006-03-20

    This review describes two mathematical approaches useful for decoding the complex signal of 2D-PAGE maps of protein mixtures. These methods are helpful for interpreting the large amount of data of each 2D-PAGE map by extracting all the analytical information hidden therein by spot overlapping. Here the basic theory and application to 2D-PAGE maps are reviewed: the means for extracting information from the experimental data and their relevance to proteomics are discussed. One method is based on the quantitative theory of statistical model of peak overlapping (SMO) using the spot experimental data (intensity and spatial coordinates). The second method is based on the study of the 2D-autocovariance function (2D-ACVF) computed on the experimental digitised map. They are two independent methods that are able to extract equal and complementary information from the 2D-PAGE map. Both methods permit to obtain fundamental information on the sample complexity and the separation performance and to single out ordered patterns present in spot positions: the availability of two independent procedures to compute the same separation parameters is a powerful tool to estimate the reliability of the obtained results. The SMO procedure is an unique tool to quantitatively estimate the degree of spot overlapping present in the map, while the 2D-ACVF method is particularly powerful in simply singling out the presence of order in the spot position from the complexity of the whole 2D map, i.e., spot trains. The procedures were validated by extensive numerical computation on computer-generated maps describing experimental 2D-PAGE gels of protein mixtures. Their applicability to real samples was tested on reference maps obtained from literature sources. The review describes the most relevant information for proteomics: sample complexity, separation performance, overlapping extent, identification of spot trains related to post-translational modifications (PTMs).

  6. An Improved Azimuth Angle Estimation Method with a Single Acoustic Vector Sensor Based on an Active Sonar Detection System.

    PubMed

    Zhao, Anbang; Ma, Lin; Ma, Xuefei; Hui, Juan

    2017-02-20

    In this paper, an improved azimuth angle estimation method with a single acoustic vector sensor (AVS) is proposed based on matched filtering theory. The proposed method is mainly applied in an active sonar detection system. According to the conventional passive method based on complex acoustic intensity measurement, the mathematical and physical model of this proposed method is described in detail. The computer simulation and lake experiments results indicate that this method can realize the azimuth angle estimation with high precision by using only a single AVS. Compared with the conventional method, the proposed method achieves better estimation performance. Moreover, the proposed method does not require complex operations in frequencydomain and achieves computational complexity reduction.

  7. [Age factor in a complex evaluation of health of air staff].

    PubMed

    Ushakov, I B; Batishcheva, G A; Chernov, Iu N; Khomenko, M N; Soldatov, S K

    2010-03-01

    Was elaborated program of a complex of estimation of health condition of air staff with determination of capability of early diagnostic of functional tension of physiological systems. According to this system there were observed 73 airmen using a complex of tests (estimation of level of pectoral control, of personal and reactive anxiety, vegetal regulation etc.). Was detected, that length of service and sympato-adrenaline activeness with vicarious decrease of adrenoreactiveness are in direct proportion. Were marked the most informative indexes of estimation of functional tension of psycho-physiological functions, vegetative regulation and cardiovascular system. Was shown that the elaborated system of individual estimation of health of air staff permits diagnose prenosological conditions and determine indexes for rehabilitation treatment.

  8. Quercetin as colorimetric reagent for determination of zirconium

    USGS Publications Warehouse

    Grimaldi, F.S.; White, C.E.

    1953-01-01

    Methods described in the literature for the determination of zirconium are generally designed for relatively large amounts of this element. A good procedure using colorimetric reagent for the determination of trace amounts is desirable. Quercetin has been found to yield a sensitive color reaction with zirconium suitable for the determination of from 0.1 to 50?? of zirconium dioxide. The procedure developed involves the separation of zirconium from interfering elements by precipitation with p-dimethylaminoazophenylarsonic acid prior to its estimation with quercetin. The quercetin reaction is carried out in 0.5N hydrochloric acid solution. Under the operating conditions it is indicated that quercetin forms a 2 to 1 complex with zirconium; however, a 2 to 1 and a 1 to 1 complex can coexist under special conditions. Approximate values for the equilibrium constants of the complexes are K1 = 0.33 ?? 10-5 and K2 = 1.3 ?? 10-9. Seven Bureau of Standards samples of glass sands and refractories were analyzed with excellent results. The method described should find considerable application in the analysis of minerals and other materials for macro as well as micro amounts of zirconium.

  9. The Complex Genetic Basis of Congenital Heart Defects

    PubMed Central

    Akhirome, Ehiole; Walton, Nephi A.; Nogee, Julie M.; Jay, Patrick Y.

    2017-01-01

    Twenty years ago, chromosomal abnormalities were the only identifiable genetic causes of a small fraction of congenital heart defects (CHD). Today, a de novo or inherited genetic abnormality can be identified as pathogenic in one-third of cases. We refer to them here as monogenic causes, insofar as the genetic abnormality has a readily detectable, large effect. What explains the other two-thirds? This review considers a complex genetic basis. That is, a combination of genetic mutations or variants that individually may have little or no detectable effect contribute to the pathogenesis of a heart defect. Genes in the embryo that act directly in cardiac developmental pathways have received the most attention, but genes in the mother that establish the gestational milieu via pathways related to metabolism and aging also have an effect. A growing body of evidence highlights the pathogenic significance of genetic interactions in the embryo and maternal effects that have a genetic basis. The investigation of CHD as guided by a complex genetic model could help estimate risk more precisely and logically lead to a means of prevention. PMID:28381817

  10. Spatiotemporal chaos and two-dimensional dissipative rogue waves in Lugiato-Lefever model

    NASA Astrophysics Data System (ADS)

    Panajotov, Krassimir; Clerc, Marcel G.; Tlidi, Mustapha

    2017-06-01

    Driven nonlinear optical cavities can exhibit complex spatiotemporal dynamics. We consider the paradigmatic Lugiato-Lefever model describing driven nonlinear optical resonator. This model is one of the most-studied nonlinear equations in optics. It describes a large spectrum of nonlinear phenomena from bistability, to periodic patterns, localized structures, self-pulsating localized structures and to a complex spatiotemporal behavior. The model is considered also as prototype model to describe several optical nonlinear devices such as Kerr media, liquid crystals, left handed materials, nonlinear fiber cavity, and frequency comb generation. We focus our analysis on a spatiotemporal chaotic dynamics in one-dimension. We identify a route to spatiotemporal chaos through an extended quasiperiodicity. We have estimated the Kaplan-Yorke dimension that provides a measure of the strange attractor complexity. Likewise, we show that the Lugiato-Leferver equation supports rogues waves in two-dimensional settings. We characterize rogue-wave formation by computing the probability distribution of the pulse height. Contribution to the Topical Issue "Theory and Applications of the Lugiato-Lefever Equation", edited by Yanne K. Chembo, Damia Gomila, Mustapha Tlidi, Curtis R. Menyuk.

  11. Application of the two-source energy balance model to partition evapotranspiration in an arid wine vineyard

    NASA Astrophysics Data System (ADS)

    Kool, Dilia; Kustas, William P.; Agam, Nurit

    2016-04-01

    The partitioning of evapotranspiration (ET) into transpiration (T), a productive water use, and soil water evaporation (E), which is generally considered a water loss, is highly relevant to agriculture in the light of increasing desertification and water scarcity. This task is challenged by the complexity of soil and plant interactions, coupled with changes in atmospheric and soil water content conditions. Many of the processes controlling water/energy exchange are not adequately modeled. The two-source energy balance model (TSEB) was evaluated and adapted for independent E and T estimations in an isolated drip-irrigated wine vineyard in the arid Negev desert. The TSEB model estimates ET by computing vegetation and soil energy fluxes using remotely sensed composite surface temperature, local weather data (solar radiation, air temperature and humidity, and wind speed), and vegetation metrics (row spacing, canopy height and width, and leaf area). The soil and vegetation energy fluxes are computed numerically using a system of temperature gradient and resistance equations; where soil and canopy temperatures are derived from the composite surface temperature. For estimation of ET, the TSEB model has been shown to perform well for various agricultural crops under a wide range of environmental conditions, but validation of T and E fluxes is limited to one study in a well-watered cotton crop. Extending the TSEB approach to water-limited vineyards demands careful consideration regarding how the complex canopy structure of vineyards will influence the accuracy of the partitioning between E and T. Data for evaluation of the TSEB model were collected over a season (bud break till harvest). Composite, canopy, and soil surface temperatures were measured using infrared thermometers. The composite vegetation and soil surface energy fluxes were assessed using independent measurements of net radiation, and soil, sensible and latent heat flux. The below canopy energy balance was assessed at the dry midrow position as well as the wet irrigated position directly underneath the vine row, where net radiation and soil heat flux were measured, sensible heat flux was computed indirectly, and E was calculated as the residual. While the below canopy energy balance approach used in this study allowed continuous assessment of E at daily intervals, instantaneous E fluxes could not be assessed due to vertical variability in shading below the canopy. Seasonal partitioning indicated that total E amounted to 9-11% of ET. Initial evaluation of the TSEB model indicated that discrepancies between modeled and measured fluxes can largely be attributed to net radiation partitioning. In addition, large diurnal variation at the soil surface requires adaptation of the soil heat flux formulations. Improved estimation of energy fluxes by accounting for the relatively complex canopy structure of vineyards will be highlighted.

  12. Genetic Characterization of Dog Personality Traits.

    PubMed

    Ilska, Joanna; Haskell, Marie J; Blott, Sarah C; Sánchez-Molano, Enrique; Polgar, Zita; Lofgren, Sarah E; Clements, Dylan N; Wiener, Pamela

    2017-06-01

    The genetic architecture of behavioral traits in dogs is of great interest to owners, breeders, and professionals involved in animal welfare, as well as to scientists studying the genetics of animal (including human) behavior. The genetic component of dog behavior is supported by between-breed differences and some evidence of within-breed variation. However, it is a challenge to gather sufficiently large datasets to dissect the genetic basis of complex traits such as behavior, which are both time-consuming and logistically difficult to measure, and known to be influenced by nongenetic factors. In this study, we exploited the knowledge that owners have of their dogs to generate a large dataset of personality traits in Labrador Retrievers. While accounting for key environmental factors, we demonstrate that genetic variance can be detected for dog personality traits assessed using questionnaire data. We identified substantial genetic variance for several traits, including fetching tendency and fear of loud noises, while other traits revealed negligibly small heritabilities. Genetic correlations were also estimated between traits; however, due to fairly large SEs, only a handful of trait pairs yielded statistically significant estimates. Genomic analyses indicated that these traits are mainly polygenic, such that individual genomic regions have small effects, and suggested chromosomal associations for six of the traits. The polygenic nature of these traits is consistent with previous behavioral genetics studies in other species, for example in mouse, and confirms that large datasets are required to quantify the genetic variance and to identify the individual genes that influence behavioral traits. Copyright © 2017 by the Genetics Society of America.

  13. Estimates of water source contributions in a dynamic urban water supply system inferred via a Bayesian stable isotope mixing model

    NASA Astrophysics Data System (ADS)

    Jameel, M. Y.; Brewer, S.; Fiorella, R.; Tipple, B. J.; Bowen, G. J.; Terry, S.

    2017-12-01

    Public water supply systems (PWSS) are complex distribution systems and critical infrastructure, making them vulnerable to physical disruption and contamination. Exploring the susceptibility of PWSS to such perturbations requires detailed knowledge of the supply system structure and operation. Although the physical structure of supply systems (i.e., pipeline connection) is usually well documented for developed cities, the actual flow patterns of water in these systems are typically unknown or estimated based on hydrodynamic models with limited observational validation. Here, we present a novel method for mapping the flow structure of water in a large, complex PWSS, building upon recent work highlighting the potential of stable isotopes of water (SIW) to document water management practices within complex PWSS. We sampled a major water distribution system of the Salt Lake Valley, Utah, measuring SIW of water sources, treatment facilities, and numerous sites within in the supply system. We then developed a hierarchical Bayesian (HB) isotope mixing model to quantify the proportion of water supplied by different sources at sites within the supply system. Known production volumes and spatial distance effects were used to define the prior probabilities for each source; however, we did not include other physical information about the supply system. Our results were in general agreement with those obtained by hydrodynamic models and provide quantitative estimates of contributions of different water sources to a given site along with robust estimates of uncertainty. Secondary properties of the supply system, such as regions of "static" and "dynamic" source (e.g., regions supplied dominantly by one source vs. those experiencing active mixing between multiple sources), can be inferred from the results. The isotope-based HB isotope mixing model offers a new investigative technique for analyzing PWSS and documenting aspects of supply system structure and operation that are otherwise challenging to observe. The method could allow water managers to document spatiotemporal variation in PWSS flow patterns, critical for interrogating the distribution system to inform operation decision making or disaster response, optimize water supply and, monitor and enforce water rights.

  14. Central Arctic Crustal Modeling Constrained by Potential Field data and recent ECS Seismic Data

    NASA Astrophysics Data System (ADS)

    Evangelatos, John; Oakey, Gordon; Saltus, Rick

    2017-04-01

    2-D gravity and magnetic models have been generated for several transects across the Alpha-Mendeleev ridge complex to study the regional variability of the crustal structure and identify large scale lateral changes. The geometry and density parameters for the models have been constrained using recently acquired seismic reflection and refraction data collected jointly by Canada and the United States as part of their collaborative Arctic ECS programs. A total of fifteen models have been generated perpendicular to the ridge complex, typically 50 to 150 km apart. A minimalist approach to modeling involved maintaining a simple, laterally continuous density structure for the crust while varying the model geometry to fit the observed gravity field. This approach is justified because low amplitude residual Bouguer anomalies suggest a relatively homogenous density structure within the ridge complex. These models have provided a new measure of the regional variability in crustal thickness. Typically, models with thinner crust correspond with deeper bathymetric depths of the ridge which is consistent with regional isostatic equilibrium. Complex "chaotic" magnetic anomalies are associated with the Alpha-Mendeleev ridge complex, which extends beneath the surrounding sedimentary basins. Pseudogravity inversion (magnetic potential) of the magnetic field provides a quantifiable areal extent of ˜1.3 x106 km2. Forward modeling confirms that the magnetic anomalies are not solely the result of magnetized bathymetric highs, but are caused to a great extent by mid- and lower crustal sources. The magnetization of the crust inferred from modeling is significantly higher than available lab measurements of onshore volcanic rocks. Although the 2-D models cannot uniquely identify whether the crustal protolith was continental or oceanic, there is a necessity for a significant content of high density and highly magnetic (ultramafic) material. Based on the crustal thickness estimates from our regional 2-D gravity models and the two possible protoliths, we determine volumetric estimates of the volcanic composition to ˜ 6 × 106 km3 for the mid- and upper-crust and between 10 × 106 and 14 × 106 km3 within the lower crust — for a total of at least ˜16 × 106 km3. This exceeds any estimates for the onshore circum-Arctic HALIP by more than an order of magnitude.

  15. A New Boundary for the High Plains - Ogallala Aquifer Complex

    NASA Astrophysics Data System (ADS)

    Haacker, E. M.; Nozari, S.; Kendall, A. D.

    2017-12-01

    In the semi-arid Great Plains, water is the key ingredient for crop growth: the difference between meager yields for many crops and an agricultural bonanza. The High Plains-Ogallala Aquifer complex (HPA) underlies 452,000 square kilometers of the region, and over 95% of water withdrawn from the aquifer is used for irrigation. Much of the HPA is being pumped unsustainably, and since the region is heavily reliant on this resource for its social and economic health, the High Plains has been a leader in groundwater management planning. However, the geographic boundary of the High Plains region fails to reflect the hydrogeological realities of the aquifer. The current boundary, recognizable from countless textbooks and news articles, is only slightly modified from a version from the 1980's, and largely follows the physiographic borders of the High Plains - defined by surface features such as escarpments and rivers - rather than the edges of water-bearing sediment sufficient for high-volume pumping. This is supported by three lines of evidence: hydrogeological observations from the original aquifer boundary determination; the extent of irrigated land, as estimated by MODIS-MIrAD data; and statistical estimates of saturated thickness, incorporating improved maps of the aquifer base and an additional 35 years of water table measurements. In this project, new maps of saturated thickness are used to create an updated aquifer boundary, which conforms with the standard definition of an aquifer as a package of sediment that yields enough water to be economically pumped. This has major implications for social and physical models, as well as water planning and estimates of sustainability for the HPA. Much of the area of the HPA that has been labeled `sustainable' based upon estimates of recharge relative to pumping estimates falls outside the updated aquifer boundary. In reality, the sustainably-pumped area of this updated aquifer boundary is far smaller—a fact that if more widely understood could help drive further regulatory action in this critical water resource region.

  16. Adaptive Green-Kubo estimates of transport coefficients from molecular dynamics based on robust error analysis.

    PubMed

    Jones, Reese E; Mandadapu, Kranthi K

    2012-04-21

    We present a rigorous Green-Kubo methodology for calculating transport coefficients based on on-the-fly estimates of: (a) statistical stationarity of the relevant process, and (b) error in the resulting coefficient. The methodology uses time samples efficiently across an ensemble of parallel replicas to yield accurate estimates, which is particularly useful for estimating the thermal conductivity of semi-conductors near their Debye temperatures where the characteristic decay times of the heat flux correlation functions are large. Employing and extending the error analysis of Zwanzig and Ailawadi [Phys. Rev. 182, 280 (1969)] and Frenkel [in Proceedings of the International School of Physics "Enrico Fermi", Course LXXV (North-Holland Publishing Company, Amsterdam, 1980)] to the integral of correlation, we are able to provide tight theoretical bounds for the error in the estimate of the transport coefficient. To demonstrate the performance of the method, four test cases of increasing computational cost and complexity are presented: the viscosity of Ar and water, and the thermal conductivity of Si and GaN. In addition to producing accurate estimates of the transport coefficients for these materials, this work demonstrates precise agreement of the computed variances in the estimates of the correlation and the transport coefficient with the extended theory based on the assumption that fluctuations follow a Gaussian process. The proposed algorithm in conjunction with the extended theory enables the calculation of transport coefficients with the Green-Kubo method accurately and efficiently.

  17. Adaptive Green-Kubo estimates of transport coefficients from molecular dynamics based on robust error analysis

    NASA Astrophysics Data System (ADS)

    Jones, Reese E.; Mandadapu, Kranthi K.

    2012-04-01

    We present a rigorous Green-Kubo methodology for calculating transport coefficients based on on-the-fly estimates of: (a) statistical stationarity of the relevant process, and (b) error in the resulting coefficient. The methodology uses time samples efficiently across an ensemble of parallel replicas to yield accurate estimates, which is particularly useful for estimating the thermal conductivity of semi-conductors near their Debye temperatures where the characteristic decay times of the heat flux correlation functions are large. Employing and extending the error analysis of Zwanzig and Ailawadi [Phys. Rev. 182, 280 (1969)], 10.1103/PhysRev.182.280 and Frenkel [in Proceedings of the International School of Physics "Enrico Fermi", Course LXXV (North-Holland Publishing Company, Amsterdam, 1980)] to the integral of correlation, we are able to provide tight theoretical bounds for the error in the estimate of the transport coefficient. To demonstrate the performance of the method, four test cases of increasing computational cost and complexity are presented: the viscosity of Ar and water, and the thermal conductivity of Si and GaN. In addition to producing accurate estimates of the transport coefficients for these materials, this work demonstrates precise agreement of the computed variances in the estimates of the correlation and the transport coefficient with the extended theory based on the assumption that fluctuations follow a Gaussian process. The proposed algorithm in conjunction with the extended theory enables the calculation of transport coefficients with the Green-Kubo method accurately and efficiently.

  18. The Relative Performance of High Resolution Quantitative Precipitation Estimates in the Russian River Basin

    NASA Astrophysics Data System (ADS)

    Bytheway, J. L.; Biswas, S.; Cifelli, R.; Hughes, M.

    2017-12-01

    The Russian River carves a 110 mile path through Mendocino and Sonoma counties in western California, providing water for thousands of residents and acres of agriculture as well as a home for several species of endangered fish. The Russian River basin receives almost all of its precipitation during the October through March wet season, and the systems bringing this precipitation are often impacted by atmospheric river events as well as the complex topography of the region. This study will examine the performance of several high resolution (hourly, < 5km) estimates of precipitation from observational products and forecasts over the 2015-2016 and 2016-2017 wet seasons. Comparisons of event total rainfall as well as hourly rainfall will be performed using 1) rain gauges operated by the National Oceanic and Atmospheric Administration (NOAA) Physical Sciences Division (PSD), 2) products from the Multi-Radar/Multi-Sensor (MRMS) QPE dataset, and 3) quantitative precipitation forecasts from the High Resolution Rapid Refresh (HRRR) model at 1, 3, 6, and 12 hour lead times. Further attention will be given to cases or locations representing large disparities between the estimates.

  19. Estimation of Stormwater Interception Rate for various LID Facilities

    NASA Astrophysics Data System (ADS)

    Kim, S.; Lee, O.; Choi, J.

    2017-12-01

    In this study, the stormwater interception rate is proposed to apply in the design of LID facilities. For this purpose, EPA-SWMM is built with some areas of Noksan National Industrial Complex where long-term observed stormwater data were monitored and stormwater interception rates for various design capacities of various LID facilities are estimated. While the sensitivity of stormwater interception rate according to design specifications of bio-retention and infiltration trench facilities is not large, the sensitivity of stormwater interception rate according to local rainfall characteristics is relatively big. As a result of comparing the present rainfall interception rate estimation method which is officially operated in Korea with the one proposed in this study, it will be presented that the present method is highly likely to overestimate the performance of the bio-retention and infiltration trench facilities. Finally, a new stormwater interception rate formulas for the bio-retention and infiltration trench LID facilities will be proposed. Acknowledgement This research was supported by a grant (2016000200002) from Public Welfare Technology Development Program funded by Ministry of Environment of Korean government.

  20. Estimation of alluvial-fill thickness in the Mimbres ground-water basin, New Mexico, from interpretation of isostatic residual gravity anomalies

    USGS Publications Warehouse

    Heywood, Charles E.

    2002-01-01

    The geologic structure of the Mimbres ground-water basin in southwest New Mexico is characterized by north- and northwest-trending structural subbasins. Sedimentation of Miocene and Pliocene age has filled and obscured the boundaries of these subbasins and formed poten- tially productive aquifers of varied thickness. The location and depth of the subbasins can be esti- mated from analysis of isostatic residual gravity anomalies. Density contrasts of various basement lithologies generate complex regional gravity trends, which are convolved with the gravity signal from the Miocene and Pliocene alluvial fill. An iterative scheme was used to separate these regional gravity trends from the alluvial-fill grav- ity signal, which was inverted with estimated depth-density relations to compute the thickness of the alluvial fill at 1-kilometer spacing. The thickness estimates were constrained by explor- atory drill-hole information, interpreted seismic- refraction profiles, and location of bedrock lithol- ogy from surficial geologic mapping. The result- ing map of alluvial-fill thickness suggests large areas of thin alluvium that separate deep structural subbasins.

  1. Target Tracking Using SePDAF under Ambiguous Angles for Distributed Array Radar

    PubMed Central

    Long, Teng; Zhang, Honggang; Zeng, Tao; Chen, Xinliang; Liu, Quanhua; Zheng, Le

    2016-01-01

    Distributed array radar can improve radar detection capability and measurement accuracy. However, it will suffer cyclic ambiguity in its angle estimates according to the spatial Nyquist sampling theorem since the large sparse array is undersampling. Consequently, the state estimation accuracy and track validity probability degrades when the ambiguous angles are directly used for target tracking. This paper proposes a second probability data association filter (SePDAF)-based tracking method for distributed array radar. Firstly, the target motion model and radar measurement model is built. Secondly, the fusion result of each radar’s estimation is employed to the extended Kalman filter (EKF) to finish the first filtering. Thirdly, taking this result as prior knowledge, and associating with the array-processed ambiguous angles, the SePDAF is applied to accomplish the second filtering, and then achieving a high accuracy and stable trajectory with relatively low computational complexity. Moreover, the azimuth filtering accuracy will be promoted dramatically and the position filtering accuracy will also improve. Finally, simulations illustrate the effectiveness of the proposed method. PMID:27618058

  2. Nested Sampling for Bayesian Model Comparison in the Context of Salmonella Disease Dynamics

    PubMed Central

    Dybowski, Richard; McKinley, Trevelyan J.; Mastroeni, Pietro; Restif, Olivier

    2013-01-01

    Understanding the mechanisms underlying the observed dynamics of complex biological systems requires the statistical assessment and comparison of multiple alternative models. Although this has traditionally been done using maximum likelihood-based methods such as Akaike's Information Criterion (AIC), Bayesian methods have gained in popularity because they provide more informative output in the form of posterior probability distributions. However, comparison between multiple models in a Bayesian framework is made difficult by the computational cost of numerical integration over large parameter spaces. A new, efficient method for the computation of posterior probabilities has recently been proposed and applied to complex problems from the physical sciences. Here we demonstrate how nested sampling can be used for inference and model comparison in biological sciences. We present a reanalysis of data from experimental infection of mice with Salmonella enterica showing the distribution of bacteria in liver cells. In addition to confirming the main finding of the original analysis, which relied on AIC, our approach provides: (a) integration across the parameter space, (b) estimation of the posterior parameter distributions (with visualisations of parameter correlations), and (c) estimation of the posterior predictive distributions for goodness-of-fit assessments of the models. The goodness-of-fit results suggest that alternative mechanistic models and a relaxation of the quasi-stationary assumption should be considered. PMID:24376528

  3. Avoiding Decline: Fostering Resilience and Sustainability in ...

    EPA Pesticide Factsheets

    Eighty-five percent of United States citizens live in urban areas. However, research surrounding the resilience and sustainability of complex urban systems focuses largely on coastal megacities (>1 million people). Midsize cities differ from their larger counterparts due to tight urban-rural feedbacks with their immediate natural environments that result from heavy reliance and close management of local ecosystem services. They also may be less path-dependent than larger cities due to shorter average connection length among system components, contributing to higher responsiveness among social, infrastructural, and ecological feedbacks. These distinct midsize city features call for a framework that organizes information and concepts concerning the sustainability of midsize cities specifically. We argue that an integrative approach is necessary to capture properties emergent from the complex interactions of the social, infrastructural, and ecological subsystems that comprise a city system. We suggest approaches to estimate the relative resilience of midsize cities, and include an example assessment to illustrate one such estimation approach. Resilience assessments of a midsize city can be used to examine why some cities end up on sustainable paths while others diverge to unsustainable paths, and which feedbacks may be partially responsible. They also provide insight into how city planners and decision makers can use information about the resilience of midsize citi

  4. Smart algorithms and adaptive methods in computational fluid dynamics

    NASA Astrophysics Data System (ADS)

    Tinsley Oden, J.

    1989-05-01

    A review is presented of the use of smart algorithms which employ adaptive methods in processing large amounts of data in computational fluid dynamics (CFD). Smart algorithms use a rationally based set of criteria for automatic decision making in an attempt to produce optimal simulations of complex fluid dynamics problems. The information needed to make these decisions is not known beforehand and evolves in structure and form during the numerical solution of flow problems. Once the code makes a decision based on the available data, the structure of the data may change, and criteria may be reapplied in order to direct the analysis toward an acceptable end. Intelligent decisions are made by processing vast amounts of data that evolve unpredictably during the calculation. The basic components of adaptive methods and their application to complex problems of fluid dynamics are reviewed. The basic components of adaptive methods are: (1) data structures, that is what approaches are available for modifying data structures of an approximation so as to reduce errors; (2) error estimation, that is what techniques exist for estimating error evolution in a CFD calculation; and (3) solvers, what algorithms are available which can function in changing meshes. Numerical examples which demonstrate the viability of these approaches are presented.

  5. The early maximum likelihood estimation model of audiovisual integration in speech perception.

    PubMed

    Andersen, Tobias S

    2015-05-01

    Speech perception is facilitated by seeing the articulatory mouth movements of the talker. This is due to perceptual audiovisual integration, which also causes the McGurk-MacDonald illusion, and for which a comprehensive computational account is still lacking. Decades of research have largely focused on the fuzzy logical model of perception (FLMP), which provides excellent fits to experimental observations but also has been criticized for being too flexible, post hoc and difficult to interpret. The current study introduces the early maximum likelihood estimation (MLE) model of audiovisual integration to speech perception along with three model variations. In early MLE, integration is based on a continuous internal representation before categorization, which can make the model more parsimonious by imposing constraints that reflect experimental designs. The study also shows that cross-validation can evaluate models of audiovisual integration based on typical data sets taking both goodness-of-fit and model flexibility into account. All models were tested on a published data set previously used for testing the FLMP. Cross-validation favored the early MLE while more conventional error measures favored more complex models. This difference between conventional error measures and cross-validation was found to be indicative of over-fitting in more complex models such as the FLMP.

  6. Regression analysis for bivariate gap time with missing first gap time data.

    PubMed

    Huang, Chia-Hui; Chen, Yi-Hau

    2017-01-01

    We consider ordered bivariate gap time while data on the first gap time are unobservable. This study is motivated by the HIV infection and AIDS study, where the initial HIV contracting time is unavailable, but the diagnosis times for HIV and AIDS are available. We are interested in studying the risk factors for the gap time between initial HIV contraction and HIV diagnosis, and gap time between HIV and AIDS diagnoses. Besides, the association between the two gap times is also of interest. Accordingly, in the data analysis we are faced with two-fold complexity, namely data on the first gap time is completely missing, and the second gap time is subject to induced informative censoring due to dependence between the two gap times. We propose a modeling framework for regression analysis of bivariate gap time under the complexity of the data. The estimating equations for the covariate effects on, as well as the association between, the two gap times are derived through maximum likelihood and suitable counting processes. Large sample properties of the resulting estimators are developed by martingale theory. Simulations are performed to examine the performance of the proposed analysis procedure. An application of data from the HIV and AIDS study mentioned above is reported for illustration.

  7. A Panel of Ancestry Informative Markers for the Complex Five-Way Admixed South African Coloured Population

    PubMed Central

    Daya, Michelle; van der Merwe, Lize; Galal, Ushma; Möller, Marlo; Salie, Muneeb; Chimusa, Emile R.; Galanter, Joshua M.; van Helden, Paul D.; Henn, Brenna M.; Gignoux, Chris R.; Hoal, Eileen

    2013-01-01

    Admixture is a well known confounder in genetic association studies. If genome-wide data is not available, as would be the case for candidate gene studies, ancestry informative markers (AIMs) are required in order to adjust for admixture. The predominant population group in the Western Cape, South Africa, is the admixed group known as the South African Coloured (SAC). A small set of AIMs that is optimized to distinguish between the five source populations of this population (African San, African non-San, European, South Asian, and East Asian) will enable researchers to cost-effectively reduce false-positive findings resulting from ignoring admixture in genetic association studies of the population. Using genome-wide data to find SNPs with large allele frequency differences between the source populations of the SAC, as quantified by Rosenberg et. al's -statistic, we developed a panel of AIMs by experimenting with various selection strategies. Subsets of different sizes were evaluated by measuring the correlation between ancestry proportions estimated by each AIM subset with ancestry proportions estimated using genome-wide data. We show that a panel of 96 AIMs can be used to assess ancestry proportions and to adjust for the confounding effect of the complex five-way admixture that occurred in the South African Coloured population. PMID:24376522

  8. New method for estimation of fluence complexity in IMRT fields and correlation with gamma analysis

    NASA Astrophysics Data System (ADS)

    Hanušová, T.; Vondráček, V.; Badraoui-Čuprová, K.; Horáková, I.; Koniarová, I.

    2015-01-01

    A new method for estimation of fluence complexity in Intensity Modulated Radiation Therapy (IMRT) fields is proposed. Unlike other previously published works, it is based on portal images calculated by the Portal Dose Calculation algorithm in Eclipse (version 8.6, Varian Medical Systems) in the plane of the EPID aS500 detector (Varian Medical Systems). Fluence complexity is given by the number and the amplitudes of dose gradients in these matrices. Our method is validated using a set of clinical plans where fluence has been smoothed manually so that each plan has a different level of complexity. Fluence complexity calculated with our tool is in accordance with the different levels of smoothing as well as results of gamma analysis, when calculated and measured dose matrices are compared. Thus, it is possible to estimate plan complexity before carrying out the measurement. If appropriate thresholds are determined which would distinguish between acceptably and overly modulated plans, this might save time in the re-planning and re-measuring process.

  9. The value of value of information: best informing research design and prioritization using current methods.

    PubMed

    Eckermann, Simon; Karnon, Jon; Willan, Andrew R

    2010-01-01

    Value of information (VOI) methods have been proposed as a systematic approach to inform optimal research design and prioritization. Four related questions arise that VOI methods could address. (i) Is further research for a health technology assessment (HTA) potentially worthwhile? (ii) Is the cost of a given research design less than its expected value? (iii) What is the optimal research design for an HTA? (iv) How can research funding be best prioritized across alternative HTAs? Following Occam's razor, we consider the usefulness of VOI methods in informing questions 1-4 relative to their simplicity of use. Expected value of perfect information (EVPI) with current information, while simple to calculate, is shown to provide neither a necessary nor a sufficient condition to address question 1, given that what EVPI needs to exceed varies with the cost of research design, which can vary from very large down to negligible. Hence, for any given HTA, EVPI does not discriminate, as it can be large and further research not worthwhile or small and further research worthwhile. In contrast, each of questions 1-4 are shown to be fully addressed (necessary and sufficient) where VOI methods are applied to maximize expected value of sample information (EVSI) minus expected costs across designs. In comparing complexity in use of VOI methods, applying the central limit theorem (CLT) simplifies analysis to enable easy estimation of EVSI and optimal overall research design, and has been shown to outperform bootstrapping, particularly with small samples. Consequently, VOI methods applying the CLT to inform optimal overall research design satisfy Occam's razor in both improving decision making and reducing complexity. Furthermore, they enable consideration of relevant decision contexts, including option value and opportunity cost of delay, time, imperfect implementation and optimal design across jurisdictions. More complex VOI methods such as bootstrapping of the expected value of partial EVPI may have potential value in refining overall research design. However, Occam's razor must be seriously considered in application of these VOI methods, given their increased complexity and current limitations in informing decision making, with restriction to EVPI rather than EVSI and not allowing for important decision-making contexts. Initial use of CLT methods to focus these more complex partial VOI methods towards where they may be useful in refining optimal overall trial design is suggested. Integrating CLT methods with such partial VOI methods to allow estimation of partial EVSI is suggested in future research to add value to the current VOI toolkit.

  10. Hardware Implementation of a MIMO Decoder Using Matrix Factorization Based Channel Estimation

    NASA Astrophysics Data System (ADS)

    Islam, Mohammad Tariqul; Numan, Mostafa Wasiuddin; Misran, Norbahiah; Ali, Mohd Alauddin Mohd; Singh, Mandeep

    2011-05-01

    This paper presents an efficient hardware realization of multiple-input multiple-output (MIMO) wireless communication decoder that utilizes the available resources by adopting the technique of parallelism. The hardware is designed and implemented on Xilinx Virtex™-4 XC4VLX60 field programmable gate arrays (FPGA) device in a modular approach which simplifies and eases hardware update, and facilitates testing of the various modules independently. The decoder involves a proficient channel estimation module that employs matrix factorization on least squares (LS) estimation to reduce a full rank matrix into a simpler form in order to eliminate matrix inversion. This results in performance improvement and complexity reduction of the MIMO system. Performance evaluation of the proposed method is validated through MATLAB simulations which indicate 2 dB improvement in terms of SNR compared to LS estimation. Moreover complexity comparison is performed in terms of mathematical operations, which shows that the proposed approach appreciably outperforms LS estimation at a lower complexity and represents a good solution for channel estimation technique.

  11. Practical aspects of modeling aircraft dynamics from flight data

    NASA Technical Reports Server (NTRS)

    Iliff, K. W.; Maine, R. E.

    1984-01-01

    The purpose of parameter estimation, a subset of system identification, is to estimate the coefficients (such as stability and control derivatives) of the aircraft differential equations of motion from sampled measured dynamic responses. In the past, the primary reason for estimating stability and control derivatives from flight tests was to make comparisons with wind tunnel estimates. As aircraft became more complex, and as flight envelopes were expanded to include flight regimes that were not well understood, new requirements for the derivative estimates evolved. For many years, the flight determined derivatives were used in simulations to aid in flight planning and in pilot training. The simulations were particularly important in research flight test programs in which an envelope expansion into new flight regimes was required. Parameter estimation techniques for estimating stability and control derivatives from flight data became more sophisticated to support the flight test programs. As knowledge of these new flight regimes increased, more complex aircraft were flown. Much of this increased complexity was in sophisticated flight control systems. The design and refinement of the control system required higher fidelity simulations than were previously required.

  12. Evaluation of assigned-value uncertainty for complex calibrator value assignment processes: a prealbumin example.

    PubMed

    Middleton, John; Vaks, Jeffrey E

    2007-04-01

    Errors of calibrator-assigned values lead to errors in the testing of patient samples. The ability to estimate the uncertainties of calibrator-assigned values and other variables minimizes errors in testing processes. International Organization of Standardization guidelines provide simple equations for the estimation of calibrator uncertainty with simple value-assignment processes, but other methods are needed to estimate uncertainty in complex processes. We estimated the assigned-value uncertainty with a Monte Carlo computer simulation of a complex value-assignment process, based on a formalized description of the process, with measurement parameters estimated experimentally. This method was applied to study uncertainty of a multilevel calibrator value assignment for a prealbumin immunoassay. The simulation results showed that the component of the uncertainty added by the process of value transfer from the reference material CRM470 to the calibrator is smaller than that of the reference material itself (<0.8% vs 3.7%). Varying the process parameters in the simulation model allowed for optimizing the process, while keeping the added uncertainty small. The patient result uncertainty caused by the calibrator uncertainty was also found to be small. This method of estimating uncertainty is a powerful tool that allows for estimation of calibrator uncertainty for optimization of various value assignment processes, with a reduced number of measurements and reagent costs, while satisfying the requirements to uncertainty. The new method expands and augments existing methods to allow estimation of uncertainty in complex processes.

  13. Grouping methods for estimating the prevalences of rare traits from complex survey data that preserve confidentiality of respondents.

    PubMed

    Hyun, Noorie; Gastwirth, Joseph L; Graubard, Barry I

    2018-03-26

    Originally, 2-stage group testing was developed for efficiently screening individuals for a disease. In response to the HIV/AIDS epidemic, 1-stage group testing was adopted for estimating prevalences of a single or multiple traits from testing groups of size q, so individuals were not tested. This paper extends the methodology of 1-stage group testing to surveys with sample weighted complex multistage-cluster designs. Sample weighted-generalized estimating equations are used to estimate the prevalences of categorical traits while accounting for the error rates inherent in the tests. Two difficulties arise when using group testing in complex samples: (1) How does one weight the results of the test on each group as the sample weights will differ among observations in the same group. Furthermore, if the sample weights are related to positivity of the diagnostic test, then group-level weighting is needed to reduce bias in the prevalence estimation; (2) How does one form groups that will allow accurate estimation of the standard errors of prevalence estimates under multistage-cluster sampling allowing for intracluster correlation of the test results. We study 5 different grouping methods to address the weighting and cluster sampling aspects of complex designed samples. Finite sample properties of the estimators of prevalences, variances, and confidence interval coverage for these grouping methods are studied using simulations. National Health and Nutrition Examination Survey data are used to illustrate the methods. Copyright © 2018 John Wiley & Sons, Ltd.

  14. Interlaboratory Study Characterizing a Yeast Performance Standard for Benchmarking LC-MS Platform Performance*

    PubMed Central

    Paulovich, Amanda G.; Billheimer, Dean; Ham, Amy-Joan L.; Vega-Montoto, Lorenzo; Rudnick, Paul A.; Tabb, David L.; Wang, Pei; Blackman, Ronald K.; Bunk, David M.; Cardasis, Helene L.; Clauser, Karl R.; Kinsinger, Christopher R.; Schilling, Birgit; Tegeler, Tony J.; Variyath, Asokan Mulayath; Wang, Mu; Whiteaker, Jeffrey R.; Zimmerman, Lisa J.; Fenyo, David; Carr, Steven A.; Fisher, Susan J.; Gibson, Bradford W.; Mesri, Mehdi; Neubert, Thomas A.; Regnier, Fred E.; Rodriguez, Henry; Spiegelman, Cliff; Stein, Stephen E.; Tempst, Paul; Liebler, Daniel C.

    2010-01-01

    Optimal performance of LC-MS/MS platforms is critical to generating high quality proteomics data. Although individual laboratories have developed quality control samples, there is no widely available performance standard of biological complexity (and associated reference data sets) for benchmarking of platform performance for analysis of complex biological proteomes across different laboratories in the community. Individual preparations of the yeast Saccharomyces cerevisiae proteome have been used extensively by laboratories in the proteomics community to characterize LC-MS platform performance. The yeast proteome is uniquely attractive as a performance standard because it is the most extensively characterized complex biological proteome and the only one associated with several large scale studies estimating the abundance of all detectable proteins. In this study, we describe a standard operating protocol for large scale production of the yeast performance standard and offer aliquots to the community through the National Institute of Standards and Technology where the yeast proteome is under development as a certified reference material to meet the long term needs of the community. Using a series of metrics that characterize LC-MS performance, we provide a reference data set demonstrating typical performance of commonly used ion trap instrument platforms in expert laboratories; the results provide a basis for laboratories to benchmark their own performance, to improve upon current methods, and to evaluate new technologies. Additionally, we demonstrate how the yeast reference, spiked with human proteins, can be used to benchmark the power of proteomics platforms for detection of differentially expressed proteins at different levels of concentration in a complex matrix, thereby providing a metric to evaluate and minimize preanalytical and analytical variation in comparative proteomics experiments. PMID:19858499

  15. Assessment of the Suitability of High Resolution Numerical Weather Model Outputs for Hydrological Modelling in Mountainous Cold Regions

    NASA Astrophysics Data System (ADS)

    Rasouli, K.; Pomeroy, J. W.; Hayashi, M.; Fang, X.; Gutmann, E. D.; Li, Y.

    2017-12-01

    The hydrology of mountainous cold regions has a large spatial variability that is driven both by climate variability and near-surface process variability associated with complex terrain and patterns of vegetation, soils, and hydrogeology. There is a need to downscale large-scale atmospheric circulations towards the fine scales that cold regions hydrological processes operate at to assess their spatial variability in complex terrain and quantify uncertainties by comparison to field observations. In this research, three high resolution numerical weather prediction models, namely, the Intermediate Complexity Atmosphere Research (ICAR), Weather Research and Forecasting (WRF), and Global Environmental Multiscale (GEM) models are used to represent spatial and temporal patterns of atmospheric conditions appropriate for hydrological modelling. An area covering high mountains and foothills of the Canadian Rockies was selected to assess and compare high resolution ICAR (1 km × 1 km), WRF (4 km × 4 km), and GEM (2.5 km × 2.5 km) model outputs with station-based meteorological measurements. ICAR with very low computational cost was run with different initial and boundary conditions and with finer spatial resolution, which allowed an assessment of modelling uncertainty and scaling that was difficult with WRF. Results show that ICAR, when compared with WRF and GEM, performs very well in precipitation and air temperature modelling in the Canadian Rockies, while all three models show a fair performance in simulating wind and humidity fields. Representation of local-scale atmospheric dynamics leading to realistic fields of temperature and precipitation by ICAR, WRF, and GEM makes these models suitable for high resolution cold regions hydrological predictions in complex terrain, which is a key factor in estimating water security in western Canada.

  16. Evaluating a complex system-wide intervention using the difference in differences method: the Delivering Choice Programme.

    PubMed

    Round, Jeff; Drake, Robyn; Kendall, Edward; Addicott, Rachael; Agelopoulos, Nicky; Jones, Louise

    2015-03-01

    We report the use of difference in differences (DiD) methodology to evaluate a complex, system-wide healthcare intervention. We use the worked example of evaluating the Marie Curie Delivering Choice Programme (DCP) for advanced illness in a large urban healthcare economy. DiD was selected because a randomised controlled trial was not feasible. The method allows for before and after comparison of changes that occur in an intervention site with a matched control site. This enables analysts to control for the effect of the intervention in the absence of a local control. Any policy, seasonal or other confounding effects over the test period are assumed to have occurred in a balanced way at both sites. Data were obtained from primary care trusts. Outcomes were place of death, inpatient admissions, length of stay and costs. Small changes were identified between pre- and post-DCP outputs in the intervention site. The proportion of home deaths and median cost increased slightly, while the number of admissions per patient and the average length of stay per admission decreased slightly. None of these changes was statistically significant. Effects estimates were limited by small numbers accessing new services and selection bias in sample population and comparator site. In evaluating the effect of a complex healthcare intervention, the choice of analysis method and output measures is crucial. Alternatives to randomised controlled trials may be required for evaluating large scale complex interventions and the DiD approach is suitable, subject to careful selection of measured outputs and control population. Published by the BMJ Publishing Group Limited. For permission to use (where not already granted under a licence) please go to http://group.bmj.com/group/rights-licensing/permissions.

  17. A CU-Level Rate and Distortion Estimation Scheme for RDO of Hardware-Friendly HEVC Encoders Using Low-Complexity Integer DCTs.

    PubMed

    Lee, Bumshik; Kim, Munchurl

    2016-08-01

    In this paper, a low complexity coding unit (CU)-level rate and distortion estimation scheme is proposed for High Efficiency Video Coding (HEVC) hardware-friendly implementation where a Walsh-Hadamard transform (WHT)-based low-complexity integer discrete cosine transform (DCT) is employed for distortion estimation. Since HEVC adopts quadtree structures of coding blocks with hierarchical coding depths, it becomes more difficult to estimate accurate rate and distortion values without actually performing transform, quantization, inverse transform, de-quantization, and entropy coding. Furthermore, DCT for rate-distortion optimization (RDO) is computationally high, because it requires a number of multiplication and addition operations for various transform block sizes of 4-, 8-, 16-, and 32-orders and requires recursive computations to decide the optimal depths of CU or transform unit. Therefore, full RDO-based encoding is highly complex, especially for low-power implementation of HEVC encoders. In this paper, a rate and distortion estimation scheme is proposed in CU levels based on a low-complexity integer DCT that can be computed in terms of WHT whose coefficients are produced in prediction stages. For rate and distortion estimation in CU levels, two orthogonal matrices of 4×4 and 8×8 , which are applied to WHT that are newly designed in a butterfly structure only with addition and shift operations. By applying the integer DCT based on the WHT and newly designed transforms in each CU block, the texture rate can precisely be estimated after quantization using the number of non-zero quantized coefficients and the distortion can also be precisely estimated in transform domain without de-quantization and inverse transform required. In addition, a non-texture rate estimation is proposed by using a pseudoentropy code to obtain accurate total rate estimates. The proposed rate and the distortion estimation scheme can effectively be used for HW-friendly implementation of HEVC encoders with 9.8% loss over HEVC full RDO, which much less than 20.3% and 30.2% loss of a conventional approach and Hadamard-only scheme, respectively.

  18. Health risk assessment of volatile organic compounds exposure near Daegu dyeing industrial complex in South Korea.

    PubMed

    Shuai, Jianfei; Kim, Sunshin; Ryu, Hyeonsu; Park, Jinhyeon; Lee, Chae Kwan; Kim, Geun-Bae; Ultra, Venecio U; Yang, Wonho

    2018-04-20

    Studying human health in areas with industrial contamination is a serious and complex issue. In recent years, attention has increasingly focused on the health implications of large industrial complexes. A variety of potential toxic chemicals have been produced during manufacturing processes and activities in industrial complexes in South Korea. A large number of dyeing industries gathered together in Daegu dyeing industrial complex. The residents near the industrial complex could be often exposed to volatile organic compounds. This study aimed to evaluate VOCs levels in the ambient air of DDIC, to assess the impact on human health risks, and to find more convincing evidences to prove these VOCs emitted from DDIC. According to deterministic risk assessment, inhalation was the most important route. Residential indoor, outdoor and personal exposure air VOCs were measured by passive samplers in exposed area and controlled area in different seasons. Satisfaction with ambient environments and self-reported diseases were also obtained by questionnaire survey. The VOCs concentrations in exposed area and controlled area was compared by t-test. The relationships among every VOC were tested by correlation. The values of hazard quotient (HQ) and life cancer risk were estimated. The concentrations of measured VOCs were presented, moreover, the variety of concentrations according the distances from the residential settings to the industrial complex site in exposed area. The residential indoor, outdoor, and personal exposure concentrations of toluene, DMF and chloroform in exposed area were significantly higher than the corresponding concentrations in controlled area both in summer and autumn. Toluene, DMF, chloroform and MEK had significantly positive correlations with each other in indoor and outdoor, and even in personal exposure. The HQ for DMF exceeded 1, and the life cancer risk of chloroform was greater than 10 - 4 in exposed area. The prevalence of respiratory diseases, anaphylactic diseases and cardiovascular diseases in exposed area were significantly higher than in controlled area. This study showed that adverse cancer and non-cancer health effects may occur by VOCs emitted from DDIC, and some risk managements are needed. Moreover, this study provides a convenient preliminarily method for pollutants source characteristics.

  19. Acquisition Program Lead Systems Integration/Lead Capabilities Integration Decision Support Methodology and Tool

    DTIC Science & Technology

    2015-04-30

    from the MIT Sloan School that provide a relative complexity score for functions (Product and Context Complexity). The PMA assesses the complexity...collection of information is estimated to average 1 hour per response, including the time for reviewing instructions, searching existing data sources...gathering and maintaining the data needed, and completing and reviewing the collection of information. Send comments regarding this burden estimate or

  20. Cryptic Species in Tropic Sands - Interactive 3D Anatomy, Molecular Phylogeny and Evolution of Meiofaunal Pseudunelidae (Gastropoda, Acochlidia)

    PubMed Central

    Neusser, Timea P.; Jörger, Katharina M.; Schrödl, Michael

    2011-01-01

    Background Towards realistic estimations of the diversity of marine animals, tiny meiofaunal species usually are underrepresented. Since the biological species concept is hardly applicable on exotic and elusive animals, it is even more important to apply a morphospecies concept on the best level of information possible, using accurate and efficient methodology such as 3D modelling from histological sections. Molecular approaches such as sequence analyses may reveal further, cryptic species. This is the first case study on meiofaunal gastropods to test diversity estimations from traditional taxonomy against results from modern microanatomical methodology and molecular systematics. Results The examined meiofaunal Pseudunela specimens from several Indo-Pacific islands cannot be distinguished by external features. Their 3D microanatomy shows differences in the organ systems and allows for taxonomic separation in some cases. Additional molecular analyses based on partial mitochondrial cytochrome c oxidase subunit I (COI) and 16S rRNA markers revealed considerable genetic structure that is largely congruent with anatomical or geographical patterns. Two new species (Pseudunela viatoris and P. marteli spp. nov.) are formally described integrating morphological and genetic analyses. Phylogenetic analysis using partial 16S rRNA, COI and the nuclear 18S rRNA markers shows a clade of Pseudunelidae species as the sister group to limnic Acochlidiidae. Within Pseudunela, two subtypes of complex excretory systems occur. A complex kidney already evolved in the ancestor of Hedylopsacea. Several habitat shifts occurred during hedylopsacean evolution. Conclusions Cryptic species occur in tropical meiofaunal Pseudunela gastropods, and likely in other meiofaunal groups with poor dispersal abilities, boosting current diversity estimations. Only a combined 3D microanatomical and molecular approach revealed actual species diversity within Pseudunela reliably. Such integrative methods are recommended for all taxonomic approaches and biodiversity surveys on soft-bodied and small-sized invertebrates. With increasing taxon sampling and details studied, the evolution of acochlidian panpulmonates is even more complex than expected. PMID:21912592

  1. The collaborative historical African rainfall model: description and evaluation

    USGS Publications Warehouse

    Funk, Christopher C.; Michaelsen, Joel C.; Verdin, James P.; Artan, Guleid A.; Husak, Gregory; Senay, Gabriel B.; Gadain, Hussein; Magadazire, Tamuka

    2003-01-01

    In Africa the variability of rainfall in space and time is high, and the general availability of historical gauge data is low. This makes many food security and hydrologic preparedness activities difficult. In order to help overcome this limitation, we have created the Collaborative Historical African Rainfall Model (CHARM). CHARM combines three sources of information: climatologically aided interpolated (CAI) rainfall grids (monthly/0.5° ), National Centers for Environmental Prediction reanalysis precipitation fields (daily/1.875° ) and orographic enhancement estimates (daily/0.1° ). The first set of weights scales the daily reanalysis precipitation fields to match the gridded CAI monthly rainfall time series. This produces data with a daily/0.5° resolution. A diagnostic model of orographic precipitation, VDELB—based on the dot-product of the surface wind V and terrain gradient (DEL) and atmospheric buoyancy B—is then used to estimate the precipitation enhancement produced by complex terrain. Although the data are produced on 0.1° grids to facilitate integration with satellite-based rainfall estimates, the ‘true’ resolution of the data will be less than this value, and varies with station density, topography, and precipitation dynamics. The CHARM is best suited, therefore, to applications that integrate rainfall or rainfall-driven model results over large regions. The CHARM time series is compared with three independent datasets: dekadal satellite-based rainfall estimates across the continent, dekadal interpolated gauge data in Mali, and daily interpolated gauge data in western Kenya. These comparisons suggest reasonable accuracies (standard errors of about half a standard deviation) when data are aggregated to regional scales, even at daily time steps. Thus constrained, numerical weather prediction precipitation fields do a reasonable job of representing large-scale diurnal variations.

  2. Development and validation of anthropometric prediction equations for estimation of lean body mass and appendicular lean soft tissue in Indian men and women.

    PubMed

    Kulkarni, Bharati; Kuper, Hannah; Taylor, Amy; Wells, Jonathan C; Radhakrishna, K V; Kinra, Sanjay; Ben-Shlomo, Yoav; Smith, George Davey; Ebrahim, Shah; Byrne, Nuala M; Hills, Andrew P

    2013-10-15

    Lean body mass (LBM) and muscle mass remain difficult to quantify in large epidemiological studies due to the unavailability of inexpensive methods. We therefore developed anthropometric prediction equations to estimate the LBM and appendicular lean soft tissue (ALST) using dual-energy X-ray absorptiometry (DXA) as a reference method. Healthy volunteers (n = 2,220; 36% women; age 18-79 yr), representing a wide range of body mass index (14-44 kg/m(2)), participated in this study. Their LBM, including ALST, was assessed by DXA along with anthropometric measurements. The sample was divided into prediction (60%) and validation (40%) sets. In the prediction set, a number of prediction models were constructed using DXA-measured LBM and ALST estimates as dependent variables and a combination of anthropometric indices as independent variables. These equations were cross-validated in the validation set. Simple equations using age, height, and weight explained >90% variation in the LBM and ALST in both men and women. Additional variables (hip and limb circumferences and sum of skinfold thicknesses) increased the explained variation by 5-8% in the fully adjusted models predicting LBM and ALST. More complex equations using all of the above anthropometric variables could predict the DXA-measured LBM and ALST accurately, as indicated by low standard error of the estimate (LBM: 1.47 kg and 1.63 kg for men and women, respectively), as well as good agreement by Bland-Altman analyses (Bland JM, Altman D. Lancet 1: 307-310, 1986). These equations could be a valuable tool in large epidemiological studies assessing these body compartments in Indians and other population groups with similar body composition.

  3. Genome-wide heterogeneity of nucleotide substitution model fit.

    PubMed

    Arbiza, Leonardo; Patricio, Mateus; Dopazo, Hernán; Posada, David

    2011-01-01

    At a genomic scale, the patterns that have shaped molecular evolution are believed to be largely heterogeneous. Consequently, comparative analyses should use appropriate probabilistic substitution models that capture the main features under which different genomic regions have evolved. While efforts have concentrated in the development and understanding of model selection techniques, no descriptions of overall relative substitution model fit at the genome level have been reported. Here, we provide a characterization of best-fit substitution models across three genomic data sets including coding regions from mammals, vertebrates, and Drosophila (24,000 alignments). According to the Akaike Information Criterion (AIC), 82 of 88 models considered were selected as best-fit models at least in one occasion, although with very different frequencies. Most parameter estimates also varied broadly among genes. Patterns found for vertebrates and Drosophila were quite similar and often more complex than those found in mammals. Phylogenetic trees derived from models in the 95% confidence interval set showed much less variance and were significantly closer to the tree estimated under the best-fit model than trees derived from models outside this interval. Although alternative criteria selected simpler models than the AIC, they suggested similar patterns. All together our results show that at a genomic scale, different gene alignments for the same set of taxa are best explained by a large variety of different substitution models and that model choice has implications on different parameter estimates including the inferred phylogenetic trees. After taking into account the differences related to sample size, our results suggest a noticeable diversity in the underlying evolutionary process. All together, we conclude that the use of model selection techniques is important to obtain consistent phylogenetic estimates from real data at a genomic scale.

  4. Development and validation of anthropometric prediction equations for estimation of lean body mass and appendicular lean soft tissue in Indian men and women

    PubMed Central

    Kuper, Hannah; Taylor, Amy; Wells, Jonathan C.; Radhakrishna, K. V.; Kinra, Sanjay; Ben-Shlomo, Yoav; Smith, George Davey; Ebrahim, Shah; Byrne, Nuala M.; Hills, Andrew P.

    2013-01-01

    Lean body mass (LBM) and muscle mass remain difficult to quantify in large epidemiological studies due to the unavailability of inexpensive methods. We therefore developed anthropometric prediction equations to estimate the LBM and appendicular lean soft tissue (ALST) using dual-energy X-ray absorptiometry (DXA) as a reference method. Healthy volunteers (n = 2,220; 36% women; age 18-79 yr), representing a wide range of body mass index (14–44 kg/m2), participated in this study. Their LBM, including ALST, was assessed by DXA along with anthropometric measurements. The sample was divided into prediction (60%) and validation (40%) sets. In the prediction set, a number of prediction models were constructed using DXA-measured LBM and ALST estimates as dependent variables and a combination of anthropometric indices as independent variables. These equations were cross-validated in the validation set. Simple equations using age, height, and weight explained >90% variation in the LBM and ALST in both men and women. Additional variables (hip and limb circumferences and sum of skinfold thicknesses) increased the explained variation by 5–8% in the fully adjusted models predicting LBM and ALST. More complex equations using all of the above anthropometric variables could predict the DXA-measured LBM and ALST accurately, as indicated by low standard error of the estimate (LBM: 1.47 kg and 1.63 kg for men and women, respectively), as well as good agreement by Bland-Altman analyses (Bland JM, Altman D. Lancet 1: 307–310, 1986). These equations could be a valuable tool in large epidemiological studies assessing these body compartments in Indians and other population groups with similar body composition. PMID:23950165

  5. Optimal estimation of large structure model errors. [in Space Shuttle controller design

    NASA Technical Reports Server (NTRS)

    Rodriguez, G.

    1979-01-01

    In-flight estimation of large structure model errors is usually required as a means of detecting inevitable deficiencies in large structure controller/estimator models. The present paper deals with a least-squares formulation which seeks to minimize a quadratic functional of the model errors. The properties of these error estimates are analyzed. It is shown that an arbitrary model error can be decomposed as the sum of two components that are orthogonal in a suitably defined function space. Relations between true and estimated errors are defined. The estimates are found to be approximations that retain many of the significant dynamics of the true model errors. Current efforts are directed toward application of the analytical results to a reference large structure model.

  6. Direct migration motion estimation and mode decision to decoder for a low-complexity decoder Wyner-Ziv video coding

    NASA Astrophysics Data System (ADS)

    Lei, Ted Chih-Wei; Tseng, Fan-Shuo

    2017-07-01

    This paper addresses the problem of high-computational complexity decoding in traditional Wyner-Ziv video coding (WZVC). The key focus is the migration of two traditionally high-computationally complex encoder algorithms, namely motion estimation and mode decision. In order to reduce the computational burden in this process, the proposed architecture adopts the partial boundary matching algorithm and four flexible types of block mode decision at the decoder. This approach does away with the need for motion estimation and mode decision at the encoder. The experimental results show that the proposed padding block-based WZVC not only decreases decoder complexity to approximately one hundredth that of the state-of-the-art DISCOVER decoding but also outperforms DISCOVER codec by up to 3 to 4 dB.

  7. An Improved Azimuth Angle Estimation Method with a Single Acoustic Vector Sensor Based on an Active Sonar Detection System

    PubMed Central

    Zhao, Anbang; Ma, Lin; Ma, Xuefei; Hui, Juan

    2017-01-01

    In this paper, an improved azimuth angle estimation method with a single acoustic vector sensor (AVS) is proposed based on matched filtering theory. The proposed method is mainly applied in an active sonar detection system. According to the conventional passive method based on complex acoustic intensity measurement, the mathematical and physical model of this proposed method is described in detail. The computer simulation and lake experiments results indicate that this method can realize the azimuth angle estimation with high precision by using only a single AVS. Compared with the conventional method, the proposed method achieves better estimation performance. Moreover, the proposed method does not require complex operations in frequency-domain and achieves computational complexity reduction. PMID:28230763

  8. Detection of circumstellar gas associated with GG Tauri

    NASA Technical Reports Server (NTRS)

    Skrutskie, M. F.; Snell, R. L.; Strom, K. M.; Strom, S. E.; Edwards, S.; Fukui, Y.; Mizuno, A.; Hayashi, M.; Ohashi, N.

    1993-01-01

    Double-peaked (C-12)O (1-0) emission centered on the young T Tauri star GG Tau possesses a line profile which may be modeled on the assumption that CO emission arises in an extended circumstellar disk. While bounds on the observed gas mass can be estimated on this basis, it is suggested that a large amount of mass could lie within a small and optically thick region, escaping detection due to beam-dilution effects. In addition, CO may no longer accurately trace the gas mass due to its dissociation, or freezing into grains, or due to the locking-up of carbon into more complex molecules.

  9. The violent interstellar medium in Messier 31

    NASA Technical Reports Server (NTRS)

    Brinks, Elias; Braun, Robert; Unger, Stephen W.

    1990-01-01

    Taurus observations in the line of H alpha and Very Large Array (VLA) HI mapping of the HII complex No. 722 in M31, reveal what seems to be a spherical cavity 330 pc in diameter blown out by a stellar association of over 20(exp 6) years old. Evidence of induced star formation which was initiated less than 5(exp 6) years ago is present in the form of bright HII emission and numerous O, B and Wolf-Rayet stars which are found within the shell surrounding the cavity. The energy necessary to create the HI shell is estimated to be about 5(exp 51) erg.

  10. BlueSNP: R package for highly scalable genome-wide association studies using Hadoop clusters.

    PubMed

    Huang, Hailiang; Tata, Sandeep; Prill, Robert J

    2013-01-01

    Computational workloads for genome-wide association studies (GWAS) are growing in scale and complexity outpacing the capabilities of single-threaded software designed for personal computers. The BlueSNP R package implements GWAS statistical tests in the R programming language and executes the calculations across computer clusters configured with Apache Hadoop, a de facto standard framework for distributed data processing using the MapReduce formalism. BlueSNP makes computationally intensive analyses, such as estimating empirical p-values via data permutation, and searching for expression quantitative trait loci over thousands of genes, feasible for large genotype-phenotype datasets. http://github.com/ibm-bioinformatics/bluesnp

  11. Constitutive law for the densification of fused silica with applications in polishing and microgrinding

    NASA Astrophysics Data System (ADS)

    Lambropoulos, John C.; Fang, Tong; Xu, Su; Gracewski, Sheryl M.

    1995-09-01

    We discuss a constitutive model describing the permanent densification of fused silica under large applied pressures and shear stresses. The constitutive law is assumed to be rate- independent, and uses a yield function coupling hydrostatic pressure and shear stress, a flow rule describing the evolution of permanent strains after initial densification, and a hardening rule describing the dependence of the incremental densification on the levels of applied stresses. The constitutive law accounts for multiaxial states of stress, since during polishing and grinding operations complex stress states occur in a thin surface layer due to the action of abrasive particles. Due to frictional and other abrasive forces, large shear stresses are present near the surface during manufacturing. We apply the constitutive law in estimating the extent of the densified layer during the mechanical interaction of an abrasive grain and a flat surface.

  12. Decentralized control of large flexible structures by joint decoupling

    NASA Technical Reports Server (NTRS)

    Su, Tzu-Jeng; Juang, Jer-Nan

    1994-01-01

    This paper presents a novel method to design decentralized controllers for large complex flexible structures by using the idea of joint decoupling. Decoupling of joint degrees of freedom from the interior degrees of freedom is achieved by setting the joint actuator commands to cancel the internal forces exerting on the joint degrees of freedom. By doing so, the interactions between substructures are eliminated. The global structure control design problem is then decomposed into several substructure control design problems. Control commands for interior actuators are set to be localized state feedback using decentralized observers for state estimation. The proposed decentralized controllers can operate successfully at the individual substructure level as well as at the global structure level. Not only control design but also control implementation is decentralized. A two-component mass-spring-damper system is used as an example to demonstrate the proposed method.

  13. A unified framework of image latent feature learning on Sina microblog

    NASA Astrophysics Data System (ADS)

    Wei, Jinjin; Jin, Zhigang; Zhou, Yuan; Zhang, Rui

    2015-10-01

    Large-scale user-contributed images with texts are rapidly increasing on the social media websites, such as Sina microblog. However, the noise and incomplete correspondence between the images and the texts give rise to the difficulty in precise image retrieval and ranking. In this paper, a hypergraph-based learning framework is proposed for image ranking, which simultaneously utilizes visual feature, textual content and social link information to estimate the relevance between images. Representing each image as a vertex in the hypergraph, complex relationship between images can be reflected exactly. Then updating the weight of hyperedges throughout the hypergraph learning process, the effect of different edges can be adaptively modulated in the constructed hypergraph. Furthermore, the popularity degree of the image is employed to re-rank the retrieval results. Comparative experiments on a large-scale Sina microblog data-set demonstrate the effectiveness of the proposed approach.

  14. Isolation with Migration Models for More Than Two Populations

    PubMed Central

    Hey, Jody

    2010-01-01

    A method for studying the divergence of multiple closely related populations is described and assessed. The approach of Hey and Nielsen (2007, Integration within the Felsenstein equation for improved Markov chain Monte Carlo methods in population genetics. Proc Natl Acad Sci USA. 104:2785–2790) for fitting an isolation-with-migration model was extended to the case of multiple populations with a known phylogeny. Analysis of simulated data sets reveals the kinds of history that are accessible with a multipopulation analysis. Necessarily, processes associated with older time periods in a phylogeny are more difficult to estimate; and histories with high levels of gene flow are particularly difficult with more than two populations. However, for histories with modest levels of gene flow, or for very large data sets, it is possible to study large complex divergence problems that involve multiple closely related populations or species. PMID:19955477

  15. Isolation with migration models for more than two populations.

    PubMed

    Hey, Jody

    2010-04-01

    A method for studying the divergence of multiple closely related populations is described and assessed. The approach of Hey and Nielsen (2007, Integration within the Felsenstein equation for improved Markov chain Monte Carlo methods in population genetics. Proc Natl Acad Sci USA. 104:2785-2790) for fitting an isolation-with-migration model was extended to the case of multiple populations with a known phylogeny. Analysis of simulated data sets reveals the kinds of history that are accessible with a multipopulation analysis. Necessarily, processes associated with older time periods in a phylogeny are more difficult to estimate; and histories with high levels of gene flow are particularly difficult with more than two populations. However, for histories with modest levels of gene flow, or for very large data sets, it is possible to study large complex divergence problems that involve multiple closely related populations or species.

  16. Potential application of machine learning in health outcomes research and some statistical cautions.

    PubMed

    Crown, William H

    2015-03-01

    Traditional analytic methods are often ill-suited to the evolving world of health care big data characterized by massive volume, complexity, and velocity. In particular, methods are needed that can estimate models efficiently using very large datasets containing healthcare utilization data, clinical data, data from personal devices, and many other sources. Although very large, such datasets can also be quite sparse (e.g., device data may only be available for a small subset of individuals), which creates problems for traditional regression models. Many machine learning methods address such limitations effectively but are still subject to the usual sources of bias that commonly arise in observational studies. Researchers using machine learning methods such as lasso or ridge regression should assess these models using conventional specification tests. Copyright © 2015 International Society for Pharmacoeconomics and Outcomes Research (ISPOR). Published by Elsevier Inc. All rights reserved.

  17. Fast adaptive flat-histogram ensemble to enhance the sampling in large systems

    NASA Astrophysics Data System (ADS)

    Xu, Shun; Zhou, Xin; Jiang, Yi; Wang, YanTing

    2015-09-01

    An efficient novel algorithm was developed to estimate the Density of States (DOS) for large systems by calculating the ensemble means of an extensive physical variable, such as the potential energy, U, in generalized canonical ensembles to interpolate the interior reverse temperature curve , where S( U) is the logarithm of the DOS. This curve is computed with different accuracies in different energy regions to capture the dependence of the reverse temperature on U without setting prior grid in the U space. By combining with a U-compression transformation, we decrease the computational complexity from O( N 3/2) in the normal Wang Landau type method to O( N 1/2) in the current algorithm, as the degrees of freedom of system N. The efficiency of the algorithm is demonstrated by applying to Lennard Jones fluids with various N, along with its ability to find different macroscopic states, including metastable states.

  18. Crustal-scale electrical conductivity anomaly beneath inflating Lazufre volcanic complex, Central Andes

    NASA Astrophysics Data System (ADS)

    Budach, Ingmar; Brasse, Heinrich; Díaz, Daniel

    2013-03-01

    Large-scale surface deformation was observed at Lazufre volcanic center in the Central Andes of Northern Chile/Northwestern Argentina by means of Interferometric Synthetic Aperture Radar (InSAR). Uplift started there after 1998 and increased dramatically in the following years up to a rate of 3 cm/a. Lazufre is now one of the largest deforming volcano systems on Earth, but the cause for uplift - likely influx of magmatic material into the crust - is still poorly understood. In the beginning of 2010 a magnetotelluric survey was conducted to delineate the electrical conductivity distribution in the area. Several long-period magnetotelluric (LMT) sites and two broadband magnetotelluric (BBMT) sites were set up on an EW trending profile crossing the volcanic center; furthermore some LMT sites were arranged circularly around Lazufre complex and adjacent Lastarria volcano. Data were processed using an algorithm for robust and remote reference transfer function estimation. Electrical strike directions were estimated and induction arrows were derived. Although electrical strike is rather ambiguous, in a first step a 2-D resistivity model was calculated. The most prominent feature of this model is a well conducting structure rising from the upper mantle to the shallow crust beneath the center of elevation. This can be interpreted as partial melts ascending from the asthenospheric wedge and feeding a potential magma reservoir beneath Lazufre volcanic center. An improved model is finally achieved by 3-D inversion, supporting this feature. We assume that these rising melts are the source of the observed uplift at Lazufre complex.

  19. Targeting a Complex Transcriptome: The Construction of the Mouse Full-Length cDNA Encyclopedia

    PubMed Central

    Carninci, Piero; Waki, Kazunori; Shiraki, Toshiyuki; Konno, Hideaki; Shibata, Kazuhiro; Itoh, Masayoshi; Aizawa, Katsunori; Arakawa, Takahiro; Ishii, Yoshiyuki; Sasaki, Daisuke; Bono, Hidemasa; Kondo, Shinji; Sugahara, Yuichi; Saito, Rintaro; Osato, Naoki; Fukuda, Shiro; Sato, Kenjiro; Watahiki, Akira; Hirozane-Kishikawa, Tomoko; Nakamura, Mari; Shibata, Yuko; Yasunishi, Ayako; Kikuchi, Noriko; Yoshiki, Atsushi; Kusakabe, Moriaki; Gustincich, Stefano; Beisel, Kirk; Pavan, William; Aidinis, Vassilis; Nakagawara, Akira; Held, William A.; Iwata, Hiroo; Kono, Tomohiro; Nakauchi, Hiromitsu; Lyons, Paul; Wells, Christine; Hume, David A.; Fagiolini, Michela; Hensch, Takao K.; Brinkmeier, Michelle; Camper, Sally; Hirota, Junji; Mombaerts, Peter; Muramatsu, Masami; Okazaki, Yasushi; Kawai, Jun; Hayashizaki, Yoshihide

    2003-01-01

    We report the construction of the mouse full-length cDNA encyclopedia,the most extensive view of a complex transcriptome,on the basis of preparing and sequencing 246 libraries. Before cloning,cDNAs were enriched in full-length by Cap-Trapper,and in most cases,aggressively subtracted/normalized. We have produced 1,442,236 successful 3′-end sequences clustered into 171,144 groups, from which 60,770 clones were fully sequenced cDNAs annotated in the FANTOM-2 annotation. We have also produced 547,149 5′ end reads,which clustered into 124,258 groups. Altogether, these cDNAs were further grouped in 70,000 transcriptional units (TU),which represent the best coverage of a transcriptome so far. By monitoring the extent of normalization/subtraction, we define the tentative equivalent coverage (TEC),which was estimated to be equivalent to >12,000,000 ESTs derived from standard libraries. High coverage explains discrepancies between the very large numbers of clusters (and TUs) of this project,which also include non-protein-coding RNAs,and the lower gene number estimation of genome annotations. Altogether,5′-end clusters identify regions that are potential promoters for 8637 known genes and 5′-end clusters suggest the presence of almost 63,000 transcriptional starting points. An estimate of the frequency of polyadenylation signals suggests that at least half of the singletons in the EST set represent real mRNAs. Clones accounting for about half of the predicted TUs await further sequencing. The continued high-discovery rate suggests that the task of transcriptome discovery is not yet complete. PMID:12819125

  20. Methods of information geometry in computational system biology (consistency between chemical and biological evolution).

    PubMed

    Astakhov, Vadim

    2009-01-01

    Interest in simulation of large-scale metabolic networks, species development, and genesis of various diseases requires new simulation techniques to accommodate the high complexity of realistic biological networks. Information geometry and topological formalisms are proposed to analyze information processes. We analyze the complexity of large-scale biological networks as well as transition of the system functionality due to modification in the system architecture, system environment, and system components. The dynamic core model is developed. The term dynamic core is used to define a set of causally related network functions. Delocalization of dynamic core model provides a mathematical formalism to analyze migration of specific functions in biosystems which undergo structure transition induced by the environment. The term delocalization is used to describe these processes of migration. We constructed a holographic model with self-poetic dynamic cores which preserves functional properties under those transitions. Topological constraints such as Ricci flow and Pfaff dimension were found for statistical manifolds which represent biological networks. These constraints can provide insight on processes of degeneration and recovery which take place in large-scale networks. We would like to suggest that therapies which are able to effectively implement estimated constraints, will successfully adjust biological systems and recover altered functionality. Also, we mathematically formulate the hypothesis that there is a direct consistency between biological and chemical evolution. Any set of causal relations within a biological network has its dual reimplementation in the chemistry of the system environment.

Top