Science.gov

Sample records for additional simplifying assumptions

  1. Impact of velocity distribution assumption on simplified laser speckle imaging equation

    PubMed Central

    Ramirez-San-Juan, Julio C; Ramos-Garcia, Ruben; Guizar-Iturbide, Ileana; Martinez-Niconoff, Gabriel; Choi, Bernard

    2012-01-01

    Since blood flow is tightly coupled to the health status of biological tissue, several instruments have been developed to monitor blood flow and perfusion dynamics. One such instrument is laser speckle imaging. The goal of this study was to evaluate the use of two velocity distribution assumptions (Lorentzian- and Gaussian-based) to calculate speckle flow index (SFI) values. When the normalized autocorrelation function for the Lorentzian and Gaussian velocity distributions satisfy the same definition of correlation time, then the same velocity range is predicted for low speckle contrast (0 < C < 0.6) and predict different flow velocity range for high contrast. Our derived equations form the basis for simplified calculations of SFI values. PMID:18542407

  2. Part-whole bias in intertemporal choice: An empirical study of additive assumption

    NASA Astrophysics Data System (ADS)

    Lu, Yang; Wu, Dongmei; Zhuang, Xintian

    2016-12-01

    Additive assumption means the overall value of multiple-dated outcomes is based on a simple aggregation of the values of each individual outcome. This assumption is generally accepted in the field of intertemporal choices. However, recent studies show additive assumption is questionable. In this paper, we experimentally tested the additive property of multiple-dated monetary rewards. Our results show: (1) additive assumption does not hold regardless of gain or loss; (2) the sum of subjective values of individual rewards is consistently larger than the valuation placed on the same rewards as a whole. This finding suggests that part-whole bias exists in the context of valuation of intertemporal monetary rewards.

  3. Feasibility of a simplified fuel additive evaluation protocol

    SciTech Connect

    Lister, S.J.; Hunzinger, R.D.; Taghizadeh, A.

    1998-12-31

    This report describes the work carried out during the four stages of the first phase of a project that involved the determination of the feasibility of replacing the Association of American Railroads Recommended Practice (ARRP) 503 protocol for testing diesel fuel oil additives with a new procedure using the single cylinder research engine SCRE-251 as the laboratory test engine, which tests for both engine performance as well as emissions compliance. The report begins with a review of the literature on fuel additive testing, then reviews the new US Environmental Protection Agency regulations regarding locomotive diesel emissions. This is followed by a review of the ARRP 503 protocol and the proposed new procedure, a comparison of the ARRP 503 test engines and the SCRE-251, and a study of the SCRE-251`s ability to represent a multi-cylinder medium-speed diesel engine. Appendices include fuel additive manufacturers` information sheets.

  4. Simplify, simplify

    PubMed Central

    Berenbaum, May R; Seufferheld, Manfredo J; Margam, Venu M; Strycharz, Joseph P; Yoon, Kyong S; Sun, Weilin; Reenan, Robert; Lee, Si Hyeock; Clark, John M

    2011-01-01

    The body louse, with its recently sequenced genome, is now primed to serve as a powerful model organism for addressing fundamental questions relating to how insects interact with their environment. One characteristic of the body louse that facilitates this research is the size of its genome—the smallest insect genome sequenced to date. This diminutive genome must nonetheless control an organism that senses and responds to its environment, reacting to threats of corporal and genomic integrity. Additionally, the body louse transmits several important human diseases compared to its very close relative, the head louse, which does not. Therefore, these two organisms comprise an excellent model system for studying molecular mechanisms associated with vector competence. To understand more fully the development of vector/pathogen interactions, we have developed an in vitro bioassay system and determined that the body louse genome appears to contain the genes necessary for RNAi. The body louse will therefore be useful for determining the set of conditions permissive to the evolution of vector competence. PMID:21655436

  5. Coordinate swapping in standard addition graphs for analytical chemistry: a simplified path for uncertainty calculation in linear and nonlinear plots.

    PubMed

    Meija, Juris; Pagliano, Enea; Mester, Zoltán

    2014-09-02

    Uncertainty of the result from the method of standard addition is often underestimated due to neglect of the covariance between the intercept and the slope. In order to simplify the data analysis from standard addition experiments, we propose x-y coordinate swapping in conventional linear regression. Unlike the ratio of the intercept and slope, which is the result of the traditional method of standard addition, the result of the inverse standard addition is obtained directly from the intercept of the swapped calibration line. Consequently, the uncertainty evaluation becomes markedly simpler. The method is also applicable to nonlinear curves, such as the quadratic model, without incurring any additional complexity.

  6. Modeling probability and additive summation for detection across multiple mechanisms under the assumptions of signal detection theory.

    PubMed

    Kingdom, Frederick A A; Baldwin, Alex S; Schmidtmann, Gunnar

    2015-01-01

    Many studies have investigated how multiple stimuli combine to reach threshold. There are broadly speaking two ways this can occur: additive summation (AS) where inputs from the different stimuli add together in a single mechanism, or probability summation (PS) where different stimuli are detected independently by separate mechanisms. PS is traditionally modeled under high threshold theory (HTT); however, tests have shown that HTT is incorrect and that signal detection theory (SDT) is the better framework for modeling summation. Modeling the equivalent of PS under SDT is, however, relatively complicated, leading many investigators to use Monte Carlo simulations for the predictions. We derive formulas that employ numerical integration to predict the proportion correct for detecting multiple stimuli assuming PS under SDT, for the situations in which stimuli are either equal or unequal in strength. Both formulas are general purpose, calculating performance for forced-choice tasks with M alternatives, n stimuli, in Q monitored mechanisms, each subject to a non-linear transducer with exponent τ. We show how the probability (and additive) summation formulas can be used to simulate psychometric functions, which when fitted with Weibull functions make signature predictions for how thresholds and psychometric function slopes vary as a function of τ, n, and Q. We also show how one can fit the formulas directly to real psychometric functions using data from a binocular summation experiment, and show how one can obtain estimates of τ and test whether binocular summation conforms more to PS or AS. The methods described here can be readily applied using software functions newly added to the Palamedes toolbox.

  7. False assumptions.

    PubMed

    Swaminathan, M

    1997-01-01

    Indian women do not have to be told the benefits of breast feeding or "rescued from the clutches of wicked multinational companies" by international agencies. There is no proof that breast feeding has declined in India; in fact, a 1987 survey revealed that 98% of Indian women breast feed. Efforts to promote breast feeding among the middle classes rely on such initiatives as the "baby friendly" hospital where breast feeding is promoted immediately after birth. This ignores the 76% of Indian women who give birth at home. Blaming this unproved decline in breast feeding on multinational companies distracts attention from more far-reaching and intractable effects of social change. While the Infant Milk Substitutes Act is helpful, it also deflects attention from more pressing issues. Another false assumption is that Indian women are abandoning breast feeding to comply with the demands of employment, but research indicates that most women give up employment for breast feeding, despite the economic cost to their families. Women also seek work in the informal sector to secure the flexibility to meet their child care responsibilities. Instead of being concerned about "teaching" women what they already know about the benefits of breast feeding, efforts should be made to remove the constraints women face as a result of their multiple roles and to empower them with the support of families, governmental policies and legislation, employers, health professionals, and the media.

  8. Simplified Vicarious Radiometric Calibration

    NASA Technical Reports Server (NTRS)

    Stanley, Thomas; Ryan, Robert; Holekamp, Kara; Pagnutti, Mary

    2010-01-01

    ground target areas having different reflectance values. The target areas can be natural or artificial and must be large enough to minimize adjacent-pixel contamination effects. The radiative coupling between the atmosphere and the terrain needs to be approximately the same for the two targets. This condition can be met for relatively uniform backgrounds when the distance between the targets is within a few hundred meters. For each target area, the radiance leaving the ground in the direction of the satellite is measured with a radiometrically calibrated spectroradiometer. Using the radiance measurements from the two targets, atmospheric adjacency and atmospheric scattering effects can be subtracted, thereby eliminating many assumptions about the atmosphere and the radiative interaction between the atmosphere and the terrain. In addition, the radiometrically calibrated spectroradiometer can be used with a known reflectance target to estimate atmospheric transmission and diffuse- to-global ratios without the need for ancillary sun photometers. Several comparisons between the simplified method and traditional techniques were found to agree within a few percent. Hence, the simplified method reduces the overall complexity of performing vicarious calibrations and can serve as a method for validating traditional radiative transfer models

  9. Use of a simplified generalized standard additions method for the analysis of cement, gypsum and basic slag by slurry nebulization ICP-OES.

    PubMed

    Marjanovic, Ljiljana; McCrindle, Robert I; Botha, Barend M; Potgieter, Herman J

    2004-05-01

    The simplified generalized standard additions method (GSAM) was investigated as an alternative method for the ICP-OES analysis of solid materials, introduced into the plasma in the form of slurries. The method is an expansion of the conventional standard additions method. It is based on the principle of varying both the sample mass and the amount of standard solution added. The relationship between the sample mass, standard solution added and signal intensity is assumed to be linear. Concentration of the analyte can be found either geometrically from the slope of the two-dimensional response plane in a three-dimensional space or mathematically from the ratio of the parameters estimated by multiple linear regression. The analysis of a series of certified reference materials (CRMs) (cement CRM-BCS No 353, gypsum CRM-Gyp A and basic slag CRM No 382/I) introduced into the plasma in the form of slurry is described. The slurries contained glycerol and hydrochloric acid and were placed in an ultrasonic bath to ensure good dispersion. "Table curve 3D" software was used to fit the data. Results obtained showed that the method could be successfully applied to the analysis of cement, gypsum and slag samples, without the need to dissolve them. In this way, we could avoid the use of hazardous chemicals (concentrated acids), incomplete dissolution and loss of some volatiles. The application of the simplified GSAM for the analysis did not require a CRM with similar chemical and mineralogical properties for the calibration of the instrument.

  10. Simplifying Podcasting

    ERIC Educational Resources Information Center

    Panday, Prachi Parashar

    2009-01-01

    The article intends to simplify the different aspects of podcasting. The article covers types of podcasts; the pedagogical benefits of podcasting; the connection between theory and podcasting; answers to questions, queries, and apprehensions. Before trying out a new tool, it is important to understand why we do things the way we do. A crucial part…

  11. The assumptions of computing

    SciTech Connect

    Huggins, J.K.

    1994-12-31

    The use of computers, like any technological activity, is not content-neutral. Users of computers constantly interact with assumptions regarding worthwhile activity which are embedded in any computing system. Directly questioning these assumptions in the context of computing allows us to develop an understanding of responsible computing.

  12. Additives

    NASA Technical Reports Server (NTRS)

    Smalheer, C. V.

    1973-01-01

    The chemistry of lubricant additives is discussed to show what the additives are chemically and what functions they perform in the lubrication of various kinds of equipment. Current theories regarding the mode of action of lubricant additives are presented. The additive groups discussed include the following: (1) detergents and dispersants, (2) corrosion inhibitors, (3) antioxidants, (4) viscosity index improvers, (5) pour point depressants, and (6) antifouling agents.

  13. Testing Our Fundamental Assumptions

    NASA Astrophysics Data System (ADS)

    Kohler, Susanna

    2016-06-01

    fundamental assumptions.A recent focus set in the Astrophysical Journal Letters, titled Focus on Exploring Fundamental Physics with Extragalactic Transients, consists of multiple published studies doing just that.Testing General RelativitySeveral of the articles focus on the 4th point above. By assuming that the delay in photon arrival times is only due to the gravitational potential of the Milky Way, these studies set constraints on the deviation of our galaxys gravitational potential from what GR would predict. The study by He Gao et al. uses the different photon arrival times from gamma-ray bursts to set constraints at eVGeV energies, and the study by Jun-Jie Wei et al. complements this by setting constraints at keV-TeV energies using photons from high-energy blazar emission.Photons or neutrinos from different extragalactic transients each set different upper limits on delta gamma, the post-Newtonian parameter, vs. particle energy or frequency. This is a test of Einsteins equivalence principle: if the principle is correct, delta gamma would be exactly zero, meaning that photons of different energies move at the same velocity through a vacuum. [Tingay Kaplan 2016]S.J. Tingay D.L. Kaplan make the case that measuring the time delay of photons from fast radio bursts (FRBs; transient radio pulses that last only a few milliseconds) will provide even tighter constraints if we are able to accurately determine distances to these FRBs.And Adi Musser argues that the large-scale structure of the universe plays an even greater role than the Milky Way gravitational potential, allowing for even stricter testing of Einsteins equivalence principle.The ever-narrower constraints from these studies all support GR as a correct set of rules through which to interpret our universe.Other Tests of Fundamental PhysicsIn addition to the above tests, Xue-Feng Wu et al. show that FRBs can be used to provide severe constraints on the rest mass of the photon, and S. Croft et al. even touches on what we

  14. Contextuality under weak assumptions

    NASA Astrophysics Data System (ADS)

    Simmons, Andrew W.; Wallman, Joel J.; Pashayan, Hakop; Bartlett, Stephen D.; Rudolph, Terry

    2017-03-01

    The presence of contextuality in quantum theory was first highlighted by Bell, Kochen and Specker, who discovered that for quantum systems of three or more dimensions, measurements could not be viewed as deterministically revealing pre-existing properties of the system. More precisely, no model can assign deterministic outcomes to the projectors of a quantum measurement in a way that depends only on the projector and not the context (the full set of projectors) in which it appeared, despite the fact that the Born rule probabilities associated with projectors are independent of the context. A more general, operational definition of contextuality introduced by Spekkens, which we will term ‘probabilistic contextuality’, drops the assumption of determinism and allows for operations other than measurements to be considered contextual. Even two-dimensional quantum mechanics can be shown to be contextual under this generalised notion. Probabilistic noncontextuality represents the postulate that elements of an operational theory that cannot be distinguished from each other based on the statistics of arbitrarily many repeated experiments (they give rise to the same operational probabilities) are ontologically identical. In this paper, we introduce a framework that enables us to distinguish between different noncontextuality assumptions in terms of the relationships between the ontological representations of objects in the theory given a certain relation between their operational representations. This framework can be used to motivate and define a ‘possibilistic’ analogue, encapsulating the idea that elements of an operational theory that cannot be unambiguously distinguished operationally can also not be unambiguously distinguished ontologically. We then prove that possibilistic noncontextuality is equivalent to an alternative notion of noncontextuality proposed by Hardy. Finally, we demonstrate that these weaker noncontextuality assumptions are sufficient to prove

  15. A simplified determination of total concentrations of Ca, Fe, Mg and Mn in addition to their bioaccessible fraction in popular instant coffee brews.

    PubMed

    Stelmach, Ewelina; Szymczycha-Madeja, Anna; Pohl, Pawel

    2016-04-15

    A direct analysis of instant coffee brews with HR-CS-FAAS spectrometry to determine the total Ca, Fe, Mg and Mn content has been developed and validated. The proposed method is simple and fast; it delivers good analytical performance; its accuracy being within -3% to 3%, its precision--2-3% and detection limits--0.03, 0.04, 0.004 and 0.01 mg l(-1) for Ca, Fe, Mg and Mn, respectively. In addition, Ca, Fe, Mg and Mn bioaccessibility in instant coffee brews was measured by means of the in vitro gastrointestinal digestion with the use of simulated gastric and intestinal juice solutions. Absorption of metals in intestinal villi was simulated by means of ultrafiltration over semi-permeable membrane with a molecular weight cut-off of 5 kDa. Ca, Fe, Mg and Mn concentrations in permeates of instant coffee gastrointestinal incubates were measured with HR-CS-FAA spectrometry.

  16. Simplified modal method for slanted grating

    NASA Astrophysics Data System (ADS)

    Li, Shubin; Zhou, Changhe; Jia, Wei

    2017-01-01

    We report the simplified modal method for the slanted grating based on the accurate dispersion equation. The vividly physical insight is presented to interpret the diffraction process for slanted grating. We also present that the simplified modal method with the two-lowest mode condition is effective for a large slanted angle up to 26°. By examining the eignefunction, the mode index, and the two-lowest mode condition, we provide new evidences to verify the assumption that a slanted grating with subwavelength period can be analyzed as an equivalent rectangular grating using the simplified modal method, which is right and convenient to use for a small slanted angle up to 20°. Numerical simulations of the simplified modal method are coincident with rigorous coupled wave analysis for small slanted angle gratings. Thus the simplified modal method can be used for small slanted angle grating since the equivalence of slanted grating and rectangular grating is verified due to its vivid physical analysis.

  17. Impact of unseen assumptions on communication of atmospheric carbon mitigation options

    NASA Astrophysics Data System (ADS)

    Elliot, T. R.; Celia, M. A.; Court, B.

    2010-12-01

    With the rapid access and dissemination of information made available through online and digital pathways, there is need for a concurrent openness and transparency in communication of scientific investigation. Even with open communication it is essential that the scientific community continue to provide impartial result-driven information. An unknown factor in climate literacy is the influence of an impartial presentation of scientific investigation that has utilized biased base-assumptions. A formal publication appendix, and additional digital material, provides active investigators a suitable framework and ancillary material to make informed statements weighted by assumptions made in a study. However, informal media and rapid communiqués rarely make such investigatory attempts, often citing headline or key phrasing within a written work. This presentation is focused on Geologic Carbon Sequestration (GCS) as a proxy for the wider field of climate science communication, wherein we primarily investigate recent publications in GCS literature that produce scenario outcomes using apparently biased pro- or con- assumptions. A general review of scenario economics, capture process efficacy and specific examination of sequestration site assumptions and processes, reveals an apparent misrepresentation of what we consider to be a base-case GCS system. The authors demonstrate the influence of the apparent bias in primary assumptions on results from commonly referenced subsurface hydrology models. By use of moderate semi-analytical model simplification and Monte Carlo analysis of outcomes, we can establish the likely reality of any GCS scenario within a pragmatic middle ground. Secondarily, we review the development of publically available web-based computational tools and recent workshops where we presented interactive educational opportunities for public and institutional participants, with the goal of base-assumption awareness playing a central role. Through a series of

  18. Simplified computer program for the analysis of phase change in liquid face seals

    NASA Technical Reports Server (NTRS)

    Birchak, M.; Hughes, W. F.

    1977-01-01

    A simplified computer program is presented which allows for the prediction of boiling (phase change) in liquid face seals. The program determines if and when boiling occurs and then calculates the location of the boiling interface, pressure and temperature profiles, and load. The main assumption which allows for a simplified analysis is the assumption of an isothermal gas phase.

  19. Simplified propagation of standard uncertainties

    SciTech Connect

    Shull, A.H.

    1997-06-09

    An essential part of any measurement control program is adequate knowledge of the uncertainties of the measurement system standards. Only with an estimate of the standards` uncertainties can one determine if the standard is adequate for its intended use or can one calculate the total uncertainty of the measurement process. Purchased standards usually have estimates of uncertainty on their certificates. However, when standards are prepared and characterized by a laboratory, variance propagation is required to estimate the uncertainty of the standard. Traditional variance propagation typically involves tedious use of partial derivatives, unfriendly software and the availability of statistical expertise. As a result, the uncertainty of prepared standards is often not determined or determined incorrectly. For situations meeting stated assumptions, easier shortcut methods of estimation are now available which eliminate the need for partial derivatives and require only a spreadsheet or calculator. A system of simplifying the calculations by dividing into subgroups of absolute and relative uncertainties is utilized. These methods also incorporate the International Standards Organization (ISO) concepts for combining systematic and random uncertainties as published in their Guide to the Expression of Measurement Uncertainty. Details of the simplified methods and examples of their use are included in the paper.

  20. Disastrous assumptions about community disasters

    SciTech Connect

    Dynes, R.R.

    1995-12-31

    Planning for local community disasters is compounded with erroneous assumptions. Six problematic models are identified: agent facts, big accident, end of the world, media, command and control, administrative. Problematic assumptions in each of them are identified. A more adequate model centered on problem solving is identified. That there is a discrepancy between disaster planning efforts and the actual response experience seems rather universal. That discrepancy is symbolized by the graffiti which predictably surfaces on many walls in post disaster locations -- ``First the earthquake, then the disaster.`` That contradiction is seldom reduced as a result of post disaster critiques, since the most usual conclusion is that the plan was adequate but the ``people`` did not follow it. Another explanation will be provided here. A more plausible explanation for failure is that most planning efforts adopt a number of erroneous assumptions which affect the outcome. Those assumptions are infrequently changed or modified by experience.

  1. Spatial Encryption under Simpler Assumption

    NASA Astrophysics Data System (ADS)

    Zhou, Muxin; Cao, Zhenfu

    Spatial encryption was first proposed by Boneh and Hamburg. They showed that many useful encryption systems can be derived from it. In this paper, we describe two variants of spatial encryption. First we present a scheme that can be proved to be secure under the decisional bilinear Diffie-Hellman assumption, which is much simpler than the BDHE assumption used by Boneh and Hamburg. However, as a compromise, our ciphertext size and private key size are larger. We also discuss some techniques to shrink the private key of this scheme in a real application. Finally, we provide a hybrid construction which allows an optimal tradeoff between efficiency and security.

  2. Sampling Assumptions in Inductive Generalization

    ERIC Educational Resources Information Center

    Navarro, Daniel J.; Dry, Matthew J.; Lee, Michael D.

    2012-01-01

    Inductive generalization, where people go beyond the data provided, is a basic cognitive capability, and it underpins theoretical accounts of learning, categorization, and decision making. To complete the inductive leap needed for generalization, people must make a key "sampling" assumption about how the available data were generated.…

  3. Simplified DM models with the full SM gauge symmetry: the case of t-channel colored scalar mediators

    NASA Astrophysics Data System (ADS)

    Ko, P.; Natale, Alexander; Park, Myeonghun; Yokoya, Hiroshi

    2017-01-01

    The general strategy for dark matter (DM) searches at colliders currently relies on simplified models. In this paper, we propose a new t-channel UV-complete simplified model that improves the existing simplified DM models in two important respects: (i) we impose the full SM gauge symmetry including the fact that the left-handed and the right-handed fermions have two independent mediators with two independent couplings, and (ii) we include the renormalization group evolution when we derive the effective Lagrangian for DM-nucleon scattering from the underlying UV complete models by integrating out the t-channel mediators. The first improvement will introduce a few more new parameters compared with the existing simplified DM models. In this study we look at the effect this broader set of free parameters has on direct detection and the mono- X + MET ( X=jet, W,Z) signatures at 13TeV LHC while maintaining gauge invariance of the simplified model under the full SM gauge group. We find that the direct detection constraints require DM masses less than 10 GeV in order to produce phenomenologically interesting collider signatures. Additionally, for a fixed mono-W cross section it is possible to see very large differences in the mono-jet cross section when the usual simplified model assumptions are loosened and isospin violation between RH and LH DM-SM quark couplings are allowed.

  4. Learning Assumptions for Compositional Verification

    NASA Technical Reports Server (NTRS)

    Cobleigh, Jamieson M.; Giannakopoulou, Dimitra; Pasareanu, Corina; Clancy, Daniel (Technical Monitor)

    2002-01-01

    Compositional verification is a promising approach to addressing the state explosion problem associated with model checking. One compositional technique advocates proving properties of a system by checking properties of its components in an assume-guarantee style. However, the application of this technique is difficult because it involves non-trivial human input. This paper presents a novel framework for performing assume-guarantee reasoning in an incremental and fully automated fashion. To check a component against a property, our approach generates assumptions that the environment needs to satisfy for the property to hold. These assumptions are then discharged on the rest of the system. Assumptions are computed by a learning algorithm. They are initially approximate, but become gradually more precise by means of counterexamples obtained by model checking the component and its environment, alternately. This iterative process may at any stage conclude that the property is either true or false in the system. We have implemented our approach in the LTSA tool and applied it to the analysis of a NASA system.

  5. Investigations in a Simplified Bracketed Grid Approach to Metrical Structure

    ERIC Educational Resources Information Center

    Liu, Patrick Pei

    2010-01-01

    In this dissertation, I examine the fundamental mechanisms and assumptions of the Simplified Bracketed Grid Theory (Idsardi 1992) in two ways: first, by comparing it with Parametric Metrical Theory (Hayes 1995), and second, by implementing it in the analysis of several case studies in stress assignment and syllabification. Throughout these…

  6. Faulty assumptions for repository requirements

    SciTech Connect

    Sutcliffe, W G

    1999-06-03

    Long term performance requirements for a geologic repository for spent nuclear fuel and high-level waste are based on assumptions concerning water use and subsequent deaths from cancer due to ingesting water contaminated with radio isotopes ten thousand years in the future. This paper argues that the assumptions underlying these requirements are faulty for a number of reasons. First, in light of the inevitable technological progress, including efficient desalination of water, over the next ten thousand years, it is inconceivable that a future society would drill for water near a repository. Second, even today we would not use water without testing its purity. Third, today many types of cancer are curable, and with the rapid progress in medical technology in general, and the prevention and treatment of cancer in particular, it is improbable that cancer caused by ingesting contaminated water will be a sign&ant killer in the far future. This paper reviews the performance requirements for geological repositories and comments on the difficulties in proving compliance in the face of inherent uncertainties. The already tiny long-term risk posed by a geologic repository is presented and contrasted with contemporary every day risks. A number of examples of technological progress, including cancer treatments, are advanced. The real and significant costs resulting from the overly conservative requirements are then assessed. Examples are given of how money (and political capital) could be put to much better use to save lives today and in the future. It is concluded that although a repository represents essentially no long-term risk, monitored retrievable dry storage (above or below ground) is the current best alternative for spent fuel and high-level nuclear waste.

  7. Scenarios Based on Shared Socioeconomic Pathway Assumptions

    NASA Astrophysics Data System (ADS)

    Edmonds, J.

    2013-12-01

    scenario with at least 8.5 Wm-2. To address this problem each SSP scenario can be treated as a reference scenario, to which emissions mitigation policies can be applied to create a set of RCP replications. These RCP replications have the underlying SSP socio-economic assumptions in addition to policy assumptions and radiative forcing levels consistent with the CMIP5 products. We report quantitative results of initial experiments from the five participating groups.

  8. 39 Questionable Assumptions in Modern Physics

    NASA Astrophysics Data System (ADS)

    Volk, Greg

    2009-03-01

    The growing body of anomalies in new energy, low energy nuclear reactions, astrophysics, atomic physics, and entanglement, combined with the failure of the Standard Model and string theory to predict many of the most basic fundamental phenomena, all point to a need for major new paradigms. Not Band-Aids, but revolutionary new ways of conceptualizing physics, in the spirit of Thomas Kuhn's The Structure of Scientific Revolutions. This paper identifies a number of long-held, but unproven assumptions currently being challenged by an increasing number of alternative scientists. Two common themes, both with venerable histories, keep recurring in the many alternative theories being proposed: (1) Mach's Principle, and (2) toroidal, vortex particles. Matter-based Mach's Principle differs from both space-based universal frames and observer-based Einsteinian relativity. Toroidal particles, in addition to explaining electron spin and the fundamental constants, satisfy the basic requirement of Gauss's misunderstood B Law, that motion itself circulates. Though a comprehensive theory is beyond the scope of this paper, it will suggest alternatives to the long list of assumptions in context.

  9. Simplified mechanistic models of gene regulation for analysis and design

    PubMed Central

    Hancock, Edward J.; Stan, Guy-Bart; Arpino, James A. J.; Papachristodoulou, Antonis

    2015-01-01

    Simplified mechanistic models of gene regulation are fundamental to systems biology and essential for synthetic biology. However, conventional simplified models typically have outputs that are not directly measurable and are based on assumptions that do not often hold under experimental conditions. To resolve these issues, we propose a ‘model reduction’ methodology and simplified kinetic models of total mRNA and total protein concentration, which link measurements, models and biochemical mechanisms. The proposed approach is based on assumptions that hold generally and include typical cases in systems and synthetic biology where conventional models do not hold. We use novel assumptions regarding the ‘speed of reactions’, which are required for the methodology to be consistent with experimental data. We also apply the methodology to propose simplified models of gene regulation in the presence of multiple protein binding sites, providing both biological insights and an illustration of the generality of the methodology. Lastly, we show that modelling total protein concentration allows us to address key questions on gene regulation, such as efficiency, burden, competition and modularity. PMID:26063825

  10. Modelling Surface Emissivity at Microwave Frequencies: Impact of the Surface Assumptions

    NASA Astrophysics Data System (ADS)

    Hermozo, Laura; Eymard, Laurence; Karbou, Fatima

    2015-12-01

    In this study, a new method is proposed to monitor and characterize Arctic sea ice. As inputs, passive microwave observations at window frequencies ranging from 23.8 to 89 GHz from the Advanced Microwave Sounding Units (AMSU) -A and -B are used. Unlike several studies which use a simplified flat surface assumption to retrieve sea ice surface emissivity, the relative roughness of the surface is taken into account to retrieve a rough ‘Lambertian’ (additionally to a flat ‘Specular’) surface emissivity over Arctic. The effect of the two surface assumptions is analyzed at different window frequencies, using a Specular/Lambertian emissivity ratio over a one-year database of near-nadir observations. A monthly ice/no ice delimitation is obtained using emissivity ratio at low frequencies and is found to be in very good agreement with other available sea ice products. The potential of emissivity ratio at 50.3 GHz over sea ice, combined with the signal at other window frequencies, is used towards a sea ice classification over Arctic.

  11. A discussion of assumptions and solution approaches of infiltration into a cracked soil

    Technology Transfer Automated Retrieval System (TEKTRAN)

    A model for predicting rain infiltration into a swelling/shrinking/cracking soil was proposed (Römkens, M.J.M., and S. N. Prasad., 2006, Agricultural Water Management. 86:196-205). Several simplifying assumptions were made. The model consists of a two-component process of Darcian matrix flow and Hor...

  12. On the underlying assumptions of threshold Boolean networks as a model for genetic regulatory network behavior

    PubMed Central

    Tran, Van; McCall, Matthew N.; McMurray, Helene R.; Almudevar, Anthony

    2013-01-01

    Boolean networks (BoN) are relatively simple and interpretable models of gene regulatory networks. Specifying these models with fewer parameters while retaining their ability to describe complex regulatory relationships is an ongoing methodological challenge. Additionally, extending these models to incorporate variable gene decay rates, asynchronous gene response, and synergistic regulation while maintaining their Markovian nature increases the applicability of these models to genetic regulatory networks (GRN). We explore a previously-proposed class of BoNs characterized by linear threshold functions, which we refer to as threshold Boolean networks (TBN). Compared to traditional BoNs with unconstrained transition functions, these models require far fewer parameters and offer a more direct interpretation. However, the functional form of a TBN does result in a reduction in the regulatory relationships which can be modeled. We show that TBNs can be readily extended to permit self-degradation, with explicitly modeled degradation rates. We note that the introduction of variable degradation compromises the Markovian property fundamental to BoN models but show that a simple state augmentation procedure restores their Markovian nature. Next, we study the effect of assumptions regarding self-degradation on the set of possible steady states. Our findings are captured in two theorems relating self-degradation and regulatory feedback to the steady state behavior of a TBN. Finally, we explore assumptions of synchronous gene response and asynergistic regulation and show that TBNs can be easily extended to relax these assumptions. Applying our methods to the budding yeast cell-cycle network revealed that although the network is complex, its steady state is simplified by the presence of self-degradation and lack of purely positive regulatory cycles. PMID:24376454

  13. An Accurate, Simplified Model Intrabeam Scattering

    SciTech Connect

    Bane, Karl LF

    2002-05-23

    Beginning with the general Bjorken-Mtingwa solution for intrabeam scattering (IBS) we derive an accurate, greatly simplified model of IBS, valid for high energy beams in normal storage ring lattices. In addition, we show that, under the same conditions, a modified version of Piwinski's IBS formulation (where {eta}{sub x,y}{sup 2}/{beta}{sub x,y} has been replaced by {Eta}{sub x,y}) asymptotically approaches the result of Bjorken-Mtingwa.

  14. 47 CFR 214.3 - Assumptions.

    Code of Federal Regulations, 2014 CFR

    2014-10-01

    ... 47 Telecommunication 5 2014-10-01 2014-10-01 false Assumptions. 214.3 Section 214.3 Telecommunication OFFICE OF SCIENCE AND TECHNOLOGY POLICY AND NATIONAL SECURITY COUNCIL PROCEDURES FOR THE USE AND COORDINATION OF THE RADIO SPECTRUM DURING A WARTIME EMERGENCY § 214.3 Assumptions. When the provisions of...

  15. 47 CFR 214.3 - Assumptions.

    Code of Federal Regulations, 2013 CFR

    2013-10-01

    ... 47 Telecommunication 5 2013-10-01 2013-10-01 false Assumptions. 214.3 Section 214.3 Telecommunication OFFICE OF SCIENCE AND TECHNOLOGY POLICY AND NATIONAL SECURITY COUNCIL PROCEDURES FOR THE USE AND COORDINATION OF THE RADIO SPECTRUM DURING A WARTIME EMERGENCY § 214.3 Assumptions. When the provisions of...

  16. 47 CFR 214.3 - Assumptions.

    Code of Federal Regulations, 2010 CFR

    2010-10-01

    ... 47 Telecommunication 5 2010-10-01 2010-10-01 false Assumptions. 214.3 Section 214.3 Telecommunication OFFICE OF SCIENCE AND TECHNOLOGY POLICY AND NATIONAL SECURITY COUNCIL PROCEDURES FOR THE USE AND COORDINATION OF THE RADIO SPECTRUM DURING A WARTIME EMERGENCY § 214.3 Assumptions. When the provisions of...

  17. 47 CFR 214.3 - Assumptions.

    Code of Federal Regulations, 2012 CFR

    2012-10-01

    ... 47 Telecommunication 5 2012-10-01 2012-10-01 false Assumptions. 214.3 Section 214.3 Telecommunication OFFICE OF SCIENCE AND TECHNOLOGY POLICY AND NATIONAL SECURITY COUNCIL PROCEDURES FOR THE USE AND COORDINATION OF THE RADIO SPECTRUM DURING A WARTIME EMERGENCY § 214.3 Assumptions. When the provisions of...

  18. 47 CFR 214.3 - Assumptions.

    Code of Federal Regulations, 2011 CFR

    2011-10-01

    ... 47 Telecommunication 5 2011-10-01 2011-10-01 false Assumptions. 214.3 Section 214.3 Telecommunication OFFICE OF SCIENCE AND TECHNOLOGY POLICY AND NATIONAL SECURITY COUNCIL PROCEDURES FOR THE USE AND COORDINATION OF THE RADIO SPECTRUM DURING A WARTIME EMERGENCY § 214.3 Assumptions. When the provisions of...

  19. The Assumptive Worlds of Fledgling Administrators.

    ERIC Educational Resources Information Center

    Marshall, Catherine; Mitchell, Barbara A.

    1991-01-01

    Studies school-site administrators' understanding about ways of gaining/maintaining power, control, and predictability. Multisite study data concerning assistant principals identify rules of the game for four micropolitical (site-level assumptive world) domains. Assumptive worlds create avoidance of value conflicts and risky change, group-think…

  20. Assumptions to the Annual Energy Outlook

    EIA Publications

    2015-01-01

    This report presents the major assumptions of the National Energy Modeling System (NEMS) used to generate the projections in the Annual Energy Outlook, including general features of the model structure, assumptions concerning energy markets, and the key input data and parameters that are the most significant in formulating the model results.

  1. 5 CFR 841.405 - Economic assumptions.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... 5 Administrative Personnel 2 2010-01-01 2010-01-01 false Economic assumptions. 841.405 Section 841... (CONTINUED) FEDERAL EMPLOYEES RETIREMENT SYSTEM-GENERAL ADMINISTRATION Government Costs § 841.405 Economic assumptions. The determinations of the normal cost percentage will be based on the economic...

  2. The Assumptive Worlds of Fledgling Administrators.

    ERIC Educational Resources Information Center

    Marshall, Catherine; Mitchell, Barbara

    School administrators' uses of subjective understandings and common language to gain and maintain power and predictability in their environments are described. Micropolitical theory, with a focus on language, is utilized to understand administrators' knowledge of the assumptive worlds of their subculture, and how these assumptive worlds constrain…

  3. 5 CFR 841.405 - Economic assumptions.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... 5 Administrative Personnel 2 2011-01-01 2011-01-01 false Economic assumptions. 841.405 Section 841... (CONTINUED) FEDERAL EMPLOYEES RETIREMENT SYSTEM-GENERAL ADMINISTRATION Government Costs § 841.405 Economic assumptions. The determinations of the normal cost percentage will be based on the economic...

  4. 5 CFR 841.405 - Economic assumptions.

    Code of Federal Regulations, 2013 CFR

    2013-01-01

    ... 5 Administrative Personnel 2 2013-01-01 2013-01-01 false Economic assumptions. 841.405 Section 841... (CONTINUED) FEDERAL EMPLOYEES RETIREMENT SYSTEM-GENERAL ADMINISTRATION Government Costs § 841.405 Economic assumptions. The determinations of the normal cost percentage will be based on the economic...

  5. 5 CFR 841.405 - Economic assumptions.

    Code of Federal Regulations, 2012 CFR

    2012-01-01

    ... 5 Administrative Personnel 2 2012-01-01 2012-01-01 false Economic assumptions. 841.405 Section 841... (CONTINUED) FEDERAL EMPLOYEES RETIREMENT SYSTEM-GENERAL ADMINISTRATION Government Costs § 841.405 Economic assumptions. The determinations of the normal cost percentage will be based on the economic...

  6. 5 CFR 841.405 - Economic assumptions.

    Code of Federal Regulations, 2014 CFR

    2014-01-01

    ... 5 Administrative Personnel 2 2014-01-01 2014-01-01 false Economic assumptions. 841.405 Section 841... (CONTINUED) FEDERAL EMPLOYEES RETIREMENT SYSTEM-GENERAL ADMINISTRATION Government Costs § 841.405 Economic assumptions. The determinations of the normal cost percentage will be based on the economic...

  7. 29 CFR 4044.53 - Mortality assumptions.

    Code of Federal Regulations, 2012 CFR

    2012-07-01

    ... 29 Labor 9 2012-07-01 2012-07-01 false Mortality assumptions. 4044.53 Section 4044.53 Labor Regulations Relating to Labor (Continued) PENSION BENEFIT GUARANTY CORPORATION PLAN TERMINATIONS ALLOCATION OF ASSETS IN SINGLE-EMPLOYER PLANS Valuation of Benefits and Assets Trusteed Plans 4044.53 Mortality assumptions. (a) General rule. Subject...

  8. 29 CFR 4044.53 - Mortality assumptions.

    Code of Federal Regulations, 2014 CFR

    2014-07-01

    ... 29 Labor 9 2014-07-01 2014-07-01 false Mortality assumptions. 4044.53 Section 4044.53 Labor Regulations Relating to Labor (Continued) PENSION BENEFIT GUARANTY CORPORATION PLAN TERMINATIONS ALLOCATION OF ASSETS IN SINGLE-EMPLOYER PLANS Valuation of Benefits and Assets Trusteed Plans § 4044.53 Mortality assumptions. (a) General rule. Subject...

  9. 29 CFR 4044.53 - Mortality assumptions.

    Code of Federal Regulations, 2011 CFR

    2011-07-01

    ... 29 Labor 9 2011-07-01 2011-07-01 false Mortality assumptions. 4044.53 Section 4044.53 Labor Regulations Relating to Labor (Continued) PENSION BENEFIT GUARANTY CORPORATION PLAN TERMINATIONS ALLOCATION OF ASSETS IN SINGLE-EMPLOYER PLANS Valuation of Benefits and Assets Trusteed Plans 4044.53 Mortality assumptions. (a) General rule. Subject...

  10. 29 CFR 4044.53 - Mortality assumptions.

    Code of Federal Regulations, 2013 CFR

    2013-07-01

    ... 29 Labor 9 2013-07-01 2013-07-01 false Mortality assumptions. 4044.53 Section 4044.53 Labor Regulations Relating to Labor (Continued) PENSION BENEFIT GUARANTY CORPORATION PLAN TERMINATIONS ALLOCATION OF ASSETS IN SINGLE-EMPLOYER PLANS Valuation of Benefits and Assets Trusteed Plans § 4044.53 Mortality assumptions. (a) General rule. Subject...

  11. Economic Analysis and Assumptions in Global Education.

    ERIC Educational Resources Information Center

    Miller, Steven L.

    Economic educators recognize the importance of a global perspective, at least in part because the international sector has become more important over the past few decades. The application of economic principles calls into question some assumptions that appear to be common among members of the global education movement. That these assumptions might…

  12. Persuasive Arguments Theory: A Test of Assumptions.

    ERIC Educational Resources Information Center

    Meyers, Renee A.

    1989-01-01

    Assesses Persuasive Arguments Theory (PAT) assumptions about argument and argument influence. Reviews and critiques PAT, advancing five hypotheses to test its assumptions. Finds little correspondence between cognitive and discussion arguments in number, content, or persuasiveness. Discusses implications for PAT and for the role of communication in…

  13. Teaching Critical Thinking by Examining Assumptions

    ERIC Educational Resources Information Center

    Yanchar, Stephen C.; Slife, Brent D.

    2004-01-01

    We describe how instructors can integrate the critical thinking skill of examining theoretical assumptions (e.g., determinism and materialism) and implications into psychology courses. In this instructional approach, students formulate questions that help them identify assumptions and implications, use those questions to identify and examine the…

  14. Debating with Untested Assumptions: The Need to Understand School Discipline.

    ERIC Educational Resources Information Center

    Lufler, Henry S., Jr.

    1979-01-01

    The use of assumptions and language which liken school rule-breakers to criminal defendants does not accurately reflect the way the discipline system functions. In addition, there is no understanding among teachers and administrators of the Supreme Court's decisions in the area of school discipline. (Author/RLV)

  15. Simple tests of the factorization assumption

    SciTech Connect

    Mannel, T. Deutsches Elektronen-Synchrotron, D-2000 Hamburg 52, Germany ); Roberts, W.; Ryzak, Z. )

    1991-07-01

    We suggest simple experimental tests to determine whether or not the factorization assumption is valid for the two-body, nonleptonic decays of heavy hadrons, when both daughter hadrons are also heavy.

  16. Examining Computational Assumptions For Godiva IV

    SciTech Connect

    Kirkland, Alexander Matthew; Jaegers, Peter James

    2016-08-11

    Over the course of summer 2016, the effects of several computational modeling assumptions with respect to the Godiva IV reactor were examined. The majority of these assumptions pertained to modeling errors existing in the control rods and burst rod. The Monte Carlo neutron transport code, MCNP, was used to investigate these modeling changes, primarily by comparing them to that of the original input deck specifications.

  17. Simplified Decoding of Convolutional Codes

    NASA Technical Reports Server (NTRS)

    Truong, T. K.; Reed, I. S.

    1986-01-01

    Some complicated intermediate steps shortened or eliminated. Decoding of convolutional error-correcting digital codes simplified by new errortrellis syndrome technique. In new technique, syndrome vector not computed. Instead, advantage taken of newly-derived mathematical identities simplify decision tree, folding it back on itself into form called "error trellis." This trellis graph of all path solutions of syndrome equations. Each path through trellis corresponds to specific set of decisions as to received digits. Existing decoding algorithms combined with new mathematical identities reduce number of combinations of errors considered and enable computation of correction vector directly from data and check bits as received.

  18. Shattering world assumptions: A prospective view of the impact of adverse events on world assumptions.

    PubMed

    Schuler, Eric R; Boals, Adriel

    2016-05-01

    Shattered Assumptions theory (Janoff-Bulman, 1992) posits that experiencing a traumatic event has the potential to diminish the degree of optimism in the assumptions of the world (assumptive world), which could lead to the development of posttraumatic stress disorder. Prior research assessed the assumptive world with a measure that was recently reported to have poor psychometric properties (Kaler et al., 2008). The current study had 3 aims: (a) to assess the psychometric properties of a recently developed measure of the assumptive world, (b) to retrospectively examine how prior adverse events affected the optimism of the assumptive world, and (c) to measure the impact of an intervening adverse event. An 8-week prospective design with a college sample (N = 882 at Time 1 and N = 511 at Time 2) was used to assess the study objectives. We split adverse events into those that were objectively or subjectively traumatic in nature. The new measure exhibited adequate psychometric properties. The report of a prior objective or subjective trauma at Time 1 was related to a less optimistic assumptive world. Furthermore, participants who experienced an intervening objectively traumatic event evidenced a decrease in optimistic views of the world compared with those who did not experience an intervening adverse event. We found support for Shattered Assumptions theory retrospectively and prospectively using a reliable measure of the assumptive world. We discuss future assessments of the measure of the assumptive world and clinical implications to help rebuild the assumptive world with current therapies. (PsycINFO Database Record

  19. The helical decomposition and the instability assumption

    NASA Technical Reports Server (NTRS)

    Waleffe, Fabian A.

    1993-01-01

    Direct numerical simulations show that the triadic transfer function T(k,p,q) peaks sharply when q (or p) is much smaller than k. The triadic transfer function T(k,p,q) gives the rate of energy input into wave number k from all interactions with modes of wave number p and q, where k, p, q form a triangle. This observation was thought to suggest that energy is cascaded downscale through non-local interactions with local transfer and that there was a strong connection between large and small scales. Both suggestions were in contradiction with the classical Kolmogorov picture of the energy cascade. The helical decomposition was found useful in distinguishing between kinematically independent interactions. That analysis has gone beyond the question of non-local interaction with local transfer. In particular, an assumption about the statistical direction of triadic energy transfer in any kinematically independent interaction was introduced (the instability assumption). That assumption is not necessary for the conclusions about non-local interactions with local transfer recalled above. In the case of turbulence under rapid rotation, the instability assumption leads to the prediction that energy is transferred in spectral space from the poles of the rotation axis toward the equator. The instability assumption is thought to be of general validity for any type of triad interactions (e.g. internal waves). The helical decomposition and the instability assumption offer detailed information about the homogeneous statistical dynamics of the Navier-Stokes equations. The objective was to explore the validity of the instability assumption and to study the contributions of the various types of helical interactions to the energy cascade and the subgrid-scale eddy-viscosity. This was done in the context of spectral closures of the Direct Interaction or Quasi-Normal type.

  20. SEA - A Simplified Employee Assessment

    SciTech Connect

    Busby, L

    2001-04-23

    This paper presents a proposal for modifying the current employee annual evaluation process in SCAD. It purports to simplify that process, primarily by breaking up the resultant document into a set of more or less independent components. It claims to reduce the overall time and effort required from each actor.

  1. Simplifying the Water Poverty Index

    ERIC Educational Resources Information Center

    Cho, Danny I.; Ogwang, Tomson; Opio, Christopher

    2010-01-01

    In this paper, principal components methodology is used to derive simplified and cost effective indexes of water poverty. Using a well known data set for 147 countries from which an earlier five-component water poverty index comprising of "Resources," "Access," "Capacity," "Use" and "Environment" was constructed, we find that a simplified…

  2. 75 FR 28223 - Simplified Proceedings

    Federal Register 2010, 2011, 2012, 2013, 2014

    2010-05-20

    ...''). Trials are held before the Commission's Administrative Law Judges and appellate review is provided by a... proposing a rule to simplify the procedures for handling certain civil penalty proceedings. DATES: Written... civil penalty for the alleged violation described in the citation or order. The procedures for this...

  3. Simplified High-Power Inverter

    NASA Technical Reports Server (NTRS)

    Edwards, D. B.; Rippel, W. E.

    1984-01-01

    Solid-state inverter simplified by use of single gate-turnoff device (GTO) to commutate multiple silicon controlled rectifiers (SCR's). By eliminating conventional commutation circuitry, GTO reduces cost, size and weight. GTO commutation applicable to inverters of greater than 1-kilowatt capacity. Applications include emergency power, load leveling, drives for traction and stationary polyphase motors, and photovoltaic-power conditioning.

  4. A Simplified Model of Intrabeam Scattering

    SciTech Connect

    Bane, Karl LF

    2002-06-03

    Beginning with the general Bjorken-Mtingwa solution, we derive a simplified model of intrabeam scattering (IBS), one valid for high energy beams in normal storage rings; our result is similar, though more accurate than a model due to Raubenheimer. In addition, we show that a modified version of Piwinski's IBS formulation (where {eta}{sub x,y}{sup 2}/{beta}{sub x,y} has been replaced by {Eta}{sub x,y}) at high energies asymptotically approaches the same result.

  5. Life Support Baseline Values and Assumptions Document

    NASA Technical Reports Server (NTRS)

    Anderson, Molly S.; Ewert, Michael K.; Keener, John F.; Wagner, Sandra A.

    2015-01-01

    The Baseline Values and Assumptions Document (BVAD) provides analysts, modelers, and other life support researchers with a common set of values and assumptions which can be used as a baseline in their studies. This baseline, in turn, provides a common point of origin from which many studies in the community may depart, making research results easier to compare and providing researchers with reasonable values to assume for areas outside their experience. With the ability to accurately compare different technologies' performance for the same function, managers will be able to make better decisions regarding technology development.

  6. Evaluation and Reliability of a Simplified Method for Identification of Food-Borne Yeasts

    PubMed Central

    Rohm, Harald; Lechner, Frieda

    1990-01-01

    A simplified identification key described by Deak and Beuchat (T. Deak and L. R. Beuchat, J. Food Prot. 50:243-264, 1987) and the computer method of Barnett et al. (J. A. Barnett, R. W. Payne, and D. Yarrow, Yeast Identification Program, 1985) were used to identify 12 reference strains and 382 yeasts isolated from cultured milk products. Because the simplified key failed to account for species variability with regard to physiological, morphological, and sexual reproduction characteristics, poor agreement of the identification results was obtained. A reevaluation of the basic theoretical assumptions of the simplified key only confirmed the practical results and indicates that this identification method is unsatisfactory PMID:16348183

  7. Simplifier: a web tool to eliminate redundant NGS contigs

    PubMed Central

    Ramos, Rommel Thiago Jucá; Carneiro, Adriana Ribeiro; Azevedo, Vasco; Schneider, Maria Paula; Barh, Debmalya; Silva, Artur

    2012-01-01

    Modern genomic sequencing technologies produce a large amount of data with reduced cost per base; however, this data consists of short reads. This reduction in the size of the reads, compared to those obtained with previous methodologies, presents new challenges, including a need for efficient algorithms for the assembly of genomes from short reads and for resolving repetitions. Additionally after abinitio assembly, curation of the hundreds or thousands of contigs generated by assemblers demands considerable time and computational resources. We developed Simplifier, a stand-alone software that selectively eliminates redundant sequences from the collection of contigs generated by ab initio assembly of genomes. Application of Simplifier to data generated by assembly of the genome of Corynebacterium pseudotuberculosis strain 258 reduced the number of contigs generated by ab initio methods from 8,004 to 5,272, a reduction of 34.14%; in addition, N50 increased from 1 kb to 1.5 kb. Processing the contigs of Escherichia coli DH10B with Simplifier reduced the mate-paired library 17.47% and the fragment library 23.91%. Simplifier removed redundant sequences from datasets produced by assemblers, thereby reducing the effort required for finalization of genome assembly in tests with data from Prokaryotic organisms. Availability Simplifier is available at http://www.genoma.ufpa.br/rramos/softwares/simplifier.xhtmlIt requires Sun jdk 6 or higher. PMID:23275695

  8. Assessing Statistical Model Assumptions under Climate Change

    NASA Astrophysics Data System (ADS)

    Varotsos, Konstantinos V.; Giannakopoulos, Christos; Tombrou, Maria

    2016-04-01

    The majority of the studies assesses climate change impacts on air-quality using chemical transport models coupled to climate ones in an off-line mode, for various horizontal resolutions and different present and future time slices. A complementary approach is based on present-day empirical relations between air-pollutants and various meteorological variables which are then extrapolated to the future. However, the extrapolation relies on various assumptions such as that these relationships will retain their main characteristics in the future. In this study we focus on the ozone-temperature relationship. It is well known that among a number of meteorological variables, temperature is found to exhibit the highest correlation with ozone concentrations. This has led, in the past years, to the development and application of statistical models with which the potential impact of increasing future temperatures on various ozone statistical targets was examined. To examine whether the ozone-temperature relationship retains its main characteristics under warmer temperatures we analyze the relationship during the heatwaves events of 2003 and 2006 in Europe. More specifically, we use available gridded daily maximum temperatures (E-OBS) and hourly ozone observations from different non-urban stations (EMEP) within the areas that were impacted from the two heatwave events. In addition, we compare the temperature distributions of the two events with temperatures from two different future time periods 2021-2050 and 2071-2100 from a number of regional climate models developed under the framework of the Cordex initiative (http://www.cordex.org) with a horizontal resolution of 12 x 12km, based on different IPCC RCPs emissions scenarios. A statistical analysis is performed on the ozone-temperature relationship for each station and for the two aforementioned years which are then compared against the ozone-temperature relationships obtained from the rest of the available dataseries. The

  9. 10 CFR 436.14 - Methodological assumptions.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... Procedures for Life Cycle Cost Analyses § 436.14 Methodological assumptions. (a) Each Federal Agency shall... the Life Cycle Costing Manual for the Federal Energy Management Program (NIST 85-3273) and determined... of the fiscal year in the Annual Supplement to the Life Cycle Costing Manual for the Federal...

  10. Mexican-American Cultural Assumptions and Implications.

    ERIC Educational Resources Information Center

    Carranza, E. Lou

    The search for presuppositions of a people's thought is not new. Octavio Paz and Samuel Ramos have both attempted to describe the assumptions underlying the Mexican character. Paz described Mexicans as private, defensive, and stoic, characteristics taken to the extreme in the "pachuco." Ramos, on the other hand, described Mexicans as…

  11. Extracurricular Business Planning Competitions: Challenging the Assumptions

    ERIC Educational Resources Information Center

    Watson, Kayleigh; McGowan, Pauric; Smith, Paul

    2014-01-01

    Business planning competitions [BPCs] are a commonly offered yet under-examined extracurricular activity. Given the extent of sceptical comment about business planning, this paper offers what the authors believe is a much-needed critical discussion of the assumptions that underpin the provision of such competitions. In doing so it is suggested…

  12. 24 CFR 58.4 - Assumption authority.

    Code of Federal Regulations, 2010 CFR

    2010-04-01

    ..., decision-making, and action that would otherwise apply to HUD under NEPA and other provisions of law that... environmental review, decision-making and action for programs authorized by the Native American Housing... separate decision regarding assumption of responsibilities for each of these Acts and communicate...

  13. Publish unexpected results that conflict with assumptions

    Technology Transfer Automated Retrieval System (TEKTRAN)

    Some widely held scientific assumptions have been discredited, whereas others are just inappropriate for many applications. Sometimes, a widely-held analysis procedure takes on a life of its own, forgetting the original purpose of the analysis. The peer-reviewed system makes it difficult to get a pa...

  14. Causal Mediation Analysis: Warning! Assumptions Ahead

    ERIC Educational Resources Information Center

    Keele, Luke

    2015-01-01

    In policy evaluations, interest may focus on why a particular treatment works. One tool for understanding why treatments work is causal mediation analysis. In this essay, I focus on the assumptions needed to estimate mediation effects. I show that there is no "gold standard" method for the identification of causal mediation effects. In…

  15. Work and Family: Testing the Assumptions.

    DTIC Science & Technology

    1982-08-01

    person career" ( Papanek , 1973). And yet, much of our thinking is still premised on the assumption that all family patterns are traditional. Take, for...fufilling prophecy. Antioch Review, Summer 1948, 193-210. Papanek , 1i. Men, women, and work: Reflections on the two-person career. American Journal of

  16. 29 CFR 4281.14 - Mortality assumptions.

    Code of Federal Regulations, 2014 CFR

    2014-07-01

    ... 29 Labor 9 2014-07-01 2014-07-01 false Mortality assumptions. 4281.14 Section 4281.14 Labor Regulations Relating to Labor (Continued) PENSION BENEFIT GUARANTY CORPORATION INSOLVENCY, REORGANIZATION, TERMINATION, AND OTHER RULES APPLICABLE TO MULTIEMPLOYER PLANS DUTIES OF PLAN SPONSOR FOLLOWING MASS WITHDRAWAL Valuation of Plan Benefits and...

  17. 29 CFR 4281.14 - Mortality assumptions.

    Code of Federal Regulations, 2012 CFR

    2012-07-01

    ... 29 Labor 9 2012-07-01 2012-07-01 false Mortality assumptions. 4281.14 Section 4281.14 Labor Regulations Relating to Labor (Continued) PENSION BENEFIT GUARANTY CORPORATION INSOLVENCY, REORGANIZATION, TERMINATION, AND OTHER RULES APPLICABLE TO MULTIEMPLOYER PLANS DUTIES OF PLAN SPONSOR FOLLOWING MASS WITHDRAWAL Valuation of Plan Benefits and...

  18. 29 CFR 4281.14 - Mortality assumptions.

    Code of Federal Regulations, 2011 CFR

    2011-07-01

    ... 29 Labor 9 2011-07-01 2011-07-01 false Mortality assumptions. 4281.14 Section 4281.14 Labor Regulations Relating to Labor (Continued) PENSION BENEFIT GUARANTY CORPORATION INSOLVENCY, REORGANIZATION, TERMINATION, AND OTHER RULES APPLICABLE TO MULTIEMPLOYER PLANS DUTIES OF PLAN SPONSOR FOLLOWING MASS WITHDRAWAL Valuation of Plan Benefits and...

  19. 29 CFR 4281.14 - Mortality assumptions.

    Code of Federal Regulations, 2013 CFR

    2013-07-01

    ... 29 Labor 9 2013-07-01 2013-07-01 false Mortality assumptions. 4281.14 Section 4281.14 Labor Regulations Relating to Labor (Continued) PENSION BENEFIT GUARANTY CORPORATION INSOLVENCY, REORGANIZATION, TERMINATION, AND OTHER RULES APPLICABLE TO MULTIEMPLOYER PLANS DUTIES OF PLAN SPONSOR FOLLOWING MASS WITHDRAWAL Valuation of Plan Benefits and...

  20. Critically Challenging Some Assumptions in HRD

    ERIC Educational Resources Information Center

    O'Donnell, David; McGuire, David; Cross, Christine

    2006-01-01

    This paper sets out to critically challenge five interrelated assumptions prominent in the (human resource development) HRD literature. These relate to: the exploitation of labour in enhancing shareholder value; the view that employees are co-contributors to and co-recipients of HRD benefits; the distinction between HRD and human resource…

  1. Assumptive Worldviews and Problematic Reactions to Bereavement

    ERIC Educational Resources Information Center

    Currier, Joseph M.; Holland, Jason M.; Neimeyer, Robert A.

    2009-01-01

    Forty-two individuals who had lost an immediate family member in the prior 2 years and 42 nonbereaved matched controls completed the World Assumptions Scale (Janoff-Bulman, 1989) and the Symptom Checklist-10-Revised (Rosen et al., 2000). Results showed that bereaved individuals were significantly more distressed than nonbereaved matched controls,…

  2. Ability and Its Educational Correlates: Three Assumptions.

    ERIC Educational Resources Information Center

    Parks-Trusz, Sandra L.; Trusz, Andrew R.

    1981-01-01

    Models which describe assumptions concerning the distribution of intellectual capacity are: (1) the differential distribution model, encompassing theories that consider differences in intelligence genetic; (2) the normal distribution model; and (3) the evenly distributed model, focusing on the similarity of intellectual potentials within the human…

  3. 10 CFR 436.14 - Methodological assumptions.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... Procedures for Life Cycle Cost Analyses § 436.14 Methodological assumptions. (a) Each Federal Agency shall... the Life Cycle Costing Manual for the Federal Energy Management Program (NIST 85-3273) and determined... of the fiscal year in the Annual Supplement to the Life Cycle Costing Manual for the Federal...

  4. 10 CFR 436.14 - Methodological assumptions.

    Code of Federal Regulations, 2013 CFR

    2013-01-01

    ... Procedures for Life Cycle Cost Analyses § 436.14 Methodological assumptions. (a) Each Federal Agency shall... the Life Cycle Costing Manual for the Federal Energy Management Program (NIST 85-3273) and determined... of the fiscal year in the Annual Supplement to the Life Cycle Costing Manual for the Federal...

  5. Assumptions of Multiple Regression: Correcting Two Misconceptions

    ERIC Educational Resources Information Center

    Williams, Matt N.; Gomez Grajales, Carlos Alberto; Kurkiewicz, Dason

    2013-01-01

    In 2002, an article entitled "Four assumptions of multiple regression that researchers should always test" by Osborne and Waters was published in "PARE." This article has gone on to be viewed more than 275,000 times (as of August 2013), and it is one of the first results displayed in a Google search for "regression…

  6. Assumptions at the philosophical and programmatic levels in evaluation.

    PubMed

    Mertens, Donna M

    2016-12-01

    Stakeholders and evaluators hold a variety of levels of assumptions at the philosophical, methodological, and programmatic levels. The use of a transformative philosophical framework is presented as a way for evaluators to become more aware of the implications of various assumptions made by themselves and program stakeholders. The argument is examined and demonstrated that evaluators who are aware of the assumptions that underlie their evaluation choices are able to provide useful support for stakeholders in the examination of the assumptions they hold with regard to the nature of the problem being addressed, the program designed to solve the problem, and the approach to evaluation that is appropriate in that context. Such an informed approach has the potential for development of more appropriate and culturally responsive programs being implemented in ways that lead to the desired impacts, as well as to lead to evaluation approaches that support effective solutions to intransigent social problems. These arguments are illustrated through examples of evaluations from multiple sectors; additional challenges are also identified.

  7. Deep Borehole Field Test Requirements and Controlled Assumptions.

    SciTech Connect

    Hardin, Ernest

    2015-07-01

    This document presents design requirements and controlled assumptions intended for use in the engineering development and testing of: 1) prototype packages for radioactive waste disposal in deep boreholes; 2) a waste package surface handling system; and 3) a subsurface system for emplacing and retrieving packages in deep boreholes. Engineering development and testing is being performed as part of the Deep Borehole Field Test (DBFT; SNL 2014a). This document presents parallel sets of requirements for a waste disposal system and for the DBFT, showing the close relationship. In addition to design, it will also inform planning for drilling, construction, and scientific characterization activities for the DBFT. The information presented here follows typical preparations for engineering design. It includes functional and operating requirements for handling and emplacement/retrieval equipment, waste package design and emplacement requirements, borehole construction requirements, sealing requirements, and performance criteria. Assumptions are included where they could impact engineering design. Design solutions are avoided in the requirements discussion. Deep Borehole Field Test Requirements and Controlled Assumptions July 21, 2015 iv ACKNOWLEDGEMENTS This set of requirements and assumptions has benefited greatly from reviews by Gordon Appel, Geoff Freeze, Kris Kuhlman, Bob MacKinnon, Steve Pye, David Sassani, Dave Sevougian, and Jiann Su.

  8. Simplifying plasma chemistry via ILDM

    NASA Astrophysics Data System (ADS)

    Rehman, T.; Kemaneci, E.; Graef, W.; van Dijk, J.

    2016-02-01

    A plasma fluid model containing a large number of chemical species and reactions yields a high computational load. One of the methods to overcome this difficulty is to apply Chemical Reduction Techniques as used in combustion engineering. The chemical reduction technique that we study here is ILDM (Intrinsic Lower Dimensional Manifold). The ILDM method is used to simplify an argon plasma model and then a comparison is made with a CRM (Collisional Radiative Model).

  9. Simplifying the complexity of resistance heterogeneity in metastasis

    PubMed Central

    Lavi, Orit; Greene, James M.; Levy, Doron; Gottesman, Michael M.

    2014-01-01

    The main goal of treatment regimens for metastasis is to control growth rates, not eradicate all cancer cells. Mathematical models offer methodologies that incorporate high-throughput data with dynamic effects on net growth. The ideal approach would simplify, but not over-simplify, a complex problem into meaningful and manageable estimators that predict a patient’s response to specific treatments. Here, we explore three fundamental approaches with different assumptions concerning resistance mechanisms, in which the cells are categorized into either discrete compartments or described by a continuous range of resistance levels. We argue in favor of modeling resistance as a continuum and demonstrate how integrating cellular growth rates, density-dependent versus exponential growth, and intratumoral heterogeneity improves predictions concerning the resistance heterogeneity of metastases. PMID:24491979

  10. Methane generation in tropical landfills: simplified methods and field results.

    PubMed

    Machado, Sandro L; Carvalho, Miriam F; Gourc, Jean-Pierre; Vilar, Orencio M; do Nascimento, Julio C F

    2009-01-01

    This paper deals with the use of simplified methods to predict methane generation in tropical landfills. Methane recovery data obtained on site as part of a research program being carried out at the Metropolitan Landfill, Salvador, Brazil, is analyzed and used to obtain field methane generation over time. Laboratory data from MSW samples of different ages are presented and discussed; and simplified procedures to estimate the methane generation potential, Lo, and the constant related to the biodegradation rate, k are applied. The first order decay method is used to fit field and laboratory results. It is demonstrated that despite the assumptions and the simplicity of the adopted laboratory procedures, the values Lo and k obtained are very close to those measured in the field, thus making this kind of analysis very attractive for first approach purposes.

  11. A note on the assumption of quasiequilibrium in semiconductor junction devices

    NASA Technical Reports Server (NTRS)

    Von Roos, O.

    1977-01-01

    It is shown that the quasi-equilibrium theory for p-n junctions, as originally proposed by Shockley (1949), does not apply under conditions involving an application of comparatively low external voltages. A numerical example indicates that the quasi-equilibrium assumption must be discarded as soon as the voltage is increased beyond a certain critical value, although the system may still be in a low-level injection regime. It is currently not known which set of simplifying assumptions may replace the quasi-equilibrium assumptions. Possible analytic simplification relations applicable to moderate or high injection levels can, perhaps, be based on an approach considered by Mari (1968) and Choo (1971, 1972).

  12. Simplified tools for evaluating domestic ventilation systems

    SciTech Connect

    Maansson, L.G.; Orme, M.

    1999-07-01

    Within an International Energy Agency (IEA) project, Annex 27, experts from 8 countries (Canada, France, Italy, Japan, The Netherlands, Sweden, UK and USA) have developed simplified tools for evaluating domestic ventilation systems during the heating season. Tools for building and user aspects, thermal comfort, noise, energy, life cycle cost, reliability and indoor air quality (IAQ) have been devised. The results can be used both for dwellings at the design stage and after construction. The tools lead to immediate answers and indications about the consequences of different choices that may arise during discussion with clients. This paper presents an introduction to these tools. Examples applications of the indoor air quality and energy simplified tools are also provided. The IAQ tool accounts for constant emission sources, CO{sub 2}, cooking products, tobacco smoke, condensation risks, humidity levels (i.e., for judging the risk for mould and house dust mites), and pressure difference (for identifying the risk for radon or land fill spillage entering the dwelling or problems with indoor combustion appliances). An elaborated set of design parameters were worked out that resulted in about 17,000 combinations. By using multi-variate analysis it was possible to reduce this to 174 combinations for IAQ. In addition, a sensitivity analysis was made using 990 combinations. The results from all the runs were used to develop a simplified tool, as well as quantifying equations relying on the design parameters. A computerized energy tool has also been developed within this project, which takes into account air tightness, climate, window airing pattern, outdoor air flow rate and heat exchange efficiency.

  13. The OPERA hypothesis: assumptions and clarifications.

    PubMed

    Patel, Aniruddh D

    2012-04-01

    Recent research suggests that musical training enhances the neural encoding of speech. Why would musical training have this effect? The OPERA hypothesis proposes an answer on the basis of the idea that musical training demands greater precision in certain aspects of auditory processing than does ordinary speech perception. This paper presents two assumptions underlying this idea, as well as two clarifications, and suggests directions for future research.

  14. On distributional assumptions and whitened cosine similarities.

    PubMed

    Loog, Marco

    2008-06-01

    Recently, an interpretation of the whitened cosine similarity measure as a Bayes decision rule was proposed (C. Liu, "The Bayes Decision Rule Induced Similarity Measures,'' IEEE Trans. Pattern Analysis and Machine Intelligence, vol. 29, no. 6, pp. 1086-1090, June 2007. This communication makes the observation that some of the distributional assumptions made to derive this measure are very restrictive and, considered simultaneously, even inconsistent.

  15. Caring for Caregivers: Challenging the Assumptions.

    PubMed

    Williams, A Paul; Peckham, Allie; Kuluski, Kerry; Lum, Janet; Warrick, Natalie; Spalding, Karen; Tam, Tommy; Bruce-Barrett, Cindy; Grasic, Marta; Im, Jennifer

    2015-01-01

    Informal and mostly unpaid caregivers - spouses, family, friends and neighbours - play a crucial role in supporting the health, well-being, functional independence and quality of life of growing numbers of persons of all ages who cannot manage on their own. Yet, informal caregiving is in decline; falling rates of engagement in caregiving are compounded by a shrinking caregiver pool. How should policymakers respond? In this paper, we draw on a growing international literature, along with findings from community-based studies conducted by our team across Ontario, to highlight six common assumptions about informal caregivers and what can be done to support them. These include the assumption that caregivers will be there to take on an increasing responsibility; that caregiving is only about an aging population; that money alone can do the job; that policymakers can simply wait and see; that front-line care professionals should be left to fill the policy void; and that caregivers should be addressed apart from cared-for persons and formal care systems. While each assumption has a different focus, all challenge policymakers to view caregivers as key players in massive social and political change, and to respond accordingly.

  16. Occupancy estimation and the closure assumption

    USGS Publications Warehouse

    Rota, Christopher T.; Fletcher, Robert J.; Dorazio, Robert M.; Betts, Matthew G.

    2009-01-01

    1. Recent advances in occupancy estimation that adjust for imperfect detection have provided substantial improvements over traditional approaches and are receiving considerable use in applied ecology. To estimate and adjust for detectability, occupancy modelling requires multiple surveys at a site and requires the assumption of 'closure' between surveys, i.e. no changes in occupancy between surveys. Violations of this assumption could bias parameter estimates; however, little work has assessed model sensitivity to violations of this assumption or how commonly such violations occur in nature. 2. We apply a modelling procedure that can test for closure to two avian point-count data sets in Montana and New Hampshire, USA, that exemplify time-scales at which closure is often assumed. These data sets illustrate different sampling designs that allow testing for closure but are currently rarely employed in field investigations. Using a simulation study, we then evaluate the sensitivity of parameter estimates to changes in site occupancy and evaluate a power analysis developed for sampling designs that is aimed at limiting the likelihood of closure. 3. Application of our approach to point-count data indicates that habitats may frequently be open to changes in site occupancy at time-scales typical of many occupancy investigations, with 71% and 100% of species investigated in Montana and New Hampshire respectively, showing violation of closure across time periods of 3 weeks and 8 days respectively. 4. Simulations suggest that models assuming closure are sensitive to changes in occupancy. Power analyses further suggest that the modelling procedure we apply can effectively test for closure. 5. Synthesis and applications. Our demonstration that sites may be open to changes in site occupancy over time-scales typical of many occupancy investigations, combined with the sensitivity of models to violations of the closure assumption, highlights the importance of properly addressing

  17. A Proposal for Testing Local Realism Without Using Assumptions Related to Hidden Variable States

    NASA Technical Reports Server (NTRS)

    Ryff, Luiz Carlos

    1996-01-01

    A feasible experiment is discussed which allows us to prove a Bell's theorem for two particles without using an inequality. The experiment could be used to test local realism against quantum mechanics without the introduction of additional assumptions related to hidden variables states. Only assumptions based on direct experimental observation are needed.

  18. Simplifying microbial electrosynthesis reactor design

    PubMed Central

    Giddings, Cloelle G. S.; Nevin, Kelly P.; Woodward, Trevor; Lovley, Derek R.; Butler, Caitlyn S.

    2015-01-01

    Microbial electrosynthesis, an artificial form of photosynthesis, can efficiently convert carbon dioxide into organic commodities; however, this process has only previously been demonstrated in reactors that have features likely to be a barrier to scale-up. Therefore, the possibility of simplifying reactor design by both eliminating potentiostatic control of the cathode and removing the membrane separating the anode and cathode was investigated with biofilms of Sporomusa ovata. S. ovata reduces carbon dioxide to acetate and acts as the microbial catalyst for plain graphite stick cathodes as the electron donor. In traditional ‘H-cell’ reactors, where the anode and cathode chambers were separated with a proton-selective membrane, the rates and columbic efficiencies of microbial electrosynthesis remained high when electron delivery at the cathode was powered with a direct current power source rather than with a potentiostat-poised cathode utilized in previous studies. A membrane-less reactor with a direct-current power source with the cathode and anode positioned to avoid oxygen exposure at the cathode, retained high rates of acetate production as well as high columbic and energetic efficiencies. The finding that microbial electrosynthesis is feasible without a membrane separating the anode from the cathode, coupled with a direct current power source supplying the energy for electron delivery, is expected to greatly simplify future reactor design and lower construction costs. PMID:26029199

  19. Water resources of Assumption Parish, Louisiana

    USGS Publications Warehouse

    Prakken, Lawrence B.; Lovelace, John K.

    2013-01-01

    Information concerning the availability, use, and quality of water in Assumption Parish, Louisiana, is critical for proper water-supply management. The purpose of this fact sheet is to present information that can be used by water managers, parish residents, and others for management of this vital resource. Information on the availability, past and current use, use trends, and water quality from groundwater and surface-water sources in the parish is presented. Previously published reports and data stored in the U.S. Geological Survey’s National Water Information System (http://waterdata.usgs.gov/nwis) are the primary sources of the information presented here. In 2010, about 21.4 million gallons per day (Mgal/d) of water were withdrawn in Assumption Parish, including about 12.4 Mgal/d from surface-water sources and 9.03 Mgal/d from groundwater sources. Withdrawals for industrial use accounted for about 16.4 Mgal/d or 76 percent of the total water withdrawn. Other categories of use included public supply, rural domestic, livestock, general irrigation, and aquaculture.Water-use data collected at 5-year intervals from 1960 to 2010 indicated that water withdrawals peaked in 2000 at about 29.7 Mgal/d.

  20. Are nest sites actively chosen? Testing a common assumption for three non-resource limited birds

    NASA Astrophysics Data System (ADS)

    Goodenough, A. E.; Elliot, S. L.; Hart, A. G.

    2009-09-01

    Many widely-accepted ecological concepts are simplified assumptions about complex situations that remain largely untested. One example is the assumption that nest-building species choose nest sites actively when they are not resource limited. This assumption has seen little direct empirical testing: most studies on nest-site selection simply assume that sites are chosen actively (and seek explanations for such behaviour) without considering that sites may be selected randomly. We used 15 years of data from a nestbox scheme in the UK to test the assumption of active nest-site choice in three cavity-nesting bird species that differ in breeding and migratory strategy: blue tit ( Cyanistes caeruleus), great tit ( Parus major) and pied flycatcher ( Ficedula hypoleuca). Nest-site selection was non-random (implying active nest-site choice) for blue and great tits, but not for pied flycatchers. We also considered the relative importance of year-specific and site-specific factors in determining occupation of nest sites. Site-specific factors were more important than year-specific factors for the tit species, while the reverse was true for pied flycatchers. Our results show that nest-site selection, in birds at least, is not always the result of active choice, such that choice should not be assumed automatically in studies of nesting behaviour. We use this example to highlight the need to test key ecological assumptions empirically, and the importance of doing so across taxa rather than for single "model" species.

  1. Effectiveness of simplified dentin bonding systems.

    PubMed

    Imai, T; Itoh, K; Tani, C; Manabe, A; Yamashita, T; Hisamitsu, H; Wakumoto, S

    1998-03-01

    The effectiveness of newly developed commercial dentin bonding systems (SB, MB II and KB) was evaluated by measuring the contraction gap width of a resin composite restored into a cylindrical dentin cavity prepared in an extracted human molar and by measuring the tensile bond strength to the flat dentin surface. In addition, calcium loss during dentin conditioning was analyzed using electron microanalyses. An experimental dentin bonding system composed of EDTA conditioning, GM solution priming and a bonding agent containing 10-MDP was employed as a control in which it was presumed that contraction gap formation was prevented completely. However, gap formation was observed using the three commercial simplified dentin bonding systems. SEM observation showed that the gap was formed between the resin composite and the top surface of the dentin cavity wall indicating that the fracture occurred at the adhesive interface, but never inside the dentin nor inside the resin composite.

  2. Simplifier cut core inductor design

    NASA Technical Reports Server (NTRS)

    Mclyman, W. T.

    1976-01-01

    Advantages of specifying C cores and cut toroids fabricated from grain oriented silicon steels for use in high frequency power converters and pulse width modulated switching regulators are discussed. A method for rating cores assigns to each core a number which is the product of its window and core cross section area, called 'Area Product A sub p.' A correlation between the A sub p numbers and current density for a given temperature rise was developed. Also, straight line relationships were developed for A sub p and volume, A sub p and surface area, and A sub p and weight. These relationships can be used to simplify and standardize the process of inductor design. They also make it possible to design inductors of small bulk and volume or to optimize efficiency.

  3. Simplified Analysis Model for Predicting Pyroshock Responses on Composite Panel

    NASA Astrophysics Data System (ADS)

    Iwasa, Takashi; Shi, Qinzhong

    A simplified analysis model based on the frequency response analysis and the wave propagation analysis was established for predicting Shock Response Spectrum (SRS) on the composite panel subjected to pyroshock loadings. The complex composite panel was modeled as an isotropic single layer panel defined in NASA Lewis Method. Through the conductance of an impact excitation test on a composite panel with no equipment mounted on, it was presented that the simplified analysis model could estimate the SRS as well as the acceleration peak values in both near and far field in an accurate way. In addition, through the simulation for actual pyroshock tests on an actual satellite system, the simplified analysis model was proved to be applicable in predicting the actual pyroshock responses, while bringing forth several technical issues to estimate the pyroshock test specifications in early design stages.

  4. Simplified Methodology for Calculating Building Heating Loads.

    DTIC Science & Technology

    1980-11-01

    an inexpensive, accurate, and reliable simplified methodology , termed the "Modified Bin Method ", for 2 calculating building heating loads. In doing so...I AD-AI01 725 AIR FORCE INST OF TECH WRIGMT-PATTERSON AFB OH F/6 13/1 SIMPLIFIED METHODOLOGY FOR CALCULATING BUILDING HEATING LOADS.(U) NOV 80 S 0...University The Graduate School ," Department of Architectural Engineering 4, Simplified Methodology for Calculating Building Heating Loads, -A /. ’.- A

  5. Model investigation overthrows assumptions of watershed research

    NASA Astrophysics Data System (ADS)

    Schultz, Colin

    2012-04-01

    A 2009 study revealed serious flaws in a standard technique used by hydrological researchers to understand how changes in watershed land use affect stream flow behaviors, such as peak flows. The study caused academics and government agencies alike to rethink decades of watershed research and prompted Kuraś et al. to reinvestigate a number of long-standing assumptions in watershed research using a complex and well-validated computer model that accounts for a range of internal watershed dynamics and hydrologic processes. For the test site at 241 Creek in British Columbia, Canada, the authors found not only that deforestation increased the severity of floods but also that it had a scaling influence on both the magnitudes and frequencies of the floods. The model showed that the larger the flood, the more its magnitude was amplified by deforestation, with 10-to 100-year-return-period floods increasing in size by 9%-25%. Following a simulated removal of half of the watershed's trees, the authors found that 10-year-return-period floods occurred twice as often, while 100-year-return-period events became 5-6.7 times more frequent. This proportional relationship between the increase in flood magnitudes and frequencies following deforestation and the size of the flood runs counter to the prevailing wisdom in hydrological science.

  6. Transsexual parenthood and new role assumptions.

    PubMed

    Faccio, Elena; Bordin, Elena; Cipolletta, Sabrina

    2013-01-01

    This study explores the parental role of transsexuals and compares this to common assumptions about transsexuality and parentage. We conducted semi-structured interviews with 14 male-to-female transsexuals and 14 men, half parents and half non-parents, in order to explore four thematic areas: self-representation of the parental role, the description of the transsexual as a parent, the common representations of transsexuals as a parent, and male and female parental stereotypes. We conducted thematic and lexical analyses of the interviews using Taltac2 software. The results indicate that social representations of transsexuality and parenthood have a strong influence on processes of self-representation. Transsexual parents accurately understood conventional male and female parental prototypes and saw themselves as competent, responsible parents. They constructed their role based on affection toward the child rather than on the complementary role of their wives. In contrast, men's descriptions of transsexual parental roles were simpler and the descriptions of their parental role coincided with their personal experiences. These results suggest that the transsexual journey toward parenthood involves a high degree of re-adjustment, because their parental role does not coincide with a conventional one.

  7. 48 CFR 453.213 - Simplified Acquisition and other simplified purchase procedures (AD-838).

    Code of Federal Regulations, 2011 CFR

    2011-10-01

    ... 48 Federal Acquisition Regulations System 4 2011-10-01 2011-10-01 false Simplified Acquisition and other simplified purchase procedures (AD-838). 453.213 Section 453.213 Federal Acquisition Regulations... Acquisition and other simplified purchase procedures (AD-838). Form AD-838, Purchase Order, is prescribed...

  8. 48 CFR 453.213 - Simplified Acquisition and other simplified purchase procedures (AD-838).

    Code of Federal Regulations, 2010 CFR

    2010-10-01

    ... 48 Federal Acquisition Regulations System 4 2010-10-01 2010-10-01 false Simplified Acquisition and other simplified purchase procedures (AD-838). 453.213 Section 453.213 Federal Acquisition Regulations... Acquisition and other simplified purchase procedures (AD-838). Form AD-838, Purchase Order, is prescribed...

  9. Assumptions and ambiguities in nonplanar acoustic soliton theory

    SciTech Connect

    Verheest, Frank; Hellberg, Manfred A.

    2014-02-15

    There have been many recent theoretical investigations of the nonlinear evolution of electrostatic modes with cylindrical or spherical symmetry. Through a reductive perturbation analysis based on a quasiplanar stretching, a modified form of the Korteweg-de Vries or related equation is derived, containing an additional term which is linear in the electrostatic potential and singular at time t = 0. Unfortunately, these analyses contain several restrictive assumptions and ambiguities which are normally neither properly explained nor discussed, and severely limit the applicability of the technique. Most glaring are the use of plane-wave stretchings, the assumption that shape-preserving cylindrical modes can exist and that, although time is homogeneous, the origin of time (which can be chosen arbitrarily) needs to be avoided. Hence, only in the domain where the nonlinear modes are quasiplanar, far from the axis of cylindrical or from the origin of spherical symmetry can acceptable but unexciting results be obtained. Nonplanar nonlinear modes are clearly an interesting topic of research, as some of these phenomena have been observed in experiments. However, it is argued that a proper study of such modes needs numerical simulations rather than ill-suited analytical approximations.

  10. Roy's specific life values and the philosophical assumption of humanism.

    PubMed

    Hanna, Debra R

    2013-01-01

    Roy's philosophical assumption of humanism, which is shaped by the veritivity assumption, is considered in terms of her specific life values and in contrast to the contemporary view of humanism. Like veritivity, Roy's philosophical assumption of humanism unites a theocentric focus with anthropological values. Roy's perspective enriches the mainly secular, anthropocentric assumption. In this manuscript, the basis for Roy's perspective of humanism will be discussed so that readers will be able to use the Roy adaptation model in an authentic manner.

  11. Lens window simplifies TDL housing

    NASA Technical Reports Server (NTRS)

    Robinson, D. M.; Rowland, C. W.

    1979-01-01

    Lens window seal in tunable-diode-laser housing replaces plan parallel window. Lens seals housing and acts as optical-output coupler, thus eliminating need for additional reimaging or collimating optics.

  12. Philosophy of Technology Assumptions in Educational Technology Leadership

    ERIC Educational Resources Information Center

    Webster, Mark David

    2017-01-01

    A qualitative study using grounded theory methods was conducted to (a) examine what philosophy of technology assumptions are present in the thinking of K-12 technology leaders, (b) investigate how the assumptions may influence technology decision making, and (c) explore whether technological determinist assumptions are present. Subjects involved…

  13. Simplified method for nonlinear structural analysis

    NASA Technical Reports Server (NTRS)

    Kaufman, A.

    1983-01-01

    A simplified inelastic analysis computer program was developed for predicting the stress-strain history of a thermomechanically cycled structure from an elastic solution. The program uses an iterative and incremental procedure to estimate the plastic strains from the material stress-strain properties and a simulated plasticity hardening model. The simplified method was exercised on a number of problems involving uniaxial and multiaxial loading, isothermal and nonisothermal conditions, and different materials and plasticity models. Good agreement was found between these analytical results and nonlinear finite element solutions for these problems. The simplified analysis program used less than 1 percent of the CPU time required for a nonlinear finite element analysis.

  14. Simplified Storm Surge Simulations Using Bernstein Polynomials

    NASA Astrophysics Data System (ADS)

    Beisiegel, Nicole; Behrens, Jörn

    2016-04-01

    Storm surge simulations are vital for forecasting, hazard assessment and eventually improving our understanding of Earth system processes. Discontinuous Galerkin (DG) methods have recently been explored in that context, because they are locally mass-conservative and in combination with suitable robust nodal filtering techniques (slope limiters) positivity-preserving and well-balanced for the still water state at rest. These filters manipulate interpolation point values in every time step in order to retain the desirable properties of the scheme. In particular, DG methods are able to represent prognostic variables such as the fluid height at high-order accuracy inside each element (triangle). For simulations that include wetting and drying, however, the high-order accuracy will destabilize the numerical model because point values on quadrature points may become negative during the computation if they do not coincide with interpolation points. This is why the model that we are presenting utilizes Bernstein polynomials as basis functions to model the wetting and drying. This has the advantage that negative pointvalues away from interpolation points are prevented, the model is stabilized and no additional time step restriction is introduced. Numerical tests show that the model is capable of simulating simplified storm surges. Furthermore, a comparison of model results with third-order Bernstein polynomials with results using traditional nodal Lagrange polynomials reveals an improvement in numerical convergence.

  15. A simplified Reynolds stress model for unsteady turbulent boundary layers

    NASA Technical Reports Server (NTRS)

    Fan, Sixin; Lakshminarayana, Budugur

    1993-01-01

    A simplified Reynolds stress model has been developed for the prediction of unsteady turbulent boundary layers. By assuming that the net transport of Reynolds stresses is locally proportional to the net transport of the turbulent kinetic energy, the time dependent full Reynolds stress model is reduced to a set of ordinary differential equations. These equations contain only time derivatives and can be readily integrated in a time dependent boundary layer or Navier-Stokes code. The turbulent kinetic energy and dissipation rate needed for the model are obtained by solving the k-epsilon equations. This simplified Reynolds stress turbulence model (SRSM) does not use the eddy viscosity assumption, which may not be valid for unsteady turbulent flows. The anisotropy of both the steady and the unsteady turbulent normal stresses can be captured by the SRSM model. Through proper damping of the shear stresses, the present model can be used in the near wall region of turbulent boundary layers. This model has been validated against data for steady and unsteady turbulent boundary layers, including periodic turbulent boundary layers subjected to a mean adverse pressure gradient. For the cases tested, the predicted unsteady velocity and turbulent stress components agree well with the experimental data. Comparison between the predictions from the SRSM model and a k-epsilon model is also presented.

  16. Statistical Tests of the PTHA Poisson Assumption for Submarine Landslides

    NASA Astrophysics Data System (ADS)

    Geist, E. L.; Chaytor, J. D.; Parsons, T.; Ten Brink, U. S.

    2012-12-01

    We demonstrate that a sequence of dated mass transport deposits (MTDs) can provide information to statistically test whether or not submarine landslides associated with these deposits conform to a Poisson model of occurrence. Probabilistic tsunami hazard analysis (PTHA) most often assumes Poissonian occurrence for all sources, with an exponential distribution of return times. Using dates that define the bounds of individual MTDs, we first describe likelihood and Monte Carlo methods of parameter estimation for a suite of candidate occurrence models (Poisson, lognormal, gamma, Brownian Passage Time). In addition to age-dating uncertainty, both methods incorporate uncertainty caused by the open time intervals: i.e., before the first and after the last event to the present. Accounting for these open intervals is critical when there are a small number of observed events. The optimal occurrence model is selected according to both the Akaike Information Criteria (AIC) and Akaike's Bayesian Information Criterion (ABIC). In addition, the likelihood ratio test can be performed on occurrence models from the same family: e.g., the gamma model relative to the exponential model of return time distribution. Parameter estimation, model selection, and hypothesis testing are performed on data from two IODP holes in the northern Gulf of Mexico that penetrated a total of 14 MTDs, some of which are correlated between the two holes. Each of these events has been assigned an age based on microfossil zonations and magnetostratigraphic datums. Results from these sites indicate that the Poisson assumption is likely valid. However, parameter estimation results using the likelihood method for one of the sites suggest that the events may have occurred quasi-periodically. Methods developed in this study provide tools with which one can determine both the rate of occurrence and the statistical validity of the Poisson assumption when submarine landslides are included in PTHA.

  17. Finite Element Modeling of a Cylindrical Contact Using Hertzian Assumptions

    NASA Technical Reports Server (NTRS)

    Knudsen, Erik

    2003-01-01

    The turbine blades in the high-pressure fuel turbopump/alternate turbopump (HPFTP/AT) are subjected to hot gases rapidly flowing around them. This flow excites vibrations in the blades. Naturally, one has to worry about resonance, so a damping device was added to dissipate some energy from the system. The foundation is now laid for a very complex problem. The damper is in contact with the blade, so now there are contact stresses (both normal and tangential) to contend with. Since these stresses can be very high, it is not all that difficult to yield the material. Friction is another non-linearity and the blade is made out of a Nickel-based single-crystal superalloy that is orthotropic. A few approaches exist to solve such a problem and computer models, using contact elements, have been built with friction, plasticity, etc. These models are quite cumbersome and require many hours to solve just one load case and material orientation. A simpler approach is required. Ideally, the model should be simplified so the analysis can be conducted faster. When working with contact problems determining the contact patch and the stresses in the material are the main concerns. Closed-form solutions for non-conforming bodies, developed by Hertz, made out of isotropic materials are readily available. More involved solutions for 3-D cases using different materials are also available. The question is this: can Hertzian1 solutions be applied, or superimposed, to more complicated problems-like those involving anisotropic materials? That is the point of the investigation here. If these results agree with the more complicated computer models, then the analytical solutions can be used in lieu of the numerical solutions that take a very long time to process. As time goes on, the analytical solution will eventually have to include things like friction and plasticity. The models in this report use no contact elements and are essentially an applied load problem using Hertzian assumptions to

  18. Veitch diagram plotter simplifies Boolean functions

    NASA Technical Reports Server (NTRS)

    Rubin, D. K.

    1964-01-01

    This device for simplifying the plotting of a Veitch diagram consists of several overlays for blocking out the unwanted squares. This method of plotting the various input combinations to a computer is used in conjunction with the Boolean functions.

  19. Simplified Derivation Of Geometrical Dynamic Range Compression

    NASA Astrophysics Data System (ADS)

    Altman, W. P.

    1982-08-01

    A simplified derivation is presented which illustrates the effects of unmatched fields of view for a lidar system. The resultant equations are easily evaluated on a programmable calculator, and comparison with previous calculations are numerically similar.

  20. Non-rigid response of the trunk to dynamic axial loading: an evaluation of current modelling assumptions.

    PubMed

    McGill, S M; Thorstensson, A; Norman, R W

    1989-02-01

    Most biomechanical models of the trunk and limb system represent the segments as rigid links joined with pins. The implication of this simplifying assumption is that dynamic loads applied to the body are instantaneously transmitted through the linkage with no distortion of the applied force-time waveform. The purpose of this study was to determine whether this assumption was justifiable for loads applied to the hands or shoulders and transmitted through the trunk to the bottom of the pelvis. If not, the intent was to find a transfer function relating load inputs applied to the top of the trunk and outputs measured at the bottom of the trunk. Loads were manually applied to a force transducer attached to a shoulder yoke or to a hand-held loading pan mounted on three subjects. The subjects sat both erect and slouched and also varied the trunk stiffness by voluntarily changing trunk muscle activation. Subjects sat on a rigid steel stool mounted on a force plate from which force outputs were recorded. In addition, a few case studies were performed; some trials utilized light impacts by a hammer to the transducer mounted on the crown of the bare head and in other trials the body was in a standing (lifting) posture. The results indicated that the mechanical response is indeed time dependent. For applied forces with a rise time less than 20 ms (hammer impact) the trunk system reduced the magnitude of the force peaks by up to 40%. However, for applied loads with rise times from about 30 to 300 ms the peak output force was larger than the applied force peak. This was due to inertial effects of the upper body mass as it 'rebounded' from the applied load. During these conditions of biologically generated applied forces, the output could be reasonably reconstructed by processing the input signal through a Butterworth second order low pass filter with a cutoff frequency of 7·5 Hz, a gain and in some cases, a shift in time up to 20 ms. The assumption of rigid body segments in

  1. Simplified solution for stresses and deformation

    NASA Technical Reports Server (NTRS)

    Hamrock, B. J.; Brewe, D. E.

    1981-01-01

    Conventional contact deformation analysis for ball bearings, gears, and cams involves tedious iterative procedures or the use of design charts. A simplified approach that makes the elastic deformation at the center of contact easy to calculate was previously reported. The range of validity in which these equations can be used is extended. A simplified approach to the calculation of the location and magnitude of subsurface stresses developed in machine element applications is included.

  2. Simplified Quantum Logic with Trapped Ions

    DTIC Science & Technology

    2016-06-23

    PHYSICAL REVIEW A ATOMIC , MOLECULAR, AND OPTICAL PHYSICS THIRD SERIES, VOLUME 55, NUMBER 4 APRIL 1997Simplified quantum logic with trapped ions C...Received 17 December 1996! We describe a simplified scheme for quantum logic with a collection of laser-cooled trapped atomic ions. Building on the...in a system of laser-cooled trapped atomic ions. In the simplest form of the ion trap quantum computer, two internal electronic levels of each ion in

  3. Simplified models for LHC new physics searches

    NASA Astrophysics Data System (ADS)

    Alves, Daniele; Arkani-Hamed, Nima; Arora, Sanjay; Bai, Yang; Baumgart, Matthew; Berger, Joshua; Buckley, Matthew; Butler, Bart; Chang, Spencer; Cheng, Hsin-Chia; Cheung, Clifford; Sekhar Chivukula, R.; Cho, Won Sang; Cotta, Randy; D'Alfonso, Mariarosaria; El Hedri, Sonia; Essig (Editor, Rouven; Evans, Jared A.; Fitzpatrick, Liam; Fox, Patrick; Franceschini, Roberto; Freitas, Ayres; Gainer, James S.; Gershtein, Yuri; Gray, Richard; Gregoire, Thomas; Gripaios, Ben; Gunion, Jack; Han, Tao; Haas, Andy; Hansson, Per; Hewett, JoAnne; Hits, Dmitry; Hubisz, Jay; Izaguirre, Eder; Kaplan, Jared; Katz, Emanuel; Kilic, Can; Kim, Hyung-Do; Kitano, Ryuichiro; Koay, Sue Ann; Ko, Pyungwon; Krohn, David; Kuflik, Eric; Lewis, Ian; Lisanti (Editor, Mariangela; Liu, Tao; Liu, Zhen; Lu, Ran; Luty, Markus; Meade, Patrick; Morrissey, David; Mrenna, Stephen; Nojiri, Mihoko; Okui, Takemichi; Padhi, Sanjay; Papucci, Michele; Park, Michael; Park, Myeonghun; Perelstein, Maxim; Peskin, Michael; Phalen, Daniel; Rehermann, Keith; Rentala, Vikram; Roy, Tuhin; Ruderman, Joshua T.; Sanz, Veronica; Schmaltz, Martin; Schnetzer, Stephen; Schuster (Editor, Philip; Schwaller, Pedro; Schwartz, Matthew D.; Schwartzman, Ariel; Shao, Jing; Shelton, Jessie; Shih, David; Shu, Jing; Silverstein, Daniel; Simmons, Elizabeth; Somalwar, Sunil; Spannowsky, Michael; Spethmann, Christian; Strassler, Matthew; Su, Shufang; Tait (Editor, Tim; Thomas, Brooks; Thomas, Scott; Toro (Editor, Natalia; Volansky, Tomer; Wacker (Editor, Jay; Waltenberger, Wolfgang; Yavin, Itay; Yu, Felix; Zhao, Yue; Zurek, Kathryn; LHC New Physics Working Group

    2012-10-01

    This document proposes a collection of simplified models relevant to the design of new-physics searches at the Large Hadron Collider (LHC) and the characterization of their results. Both ATLAS and CMS have already presented some results in terms of simplified models, and we encourage them to continue and expand this effort, which supplements both signature-based results and benchmark model interpretations. A simplified model is defined by an effective Lagrangian describing the interactions of a small number of new particles. Simplified models can equally well be described by a small number of masses and cross-sections. These parameters are directly related to collider physics observables, making simplified models a particularly effective framework for evaluating searches and a useful starting point for characterizing positive signals of new physics. This document serves as an official summary of the results from the ‘Topologies for Early LHC Searches’ workshop, held at SLAC in September of 2010, the purpose of which was to develop a set of representative models that can be used to cover all relevant phase space in experimental searches. Particular emphasis is placed on searches relevant for the first ˜50-500 pb-1 of data and those motivated by supersymmetric models. This note largely summarizes material posted at http://lhcnewphysics.org/, which includes simplified model definitions, Monte Carlo material, and supporting contacts within the theory community. We also comment on future developments that may be useful as more data is gathered and analyzed by the experiments.

  4. Simplified Models for LHC New Physics Searches

    SciTech Connect

    Alves, Daniele; Arkani-Hamed, Nima; Arora, Sanjay; Bai, Yang; Baumgart, Matthew; Berger, Joshua; Buckley, Matthew; Butler, Bart; Chang, Spencer; Cheng, Hsin-Chia; Cheung, Clifford; Chivukula, R.Sekhar; Cho, Won Sang; Cotta, Randy; D'Alfonso, Mariarosaria; El Hedri, Sonia; Essig, Rouven,; Evans, Jared A.; Fitzpatrick, Liam; Fox, Patrick; Franceschini, Roberto; /more authors..

    2012-06-01

    This document proposes a collection of simplified models relevant to the design of new-physics searches at the LHC and the characterization of their results. Both ATLAS and CMS have already presented some results in terms of simplified models, and we encourage them to continue and expand this effort, which supplements both signature-based results and benchmark model interpretations. A simplified model is defined by an effective Lagrangian describing the interactions of a small number of new particles. Simplified models can equally well be described by a small number of masses and cross-sections. These parameters are directly related to collider physics observables, making simplified models a particularly effective framework for evaluating searches and a useful starting point for characterizing positive signals of new physics. This document serves as an official summary of the results from the 'Topologies for Early LHC Searches' workshop, held at SLAC in September of 2010, the purpose of which was to develop a set of representative models that can be used to cover all relevant phase space in experimental searches. Particular emphasis is placed on searches relevant for the first {approx} 50-500 pb{sup -1} of data and those motivated by supersymmetric models. This note largely summarizes material posted at http://lhcnewphysics.org/, which includes simplified model definitions, Monte Carlo material, and supporting contacts within the theory community. We also comment on future developments that may be useful as more data is gathered and analyzed by the experiments.

  5. Post-traumatic stress and world assumptions: the effects of religious coping.

    PubMed

    Zukerman, Gil; Korn, Liat

    2014-12-01

    Religiosity has been shown to moderate the negative effects of traumatic event experiences. The current study was designed to examine the relationship between post-traumatic stress (PTS) following traumatic event exposure; world assumptions defined as basic cognitive schemas regarding the world; and self and religious coping conceptualized as drawing on religious beliefs and practices for understanding and dealing with life stressors. This study examined 777 Israeli undergraduate students who completed several questionnaires which sampled individual world assumptions and religious coping in addition to measuring PTS, as manifested by the PTSD check list. Results indicate that positive religious coping was significantly associated with more positive world assumptions, while negative religious coping was significantly associated with more negative world assumptions. Additionally, negative world assumptions were significantly associated with more avoidance symptoms, while reporting higher rates of traumatic event exposure was significantly associated with more hyper-arousal. These findings suggest that religious-related cognitive schemas directly affect world assumptions by creating protective shields that may prevent the negative effects of confronting an extreme negative experience.

  6. Green technologies--assumption of economic recovery.

    PubMed

    Siljeg, Mario; Zorić, Sandra Tucak; Vucinić, Aleksandra Anić; Kalambura, Sanja; Cemerin, Vedrana; Jovicić, Nives

    2014-03-01

    Green technologies include implementation of technological projects in the field of environmental protection through all associated components, such as: waters, soil, air or biodiversity. Hence, such projects potentially become a driving force of new economic momentum in the conditions of post-crisis recovery. In addition, the support of this segment by the institutions of the European Union, either in terms of organization through the establishment of rules and systems for monitoring and control of environmental protection measures, and most importantly in terms of finances, by supporting the development of infrastructure for environmental protection, is today an indisputable category. The aim of the research is to show the potential of green technologies in the initiation of economic activities based on content analysis of the collected literature, as well as to determine the correlation between green technologies and environmental protection and the measures for the reduction of the impact of energy sector on the greenhouse gas emissions.

  7. Fluid-Structure Interaction Modeling of Intracranial Aneurysm Hemodynamics: Effects of Different Assumptions

    NASA Astrophysics Data System (ADS)

    Rajabzadeh Oghaz, Hamidreza; Damiano, Robert; Meng, Hui

    2015-11-01

    Intracranial aneurysms (IAs) are pathological outpouchings of cerebral vessels, the progression of which are mediated by complex interactions between the blood flow and vasculature. Image-based computational fluid dynamics (CFD) has been used for decades to investigate IA hemodynamics. However, the commonly adopted simplifying assumptions in CFD (e.g. rigid wall) compromise the simulation accuracy and mask the complex physics involved in IA progression and eventual rupture. Several groups have considered the wall compliance by using fluid-structure interaction (FSI) modeling. However, FSI simulation is highly sensitive to numerical assumptions (e.g. linear-elastic wall material, Newtonian fluid, initial vessel configuration, and constant pressure outlet), the effects of which are poorly understood. In this study, a comprehensive investigation of the sensitivity of FSI simulations in patient-specific IAs is investigated using a multi-stage approach with a varying level of complexity. We start with simulations incorporating several common simplifications: rigid wall, Newtonian fluid, and constant pressure at the outlets, and then we stepwise remove these simplifications until the most comprehensive FSI simulations. Hemodynamic parameters such as wall shear stress and oscillatory shear index are assessed and compared at each stage to better understand the sensitivity of in FSI simulations for IA to model assumptions. Supported by the National Institutes of Health (1R01 NS 091075-01).

  8. Effects of various assumptions on the calculated liquid fraction in isentropic saturated equilibrium expansions

    NASA Technical Reports Server (NTRS)

    Bursik, J. W.; Hall, R. M.

    1980-01-01

    The saturated equilibrium expansion approximation for two phase flow often involves ideal-gas and latent-heat assumptions to simplify the solution procedure. This approach is well documented by Wegener and Mack and works best at low pressures where deviations from ideal-gas behavior are small. A thermodynamic expression for liquid mass fraction that is decoupled from the equations of fluid mechanics is used to compare the effects of the various assumptions on nitrogen-gas saturated equilibrium expansion flow starting at 8.81 atm, 2.99 atm, and 0.45 atm, which are conditions representative of transonic cryogenic wind tunnels. For the highest pressure case, the entire set of ideal-gas and latent-heat assumptions are shown to be in error by 62 percent for the values of heat capacity and latent heat. An approximation of the exact, real-gas expression is also developed using a constant, two phase isentropic expansion coefficient which results in an error of only 2 percent for the high pressure case.

  9. The influence of computational assumptions on analysing abdominal aortic aneurysm haemodynamics.

    PubMed

    Ene, Florentina; Delassus, Patrick; Morris, Liam

    2014-08-01

    The variation in computational assumptions for analysing abdominal aortic aneurysm haemodynamics can influence the desired output results and computational cost. Such assumptions for abdominal aortic aneurysm modelling include static/transient pressures, steady/transient flows and rigid/compliant walls. Six computational methods and these various assumptions were simulated and compared within a realistic abdominal aortic aneurysm model with and without intraluminal thrombus. A full transient fluid-structure interaction was required to analyse the flow patterns within the compliant abdominal aortic aneurysms models. Rigid wall computational fluid dynamics overestimates the velocity magnitude by as much as 40%-65% and the wall shear stress by 30%-50%. These differences were attributed to the deforming walls which reduced the outlet volumetric flow rate for the transient fluid-structure interaction during the majority of the systolic phase. Static finite element analysis accurately approximates the deformations and von Mises stresses when compared with transient fluid-structure interaction. Simplifying the modelling complexity reduces the computational cost significantly. In conclusion, the deformation and von Mises stress can be approximately found by static finite element analysis, while for compliant models a full transient fluid-structure interaction analysis is required for acquiring the fluid flow phenomenon.

  10. Impact of actuarial assumptions on pension costs: A simulation analysis

    NASA Astrophysics Data System (ADS)

    Yusof, Shaira; Ibrahim, Rose Irnawaty

    2013-04-01

    This study investigates the sensitivity of pension costs to changes in the underlying assumptions of a hypothetical pension plan in order to gain a perspective on the relative importance of the various actuarial assumptions via a simulation analysis. Simulation analyses are used to examine the impact of actuarial assumptions on pension costs. There are two actuarial assumptions will be considered in this study which are mortality rates and interest rates. To calculate pension costs, Accrued Benefit Cost Method, constant amount (CA) modification, constant percentage of salary (CS) modification are used in the study. The mortality assumptions and the implied mortality experience of the plan can potentially have a significant impact on pension costs. While for interest rate assumptions, it is inversely related to the pension costs. Results of the study have important implications for analyst of pension costs.

  11. A Simplified HTTR Diffusion Theory Benchmark

    SciTech Connect

    Rodolfo M. Ferrer; Abderrafi M. Ougouag; Farzad Rahnema

    2010-10-01

    The Georgia Institute of Technology (GA-Tech) recently developed a transport theory benchmark based closely on the geometry and the features of the HTTR reactor that is operational in Japan. Though simplified, the benchmark retains all the principal physical features of the reactor and thus provides a realistic and challenging test for the codes. The purpose of this paper is twofold. The first goal is an extension of the benchmark to diffusion theory applications by generating the additional data not provided in the GA-Tech prior work. The second goal is to use the benchmark on the HEXPEDITE code available to the INL. The HEXPEDITE code is a Green’s function-based neutron diffusion code in 3D hexagonal-z geometry. The results showed that the HEXPEDITE code accurately reproduces the effective multiplication factor of the reference HELIOS solution. A secondary, but no less important, conclusion is that in the testing against actual HTTR data of a full sequence of codes that would include HEXPEDITE, in the apportioning of inevitable discrepancies between experiment and models, the portion of error attributable to HEXPEDITE would be expected to be modest. If large discrepancies are observed, they would have to be explained by errors in the data fed into HEXPEDITE. Results based on a fully realistic model of the HTTR reactor are presented in a companion paper. The suite of codes used in that paper also includes HEXPEDITE. The results shown here should help that effort in the decision making process for refining the modeling steps in the full sequence of codes.

  12. Simplified Warfarin Dose-response Pharmacodynamic Models

    PubMed Central

    Kim, Seongho; Gaweda, Adam E.; Wu, Dongfeng; Li, Lang; Rai, Shesh N.; Brier, Michael E.

    2014-01-01

    Warfarin is a frequently used oral anticoagulant for long-term prevention and treatment of thromboembolic events. Due to its narrow therapeutic range and large inter-individual dose-response variability, it is highly desirable to personalize warfarin dosing. However, the complexity of the conventional kinetic-pharmacodynamic (K-PD) models hampers the development of the personalized dose management. To avert this challenge, we propose simplified PD models for warfarin dose-response relationship, which is motivated by ideas from control theory. The simplified models were further applied to longitudinal data of 37 patients undergoing anticoagulation treatment using the standard two-stage approach and then compared with the conventional K-PD models. Data analysis shows that all models have a similar predictive ability, but the simplified models are most parsimonious. PMID:25750489

  13. Hypersonic Vehicle Propulsion System Simplified Model Development

    NASA Technical Reports Server (NTRS)

    Stueber, Thomas J.; Raitano, Paul; Le, Dzu K.; Ouzts, Peter

    2007-01-01

    This document addresses the modeling task plan for the hypersonic GN&C GRC team members. The overall propulsion system modeling task plan is a multi-step process and the task plan identified in this document addresses the first steps (short term modeling goals). The procedures and tools produced from this effort will be useful for creating simplified dynamic models applicable to a hypersonic vehicle propulsion system. The document continues with the GRC short term modeling goal. Next, a general description of the desired simplified model is presented along with simulations that are available to varying degrees. The simulations may be available in electronic form (FORTRAN, CFD, MatLab,...) or in paper form in published documents. Finally, roadmaps outlining possible avenues towards realizing simplified model are presented.

  14. 14 CFR 29.473 - Ground loading conditions and assumptions.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... 14 Aeronautics and Space 1 2011-01-01 2011-01-01 false Ground loading conditions and assumptions... TRANSPORTATION AIRCRAFT AIRWORTHINESS STANDARDS: TRANSPORT CATEGORY ROTORCRAFT Strength Requirements Ground Loads § 29.473 Ground loading conditions and assumptions. (a) For specified landing conditions, a...

  15. 14 CFR 29.473 - Ground loading conditions and assumptions.

    Code of Federal Regulations, 2013 CFR

    2013-01-01

    ... 14 Aeronautics and Space 1 2013-01-01 2013-01-01 false Ground loading conditions and assumptions... TRANSPORTATION AIRCRAFT AIRWORTHINESS STANDARDS: TRANSPORT CATEGORY ROTORCRAFT Strength Requirements Ground Loads § 29.473 Ground loading conditions and assumptions. (a) For specified landing conditions, a...

  16. 14 CFR 27.473 - Ground loading conditions and assumptions.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... 14 Aeronautics and Space 1 2011-01-01 2011-01-01 false Ground loading conditions and assumptions... TRANSPORTATION AIRCRAFT AIRWORTHINESS STANDARDS: NORMAL CATEGORY ROTORCRAFT Strength Requirements Ground Loads § 27.473 Ground loading conditions and assumptions. (a) For specified landing conditions, a...

  17. 14 CFR 29.473 - Ground loading conditions and assumptions.

    Code of Federal Regulations, 2012 CFR

    2012-01-01

    ... 14 Aeronautics and Space 1 2012-01-01 2012-01-01 false Ground loading conditions and assumptions... TRANSPORTATION AIRCRAFT AIRWORTHINESS STANDARDS: TRANSPORT CATEGORY ROTORCRAFT Strength Requirements Ground Loads § 29.473 Ground loading conditions and assumptions. (a) For specified landing conditions, a...

  18. 14 CFR 27.473 - Ground loading conditions and assumptions.

    Code of Federal Regulations, 2012 CFR

    2012-01-01

    ... 14 Aeronautics and Space 1 2012-01-01 2012-01-01 false Ground loading conditions and assumptions... TRANSPORTATION AIRCRAFT AIRWORTHINESS STANDARDS: NORMAL CATEGORY ROTORCRAFT Strength Requirements Ground Loads § 27.473 Ground loading conditions and assumptions. (a) For specified landing conditions, a...

  19. 14 CFR 27.473 - Ground loading conditions and assumptions.

    Code of Federal Regulations, 2013 CFR

    2013-01-01

    ... 14 Aeronautics and Space 1 2013-01-01 2013-01-01 false Ground loading conditions and assumptions... TRANSPORTATION AIRCRAFT AIRWORTHINESS STANDARDS: NORMAL CATEGORY ROTORCRAFT Strength Requirements Ground Loads § 27.473 Ground loading conditions and assumptions. (a) For specified landing conditions, a...

  20. Exposing Trust Assumptions in Distributed Policy Enforcement (Briefing Charts)

    DTIC Science & Technology

    2009-11-04

    506. September 2009, Pisa , Italy. – "Path-based Access Control for Enterprise Networks” Matthew Burnside and Angelos D. Keromytis. In Proceedings of...assumptions? 11/04/09 ONR MURI Review 14 Summary • Exploring systems that allow (and require) explicit assumption (trust) declarations • All

  1. Testing Assumptions: The Impact of Two Study Abroad Program Models

    ERIC Educational Resources Information Center

    Norris, Emily Mohajeri; Dwyer, Mary M.

    2005-01-01

    There are many untested, long-held assumptions within the field of study abroad concerning the impact of program elements such as study duration, language of instruction, program models, and student housing choices. One assumption embraced within the field is that direct enrollment (or full immersion) programs are more effective at achieving a…

  2. When Proofs Reflect More on Assumptions than Conclusions

    ERIC Educational Resources Information Center

    Dawkins, Paul Christian

    2014-01-01

    This paper demonstrates how questions of "provability" can help students engaged in reinvention of mathematical theory to understand the axiomatic game. While proof demonstrates how conclusions follow from assumptions, "provability" characterizes the dual relation that assumptions are "justified" when they afford…

  3. 14 CFR 29.473 - Ground loading conditions and assumptions.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... 14 Aeronautics and Space 1 2010-01-01 2010-01-01 false Ground loading conditions and assumptions... TRANSPORTATION AIRCRAFT AIRWORTHINESS STANDARDS: TRANSPORT CATEGORY ROTORCRAFT Strength Requirements Ground Loads § 29.473 Ground loading conditions and assumptions. (a) For specified landing conditions, a...

  4. Co-Dependency: An Examination of Underlying Assumptions.

    ERIC Educational Resources Information Center

    Myer, Rick A.; And Others

    1991-01-01

    Discusses need for careful examination of codependency as diagnostic category. Critically examines assumptions that codependency is disease, addiction, or predetermined by the environment. Discusses implications of assumptions. Offers recommendations for mental health counselors focusing on need for systematic research, redirection of efforts to…

  5. 26 CFR 1.357-1 - Assumption of liability.

    Code of Federal Regulations, 2013 CFR

    2013-04-01

    ... 26 Internal Revenue 4 2013-04-01 2013-04-01 false Assumption of liability. 1.357-1 Section 1.357-1 Internal Revenue INTERNAL REVENUE SERVICE, DEPARTMENT OF THE TREASURY (CONTINUED) INCOME TAX (CONTINUED) INCOME TAXES (CONTINUED) Effects on Shareholders and Security Holders § 1.357-1 Assumption of...

  6. 26 CFR 1.357-1 - Assumption of liability.

    Code of Federal Regulations, 2010 CFR

    2010-04-01

    ... 26 Internal Revenue 4 2010-04-01 2010-04-01 false Assumption of liability. 1.357-1 Section 1.357-1 Internal Revenue INTERNAL REVENUE SERVICE, DEPARTMENT OF THE TREASURY (CONTINUED) INCOME TAX (CONTINUED) INCOME TAXES Effects on Shareholders and Security Holders § 1.357-1 Assumption of liability. (a)...

  7. 26 CFR 1.357-1 - Assumption of liability.

    Code of Federal Regulations, 2014 CFR

    2014-04-01

    ... 26 Internal Revenue 4 2014-04-01 2014-04-01 false Assumption of liability. 1.357-1 Section 1.357-1 Internal Revenue INTERNAL REVENUE SERVICE, DEPARTMENT OF THE TREASURY (CONTINUED) INCOME TAX (CONTINUED) INCOME TAXES (CONTINUED) Effects on Shareholders and Security Holders § 1.357-1 Assumption of...

  8. 26 CFR 1.357-1 - Assumption of liability.

    Code of Federal Regulations, 2012 CFR

    2012-04-01

    ... 26 Internal Revenue 4 2012-04-01 2012-04-01 false Assumption of liability. 1.357-1 Section 1.357-1 Internal Revenue INTERNAL REVENUE SERVICE, DEPARTMENT OF THE TREASURY (CONTINUED) INCOME TAX (CONTINUED) INCOME TAXES (Continued) Effects on Shareholders and Security Holders § 1.357-1 Assumption of...

  9. 26 CFR 1.357-1 - Assumption of liability.

    Code of Federal Regulations, 2011 CFR

    2011-04-01

    ... 26 Internal Revenue 4 2011-04-01 2011-04-01 false Assumption of liability. 1.357-1 Section 1.357-1 Internal Revenue INTERNAL REVENUE SERVICE, DEPARTMENT OF THE TREASURY (CONTINUED) INCOME TAX (CONTINUED) INCOME TAXES Effects on Shareholders and Security Holders § 1.357-1 Assumption of liability. (a)...

  10. 40 CFR 265.150 - State assumption of responsibility.

    Code of Federal Regulations, 2010 CFR

    2010-07-01

    ... 40 Protection of Environment 25 2010-07-01 2010-07-01 false State assumption of responsibility. 265.150 Section 265.150 Protection of Environment ENVIRONMENTAL PROTECTION AGENCY (CONTINUED) SOLID..., STORAGE, AND DISPOSAL FACILITIES Financial Requirements § 265.150 State assumption of responsibility....

  11. 40 CFR 264.150 - State assumption of responsibility.

    Code of Federal Regulations, 2010 CFR

    2010-07-01

    ... 40 Protection of Environment 25 2010-07-01 2010-07-01 false State assumption of responsibility. 264.150 Section 264.150 Protection of Environment ENVIRONMENTAL PROTECTION AGENCY (CONTINUED) SOLID... FACILITIES Financial Requirements § 264.150 State assumption of responsibility. (a) If a State either...

  12. 29 CFR 4231.10 - Actuarial calculations and assumptions.

    Code of Federal Regulations, 2010 CFR

    2010-07-01

    ... 29 Labor 9 2010-07-01 2010-07-01 false Actuarial calculations and assumptions. 4231.10 Section... MULTIEMPLOYER PLANS § 4231.10 Actuarial calculations and assumptions. (a) Most recent valuation. All calculations required by this part must be based on the most recent actuarial valuation as of the date...

  13. 29 CFR Appendix C to Part 4044 - Loading Assumptions

    Code of Federal Regulations, 2010 CFR

    2010-07-01

    ... 29 Labor 9 2010-07-01 2010-07-01 false Loading Assumptions C Appendix C to Part 4044 Labor... ASSETS IN SINGLE-EMPLOYER PLANS Pt. 4044, App. C Appendix C to Part 4044—Loading Assumptions If the total value of the plan's benefit liabilities (as defined in 29 U.S.C. § 1301(a)(16)), exclusive of...

  14. 29 CFR Appendix C to Part 4044 - Loading Assumptions

    Code of Federal Regulations, 2013 CFR

    2013-07-01

    ... 29 Labor 9 2013-07-01 2013-07-01 false Loading Assumptions C Appendix C to Part 4044 Labor... ASSETS IN SINGLE-EMPLOYER PLANS Pt. 4044, App. C Appendix C to Part 4044—Loading Assumptions If the total value of the plan's benefit liabilities (as defined in 29 U.S.C. § 1301(a)(16)), exclusive of...

  15. 29 CFR Appendix C to Part 4044 - Loading Assumptions

    Code of Federal Regulations, 2012 CFR

    2012-07-01

    ... 29 Labor 9 2012-07-01 2012-07-01 false Loading Assumptions C Appendix C to Part 4044 Labor... ASSETS IN SINGLE-EMPLOYER PLANS Pt. 4044, App. C Appendix C to Part 4044—Loading Assumptions If the total value of the plan's benefit liabilities (as defined in 29 U.S.C. § 1301(a)(16)), exclusive of...

  16. SIMPLIFIED LAPAROSCOPIC CHOLECYSTECTOMY WITH TWO INCISIONS

    PubMed Central

    ABAID, Rafael Antoniazzi; CECCONELLO, Ivan; ZILBERSTEIN, Bruno

    2014-01-01

    Background Laparoscopic cholecystectomy has traditionally been performed with four incisions to insert four trocars, in a simple, efficient and safe way. Aim To describe a simplified technique of laparoscopic cholecystectomy with two incisions, using basic conventional instrumental. Technique In one incision in the umbilicus are applied two trocars and in epigastrium one more. The use of two trocars on the same incision, working in "x" does not hinder the procedure and does not require special instruments. Conclusion Simplified laparoscopic cholecystectomy with two incisions is feasible and easy to perform, allowing to operate with ergonomy and safety, with good cosmetic result. PMID:25004296

  17. Stream of consciousness: Quantum and biochemical assumptions regarding psychopathology.

    PubMed

    Tonello, Lucio; Cocchi, Massimo; Gabrielli, Fabio; Tuszynski, Jack A

    2017-04-01

    The accepted paradigms of mainstream neuropsychiatry appear to be incompletely adequate and in various cases offer equivocal analyses. However, a growing number of new approaches are being proposed that suggest the emergence of paradigm shifts in this area. In particular, quantum theories of mind, brain and consciousness seem to offer a profound change to the current approaches. Unfortunately these quantum paradigms harbor at least two serious problems. First, they are simply models, theories, and assumptions, with no convincing experiments supporting their claims. Second, they deviate from contemporary mainstream views of psychiatric illness and do so in revolutionary ways. We suggest a possible way to integrate experimental neuroscience with quantum models in order to address outstanding issues in psychopathology. A key role is played by the phenomenon called the "stream of consciousness", which can be linked to the so-called "Gamma Synchrony" (GS), which is clearly demonstrated by EEG data. In our novel proposal, a unipolar depressed patient could be seen as a subject with an altered stream of consciousness. In particular, some clues suggest that depression is linked to an "increased power" stream of consciousness. It is additionally suggested that such an approach to depression might be extended to psychopathology in general with potential benefits to diagnostics and therapeutics in neuropsychiatry.

  18. Simplified Recipes for Day Care Centers.

    ERIC Educational Resources Information Center

    Asmussen, Patricia D.

    The spiral-bound collection of 156 simplified recipes is designed to help those who prepare food for groups of children at day care centers. The recipes provide for 25 child-size servings to meet the nutritional needs and appetites of children from 2 to 6 years of age. The first section gives general information on ladle and scoop sizes, weights…

  19. Simplified Fabrication of Helical Copper Antennas

    NASA Technical Reports Server (NTRS)

    Petro, Andrew

    2006-01-01

    A simplified technique has been devised for fabricating helical antennas for use in experiments on radio-frequency generation and acceleration of plasmas. These antennas are typically made of copper (for electrical conductivity) and must have a specific helical shape and precise diameter.

  20. Simplified procedures for designing composite bolted joints

    NASA Technical Reports Server (NTRS)

    Chamis, Christos C.

    1988-01-01

    Simplified procedures are described to design and analyze single and multi-bolt composite joints. Numerical examples illustrate the use of these methods. Factors affecting composite bolted joints are summarized. References are cited where more detailed discussion is presented on specific aspects of composite bolted joints. Design variables associated with these joints are summarized in the appendix.

  1. Simplified procedures for designing composite bolted joints

    NASA Technical Reports Server (NTRS)

    Chamis, Christos C.

    1990-01-01

    Simplified procedures are described to design and analyze single and multi-bolt composite joints. Numerical examples illustrate the use of these methods. Factors affecting composite bolted joints are summarized. References are cited where more detailed discussion is presented on specific aspects of composite bolted joints. Design variables associated with these joints are summarized in the appendix.

  2. Simplified modeling for infiltration and radon entry

    SciTech Connect

    Sherman, M.H.

    1992-08-01

    Air leakage in the envelopes of residential buildings is the primary mechanism for provided ventilation to those buildings. For radon the same mechanisms that drive the ventilation, drive the radon entry This paper attempts to provide a simplified physical model that can be used to understand the interactions between the building leakage distribution, the forces that drive infiltration and ventilation, and indoor radon concentrations, Combining both ventilation and entry modeling together allows an estimation of Radon concentration and exposure to be made and demonstrates how changes in the envelope or ventilation system would affect it. This paper will develop simplified modeling approaches for estimating both ventilation rate and radon entry rate based on the air tightness of the envelope and the driving forces. These approaches will use conventional leakage values (i.e. effective leakage area ) to quantify the air tightness and include natural and mechanical driving forces. This paper will introduce a simplified parameter, the Radon Leakage Area, that quantifies the resistance to radon entry. To be practical for dwellings, modeling of the occupant exposures to indoor pollutants must be simple to use and not require unreasonable input data. This paper presents the derivation of the simplified physical model, and applies that model to representative situations to explore the tendencies to be expected under different circumstances.

  3. Marking and Moderation in the UK: False Assumptions and Wasted Resources

    ERIC Educational Resources Information Center

    Bloxham, Sue

    2009-01-01

    This article challenges a number of assumptions underlying marking of student work in British universities. It argues that, in developing rigorous moderation procedures, we have created a huge burden for markers which adds little to accuracy and reliability but creates additional work for staff, constrains assessment choices and slows down…

  4. The Impact of Modeling Assumptions in Galactic Chemical Evolution Models

    NASA Astrophysics Data System (ADS)

    Côté, Benoit; O’Shea, Brian W.; Ritter, Christian; Herwig, Falk; Venn, Kim A.

    2017-02-01

    We use the OMEGA galactic chemical evolution code to investigate how the assumptions used for the treatment of galactic inflows and outflows impact numerical predictions. The goal is to determine how our capacity to reproduce the chemical evolution trends of a galaxy is affected by the choice of implementation used to include those physical processes. In pursuit of this goal, we experiment with three different prescriptions for galactic inflows and outflows and use OMEGA within a Markov Chain Monte Carlo code to recover the set of input parameters that best reproduces the chemical evolution of nine elements in the dwarf spheroidal galaxy Sculptor. This provides a consistent framework for comparing the best-fit solutions generated by our different models. Despite their different degrees of intended physical realism, we found that all three prescriptions can reproduce in an almost identical way the stellar abundance trends observed in Sculptor. This result supports the similar conclusions originally claimed by Romano & Starkenburg for Sculptor. While the three models have the same capacity to fit the data, the best values recovered for the parameters controlling the number of SNe Ia and the strength of galactic outflows, are substantially different and in fact mutually exclusive from one model to another. For the purpose of understanding how a galaxy evolves, we conclude that only reproducing the evolution of a limited number of elements is insufficient and can lead to misleading conclusions. More elements or additional constraints such as the Galaxy’s star-formation efficiency and the gas fraction are needed in order to break the degeneracy between the different modeling assumptions. Our results show that the successes and failures of chemical evolution models are predominantly driven by the input stellar yields, rather than by the complexity of the Galaxy model itself. Simple models such as OMEGA are therefore sufficient to test and validate stellar yields. OMEGA

  5. Simplifying CEA through Excel, VBA, and Subeq

    NASA Technical Reports Server (NTRS)

    Foster, Ryan

    2004-01-01

    Many people use compound equilibrium programs for very different reasons, varying from refrigerators to light bulbs to rockets. A commonly used equilibrium program is CEA. CEA can take various inputs such as pressure, temperature, and volume along with numerous reactants and run them through equilibrium equations to obtain valuable output information, including products formed and their relative amounts. A little over a year ago, Bonnie McBride created the program subeq with the goal to simplify the calling of CEA. Subeq was also designed to be called by other programs, including Excel, through the use of Visual Basic for Applications (VBA). The largest advantage of using Excel is that it allows the user to input the information in a colorful and user-friendly environment while allowing VBA to run subeq, which is in the form of a FORTRAN DLL (Dynamic Link Library). Calling subeq in this form makes it much faster than if it were converted to VBA. Since subeq requires such large lists of reactant and product names, all of which can't be passed in as an array, subeq had to be changed to accept very long strings of reactants and products. To pass this string and adjust the transfer of input and output parameters, the subeq DLL had to be changed. One program that does this is Compaq Visual FORTRAN, which allows DLLs to be edited, debugged, and compiled. Compaq Visual FORTRAN uses FORTRAN 90/95, which has additional features to that of FORTRAN 77. My goals this summer include finishing up the excel spreadsheet of subeq, which I started last summer, and putting it on the Internet so that others can use it without having to download my spreadsheet. To finish up the spreadsheet I will need to work on debugging current options and problems. I will also work on making it as robust as possible, so that all errors that may arise will be clearly communicated to the user. New features will be added old ones will be changed as I receive comments from people using the spreadsheet

  6. Sensitivity of Rooftop PV Projections in the SunShot Vision Study to Market Assumptions

    SciTech Connect

    Drury, E.; Denholm, P.; Margolis, R.

    2013-01-01

    The SunShot Vision Study explored the potential growth of solar markets if solar prices decreased by about 75% from 2010 to 2020. The SolarDS model was used to simulate rooftop PV demand for this study, based on several PV market assumptions--future electricity rates, customer access to financing, and others--in addition to the SunShot PV price projections. This paper finds that modeled PV demand is highly sensitive to several non-price market assumptions, particularly PV financing parameters.

  7. A VLSI architecture for simplified arithmetic Fourier transform algorithm

    NASA Technical Reports Server (NTRS)

    Reed, Irving S.; Shih, Ming-Tang; Truong, T. K.; Hendon, E.; Tufts, D. W.

    1992-01-01

    The arithmetic Fourier transform (AFT) is a number-theoretic approach to Fourier analysis which has been shown to perform competitively with the classical FFT in terms of accuracy, complexity, and speed. Theorems developed in a previous paper for the AFT algorithm are used here to derive the original AFT algorithm which Bruns found in 1903. This is shown to yield an algorithm of less complexity and of improved performance over certain recent AFT algorithms. A VLSI architecture is suggested for this simplified AFT algorithm. This architecture uses a butterfly structure which reduces the number of additions by 25 percent of that used in the direct method.

  8. Supporting calculations and assumptions for use in WESF safetyanalysis

    SciTech Connect

    Hey, B.E.

    1997-03-07

    This document provides a single location for calculations and assumptions used in support of Waste Encapsulation and Storage Facility (WESF) safety analyses. It also provides the technical details and bases necessary to justify the contained results.

  9. DATMS: A framework for distributed assumption based reasoning

    SciTech Connect

    Mason, C.L.; Johnson, R.R.

    1988-01-01

    The Distributed ATMS, DATMS, is a problem solving framework for multi-agent assumption based reasoning using Assumption-based Truth Maintenance Systems. It is based on the problem solving paradigm of result sharing rule-based expert systems using assumption based reasoning. We are implementing and experimenting with the DATMS under MATE, a Multi-Agent Test Environment, using C and COMMON-LISP on a network of SUN family workstations. This framework was motivated by the problem of seismic interpretation for Comprehensive or Low-Yield Test Ban Treaty verification, where a wide spread network of seismic sensor stations are required to monitor treaty compliance, and seismologists use assumption based reasoning in a collaborative fashion to interpret the seismic data. The DATMS framework differs from other previously designed problem solving organizations in its method of reasoning the ability to support an explanation facility and in addressing the problem of culpability. 13 refs.

  10. The Validity of Conventional Assumptions Concerning Flexible Response

    DTIC Science & Technology

    1989-01-01

    3AIR_ FWLE OOLEY RESEARCH REPORT c~J THE VALIDITY OF CONVENTIONAL ASSUMPTIONS ___ CONCERNING FLEXIBLE RESPONSE LT COL MARIE 3. GUTIERREZ DTIC...Validity of Conventional Assumptions Concerning Flexible Response AUTHOR: Marie J. Gutierrez, Lieutenant Colonel, USAF Thc North Atlantic Treaty...Europe since 1945. In 1967, NATO adopted the strategy of flexible response , a Etrategy dependent upon conventional, tactical nuclear, and stra- tegic

  11. US Intervention in Failed States: Bad Assumptions=Poor Outcomes

    DTIC Science & Technology

    2002-01-01

    NATIONAL DEFENSE UNIVERSITY NATIONAL WAR COLLEGE STRATEGIC LOGIC ESSAY US INTERVENTION IN FAILED STATES: BAD ASSUMPTIONS = POOR ...2002 2. REPORT TYPE 3. DATES COVERED 00-00-2002 to 00-00-2002 4. TITLE AND SUBTITLE US Intervention in Failed States: Bad Assumptions= Poor ...country remains in the grip of poverty , natural disasters, and stagnation. Rwanda Rwanda, another small African country, is populated principally

  12. Simplified models of mixed dark matter

    SciTech Connect

    Cheung, Clifford; Sanford, David E-mail: dsanford@caltech.edu

    2014-02-01

    We explore simplified models of mixed dark matter (DM), defined here to be a stable relic composed of a singlet and an electroweak charged state. Our setup describes a broad spectrum of thermal DM candidates that can naturally accommodate the observed DM abundance but are subject to substantial constraints from current and upcoming direct detection experiments. We identify ''blind spots'' at which the DM-Higgs coupling is identically zero, thus nullifying direct detection constraints on spin independent scattering. Furthermore, we characterize the fine-tuning in mixing angles, i.e. well-tempering, required for thermal freeze-out to accommodate the observed abundance. Present and projected limits from LUX and XENON1T force many thermal relic models into blind spot tuning, well-tempering, or both. This simplified model framework generalizes bino-Higgsino DM in the MSSM, singlino-Higgsino DM in the NMSSM, and scalar DM candidates that appear in models of extended Higgs sectors.

  13. Simplified analysis of complex conductivity data

    NASA Astrophysics Data System (ADS)

    Salomon, M.

    1981-03-01

    A simplified method is presented for the analysis of conductivity data for complex situations which involve ion pairing. The method was tested by comparing results with more elaborate analyses using published data for a number of salts in various solvents. The method is then used to analyze the conductivity data for several salts in thionyl chloride. These latter analyses are important for selecting electrolytes for lithium primary batteries.

  14. Simplified Learning Scheme For Analog Neural Network

    NASA Technical Reports Server (NTRS)

    Eberhardt, Silvio P.

    1991-01-01

    Synaptic connections adjusted one at a time in small increments. Simplified gradient-descent learning scheme for electronic neural-network processor less efficient than better-known back-propagation scheme, but offers two advantages: easily implemented in circuitry because data-access circuitry separated from learning circuitry; and independence of data-access circuitry makes possible to implement feedforward as well as feedback networks, including those of multiple-attractor type. Important in such applications as recognition of patterns.

  15. Simplified dichromated gelatin hologram recording process

    NASA Technical Reports Server (NTRS)

    Georgekutty, Tharayil G.; Liu, Hua-Kuang

    1987-01-01

    A simplified method for making dichromated gelatin (DCG) holographic optical elements (HOE) has been discovered. The method is much less tedious and it requires a period of processing time comparable with that for processing a silver halide hologram. HOE characteristics including diffraction efficiency (DE), linearity, and spectral sensitivity have been quantitatively investigated. The quality of the holographic grating is very high. Ninety percent or higher diffraction efficiency has been achieved in simple plane gratings made by this process.

  16. Simplified robot arm dynamics for control

    NASA Technical Reports Server (NTRS)

    Bejczy, A. K.; Paul, R. P.

    1981-01-01

    A brief summary and evaluation is presented on the use of symbolic state equation techniques in order to represent robot arm dynamics with sufficient accuracy for controlling arm motion. The use of homogeneous transformations and the Lagrangian formulation of mechanics offers a convenient frame for the derivation, analysis and simplification of complex robot dynamics equations. It is pointed out that simplified state equations can represent robot arm dynamics with good accuracy.

  17. A Simplified Inverse Approach for the Simulation of Axi-Symmetrical Cold Forging Process

    NASA Astrophysics Data System (ADS)

    Halouani, A.; Li, Y. M.; Abbès, B.; Guo, Y. Q.

    2011-01-01

    This paper presents the formulation of an axi-symmetric element based on an efficient method called "Inverse Approach" (I.A.) for the numerical modeling of cold forging process. In contrast to the classical incremental methods, the Inverse Approach exploits the known shape of the final part and executes the calculation from the final part to the initial billet. The assumptions of the proportional loading and the simplified tool actions make the I.A. calculation very fast. The metal's incompressibility is ensured by the penalty method. The comparison with ABAQUS® and FORGE® shows the efficiency and limitations of the I.A. This simplified method will be a good tool for the preliminary preform design.

  18. Two simplified procedures for predicting cyclic material response from a strain history

    NASA Technical Reports Server (NTRS)

    Kaufman, A.; Moreno, V.

    1985-01-01

    Simplified inelastic analysis procedures were developed at NASA Lewis and Pratt & Whitney Aircraft for predicting the stress-strain response at the critical location of a thermomechanically cycled structure. These procedures are intended primarily for use as economical structural analysis tools in the early design stages of aircraft engine hot section components where nonlinear finite-element analyses would be prohibitively expensive. Both simplified methods use as input the total strain history calculated from a linear elastic analysis. The elastic results are modified to approximate the characteristics of the inelastic cycle by incremental solution techniques. A von Mises yield criterion is used to determine the onset of active plasticity. The fundamental assumption of these methods is that the inelastic strain is local and constrained from redistribution by the surrounding elastic material.

  19. Simplified strategies based on damage mechanics for concrete under dynamic loading.

    PubMed

    Mazars, Jacky; Grange, Stéphane

    2017-01-28

    Based on previous work, the µ damage model has been designed to figure out the various damage effects in concrete correlated with monotonic and cyclic loading, including unilateral effects. Assumptions are formulated to simplify constitutive relationships while still allowing for a correct description of the main nonlinear effects. In this context, the paper presents an enhanced simplified finite-element description including a damage description, based on the use of multifibre beam elements and including strain rate effects. Applications show that such a strategy leads to an efficient tool to simulate dynamic loading at low, medium and high velocities.This article is part of the themed issue 'Experimental testing and modelling of brittle materials at high strain rates'.

  20. Simplified Modeling of Active Magnetic Regenerators

    NASA Astrophysics Data System (ADS)

    Burdyny, Thomas

    Active magnetic regenerator (AMR) refrigeration is an alternative technology to conventional vapor-compression refrigerators that has the potential to operate at higher efficiencies. Based on the magnetocaloric effect, this technology uses the magnetization and demagnetization of environmentally neutral solid refrigerants to produce a cooling effect. To become competitive however, a large amount of research into the optimal device configurations, operating parameters and refrigerants is still needed. To aid in this research, a simplified model for predicting the general trends of AMR devices at a low computational cost is developed. The derivation and implementation of the model for an arbitrary AMR is presented. Simulations from the model are compared to experimental results from two different devices and show good agreement across a wide range of operating parameters. The simplified model is also used to study the impacts of Curie temperature spacing, material weighting and devices on the performance of multilayered regenerators. Future applications of the simplified AMR model include costing and optimization programs where the low computational demand of the model can be fully exploited.

  1. The steady-state assumption in oscillating and growing systems.

    PubMed

    Reimers, Alexandra-M; Reimers, Arne C

    2016-10-07

    The steady-state assumption, which states that the production and consumption of metabolites inside the cell are balanced, is one of the key aspects that makes an efficient analysis of genome-scale metabolic networks possible. It can be motivated from two different perspectives. In the time-scales perspective, we use the fact that metabolism is much faster than other cellular processes such as gene expression. Hence, the steady-state assumption is derived as a quasi-steady-state approximation of the metabolism that adapts to the changing cellular conditions. In this article we focus on the second perspective, stating that on the long run no metabolite can accumulate or deplete. In contrast to the first perspective it is not immediately clear how this perspective can be captured mathematically and what assumptions are required to obtain the steady-state condition. By presenting a mathematical framework based on the second perspective we demonstrate that the assumption of steady-state also applies to oscillating and growing systems without requiring quasi-steady-state at any time point. However, we also show that the average concentrations may not be compatible with the average fluxes. In summary, we establish a mathematical foundation for the steady-state assumption for long time periods that justifies its successful use in many applications. Furthermore, this mathematical foundation also pinpoints unintuitive effects in the integration of metabolite concentrations using nonlinear constraints into steady-state models for long time periods.

  2. Simplified Analysis of Pulse Detonation Rocket Engine Blowdown Gasdynamics and Performance

    NASA Technical Reports Server (NTRS)

    Morris, C. I.; Rodgers, Stephen L. (Technical Monitor)

    2002-01-01

    Pulse detonation rocket engines (PDREs) offer potential performance improvements over conventional designs, but represent a challenging modellng task. A simplified model for an idealized, straight-tube, single-shot PDRE blowdown process and thrust determination is described and implemented. In order to form an assessment of the accuracy of the model, the flowfield time history is compared to experimental data from Stanford University. Parametric Studies of the effect of mixture stoichiometry, initial fill temperature, and blowdown pressure ratio on the performance of a PDRE are performed using the model. PDRE performance is also compared with a conventional steady-state rocket engine over a range of pressure ratios using similar gasdynamic assumptions.

  3. Provably-Secure (Chinese Government) SM2 and Simplified SM2 Key Exchange Protocols

    PubMed Central

    Nam, Junghyun; Kim, Moonseong

    2014-01-01

    We revisit the SM2 protocol, which is widely used in Chinese commercial applications and by Chinese government agencies. Although it is by now standard practice for protocol designers to provide security proofs in widely accepted security models in order to assure protocol implementers of their security properties, the SM2 protocol does not have a proof of security. In this paper, we prove the security of the SM2 protocol in the widely accepted indistinguishability-based Bellare-Rogaway model under the elliptic curve discrete logarithm problem (ECDLP) assumption. We also present a simplified and more efficient version of the SM2 protocol with an accompanying security proof. PMID:25276863

  4. Evolution of Requirements and Assumptions for Future Exploration Missions

    NASA Technical Reports Server (NTRS)

    Anderson, Molly; Sargusingh, Miriam; Perry, Jay

    2017-01-01

    NASA programs are maturing technologies, systems, and architectures to enabling future exploration missions. To increase fidelity as technologies mature, developers must make assumptions that represent the requirements of a future program. Multiple efforts have begun to define these requirements, including team internal assumptions, planning system integration for early demonstrations, and discussions between international partners planning future collaborations. For many detailed life support system requirements, existing NASA documents set limits of acceptable values, but a future vehicle may be constrained in other ways, and select a limited range of conditions. Other requirements are effectively set by interfaces or operations, and may be different for the same technology depending on whether the hard-ware is a demonstration system on the International Space Station, or a critical component of a future vehicle. This paper highlights key assumptions representing potential life support requirements and explanations of the driving scenarios, constraints, or other issues that drive them.

  5. A Simplified Model of Choice Behavior under Uncertainty

    PubMed Central

    Lin, Ching-Hung; Lin, Yu-Kai; Song, Tzu-Jiun; Huang, Jong-Tsun; Chiu, Yao-Chu

    2016-01-01

    The Iowa Gambling Task (IGT) has been standardized as a clinical assessment tool (Bechara, 2007). Nonetheless, numerous research groups have attempted to modify IGT models to optimize parameters for predicting the choice behavior of normal controls and patients. A decade ago, most researchers considered the expected utility (EU) model (Busemeyer and Stout, 2002) to be the optimal model for predicting choice behavior under uncertainty. However, in recent years, studies have demonstrated that models with the prospect utility (PU) function are more effective than the EU models in the IGT (Ahn et al., 2008). Nevertheless, after some preliminary tests based on our behavioral dataset and modeling, it was determined that the Ahn et al. (2008) PU model is not optimal due to some incompatible results. This study aims to modify the Ahn et al. (2008) PU model to a simplified model and used the IGT performance of 145 subjects as the benchmark data for comparison. In our simplified PU model, the best goodness-of-fit was found mostly as the value of α approached zero. More specifically, we retested the key parameters α, λ, and A in the PU model. Notably, the influence of the parameters α, λ, and A has a hierarchical power structure in terms of manipulating the goodness-of-fit in the PU model. Additionally, we found that the parameters λ and A may be ineffective when the parameter α is close to zero in the PU model. The present simplified model demonstrated that decision makers mostly adopted the strategy of gain-stay loss-shift rather than foreseeing the long-term outcome. However, there are other behavioral variables that are not well revealed under these dynamic-uncertainty situations. Therefore, the optimal behavioral models may not have been found yet. In short, the best model for predicting choice behavior under dynamic-uncertainty situations should be further evaluated. PMID:27582715

  6. A Simplified Model of Choice Behavior under Uncertainty.

    PubMed

    Lin, Ching-Hung; Lin, Yu-Kai; Song, Tzu-Jiun; Huang, Jong-Tsun; Chiu, Yao-Chu

    2016-01-01

    The Iowa Gambling Task (IGT) has been standardized as a clinical assessment tool (Bechara, 2007). Nonetheless, numerous research groups have attempted to modify IGT models to optimize parameters for predicting the choice behavior of normal controls and patients. A decade ago, most researchers considered the expected utility (EU) model (Busemeyer and Stout, 2002) to be the optimal model for predicting choice behavior under uncertainty. However, in recent years, studies have demonstrated that models with the prospect utility (PU) function are more effective than the EU models in the IGT (Ahn et al., 2008). Nevertheless, after some preliminary tests based on our behavioral dataset and modeling, it was determined that the Ahn et al. (2008) PU model is not optimal due to some incompatible results. This study aims to modify the Ahn et al. (2008) PU model to a simplified model and used the IGT performance of 145 subjects as the benchmark data for comparison. In our simplified PU model, the best goodness-of-fit was found mostly as the value of α approached zero. More specifically, we retested the key parameters α, λ, and A in the PU model. Notably, the influence of the parameters α, λ, and A has a hierarchical power structure in terms of manipulating the goodness-of-fit in the PU model. Additionally, we found that the parameters λ and A may be ineffective when the parameter α is close to zero in the PU model. The present simplified model demonstrated that decision makers mostly adopted the strategy of gain-stay loss-shift rather than foreseeing the long-term outcome. However, there are other behavioral variables that are not well revealed under these dynamic-uncertainty situations. Therefore, the optimal behavioral models may not have been found yet. In short, the best model for predicting choice behavior under dynamic-uncertainty situations should be further evaluated.

  7. Pullback attractors of the two-dimensional non-autonomous simplified Ericksen-Leslie system for nematic liquid crystal flows

    NASA Astrophysics Data System (ADS)

    You, Bo; Li, Fang

    2016-08-01

    This paper is concerned with the long-time behaviour of the two-dimensional non-autonomous simplified Ericksen-Leslie system for nematic liquid crystal flows introduced in Lin and Liu (Commun Pure Appl Math, 48:501-537, 1995) with a non-autonomous forcing bulk term and order parameter field boundary conditions. In this paper, we prove the existence of pullback attractors and estimate the upper bound of its fractal dimension under some suitable assumptions.

  8. Effect of Selected Modeling Assumptions on Subsurface Radionuclide Transport Projections for the Potential Environmental Management Disposal Facility at Oak Ridge, Tennessee

    SciTech Connect

    Painter, Scott L.

    2016-06-28

    The Department of Energy’s Office of Environmental Management recently revised a Remedial Investigation/ Feasibility Study (RI/FS) that included an analysis of subsurface radionuclide transport at a potential new Environmental Management Disposal Facility (EMDF) in East Bear Creek Valley near Oak Ridge, Tennessee. The effect of three simplifying assumptions used in the RI/FS analyses are investigated using the same subsurface pathway conceptualization but with more flexible modeling tools. Neglect of vadose zone dispersion was found to be conservative or non-conservative, depending on the retarded travel time and the half-life. For a given equilibrium distribution coefficient, a relatively narrow range of half-life was identified for which neglect of vadose zone transport is non-conservative and radionuclide discharge into surface water is non-negligible. However, there are two additional conservative simplifications in the reference case that compensate for the non-conservative effect of neglecting vadose zone dispersion: the use of a steady infiltration rate and vadose zone velocity, and the way equilibrium sorption is used to represent transport in the fractured material of the saturated aquifer. With more realistic representations of all three processes, the RI/FS reference case was found to either provide a reasonably good approximation to the peak concentration or was significantly conservative (pessimistic) for all parameter combinations considered.

  9. Simplified solution of diffraction from a Lyot system

    NASA Astrophysics Data System (ADS)

    Wang, Yaujen; Vaughan, Arthur H.

    1988-01-01

    This paper presents a derivation of a simplified analytical solution of diffraction from a Lyot (1939) system designed for observation of the solar corona outside of eclipses. Applying the theorem of Papoulis (1986) to simplify the calculations, a simplified solution is derived which is found to be in reasonable agreement with the exact solution. The simplified solution suffices for the preliminary evaluation of the amount of diffraction reduction needed to meet certain system requirements and also serves as a guideline for further apodization.

  10. Food additives

    PubMed Central

    Spencer, Michael

    1974-01-01

    Food additives are discussed from the food technology point of view. The reasons for their use are summarized: (1) to protect food from chemical and microbiological attack; (2) to even out seasonal supplies; (3) to improve their eating quality; (4) to improve their nutritional value. The various types of food additives are considered, e.g. colours, flavours, emulsifiers, bread and flour additives, preservatives, and nutritional additives. The paper concludes with consideration of those circumstances in which the use of additives is (a) justified and (b) unjustified. PMID:4467857

  11. The electrostatic interaction of an external charged system with a metal surface: a simplified density functional theory approach

    NASA Astrophysics Data System (ADS)

    Scivetti, Iván; Persson, Mats

    2013-09-01

    As a first step to meet the challenge to calculate the electronic structure and total energy of charged states of atoms and molecules adsorbed on ultrathin insulating films supported by a metallic substrate using density functional theory (DFT), we have developed a simplified new DFT scheme that only describes the electrostatic interaction of an external charged system with a metal surface. This purely electrostatic interaction is obtained from the assumption that the electron densities of the two fragments (charged system and metal surface) are non-overlapping and by neglecting non-local exchange-correlation effects such as the van der Waals interactions between the two fragments. In addition, the response of the metal surface to the electrostatic potential from the charged system is treated to linear order, whereas the charged system is treated fully within DFT. In particular, we consider the classical perfect conductor model for the metal response, although our formalism is not limited to this approximation. To test the computational implementation of this new scheme, we have considered the case of a Na+ cation interacting with a perfect conductor. The application of this new methodology to realistic problems involving charged systems adsorbed on insulating films supported by a metal surface are deferred to a separate following publication.

  12. 29 CFR 2700.104 - Discontinuance of simplified proceedings.

    Code of Federal Regulations, 2012 CFR

    2012-07-01

    ...) Procedure. If it becomes apparent at any time that a case is not appropriate for Simplified Proceedings, the... Simplified Proceedings and order the case to continue under conventional rules. (b) Party motion. At any time... Simplified Proceedings be discontinued and that the matter continue under conventional procedures. A...

  13. 29 CFR 2700.104 - Discontinuance of simplified proceedings.

    Code of Federal Regulations, 2013 CFR

    2013-07-01

    ...) Procedure. If it becomes apparent at any time that a case is not appropriate for Simplified Proceedings, the... Simplified Proceedings and order the case to continue under conventional rules. (b) Party motion. At any time... Simplified Proceedings be discontinued and that the matter continue under conventional procedures. A...

  14. Chronic Meningitis: Simplifying a Diagnostic Challenge.

    PubMed

    Baldwin, Kelly; Whiting, Chris

    2016-03-01

    Chronic meningitis can be a diagnostic dilemma for even the most experienced clinician. Many times, the differential diagnosis is broad and encompasses autoimmune, neoplastic, and infectious etiologies. This review will focus on a general approach to chronic meningitis to simplify the diagnostic challenges many clinicians face. The article will also review the most common etiologies of chronic meningitis in some detail including clinical presentation, diagnostic testing, treatment, and outcomes. By using a case-based approach, we will focus on the key elements of clinical presentation and laboratory analysis that will yield the most rapid and accurate diagnosis in these complicated cases.

  15. Simplified Explosive Joining of Tubes to Fittings

    NASA Technical Reports Server (NTRS)

    Bement, L. J.; Bailey, J. W.; Perry, R.; Finch, M. S.

    1987-01-01

    Technique simplifies tube-to-fitting joining, as compared to fusion welding, and provides improvement on standard procedures used to join tubes explosively to tube fittings. Special tool inserted into tube to be joined. Tool allows strip of ribbon explosive to be placed right at joint. Ribbon explosive and mild detonating fuse allows use of smaller charge. Assembled tool storable, and process amenable to automation. Assembly of components, insertion of tool into weld site, and joining operation mechanized without human contact. Used to assemble components in nuclear reactors or in other environments hostile to humans.

  16. Simplified dynamic buckling assessment of steel containments

    SciTech Connect

    Farrar, C.R.; Duffey, T.A.; Renick, D.H.

    1993-02-01

    A simplified, three-degree-of-freedom analytical procedure for performing a response spectrum buckling analysis of a thin containment shell is developed. Two numerical examples with R/t values which bound many existing steel containments are used to illustrate the procedure. The role of damping on incipient buckling acceleration level is evaluated for a regulatory seismic spectrum using the two numerical examples. The zero-period acceleration level that causes incipient buckling in either of the two containments increases 31% when damping is increased from 1% to 4% of critical. Comparisons with finite element results on incipient buckling levels are favorable.

  17. Simplified definition system: magnetic products fabrication

    SciTech Connect

    Morris, R.S.

    1981-06-01

    The Simplified Definition System, a product definition approach that differentiates between design and production agency manufacturing requirements, has been used in producing 50 types of magnetic products. This system was formed as a result of cooperative work and proposed modifications by engineers from Bendix Kansas City and Sandia National Laboratories Albuquerque (SNLA) to reduce product costs. The system places responsibility for production-related requirements with a production agency, a procedure that has realized both direct and indirect cost savings. This report is a documentation of the system's description and projected savings on magnetic products.

  18. Hyperbolic Structure for a Simplified Model of Dynamical Perfect Plasticity

    NASA Astrophysics Data System (ADS)

    Babadjian, Jean-François; Mifsud, Clément

    2017-02-01

    This paper is devoted to confronting two different approaches to the problem of dynamical perfect plasticity. Interpreting this model as a constrained boundary value Friedrichs' system enables one to derive admissible hyperbolic boundary conditions. Using variational methods, we show the well-posedness of this problem in a suitable weak measure theoretical setting. Thanks to the property of finite speed propagation, we establish a new regularity result for the solution in short time. Finally, we prove that this variational solution is actually a solution of the hyperbolic formulation in a suitable dissipative/entropic sense, and that a partial converse statement holds under an additional time regularity assumption for the dissipative solutions.

  19. Teacher Incentives: Goals, Strategies, Assumptions and Their Feasibility.

    ERIC Educational Resources Information Center

    Duttweiler, Patricia C.

    This presentation discusses strategies upon which to base teacher incentives, and also assumptions within these strategies. The first strategy, influencing teacher turnover rates so that good teachers remain in the profession longer and poor teachers leave teaching sooner, includes consideration of occupational rewards, career structure, benefits,…

  20. Viruses, Murphy's Law, and the Dangers of Assumptions....

    ERIC Educational Resources Information Center

    Lester, Dan

    1999-01-01

    An experienced library technology manager relates what happened in the wake of a serious library computer virus attack, which he accidentally unleashed. The narrative describes the combination of coincidences, mistakes, assumptions, and delays that caused the incident, and outlines the 10 key lessons learned. (AEF)

  1. Information and Democracy: An Examination of Underlying Assumptions.

    ERIC Educational Resources Information Center

    Dervin, Brenda

    1994-01-01

    Examines the relationship between information and democracy through a framework of six stereotyped assumptions regarding ontology and epistemology, i.e., dogma, naturalism, cultural relativity, constructivism, postmodernism, and communitarianism. Implications for the design and implementation of information systems are discussed. (Contains 93…

  2. The Metatheoretical Assumptions of Literacy Engagement: A Preliminary Centennial History

    ERIC Educational Resources Information Center

    Hruby, George G.; Burns, Leslie D.; Botzakis, Stergios; Groenke, Susan L.; Hall, Leigh A.; Laughter, Judson; Allington, Richard L.

    2016-01-01

    In this review of literacy education research in North America over the past century, the authors examined the historical succession of theoretical frameworks on students' active participation in their own literacy learning, and in particular the metatheoretical assumptions that justify those frameworks. The authors used "motivation" and…

  3. Assumptions Underlying the Identification of Gifted and Talented Students

    ERIC Educational Resources Information Center

    Brown, Scott W.; Renzuli, Joseph S.; Gubbins, E. Jean; Siegle, Del; Zhang, Wanli; Chen, Ching-Hui

    2005-01-01

    This study examined a national sample of classroom teachers, teachers of the gifted, administrators, and consultants from rural, suburban, and urban areas regarding their assumptions about the gifted identification process. Respondents indicated the degree to which they agreed or disagreed with 20 items that reflected guidelines for a…

  4. The Ubiquitous Laplacian Assumption: Reply to Lee and Wagenmakers (2005)

    ERIC Educational Resources Information Center

    Trafimow, David

    2005-01-01

    In their comment on D. Trafimow, M. D. Lee and E. Wagenmakers argued that the requisite probabilities to use in Bayes's theorem can always be found. In the present reply, the author asserts that M. D. Lee and E. Wagenmakers use a problematic assumption and that finding the requisite probabilities is not straightforward. After describing the…

  5. Seven Assumptions of a Solution-Focused Conversational Leader.

    ERIC Educational Resources Information Center

    Paull, Robert C.; McGrevin, Carol Z.

    1996-01-01

    Effective psychologists and school leaders know how to manage conversations to help clients or stakeholders move toward solutions. This article presents the assumptions of solution-focused brief therapy in a school leadership context. Key components are focusing on solutions, finding exceptions, identifying changes, starting small, listening to…

  6. 7 CFR 1779.88 - Transfers and assumptions.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... at least equal to either the present market value or the debt, whichever is less. (b) Transfers to an... with prior Agency written concurrence and only when the value of the collateral being transferred is at... 7 Agriculture 12 2011-01-01 2011-01-01 false Transfers and assumptions. 1779.88 Section...

  7. Challenging Educational Assumptions: Lessons from an Italian-American Collaboration.

    ERIC Educational Resources Information Center

    Krechevsky, Mara; Stork, Janet

    2000-01-01

    Argues that four key assumptions about teaching and learning must be reevaluated to create a curriculum of the future: (1) learning is the result of individual activity; (2) teachers are consumers of theory; (3) assessment is concerned with evaluating learning outcomes; and (4) teaching and learning are primarily cognitive, ethical, or affective…

  8. 7 CFR 4287.134 - Transfer and assumption.

    Code of Federal Regulations, 2012 CFR

    2012-01-01

    ... for the appraisal. If the transfer is for less than the debt, the lender must demonstrate to the Agency that the transferor and guarantors have no reasonable debt-paying ability considering their assets... before a transfer and assumption will be credited to the transferor's guaranteed loan debt in...

  9. 7 CFR 4287.134 - Transfer and assumption.

    Code of Federal Regulations, 2013 CFR

    2013-01-01

    ... for the appraisal. If the transfer is for less than the debt, the lender must demonstrate to the Agency that the transferor and guarantors have no reasonable debt-paying ability considering their assets... before a transfer and assumption will be credited to the transferor's guaranteed loan debt in...

  10. 7 CFR 4287.134 - Transfer and assumption.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... for the appraisal. If the transfer is for less than the debt, the lender must demonstrate to the Agency that the transferor and guarantors have no reasonable debt-paying ability considering their assets... before a transfer and assumption will be credited to the transferor's guaranteed loan debt in...

  11. 7 CFR 4287.134 - Transfer and assumption.

    Code of Federal Regulations, 2014 CFR

    2014-01-01

    ... for the appraisal. If the transfer is for less than the debt, the lender must demonstrate to the Agency that the transferor and guarantors have no reasonable debt-paying ability considering their assets... before a transfer and assumption will be credited to the transferor's guaranteed loan debt in...

  12. 40 CFR 144.66 - State assumption of responsibility.

    Code of Federal Regulations, 2014 CFR

    2014-07-01

    .... 144.66 Section 144.66 Protection of Environment ENVIRONMENTAL PROTECTION AGENCY (CONTINUED) WATER PROGRAMS (CONTINUED) UNDERGROUND INJECTION CONTROL PROGRAM Financial Responsibility: Class I Hazardous Waste Injection Wells § 144.66 State assumption of responsibility. (a) If a State either assumes...

  13. 40 CFR 144.66 - State assumption of responsibility.

    Code of Federal Regulations, 2011 CFR

    2011-07-01

    .... 144.66 Section 144.66 Protection of Environment ENVIRONMENTAL PROTECTION AGENCY (CONTINUED) WATER PROGRAMS (CONTINUED) UNDERGROUND INJECTION CONTROL PROGRAM Financial Responsibility: Class I Hazardous Waste Injection Wells § 144.66 State assumption of responsibility. (a) If a State either assumes...

  14. 40 CFR 144.66 - State assumption of responsibility.

    Code of Federal Regulations, 2013 CFR

    2013-07-01

    .... 144.66 Section 144.66 Protection of Environment ENVIRONMENTAL PROTECTION AGENCY (CONTINUED) WATER PROGRAMS (CONTINUED) UNDERGROUND INJECTION CONTROL PROGRAM Financial Responsibility: Class I Hazardous Waste Injection Wells § 144.66 State assumption of responsibility. (a) If a State either assumes...

  15. 40 CFR 144.66 - State assumption of responsibility.

    Code of Federal Regulations, 2012 CFR

    2012-07-01

    .... 144.66 Section 144.66 Protection of Environment ENVIRONMENTAL PROTECTION AGENCY (CONTINUED) WATER PROGRAMS (CONTINUED) UNDERGROUND INJECTION CONTROL PROGRAM Financial Responsibility: Class I Hazardous Waste Injection Wells § 144.66 State assumption of responsibility. (a) If a State either assumes...

  16. 40 CFR 144.66 - State assumption of responsibility.

    Code of Federal Regulations, 2010 CFR

    2010-07-01

    .... 144.66 Section 144.66 Protection of Environment ENVIRONMENTAL PROTECTION AGENCY (CONTINUED) WATER PROGRAMS (CONTINUED) UNDERGROUND INJECTION CONTROL PROGRAM Financial Responsibility: Class I Hazardous Waste Injection Wells § 144.66 State assumption of responsibility. (a) If a State either assumes...

  17. 41 CFR 60-3.9 - No assumption of validity.

    Code of Federal Regulations, 2010 CFR

    2010-07-01

    ... 41 Public Contracts and Property Management 1 2010-07-01 2010-07-01 true No assumption of validity. 60-3.9 Section 60-3.9 Public Contracts and Property Management Other Provisions Relating to Public... 3-UNIFORM GUIDELINES ON EMPLOYEE SELECTION PROCEDURES (1978) General Principles § 60-3.9...

  18. Checking the Assumptions of Rasch's Model for Speed Tests

    ERIC Educational Resources Information Center

    Jansen, M. G. H.; Glas, C. A. W.

    2005-01-01

    Two new tests for a model for the response times on pure speed tests by Rasch (1960) are proposed. The model is based on the assumption that the test response times are approximately gamma distributed, with known index parameters and unknown rate parameters. The rate parameters are decomposed in a subject ability parameter and a test difficulty…

  19. 10 CFR 71.83 - Assumptions as to unknown properties.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... 10 Energy 2 2010-01-01 2010-01-01 false Assumptions as to unknown properties. 71.83 Section 71.83 Energy NUCLEAR REGULATORY COMMISSION (CONTINUED) PACKAGING AND TRANSPORTATION OF RADIOACTIVE MATERIAL... fissile material in any package is not known, the licensee shall package the fissile material as if...

  20. Transformation in Reverse: Naive Assumptions of an Urban Educator

    ERIC Educational Resources Information Center

    Hagiwara, Sumi; Wray, Susan

    2009-01-01

    The complexity of urban contexts is often subsumed into generalizations and deficit assumptions of urban communities and its members by those unfamiliar with urban culture. This is especially true for teachers seeking work in urban schools. This article addresses the complex interpretations of urban through the lens of a White male graduate…

  1. 14 CFR 25.473 - Landing load conditions and assumptions.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... and landing gear loads must take into account at least the following elements: (1) Landing gear... response of the airframe, if significant. (d) The landing gear dynamic characteristics must be validated by... 14 Aeronautics and Space 1 2011-01-01 2011-01-01 false Landing load conditions and assumptions....

  2. 46 CFR 174.070 - General damage stability assumptions.

    Code of Federal Regulations, 2010 CFR

    2010-10-01

    ... 46 Shipping 7 2010-10-01 2010-10-01 false General damage stability assumptions. 174.070 Section 174.070 Shipping COAST GUARD, DEPARTMENT OF HOMELAND SECURITY (CONTINUED) SUBDIVISION AND STABILITY SPECIAL RULES PERTAINING TO SPECIFIC VESSEL TYPES Special Rules Pertaining to Mobile Offshore...

  3. 46 CFR 174.070 - General damage stability assumptions.

    Code of Federal Regulations, 2013 CFR

    2013-10-01

    ... 46 Shipping 7 2013-10-01 2013-10-01 false General damage stability assumptions. 174.070 Section 174.070 Shipping COAST GUARD, DEPARTMENT OF HOMELAND SECURITY (CONTINUED) SUBDIVISION AND STABILITY SPECIAL RULES PERTAINING TO SPECIFIC VESSEL TYPES Special Rules Pertaining to Mobile Offshore...

  4. Making Predictions about Chemical Reactivity: Assumptions and Heuristics

    ERIC Educational Resources Information Center

    Maeyer, Jenine; Talanquer, Vicente

    2013-01-01

    Diverse implicit cognitive elements seem to support but also constrain reasoning in different domains. Many of these cognitive constraints can be thought of as either implicit assumptions about the nature of things or reasoning heuristics for decision-making. In this study we applied this framework to investigate college students' understanding of…

  5. 40 CFR 261.150 - State assumption of responsibility.

    Code of Federal Regulations, 2010 CFR

    2010-07-01

    ... 40 Protection of Environment 25 2010-07-01 2010-07-01 false State assumption of responsibility. 261.150 Section 261.150 Protection of Environment ENVIRONMENTAL PROTECTION AGENCY (CONTINUED) SOLID WASTES (CONTINUED) IDENTIFICATION AND LISTING OF HAZARDOUS WASTE Financial Requirements for Management...

  6. Challenging Old Assumptions: Preparing Teachers for Inner City Schools.

    ERIC Educational Resources Information Center

    Olmedo, Irma M.

    1997-01-01

    Researchers analyzed journals and essays from an elementary teacher education course, examining white prospective teachers' changing views about inner-city schools with minority children as they completed fieldwork and relevant readings. The experiences helped them question old assumptions about urban students and teaching and about the value of…

  7. 7 CFR 3555.256 - Transfer and assumptions.

    Code of Federal Regulations, 2014 CFR

    2014-01-01

    ... 7 Agriculture 15 2014-01-01 2014-01-01 false Transfer and assumptions. 3555.256 Section 3555.256 Agriculture Regulations of the Department of Agriculture (Continued) RURAL HOUSING SERVICE, DEPARTMENT OF AGRICULTURE GUARANTEED RURAL HOUSING PROGRAM (Eff. 9-1-14) Servicing Performing Loans § 3555.256 Transfer...

  8. Evaluation of assumptions in soil moisture triple collocation analysis

    Technology Transfer Automated Retrieval System (TEKTRAN)

    Triple collocation analysis (TCA) enables estimation of error variances for three or more products that retrieve or estimate the same geophysical variable using mutually-independent methods. Several statistical assumptions regarding the statistical nature of errors (e.g., mutual independence and ort...

  9. Qualifications and Assignments of Alternatively Certified Teachers: Testing Core Assumptions

    ERIC Educational Resources Information Center

    Cohen-Vogel, Lora; Smith, Thomas M.

    2007-01-01

    By analyzing data from the Schools and Staffing Survey, the authors empirically test four of the core assumptions embedded in current arguments for expanding alternative teacher certification (AC): AC attracts experienced candidates from fields outside of education; AC attracts top-quality, well-trained teachers; AC disproportionately trains…

  10. 36 CFR 254.7 - Assumption of costs.

    Code of Federal Regulations, 2010 CFR

    2010-07-01

    ... 36 Parks, Forests, and Public Property 2 2010-07-01 2010-07-01 false Assumption of costs. 254.7 Section 254.7 Parks, Forests, and Public Property FOREST SERVICE, DEPARTMENT OF AGRICULTURE LANDOWNERSHIP... species habitat; (4) Cash equalization funds are available for compensation of the non-Federal party;...

  11. 36 CFR 254.7 - Assumption of costs.

    Code of Federal Regulations, 2013 CFR

    2013-07-01

    ... 36 Parks, Forests, and Public Property 2 2013-07-01 2013-07-01 false Assumption of costs. 254.7 Section 254.7 Parks, Forests, and Public Property FOREST SERVICE, DEPARTMENT OF AGRICULTURE LANDOWNERSHIP... species habitat; (4) Cash equalization funds are available for compensation of the non-Federal party;...

  12. 36 CFR 254.7 - Assumption of costs.

    Code of Federal Regulations, 2012 CFR

    2012-07-01

    ... 36 Parks, Forests, and Public Property 2 2012-07-01 2012-07-01 false Assumption of costs. 254.7 Section 254.7 Parks, Forests, and Public Property FOREST SERVICE, DEPARTMENT OF AGRICULTURE LANDOWNERSHIP... species habitat; (4) Cash equalization funds are available for compensation of the non-Federal party;...

  13. 14 CFR 27.473 - Ground loading conditions and assumptions.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... 14 Aeronautics and Space 1 2010-01-01 2010-01-01 false Ground loading conditions and assumptions. 27.473 Section 27.473 Aeronautics and Space FEDERAL AVIATION ADMINISTRATION, DEPARTMENT OF TRANSPORTATION AIRCRAFT AIRWORTHINESS STANDARDS: NORMAL CATEGORY ROTORCRAFT Strength Requirements Ground...

  14. Impact of one-layer assumption on diffuse reflectance spectroscopy of skin

    NASA Astrophysics Data System (ADS)

    Hennessy, Ricky; Markey, Mia K.; Tunnell, James W.

    2015-02-01

    Diffuse reflectance spectroscopy (DRS) can be used to noninvasively measure skin properties. To extract skin properties from DRS spectra, you need a model that relates the reflectance to the tissue properties. Most models are based on the assumption that skin is homogenous. In reality, skin is composed of multiple layers, and the homogeneity assumption can lead to errors. In this study, we analyze the errors caused by the homogeneity assumption. This is accomplished by creating realistic skin spectra using a computational model, then extracting properties from those spectra using a one-layer model. The extracted parameters are then compared to the parameters used to create the modeled spectra. We used a wavelength range of 400 to 750 nm and a source detector separation of 250 μm. Our results show that use of a one-layer skin model causes underestimation of hemoglobin concentration [Hb] and melanin concentration [mel]. Additionally, the magnitude of the error is dependent on epidermal thickness. The one-layer assumption also causes [Hb] and [mel] to be correlated. Oxygen saturation is overestimated when it is below 50% and underestimated when it is above 50%. We also found that the vessel radius factor used to account for pigment packaging is correlated with epidermal thickness.

  15. Impact of one-layer assumption on diffuse reflectance spectroscopy of skin.

    PubMed

    Hennessy, Ricky; Markey, Mia K; Tunnell, James W

    2015-02-01

    Diffuse reflectance spectroscopy (DRS) can be used to noninvasively measure skin properties. To extract skin properties from DRS spectra, you need a model that relates the reflectance to the tissue properties. Most models are based on the assumption that skin is homogenous. In reality, skin is composed of multiple layers, and the homogeneity assumption can lead to errors. In this study, we analyze the errors caused by the homogeneity assumption. This is accomplished by creating realistic skin spectra using a computational model, then extracting properties from those spectra using a one-layer model. The extracted parameters are then compared to the parameters used to create the modeled spectra. We used a wavelength range of 400 to 750 nm and a source detector separation of 250 μm. Our results show that use of a one-layer skin model causes underestimation of hemoglobin concentration [Hb] and melanin concentration [mel]. Additionally, the magnitude of the error is dependent on epidermal thickness. The one-layer assumption also causes [Hb] and [mel] to be correlated. Oxygen saturation is overestimated when it is below 50% and underestimated when it is above 50%. We also found that the vessel radius factor used to account for pigment packaging is correlated with epidermal thickness.

  16. A Simplified Scheme for Kinematic Source Inversion

    NASA Astrophysics Data System (ADS)

    Iglesias, A.; Castro-Artola, O.; Singh, S.; Hjorleifsdottir, V.; Legrand, D.

    2013-05-01

    It is well known that different kinematic source inversion schemes lead to non-unique solutions. For this reason, a simplified scheme, which yields the main characteristics of the rupture process, rather than the details, may be desirable. In this work we propose a modification of the frequency-domain inversion scheme of Cotton & Campillo (1995) to extract kinematic parameters using simplified geometries (ellipses). The forward problem is re-parameterized by including one or two ellipses in which the displacement is smoothly distributed. For the ellipses we invert for the position of the centers within the fault plane, the major and minor semi-axes, the maximum displacements, the angles of rotation and a parameter that controls the distribution of slip. A simulated annealing scheme is used to invert near-source displacements. We first test the method on synthetic displacement records corresponding to the Guerrero-Oaxaca earthquake (20/03/2012, Mw=7.5) by comparing the results obtained from the modified technique with the original method. In the next step, we use displacements obtained by double numerical integration of recorded accelerograms. We find that, in spite of the simple geometry, the modified method leads to a good fit between observed and synthetic displacements and recovers the main rupture characteristics.

  17. Food additives

    MedlinePlus

    ... or natural. Natural food additives include: Herbs or spices to add flavor to foods Vinegar for pickling ... Certain colors improve the appearance of foods. Many spices, as well as natural and man-made flavors, ...

  18. Office of River Protection: Simplifying Project management tools

    SciTech Connect

    TAYLOR, D.G.

    2000-09-24

    The primary approach to the effort was to form a multi-organizational team comprised of federal and contractor staff to develop and implement the necessary tools and systems to manage the project. In late 1999 the DOE Manager of the Office of River Protection formed the Project Integration Office to achieve the objective of managing the efforts as a single project. The first major task, and the foundation upon which to base the development of all other tools, was the establishment of a single baseline of activities. However, defining a single scope schedule and cost was a difficult matter indeed. Work scopes were available throughout the project, but the level of detail and the integration of the activities existed primarily between working groups and individuals and not on a project-wide basis. This creates a situation where technical needs, logic flaws, resource balancing, and other similar integration needs are not elevated for management attention and resolution. It should be noted that probably 90% of the interface issues were known and being addressed. The key is simplifying the process and providing tangible assurance that the other 10% does not contain issues that can delay the project. Fortunately all of the contractors employed a common scheduling tool, which served as the basis for first communicating and then integrating baseline activities. Utilizing a powerful computer-based scheduling tool, it was soon possible to integrate the various schedules after the following was accomplished: Establishment of a scheduling specification (standardized input, coding, and approach to logic); and Clearly defined project assumptions.

  19. Detailed and simplified nonequilibrium helium ionization in the solar atmosphere

    SciTech Connect

    Golding, Thomas Peter; Carlsson, Mats; Leenaarts, Jorrit E-mail: mats.carlsson@astro.uio.no

    2014-03-20

    Helium ionization plays an important role in the energy balance of the upper chromosphere and transition region. Helium spectral lines are also often used as diagnostics of these regions. We carry out one-dimensional radiation-hydrodynamics simulations of the solar atmosphere and find that the helium ionization is set mostly by photoionization and direct collisional ionization, counteracted by radiative recombination cascades. By introducing an additional recombination rate mimicking the recombination cascades, we construct a simplified three-level helium model atom consisting of only the ground states. This model atom is suitable for modeling nonequilibrium helium ionization in three-dimensional numerical models. We perform a brief investigation of the formation of the He I 10830 and He II 304 spectral lines. Both lines show nonequilibrium features that are not recovered with statistical equilibrium models, and caution should therefore be exercised when such models are used as a basis for interpretating observations.

  20. Simplified Method for Preparing Methylene-Blue-Sensitized Dichromated Gelatin

    NASA Astrophysics Data System (ADS)

    Kurokawa, Kazumasa; Koike, Satoshi; Namba, Sinji; Mizuno, Toru; Kubota, Toshihiro

    1998-05-01

    Methylene-blue-sensitized dichromated gelatin (MBDCG) is a suitable material for recording full-color holograms in a single layer. However, a drying process in an ammonia atmosphere is necessary to prepare the MBDCG plate. This process is time-consuming and unstable. A simplified method for preparing the MBDCG plate is presented in which the MBDCG can be dried without ammonia. Elimination of the drying process is possible when the methylene blue in MBDCG does not separate. This is achieved by a decrease in the concentration of dichromate in the photosensitized solution and the addition of an ammonia solution to the photosensitized solution. Last, the gelatin is allowed to gel. A Lippmann color hologram grating with a diffraction efficiency of more than 80% is obtained by use of this MBDCG.

  1. The simplified Fermi accelerator in classical and quantum mechanics

    NASA Astrophysics Data System (ADS)

    Karner, Gunther

    1994-11-01

    We review the simplified classical Fermi acceleration mechanism and construct a quantum counterpart by imposing time-dependent boundary conditions on solutions of the "free" Schrödinger equation at the unit interval. We find similiar dynamical features in the sense that limiting KAM curves, respectively purely singular quasienergy spectrum, exist(s) for sufficiently smooth "wall oscillations" (typically of C 2 type). In addition, we investigate quantum analogs to local approximations of the Fermi map both in its quasiperiodic and irregular phase space regions. In particular, we find pure point q.e. spectrum in the former case and conjecture that "random boundary conditions" are necessary to model a quantum analog to the chaotic regime of the classical accelerator.

  2. Combustion Safety Simplified Test Protocol Field Study

    SciTech Connect

    Brand, L.; Cautley, D.; Bohac, D.; Francisco, P.; Shen, L.; Gloss, S.

    2015-11-01

    Combustions safety is an important step in the process of upgrading homes for energy efficiency. There are several approaches used by field practitioners, but researchers have indicated that the test procedures in use are complex to implement and provide too many false positives. Field failures often mean that the house is not upgraded until after remediation or not at all, if not include in the program. In this report the PARR and NorthernSTAR DOE Building America Teams provide a simplified test procedure that is easier to implement and should produce fewer false positives. A survey of state weatherization agencies on combustion safety issues, details of a field data collection instrumentation package, summary of data collected over seven months, data analysis and results are included. The project team collected field data on 11 houses in 2015.

  3. Constructing Simplified Plans via Truth Criteria Approximation

    NASA Technical Reports Server (NTRS)

    Chien, S.; DeJong, G.

    1994-01-01

    This paper has presented an approach to dealing with the complexity of explanation-based learning plans in complex domains. This approach uses a simplified algorithm to construct plans, and employs later refinements to repair bugs in constructed plans. This algorithm has the theoretical properties of completeness and convergence upon soundness. This incremental reasoning planning and learning algorithm has been implemented using a partial-order constraint posting planner and empirically compared to a conventional exhaustive reasoning partial-order constraint-posting planner and learning algorithm. This comparison showed that 1) incremental reasoning significantly reduced learning costs compared to exhaustive reasoning, 2) Explanation-based Learning (EBL) reduced failures from incremental reasoning, 3) EBL with incremental reasoning required less search to solve problems than EBL with exhaustive reasoning.

  4. Nonlinear optimization simplified by hypersurface deformation

    SciTech Connect

    Stillinger, F.H.; Weber, T.A.

    1988-09-01

    A general strategy is advanced for simplifying nonlinear optimization problems, the ant-lion method. This approach exploits shape modifications of the cost-function hypersurface which distend basins surrounding low-lying minima (including global minima). By intertwining hypersurface deformations with steepest-descent displacements, the search is concentrated on a small relevant subset of all minima. Specific calculations demonstrating the value of this method are reported for the partitioning of two classes of irregular but nonrandom graphs, the prime-factor graphs and the pi graphs. We also indicate how this approach can be applied to the traveling salesman problem and to design layout optimization, and that it may be useful in combination with simulated annealing strategies.

  5. Aeroacoustic Analysis of a Simplified Landing Gear

    NASA Technical Reports Server (NTRS)

    Lockard, David P.; Khorrami, Mehdi, R.; Li, Fei

    2004-01-01

    A hybrid approach is used to investigate the noise generated by a simplified landing gear without small scale parts such as hydraulic lines and fasteners. The Ffowcs Williams and Hawkings equation is used to predict the noise at far-field observer locations from flow data provided by an unsteady computational fluid dynamics calculation. A simulation with 13 million grid points has been completed, and comparisons are made between calculations with different turbulence models. Results indicate that the turbulence model has a profound effect on the levels and character of the unsteadiness. Flow data on solid surfaces and a set of permeable surfaces surrounding the gear have been collected. Noise predictions using the porous surfaces appear to be contaminated by errors caused by large wake fluctuations passing through the surfaces. However, comparisons between predictions using the solid surfaces with the near-field CFD solution are in good agreement giving confidence in the far-field results.

  6. Simplified Model of Nonlinear Landau Damping

    SciTech Connect

    N. A. Yampolsky and N. J. Fisch

    2009-07-16

    The nonlinear interaction of a plasma wave with resonant electrons results in a plateau in the electron distribution function close to the phase velocity of the plasma wave. As a result, Landau damping of the plasma wave vanishes and the resonant frequency of the plasma wave downshifts. However, this simple picture is invalid when the external driving force changes the plasma wave fast enough so that the plateau cannot be fully developed. A new model to describe amplification of the plasma wave including the saturation of Landau damping and the nonlinear frequency shift is proposed. The proposed model takes into account the change of the plasma wave amplitude and describes saturation of the Landau damping rate in terms of a single fluid equation, which simplifies the description of the inherently kinetic nature of Landau damping. A proposed fluid model, incorporating these simplifications, is verified numerically using a kinetic Vlasov code.

  7. Entropy reduction via simplified image contourization

    NASA Technical Reports Server (NTRS)

    Turner, Martin J.

    1993-01-01

    The process of contourization is presented which converts a raster image into a set of plateaux or contours. These contours can be grouped into a hierarchical structure, defining total spatial inclusion, called a contour tree. A contour coder has been developed which fully describes these contours in a compact and efficient manner and is the basis for an image compression method. Simplification of the contour tree has been undertaken by merging contour tree nodes thus lowering the contour tree's entropy. This can be exploited by the contour coder to increase the image compression ratio. By applying general and simple rules derived from physiological experiments on the human vision system, lossy image compression can be achieved which minimizes noticeable artifacts in the simplified image.

  8. Space station ECLSS simplified integrated test

    NASA Technical Reports Server (NTRS)

    Schunk, Richard G.; Bagdigian, Robert M.; Carrasquillo, Robyn L.; Ogle, Kathyrn Y.; Wieland, Paul O.

    1989-01-01

    A discussion of the Space Station Simplified Integrated Test (SIT) was conducted. The first in a series of three integrated Environmental Control and Life Support (ECLS) system tests, the primary objectives of the SIT were to verify proper operation of ECLS subsystems functioning in an integrated fashion as well as to gather preliminary performance data for the partial ECLS system used in the test. A description of the SIT configuration, a summary of events, a discussion of anomalies that occurred during the test, and detailed results and analysis from individual measurements and water and gas samples taken during the test are included. The preprototype ECLS hardware used in the test is reported providing an overall process description and theory of operation for each hardware item.

  9. Simplifying the circuit of Josephson parametric converters

    NASA Astrophysics Data System (ADS)

    Abdo, Baleegh; Brink, Markus; Chavez-Garcia, Jose; Keefe, George

    Josephson parametric converters (JPCs) are quantum-limited three-wave mixing devices that can play various important roles in quantum information processing in the microwave domain, including amplification of quantum signals, transduction of quantum information, remote entanglement of qubits, nonreciprocal amplification, and circulation of signals. However, the input-output and biasing circuit of a state-of-the-art JPC consists of bulky components, i.e. two commercial off-chip broadband 180-degree hybrids, four phase-matched short coax cables, and one superconducting magnetic coil. Such bulky hardware significantly hinders the integration of JPCs in scalable quantum computing architectures. In my talk, I will present ideas on how to simplify the JPC circuit and show preliminary experimental results

  10. Testing Modeling Assumptions in the West Africa Ebola Outbreak

    NASA Astrophysics Data System (ADS)

    Burghardt, Keith; Verzijl, Christopher; Huang, Junming; Ingram, Matthew; Song, Binyang; Hasne, Marie-Pierre

    2016-10-01

    The Ebola virus in West Africa has infected almost 30,000 and killed over 11,000 people. Recent models of Ebola Virus Disease (EVD) have often made assumptions about how the disease spreads, such as uniform transmissibility and homogeneous mixing within a population. In this paper, we test whether these assumptions are necessarily correct, and offer simple solutions that may improve disease model accuracy. First, we use data and models of West African migration to show that EVD does not homogeneously mix, but spreads in a predictable manner. Next, we estimate the initial growth rate of EVD within country administrative divisions and find that it significantly decreases with population density. Finally, we test whether EVD strains have uniform transmissibility through a novel statistical test, and find that certain strains appear more often than expected by chance.

  11. Testing Modeling Assumptions in the West Africa Ebola Outbreak

    PubMed Central

    Burghardt, Keith; Verzijl, Christopher; Huang, Junming; Ingram, Matthew; Song, Binyang; Hasne, Marie-Pierre

    2016-01-01

    The Ebola virus in West Africa has infected almost 30,000 and killed over 11,000 people. Recent models of Ebola Virus Disease (EVD) have often made assumptions about how the disease spreads, such as uniform transmissibility and homogeneous mixing within a population. In this paper, we test whether these assumptions are necessarily correct, and offer simple solutions that may improve disease model accuracy. First, we use data and models of West African migration to show that EVD does not homogeneously mix, but spreads in a predictable manner. Next, we estimate the initial growth rate of EVD within country administrative divisions and find that it significantly decreases with population density. Finally, we test whether EVD strains have uniform transmissibility through a novel statistical test, and find that certain strains appear more often than expected by chance. PMID:27721505

  12. An epidemic model to evaluate the homogeneous mixing assumption

    NASA Astrophysics Data System (ADS)

    Turnes, P. P.; Monteiro, L. H. A.

    2014-11-01

    Many epidemic models are written in terms of ordinary differential equations (ODE). This approach relies on the homogeneous mixing assumption; that is, the topological structure of the contact network established by the individuals of the host population is not relevant to predict the spread of a pathogen in this population. Here, we propose an epidemic model based on ODE to study the propagation of contagious diseases conferring no immunity. The state variables of this model are the percentages of susceptible individuals, infectious individuals and empty space. We show that this dynamical system can experience transcritical and Hopf bifurcations. Then, we employ this model to evaluate the validity of the homogeneous mixing assumption by using real data related to the transmission of gonorrhea, hepatitis C virus, human immunodeficiency virus, and obesity.

  13. Testing Modeling Assumptions in the West Africa Ebola Outbreak.

    PubMed

    Burghardt, Keith; Verzijl, Christopher; Huang, Junming; Ingram, Matthew; Song, Binyang; Hasne, Marie-Pierre

    2016-10-10

    The Ebola virus in West Africa has infected almost 30,000 and killed over 11,000 people. Recent models of Ebola Virus Disease (EVD) have often made assumptions about how the disease spreads, such as uniform transmissibility and homogeneous mixing within a population. In this paper, we test whether these assumptions are necessarily correct, and offer simple solutions that may improve disease model accuracy. First, we use data and models of West African migration to show that EVD does not homogeneously mix, but spreads in a predictable manner. Next, we estimate the initial growth rate of EVD within country administrative divisions and find that it significantly decreases with population density. Finally, we test whether EVD strains have uniform transmissibility through a novel statistical test, and find that certain strains appear more often than expected by chance.

  14. Effects of internal gain assumptions in building energy calculations

    SciTech Connect

    Christensen, C.; Perkins, R.

    1981-01-01

    The utilization of direct solar gains in buildings can be affected by operating profiles, such as schedules for internal gains, thermostat controls, and ventilation rates. Building energy analysis methods use various assumptions about these profiles. The effects of typical internal gain assumptions in energy calculations are described. Heating and cooling loads from simulations using the DOE 2.1 computer code are compared for various internal-gain inputs: typical hourly profiles, constant average profiles, and zero gain profiles. Prototype single-family-detached and multi-family-attached residential units are studied with various levels of insulation and infiltration. Small detached commercial buildings and attached zones in large commercial buildings are studied with various levels of internal gains. The results of this study indicate that calculations of annual heating and cooling loads are sensitive to internal gains, but in most cases are relatively insensitive to hourly variations in internal gains.

  15. Effects of internal gain assumptions in building energy calculations

    NASA Astrophysics Data System (ADS)

    Christensen, C.; Perkins, R.

    1981-01-01

    The utilization of direct solar gains in buildings can be affected by operating profiles, such as schedules for internal gains, thermostat controls, and ventilation rates. Building energy analysis methods use various assumptions about these profiles. The effects of typical internal gain assumptions in energy calculations are described. Heating and cooling loads from simulations using the DOE 2.1 computer code are compared for various internal gain inputs: typical hourly profiles, constant average profiles, and zero gain profiles. Prototype single-family-detached and multifamily-attached residential units are studied with various levels of insulation and infiltration. Small detached commercial buildings and attached zones in large commercial buildings are studied with various levels of internal gains. The results indicate that calculations of annual heating and cooling loads are sensitive to internal gains, but in most cases are relatively insensitive to hourly variations in internal gains.

  16. Effects of internal gain assumptions in building energy calculations

    NASA Astrophysics Data System (ADS)

    Christensen, C.; Perkins, R.

    The utilization of direct solar gains in buildings can be affected by operating profiles, such as schedules for internal gains, thermostat controls, and ventilation rates. Building energy analysis methods use various assumptions about these profiles. The effects of typical internal gain assumptions in energy calculations. Heating and cooling loads from simulations using the Department of Energy 2.1 computer code are compared for various internal-gain inputs: typical hourly profiles, constant average profiles, and zero gain profiles. Prototype single-family-detached and multi-family-attached residential units are studied with various levels of insulation and infiltration. Small detached commercial buildings and attached zones in large commercial buildings are studied with various levels of internal gains. The results indicate that calculations of annual heating and cooling loads are sensitive to internal gains, but in most cases are relatively insensitive to hourly variations in internal gains.

  17. Miniaturized photoelectric angular sensor with simplified design

    NASA Astrophysics Data System (ADS)

    Dumbravescu, Niculae; Schiaua, Silviu

    1999-09-01

    In building the movable elements of robots, peripheral devices and measuring apparata, increasing the resolution of the angular sensor systems, based on incremental rotary encoders, is essential, together with decreasing the complexity, dimensions and weight. Especially when the angular sensor is integrated in a measuring system, belonging to a programmed light airplane for surveillance, the key issue is to reduce both dimensions and weight. This can be done using a simplified design, which consists in the following solutions: replacement of the fragile Cr on glass substrate, 1.5 mm thick (normally used for the fabrication of incremental disks), with light Cr on polycarbonate substrate, with only 0.15 mm thick; the absence of collimating optics (based on microlenses, used in IR emitter-photocell receiver assembly), as a result of the good coupling efficiency (due to the possible approaching of these elements at minimum 0.45 mm); the shrinkage of the disk's diameters to only 14 mm; the use of surface mounting devices and the related surface mounting technology, enabling to reduce dimensions and weight. The maximum number of slits on a 14 mm diameter dividing disk, usually obtained in a Cr on polycarbonate version, being approx. 1000, no problem occurs in our case, for 360 slits. The requested angular resolution (only 0.5 degrees for the light airplane), using the whole classical '4x digital multiplication' is not necessary, but a lower one of only 2x, resulting in a simplified electronics. The proposed design permitted, that an original arrangement, for building a small size, lightweight, heavy-duty incremental transducer based angular sensor system, to be obtained, useful not only in avionics, but also in robotics, or other special applications. Besides, extending the number of fixed gratings (masks) allows, that many primary signals to be derived, and a further increase in resolution of even 6 angular minutes to be obtained from the initial 360 slits.

  18. Tax Subsidies for Employer-Sponsored Health Insurance: Updated Microsimulation Estimates and Sensitivity to Alternative Incidence Assumptions

    PubMed Central

    Miller, G Edward; Selden, Thomas M

    2013-01-01

    Objective To estimate 2012 tax expenditures for employer-sponsored insurance (ESI) in the United States and to explore the sensitivity of estimates to assumptions regarding the incidence of employer premium contributions. Data Sources Nationally representative Medical Expenditure Panel Survey data from the 2005–2007 Household Component (MEPS-HC) and the 2009–2010 Insurance Component (MEPS IC). Study Design We use MEPS HC workers to construct synthetic workforces for MEPS IC establishments, applying the workers' marginal tax rates to the establishments' insurance premiums to compute the tax subsidy, in aggregate and by establishment characteristics. Simulation enables us to examine the sensitivity of ESI tax subsidy estimates to a range of scenarios for the within-firm incidence of employer premium contributions when workers have heterogeneous health risks and make heterogeneous plan choices. Principal Findings We simulate the total ESI tax subsidy for all active, civilian U.S. workers to be $257.4 billion in 2012. In the private sector, the subsidy disproportionately flows to workers in large establishments and establishments with predominantly high wage or full-time workforces. The estimates are remarkably robust to alternative incidence assumptions. Conclusions The aggregate value of the ESI tax subsidy and its distribution across firms can be reliably estimated using simplified incidence assumptions. PMID:23398400

  19. Sensitivity of probabilistic MCO water content estimates to key assumptions

    SciTech Connect

    DUNCAN, D.R.

    1999-02-25

    Sensitivity of probabilistic multi-canister overpack (MCO) water content estimates to key assumptions is evaluated with emphasis on the largest non-cladding film-contributors, water borne by particulates adhering to damage sites, and water borne by canister particulate. Calculations considered different choices of damage state degree of independence, different choices of percentile for reference high inputs, three types of input probability density function (pdfs): triangular, log-normal, and Weibull, and the number of scrap baskets in an MCO.

  20. The parallel flow assumption in Greenland outlet glaciers

    NASA Astrophysics Data System (ADS)

    Gourmelen, N.; Shepherd, A.; Park, J. W.

    2012-04-01

    Climate warming over the 20th century has forced dramatic changes in the Greenland Ice Sheet (GrIS). These changes have led to a reduction in the mass of the GrIS and a consequent rise in global sea level. Satellite observations have revealed an increased flow of the glaciers to the sea [Rignot et al., 2008], increased surface melting [Steffen et al., 2004], lowering of the Ice Sheet surface [Zwally and Giovinetto, 2001], retreat of the glaciers' fronts [Box et al., 2006], and gravity anomaly related to ice mass loss [Velicogna and Wahr, 2006]. When measuring the flow of ice from spaceborne sensors, it is often assumed that the direction and dip of flow follow the direction and dip of the ice surface [Rignot et al., 2011]; this is known as the surface parallel flow assumption [Joughin et al., 1996]. This assumption is often the only way to constrain glaciers' flow, as observation by spaceborne sensors in Polar Regions is limited. Departure from the parallel flow assumption means that the magnitude of ice flow and its changes will be misestimated. Departure from the ice parallel flow assumption also provides clues on the mechanisms leading to the flow pattern and to the change in flow magnitude. Here we exploit 20 years of Synthetic Aperture Radar spaceborn missions to constrain the three-dimensional flow of marine terminating glaciers in the GrIS. We use datasets from past and present ERS1, ERS2, ENVISAT, ALOS and TerraSAR-X missions to construct 3-dimensional flow maps of selected outlet glaciers of the GrIS. This dataset is used to explore the relationship between flow patterns, its temporal evolution, and processes at play at the margins of the GrIS.

  1. Assumptions and realities of the NCLEX-RN.

    PubMed

    Aucoin, Julia W; Treas, Leslie

    2005-01-01

    Every three years the National Council of State Boards of Nursing conducts a practice analysis to verify the activities that are tested on the licensure exam (NCLEX-RN). Faculty can benefit from information in the practice analysis to ensure that courses and experiences adequately prepare graduates for the NCLEX-RN. This summary of the practice analysis challenges common assumptions and provides recommendations for faculty.

  2. Validity of the "thin" and "thick" double-layer assumptions to model streaming currents in porous media

    NASA Astrophysics Data System (ADS)

    Leinov, E.; Jackson, M.

    2012-12-01

    Measurements of the streaming potential component of the spontaneous potential have been used to characterize groundwater flow and subsurface hydraulic properties in numerous studies. Streaming potentials in porous media arise from the electrical double layer which forms at solid-fluid interfaces. The solid surfaces typically become electrically charged, in which case an excess of counter-charge accumulates in the adjacent fluid. If the fluid is induced to flow by an external pressure gradient, then some of the excess charge within the diffuse part of the double layer is transported with the flow, giving rise to a streaming current. Divergence of the streaming current density establishes an electrical potential, termed the streaming potential. Within the diffuse layer, the Poisson-Boltzmann equation is typically used to describe the variation in electrical potential with distance from the solid surface. In many subsurface settings, it is reasonable to assume that the thickness of the diffuse layer is small compared to the pore radius. This is the so-called 'thin double layer assumption', which has been invoked by numerous authors to model streaming potentials in porous media. However, a number of recent papers have proposed a different approach, in which the thickness of the diffuse layer is assumed to be large compared to the pore radius. This is the so-called 'thick double layer assumption' in which the excess charge density within the pore is assumed to be constant and independent of distance from the solid surface. The advantage of both the 'thin' and 'thick' double layer assumptions is that calculation of the streaming current is greatly simplified. However, perhaps surprisingly, the conditions for which these assumptions are valid have not been determined quantitatively, yet they have a significant impact on the interpretation of streaming potential measurements in natural systems. We use a simple capillary tubes to model investigate the validity of the thin

  3. The incompressibility assumption in computational simulations of nasal airflow.

    PubMed

    Cal, Ismael R; Cercos-Pita, Jose Luis; Duque, Daniel

    2017-04-03

    Most of the computational works on nasal airflow up to date have assumed incompressibility, given the low Mach number of these flows. However, for high temperature gradients, the incompressibility assumption could lead to a loss of accuracy, due to the temperature dependence of air density and viscosity. In this article we aim to shed some light on the influence of this assumption in a model of calm breathing in an Asian nasal cavity, by solving the fluid flow equations in compressible and incompressible formulation for different ambient air temperatures using the OpenFOAM package. At low flow rates and warm climatological conditions, similar results were obtained from both approaches, showing that density variations need not be taken into account to obtain a good prediction of all flow features, at least for usual breathing conditions. This agrees with most of the simulations previously reported, at least as far as the incompressibility assumption is concerned. However, parameters like nasal resistance and wall shear stress distribution differ for air temperatures below [Formula: see text]C approximately. Therefore, density variations should be considered for simulations at such low temperatures.

  4. Some Considerations on the Basic Assumptions in Rotordynamics

    NASA Astrophysics Data System (ADS)

    GENTA, G.; DELPRETE, C.; BRUSA, E.

    1999-10-01

    The dynamic study of rotors is usually performed under a number of assumptions, namely small displacements and rotations, small unbalance and constant angular velocity. The latter assumption can be substituted by a known time history of the spin speed. The present paper develops a general non-linear model which can be used to study the rotordynamic behaviour of both fixed and free rotors without resorting to the mentioned assumptions and compares the results obtained from a number of non-linear numerical simulations with those computed through the usual linearized approach. It is so possible to verify that the validity of the rotordynamic models extends to situations in which fairly large unbalances and whirling motions are present and, above all, it is shown that the doubts forwarded about the application of a model which is based on constant spin speed to the case of free rotors in which the angular momentum is constant have no ground. Rotordynamic models can thus be used to study the stability in the small of spinning spacecrafts and the insight obtained from the study of rotors is useful to understand their attitude dynamics and its interactions with the vibration dynamics.

  5. A CRITICAL EXAMINATION OF THE FUNDAMENTAL ASSUMPTIONS OF SOLAR FLARE AND CORONAL MASS EJECTION MODELS

    SciTech Connect

    Spicer, D. S.; Bingham, R.; Harrison, R.

    2013-05-01

    The fundamental assumptions of conventional solar flare and coronal mass ejection (CME) theory are re-examined. In particular, the common theoretical assumption that magnetic energy that drives flares and CMEs can be stored in situ in the corona with sufficient energy density is found wanting. In addition, the observational constraint that flares and CMEs produce non-thermal electrons with fluxes of order 10{sup 34}-10{sup 36} electrons s{sup -1}, with energies of order 10-20 keV, must also be explained. This constraint when imposed on the ''standard model'' for flares and CMEs is found to miss the mark by many orders of magnitude. We suggest, in conclusion, there are really only two possible ways to explain the requirements of observations and theory: flares and CMEs are caused by mass-loaded prominences or driven directly by emerging magnetized flux.

  6. When does ``like'' like ``like''? How does the repulsion-only assumption fail?

    NASA Astrophysics Data System (ADS)

    Ise, Norio

    2000-06-01

    When dispersed particles or solute ions have high charge densities, macroscopically homogeneous systems become microscopically inhomogeneous. Examples are the two-state structure of ordered structures in disordered region without boundary and the void structures. When the charge density is increased, the reentrant phase (liquid-solid-liquid) transition is found, which is not explainable in terms of the repulsion-only assumption. Furthermore, for relatively small charge particles and at an early stage of crystallization, space-filling ordered states are first formed, disordered regions are then created inside the ordered domains, and crystal contraction thereafter takes place, causing the two-state structure. These may be explained by invoking a counterion-mediated attractive interaction between like-charged entities in addition to the widely accepted repulsion-only assumption. .

  7. Simplified formulae to investigate flexural vibration characteristics of piezoelectric tubes in ultrasonic micro-actuators.

    PubMed

    Zhang, Hui; Zhang, Shu-Yi; Fan, Li

    2010-03-01

    Based on the Rayleigh energy theory combining with Timoshenko beam model, the flexural vibration characteristics of piezoelectric tubes in ultrasonic micro-actuators are investigated. Additionally, the simplified formulae are derived to study the fundamental flexural resonance frequencies of the piezoelectric tubes with free-free ends and cantilevers. By changing the sizes of the tubes and the mass loads at the free ends, the variations of the flexural resonance frequencies of the piezoelectric tubes and cantilevers are calculated theoretically. To verify accuracy of the simplified formulae, by changing the lengths of the tubes and the mass loads the flexural resonance frequencies of the piezoelectric tube with free-free ends are measured experimentally. The theoretical results agree well with the experimental measurements, which demonstrate that the simplified formulae are accurate and effective for analyzing the flexural vibration characteristics of the piezoelectric tubes in the ultrasonic micro-actuators.

  8. Simplified Ion Thruster Xenon Feed System for NASA Science Missions

    NASA Technical Reports Server (NTRS)

    Snyder, John Steven; Randolph, Thomas M.; Hofer, Richard R.; Goebel, Dan M.

    2009-01-01

    The successful implementation of ion thruster technology on the Deep Space 1 technology demonstration mission paved the way for its first use on the Dawn science mission, which launched in September 2007. Both Deep Space 1 and Dawn used a "bang-bang" xenon feed system which has proven to be highly successful. This type of feed system, however, is complex with many parts and requires a significant amount of engineering work for architecture changes. A simplified feed system, with fewer parts and less engineering work for architecture changes, is desirable to reduce the feed system cost to future missions. An attractive new path for ion thruster feed systems is based on new components developed by industry in support of commercial applications of electric propulsion systems. For example, since the launch of Deep Space 1 tens of mechanical xenon pressure regulators have successfully flown on commercial spacecraft using electric propulsion. In addition, active proportional flow controllers have flown on the Hall-thruster-equipped Tacsat-2, are flying on the ion thruster GOCE mission, and will fly next year on the Advanced EHF spacecraft. This present paper briefly reviews the Dawn xenon feed system and those implemented on other xenon electric propulsion flight missions. A simplified feed system architecture is presented that is based on assembling flight-qualified components in a manner that will reduce non-recurring engineering associated with propulsion system architecture changes, and is compared to the NASA Dawn standard. The simplified feed system includes, compared to Dawn, passive high-pressure regulation, a reduced part count, reduced complexity due to cross-strapping, and reduced non-recurring engineering work required for feed system changes. A demonstration feed system was assembled using flight-like components and used to operate a laboratory NSTAR-class ion engine. Feed system components integrated into a single-string architecture successfully operated

  9. Potlining Additives

    SciTech Connect

    Rudolf Keller

    2004-08-10

    In this project, a concept to improve the performance of aluminum production cells by introducing potlining additives was examined and tested. Boron oxide was added to cathode blocks, and titanium was dissolved in the metal pool; this resulted in the formation of titanium diboride and caused the molten aluminum to wet the carbonaceous cathode surface. Such wetting reportedly leads to operational improvements and extended cell life. In addition, boron oxide suppresses cyanide formation. This final report presents and discusses the results of this project. Substantial economic benefits for the practical implementation of the technology are projected, especially for modern cells with graphitized blocks. For example, with an energy savings of about 5% and an increase in pot life from 1500 to 2500 days, a cost savings of $ 0.023 per pound of aluminum produced is projected for a 200 kA pot.

  10. Phosphazene additives

    DOEpatents

    Harrup, Mason K; Rollins, Harry W

    2013-11-26

    An additive comprising a phosphazene compound that has at least two reactive functional groups and at least one capping functional group bonded to phosphorus atoms of the phosphazene compound. One of the at least two reactive functional groups is configured to react with cellulose and the other of the at least two reactive functional groups is configured to react with a resin, such as an amine resin of a polycarboxylic acid resin. The at least one capping functional group is selected from the group consisting of a short chain ether group, an alkoxy group, or an aryloxy group. Also disclosed are an additive-resin admixture, a method of treating a wood product, and a wood product.

  11. A simplified electrostatic model for hydrolase catalysis.

    PubMed

    Pessoa Filho, Pedro de Alcantara; Prausnitz, John M

    2015-07-01

    Toward the development of an electrostatic model for enzyme catalysis, the active site of the enzyme is represented by a cavity whose surface (and beyond) is populated by electric charges as determined by pH and the enzyme's structure. The electric field in the cavity is obtained from electrostatics and a suitable computer program. The key chemical bond in the substrate, at its ends, has partial charges with opposite signs determined from published force-field parameters. The electric field attracts one end of the bond and repels the other, causing bond tension. If that tension exceeds the attractive force between the atoms, the bond breaks; the enzyme is then a successful catalyst. To illustrate this very simple model, based on numerous assumptions, some results are presented for three hydrolases: hen-egg white lysozyme, bovine trypsin and bovine ribonuclease. Attention is given to the effect of pH.

  12. Simplified Modeling of Oxidation of Hydrocarbons

    NASA Technical Reports Server (NTRS)

    Bellan, Josette; Harstad, Kenneth

    2008-01-01

    A method of simplified computational modeling of oxidation of hydrocarbons is undergoing development. This is one of several developments needed to enable accurate computational simulation of turbulent, chemically reacting flows. At present, accurate computational simulation of such flows is difficult or impossible in most cases because (1) the numbers of grid points needed for adequate spatial resolution of turbulent flows in realistically complex geometries are beyond the capabilities of typical supercomputers now in use and (2) the combustion of typical hydrocarbons proceeds through decomposition into hundreds of molecular species interacting through thousands of reactions. Hence, the combination of detailed reaction- rate models with the fundamental flow equations yields flow models that are computationally prohibitive. Hence, further, a reduction of at least an order of magnitude in the dimension of reaction kinetics is one of the prerequisites for feasibility of computational simulation of turbulent, chemically reacting flows. In the present method of simplified modeling, all molecular species involved in the oxidation of hydrocarbons are classified as either light or heavy; heavy molecules are those having 3 or more carbon atoms. The light molecules are not subject to meaningful decomposition, and the heavy molecules are considered to decompose into only 13 specified constituent radicals, a few of which are listed in the table. One constructs a reduced-order model, suitable for use in estimating the release of heat and the evolution of temperature in combustion, from a base comprising the 13 constituent radicals plus a total of 26 other species that include the light molecules and related light free radicals. Then rather than following all possible species through their reaction coordinates, one follows only the reduced set of reaction coordinates of the base. The behavior of the base was examined in test computational simulations of the combustion of

  13. User assumptions about information retrieval systems: Ethical concerns

    SciTech Connect

    Froehlich, T.J.

    1994-12-31

    Information professionals, whether designers, intermediaries, database producers or vendors, bear some responsibility for the information that they make available to users of information systems. The users of such systems may tend to make many assumptions about the information that a system provides, such as believing: that the data are comprehensive, current and accurate, that the information resources or databases have same degree of quality and consistency of indexing; that the abstracts, if they exist, correctly and adequate reflect the content of the article; that there is consistency informs of author names or journal titles or indexing within and across databases; that there is standardization in and across databases; that once errors are detected, they are corrected; that appropriate choices of databases or information resources are a relatively easy matter, etc. The truth is that few of these assumptions are valid in commercia or corporate or organizational databases. However, given these beliefs and assumptions by many users, often promoted by information providers, information professionals, impossible, should intervene to warn users about the limitations and constraints of the databases they are using. With the growth of the Internet and end-user products (e.g., CD-ROMs), such interventions have significantly declined. In such cases, information should be provided on start-up or through interface screens, indicating to users, the constraints and orientation of the system they are using. The principle of {open_quotes}caveat emptor{close_quotes} is naive and socially irresponsible: information professionals or systems have an obligation to provide some framework or context for the information that users are accessing.

  14. The vulnerabilities of teenage mothers: challenging prevailing assumptions.

    PubMed

    SmithBattle, L

    2000-09-01

    The belief that early childbearing leads to poverty permeates our collective understanding. However, recent findings reveal that for many teens, mothering makes sense of the limited life options that precede their pregnancies. The author challenges several assumptions about teenage mothers and offers an alternative to the modern view of the unencumbered self that drives current responses to teen childbearing. This alternative perspective entails a situated view of the self and a broader notion of parenting and citizenship that supports teen mothers and affirms our mutual interdependence.

  15. Sensitivity of fine sediment source apportionment to mixing model assumptions

    NASA Astrophysics Data System (ADS)

    Cooper, Richard; Krueger, Tobias; Hiscock, Kevin; Rawlins, Barry

    2015-04-01

    Mixing models have become increasingly common tools for quantifying fine sediment redistribution in river catchments. The associated uncertainties may be modelled coherently and flexibly within a Bayesian statistical framework (Cooper et al., 2015). However, there is more than one way to represent these uncertainties because the modeller has considerable leeway in making error assumptions and model structural choices. In this presentation, we demonstrate how different mixing model setups can impact upon fine sediment source apportionment estimates via a one-factor-at-a-time (OFAT) sensitivity analysis. We formulate 13 versions of a mixing model, each with different error assumptions and model structural choices, and apply them to sediment geochemistry data from the River Blackwater, Norfolk, UK, to apportion suspended particulate matter (SPM) contributions from three sources (arable topsoils, road verges and subsurface material) under base flow conditions between August 2012 and August 2013 (Cooper et al., 2014). Whilst all 13 models estimate subsurface sources to be the largest contributor of SPM (median ~76%), comparison of apportionment estimates reveals varying degrees of sensitivity to changing prior parameter distributions, inclusion of covariance terms, incorporation of time-variant distributions and methods of proportion characterisation. We also demonstrate differences in apportionment results between a full and an empirical Bayesian setup and between a Bayesian and a popular Least Squares optimisation approach. Our OFAT sensitivity analysis reveals that mixing model structural choices and error assumptions can significantly impact upon fine sediment source apportionment results, with estimated median contributions in this study varying by up to 21% between model versions. Users of mixing models are therefore strongly advised to carefully consider and justify their choice of model setup prior to conducting fine sediment source apportionment investigations

  16. Constant-Round Concurrent Zero Knowledge From Falsifiable Assumptions

    DTIC Science & Technology

    2013-01-01

    black-box zero knowledge. J. Comput. Syst. Sci., 72(2):321–391, 2006. [Bon03] Dan Boneh , editor. Advances in Cryptology - CRYPTO 2003, 23rd Annual...certificates that are not certified by a higher-level certificate; we refer to these as “ dan - gling” certificates. See Figure 2 for an illustration of the tree...Nao03] Moni Naor. On cryptographic assumptions and challenges. In Boneh [Bon03], pages 96–109. [Pas03a] Rafael Pass. On deniability in the common

  17. Neurotic verbalizations: an exploration of a Gestalt therapy assumption.

    PubMed

    Nelson, W M; Groman, W D

    1975-10-01

    Based upon assumptions of Gestalt Therapy, an objective, reliable scoring system for analyzing a specific aspect of verbal behavior, avoidant verbalization, was developed. Verbal samples were elicited from 30 female college students who previously had been rank ordered on the Neuroticism scale of the Maudsley Personality Inventory. The expectation that psychologically stressful questions would result in a significantly greater number of avoidant verbalizations was confirmed. Contrary to another expectation, no significant correlation between neuroticism as measured by the MPI and total number of avoidant verbalizations uttered by the Ss was found. Results were interpreted on the basis of the Gestalt Therapy theory of personality.

  18. The Influence of Microphysical Assumptions on Simulated Brightness Temperatures

    NASA Astrophysics Data System (ADS)

    Adams, I. S.; Bettenhausen, M. H.

    2012-12-01

    There have been many studies that look at the microphysical properties of snow, as these assumptions about the properties are key components to bulk microphysical parameterizations. Temperature and supersaturation affect crystal growth, and the resulting crystals collide to form larger aggregates. Ice crystal habit, latent heat conditions, and the presence of supercooled liquid all affect aggregation and riming, resulting in a wide range of irregular morphologies of varying density and fall velocity. If melting does occur, hydrometeor properties will also influence the melting process. And, the presence of melting hydrometeors directly influences the size distribution of raindrops. As the characteristics of clouds vary from system to system, understanding how microphysical assumptions influence the interpretation of passive microwave brightness temperatures is an important step in developing methods for observing precipitation. Particle properties such as shape, number density, mass density, composition, and alignment all impact the intensity and polarization state of radiation. In this work, we look at the impact of various microphysical assumptions of snow on simulated microwave brightness temperatures. We include various parameterizations of mass density. Careful attention will be made in constructing particle size distributions, using profiles of liquid and ice water content. We include non-spherical geometries using directional statistics to consider distributions of canting angles, from horizontally-aligned to completely random to quantify polarization effects. In order to accommodate the wide range of various assumptions, we limit our particle models to rotationally symmetric geometries to facilitate the use of T-Matrix code. Densities in these cases are accounted for using a dielectric mixing formula. While higher fidelity scattering calculations using realistic particle models would better represent remotely sensed measurements, such calculations are

  19. Dental Education: Trends and Assumptions for the 21st Century

    PubMed Central

    Sinkford, Jeanne C.

    1987-01-01

    Dental educational institutions, as components of university systems, must develop strategic plans for program development, resource allocation, evaluation, and continued financial support. This dynamic process will be accomplished in a competitive academic arena where program excellence and program relevance are key issues in the game of survival. This article focuses on issues and trends that form the basis for planning assumptions and initiatives into the next decade and into the 21st century. This is our challenge, this is our mission if we are to be catalysts for change in the future. PMID:3560255

  20. Combustion Safety Simplified Test Protocol Field Study

    SciTech Connect

    Brand, L; Cautley, D.; Bohac, D.; Francisco, P.; Shen, L.; Gloss, S.

    2015-11-05

    "9Combustions safety is an important step in the process of upgrading homes for energy efficiency. There are several approaches used by field practitioners, but researchers have indicated that the test procedures in use are complex to implement and provide too many false positives. Field failures often mean that the house is not upgraded until after remediation or not at all, if not include in the program. In this report the PARR and NorthernSTAR DOE Building America Teams provide a simplified test procedure that is easier to implement and should produce fewer false positives. A survey of state weatherization agencies on combustion safety issues, details of a field data collection instrumentation package, summary of data collected over seven months, data analysis and results are included. The project provides several key results. State weatherization agencies do not generally track combustion safety failures, the data from those that do suggest that there is little actual evidence that combustion safety failures due to spillage from non-dryer exhaust are common and that only a very small number of homes are subject to the failures. The project team collected field data on 11 houses in 2015. Of these homes, two houses that demonstrated prolonged and excessive spillage were also the only two with venting systems out of compliance with the National Fuel Gas Code. The remaining homes experienced spillage that only occasionally extended beyond the first minute of operation. Combustion zone depressurization, outdoor temperature, and operation of individual fans all provide statistically significant predictors of spillage.

  1. Simplified liquid oxygen propellant conditioning concepts

    NASA Technical Reports Server (NTRS)

    Cleary, N. L.; Holt, K. A.; Flachbart, R. H.

    1995-01-01

    Current liquid oxygen feed systems waste propellant and use hardware, unnecessary during flight, to condition the propellant at the engine turbopumps prior to launch. Simplified liquid oxygen propellant conditioning concepts are being sought for future launch vehicles. During a joint program, four alternative propellant conditioning options were studied: (1) passive recirculation; (2) low bleed through the engine; (3) recirculation lines; and (4) helium bubbling. The test configuration for this program was based on a vehicle design which used a main recirculation loop that was insulated on the downcomer and uninsulated on the upcomer. This produces a natural convection recirculation flow. The test article for this program simulated a feedline which ran from the main recirculation loop to the turbopump. The objective was to measure the temperature profile of this test article. Several parameters were varied from the baseline case to determine their effects on the temperature profile. These parameters included: flow configuration, feedline slope, heat flux, main recirculation loop velocity, pressure, bleed rate, helium bubbling, and recirculation lines. The heat flux, bleed rate, and recirculation configurations produced the greatest changes from the baseline temperature profile. However, the temperatures in the feedline remained subcooled. Any of the options studied could be used in future vehicles.

  2. Simplified limits on resonances at the LHC

    NASA Astrophysics Data System (ADS)

    Chivukula, R. Sekhar; Ittisamai, Pawin; Mohan, Kirtimaan; Simmons, Elizabeth H.

    2016-11-01

    In the earliest stages of evaluating new collider data, especially if a small excess may be present, it would be useful to have a method for comparing the data with entire classes of models, to get an immediate sense of which classes could conceivably be relevant. In this paper, we propose a method that applies when the new physics invoked to explain the excess corresponds to the production and decay of a single, relatively narrow, s -channel resonance. A simplifed model of the resonance allows us to convert an estimated signal cross section into general bounds on the product of the branching ratios corresponding to the dominant production and decay modes. This quickly reveals whether a given class of models could possibly produce a signal of the required size at the LHC. Our work sets up a general framework, outlines how it operates for resonances with different numbers of production and decay modes, and analyzes cases of current experimental interest, including resonances decaying to dibosons, diphotons, dileptons, or dijets. If the LHC experiments were to report their searches for new resonances beyond the standard model in the simplified limits variable ζ defined in this paper, that would make it far easier to avoid blind alleys and home in on the most likely candidate models to explain any observed excesses.

  3. Simplified Optics and Controls for Laser Communications

    NASA Technical Reports Server (NTRS)

    Chen, Chien-Chung; Hemmati, Hamid

    2006-01-01

    A document discusses an architecture of a spaceborne laser communication system that provides for a simplified control subsystem that stabilizes the line of sight in a desired direction. Heretofore, a typical design for a spaceborne laser communication system has called for a high-bandwidth control loop, a steering mirror and associated optics, and a fast steering-mirror actuator to stabilize the line of sight in the presence of vibrations. In the present architecture, the need for this fast steering-mirror subsystem is eliminated by mounting the laser-communication optics on a disturbance-free platform (DFP) that suppresses coupling of vibrations to the optics by 60 dB. Taking advantage of microgravitation, in the DFP, the optical assembly is free-flying relative to the rest of the spacecraft, and a low-spring-constant pointing control subsystem exerts small forces to regulate the position and orientation of the optics via voice coils. All steering is effected via the DFP, which can be controlled in all six degrees of freedom relative to the spacecraft. A second control loop, closed around a position sensor and the spacecraft attitude-control system, moves the spacecraft as needed to prevent mechanical contact with the optical assembly.

  4. A clinically useful simplified blastocyst grading system.

    PubMed

    Richardson, Alison; Brearley, Sophie; Ahitan, Saran; Chamberlain, Sarah; Davey, Tracey; Zujovic, Lyndsey; Hopkisson, James; Campbell, Bruce; Raine-Fenning, Nick

    2015-10-01

    The aim of this study was to investigate whether a new simplified blastocyst grading system (A: fully expanded, clear inner cell mass, cohesive trophectoderm; B: not yet expanded, clear inner cell mass, cohesive trophectoderm; C: small inner cell mass ± irregular trophectoderm ± excluded/degenerate cells) was clinically useful. All day-5 single embryo transfers between 15 June 2009 and 29 June 2012 were reviewed. Implantation, clinical pregnancy and live birth rates were related to embryo quality. Five embryologists were asked to grade and decide the clinical fate of 80 images of day-5 embryos on two occasions 4-6 weeks apart. Implantation, clinical pregnancy and live birth rates decreased with deteriorating embryo quality. A highly significant (P < 0.01) difference was observed between the groups. Inter-observer agreement was substantial for grade allocation (K = 0.63) and clinical decision-making (K = 0.66). Intra-observer agreement ranged from substantial (K = 0.71) to almost perfect (K = 0.88) for grade allocation, and was almost perfect for clinical fate determination (K ≥ 0.84). This grading system is quick and easy to use, effectively predicts IVF outcome and has levels of agreement similar to, if not better than, those associated with more complex grading systems.

  5. Interferometric phase reconstruction using simplified coherence network

    NASA Astrophysics Data System (ADS)

    Zhang, Kui; Song, Ruiqing; Wang, Hui; Wu, Di; Wang, Hua

    2016-09-01

    Interferometric time-series analysis techniques, which extend the traditional differential radar interferometry, have demonstrated a strong capability for monitoring ground surface displacement. Such techniques are able to obtain the temporal evolution of ground deformation within millimeter accuracy by using a stack of synthetic aperture radar (SAR) images. In order to minimize decorrelation between stacked SAR images, the phase reconstruction technique has been developed recently. The main idea of this technique is to reform phase observations along a SAR stack by taking advantage of a maximum likelihood estimator which is defined on the coherence matrix estimated from each target. However, the phase value of a coherence matrix element might be considerably biased when its corresponding coherence is low. In this case, it will turn to an outlying sample affecting the corresponding phase reconstruction process. In order to avoid this problem, a new approach is developed in this paper. This approach considers a coherence matrix element to be an arc in a network. A so-called simplified coherence network (SCN) is constructed to decrease the negative impact of outlying samples. Moreover, a pointed iterative strategy is designed to resolve the transformed phase reconstruction problem defined on a SCN. For validation purposes, the proposed method is applied to 29 real SAR images. The results demonstrate that the proposed method has an excellent computational efficiency and could obtain more reliable phase reconstruction solutions compared to the traditional method using phase triangulation algorithm.

  6. Simplified Dynamic Analysis of Grinders Spindle Node

    NASA Astrophysics Data System (ADS)

    Demec, Peter

    2014-12-01

    The contribution deals with the simplified dynamic analysis of surface grinding machine spindle node. Dynamic analysis is based on the use of the transfer matrix method, which is essentially a matrix form of method of initial parameters. The advantage of the described method, despite the seemingly complex mathematical apparatus, is primarily, that it does not require for solve the problem of costly commercial software using finite element method. All calculations can be made for example in MS Excel, which is advantageous especially in the initial stages of constructing of spindle node for the rapid assessment of the suitability its design. After detailing the entire structure of spindle node is then also necessary to perform the refined dynamic analysis in the environment of FEM, which it requires the necessary skills and experience and it is therefore economically difficult. This work was developed within grant project KEGA No. 023TUKE-4/2012 Creation of a comprehensive educational - teaching material for the article Production technique using a combination of traditional and modern information technology and e-learning.

  7. Simplified lattice model for polypeptide fibrillar transitions

    NASA Astrophysics Data System (ADS)

    Xiao, Xuhui; Wu, Ming-Chya

    2014-10-01

    Polypeptide fibrillar transitions are studied using a simplified lattice model, modified from the three-state Potts model, where uniform residues as spins, placed on a cubic lattice, can interact with neighbors to form coil, helical, sheet, or fibrillar structure. Using the transfer matrix method and numerical calculations, we analyzed the partition function and construct phase diagrams. The model manifests phase transitions among coil, helix, sheet, and fibril through parameterizing bond coupling energy ɛh,ɛs,ɛf, structural entropies sh,ss,sf of helical, sheet, and fibrillar states, and number density ρ. The phase diagrams show the transition sequence is basically governed by ɛh, ɛs, and ɛf, while the transition temperature is determined by the competition among ɛh, ɛs, and ɛf, as well as sh, ss, sf, and ρ. Furthermore, the fibrillation is accompanied with an abrupt phase transition from coil, helix, or sheet to fibril even for short polypeptide length, resembling the feature of nucleation-growth process. The finite-size effect in specific heat at transitions for the nonfibrillation case can be described by the scaling form of lattice model. With rich phase-transition properties, our model provides a useful reference for protein aggregation experiments and modeling.

  8. Predictive processing simplified: The infotropic machine.

    PubMed

    Thornton, Chris

    2017-03-01

    On a traditional view of cognition, we see the agent acquiring stimuli, interpreting these in some way, and producing behavior in response. An increasingly popular alternative is the predictive processing framework. This sees the agent as continually generating predictions about the world, and responding productively to any errors made. Partly because of its heritage in the Bayesian brain theory, predictive processing has generally been seen as an inherently Bayesian process. The 'hierarchical prediction machine' which mediates it is envisaged to be a specifically Bayesian device. But as this paper shows, a specification for this machine can also be derived directly from information theory, using the metric of predictive payoff as an organizing concept. Hierarchical prediction machines can be built along purely information-theoretic lines, without referencing Bayesian theory in any way; this simplifies the account to some degree. The present paper describes what is involved and presents a series of working models. An experiment involving the conversion of a Braitenberg vehicle to use a controller of this type is also described.

  9. Analysis of Modeling Assumptions used in Production Cost Models for Renewable Integration Studies

    SciTech Connect

    Stoll, Brady; Brinkman, Gregory; Townsend, Aaron; Bloom, Aaron

    2016-01-01

    Renewable energy integration studies have been published for many different regions exploring the question of how higher penetration of renewable energy will impact the electric grid. These studies each make assumptions about the systems they are analyzing; however the effect of many of these assumptions has not been yet been examined and published. In this paper we analyze the impact of modeling assumptions in renewable integration studies, including the optimization method used (linear or mixed-integer programming) and the temporal resolution of the dispatch stage (hourly or sub-hourly). We analyze each of these assumptions on a large and a small system and determine the impact of each assumption on key metrics including the total production cost, curtailment of renewables, CO2 emissions, and generator starts and ramps. Additionally, we identified the impact on these metrics if a four-hour ahead commitment step is included before the dispatch step and the impact of retiring generators to reduce the degree to which the system is overbuilt. We find that the largest effect of these assumptions is at the unit level on starts and ramps, particularly for the temporal resolution, and saw a smaller impact at the aggregate level on system costs and emissions. For each fossil fuel generator type we measured the average capacity started, average run-time per start, and average number of ramps. Linear programming results saw up to a 20% difference in number of starts and average run time of traditional generators, and up to a 4% difference in the number of ramps, when compared to mixed-integer programming. Utilizing hourly dispatch instead of sub-hourly dispatch saw no difference in coal or gas CC units for either start metric, while gas CT units had a 5% increase in the number of starts and 2% increase in the average on-time per start. The number of ramps decreased up to 44%. The smallest effect seen was on the CO2 emissions and total production cost, with a 0.8% and 0

  10. Differentiating Different Modeling Assumptions in Simulations of MagLIF loads on the Z Generator

    NASA Astrophysics Data System (ADS)

    Jennings, C. A.; Gomez, M. R.; Harding, E. C.; Knapp, P. F.; Ampleford, D. J.; Hansen, S. B.; Weis, M. R.; Glinsky, M. E.; Peterson, K.; Chittenden, J. P.

    2016-10-01

    Metal liners imploded by a fast rising (<100ns) current to compress a magnetized, preheated fuel offer the potential to efficiently reach fusion conditions. These MagLIF experiments have had some success. While experiments are increasingly well diagnosed, many of the measurements (particularly during stagnation) are time integrated, limited in spatial resolution or require additional assumptions to interpret in the context of a structured, rapidly evolving system. As such, in validating MHD calculations, there is the potential for the same observables in the experimental data to be reproduced under different modeling assumptions. Using synthetic diagnostics of the results of different pre-heat, implosion and stagnation simulations run with the Gorgon MHD code, we discuss how the interpretation of typical Z diagnostics relate to more fundamental simulation parameters. We then explore the extent to which different assumptions on instability development, current delivery, high-Z mix into the fuel and initial laser deposition can be differentiated in our existing measurements. Sandia National Laboratories is a multi-program laboratory managed and operated by Sandia Corporation, a wholly owned subsidiary of Lockheed Martin Corporation, for the U.S. DoE's NNSA under contract DE-AC04-94AL85000.

  11. Underlying assumptions of human spaceflight in the United States

    NASA Astrophysics Data System (ADS)

    Launius, Roger D.

    2008-03-01

    Throughout the 50 years of the space age, a basic assumption has reigned that human exploration represented the epitome of the endeavor. Both the United States and the Soviet Union founded their efforts on that belief and remain to the date dominant forces in the human exploration and development of space. But why did the two superpowers locked in the cold war struggle pursue human spaceflight? The answer usually given involves the quest for pride and prestige in the cold war rivalry as each sought to best the other in a complex and expensive game of "one-upmanship." Without denying the significance of the cold war in prying open the treasury of the United States for human spaceflight, the endeavor seems to have been attractive in the first place because of several underlying assumptions that are deeply seated in the American values. This essay suggests that Americans embraced human space exploration because of its potential for extending human dominion into space and for the promise of colonization and expansion, although that has usually been at best a subtext for the effort. When thinking about these human activities over the long term, moreover, it raises important and difficult questions about the evolution and survival of the species.

  12. The extended evolutionary synthesis: its structure, assumptions and predictions

    PubMed Central

    Laland, Kevin N.; Uller, Tobias; Feldman, Marcus W.; Sterelny, Kim; Müller, Gerd B.; Moczek, Armin; Jablonka, Eva; Odling-Smee, John

    2015-01-01

    Scientific activities take place within the structured sets of ideas and assumptions that define a field and its practices. The conceptual framework of evolutionary biology emerged with the Modern Synthesis in the early twentieth century and has since expanded into a highly successful research program to explore the processes of diversification and adaptation. Nonetheless, the ability of that framework satisfactorily to accommodate the rapid advances in developmental biology, genomics and ecology has been questioned. We review some of these arguments, focusing on literatures (evo-devo, developmental plasticity, inclusive inheritance and niche construction) whose implications for evolution can be interpreted in two ways—one that preserves the internal structure of contemporary evolutionary theory and one that points towards an alternative conceptual framework. The latter, which we label the ‘extended evolutionary synthesis' (EES), retains the fundaments of evolutionary theory, but differs in its emphasis on the role of constructive processes in development and evolution, and reciprocal portrayals of causation. In the EES, developmental processes, operating through developmental bias, inclusive inheritance and niche construction, share responsibility for the direction and rate of evolution, the origin of character variation and organism–environment complementarity. We spell out the structure, core assumptions and novel predictions of the EES, and show how it can be deployed to stimulate and advance research in those fields that study or use evolutionary biology. PMID:26246559

  13. The extended evolutionary synthesis: its structure, assumptions and predictions.

    PubMed

    Laland, Kevin N; Uller, Tobias; Feldman, Marcus W; Sterelny, Kim; Müller, Gerd B; Moczek, Armin; Jablonka, Eva; Odling-Smee, John

    2015-08-22

    Scientific activities take place within the structured sets of ideas and assumptions that define a field and its practices. The conceptual framework of evolutionary biology emerged with the Modern Synthesis in the early twentieth century and has since expanded into a highly successful research program to explore the processes of diversification and adaptation. Nonetheless, the ability of that framework satisfactorily to accommodate the rapid advances in developmental biology, genomics and ecology has been questioned. We review some of these arguments, focusing on literatures (evo-devo, developmental plasticity, inclusive inheritance and niche construction) whose implications for evolution can be interpreted in two ways—one that preserves the internal structure of contemporary evolutionary theory and one that points towards an alternative conceptual framework. The latter, which we label the 'extended evolutionary synthesis' (EES), retains the fundaments of evolutionary theory, but differs in its emphasis on the role of constructive processes in development and evolution, and reciprocal portrayals of causation. In the EES, developmental processes, operating through developmental bias, inclusive inheritance and niche construction, share responsibility for the direction and rate of evolution, the origin of character variation and organism-environment complementarity. We spell out the structure, core assumptions and novel predictions of the EES, and show how it can be deployed to stimulate and advance research in those fields that study or use evolutionary biology.

  14. Do we need various assumptions to get a good FCN?

    NASA Astrophysics Data System (ADS)

    Huang, C.; Zhang, M.

    2015-08-01

    Free core nutation (FCN) is a rotational modes of the Earth with fluid core. All traditional theoretical methods produce FCN period near 460 sidereal days with PREM Earth model, while precise observations (VLBI + SG tides) say it is approximately 430 days. In order to fill this big gap, astronomers and geophysicists give various assumptions, e.g., increasing core-mantle-boundary (CMB) flattening by about 5%, a strong coupling between nutation and geomagnetic field near CMB, viscous coupling, or topographical coupling cross CMB, etc. Do we really need these unproved assumptions? or is it only the problem of these traditional theoretical methods themselves? Earth models (e.g., PREM) provide accurate and robust profiles of physical parameters, like density and Lame parameters, but their radial derivatives, which are also used in all traditional methods to calculate normal modes (e.g., FCN), nutation and tides of non-rigid Earth theoretically, are not so trustable as the parameters themselves. Moreover, the truncation of the expansion series of displacement vector and stress tensor in traditional methods is also of question. A new stratified spectral method is proposed and applied to the computation of normal modes, to avoid these problems. Our primary result of the FCN period is 435 ± 3 sidereal days.

  15. The contour method cutting assumption: error minimization and correction

    SciTech Connect

    Prime, Michael B; Kastengren, Alan L

    2010-01-01

    The recently developed contour method can measure 2-D, cross-sectional residual-stress map. A part is cut in two using a precise and low-stress cutting technique such as electric discharge machining. The contours of the new surfaces created by the cut, which will not be flat if residual stresses are relaxed by the cutting, are then measured and used to calculate the original residual stresses. The precise nature of the assumption about the cut is presented theoretically and is evaluated experimentally. Simply assuming a flat cut is overly restrictive and misleading. The critical assumption is that the width of the cut, when measured in the original, undeformed configuration of the body is constant. Stresses at the cut tip during cutting cause the material to deform, which causes errors. The effect of such cutting errors on the measured stresses is presented. The important parameters are quantified. Experimental procedures for minimizing these errors are presented. An iterative finite element procedure to correct for the errors is also presented. The correction procedure is demonstrated on experimental data from a steel beam that was plastically bent to put in a known profile of residual stresses.

  16. Make no assumptions: communication between persons with disabilities and clinicians.

    PubMed

    Iezzoni, Lisa I

    2006-01-01

    Fundamentally improving health care quality requires providing care that respects patients' preferences, needs, and values. This goal holds special resonance for persons with disabilities, many of whom find others defining and circumscribing their lives and opportunities. Achieving patient centeredness demands open communication between patients and clinicians, unhampered by prior and often erroneous assumptions about patients' goals, aspirations, and abilities. Building on this communication, optimal care involves collaboration between patients and clinicians, each bringing his or her particular expertise to the table. Interviews with individuals with diverse disabilities revealed a common theme of faulty communication between patients and clinicians. Some shortfalls relate to basic failures to accommodate communication needs, whereas others result from clinicians' erroneous perceptions of medical aspects of persons' underlying conditions, the role of assistive technologies, and how disability affects people's daily lives. Crafting collaborative care partnerships between patients and clinicians requires transforming traditional patient-clinician relationships. Following two basic precepts immeasurably improves communication between clinicians and patients with disabilities: first, make no assumptions, and second, just ask patients about their needs and preferences.

  17. Camera traps and mark-resight models: The value of ancillary data for evaluating assumptions

    USGS Publications Warehouse

    Parsons, Arielle W.; Simons, Theodore R.; Pollock, Kenneth H.; Stoskopf, Michael K.; Stocking, Jessica J.; O'Connell, Allan F.

    2015-01-01

    Unbiased estimators of abundance and density are fundamental to the study of animal ecology and critical for making sound management decisions. Capture–recapture models are generally considered the most robust approach for estimating these parameters but rely on a number of assumptions that are often violated but rarely validated. Mark-resight models, a form of capture–recapture, are well suited for use with noninvasive sampling methods and allow for a number of assumptions to be relaxed. We used ancillary data from continuous video and radio telemetry to evaluate the assumptions of mark-resight models for abundance estimation on a barrier island raccoon (Procyon lotor) population using camera traps. Our island study site was geographically closed, allowing us to estimate real survival and in situ recruitment in addition to population size. We found several sources of bias due to heterogeneity of capture probabilities in our study, including camera placement, animal movement, island physiography, and animal behavior. Almost all sources of heterogeneity could be accounted for using the sophisticated mark-resight models developed by McClintock et al. (2009b) and this model generated estimates similar to a spatially explicit mark-resight model previously developed for this population during our study. Spatially explicit capture–recapture models have become an important tool in ecology and confer a number of advantages; however, non-spatial models that account for inherent individual heterogeneity may perform nearly as well, especially where immigration and emigration are limited. Non-spatial models are computationally less demanding, do not make implicit assumptions related to the isotropy of home ranges, and can provide insights with respect to the biological traits of the local population.

  18. Simplified overturn stability monitoring of agricultural tractors.

    PubMed

    Nichol, C I; Sommer, H J; Murphy, D J

    2005-02-01

    Agricultural tractors are the most common source of farm work fatalities in the U.S., with overturns the most common type of incident. For the year 2001, there were 15 tractor-related fatalities in Pennsylvania, 9 of which were due to tractor rollover. A new device using low-cost sensors and microcomputers was developed around a simplified mathematical model of an agricultural tractor to inform the operator of potential tractor instability. This device communicates the current rollover potential, along with a recent history of rollover potential, to the operator of the tractor via a simple bar-graph display. The device uses a single-chip accelerometer to sense the current rollover potential and a small microprocessor to analyze the accelerometer data, compensate for variations due to temperature, and then send this information to a visual display. The use of these low-cost "off the shelf" components enabled the fabrication of a very inexpensive sensor system. Because agricultural tractors have a long service life, it was important to make the device low cost and flexible. This could enable it to be sold as an aftermarket add-on for a variety of tractor models. The device is also capable of interfacing with newer on-board tractor systems via a CAN bus to make it more attractive to tractor manufacturers who may want to incorporate this device into new models. Work is continuing on the development of an improved display to inform the tractor operator of possible instability, including display ergonomic studies, investigation of threshold levels for alerting an operator of potential instability, and investigation into audible warning signals.

  19. Simplified models for heat transfer in rooms

    NASA Astrophysics Data System (ADS)

    Graca, Guilherme C. C. Carrilho Da

    Buildings protect their occupants from the outside environment. As a semi-enclosed environment, buildings tend to contain the internally generated heat and air pollutants, as well as the solar and conductive heat gains that can occur in the facade. In the warmer months of the year this generally leads to overheating, creating a need for a cooling system. Ventilation air replaces contaminated air in the building and is often used as the dominant medium for heat transfer between indoor and outdoor environments. The goal of the research presented in this thesis is to develop a better understanding of the important parameters in the performance of ventilation systems and to develop simplified convective heat transfer models. The general approach used in this study seeks to capture the dominant physical processes for these problems with first order accuracy, and develop simple models that show the correct system behavior trends. Dimensional analysis, in conjunction with simple momentum and energy conservation, scaled model experiments and numerical simulations, is used to improve airflow and heat transfer rate predictions in both single and multi room ventilation systems. This study includes the three commonly used room ventilation modes: mixing, displacement and cross-ventilation. A new modeling approach to convective heat transfer between the building and the outside is presented: the concept of equivalent room heat transfer coefficient. The new model quantifies the reduction in heat transfer between ventilation air and internal room surfaces caused by limited thermal capacity and temperature variation of the air for the three modes studied. Particular emphasis is placed on cross-ventilation, and on the development of a simple model to characterize the airflow patterns that occur in this case. The implementation of the models in a building thermal simulation software tool is presented as well as comparisons between model predictions, experimental results and complex

  20. 12 CFR 217.144 - Simplified supervisory formula approach (SSFA).

    Code of Federal Regulations, 2014 CFR

    2014-01-01

    ... 12 Banks and Banking 2 2014-01-01 2014-01-01 false Simplified supervisory formula approach (SSFA... Simplified supervisory formula approach (SSFA). (a) General requirements for the SSFA. To use the SSFA to determine the risk weight for a securitization exposure, a Board-regulated institution must have data...

  1. 26 CFR 1.41-9 - Alternative simplified credit.

    Code of Federal Regulations, 2011 CFR

    2011-04-01

    ... 26 Internal Revenue 1 2011-04-01 2009-04-01 true Alternative simplified credit. 1.41-9 Section 1.41-9 Internal Revenue INTERNAL REVENUE SERVICE, DEPARTMENT OF THE TREASURY INCOME TAX INCOME TAXES Credits Against Tax § 1.41-9 Alternative simplified credit. For further guidance, see § 1.41-9T....

  2. 26 CFR 1.41-9 - Alternative simplified credit.

    Code of Federal Regulations, 2010 CFR

    2010-04-01

    ... 26 Internal Revenue 1 2010-04-01 2010-04-01 true Alternative simplified credit. 1.41-9 Section 1.41-9 Internal Revenue INTERNAL REVENUE SERVICE, DEPARTMENT OF THE TREASURY INCOME TAX INCOME TAXES Credits Against Tax § 1.41-9 Alternative simplified credit. For further guidance, see § 1.41-9T....

  3. Naturally Simplified Input, Comprehension, and Second Language Acquisition.

    ERIC Educational Resources Information Center

    Ellis, Rod

    This article examines the concept of simplification in second language (SL) learning, reviewing research on the simplified input that both naturalistic and classroom SL learners receive. Research indicates that simplified input, particularly if derived from naturally occurring interactions, does aid comprehension but has not been shown to…

  4. Communication: A simplified coupled-cluster Lagrangian for polarizable embedding

    SciTech Connect

    Krause, Katharina; Klopper, Wim

    2016-01-28

    A simplified coupled-cluster Lagrangian, which is linear in the Lagrangian multipliers, is proposed for the coupled-cluster treatment of a quantum mechanical system in a polarizable environment. In the simplified approach, the amplitude equations are decoupled from the Lagrangian multipliers and the energy obtained from the projected coupled-cluster equation corresponds to a stationary point of the Lagrangian.

  5. Potentialities of TEC topping: A simplified view of parametric effects

    NASA Technical Reports Server (NTRS)

    Morris, J. F.

    1980-01-01

    An examination of the benefits of thermionic-energy-conversion (TEC)-topped power plants and methods of increasing conversion efficiency are discussed. Reductions in the cost of TEC modules yield direct decreases in the cost of electricity (COE) from TEC-topped central station power plants. Simplified COE, overall-efficiency charts presented illustrate this trend. Additional capital-cost diminution results from designing more compact furnaces with considerably increased heat transfer rates allowable and desirable for high temperature TEC and heat pipes. Such improvements can evolve of the protection from hot corrosion and slag as well as the thermal expansion compatibilities offered by silicon-carbide clads on TEC-heating surfaces. Greater efficiencies and far fewer modules are possible with high-temperature, high-power-density TEC: This decreases capital and fuel costs much more and substantially increases electric power outputs for fixed fuel inputs. In addition to more electricity, less pollution, and lower costs, TEC topping used directly in coal-combustion products contributes balance-of-payment gains.

  6. Assessing the robustness of quantitative fatty acid signature analysis to assumption violations

    USGS Publications Warehouse

    Bromaghin, Jeffrey; Budge, Suzanne M.; Thiemann, Gregory W.; Rode, Karyn D.

    2016-01-01

    In most QFASA applications, investigators will generally have some knowledge of the prey available to predators and be able to assess the completeness of prey signature data and sample additional prey as necessary. Conversely, because calibration coefficients are derived from feeding trials with captive animals and their values may be sensitive to consumer physiology and nutritional status, their applicability to free-ranging animals is difficult to establish. We therefore recommend that investigators first make any improvements to the prey signature data that seem warranted and then base estimation on the Aitchison distance measure, as it appears to minimize risk from violations of the assumption that is most difficult to verify.

  7. Experimental assessment of unvalidated assumptions in classical plasticity theory.

    SciTech Connect

    Brannon, Rebecca Moss; Burghardt, Jeffrey A.; Bauer, Stephen J.; Bronowski, David R.

    2009-01-01

    This report investigates the validity of several key assumptions in classical plasticity theory regarding material response to changes in the loading direction. Three metals, two rock types, and one ceramic were subjected to non-standard loading directions, and the resulting strain response increments were displayed in Gudehus diagrams to illustrate the approximation error of classical plasticity theories. A rigorous mathematical framework for fitting classical theories to the data, thus quantifying the error, is provided. Further data analysis techniques are presented that allow testing for the effect of changes in loading direction without having to use a new sample and for inferring the yield normal and flow directions without having to measure the yield surface. Though the data are inconclusive, there is indication that classical, incrementally linear, plasticity theory may be inadequate over a certain range of loading directions. This range of loading directions also coincides with loading directions that are known to produce a physically inadmissible instability for any nonassociative plasticity model.

  8. Uncovering Metaethical Assumptions in Bioethical Discourse across Cultures.

    PubMed

    Sullivan, Laura Specker

    2016-03-01

    Much of bioethical discourse now takes place across cultures. This does not mean that cross-cultural understanding has increased. Many cross-cultural bioethical discussions are marked by entrenched disagreement about whether and why local practices are justified. In this paper, I argue that a major reason for these entrenched disagreements is that problematic metaethical commitments are hidden in these cross-cultural discourses. Using the issue of informed consent in East Asia as an example of one such discourse, I analyze two representative positions in the discussion and identify their metaethical commitments. I suggest that the metaethical assumptions of these positions result from their shared method of ethical justification: moral principlism. I then show why moral principlism is problematic in cross-cultural analyses and propose a more useful method for pursuing ethical justification across cultures.

  9. What lies beneath: underlying assumptions in bioimage analysis.

    PubMed

    Pridmore, Tony P; French, Andrew P; Pound, Michael P

    2012-12-01

    The need for plant image analysis tools is established and has led to a steadily expanding literature and set of software tools. This is encouraging, but raises a question: how does a plant scientist with no detailed knowledge or experience of image analysis methods choose the right tool(s) for the task at hand, or satisfy themselves that a suggested approach is appropriate? We believe that too great an emphasis is currently being placed on low-level mechanisms and software environments. In this opinion article we propose that a renewed focus on the core theories and algorithms used, and in particular the assumptions upon which they rely, will better equip plant scientists to evaluate the available resources.

  10. Deconstructing Community for Conservation: Why Simple Assumptions are Not Sufficient.

    PubMed

    Waylen, Kerry Ann; Fischer, Anke; McGowan, Philip J K; Milner-Gulland, E J

    2013-01-01

    Many conservation policies advocate engagement with local people, but conservation practice has sometimes been criticised for a simplistic understanding of communities and social context. To counter this, this paper explores social structuring and its influences on conservation-related behaviours at the site of a conservation intervention near Pipar forest, within the Seti Khola valley, Nepal. Qualitative and quantitative data from questionnaires and Rapid Rural Appraisal demonstrate how links between groups directly and indirectly influence behaviours of conservation relevance (including existing and potential resource-use and proconservation activities). For low-status groups the harvesting of resources can be driven by others' preference for wild foods, whilst perceptions of elite benefit-capture may cause reluctance to engage with future conservation interventions. The findings reiterate the need to avoid relying on simple assumptions about 'community' in conservation, and particularly the relevance of understanding relationships between groups, in order to understand natural resource use and implications for conservation.

  11. Cost and Performance Assumptions for Modeling Electricity Generation Technologies

    SciTech Connect

    Tidball, Rick; Bluestein, Joel; Rodriguez, Nick; Knoke, Stu

    2010-11-01

    The goal of this project was to compare and contrast utility scale power plant characteristics used in data sets that support energy market models. Characteristics include both technology cost and technology performance projections to the year 2050. Cost parameters include installed capital costs and operation and maintenance (O&M) costs. Performance parameters include plant size, heat rate, capacity factor or availability factor, and plant lifetime. Conventional, renewable, and emerging electricity generating technologies were considered. Six data sets, each associated with a different model, were selected. Two of the data sets represent modeled results, not direct model inputs. These two data sets include cost and performance improvements that result from increased deployment as well as resulting capacity factors estimated from particular model runs; other data sets represent model input data. For the technologies contained in each data set, the levelized cost of energy (LCOE) was also evaluated, according to published cost, performance, and fuel assumptions.

  12. Dynamic Group Diffie-Hellman Key Exchange under standard assumptions

    SciTech Connect

    Bresson, Emmanuel; Chevassut, Olivier; Pointcheval, David

    2002-02-14

    Authenticated Diffie-Hellman key exchange allows two principals communicating over a public network, and each holding public-private keys, to agree on a shared secret value. In this paper we study the natural extension of this cryptographic problem to a group of principals. We begin from existing formal security models and refine them to incorporate major missing details (e.g., strong-corruption and concurrent sessions). Within this model we define the execution of a protocol for authenticated dynamic group Diffie-Hellman and show that it is provably secure under the decisional Diffie-Hellman assumption. Our security result holds in the standard model and thus provides better security guarantees than previously published results in the random oracle model.

  13. Linear irreversible heat engines based on local equilibrium assumptions

    NASA Astrophysics Data System (ADS)

    Izumida, Yuki; Okuda, Koji

    2015-08-01

    We formulate an endoreversible finite-time Carnot cycle model based on the assumptions of local equilibrium and constant energy flux, where the efficiency and the power are expressed in terms of the thermodynamic variables of the working substance. By analyzing the entropy production rate caused by the heat transfer in each isothermal process during the cycle, and using the endoreversible condition applied to the linear response regime, we identify the thermodynamic flux and force of the present system and obtain a linear relation that connects them. We calculate the efficiency at maximum power in the linear response regime by using the linear relation, which agrees with the Curzon-Ahlborn (CA) efficiency known as the upper bound in this regime. This reason is also elucidated by rewriting our model into the form of the Onsager relations, where our model turns out to satisfy the tight-coupling condition leading to the CA efficiency.

  14. Polarized BRDF for coatings based on three-component assumption

    NASA Astrophysics Data System (ADS)

    Liu, Hong; Zhu, Jingping; Wang, Kai; Xu, Rong

    2017-02-01

    A pBRDF(polarized bidirectional reflection distribution function) model for coatings is given based on three-component reflection assumption in order to improve the polarized scattering simulation capability for space objects. In this model, the specular reflection is given based on microfacet theory, the multiple reflection and volume scattering are given separately according to experimental results. The polarization of specular reflection is considered from Fresnel's law, and both multiple reflection and volume scattering are assumed depolarized. Simulation and measurement results of two satellite coating samples SR107 and S781 are given to validate that the pBRDF modeling accuracy can be significantly improved by the three-component model given in this paper.

  15. Halo-independent direct detection analyses without mass assumptions

    SciTech Connect

    Anderson, Adam J.; Fox, Patrick J.; Kahn, Yonatan; McCullough, Matthew

    2015-10-06

    Results from direct detection experiments are typically interpreted by employing an assumption about the dark matter velocity distribution, with results presented in the m{sub χ}−σ{sub n} plane. Recently methods which are independent of the DM halo velocity distribution have been developed which present results in the v{sub min}−g-tilde plane, but these in turn require an assumption on the dark matter mass. Here we present an extension of these halo-independent methods for dark matter direct detection which does not require a fiducial choice of the dark matter mass. With a change of variables from v{sub min} to nuclear recoil momentum (p{sub R}), the full halo-independent content of an experimental result for any dark matter mass can be condensed into a single plot as a function of a new halo integral variable, which we call h-til-tilde(p{sub R}). The entire family of conventional halo-independent g-tilde(v{sub min}) plots for all DM masses are directly found from the single h-tilde(p{sub R}) plot through a simple rescaling of axes. By considering results in h-tilde(p{sub R}) space, one can determine if two experiments are inconsistent for all masses and all physically possible halos, or for what range of dark matter masses the results are inconsistent for all halos, without the necessity of multiple g-tilde(v{sub min}) plots for different DM masses. We conduct a sample analysis comparing the CDMS II Si events to the null results from LUX, XENON10, and SuperCDMS using our method and discuss how the results can be strengthened by imposing the physically reasonable requirement of a finite halo escape velocity.

  16. Halo-independent direct detection analyses without mass assumptions

    DOE PAGES

    Anderson, Adam J.; Fox, Patrick J.; Kahn, Yonatan; ...

    2015-10-06

    Results from direct detection experiments are typically interpreted by employing an assumption about the dark matter velocity distribution, with results presented in the mχ – σn plane. Recently methods which are independent of the DM halo velocity distribution have been developed which present results in the vmin – g~ plane, but these in turn require an assumption on the dark matter mass. Here we present an extension of these halo-independent methods for dark matter direct detection which does not require a fiducial choice of the dark matter mass. With a change of variables from vmin to nuclear recoil momentum (pR),more » the full halo-independent content of an experimental result for any dark matter mass can be condensed into a single plot as a function of a new halo integral variable, which we call tilde h(pR). The entire family of conventional halo-independent tilde g~(vmin) plots for all DM masses are directly found from the single tilde h~(pR) plot through a simple rescaling of axes. By considering results in tildeh~(pR) space, one can determine if two experiments are inconsistent for all masses and all physically possible halos, or for what range of dark matter masses the results are inconsistent for all halos, without the necessity of multiple tilde g~(vmin) plots for different DM masses. As a result, we conduct a sample analysis comparing the CDMS II Si events to the null results from LUX, XENON10, and SuperCDMS using our method and discuss how the results can be strengthened by imposing the physically reasonable requirement of a finite halo escape velocity.« less

  17. Halo-independent direct detection analyses without mass assumptions

    SciTech Connect

    Anderson, Adam J.; Fox, Patrick J.; Kahn, Yonatan; McCullough, Matthew

    2015-10-06

    Results from direct detection experiments are typically interpreted by employing an assumption about the dark matter velocity distribution, with results presented in the mχ – σn plane. Recently methods which are independent of the DM halo velocity distribution have been developed which present results in the vmin – g~ plane, but these in turn require an assumption on the dark matter mass. Here we present an extension of these halo-independent methods for dark matter direct detection which does not require a fiducial choice of the dark matter mass. With a change of variables from vmin to nuclear recoil momentum (pR), the full halo-independent content of an experimental result for any dark matter mass can be condensed into a single plot as a function of a new halo integral variable, which we call tilde h(pR). The entire family of conventional halo-independent tilde g~(vmin) plots for all DM masses are directly found from the single tilde h~(pR) plot through a simple rescaling of axes. By considering results in tildeh~(pR) space, one can determine if two experiments are inconsistent for all masses and all physically possible halos, or for what range of dark matter masses the results are inconsistent for all halos, without the necessity of multiple tilde g~(vmin) plots for different DM masses. As a result, we conduct a sample analysis comparing the CDMS II Si events to the null results from LUX, XENON10, and SuperCDMS using our method and discuss how the results can be strengthened by imposing the physically reasonable requirement of a finite halo escape velocity.

  18. Halo-independent direct detection analyses without mass assumptions

    SciTech Connect

    Anderson, Adam J.; Fox, Patrick J.; Kahn, Yonatan; McCullough, Matthew E-mail: pjfox@fnal.gov E-mail: matthew.mccullough@cern.ch

    2015-10-01

    Results from direct detection experiments are typically interpreted by employing an assumption about the dark matter velocity distribution, with results presented in the m{sub χ}−σ{sub n} plane. Recently methods which are independent of the DM halo velocity distribution have been developed which present results in the v{sub min}− g-tilde plane, but these in turn require an assumption on the dark matter mass. Here we present an extension of these halo-independent methods for dark matter direct detection which does not require a fiducial choice of the dark matter mass. With a change of variables from v{sub min} to nuclear recoil momentum (p{sub R}), the full halo-independent content of an experimental result for any dark matter mass can be condensed into a single plot as a function of a new halo integral variable, which we call h-tilde (p{sub R}). The entire family of conventional halo-independent g-tilde (v{sub min}) plots for all DM masses are directly found from the single h-tilde (p{sub R}) plot through a simple rescaling of axes. By considering results in h-tilde (p{sub R}) space, one can determine if two experiments are inconsistent for all masses and all physically possible halos, or for what range of dark matter masses the results are inconsistent for all halos, without the necessity of multiple g-tilde (v{sub min}) plots for different DM masses. We conduct a sample analysis comparing the CDMS II Si events to the null results from LUX, XENON10, and SuperCDMS using our method and discuss how the results can be strengthened by imposing the physically reasonable requirement of a finite halo escape velocity.

  19. Halo-independent direct detection analyses without mass assumptions

    NASA Astrophysics Data System (ADS)

    Anderson, Adam J.; Fox, Patrick J.; Kahn, Yonatan; McCullough, Matthew

    2015-10-01

    Results from direct detection experiments are typically interpreted by employing an assumption about the dark matter velocity distribution, with results presented in the mχ-σn plane. Recently methods which are independent of the DM halo velocity distribution have been developed which present results in the vmin-tilde g plane, but these in turn require an assumption on the dark matter mass. Here we present an extension of these halo-independent methods for dark matter direct detection which does not require a fiducial choice of the dark matter mass. With a change of variables from vmin to nuclear recoil momentum (pR), the full halo-independent content of an experimental result for any dark matter mass can be condensed into a single plot as a function of a new halo integral variable, which we call tilde h(pR). The entire family of conventional halo-independent tilde g(vmin) plots for all DM masses are directly found from the single tilde h(pR) plot through a simple rescaling of axes. By considering results in tilde h(pR) space, one can determine if two experiments are inconsistent for all masses and all physically possible halos, or for what range of dark matter masses the results are inconsistent for all halos, without the necessity of multiple tilde g(vmin) plots for different DM masses. We conduct a sample analysis comparing the CDMS II Si events to the null results from LUX, XENON10, and SuperCDMS using our method and discuss how the results can be strengthened by imposing the physically reasonable requirement of a finite halo escape velocity.

  20. Estimating ETAS: the effects of truncation, missing data, and model assumptions

    NASA Astrophysics Data System (ADS)

    Seif, Stefanie; Mignan, Arnaud; Zechar, Jeremy; Werner, Maximilian; Wiemer, Stefan

    2016-04-01

    The Epidemic-Type Aftershock Sequence (ETAS) model is widely used to describe the occurrence of earthquakes in space and time, but there has been little discussion of the limits of, and influences on, its estimation. What has been established is that ETAS parameter estimates are influenced by missing data (e.g., earthquakes are not reliably detected during lively aftershock sequences) and by simplifying assumptions (e.g., that aftershocks are isotropically distributed). In this article, we investigate the effect of truncation: how do parameter estimates depend on the cut-off magnitude, Mcut, above which parameters are estimated? We analyze catalogs from southern California and Italy and find that parameter variations as a function of Mcut are caused by (i) changing sample size (which affects e.g. Omori's cconstant) or (ii) an intrinsic dependence on Mcut (as Mcut increases, absolute productivity and background rate decrease). We also explore the influence of another form of truncation - the finite catalog length - that can bias estimators of the branching ratio. Being also a function of Omori's p-value, the true branching ratio is underestimated by 45% to 5% for 1.05< p <1.2. Finite sample size affects the variation of the branching ratio estimates. Moreover, we investigate the effect of missing aftershocks and find that the ETAS productivity parameters (α and K0) and the Omoris c-value are significantly changed only for low Mcut=2.5. We further find that conventional estimation errors for these parameters, inferred from simulations that do not account for aftershock incompleteness, are underestimated by, on average, a factor of six.

  1. Robustness of closed capture-recapture methods to violations of the closure assumption

    USGS Publications Warehouse

    Kendall, W.L.

    1999-01-01

    Closed-population capture-recapture methods have been used extensively in animal ecology, both by themselves and within the context of Pollock?s robust design and multistate models, to estimate various parameters of population and community dynamics. The defining assumption of geographic and demographic closure (i.e., no births, deaths, immigration, or emigration) for the duration of sampling is restrictive, and likely to be violated in many field situations. I evaluated several types of violations of the closure assumption and found that completely random movement in and out of a study area does not introduce bias to estimators from closed-population methods, although it decreases precision. In addition, if capture probabilities vary only with time, the closed-population Lincoln-Petersen estimator is unbiased for the size of the superpopulation when there are only births/immigration or only deaths/emigration. However, for other cases of nonrandom movement closed-population estimators were biased when movement was Markovian (dependent on the presence/absence of the animal in the previous time period), when an animal was allowed one entry to and one exit from the study area, or when there was trap response or heterogeneity among animals in capture probability. In addition, the probability that an animal is present and available for capture (e.g., breeding propensity) can be estimated using Pollock?s robust design only when movement occurs at a broader temporal scale than that of sampling.

  2. Simplified Models for Dark Matter and Missing Energy Searches at the LHC

    SciTech Connect

    Abdallah, Jalal; Ashkenazi, Adi; Boveia, Antonio; Busoni, Giorgio; De Simone, Andrea; Doglioni, Caterina; Efrati, Aielet; Etzion, Erez; Gramling, Johanna; Jacques, Thomas; Lin, Tongyan; Morgante, Enrico; Papucci, Michele; Penning, Bjoern; Riotto, Antonio Walter; Rizzo, Thomas; Salek, David; Schramm, Steven; Slone, Oren; Soreq, Yotam; Vichi, Alessandro; Volansky, Tomer; Yavin, Itay; Zhou, Ning; Zurek, Kathryn

    2014-10-01

    The study of collision events with missing energy as searches for the dark matter (DM) component of the Universe are an essential part of the extensive program looking for new physics at the LHC. Given the unknown nature of DM, the interpretation of such searches should be made broad and inclusive. This report reviews the usage of simplified models in the interpretation of missing energy searches. We begin with a brief discussion of the utility and limitation of the effective field theory approach to this problem. The bulk of the report is then devoted to several different simplified models and their signatures, including s-channel and t-channel processes. A common feature of simplified models for DM is the presence of additional particles that mediate the interactions between the Standard Model and the particle that makes up DM. We consider these in detail and emphasize the importance of their inclusion as final states in any coherent interpretation. We also review some of the experimental progress in the field, new signatures, and other aspects of the searches themselves. We conclude with comments and recommendations regarding the use of simplified models in Run-II of the LHC.

  3. Analysis of Multivariate Experimental Data Using A Simplified Regression Model Search Algorithm

    NASA Technical Reports Server (NTRS)

    Ulbrich, Norbert M.

    2013-01-01

    A new regression model search algorithm was developed that may be applied to both general multivariate experimental data sets and wind tunnel strain-gage balance calibration data. The algorithm is a simplified version of a more complex algorithm that was originally developed for the NASA Ames Balance Calibration Laboratory. The new algorithm performs regression model term reduction to prevent overfitting of data. It has the advantage that it needs only about one tenth of the original algorithm's CPU time for the completion of a regression model search. In addition, extensive testing showed that the prediction accuracy of math models obtained from the simplified algorithm is similar to the prediction accuracy of math models obtained from the original algorithm. The simplified algorithm, however, cannot guarantee that search constraints related to a set of statistical quality requirements are always satisfied in the optimized regression model. Therefore, the simplified algorithm is not intended to replace the original algorithm. Instead, it may be used to generate an alternate optimized regression model of experimental data whenever the application of the original search algorithm fails or requires too much CPU time. Data from a machine calibration of NASA's MK40 force balance is used to illustrate the application of the new search algorithm.

  4. Analysis of Multivariate Experimental Data Using A Simplified Regression Model Search Algorithm

    NASA Technical Reports Server (NTRS)

    Ulbrich, Norbert Manfred

    2013-01-01

    A new regression model search algorithm was developed in 2011 that may be used to analyze both general multivariate experimental data sets and wind tunnel strain-gage balance calibration data. The new algorithm is a simplified version of a more complex search algorithm that was originally developed at the NASA Ames Balance Calibration Laboratory. The new algorithm has the advantage that it needs only about one tenth of the original algorithm's CPU time for the completion of a search. In addition, extensive testing showed that the prediction accuracy of math models obtained from the simplified algorithm is similar to the prediction accuracy of math models obtained from the original algorithm. The simplified algorithm, however, cannot guarantee that search constraints related to a set of statistical quality requirements are always satisfied in the optimized regression models. Therefore, the simplified search algorithm is not intended to replace the original search algorithm. Instead, it may be used to generate an alternate optimized regression model of experimental data whenever the application of the original search algorithm either fails or requires too much CPU time. Data from a machine calibration of NASA's MK40 force balance is used to illustrate the application of the new regression model search algorithm.

  5. Simplified multi-track detection schemes using a priori information for bit patterned media recording

    NASA Astrophysics Data System (ADS)

    Kong, Gyuyeol; Choi, Sooyong

    2012-04-01

    Simplified multi-track detection schemes using a priori information for bit patterned magnetic recording (BPMR) are proposed in this paper. The proposed detection schemes adopt the simplified trellis diagram, use a priori information, and detect the main-track data in the along- and cross-track directions. The simplified trellis diagram, which has 4 states and 8 branches, can be obtained by setting the corner entries of the generalized partial response (GPR) target to zero and replacing the four parallel branches with a single branch. However, these simplified techniques seriously suffer from performance degradation in high density BPMR channels. To overcome the performance degradation, a priori information is used to give higher reliability to the branch metric. In addition, to fully use the characteristics of channel detection with a two-dimensional (2D) GPR target, the proposed schemes estimate a priori information and detect the main-track data in the along- and cross-track directions by using a 2D equalizer with a 2D GPR target. The bit error rate performances of the proposed schemes are compared with the previous detection schemes when areal density is 3 Tb/in2. Simulation results show that the proposed schemes with simpler structures have more than 2 dB gains compared with the other detection schemes.

  6. Simplified Models for Dark Matter Face their Consistent Completions

    SciTech Connect

    Goncalves, Dorival; Machado, Pedro N.; No, Jose Miguel

    2016-11-14

    Simplified dark matter models have been recently advocated as a powerful tool to exploit the complementarity between dark matter direct detection, indirect detection and LHC experimental probes. Focusing on pseudoscalar mediators between the dark and visible sectors, we show that the simplified dark matter model phenomenology departs significantly from that of consistent ${SU(2)_{\\mathrm{L}} \\times U(1)_{\\mathrm{Y}}}$ gauge invariant completions. We discuss the key physics simplified models fail to capture, and its impact on LHC searches. Notably, we show that resonant mono-Z searches provide competitive sensitivities to standard mono-jet analyses at $13$ TeV LHC.

  7. Simplified models for dark matter searches at the LHC

    SciTech Connect

    Abdallah, Jalal; Araujo, Henrique; Arbey, Alexandre; Ashkenazi, Adi; Belyaev, Alexander; Berger, Joshua; Boehm, Celine; Boveia, Antonio; Brennan, Amelia; Brooke, Jim; Buchmueller, Oliver; Buckley, Matthew; Busoni, Giorgio; Calibbi, Lorenzo; Chauhan, Sushil; Daci, Nadir; Davies, Gavin; De Bruyn, Isabelle; De Jong, Paul; De Roeck, Albert; de Vries, Kees; Del Re, Daniele; De Simone, Andrea; Di Simone, Andrea; Doglioni, Caterina; Dolan, Matthew; Dreiner, Herbi K.; Ellis, John; Eno, Sarah; Etzion, Erez; Fairbairn, Malcolm; Feldstein, Brian; Flaecher, Henning; Feng, Eric; Fox, Patrick; Genest, Marie-Hélène; Gouskos, Loukas; Gramling, Johanna; Haisch, Ulrich; Harnik, Roni; Hibbs, Anthony; Hoh, Siewyan; Hopkins, Walter; Ippolito, Valerio; Jacques, Thomas; Kahlhoefer, Felix; Khoze, Valentin V.; Kirk, Russell; Korn, Andreas; Kotov, Khristian; Kunori, Shuichi; Landsberg, Greg; Liem, Sebastian; Lin, Tongyan; Lowette, Steven; Lucas, Robyn; Malgeri, Luca; Malik, Sarah; McCabe, Christopher; Mete, Alaettin Serhan; Morgante, Enrico; Mrenna, Stephen; Nakahama, Yu; Newbold, Dave; Nordstrom, Karl; Pani, Priscilla; Papucci, Michele; Pataraia, Sophio; Penning, Bjoern; Pinna, Deborah; Polesello, Giacomo; Racco, Davide; Re, Emanuele; Riotto, Antonio Walter; Rizzo, Thomas; Salek, David; Sarkar, Subir; Schramm, Steven; Skubic, Patrick; Slone, Oren; Smirnov, Juri; Soreq, Yotam; Sumner, Timothy; Tait, Tim M. P.; Thomas, Marc; Tomalin, Ian; Tunnell, Christopher; Vichi, Alessandro; Volansky, Tomer; Weiner, Neal; West, Stephen M.; Wielers, Monika; Worm, Steven; Yavin, Itay; Zaldivar, Bryan; Zhou, Ning; Zurek, Kathryn

    2015-09-01

    This document a outlines a set of simplified models for dark matter and its interactions with Standard Model particles. It is intended to summarize the main characteristics that these simplified models have when applied to dark matter searches at the LHC, and to provide a number of useful expressions for reference. The list of models includes both s-channel and t-channel scenarios. For s-channel, spin-0 and spin-1 mediations are discussed, and also realizations where the Higgs particle provides a portal between the dark and visible sectors. The guiding principles underpinning the proposed simplified models are spelled out, and some suggestions for implementation are presented.

  8. Liquid-filled simplified hollow-core photonic crystal fiber

    NASA Astrophysics Data System (ADS)

    Liu, Shengnan; Gao, Wei; Li, Hongwei; Dong, Yongkang; Zhang, Hongying

    2014-12-01

    We report on a novel type of liquid-filled simplified hollow-core photonic crystal fibers (HC-PCFs), and investigate their transmission properties with various filling liquids, including water, ethanol and FC-40. The loss and dispersion characterizations are calculated for different fiber parameters including strut thickness and core diameter. The results show that there are still low-loss windows existing for liquid-filled simplified HC-PCFs, and the low-loss windows and dispersions can be easily tailored by filling different liquids. Such liquid-filled simplified HC-PCFs open up many possibilities for nonlinear fiber optics, optical, biochemical and medical sensing.

  9. Simplified models for dark matter searches at the LHC

    NASA Astrophysics Data System (ADS)

    Abdallah, Jalal; Araujo, Henrique; Arbey, Alexandre; Ashkenazi, Adi; Belyaev, Alexander; Berger, Joshua; Boehm, Celine; Boveia, Antonio; Brennan, Amelia; Brooke, Jim; Buchmueller, Oliver; Buckley, Matthew; Busoni, Giorgio; Calibbi, Lorenzo; Chauhan, Sushil; Daci, Nadir; Davies, Gavin; De Bruyn, Isabelle; De Jong, Paul; De Roeck, Albert; de Vries, Kees; Del Re, Daniele; De Simone, Andrea; Di Simone, Andrea; Doglioni, Caterina; Dolan, Matthew; Dreiner, Herbi K.; Ellis, John; Eno, Sarah; Etzion, Erez; Fairbairn, Malcolm; Feldstein, Brian; Flaecher, Henning; Feng, Eric; Fox, Patrick; Genest, Marie-Hélène; Gouskos, Loukas; Gramling, Johanna; Haisch, Ulrich; Harnik, Roni; Hibbs, Anthony; Hoh, Siewyan; Hopkins, Walter; Ippolito, Valerio; Jacques, Thomas; Kahlhoefer, Felix; Khoze, Valentin V.; Kirk, Russell; Korn, Andreas; Kotov, Khristian; Kunori, Shuichi; Landsberg, Greg; Liem, Sebastian; Lin, Tongyan; Lowette, Steven; Lucas, Robyn; Malgeri, Luca; Malik, Sarah; McCabe, Christopher; Mete, Alaettin Serhan; Morgante, Enrico; Mrenna, Stephen; Nakahama, Yu; Newbold, Dave; Nordstrom, Karl; Pani, Priscilla; Papucci, Michele; Pataraia, Sophio; Penning, Bjoern; Pinna, Deborah; Polesello, Giacomo; Racco, Davide; Re, Emanuele; Riotto, Antonio Walter; Rizzo, Thomas; Salek, David; Sarkar, Subir; Schramm, Steven; Skubic, Patrick; Slone, Oren; Smirnov, Juri; Soreq, Yotam; Sumner, Timothy; Tait, Tim M. P.; Thomas, Marc; Tomalin, Ian; Tunnell, Christopher; Vichi, Alessandro; Volansky, Tomer; Weiner, Neal; West, Stephen M.; Wielers, Monika; Worm, Steven; Yavin, Itay; Zaldivar, Bryan; Zhou, Ning; Zurek, Kathryn

    2015-09-01

    This document outlines a set of simplified models for dark matter and its interactions with Standard Model particles. It is intended to summarize the main characteristics that these simplified models have when applied to dark matter searches at the LHC, and to provide a number of useful expressions for reference. The list of models includes both ss-channel and tt-channel scenarios. For ss-channel, spin-0 and spin-1 mediations are discussed, and also realizations where the Higgs particle provides a portal between the dark and visible sectors. The guiding principles underpinning the proposed simplified models are spelled out, and some suggestions for implementation are presented.

  10. Simplified model for fouling of a pleated membrane filter

    NASA Astrophysics Data System (ADS)

    Sanaei, Pejman; Cummings, Linda

    2014-11-01

    Pleated filter cartridge are widely used to remove undesired impurities from a fluid. A filter membrane is sandwiched between porous support layers, then pleated and packed in to an annular cylindrical cartridge. Although this arrangement offers a high ratio of surface filtration area to volume, the filter performance (measured, e.g., by graph of total flux versus throughput for a given pressure drop), is not as good as a flat filter membrane. The reasons for this difference in performance are currently unclear, but likely factors include the additional resistance of the porous support layers upstream and downstream of the membrane, the pleat packing density (PPD) and possible damage to the membrane during the pleating process. To investigate this, we propose a simplified mathematical model of the filtration within a single pleat. We consider the fluid dynamics through the membrane and support layers, and propose a model by which the pores of the membrane become fouled (i) by particles smaller than the membrane pore size; and (ii) by particles larger than the pores.We present some simulations of our model, investigating how flow and fouling differ between not only flat and pleated membranes, but also for support layers of different permeability profiles. NSF DMS-1261596.

  11. Simplified two and three dimensional HTTR benchmark problems

    SciTech Connect

    Zhan Zhang; Dingkang Zhang; Justin M. Pounders; Abderrafi M. Ougouag

    2011-05-01

    To assess the accuracy of diffusion or transport methods for reactor calculations, it is desirable to create heterogeneous benchmark problems that are typical of whole core configurations. In this paper we have created two and three dimensional numerical benchmark problems typical of high temperature gas cooled prismatic cores. Additionally, a single cell and single block benchmark problems are also included. These problems were derived from the HTTR start-up experiment. Since the primary utility of the benchmark problems is in code-to-code verification, minor details regarding geometry and material specification of the original experiment have been simplified while retaining the heterogeneity and the major physics properties of the core from a neutronics viewpoint. A six-group material (macroscopic) cross section library has been generated for the benchmark problems using the lattice depletion code HELIOS. Using this library, Monte Carlo solutions are presented for three configurations (all-rods-in, partially-controlled and all-rods-out) for both the 2D and 3D problems. These solutions include the core eigenvalues, the block (assembly) averaged fission densities, local peaking factors, the absorption densities in the burnable poison and control rods, and pin fission density distribution for selected blocks. Also included are the solutions for the single cell and single block problems.

  12. Triplet repeat primed PCR simplifies testing for Huntington disease.

    PubMed

    Jama, Mohamed; Millson, Alison; Miller, Christine E; Lyon, Elaine

    2013-03-01

    Diagnostic and predictive testing for Huntington disease (HD) requires an accurate determination of the number of CAG repeats in the Huntingtin (HHT) gene. Currently, when a sample appears to be homozygous for a normal allele, additional testing is required to confirm amplification from both alleles. If the sample still appears homozygous, Southern blot analysis is performed to rule out an undetected expanded HTT allele. Southern blot analysis is expensive, time-consuming, and labor intensive and requires high concentrations of DNA. We have developed a chimeric PCR process to help streamline workflow; true homozygous alleles are easily distinguished by this simplified method, and only very large expanded alleles still require Southern blot analysis. Two hundred forty-six HD samples, previously run with a different fragment analysis method, were analyzed with our new method. All samples were correctly genotyped, resulting in 100% concordance between the methods. The chimeric PCR assay was able to identify expanded alleles up to >150 CAG repeats. This method offers a simple strategy to differentiate normal from expanded CAG alleles, thereby reducing the number of samples reflexed to Southern blot analysis. It also provides assurance that expanded alleles are not routinely missed because of allele dropout.

  13. Simplified training for hazardous materials management in developing countries

    SciTech Connect

    Braithwaite, J.

    1994-12-31

    There are thousands of dangerous situations happening daily in developing countries around the world involving untrained workers and hazardous materials. There are very few if any agencies in developing countries that are charged with ensuring safe and healthful working conditions. In addition to the problem of regulation and enforcement, there are potential training problems due to the level of literacy and degree of scientific background of these workers. Many of these workers are refugees from poorly developed countries who are willing to work no matter what the conditions. Training methods (standards) accepted as state of the art in the United States and other developed countries may not work well under the conditions found in developing countries. Because these methods may not be appropriate, new and novel ways to train workers quickly, precisely and economically in hazardous materials management should be developed. One approach is to develop training programs that use easily recognizable graphics with minimal verbal instruction, programs similar to the type used to teach universal international driving regulations and safety. The program as outlined in this paper could be tailored to any sized plant and any hazardous material handling or exposure situation. The situation in many developing countries is critical, development of simplified training methods for workers exposed to hazardous materials hold valuable market potential and are an opportunity for many underdeveloped countries to develop indigenous expertise in hazardous materials management.

  14. Description of an exact recursive method to simplify shading calculations

    SciTech Connect

    Nawrocki, A.D.; Kammerud, R.

    1981-04-01

    An exact, recursive method called SHADE is described which attempts to simplify shading calculations as performed by a programmable calculator or microcomputer. Preliminary applications of SHADE using a Hewlett Packard HP-41C programmable calculation are outlined. In particular, for a given solar hour, SHADE is used to compute the following quantities for overhang and side fin combinations which shade various openings: the percentage of the total area of the opening which is shaded; the shaded area itself; the cosine of the angle of incidence between sun and glazing surface; the direct insolation at this surface, with and without shading; and the direct solar power at this surface, with and without shading. Hence, in its present HP-41C application, SHADE can be used in preliminary design and comparative analyses of shading devices on an hourly, daily, or seasonal basis, provided that: (1) the fins and overhangs be square or rectangular, and lie in planes perpendicular to the plane of the opening; and (2) the opening itself be vertical and rectangular with arbitrary building azimuths. Design candidates include conventional overhangs and side fins, porches, and reveals. In principle, SHADE can be extended to awnings, slatted sun screens, and bevelled recesses; in addition, its HP-41C application can be extended to calculations of direct solar gain through vertical and non-vertical glazings, thereby providing a more useful tool in building heating and cooling load calculations.

  15. Review of simplified Pseudo-two-Dimensional models of lithium-ion batteries

    NASA Astrophysics Data System (ADS)

    Jokar, Ali; Rajabloo, Barzin; Désilets, Martin; Lacroix, Marcel

    2016-09-01

    Over the last decade, many efforts have been deployed to develop models for the prediction, the control, the optimization and the parameter estimation of Lithium-ion (Li-ion) batteries. It appears that the most successful electrochemical-based model for Li-ion battery is the Pseudo-two-Dimensional model (P2D). Due to the fact that the governing equations are complex, this model cannot be used in real-time applications like Battery Management Systems (BMSs). To remedy the situation, several investigations have been carried out to simplify the P2D model. Mathematical and physical techniques are employed to reduce the order of magnitude of the P2D governing equations. The present paper is a review of the studies on the modeling of Li-ion batteries with simplified P2D models. The assumptions on which these models rest are stated, the calculation methods are examined, the advantages and the drawbacks of the models are discussed and their applications are presented. Suggestions for overcoming the shortcomings of the models are made. Challenges and future directions in the modeling of Li-ion batteries are also discussed.

  16. Testing the habituation assumption underlying models of parasitoid foraging behavior

    PubMed Central

    Abram, Katrina; Colazza, Stefano; Peri, Ezio

    2017-01-01

    Background Habituation, a form of non-associative learning, has several well-defined characteristics that apply to a wide range of physiological and behavioral responses in many organisms. In classic patch time allocation models, habituation is considered to be a major mechanistic component of parasitoid behavioral strategies. However, parasitoid behavioral responses to host cues have not previously been tested for the known, specific characteristics of habituation. Methods In the laboratory, we tested whether the foraging behavior of the egg parasitoid Trissolcus basalis shows specific characteristics of habituation in response to consecutive encounters with patches of host (Nezara viridula) chemical contact cues (footprints), in particular: (i) a training interval-dependent decline in response intensity, and (ii) a training interval-dependent recovery of the response. Results As would be expected of a habituated response, wasps trained at higher frequencies decreased their behavioral response to host footprints more quickly and to a greater degree than those trained at low frequencies, and subsequently showed a more rapid, although partial, recovery of their behavioral response to host footprints. This putative habituation learning could not be blocked by cold anesthesia, ingestion of an ATPase inhibitor, or ingestion of a protein synthesis inhibitor. Discussion Our study provides support for the assumption that diminishing responses of parasitoids to chemical indicators of host presence constitutes habituation as opposed to sensory fatigue, and provides a preliminary basis for exploring the underlying mechanisms. PMID:28321365

  17. Finite Element Simulations to Explore Assumptions in Kolsky Bar Experiments.

    SciTech Connect

    Crum, Justin

    2015-08-05

    The chief purpose of this project has been to develop a set of finite element models that attempt to explore some of the assumptions in the experimental set-up and data reduction of the Kolsky bar experiment. In brief, the Kolsky bar, sometimes referred to as the split Hopkinson pressure bar, is an experimental apparatus used to study the mechanical properties of materials at high strain rates. Kolsky bars can be constructed to conduct experiments in tension or compression, both of which are studied in this paper. The basic operation of the tension Kolsky bar is as follows: compressed air is inserted into the barrel that contains the striker; the striker accelerates towards the left and strikes the left end of the barrel producing a tensile stress wave that propogates first through the barrel and then down the incident bar, into the specimen, and finally the transmission bar. In the compression case, the striker instead travels to the right and impacts the incident bar directly. As the stress wave travels through an interface (e.g., the incident bar to specimen connection), a portion of the pulse is transmitted and the rest reflected. The incident pulse, as well as the transmitted and reflected pulses are picked up by two strain gauges installed on the incident and transmitted bars as shown. By interpreting the data acquired by these strain gauges, the stress/strain behavior of the specimen can be determined.

  18. Cleanup of contaminated soil -- Unreal risk assumptions: Contaminant degradation

    SciTech Connect

    Schiffman, A.

    1995-12-31

    Exposure assessments for development of risk-based soil cleanup standards or criteria assume that contaminant mass in soil is infinite and conservative (constant concentration). This assumption is not real for most organic chemicals. Contaminant mass is lost from soil and ground water when organic chemicals degrade. Factors to correct for chemical mass lost by degradation are derived from first-order kinetics for 85 organic chemicals commonly listed by USEPA and state agencies. Soil cleanup criteria, based on constant concentration, are then corrected for contaminant mass lost. For many chemicals, accounting for mass lost yields large correction factors to risk-based soil concentrations. For degradation in ground water and soil, correction factors range from greater than one to several orders of magnitude. The long exposure durations normally used in exposure assessments (25 to 70 years) result in large correction factors to standards even for carcinogenic chemicals with long half-lives. For the ground water pathway, a typical soil criterion for TCE of 1 mg/kg would be corrected to 11 mg/kg. For noncarcinogens, correcting for mass lost means that risk algorithms used to set soil cleanup requirements are inapplicable for many chemicals, especially for long periods of exposure.

  19. On Some Unwarranted Tacit Assumptions in Cognitive Neuroscience†

    PubMed Central

    Mausfeld, Rainer

    2011-01-01

    The cognitive neurosciences are based on the idea that the level of neurons or neural networks constitutes a privileged level of analysis for the explanation of mental phenomena. This paper brings to mind several arguments to the effect that this presumption is ill-conceived and unwarranted in light of what is currently understood about the physical principles underlying mental achievements. It then scrutinizes the question why such conceptions are nevertheless currently prevailing in many areas of psychology. The paper argues that corresponding conceptions are rooted in four different aspects of our common-sense conception of mental phenomena and their explanation, which are illegitimately transferred to scientific enquiry. These four aspects pertain to the notion of explanation, to conceptions about which mental phenomena are singled out for enquiry, to an inductivist epistemology, and, in the wake of behavioristic conceptions, to a bias favoring investigations of input–output relations at the expense of enquiries into internal principles. To the extent that the cognitive neurosciences methodologically adhere to these tacit assumptions, they are prone to turn into a largely a-theoretical and data-driven endeavor while at the same time enhancing the prospects for receiving widespread public appreciation of their empirical findings. PMID:22435062

  20. Observing gravitational-wave transient GW150914 with minimal assumptions

    NASA Astrophysics Data System (ADS)

    Abbott, B. P.; Abbott, R.; Abbott, T. D.; Abernathy, M. R.; Acernese, F.; Ackley, K.; Adams, C.; Adams, T.; Addesso, P.; Adhikari, R. X.; Adya, V. B.; Affeldt, C.; Agathos, M.; Agatsuma, K.; Aggarwal, N.; Aguiar, O. D.; Aiello, L.; Ain, A.; Ajith, P.; Allen, B.; Allocca, A.; Altin, P. A.; Anderson, S. B.; Anderson, W. G.; Arai, K.; Araya, M. C.; Arceneaux, C. C.; Areeda, J. S.; Arnaud, N.; Arun, K. G.; Ascenzi, S.; Ashton, G.; Ast, M.; Aston, S. M.; Astone, P.; Aufmuth, P.; Aulbert, C.; Babak, S.; Bacon, P.; Bader, M. K. M.; Baker, P. T.; Baldaccini, F.; Ballardin, G.; Ballmer, S. W.; Barayoga, J. C.; Barclay, S. E.; Barish, B. C.; Barker, D.; Barone, F.; Barr, B.; Barsotti, L.; Barsuglia, M.; Barta, D.; Bartlett, J.; Bartos, I.; Bassiri, R.; Basti, A.; Batch, J. C.; Baune, C.; Bavigadda, V.; Bazzan, M.; Behnke, B.; Bejger, M.; Bell, A. S.; Bell, C. J.; Berger, B. K.; Bergman, J.; Bergmann, G.; Berry, C. P. L.; Bersanetti, D.; Bertolini, A.; Betzwieser, J.; Bhagwat, S.; Bhandare, R.; Bilenko, I. A.; Billingsley, G.; Birch, J.; Birney, R.; Biscans, S.; Bisht, A.; Bitossi, M.; Biwer, C.; Bizouard, M. A.; Blackburn, J. K.; Blackburn, L.; Blair, C. D.; Blair, D. G.; Blair, R. M.; Bloemen, S.; Bock, O.; Bodiya, T. P.; Boer, M.; Bogaert, G.; Bogan, C.; Bohe, A.; Bojtos, P.; Bond, C.; Bondu, F.; Bonnand, R.; Boom, B. A.; Bork, R.; Boschi, V.; Bose, S.; Bouffanais, Y.; Bozzi, A.; Bradaschia, C.; Brady, P. R.; Braginsky, V. B.; Branchesi, M.; Brau, J. E.; Briant, T.; Brillet, A.; Brinkmann, M.; Brisson, V.; Brockill, P.; Brooks, A. F.; Brown, D. A.; Brown, D. D.; Brown, N. M.; Buchanan, C. C.; Buikema, A.; Bulik, T.; Bulten, H. J.; Buonanno, A.; Buskulic, D.; Buy, C.; Byer, R. L.; Cadonati, L.; Cagnoli, G.; Cahillane, C.; Calderón Bustillo, J.; Callister, T.; Calloni, E.; Camp, J. B.; Cannon, K. C.; Cao, J.; Capano, C. D.; Capocasa, E.; Carbognani, F.; Caride, S.; Casanueva Diaz, J.; Casentini, C.; Caudill, S.; Cavaglià, M.; Cavalier, F.; Cavalieri, R.; Cella, G.; Cepeda, C. B.; Cerboni Baiardi, L.; Cerretani, G.; Cesarini, E.; Chakraborty, R.; Chatterji, S.; Chalermsongsak, T.; Chamberlin, S. J.; Chan, M.; Chao, S.; Charlton, P.; Chassande-Mottin, E.; Chen, H. Y.; Chen, Y.; Cheng, C.; Chincarini, A.; Chiummo, A.; Cho, H. S.; Cho, M.; Chow, J. H.; Christensen, N.; Chu, Q.; Chua, S.; Chung, S.; Ciani, G.; Clara, F.; Clark, J. A.; Clark, M.; Cleva, F.; Coccia, E.; Cohadon, P.-F.; Colla, A.; Collette, C. G.; Cominsky, L.; Constancio, M.; Conte, A.; Conti, L.; Cook, D.; Corbitt, T. R.; Cornish, N.; Corsi, A.; Cortese, S.; Costa, C. A.; Coughlin, M. W.; Coughlin, S. B.; Coulon, J.-P.; Countryman, S. T.; Couvares, P.; Cowan, E. E.; Coward, D. M.; Cowart, M. J.; Coyne, D. C.; Coyne, R.; Craig, K.; Creighton, J. D. E.; Cripe, J.; Crowder, S. G.; Cumming, A.; Cunningham, L.; Cuoco, E.; Dal Canton, T.; Danilishin, S. L.; D'Antonio, S.; Danzmann, K.; Darman, N. S.; Dattilo, V.; Dave, I.; Daveloza, H. P.; Davier, M.; Davies, G. S.; Daw, E. J.; Day, R.; DeBra, D.; Debreczeni, G.; Degallaix, J.; De Laurentis, M.; Deléglise, S.; Del Pozzo, W.; Denker, T.; Dent, T.; Dereli, H.; Dergachev, V.; DeRosa, R. T.; De Rosa, R.; DeSalvo, R.; Dhurandhar, S.; Díaz, M. C.; Di Fiore, L.; Di Giovanni, M.; Di Lieto, A.; Di Pace, S.; Di Palma, I.; Di Virgilio, A.; Dojcinoski, G.; Dolique, V.; Donovan, F.; Dooley, K. L.; Doravari, S.; Douglas, R.; Downes, T. P.; Drago, M.; Drever, R. W. P.; Driggers, J. C.; Du, Z.; Ducrot, M.; Dwyer, S. E.; Edo, T. B.; Edwards, M. C.; Effler, A.; Eggenstein, H.-B.; Ehrens, P.; Eichholz, J.; Eikenberry, S. S.; Engels, W.; Essick, R. C.; Etzel, T.; Evans, M.; Evans, T. M.; Everett, R.; Factourovich, M.; Fafone, V.; Fair, H.; Fairhurst, S.; Fan, X.; Fang, Q.; Farinon, S.; Farr, B.; Farr, W. M.; Favata, M.; Fays, M.; Fehrmann, H.; Fejer, M. M.; Ferrante, I.; Ferreira, E. C.; Ferrini, F.; Fidecaro, F.; Fiori, I.; Fiorucci, D.; Fisher, R. P.; Flaminio, R.; Fletcher, M.; Fournier, J.-D.; Franco, S.; Frasca, S.; Frasconi, F.; Frei, Z.; Freise, A.; Frey, R.; Frey, V.; Fricke, T. T.; Fritschel, P.; Frolov, V. V.; Fulda, P.; Fyffe, M.; Gabbard, H. A. G.; Gair, J. R.; Gammaitoni, L.; Gaonkar, S. G.; Garufi, F.; Gatto, A.; Gaur, G.; Gehrels, N.; Gemme, G.; Gendre, B.; Genin, E.; Gennai, A.; George, J.; Gergely, L.; Germain, V.; Ghosh, Archisman; Ghosh, S.; Giaime, J. A.; Giardina, K. D.; Giazotto, A.; Gill, K.; Glaefke, A.; Goetz, E.; Goetz, R.; Gondan, L.; González, G.; Gonzalez Castro, J. M.; Gopakumar, A.; Gordon, N. A.; Gorodetsky, M. L.; Gossan, S. E.; Gosselin, M.; Gouaty, R.; Graef, C.; Graff, P. B.; Granata, M.; Grant, A.; Gras, S.; Gray, C.; Greco, G.; Green, A. C.; Groot, P.; Grote, H.; Grunewald, S.; Guidi, G. M.; Guo, X.; Gupta, A.; Gupta, M. K.; Gushwa, K. E.; Gustafson, E. K.; Gustafson, R.; Haas, R.; Hacker, J. J.; Hall, B. R.; Hall, E. D.; Hammond, G.; Haney, M.; Hanke, M. M.; Hanks, J.; Hanna, C.; Hannam, M. D.; Hanson, J.; Hardwick, T.; Harms, J.; Harry, G. M.; Harry, I. W.; Hart, M. J.; Hartman, M. T.; Haster, C.-J.; Haughian, K.; Healy, J.; Heidmann, A.; Heintze, M. C.; Heitmann, H.; Hello, P.; Hemming, G.; Hendry, M.; Heng, I. S.; Hennig, J.; Heptonstall, A. W.; Heurs, M.; Hild, S.; Hinder, I.; Hoak, D.; Hodge, K. A.; Hofman, D.; Hollitt, S. E.; Holt, K.; Holz, D. E.; Hopkins, P.; Hosken, D. J.; Hough, J.; Houston, E. A.; Howell, E. J.; Hu, Y. M.; Huang, S.; Huerta, E. A.; Huet, D.; Hughey, B.; Husa, S.; Huttner, S. H.; Huynh-Dinh, T.; Idrisy, A.; Indik, N.; Ingram, D. R.; Inta, R.; Isa, H. N.; Isac, J.-M.; Isi, M.; Islas, G.; Isogai, T.; Iyer, B. R.; Izumi, K.; Jacqmin, T.; Jang, H.; Jani, K.; Jaranowski, P.; Jawahar, S.; Jiménez-Forteza, F.; Johnson, W. W.; Jones, D. I.; Jones, R.; Jonker, R. J. G.; Ju, L.; Haris, K.; Kalaghatgi, C. V.; Kalogera, V.; Kandhasamy, S.; Kang, G.; Kanner, J. B.; Karki, S.; Kasprzack, M.; Katsavounidis, E.; Katzman, W.; Kaufer, S.; Kaur, T.; Kawabe, K.; Kawazoe, F.; Kéfélian, F.; Kehl, M. S.; Keitel, D.; Kelley, D. B.; Kells, W.; Kennedy, R.; Key, J. S.; Khalaidovski, A.; Khalili, F. Y.; Khan, I.; Khan, S.; Khan, Z.; Khazanov, E. A.; Kijbunchoo, N.; Kim, C.; Kim, J.; Kim, K.; Kim, Nam-Gyu; Kim, Namjun; Kim, Y.-M.; King, E. J.; King, P. J.; Kinsey, M.; Kinzel, D. L.; Kissel, J. S.; Kleybolte, L.; Klimenko, S.; Koehlenbeck, S. M.; Kokeyama, K.; Koley, S.; Kondrashov, V.; Kontos, A.; Korobko, M.; Korth, W. Z.; Kowalska, I.; Kozak, D. B.; Kringel, V.; Królak, A.; Krueger, C.; Kuehn, G.; Kumar, P.; Kuo, L.; Kutynia, A.; Lackey, B. D.; Laguna, P.; Landry, M.; Lange, J.; Lantz, B.; Lasky, P. D.; Lazzarini, A.; Lazzaro, C.; Leaci, P.; Leavey, S.; Lebigot, E. O.; Lee, C. H.; Lee, H. K.; Lee, H. M.; Lee, K.; Lenon, A.; Leonardi, M.; Leong, J. R.; Leroy, N.; Letendre, N.; Levin, Y.; Levine, B. M.; Li, T. G. F.; Libson, A.; Littenberg, T. B.; Lockerbie, N. A.; Logue, J.; Lombardi, A. L.; Lord, J. E.; Lorenzini, M.; Loriette, V.; Lormand, M.; Losurdo, G.; Lough, J. D.; Lück, H.; Lundgren, A. P.; Luo, J.; Lynch, R.; Ma, Y.; MacDonald, T.; Machenschalk, B.; MacInnis, M.; Macleod, D. M.; Magaña-Sandoval, F.; Magee, R. M.; Mageswaran, M.; Majorana, E.; Maksimovic, I.; Malvezzi, V.; Man, N.; Mandel, I.; Mandic, V.; Mangano, V.; Mansell, G. L.; Manske, M.; Mantovani, M.; Marchesoni, F.; Marion, F.; Márka, S.; Márka, Z.; Markosyan, A. S.; Maros, E.; Martelli, F.; Martellini, L.; Martin, I. W.; Martin, R. M.; Martynov, D. V.; Marx, J. N.; Mason, K.; Masserot, A.; Massinger, T. J.; Masso-Reid, M.; Matichard, F.; Matone, L.; Mavalvala, N.; Mazumder, N.; Mazzolo, G.; McCarthy, R.; McClelland, D. E.; McCormick, S.; McGuire, S. C.; McIntyre, G.; McIver, J.; McManus, D. J.; McWilliams, S. T.; Meacher, D.; Meadors, G. D.; Meidam, J.; Melatos, A.; Mendell, G.; Mendoza-Gandara, D.; Mercer, R. A.; Merilh, E.; Merzougui, M.; Meshkov, S.; Messenger, C.; Messick, C.; Meyers, P. M.; Mezzani, F.; Miao, H.; Michel, C.; Middleton, H.; Mikhailov, E. E.; Milano, L.; Miller, J.; Millhouse, M.; Minenkov, Y.; Ming, J.; Mirshekari, S.; Mishra, C.; Mitra, S.; Mitrofanov, V. P.; Mitselmakher, G.; Mittleman, R.; Moggi, A.; Mohan, M.; Mohapatra, S. R. P.; Montani, M.; Moore, B. C.; Moore, C. J.; Moraru, D.; Moreno, G.; Morriss, S. R.; Mossavi, K.; Mours, B.; Mow-Lowry, C. M.; Mueller, C. L.; Mueller, G.; Muir, A. W.; Mukherjee, Arunava; Mukherjee, D.; Mukherjee, S.; Mukund, N.; Mullavey, A.; Munch, J.; Murphy, D. J.; Murray, P. G.; Mytidis, A.; Nardecchia, I.; Naticchioni, L.; Nayak, R. K.; Necula, V.; Nedkova, K.; Nelemans, G.; Neri, M.; Neunzert, A.; Newton, G.; Nguyen, T. T.; Nielsen, A. B.; Nissanke, S.; Nitz, A.; Nocera, F.; Nolting, D.; Normandin, M. E.; Nuttall, L. K.; Oberling, J.; Ochsner, E.; O'Dell, J.; Oelker, E.; Ogin, G. H.; Oh, J. J.; Oh, S. H.; Ohme, F.; Oliver, M.; Oppermann, P.; Oram, Richard J.; O'Reilly, B.; O'Shaughnessy, R.; Ottaway, D. J.; Ottens, R. S.; Overmier, H.; Owen, B. J.; Pai, A.; Pai, S. A.; Palamos, J. R.; Palashov, O.; Palomba, C.; Pal-Singh, A.; Pan, H.; Pankow, C.; Pannarale, F.; Pant, B. C.; Paoletti, F.; Paoli, A.; Papa, M. A.; Page, J.; Paris, H. R.; Parker, W.; Pascucci, D.; Pasqualetti, A.; Passaquieti, R.; Passuello, D.; Patricelli, B.; Patrick, Z.; Pearlstone, B. L.; Pedraza, M.; Pedurand, R.; Pekowsky, L.; Pele, A.; Penn, S.; Perreca, A.; Phelps, M.; Piccinni, O.; Pichot, M.; Piergiovanni, F.; Pierro, V.; Pillant, G.; Pinard, L.; Pinto, I. M.; Pitkin, M.; Poggiani, R.; Popolizio, P.; Post, A.; Powell, J.; Prasad, J.; Predoi, V.; Premachandra, S. S.; Prestegard, T.; Price, L. R.; Prijatelj, M.; Principe, M.; Privitera, S.; Prodi, G. A.; Prokhorov, L.; Puncken, O.; Punturo, M.; Puppo, P.; Pürrer, M.; Qi, H.; Qin, J.; Quetschke, V.; Quintero, E. A.; Quitzow-James, R.; Raab, F. J.; Rabeling, D. S.; Radkins, H.; Raffai, P.; Raja, S.; Rakhmanov, M.; Rapagnani, P.; Raymond, V.; Razzano, M.; Re, V.; Read, J.; Reed, C. M.; Regimbau, T.; Rei, L.; Reid, S.; Reitze, D. H.; Rew, H.; Reyes, S. D.; Ricci, F.; Riles, K.; Robertson, N. A.; Robie, R.; Robinet, F.; Rocchi, A.; Rolland, L.; Rollins, J. G.; Roma, V. J.; Romano, R.; Romanov, G.; Romie, J. H.; Rosińska, D.; Rowan, S.; Rüdiger, A.; Ruggi, P.; Ryan, K.; Sachdev, S.; Sadecki, T.; Sadeghian, L.; Salconi, L.; Saleem, M.; Salemi, F.; Samajdar, A.; Sammut, L.; Sanchez, E. J.; Sandberg, V.; Sandeen, B.; Sanders, J. R.; Sassolas, B.; Sathyaprakash, B. S.; Saulson, P. R.; Sauter, O.; Savage, R. L.; Sawadsky, A.; Schale, P.; Schilling, R.; Schmidt, J.; Schmidt, P.; Schnabel, R.; Schofield, R. M. S.; Schönbeck, A.; Schreiber, E.; Schuette, D.; Schutz, B. F.; Scott, J.; Scott, S. M.; Sellers, D.; Sengupta, A. S.; Sentenac, D.; Sequino, V.; Sergeev, A.; Serna, G.; Setyawati, Y.; Sevigny, A.; Shaddock, D. A.; Shah, S.; Shahriar, M. S.; Shaltev, M.; Shao, Z.; Shapiro, B.; Shawhan, P.; Sheperd, A.; Shoemaker, D. H.; Shoemaker, D. M.; Siellez, K.; Siemens, X.; Sigg, D.; Silva, A. D.; Simakov, D.; Singer, A.; Singer, L. P.; Singh, A.; Singh, R.; Singhal, A.; Sintes, A. M.; Slagmolen, B. J. J.; Smith, J. R.; Smith, N. D.; Smith, R. J. E.; Son, E. J.; Sorazu, B.; Sorrentino, F.; Souradeep, T.; Srivastava, A. K.; Staley, A.; Steinke, M.; Steinlechner, J.; Steinlechner, S.; Steinmeyer, D.; Stephens, B. C.; Stone, R.; Strain, K. A.; Straniero, N.; Stratta, G.; Strauss, N. A.; Strigin, S.; Sturani, R.; Stuver, A. L.; Summerscales, T. Z.; Sun, L.; Sutton, P. J.; Swinkels, B. L.; Szczepańczyk, M. J.; Tacca, M.; Talukder, D.; Tanner, D. B.; Tápai, M.; Tarabrin, S. P.; Taracchini, A.; Taylor, R.; Theeg, T.; Thirugnanasambandam, M. P.; Thomas, E. G.; Thomas, M.; Thomas, P.; Thorne, K. A.; Thorne, K. S.; Thrane, E.; Tiwari, S.; Tiwari, V.; Tokmakov, K. V.; Tomlinson, C.; Tonelli, M.; Torres, C. V.; Torrie, C. I.; Töyrä, D.; Travasso, F.; Traylor, G.; Trifirò, D.; Tringali, M. C.; Trozzo, L.; Tse, M.; Turconi, M.; Tuyenbayev, D.; Ugolini, D.; Unnikrishnan, C. S.; Urban, A. L.; Usman, S. A.; Vahlbruch, H.; Vajente, G.; Valdes, G.; van Bakel, N.; van Beuzekom, M.; van den Brand, J. F. J.; Van Den Broeck, C.; Vander-Hyde, D. C.; van der Schaaf, L.; van Heijningen, J. V.; van Veggel, A. A.; Vardaro, M.; Vass, S.; Vasúth, M.; Vaulin, R.; Vecchio, A.; Vedovato, G.; Veitch, J.; Veitch, P. J.; Venkateswara, K.; Verkindt, D.; Vetrano, F.; Viceré, A.; Vinciguerra, S.; Vine, D. J.; Vinet, J.-Y.; Vitale, S.; Vo, T.; Vocca, H.; Vorvick, C.; Voss, D.; Vousden, W. D.; Vyatchanin, S. P.; Wade, A. R.; Wade, L. E.; Wade, M.; Walker, M.; Wallace, L.; Walsh, S.; Wang, G.; Wang, H.; Wang, M.; Wang, X.; Wang, Y.; Ward, R. L.; Warner, J.; Was, M.; Weaver, B.; Wei, L.-W.; Weinert, M.; Weinstein, A. J.; Weiss, R.; Welborn, T.; Wen, L.; Weßels, P.; Westphal, T.; Wette, K.; Whelan, J. T.; White, D. J.; Whiting, B. F.; Williams, D.; Williams, R. D.; Williamson, A. R.; Willis, J. L.; Willke, B.; Wimmer, M. H.; Winkler, W.; Wipf, C. C.; Wittel, H.; Woan, G.; Worden, J.; Wright, J. L.; Wu, G.; Yablon, J.; Yam, W.; Yamamoto, H.; Yancey, C. C.; Yap, M. J.; Yu, H.; Yvert, M.; ZadroŻny, A.; Zangrando, L.; Zanolin, M.; Zendri, J.-P.; Zevin, M.; Zhang, F.; Zhang, L.; Zhang, M.; Zhang, Y.; Zhao, C.; Zhou, M.; Zhou, Z.; Zhu, X. J.; Zucker, M. E.; Zuraw, S. E.; Zweizig, J.; LIGO Scientific Collaboration; Virgo Collaboration

    2016-06-01

    The gravitational-wave signal GW150914 was first identified on September 14, 2015, by searches for short-duration gravitational-wave transients. These searches identify time-correlated transients in multiple detectors with minimal assumptions about the signal morphology, allowing them to be sensitive to gravitational waves emitted by a wide range of sources including binary black hole mergers. Over the observational period from September 12 to October 20, 2015, these transient searches were sensitive to binary black hole mergers similar to GW150914 to an average distance of ˜600 Mpc . In this paper, we describe the analyses that first detected GW150914 as well as the parameter estimation and waveform reconstruction techniques that initially identified GW150914 as the merger of two black holes. We find that the reconstructed waveform is consistent with the signal from a binary black hole merger with a chirp mass of ˜30 M⊙ and a total mass before merger of ˜70 M⊙ in the detector frame.

  1. Speakers' assumptions about the lexical flexibility of idioms.

    PubMed

    Gibbs, R W; Nayak, N P; Bolton, J L; Keppel, M E

    1989-01-01

    In three experiments, we examined why some idioms can be lexically altered and still retain their figurative meanings (e.g., John buttoned his lips about Mary can be changed into John fastened his lips about Mary and still mean "John didn't say anything about Mary"), whereas other idioms cannot be lexically altered without losing their figurative meanings (e.g., John kicked the bucket, meaning "John died," loses its idiomatic meaning when changed into John kicked the pail). Our hypothesis was that the lexical flexibility of idioms is determined by speakers' assumptions about the ways in which parts of idioms contribute to their figurative interpretations as a whole. The results of the three experiments indicated that idioms whose individual semantic components contribute to their overall figurative meanings (e.g., go out on a limb) were judged as less disrupted by changes in their lexical items (e.g., go out on a branch) than were nondecomposable idioms (e.g., kick the bucket) when their individual words were altered (e.g., punt the pail). These findings lend support to the idea that both the syntactic productivity and the lexical makeup of idioms are matters of degree, depending on the idioms' compositional properties. This conclusion suggests that idioms do not form a unique class of linguistic items, but share many of the properties of more literal language.

  2. Niches and distributional areas: concepts, methods, and assumptions.

    PubMed

    Soberón, Jorge; Nakamura, Miguel

    2009-11-17

    Estimating actual and potential areas of distribution of species via ecological niche modeling has become a very active field of research, yet important conceptual issues in this field remain confused. We argue that conceptual clarity is enhanced by adopting restricted definitions of "niche" that enable operational definitions of basic concepts like fundamental, potential, and realized niches and potential and actual distributional areas. We apply these definitions to the question of niche conservatism, addressing what it is that is conserved and showing with a quantitative example how niche change can be measured. In this example, we display the extremely irregular structure of niche space, arguing that it is an important factor in understanding niche evolution. Many cases of apparently successful models of distributions ignore biotic factors: we suggest explanations to account for this paradox. Finally, relating the probability of observing a species to ecological factors, we address the issue of what objects are actually calculated by different niche modeling algorithms and stress the fact that methods that use only presence data calculate very different quantities than methods that use absence data. We conclude that the results of niche modeling exercises can be interpreted much better if the ecological and mathematical assumptions of the modeling process are made explicit.

  3. Simplified cost models for prefeasibility mineral evaluations

    USGS Publications Warehouse

    Camm, Thomas W.

    1991-01-01

    This report contains 2 open pit models, 6 underground mine models, 11 mill models, and cost equations for access roads, power lines, and tailings ponds. In addition, adjustment factors for variation in haulage distances are provided for open pit models and variation in mining depths for underground models.

  4. Photographic and drafting techniques simplify method of producing engineering drawings

    NASA Technical Reports Server (NTRS)

    Provisor, H.

    1968-01-01

    Combination of photographic and drafting techniques has been developed to simplify the preparation of three dimensional and dimetric engineering drawings. Conventional photographs can be converted to line drawings by making copy negatives on high contrast film.

  5. Simplified circuit corrects faults in parallel binary information channels

    NASA Technical Reports Server (NTRS)

    Goldberg, J.

    1966-01-01

    Corrective circuit prevents the appearance of erroneous output signals from the possible failure of any single-channel element interconnected in parallel binary information channels. The circuit is simplified and economical because it does not use redundant channels.

  6. Simplified method for numerical modeling of fiber lasers.

    PubMed

    Shtyrina, O V; Yarutkina, I A; Fedoruk, M P

    2014-12-29

    A simplified numerical approach to modeling of dissipative dispersion-managed fiber lasers is examined. We present a new numerical iteration algorithm for finding the periodic solutions of the system of nonlinear ordinary differential equations describing the intra-cavity dynamics of the dissipative soliton characteristics in dispersion-managed fiber lasers. We demonstrate that results obtained using simplified model are in good agreement with full numerical modeling based on the corresponding partial differential equations.

  7. Simplified Correction Of Errors In Reed-Solomon Codes

    NASA Technical Reports Server (NTRS)

    Truong, T. K.; Hsu, I. S.; Eastman, W. L.; Reed, I. S.

    1990-01-01

    New decoder realized by simplified pipeline architecture. Simplified procedure for correction of errors and erasures in Reed-Solomon codes expected to result in simpler decoding equipment. Development widens commercial applicability of Reed-Solomon codes, used to correct bursts of errors in digital communication and recording systems. Improved decoder less complex. Made more regular, simple, and suitable for implementation in both VLSI and software.

  8. S-channel dark matter simplified models and unitarity

    NASA Astrophysics Data System (ADS)

    Englert, Christoph; McCullough, Matthew; Spannowsky, Michael

    2016-12-01

    The ultraviolet structure of s-channel mediator dark matter simplified models at hadron colliders is considered. In terms of commonly studied s-channel mediator simplified models it is argued that at arbitrarily high energies the perturbative description of dark matter production in high energy scattering can break down. This is analogous to the well documented breakdown of an EFT description of dark matter collider production. With this in mind, to diagnose whether or not the use of simplified models at the LHC is valid, perturbative unitarity of the scattering amplitude in the processes relevant to LHC dark matter searches is studied. The results are as one would expect: at the LHC and future proton colliders the simplified model descriptions of dark matter production are in general valid. As a result of the general discussion, a simple class of 'Fermiophobic Scalar' simplified models is proposed, in which a scalar mediator couples to electroweak vector bosons. The Fermiophobic simplified model is well motivated and exhibits interesting collider and direct detection phenomenology.

  9. Evaluation of Simplified Models for Estimating Public Dose from Spent Nuclear Fuel Shipments

    SciTech Connect

    Connolly, Kevin J.; Radulescu, Georgeta

    2015-01-01

    This paper investigates the dose rate as a function of distance from a representative high-capacity SNF rail-type transportation cask. It uses the SCALE suite of radiation transport modeling and simulation codes to determine neutron and gamma radiation dose rates. The SCALE calculated dose rate is compared with simplified analytical methods historically used for these calculations. The SCALE dose rate calculation presented in this paper employs a very detailed transportation cask model (e.g., pin-by-pin modeling of fuel assembly) and a new hybrid computational transport method. Because it includes pin-level heterogeneity and models ample air and soil outside the cask to simulate scattering of gamma and neutron radiation, this detailed SCALE model is expected to yield more accurate results than previously used models which made more simplistic assumptions (e.g., fuel assembly treated as a point or line source, simple 1-D model of environment outside of cask). The results in this paper are preliminary and, as progress is made on developing and validating improved models, results may be subject to change as models and estimates become more refined and better information leads to more accurate assumptions.

  10. Implications of genome wide association studies for addiction: Are our a priori assumptions all wrong?

    PubMed Central

    Hall, F. Scott; Drgonova, Jana; Jain, Siddharth; Uhl, George R.

    2013-01-01

    Substantial genetic contributions to addiction vulnerability are supported by data from twin studies, linkage studies, candidate gene association studies and, more recently, Genome Wide Association Studies (GWAS). Parallel to this work, animal studies have attempted to identify the genes that may contribute to responses to addictive drugs and addiction liability, initially focusing upon genes for the targets of the major drugs of abuse. These studies identified genes/proteins that affect responses to drugs of abuse; however, this does not necessarily mean that variation in these genes contributes to the genetic component of addiction liability. One of the major problems with initial linkage and candidate gene studies was an a priori focus on the genes thought to be involved in addiction based upon the known contributions of those proteins to drug actions, making the identification of novel genes unlikely. The GWAS approach is systematic and agnostic to such a priori assumptions. From the numerous GWAS now completed several conclusions may be drawn: (1) addiction is highly polygenic; each allelic variant contributing in a small, additive fashion to addiction vulnerability; (2) unexpected, compared to our a priori assumptions, classes of genes are most important in explaining addiction vulnerability; (3) although substantial genetic heterogeneity exists, there is substantial convergence of GWAS signals on particular genes. This review traces the history of this research; from initial transgenic mouse models based upon candidate gene and linkage studies, through the progression of GWAS for addiction and nicotine cessation, to the current human and transgenic mouse studies post-GWAS. PMID:23872493

  11. Tolerance values of benthic macroinvertebrates for stream biomonitoring: assessment of assumptions underlying scoring systems worldwide.

    PubMed

    Chang, Feng-Hsun; Lawrence, Justin E; Rios-Touma, Blanca; Resh, Vincent H

    2014-04-01

    Tolerance values (TVs) based on benthic macroinvertebrates are one of the most widely used tools for monitoring the biological impacts of water pollution, particularly in streams and rivers. We compiled TVs of benthic macroinvertebrates from 29 regions around the world to test 11 basic assumptions about pollution tolerance, that: (1) Arthropoda are < tolerant than non-Arthropoda; (2) Insecta < non-Insecta; (3) non-Oligochaeta < Oligochaeta; (4) other macroinvertebrates < Oligochaeta + Chironomidae; (5) other macroinvertebrate taxa < Isopoda + Gastropoda + Hirudinea; (6) Ephemeroptera + Plecoptera + Trichoptera (EPT) < Odonata + Coleoptera + Heteroptera (OCH); (7) EPT < non-EPT insects; (8) Diptera < Insecta; (9) Bivalvia < Gastropoda; (10) Baetidae < other Ephemeroptera; and (11) Hydropsychidae < other Trichoptera. We found that the first eight of these 11 assumptions were supported despite regional variability. In addition, we examined the effect of Best Professional Judgment (BPJ) and non-independence of TVs among countries by performing all analyses using subsets of the original dataset. These subsets included a group based on those systems using TVs that were derived from techniques other than BPJ, and groups based on methods used for TV assignment. The results obtained from these subsets and the entire dataset are similar. We also made seven a priori hypotheses about the regional similarity of TVs based on geography. Only one of these was supported. Development of TVs and the reporting of how they are assigned need to be more rigorous and be better described.

  12. Implications of genome wide association studies for addiction: are our a priori assumptions all wrong?

    PubMed

    Hall, F Scott; Drgonova, Jana; Jain, Siddharth; Uhl, George R

    2013-12-01

    Substantial genetic contributions to addiction vulnerability are supported by data from twin studies, linkage studies, candidate gene association studies and, more recently, Genome Wide Association Studies (GWAS). Parallel to this work, animal studies have attempted to identify the genes that may contribute to responses to addictive drugs and addiction liability, initially focusing upon genes for the targets of the major drugs of abuse. These studies identified genes/proteins that affect responses to drugs of abuse; however, this does not necessarily mean that variation in these genes contributes to the genetic component of addiction liability. One of the major problems with initial linkage and candidate gene studies was an a priori focus on the genes thought to be involved in addiction based upon the known contributions of those proteins to drug actions, making the identification of novel genes unlikely. The GWAS approach is systematic and agnostic to such a priori assumptions. From the numerous GWAS now completed several conclusions may be drawn: (1) addiction is highly polygenic; each allelic variant contributing in a small, additive fashion to addiction vulnerability; (2) unexpected, compared to our a priori assumptions, classes of genes are most important in explaining addiction vulnerability; (3) although substantial genetic heterogeneity exists, there is substantial convergence of GWAS signals on particular genes. This review traces the history of this research; from initial transgenic mouse models based upon candidate gene and linkage studies, through the progression of GWAS for addiction and nicotine cessation, to the current human and transgenic mouse studies post-GWAS.

  13. Simplified stereo-optical ultrasound plane calibration

    NASA Astrophysics Data System (ADS)

    Hoßbach, Martin; Noll, Matthias; Wesarg, Stefan

    2013-03-01

    Image guided therapy is a natural concept and commonly used in medicine. In anesthesia, a common task is the injection of an anesthetic close to a nerve under freehand ultrasound guidance. Several guidance systems exist using electromagnetic tracking of the ultrasound probe as well as the needle, providing the physician with a precise projection of the needle into the ultrasound image. This, however, requires additional expensive devices. We suggest using optical tracking with miniature cameras attached to a 2D ultrasound probe to achieve a higher acceptance among physicians. The purpose of this paper is to present an intuitive method to calibrate freehand ultrasound needle guidance systems employing a rigid stereo camera system. State of the art methods are based on a complex series of error prone coordinate system transformations which makes them susceptible to error accumulation. By reducing the amount of calibration steps to a single calibration procedure we provide a calibration method that is equivalent, yet not prone to error accumulation. It requires a linear calibration object and is validated on three datasets utilizing di erent calibration objects: a 6mm metal bar and a 1:25mm biopsy needle were used for experiments. Compared to existing calibration methods for freehand ultrasound needle guidance systems, we are able to achieve higher accuracy results while additionally reducing the overall calibration complexity. Ke

  14. 77 FR 24516 - Solicitation of Comments on Request for United States Assumption of Concurrent Federal Criminal...

    Federal Register 2010, 2011, 2012, 2013, 2014

    2012-04-24

    ... Solicitation of Comments on Request for United States Assumption of Concurrent Federal Criminal Jurisdiction... Assumption of Concurrent Federal Criminal Jurisdiction recently submitted to the Office of Tribal Justice... electronic copy of the request for United States assumption of concurrent federal criminal...

  15. 7 CFR 1782.13 - Transfer of security and assumption of loans.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... 7 Agriculture 12 2011-01-01 2011-01-01 false Transfer of security and assumption of loans. 1782.13... SERVICE, DEPARTMENT OF AGRICULTURE (CONTINUED) SERVICING OF WATER AND WASTE PROGRAMS § 1782.13 Transfer of security and assumption of loans. It is the Agency's policy to approve transfers and assumptions...

  16. Coefficient Alpha as an Estimate of Test Reliability under Violation of Two Assumptions.

    ERIC Educational Resources Information Center

    Zimmerman, Donald W.; And Others

    1993-01-01

    Coefficient alpha was examined through computer simulation as an estimate of test reliability under violation of two assumptions. Coefficient alpha underestimated reliability under violation of the assumption of essential tau-equivalence of subtest scores and overestimated it under violation of the assumption of uncorrelated subtest error scores.…

  17. Exploring the Influence of Ethnicity, Age, and Trauma on Prisoners' World Assumptions

    ERIC Educational Resources Information Center

    Gibson, Sandy

    2011-01-01

    In this study, the author explores world assumptions of prisoners, how these assumptions vary by ethnicity and age, and whether trauma history affects world assumptions. A random sample of young and old prisoners, matched for prison location, was drawn from the New Jersey Department of Corrections prison population. Age and ethnicity had…

  18. Philosophy of Technology Assumptions in Educational Technology Leadership: Questioning Technological Determinism

    ERIC Educational Resources Information Center

    Webster, Mark David

    2013-01-01

    Scholars have emphasized that decisions about technology can be influenced by philosophy of technology assumptions, and have argued for research that critically questions technological determinist assumptions. Empirical studies of technology management in fields other than K-12 education provided evidence that philosophy of technology assumptions,…

  19. Educational Technology as a Subversive Activity: Questioning Assumptions Related to Teaching and Leading with Technology

    ERIC Educational Resources Information Center

    Kruger-Ross, Matthew J.; Holcomb, Lori B.

    2012-01-01

    The use of educational technologies is grounded in the assumptions of teachers, learners, and administrators. Assumptions are choices that structure our understandings and help us make meaning. Current advances in Web 2.0 and social media technologies challenge our assumptions about teaching and learning. The intersection of technology and…

  20. Simplified three-dimensional tissue clearing and incorporation of colorimetric phenotyping

    PubMed Central

    Sung, Kevin; Ding, Yichen; Ma, Jianguo; Chen, Harrison; Huang, Vincent; Cheng, Michelle; Yang, Cindy F.; Kim, Jocelyn T.; Eguchi, Daniel; Di Carlo, Dino; Hsiai, Tzung K.; Nakano, Atsushi; Kulkarni, Rajan P.

    2016-01-01

    Tissue clearing methods promise to provide exquisite three-dimensional imaging information; however, there is a need for simplified methods for lower resource settings and for non-fluorescence based phenotyping to enable light microscopic imaging modalities. Here we describe the simplified CLARITY method (SCM) for tissue clearing that preserves epitopes of interest. We imaged the resulting tissues using light sheet microscopy to generate rapid 3D reconstructions of entire tissues and organs. In addition, to enable clearing and 3D tissue imaging with light microscopy methods, we developed a colorimetric, non-fluorescent method for specifically labeling cleared tissues based on horseradish peroxidase conversion of diaminobenzidine to a colored insoluble product. The methods we describe here are portable and can be accomplished at low cost, and can allow light microscopic imaging of cleared tissues, thus enabling tissue clearing and imaging in a wide variety of settings. PMID:27498769

  1. Simplified generalized-gradient approximation and anharmonicity: Benchmark calculations on molecules

    NASA Astrophysics Data System (ADS)

    Patton, David C.; Porezag, Dirk V.; Pederson, Mark R.

    1997-03-01

    Recent implementational improvements of the generalized-gradient approximation (GGA) have led to a simplified version which is parametrized entirely from fundamental constants, easier to use, and possibly easier to improve. We have performed detailed calculations on the geometries, atomization energies, vibrational energies, and infrared and Raman spectra of many first- and second-row dimers as well as some polyatomic molecules. For atomization and vibrational energies, we find that the simplified version of GGA leads to results similar to the original version. We comment on the fact that GGA-induced changes of hydrogenic bonding are different than for the other atoms in the periodic table but still an improvement over the local approximations to density-functional theory. In addition to a harmonic treatment of the vibrational modes we include the contributions of anharmonicity as well. With the exception of the light hydrogen containing molecules anharmonic corrections are quite small.

  2. SALE-3D: a simplified ALE computer program for calculating three-dimensional fluid flow

    SciTech Connect

    Amsden, A.A.; Ruppel, H.M.

    1981-11-01

    This report presents a simplified numerical fluid-dynamics computing technique for calculating time-dependent flows in three dimensions. An implicit treatment of the pressure equation permits calculation of flows far subsonic without stringent constraints on the time step. In addition, the grid vertices may be moved with the fluid in Lagrangian fashion or held fixed in an Eulerian manner, or moved in some prescribed manner to give a continuous rezoning capability. This report describes the combination of Implicit Continuous-fluid Eulerian (ICE) and Arbitrary Lagrangian-Eulerian (ALE) to form the ICEd-ALE technique in the framework of the Simplified-ALE (SALE-3D) computer program, for which a general flow diagram and complete FORTRAN listing are included. Sample problems show how to modify the code for a variety of applications. SALE-3D is patterned as closely as possible on the previously reported two-dimensional SALE program.

  3. Transient Stress- and Strain-Based Hemolysis Estimation in a Simplified Blood Pump

    PubMed Central

    Pauli, L.; Nam, J.; Pasquali, M.; Behr, M.

    2014-01-01

    SUMMARY We compare two approaches to numerical estimation of mechanical hemolysis in a simplified blood pump model. The stress-based model relies on the instantaneous shear stress in the blood flow, whereas the strain-based model uses an additional tensor equation to relate distortion of red blood cells to a shear stress measure. We use the newly proposed least-squares finite element method (LSFEM) to prevent negative concentration fields and show a stable and volume preserving LSFEM for the tensor equation. Application of both models to a simplified centrifugal blood pump at three different operating conditions show that the stress-based model overestimates the rate of hemolysis. The strain-based model is found to deliver lower hemolysis rates since it incorporates a more detailed description of biophysical phenomena into the simulation process. PMID:23922311

  4. A simplified approach for solving coagulation-diffusion equation to estimate atmospheric background particle number loading factors contributed by emissions from localized sources

    NASA Astrophysics Data System (ADS)

    Anand, S.; Mayya, Y. S.

    2011-08-01

    Coagulation and condensation/evaporation combined with atmospheric dispersion are the main processes responsible for the evolution of aerosol particle size distributions and number concentrations emitted from localized sources. A crucial question is: what fraction of freshly emitted particles survive intra-coagulation effect to persist in the atmosphere and become available for further interaction with background aerosols?. The difficulty in estimating this quantity, designated as the number survival fraction, arises due chiefly to the joint action of atmospheric diffusion with nonlinear coagulation effects which are computationally intensive to handle. We provide a simplified approach to evaluate this quantity in the context of instantaneous (puff) and continuous (plume) releases based on a reduction of the respective coagulation-diffusion equations under the assumption of a constant coagulation kernel ( K). The condensation/evaporation processes, being number conserving, are not included in the study. The approach consists of constructing moment equations for the evolution of number concentration and variance of the spatial extension of puff or plume in terms of either time or downstream distance. The puff model, applicable to instantaneous releases is solved within a 3-D, spherically symmetric framework, under an additional assumption of a constant diffusion coefficient ( D) which renders itself amenable to a closed form solution that provides a benchmark for developing the solution to the plume model. The latter case, corresponding to continuous releases, is discussed within a 2-D framework under the assumptions of constant advection velocity ( U) and space dependent diffusion coefficient expressed in terms of turbulent energy dissipation rate ( ɛ). The study brings out the special effect of the coagulation-induced flattening of the spatial concentration profiles because of which particle sizes will be larger at the centre of a Gaussian puff. For a puff of

  5. Is the assumption of normality or log-normality for continuous response data critical for benchmark dose estimation?

    PubMed

    Shao, Kan; Gift, Jeffrey S; Setzer, R Woodrow

    2013-11-01

    Continuous responses (e.g. body weight) are widely used in risk assessment for determining the benchmark dose (BMD) which is used to derive a U.S. EPA reference dose. One critical question that is not often addressed in dose-response assessments is whether to model the continuous data as normally or log-normally distributed. Additionally, if lognormality is assumed, and only summarized response data (i.e., mean±standard deviation) are available as is usual in the peer-reviewed literature, the BMD can only be approximated. In this study, using the "hybrid" method and relative deviation approach, we first evaluate six representative continuous dose-response datasets reporting individual animal responses to investigate the impact on BMD/BMDL estimates of (1) the distribution assumption and (2) the use of summarized versus individual animal data when a log-normal distribution is assumed. We also conduct simulation studies evaluating model fits to various known distributions to investigate whether the distribution assumption has influence on BMD/BMDL estimates. Our results indicate that BMDs estimated using the hybrid method are more sensitive to the distribution assumption than counterpart BMDs estimated using the relative deviation approach. The choice of distribution assumption has limited impact on the BMD/BMDL estimates when the within dose-group variance is small, while the lognormality assumption is a better choice for relative deviation method when data are more skewed because of its appropriateness in describing the relationship between mean and standard deviation. Additionally, the results suggest that the use of summarized data versus individual response data to characterize log-normal distributions has minimal impact on BMD estimates.

  6. Simplified Approach to Predicting Rough Surface Transition

    NASA Technical Reports Server (NTRS)

    Boyle, Robert J.; Stripf, Matthias

    2009-01-01

    Turbine vane heat transfer predictions are given for smooth and rough vanes where the experimental data show transition moving forward on the vane as the surface roughness physical height increases. Consiste nt with smooth vane heat transfer, the transition moves forward for a fixed roughness height as the Reynolds number increases. Comparison s are presented with published experimental data. Some of the data ar e for a regular roughness geometry with a range of roughness heights, Reynolds numbers, and inlet turbulence intensities. The approach ta ken in this analysis is to treat the roughness in a statistical sense , consistent with what would be obtained from blades measured after e xposure to actual engine environments. An approach is given to determ ine the equivalent sand grain roughness from the statistics of the re gular geometry. This approach is guided by the experimental data. A roughness transition criterion is developed, and comparisons are made with experimental data over the entire range of experimental test co nditions. Additional comparisons are made with experimental heat tran sfer data, where the roughness geometries are both regular as well a s statistical. Using the developed analysis, heat transfer calculatio ns are presented for the second stage vane of a high pressure turbine at hypothetical engine conditions.

  7. Simplified Approach to Predicting Rough Surface Transition

    NASA Technical Reports Server (NTRS)

    Boyle, R. J.; Stripf, M.

    2009-01-01

    Turbine vane heat transfer predictions are given for smooth and rough vanes where the experimental data show transition moving forward on the vane as the surface roughness physical height increases. Consistent with smooth vane heat transfer, the transition moves forward for a fixed roughness height as the Reynolds number increases. Comparisons are presented with published experimental data. Some of the data are for a regular roughness geometry with a range of roughness heights, Reynolds numbers, and inlet turbulence intensities. The approach taken in this analysis is to treat the roughness in a statistical sense, consistent with what would be obtained from blades measured after exposure to actual engine environments. An approach is given to determine the equivalent sand grain roughness from the statistics of the regular geometry. This approach is guided by the experimental data. A roughness transition criterion is developed, and comparisons are made with experimental data over the entire range of experimental test conditions. Additional comparisons are made with experimental heat transfer data, where the roughness geometries are both regular and statistical. Using the developed analysis, heat transfer calculations are presented for the second stage vane of a high pressure turbine at hypothetical engine conditions.

  8. A Simplified Model of Tropical Cyclone Intensification

    NASA Astrophysics Data System (ADS)

    Schubert, W. H.

    2015-12-01

    An axisymmetric model of tropical cyclone intensification is presented. The model is based on Salmon's wave-vortex approximation, which can describe flows with high Rossby number and low Froude number. After introducing an additional approximation designed to filter propagating inertia-gravity waves, the problem is reduced to the prediction of potential vorticity (PV) and the inversion of this PV to obtain the balanced wind and mass fields. This PV prediction/inversion problem is solved analytically for two types of forcing: a two-region model in which there is nonzero forcing in the cyclone core and zero forcing in the far-field; a three-region model in which there is non-zero forcing in both the cyclone core and the eyewall, with zero forcing in the far-field. The solutions of the two-region model provide insight into why tropical cyclones can have long incubation times before rapid intensification and how the size of the mature vortex can be influenced by the size of the initial vortex. The solutions of the three-region model provide insight into the formation of hollow PV structures and the inward movement of angular momentum surfaces across the radius of maximum wind.

  9. Administrative Simplification, Simplified for Hawai‘i

    PubMed Central

    2012-01-01

    Background The American health care system contains a layer of administrative controls that has become increasingly burdensome to medical practices in terms of uncompensated physician and staff time and practice costs. A primary care physician in solo practice spends between 4 and 10 hours a week directly interacting with health insurance companies and his or her staff will spend an additional 60 hours a week. This reduces patient-care availability, net practice income and physician job satisfaction. Methods A literature review was conducted to determine possible solutions to administrative burdens physicians face in Hawai‘i. A total of 51 articles were found matching search criteria with five being reports from major organizations. Results Twenty-seven articles were found that related to administrative simplification. The “administrative complexity” problem has been defined and its financial impact quantified. Promising solutions have been developed and proposed by private not-for-profit organizations and by the government, both state and federal. Discussion A successful administrative simplification plan would: (1) Provide rapid access to insurance information; (2) Allow medical practices to readily track specific claims; (3) Streamline the preauthorization process through the use of decision-support tools at the practice level and by directing interactions through real-time network connections between insurers and provider electronic health records, thus minimizing phone time; (4) Adopt the Universal Provider Datasource system for provider credentialing; and (5) Standardize (to the greatest degree possible) provider/insurer contracts. These solutions are outlined in detail. PMID:22737644

  10. A Simplified Shuttle Irradiation Facility for ATR

    SciTech Connect

    A. J. Palmer; S. T. Laflin

    1999-08-01

    During the past fifteen years there has been a steady increase in the demand for radioisotopes in nuclear medicine and a corresponding decline in the number of reactors within the U.S. capable of producing them. The Advanced Test Reactor (ATR) is the largest operating test reactor in the U.S., but its isotope production capabilities have been limited by the lack of an installed isotope shuttle irradiation system. A concept for a simple "low cost" shuttle irradiation facility for ATR has been developed. Cost were reduced (in comparison to previous ATR designs) by using a shielded trough of water installed in an occupiable cubicle as a shielding and contamination control barrier for the send and receive station. This shielding concept also allows all control valves to be operated by hand and thus the need for an automatic control system was eliminated. It was determined that 4-5 ft of water would be adequate to shield the isotopes of interest while shuttles are transferred to a small carrier. An additional feature of the current design is a non-isolatable by-pass line, which provides a minimum coolant flow to the test region regardless of which control valves are opened or closed. This by-pass line allows the shuttle facility to be operated without bringing reactor coolant water into the cubicle except for send and receive operations.

  11. Regional and longitudinal estimation of product lifespan distribution: a case study for automobiles and a simplified estimation method.

    PubMed

    Oguchi, Masahiro; Fuse, Masaaki

    2015-02-03

    Product lifespan estimates are important information for understanding progress toward sustainable consumption and estimating the stocks and end-of-life flows of products. Publications reported actual lifespan of products; however, quantitative data are still limited for many countries and years. This study presents regional and longitudinal estimation of lifespan distribution of consumer durables, taking passenger cars as an example, and proposes a simplified method for estimating product lifespan distribution. We estimated lifespan distribution parameters for 17 countries based on the age profile of in-use cars. Sensitivity analysis demonstrated that the shape parameter of the lifespan distribution can be replaced by a constant value for all the countries and years. This enabled a simplified estimation that does not require detailed data on the age profile. Applying the simplified method, we estimated the trend in average lifespans of passenger cars from 2000 to 2009 for 20 countries. Average lifespan differed greatly between countries (9-23 years) and was increasing in many countries. This suggests consumer behavior differs greatly among countries and has changed over time, even in developed countries. The results suggest that inappropriate assumptions of average lifespan may cause significant inaccuracy in estimating the stocks and end-of-life flows of products.

  12. Simplified Habit Reversal Plus Adjunct Contingencies in the Treatment of Thumb Sucking and Hair Pulling in a Young Child.

    ERIC Educational Resources Information Center

    Long, Ethan S.; Miltenberger, Raymond G.; Rapp, John T.

    1999-01-01

    Using simplified reversal treatment consisting of awareness training, competing response training, and social support procedures, minimal results were initially obtained in thumb sucking and hair pulling behaviors. Additional treatment phases involving differential reinforcement and response cost resulted in near zero levels of the behavior when…

  13. A Simplified Shuttle Irradiation Facility for ATR

    SciTech Connect

    Palmer, Alma Joseph; Laflin, S. T.

    1999-09-01

    During the past fifteen years there has been a steady increase in the demand for radioisotopes in nuclear medicine and a corresponding decline in the number of reactors within the U.S. capable of producing them. The Advanced Test Reactor (ATR) is the largest operating test reactor in the U.S., but its isotope production capabilities have been limited by the lack of an installed isotope shuttle irradiation system. A concept for a simple “low cost” shuttle irradiation facility for ATR has been developed. Costs were reduced (in comparison to previous ATR designs) by using a shielded trough of water installed in an occupiable cubicle as a shielding and contamination control barrier for the send and receive station. This shielding concept also allows all control valves to be operated by hand and thus the need for an automatic control system was eliminated. It was determined that 4 – 5 ft of water would be adequate to shield the isotopes of interest while shuttles are transferred to a small carrier. An additional feature of the current design is a non-isolatable by-pass line, which provides a minimum coolant flow to the test region regardless of which control valves are opened or closed. This by-pass line allows the shuttle facility to be operated without bringing reactor coolant water into the cubicle except for send and receive operations. The irradiation position selected for this concept is a 1.5 inch “B” hole (B-11). This position provides neutron fluxes of approximately: 1.6 x 1014 (<0.5 eV) and 4.0 x 1013 (>0.8 MeV) n/cm2*sec.

  14. A simplified method for elastic-plastic-creep structural analysis

    NASA Technical Reports Server (NTRS)

    Kaufman, A.

    1984-01-01

    A simplified inelastic analysis computer program (ANSYPM) was developed for predicting the stress-strain history at the critical location of a thermomechanically cycled structure from an elastic solution. The program uses an iterative and incremental procedure to estimate the plastic strains from the material stress-strain properties and a plasticity hardening model. Creep effects are calculated on the basis of stress relaxation at constant strain, creep at constant stress or a combination of stress relaxation and creep accumulation. The simplified method was exercised on a number of problems involving uniaxial and multiaxial loading, isothermal and nonisothermal conditions, dwell times at various points in the cycles, different materials and kinematic hardening. Good agreement was found between these analytical results and nonlinear finite element solutions for these problems. The simplified analysis program used less than 1 percent of the CPU time required for a nonlinear finite element analysis.

  15. A simplified method for elastic-plastic-creep structural analysis

    NASA Technical Reports Server (NTRS)

    Kaufman, A.

    1985-01-01

    A simplified inelastic analysis computer program (ANSYPM) was developed for predicting the stress-strain history at the critical location of a thermomechanically cycled structure from an elastic solution. The program uses an iterative and incremental procedure to estimate the plastic strains from the material stress-strain properties and a plasticity hardening model. Creep effects are calculated on the basis of stress relaxation at constant strain, creep at constant stress or a combination of stress relaxation and creep accumulation. The simplified method was exercised on a number of problems involving uniaxial and multiaxial loading, isothermal and nonisothermal conditions, dwell times at various points in the cycles, different materials and kinematic hardening. Good agreement was found between these analytical results and nonlinear finite element solutions for these problems. The simplified analysis program used less than 1 percent of the CPU time required for a nonlinear finite element analysis.

  16. Development of a simplified procedure for cyclic structural analysis

    NASA Technical Reports Server (NTRS)

    Kaufman, A.

    1984-01-01

    Development was extended of a simplified inelastic analysis computer program (ANSYMP) for predicting the stress-strain history at the critical location of a thermomechanically cycled structure from an elastic solution. The program uses an iterative and incremental procedure to estimate the plastic strains from the material stress-strain properties and a plasticity hardening model. Creep effects can be calculated on the basis of stress relaxation at constant strain, creep at constant stress, or a combination of stress relaxation and creep accumulation. The simplified method was exercised on a number of problems involving uniaxial and multiaxial loading, isothermal and nonisothermal conditions, dwell times at various points in the cycles, different materials, and kinematic hardening. Good agreement was found between these analytical results and nonlinear finite-element solutions for these problems. The simplified analysis program used less than 1 percent of the CPU time required for a nonlinear finite-element analysis.

  17. Optical chirp z-transform processor with a simplified architecture.

    PubMed

    Ngo, Nam Quoc

    2014-12-29

    Using a simplified chirp z-transform (CZT) algorithm based on the discrete-time convolution method, this paper presents the synthesis of a simplified architecture of a reconfigurable optical chirp z-transform (OCZT) processor based on the silica-based planar lightwave circuit (PLC) technology. In the simplified architecture of the reconfigurable OCZT, the required number of optical components is small and there are no waveguide crossings which make fabrication easy. The design of a novel type of optical discrete Fourier transform (ODFT) processor as a special case of the synthesized OCZT is then presented to demonstrate its effectiveness. The designed ODFT can be potentially used as an optical demultiplexer at the receiver of an optical fiber orthogonal frequency division multiplexing (OFDM) transmission system.

  18. Cylindrical Invisibility Cloak with Simplified Material Parameters is Inherently Visible

    NASA Astrophysics Data System (ADS)

    Yan, Min; Ruan, Zhichao; Qiu, Min

    2007-12-01

    It was proposed that perfect invisibility cloaks can be constructed for hiding objects from electromagnetic illumination [J. B. Pendry, D. Schurig, and D. R. Smith, Science 312, 1780 (2006)SCIEAS0036-807510.1126/science.1125907]. The cylindrical cloaks experimentally demonstrated [D. Schurig , Science 314, 977 (2006)SCIEAS0036-807510.1126/science.1133628] and theoretically proposed [W. Cai , Nat. Photon. 1, 224 (2007)1749-488510.1038/nphoton.2007.28] have however simplified material parameters in order to facilitate easier realization as well as to avoid infinities in optical constants. Here we show that the cylindrical cloaks with simplified material parameters inherently allow the zeroth-order cylindrical wave to pass through the cloak as if the cloak is made of a homogeneous isotropic medium, and thus visible. To all high-order cylindrical waves, our numerical simulation suggests that the simplified cloak inherits some properties of the ideal cloak, but finite scatterings exist.

  19. Simplifying EPID dosimetry for IMRT treatment verification

    SciTech Connect

    Pecharroman-Gallego, R.; Mans, Anton; Sonke, Jan-Jakob; Stroom, Joep C.; Olaciregui-Ruiz, Igor; Herk, Marcel van; Mijnheer, Ben J.

    2011-02-15

    Purpose: Electronic portal imaging devices (EPIDs) are increasingly used for IMRT dose verification, both pretreatment and in vivo. In this study, an earlier developed backprojection model has been modified to avoid the need for patient-specific transmission measurements and, consequently, leads to a faster procedure. Methods: Currently, the transmission, an essential ingredient of the backprojection model, is estimated from the ratio of EPID measurements with and without a phantom/patient in the beam. Thus, an additional irradiation to obtain ''open images'' under the same conditions as the actual phantom/patient irradiation is required. However, by calculating the transmission of the phantom/patient in the direction of the beam instead of using open images, this extra measurement can be avoided. This was achieved by using a model that includes the effect of beam hardening and off-axis dependence of the EPID response on photon beam spectral changes. The parameters in the model were empirically obtained by performing EPID measurements using polystyrene slab phantoms of different thickness in 6, 10, and 18 MV photon beams. A theoretical analysis to verify the sensitivity of the model with patient thickness changes was performed. The new model was finally applied for the analysis of EPID dose verification measurements of step-and-shoot IMRT treatments of head and neck, lung, breast, cervix, prostate, and rectum patients. All measurements were carried out using Elekta SL20i linear accelerators equipped with a hydrogenated amorphous silicon EPID, and the IMRT plans were made using PINNACLE software (Philips Medical Systems). Results: The results showed generally good agreement with the dose determined using the old model applying the measured transmission. The average differences between EPID-based in vivo dose at the isocenter determined using either the new model for transmission and its measured value were 2.6{+-}3.1%, 0.2{+-}3.1%, and 2.2{+-}3.9% for 47 patients

  20. Simulation and simplified design studies of photovoltaic systems

    SciTech Connect

    Evans, D.L.; Facinelli, W.A.; Koehler, L.P.

    1980-09-01

    Results of TRNSYS simulations of photovoltaic systems with electrical storage are described. Studies of the sensitivity of system performance, in terms of the fraction of the electrical load supplied by the solar energy system, to variables such as array size, battery size, location, time of year, and load shape are reported. An accurate simplified method for predicting array output of max-power photovoltaic systems is presented. A second simplified method, which estimates the overall performance of max-power systems, is developed. Finally, a preliminary technique for predicting clamped-voltage system performance is discussed.

  1. A simplified dynamic model of the T700 turboshaft engine

    NASA Technical Reports Server (NTRS)

    Duyar, Ahmet; Gu, Zhen; Litt, Jonathan S.

    1992-01-01

    A simplified open-loop dynamic model of the T700 turboshaft engine, valid within the normal operating range of the engine, is developed. This model is obtained by linking linear state space models obtained at different engine operating points. Each linear model is developed from a detailed nonlinear engine simulation using a multivariable system identification and realization method. The simplified model may be used with a model-based real time diagnostic scheme for fault detection and diagnostics, as well as for open loop engine dynamics studies and closed loop control analysis utilizing a user generated control law.

  2. Simplifying Probability Elicitation and Uncertainty Modeling in Bayesian Networks

    SciTech Connect

    Paulson, Patrick R; Carroll, Thomas E; Sivaraman, Chitra; Neorr, Peter A; Unwin, Stephen D; Hossain, Shamina S

    2011-04-16

    In this paper we contribute two methods that simplify the demands of knowledge elicitation for particular types of Bayesian networks. The first method simplify the task of providing probabilities when the states that a random variable takes can be described by a new, fully ordered state set in which a state implies all the preceding states. The second method leverages Dempster-Shafer theory of evidence to provide a way for the expert to express the degree of ignorance that they feel about the estimates being provided.

  3. Simplified pregnant woman models for the fetus exposure assessment

    NASA Astrophysics Data System (ADS)

    Jala, Marjorie; Conil, Emmanuelle; Varsier, Nadège; Wiart, Joe; Hadjem, Abdelhamid; Moulines, Éric; Lévy-Leduc, Céline

    2013-05-01

    In this paper, we introduce a study that we carried out in order to validate the use of a simplified pregnant woman model for the assessment of the fetus exposure to radio frequency waves. This simplified model, based on the use of a homogeneous tissue to replace most of the inner organs of the virtual mother, would allow us to deal with many issues that are raised because of the lack of pregnant woman models for numerical dosimetry. Using specific absorption rate comparisons, we show that this model could be used to estimate the fetus exposure to plane waves.

  4. Simplified ontologies allowing comparison of developmental mammalian gene expression

    PubMed Central

    Kruger, Adele; Hofmann, Oliver; Carninci, Piero; Hayashizaki, Yoshihide; Hide, Winston

    2007-01-01

    Model organisms represent an important resource for understanding the fundamental aspects of mammalian biology. Mapping of biological phenomena between model organisms is complex and if it is to be meaningful, a simplified representation can be a powerful means for comparison. The Developmental eVOC ontologies presented here are simplified orthogonal ontologies describing the temporal and spatial distribution of developmental human and mouse anatomy. We demonstrate the ontologies by identifying genes showing a bias for developmental brain expression in human and mouse. PMID:17961239

  5. Are Assumptions of Well-Known Statistical Techniques Checked, and Why (Not)?

    PubMed Central

    Hoekstra, Rink; Kiers, Henk A. L.; Johnson, Addie

    2012-01-01

    A valid interpretation of most statistical techniques requires that one or more assumptions be met. In published articles, however, little information tends to be reported on whether the data satisfy the assumptions underlying the statistical techniques used. This could be due to self-selection: Only manuscripts with data fulfilling the assumptions are submitted. Another explanation could be that violations of assumptions are rarely checked for in the first place. We studied whether and how 30 researchers checked fictitious data for violations of assumptions in their own working environment. Participants were asked to analyze the data as they would their own data, for which often used and well-known techniques such as the t-procedure, ANOVA and regression (or non-parametric alternatives) were required. It was found that the assumptions of the techniques were rarely checked, and that if they were, it was regularly by means of a statistical test. Interviews afterward revealed a general lack of knowledge about assumptions, the robustness of the techniques with regards to the assumptions, and how (or whether) assumptions should be checked. These data suggest that checking for violations of assumptions is not a well-considered choice, and that the use of statistics can be described as opportunistic. PMID:22593746

  6. Dissecting jets and missing energy searches using n-body extended simplified models

    NASA Astrophysics Data System (ADS)

    Cohen, Timothy; Dolan, Matthew J.; El Hedri, Sonia; Hirschauer, James; Tran, Nhan; Whitbeck, Andrew

    2016-08-01

    Simplified Models are a useful way to characterize new physics scenarios for the LHC. Particle decays are often represented using non-renormalizable operators that involve the minimal number of fields required by symmetries. Generalizing to a wider class of decay operators allows one to model a variety of final states. This approach, which we dub the n-body extension of Simplified Models, provides a unifying treatment of the signal phase space resulting from a variety of signals. In this paper, we present the first application of this framework in the context of multijet plus missing energy searches. The main result of this work is a global performance study with the goal of identifying which set of observables yields the best discriminating power against the largest Standard Model backgrounds for a wide range of signal jet multiplicities. Our analysis compares combinations of one, two and three variables, placing emphasis on the enhanced sensitivity gain resulting from non-trivial correlations. Utilizing boosted decision trees, we compare and classify the performance of missing energy, energy scale and energy structure observables. We demonstrate that including an observable from each of these three classes is required to achieve optimal performance. This work additionally serves to establish the utility of n-body extended Simplified Models as a diagnostic for unpacking the relative merits of different search strategies, thereby motivating their application to new physics signatures beyond jets and missing energy.

  7. Dissecting jets and missing energy searches using $n$-body extended simplified models

    DOE PAGES

    Cohen, Timothy; Dolan, Matthew J.; El Hedri, Sonia; ...

    2016-08-04

    Simplified Models are a useful way to characterize new physics scenarios for the LHC. Particle decays are often represented using non-renormalizable operators that involve the minimal number of fields required by symmetries. Generalizing to a wider class of decay operators allows one to model a variety of final states. This approach, which we dub the $n$-body extension of Simplified Models, provides a unifying treatment of the signal phase space resulting from a variety of signals. In this paper, we present the first application of this framework in the context of multijet plus missing energy searches. The main result of thismore » work is a global performance study with the goal of identifying which set of observables yields the best discriminating power against the largest Standard Model backgrounds for a wide range of signal jet multiplicities. Our analysis compares combinations of one, two and three variables, placing emphasis on the enhanced sensitivity gain resulting from non-trivial correlations. Utilizing boosted decision trees, we compare and classify the performance of missing energy, energy scale and energy structure observables. We demonstrate that including an observable from each of these three classes is required to achieve optimal performance. In conclusion, this work additionally serves to establish the utility of $n$-body extended Simplified Models as a diagnostic for unpacking the relative merits of different search strategies, thereby motivating their application to new physics signatures beyond jets and missing energy.« less

  8. Dissecting jets and missing energy searches using $n$-body extended simplified models

    SciTech Connect

    Cohen, Timothy; Dolan, Matthew J.; El Hedri, Sonia; Hirschauer, James; Tran, Nhan; Whitbeck, Andrew

    2016-08-04

    Simplified Models are a useful way to characterize new physics scenarios for the LHC. Particle decays are often represented using non-renormalizable operators that involve the minimal number of fields required by symmetries. Generalizing to a wider class of decay operators allows one to model a variety of final states. This approach, which we dub the $n$-body extension of Simplified Models, provides a unifying treatment of the signal phase space resulting from a variety of signals. In this paper, we present the first application of this framework in the context of multijet plus missing energy searches. The main result of this work is a global performance study with the goal of identifying which set of observables yields the best discriminating power against the largest Standard Model backgrounds for a wide range of signal jet multiplicities. Our analysis compares combinations of one, two and three variables, placing emphasis on the enhanced sensitivity gain resulting from non-trivial correlations. Utilizing boosted decision trees, we compare and classify the performance of missing energy, energy scale and energy structure observables. We demonstrate that including an observable from each of these three classes is required to achieve optimal performance. In conclusion, this work additionally serves to establish the utility of $n$-body extended Simplified Models as a diagnostic for unpacking the relative merits of different search strategies, thereby motivating their application to new physics signatures beyond jets and missing energy.

  9. Prenatal Substance Use: Exploring Assumptions of Maternal Unfitness

    PubMed Central

    Terplan, Mishka; Kennedy-Hendricks, Alene; Chisolm, Margaret S

    2015-01-01

    In spite of the growing knowledge and understanding of addiction as a chronic relapsing medical condition, individuals with substance use disorders (SUD) continue to experience stigmatization. Pregnant women who use substances suffer additional stigma as their use has the potential to cause fetal harm, calling into question their maternal fitness and often leading to punitive responses. Punishing pregnant women denies the integral interconnectedness of the maternal-fetal dyad. Linking substance use with maternal unfitness is not supported by the balance of the scientific evidence regarding the actual harms associated with substance use during pregnancy. Such linkage adversely impacts maternal, child, and family health by deterring pregnant women from seeking both obstetrical care and SUD treatment. Pregnant women who use substances deserve compassion and care, not pariah-status and punishment. PMID:26448685

  10. Simplified versus geometrically accurate models of forefoot anatomy to predict plantar pressures: A finite element study.

    PubMed

    Telfer, Scott; Erdemir, Ahmet; Woodburn, James; Cavanagh, Peter R

    2016-01-25

    Integration of patient-specific biomechanical measurements into the design of therapeutic footwear has been shown to improve clinical outcomes in patients with diabetic foot disease. The addition of numerical simulations intended to optimise intervention design may help to build on these advances, however at present the time and labour required to generate and run personalised models of foot anatomy restrict their routine clinical utility. In this study we developed second-generation personalised simple finite element (FE) models of the forefoot with varying geometric fidelities. Plantar pressure predictions from barefoot, shod, and shod with insole simulations using simplified models were compared to those obtained from CT-based FE models incorporating more detailed representations of bone and tissue geometry. A simplified model including representations of metatarsals based on simple geometric shapes, embedded within a contoured soft tissue block with outer geometry acquired from a 3D surface scan was found to provide pressure predictions closest to the more complex model, with mean differences of 13.3kPa (SD 13.4), 12.52kPa (SD 11.9) and 9.6kPa (SD 9.3) for barefoot, shod, and insole conditions respectively. The simplified model design could be produced in <1h compared to >3h in the case of the more detailed model, and solved on average 24% faster. FE models of the forefoot based on simplified geometric representations of the metatarsal bones and soft tissue surface geometry from 3D surface scans may potentially provide a simulation approach with improved clinical utility, however further validity testing around a range of therapeutic footwear types is required.

  11. Sensitivity of Earthquake Loss Estimates to Source Modeling Assumptions and Uncertainty

    USGS Publications Warehouse

    Reasenberg, Paul A.; Shostak, Nan; Terwilliger, Sharon

    2006-01-01

    Introduction: This report explores how uncertainty in an earthquake source model may affect estimates of earthquake economic loss. Specifically, it focuses on the earthquake source model for the San Francisco Bay region (SFBR) created by the Working Group on California Earthquake Probabilities. The loss calculations are made using HAZUS-MH, a publicly available computer program developed by the Federal Emergency Management Agency (FEMA) for calculating future losses from earthquakes, floods and hurricanes within the United States. The database built into HAZUS-MH includes a detailed building inventory, population data, data on transportation corridors, bridges, utility lifelines, etc. Earthquake hazard in the loss calculations is based upon expected (median value) ground motion maps called ShakeMaps calculated for the scenario earthquake sources defined in WGCEP. The study considers the effect of relaxing certain assumptions in the WG02 model, and explores the effect of hypothetical reductions in epistemic uncertainty in parts of the model. For example, it addresses questions such as what would happen to the calculated loss distribution if the uncertainty in slip rate in the WG02 model were reduced (say, by obtaining additional geologic data)? What would happen if the geometry or amount of aseismic slip (creep) on the region's faults were better known? And what would be the effect on the calculated loss distribution if the time-dependent earthquake probability were better constrained, either by eliminating certain probability models or by better constraining the inherent randomness in earthquake recurrence? The study does not consider the effect of reducing uncertainty in the hazard introduced through models of attenuation and local site characteristics, although these may have a comparable or greater effect than does source-related uncertainty. Nor does it consider sources of uncertainty in the building inventory, building fragility curves, and other assumptions

  12. Simplified preliminary economic analysis for passive solar heating. Master's thesis

    SciTech Connect

    Baldetti, P.J.; Lockard, M.A.

    1983-09-01

    This report establishes economic feasibility criteria for considering the use of passive solar design. In light of the growing cost of supplying the energy demands of the Air Force, a method is needed to simplify the adaptation of passive solar heating and cooling in future building construction.

  13. Simplified physically based model of earthen embankment breaching

    Technology Transfer Automated Retrieval System (TEKTRAN)

    A simplified physically based model has been developed to simulate the breaching processes of homogenous and composite earthen embankments owing to overtopping and piping. The breach caused by overtopping flow is approximated as a flat broad-crested weir with a trapezoidal cross section, downstream ...

  14. 77 FR 54482 - Allocation of Costs Under the Simplified Methods

    Federal Register 2010, 2011, 2012, 2013, 2014

    2012-09-05

    ... 26 CFR Part 1 [REG-126770-06] RIN 1545-BG07 Allocation of Costs Under the Simplified Methods AGENCY... contains proposed regulations on allocating costs to certain property produced by the taxpayer or acquired... of property that are required to capitalize certain costs to the property and that allocate...

  15. A simplified ductile-brittle transition temperature tester

    NASA Technical Reports Server (NTRS)

    Arias, A.

    1973-01-01

    The construction and operation of a versatile, simplified bend tester is described. The tester is usable at temperatures from - 192 to 650 C in air. Features of the tester include a single test chamber for cryogenic or elevated temperatures, specimen alining support rollers, and either manual or motorized operation.

  16. Some Observations on Cohesion and Coherence in Simplified Texts.

    ERIC Educational Resources Information Center

    Lautamatti, Liisa

    Analysis of several texts simplified to teach reading comprehension to college second language students shows that simplification affects both textual cohesion and coherence. Moreover, simplification restricts the amount of significant semantic information and creates a more transparent pattern of reference. The use of cohesive conjunctions varies…

  17. Simplified Load-Following Control for a Fuel Cell System

    NASA Technical Reports Server (NTRS)

    Vasquez, Arturo

    2010-01-01

    A simplified load-following control scheme has been proposed for a fuel cell power system. The scheme could be used to control devices that are important parts of a fuel cell system but are sometimes characterized as parasitic because they consume some of the power generated by the fuel cells.

  18. A Simplified Technique for Evaluating Human "CCR5" Genetic Polymorphism

    ERIC Educational Resources Information Center

    Falteisek, Lukáš; Cerný, Jan; Janštová, Vanda

    2013-01-01

    To involve students in thinking about the problem of AIDS (which is important in the view of nondecreasing infection rates), we established a practical lab using a simplified adaptation of Thomas's (2004) method to determine the polymorphism of HIV co-receptor CCR5 from students' own epithelial cells. CCR5 is a receptor involved in inflammatory…

  19. Simplified seismic performance assessment and implications for seismic design

    NASA Astrophysics Data System (ADS)

    Sullivan, Timothy J.; Welch, David P.; Calvi, Gian Michele

    2014-08-01

    The last decade or so has seen the development of refined performance-based earthquake engineering (PBEE) approaches that now provide a framework for estimation of a range of important decision variables, such as repair costs, repair time and number of casualties. This paper reviews current tools for PBEE, including the PACT software, and examines the possibility of extending the innovative displacement-based assessment approach as a simplified structural analysis option for performance assessment. Details of the displacement-based s+eismic assessment method are reviewed and a simple means of quickly assessing multiple hazard levels is proposed. Furthermore, proposals for a simple definition of collapse fragility and relations between equivalent single-degree-of-freedom characteristics and multi-degree-of-freedom story drift and floor acceleration demands are discussed, highlighting needs for future research. To illustrate the potential of the methodology, performance measures obtained from the simplified method are compared with those computed using the results of incremental dynamic analyses within the PEER performance-based earthquake engineering framework, applied to a benchmark building. The comparison illustrates that the simplified method could be a very effective conceptual seismic design tool. The advantages and disadvantages of the simplified approach are discussed and potential implications of advanced seismic performance assessments for conceptual seismic design are highlighted through examination of different case study scenarios including different structural configurations.

  20. A Simplified Diagnostic Method for Elastomer Bond Durability

    NASA Technical Reports Server (NTRS)

    White, Paul

    2009-01-01

    A simplified method has been developed for determining bond durability under exposure to water or high humidity conditions. It uses a small number of test specimens with relatively short times of water exposure at elevated temperature. The method is also gravimetric; the only equipment being required is an oven, specimen jars, and a conventional laboratory balance.

  1. 48 CFR 1532.003 - Simplified acquisition procedures financing.

    Code of Federal Regulations, 2010 CFR

    2010-10-01

    ... procedures financing. 1532.003 Section 1532.003 Federal Acquisition Regulations System ENVIRONMENTAL... procedures financing. (a) Scope. This subpart provides for authorization of advance and interim payments on commercial item orders not exceeding the simplified acquisition threshold. Advance payments are payments...

  2. 48 CFR 1532.003 - Simplified acquisition procedures financing.

    Code of Federal Regulations, 2012 CFR

    2012-10-01

    ... procedures financing. 1532.003 Section 1532.003 Federal Acquisition Regulations System ENVIRONMENTAL... procedures financing. (a) Scope. This subpart provides for authorization of advance and interim payments on commercial item orders not exceeding the simplified acquisition threshold. Advance payments are payments...

  3. 48 CFR 1532.003 - Simplified acquisition procedures financing.

    Code of Federal Regulations, 2014 CFR

    2014-10-01

    ... procedures financing. 1532.003 Section 1532.003 Federal Acquisition Regulations System ENVIRONMENTAL... procedures financing. (a) Scope. This subpart provides for authorization of advance and interim payments on commercial item orders not exceeding the simplified acquisition threshold. Advance payments are payments...

  4. Measuring Phantom Recollection in the Simplified Conjoint Recognition Paradigm

    ERIC Educational Resources Information Center

    Stahl, Christoph; Klauer, Karl Christoph

    2009-01-01

    False memories are sometimes strong enough to elicit recollective experiences. This phenomenon has been termed Phantom Recollection (PR). The Conjoint Recognition (CR) paradigm has been used to empirically separate PR from other memory processes. Recently, a simplification of the CR procedure has been proposed. We herein extend the simplified CR…

  5. Psychometric Evaluation of the Simplified Chinese Version of Flourishing Scale

    ERIC Educational Resources Information Center

    Tang, Xiaoqing; Duan, Wenjie; Wang, Zhizhang; Liu, Tianyuan

    2016-01-01

    Objectives: The Flourishing Scale (FS) was developed to measure psychological well-being from the eudaimonic perspective, highlighting the flourishing of human functioning. This article evaluated the psychometric characteristics of the simplified Chinese version of FS among a Chinese community population. Method: A total of 433 participants from…

  6. Synthesis of (+)-crocacin D and simplified bioactive analogues.

    PubMed

    Pasqua, Adele E; Ferrari, Frank D; Crawford, James J; Whittingham, William G; Marquez, Rodolfo

    2015-03-01

    The total synthesis of (+)-crocacin D has been achieved in 15 steps (9 isolated intermediates) and 14% overall yield from commercially available starting materials and using (+)-crocacin C as a key intermediate. A number of simplified analogues and their biological activities are also reported.

  7. The new soliton solutions via modified Hereman's simplified method

    NASA Astrophysics Data System (ADS)

    Pınar, Zehra

    2017-02-01

    It is well recognized that new types of exact travelling wave solutions to nonlinear partial differential equations can be obtained modifications of the methods which are in hand. In this study, the new ansatz which was proposed by Wazwaz is considered to get new soliton solutions modifying the methodology which is known as Hereman's simplified method.

  8. Simplify Web Development for Faculty and Promote Instructional Design.

    ERIC Educational Resources Information Center

    Pedersen, David C.

    Faculty members are often overwhelmed with the prospect of implementing Web-based instruction. In an effort to simplify the process and incorporate some basic instructional design elements, the Educational Technology Team at Embry Riddle Aeronautical University created a course template for WebCT. Utilizing rapid prototyping, the template…

  9. 48 CFR 14.201-9 - Simplified contract format.

    Code of Federal Regulations, 2013 CFR

    2013-10-01

    ... CONTRACTING METHODS AND CONTRACT TYPES SEALED BIDDING Solicitation of Bids 14.201-9 Simplified contract format...) Contract line item number. (2) Description of supplies or services, or data sufficient to identify the..., but shall retain them in the contract file. Award by acceptance of a bid on the award portion of...

  10. 48 CFR 14.201-9 - Simplified contract format.

    Code of Federal Regulations, 2011 CFR

    2011-10-01

    ... CONTRACTING METHODS AND CONTRACT TYPES SEALED BIDDING Solicitation of Bids 14.201-9 Simplified contract format...) Contract line item number. (2) Description of supplies or services, or data sufficient to identify the..., but shall retain them in the contract file. Award by acceptance of a bid on the award portion of...

  11. 7 CFR 273.25 - Simplified Food Stamp Program.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... 7 Agriculture 4 2011-01-01 2011-01-01 false Simplified Food Stamp Program. 273.25 Section 273.25 Agriculture Regulations of the Department of Agriculture (Continued) FOOD AND NUTRITION SERVICE, DEPARTMENT OF AGRICULTURE FOOD STAMP AND FOOD DISTRIBUTION PROGRAM CERTIFICATION OF ELIGIBLE HOUSEHOLDS Program...

  12. 7 CFR 273.25 - Simplified Food Stamp Program.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... 7 Agriculture 4 2010-01-01 2010-01-01 false Simplified Food Stamp Program. 273.25 Section 273.25 Agriculture Regulations of the Department of Agriculture (Continued) FOOD AND NUTRITION SERVICE, DEPARTMENT OF AGRICULTURE FOOD STAMP AND FOOD DISTRIBUTION PROGRAM CERTIFICATION OF ELIGIBLE HOUSEHOLDS § 273.25...

  13. 7 CFR 273.25 - Simplified Food Stamp Program.

    Code of Federal Regulations, 2012 CFR

    2012-01-01

    ... 7 Agriculture 4 2012-01-01 2012-01-01 false Simplified Food Stamp Program. 273.25 Section 273.25 Agriculture Regulations of the Department of Agriculture (Continued) FOOD AND NUTRITION SERVICE, DEPARTMENT OF AGRICULTURE FOOD STAMP AND FOOD DISTRIBUTION PROGRAM CERTIFICATION OF ELIGIBLE HOUSEHOLDS Program...

  14. 7 CFR 273.25 - Simplified Food Stamp Program.

    Code of Federal Regulations, 2013 CFR

    2013-01-01

    ... 7 Agriculture 4 2013-01-01 2013-01-01 false Simplified Food Stamp Program. 273.25 Section 273.25 Agriculture Regulations of the Department of Agriculture (Continued) FOOD AND NUTRITION SERVICE, DEPARTMENT OF AGRICULTURE FOOD STAMP AND FOOD DISTRIBUTION PROGRAM CERTIFICATION OF ELIGIBLE HOUSEHOLDS Program...

  15. 7 CFR 273.25 - Simplified Food Stamp Program.

    Code of Federal Regulations, 2014 CFR

    2014-01-01

    ... 7 Agriculture 4 2014-01-01 2014-01-01 false Simplified Food Stamp Program. 273.25 Section 273.25 Agriculture Regulations of the Department of Agriculture (Continued) FOOD AND NUTRITION SERVICE, DEPARTMENT OF AGRICULTURE FOOD STAMP AND FOOD DISTRIBUTION PROGRAM CERTIFICATION OF ELIGIBLE HOUSEHOLDS Program...

  16. Molded elastomer provides compact ferrite-core holder, simplifies assembly

    NASA Technical Reports Server (NTRS)

    Hayden, R. R.

    1964-01-01

    A ferrite-core holder, fabricated by casting an elastomer in a simple mold, simplifies the assembly of modular matrix units for computers. Use of the device permits the core leads to be multiply threaded and soldered to terminals, without requiring intermediate terminals.

  17. Cosmological perturbations and quasistatic assumption in f (R ) theories

    NASA Astrophysics Data System (ADS)

    Chiu, Mu-Chen; Taylor, Andy; Shu, Chenggang; Tu, Hong

    2015-11-01

    f (R ) gravity is one of the simplest theories of modified gravity to explain the accelerated cosmic expansion. Although it is usually assumed that the quasi-Newtonian approach (a combination of the quasistatic approximation and sub-Hubble limit) for cosmic perturbations is good enough to describe the evolution of large scale structure in f (R ) models, some studies have suggested that this method is not valid for all f (R ) models. Here, we show that in the matter-dominated era, the pressure and shear equations alone, which can be recast into four first-order equations to solve for cosmological perturbations exactly, are sufficient to solve for the Newtonian potential, Ψ , and the curvature potential, Φ . Based on these two equations, we are able to clarify how the exact linear perturbations fit into different limits. We find that the Compton length controls the quasistatic behaviors in f (R ) gravity. In addition, regardless the validity of quasistatic approximation, a strong version of the sub-Hubble limit alone is sufficient to reduce the exact linear perturbations in any viable f (R ) gravity to second order. Our findings disagree with some previous studies where we find little difference between our exact and quasi-Newtonian solutions even up to k =10 c-1H0.

  18. On Simplifying Features in OpenStreetMap database

    NASA Astrophysics Data System (ADS)

    Qian, Xinlin; Tao, Kunwang; Wang, Liang

    2015-04-01

    Currently the visualization of OpenStreetMap data is using a tile server which stores map tiles that have been rendered from vector data in advance. However, tiled map are short of functionalities such as data editing and customized styling. To enable these advanced functionality, Client-side processing and rendering of geospatial data is needed. Considering the voluminous size of the OpenStreetMap data, simply sending region queries results of OSM database to client is prohibitive. To make the OSM data retrieved from database adapted for client receiving and rendering, It must be filtered and simplified at server-side to limit its volume. We propose a database extension for OSM database to make it possible to simplifying geospatial objects such as ways and relations during data queries. Several auxiliary tables and PL/pgSQL functions are presented to make the geospatial features can be simplified by omitting unimportant vertices. There are five components in the database extension: Vertices weight computation by polyline and polygon simplification algorithm, Vertices weight storage in auxiliary tables. filtering and selecting of vertices using specific threshold value during spatial queries, assembling of simplified geospatial objects using filtered vertices, vertices weight updating after geospatial objects editing. The database extension is implemented on an OSM APIDB using PL/pgSQL. The database contains a subset of OSM database. The experimental database contains geographic data of United Kingdom which is about 100 million vertices and roughly occupy 100GB disk. JOSM are used to retrieve the data from the database using a revised data accessing API and render the geospatial objects in real-time. When serving simplified data to client, The database allows user to set the bound of the error of simplification or the bound of responding time in each data query. Experimental results show the effectiveness and efficiency of the proposed methods in building a

  19. The Consistency Assumption for Causal Inference in Social Epidemiology: When a Rose is Not a Rose

    PubMed Central

    Rehkopf, David H.; Glymour, M. Maria; Osypuk, Theresa L.

    2016-01-01

    The assumption that exposures as measured in observational settings have clear and specific definitions underpins epidemiologic research and allows us to use observational data to predict outcomes in interventions. This leap between exposures as measured and exposures as intervened upon is typically supported by the consistency assumption. The consistency assumption has received extensive attention in risk factor epidemiology but relatively little emphasis in social epidemiology. However, violations of the consistency assumption may be especially important to consider when understanding how social and economic exposures influence health. Efforts to clarify the definitions of our exposures, thus bolstering the consistency assumption, will help guide interventions to improve population health and reduce health disparities. This article focuses on the consistency assumption as considered within social epidemiology. We explain how this assumption is articulated in the causal inference literature and give examples of how it might be violated for three common exposure in social epidemiology research: income, education and neighborhood characteristics. We conclude that there is good reason to worry about consistency assumption violations in much of social epidemiology research. Theoretically motivated explorations of mechanisms along with empirical comparisons of research findings under alternative operationalizations of exposure can help identify consistency violations. We recommend that future social epidemiology studies be more explicit to name and discuss the consistency assumption when describing the exposure of interest, including reconciling disparate results in the literature. PMID:27326386

  20. Testing assumptions and predictions of star formation theories

    NASA Astrophysics Data System (ADS)

    González-Samaniego, Alejandro; Vázquez-Semadeni, Enrique; González, Ricardo F.; Kim, Jongsoo

    2014-05-01

    We present numerical simulations of isothermal, magnetohydrodynamic (MHD), supersonic turbulence, designed to test various hypotheses frequently assumed in star formation (SF) theories. This study complements our previous one in the non-magnetic (HD) case. We consider three simulations, each with different values of its physical size, rms sonic Mach number M_s, and Jeans parameter J, but so that all three have the same value of the virial parameter and conform with Larson's scaling relations. As in the non-magnetic case, we find that (1) no structures that are both subsonic and super-Jeans are produced; (2) that the fraction of small-scale super-Jeans structures increases when self-gravity is turned on, and the production of very dense cores by turbulence alone is very low. This implies that self-gravity is involved not only in the collapse of Jeans-unstable cores, but also in their formation. (3) We also find that denser regions tend to have a stronger velocity convergence, implying a net inwards flow towards the regions' centres. Contrary to the non-magnetic case, we find that the magnetic simulation with lowest values of M_s and J (respectively, 5 and 2) does not produce any collapsing regions for over three simulation free-fall times, in spite of being both Jeans-unstable and magnetically supercritical. We attribute this result to the combined thermal and magnetic support. Next, we compare the results of our HD and MHD simulations with the predictions from the recent SF theories by Krumholz & McKee, Padoan & Nordlund, and Hennebelle & Chabrier, using expressions recently provided by Federrath & Klessen, which extend those theories to the magnetic case. In both the HD and MHD cases, we find that the theoretical predictions tend to be larger than the SFEff measured in the simulations. In the MHD case, none of the theories captures the suppression of collapse at low values of Jeff by the additional support from the magnetic field. We conclude that randomly driven

  1. Simplified stratigraphic cross sections of the Eocene Green River Formation in the Piceance Basin, northwestern Colorado

    USGS Publications Warehouse

    Dietrich, John D.; Johnson, Ronald C.

    2013-01-01

    Thirteen stratigraphic cross sections of the Eocene Green River Formation in the Piceance Basin of northwestern Colorado are presented in this report. Originally published in a much larger and more detailed form by Self and others (2010), they are shown here in simplified, page-size versions that are easily accessed and used for presentation purposes. Modifications to the original versions include the elimination of the detailed lithologic columns and oil-yield histograms from Fischer assay data and the addition of ground-surface lines to give the depth of the various oil shale units shown on the cross section.

  2. FUNDAMENTAL AREAS OF PHENOMENOLOGY (INCLUDING APPLICATIONS): Optical Image Encryption with Simplified Fractional Hartley Transform

    NASA Astrophysics Data System (ADS)

    Li, Xin-Xin; Zhao, Dao-Mu

    2008-07-01

    We present a new method for image encryption on the basis of simplified fractional Hartley transform (SFRHT). SFRHT is a real transform as Hartley transform (HT) and furthermore, superior to HT in virtue of the advantage that it can also append fractional orders as additional keys for the purpose of improving the system security to some extent. With this method, one can encrypt an image with an intensity-only medium such as a photographic film or a CCD camera by spatially incoherent or coherent illumination. The optical realization is then proposed and computer simulations are also performed to verify the feasibility of this method.

  3. Affi-gel blue treatment simplifies the protein composition of sarcoplasmic reticulum vesicles.

    PubMed

    Papp, S; Dux, L; Martonosi, A

    1986-04-01

    Sarcoplasmic reticulum vesicles isolated by conventional techniques usually contain, in addition to the recognized sarcoplasmic reticulum components, several other proteins (phosphorylase, myosin, glyceraldehyde-3-phosphate dehydrogenase, etc.) in variable amounts; these proteins complicate the interpretation of chemical modification data. Incubation of sarcoplasmic reticulum vesicles with Affi-Gel blue particles for 1-4 h at 2 degrees C, followed by sedimentation of the Affi-Gel in a clinical centrifuge, simplifies the protein composition by selective adsorption of the accessory proteins, and improves the consistency of the preparations. The Affi-Gel blue treatment is recommended as part of the standard procedure for the isolation of sarcoplasmic reticulum vesicles.

  4. Testing assumptions for conservation of migratory shorebirds and coastal managed wetlands

    USGS Publications Warehouse

    Collazo, Jaime; James Lyons,; Herring, Garth

    2015-01-01

    Managed wetlands provide critical foraging and roosting habitats for shorebirds during migration; therefore, ensuring their availability is a priority action in shorebird conservation plans. Contemporary shorebird conservation plans rely on a number of assumptions about shorebird prey resources and migratory behavior to determine stopover habitat requirements. For example, the US Shorebird Conservation Plan for the Southeast-Caribbean region assumes that average benthic invertebrate biomass in foraging habitats is 2.4 g dry mass m−2 and that the dominant prey item of shorebirds in the region is Chironomid larvae. For effective conservation and management, it is important to test working assumptions and update predictive models that are used to estimate habitat requirements. We surveyed migratory shorebirds and sampled the benthic invertebrate community in coastal managed wetlands of South Carolina. We sampled invertebrates at three points in time representing early, middle, and late stages of spring migration, and concurrently surveyed shorebird stopover populations at approximately 7-day intervals throughout migration. We used analysis of variance by ranks to test for temporal variation in invertebrate biomass and density, and we used a model based approach (linear mixed model and Monte Carlo simulation) to estimate mean biomass and density. There was little evidence of a temporal variation in biomass or density during the course of spring shorebird migration, suggesting that shorebirds did not deplete invertebrate prey resources at our site. Estimated biomass was 1.47 g dry mass m−2 (95 % credible interval 0.13–3.55), approximately 39 % lower than values used in the regional shorebird conservation plan. An additional 4728 ha (a 63 % increase) would be required if habitat objectives were derived from biomass levels observed in our study. Polychaetes, especially Laeonereis culveri(2569 individuals m−2), were the most abundant prey in foraging

  5. Volume overview: Working with assumptions. Existing and emerging approaches for improved program design, monitoring and evaluation.

    PubMed

    Nkwake, Apollo M; Morrow, Nathan

    2016-12-01

    This volume attempts to systematically capture the state of practice, highlight commonalities linking existing and emerging approaches to assumption-making and evaluation. It tries to organize existing and emerging knowledge, tools and terminology into an emergent but useful typology for working with assumptions and complexity in program designs, monitoring and evaluation.

  6. Assessing Key Assumptions of Network Meta-Analysis: A Review of Methods

    ERIC Educational Resources Information Center

    Donegan, Sarah; Williamson, Paula; D'Alessandro, Umberto; Tudur Smith, Catrin

    2013-01-01

    Background: Homogeneity and consistency assumptions underlie network meta-analysis (NMA). Methods exist to assess the assumptions but they are rarely and poorly applied. We review and illustrate methods to assess homogeneity and consistency. Methods: Eligible articles focussed on indirect comparison or NMA methodology. Articles were sought by…

  7. The Holmes Report: Epistemological Assumptions that Impact Art Teacher Assessment and Preparation.

    ERIC Educational Resources Information Center

    Maitland-Gholson, Jane

    1988-01-01

    Examines implicit assumptions about knowledge and learning found in the "Holmes Report" on teacher preparation. Considers the impact of these epistemological assumptions on conceptions of art knowledge, learning, and teacher assessment. Suggests that art education could provide a needed evaluative counterpoint to current trends toward…

  8. 33 CFR Appendix B to Part 157 - Subdivision and Stability Assumptions

    Code of Federal Regulations, 2011 CFR

    2011-07-01

    ... space located aft is involved in the damage assumption. The machinery space is calculated as a single... between adjacent transverse bulkheads except the machinery space. (b) The extent and the character of the... the outer hull. 4. Characteristic and Condition Assumption for Calculations. (a) Account must be...

  9. A Test of the Dimensionality Assumptions of Rotter's Internal-External Scale

    ERIC Educational Resources Information Center

    Klockars, Alan J.; Varnum, Susan W.

    1975-01-01

    Examined two assumptions about the dimensionality of Rotters' Internal-External (I-E) scale: First, the bipolarity of the two statements within each item pair; second, the unidimensionality of the overall construct. Both assumptions regarding Rotters' I-E Scale were found untenable. (Author/BJG)

  10. 77 FR 24517 - Solicitation of Comments on Request for United States Assumption of Concurrent Federal Criminal...

    Federal Register 2010, 2011, 2012, 2013, 2014

    2012-04-24

    ... Solicitation of Comments on Request for United States Assumption of Concurrent Federal Criminal Jurisdiction... Criminal Jurisdiction recently submitted to the Office of Tribal Justice, Department of Justice by the... States assumption of concurrent federal criminal jurisdiction submitted by the Hoopa Valley Tribe is...

  11. 77 FR 64548 - Solicitation of Comments on Request for United States Assumption of Concurrent Federal Criminal...

    Federal Register 2010, 2011, 2012, 2013, 2014

    2012-10-22

    ... Solicitation of Comments on Request for United States Assumption of Concurrent Federal Criminal Jurisdiction... Criminal Jurisdiction recently submitted to the Office of Tribal Justice, Department of Justice by the... electronic copy of the request for United States assumption of concurrent federal criminal...

  12. 12 CFR 307.2 - Certification of assumption of deposit liabilities.

    Code of Federal Regulations, 2014 CFR

    2014-01-01

    ... depository institution in default, as defined in section 3(x)(1) of the FDI Act (12 U.S.C. 1813(x)(1)), and... satisfactory evidence of such deposit assumption, as required by section 8(q) of the FDI Act (12 U.S.C. 1818(q... satisfactory evidence of such assumption, pursuant to section 8(q) of the FDI Act and this...

  13. 12 CFR 307.2 - Certification of assumption of deposit liabilities.

    Code of Federal Regulations, 2013 CFR

    2013-01-01

    ... depository institution in default, as defined in section 3(x)(1) of the FDI Act (12 U.S.C. 1813(x)(1)), and... satisfactory evidence of such deposit assumption, as required by section 8(q) of the FDI Act (12 U.S.C. 1818(q... satisfactory evidence of such assumption, pursuant to section 8(q) of the FDI Act and this...

  14. 12 CFR 307.2 - Certification of assumption of deposit liabilities.

    Code of Federal Regulations, 2012 CFR

    2012-01-01

    ... depository institution in default, as defined in section 3(x)(1) of the FDI Act (12 U.S.C. 1813(x)(1)), and... satisfactory evidence of such deposit assumption, as required by section 8(q) of the FDI Act (12 U.S.C. 1818(q... satisfactory evidence of such assumption, pursuant to section 8(q) of the FDI Act and this...

  15. 12 CFR 307.2 - Certification of assumption of deposit liabilities.

    Code of Federal Regulations, 2011 CFR

    2011-01-01

    ... depository institution in default, as defined in section 3(x)(1) of the FDI Act (12 U.S.C. 1813(x)(1)), and... satisfactory evidence of such deposit assumption, as required by section 8(q) of the FDI Act (12 U.S.C. 1818(q... satisfactory evidence of such assumption, pursuant to section 8(q) of the FDI Act and this...

  16. Making Sense out of Sex Stereotypes in Advertising: A Feminist Analysis of Assumptions.

    ERIC Educational Resources Information Center

    Ferrante, Karlene

    Sexism and racism in advertising have been well documented, but feminist research aimed at social change must go beyond existing content analyses to ask how advertising is created. Analysis of the "mirror assumption" (advertising reflects society) and the "gender assumption" (advertising speaks in a male voice to female…

  17. 12 CFR 563.22 - Merger, consolidation, purchase or sale of assets, or assumption of liabilities.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... assets, or assumption of liabilities. 563.22 Section 563.22 Banks and Banking OFFICE OF THRIFT... Merger, consolidation, purchase or sale of assets, or assumption of liabilities. (a) No savings..., the term “transfer” means purchases or sales of assets or liabilities in bulk not made in the...

  18. 12 CFR 307.2 - Certification of assumption of deposit liabilities.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... depository institution in default, as defined in section 3(x)(1) of the FDI Act (12 U.S.C. 1813(x)(1)), and... satisfactory evidence of such deposit assumption, as required by section 8(q) of the FDI Act (12 U.S.C. 1818(q... satisfactory evidence of such assumption, pursuant to section 8(q) of the FDI Act and this...

  19. Robustness of the Polytomous IRT Model to Violations of the Unidimensionality Assumption.

    ERIC Educational Resources Information Center

    Dawadi, Bhaskar R.

    The robustness of the polytomous Item Response Theory (IRT) model to violations of the unidimensionality assumption was studied. A secondary purpose was to provide guidelines to practitioners to help in deciding whether to use an IRT model to analyze their data. In a simulation study, the unidimensionality assumption was deliberately violated by…

  20. 33 CFR Appendix B to Part 157 - Subdivision and Stability Assumptions

    Code of Federal Regulations, 2010 CFR

    2010-07-01

    ... Assumptions B Appendix B to Part 157 Navigation and Navigable Waters COAST GUARD, DEPARTMENT OF HOMELAND... CARRYING OIL IN BULK Pt. 157, App. B Appendix B to Part 157—Subdivision and Stability Assumptions 1. Source... between adjacent transverse bulkheads except the machinery space. (b) The extent and the character of...

  1. 42 CFR 417.120 - Fiscally sound operation and assumption of financial risk.

    Code of Federal Regulations, 2010 CFR

    2010-10-01

    ... 42 Public Health 3 2010-10-01 2010-10-01 false Fiscally sound operation and assumption of...: Organization and Operation § 417.120 Fiscally sound operation and assumption of financial risk. (a) Fiscally sound operation—(1) General requirements. Each HMO must have a fiscally sound operation, as...

  2. 42 CFR 417.120 - Fiscally sound operation and assumption of financial risk.

    Code of Federal Regulations, 2013 CFR

    2013-10-01

    ... 42 Public Health 3 2013-10-01 2013-10-01 false Fiscally sound operation and assumption of... Organizations: Organization and Operation § 417.120 Fiscally sound operation and assumption of financial risk. (a) Fiscally sound operation—(1) General requirements. Each HMO must have a fiscally sound...

  3. 42 CFR 417.120 - Fiscally sound operation and assumption of financial risk.

    Code of Federal Regulations, 2011 CFR

    2011-10-01

    ... 42 Public Health 3 2011-10-01 2011-10-01 false Fiscally sound operation and assumption of...: Organization and Operation § 417.120 Fiscally sound operation and assumption of financial risk. (a) Fiscally sound operation—(1) General requirements. Each HMO must have a fiscally sound operation, as...

  4. 42 CFR 417.120 - Fiscally sound operation and assumption of financial risk.

    Code of Federal Regulations, 2012 CFR

    2012-10-01

    ... 42 Public Health 3 2012-10-01 2012-10-01 false Fiscally sound operation and assumption of... Organizations: Organization and Operation § 417.120 Fiscally sound operation and assumption of financial risk. (a) Fiscally sound operation—(1) General requirements. Each HMO must have a fiscally sound...

  5. 42 CFR 417.120 - Fiscally sound operation and assumption of financial risk.

    Code of Federal Regulations, 2014 CFR

    2014-10-01

    ... 42 Public Health 3 2014-10-01 2014-10-01 false Fiscally sound operation and assumption of... Organizations: Organization and Operation § 417.120 Fiscally sound operation and assumption of financial risk. (a) Fiscally sound operation—(1) General requirements. Each HMO must have a fiscally sound...

  6. An Azulene-Based Discovery Experiment: Challenging Students to Watch for the "False Assumption"

    ERIC Educational Resources Information Center

    Garner, Charles M.

    2005-01-01

    A discovery-based experiment is developed depending on a "false assumption" that the students mistakenly assume they know the structure of a reaction product and are forced to reconcile observations that are inconsistent with this assumption. This experiment involves the chemistry of azulenes, an interesting class of intensely colored aromatic…

  7. Female Offenders: Three Assumptions about Self-Esteem, Sex-Role Identity, and Feminism.

    ERIC Educational Resources Information Center

    Widom, Cathy Spatz

    1979-01-01

    Investigates the validity of three assumptions about self-esteem, sex-role identity, and feminism in female offenders in a study of women awaiting trial in Massachusetts. Results did not support assumptions regarding low self-esteem and increased masculinity in female offenders. Speculations were made about the role femininity plays in…

  8. Examining Design Assumptions for an Information Retrieval Service: SDI Use for Scientific and Technical Databases.

    ERIC Educational Resources Information Center

    Cole, Elliot

    1981-01-01

    Examines the assumptions which underlie the design of SDI systems with respect to context of use, activities supported by SDI, and characteristics which make documents useful. Relates these assumptions to the failure of SDI to gain widespread acceptance among scientific information users. A 43-item reference list is included. (JL)

  9. On TESOL '82. Pacific Perspectives on Language Learning and Teaching. II: Challenging Assumptions.

    ERIC Educational Resources Information Center

    Stevick, Earl W.; And Others

    This section of the TESOL convention volume challenges basic assumptions which are held by language teachers and researchers while at the same time providing other assumptions for professionals to challenge. The following papers are presented: (1) My View of "Teaching Languages: A Way and Ways," by E. Stevick; (2) "'I Got…

  10. Conclusion: Agency in the face of complexity and the future of assumption-aware evaluation practice.

    PubMed

    Morrow, Nathan; Nkwake, Apollo M

    2016-12-01

    This final chapter in the volume pulls together common themes from the diverse set of articles by a group of eight authors in this issue, and presents some reflections on the next steps for improving the ways in which evaluators work with assumptions. Collectively, the authors provide a broad overview of existing and emerging approaches to the articulation and use of assumptions in evaluation theory and practice. The authors reiterate the rationale and key terminology as a common basis for working with assumption in program design and evaluation. They highlight some useful concepts and categorizations to promote more rigorous treatment of assumptions in evaluation. A three-tier framework for fostering agency for assumption-aware evaluation practice is proposed-agency for themselves (evaluators); agency for others (stakeholders); and agency for standards and principles.

  11. A new scenario framework for climate change research: The concept of Shared Climate Policy Assumptions

    SciTech Connect

    Kriegler, Elmar; Edmonds, James A.; Hallegatte, Stephane; Ebi, Kristie L.; Kram, Tom; Riahi, Keywan; Winkler, Harald; Van Vuuren, Detlef

    2014-04-01

    The paper presents the concept of shared climate policy assumptions as an important element of the new scenario framework. Shared climate policy assumptions capture key climate policy dimensions such as the type and scale of mitigation and adaptation measures. They are not specified in the socio-economic reference pathways, and therefore introduce an important third dimension to the scenario matrix architecture. Climate policy assumptions will have to be made in any climate policy scenario, and can have a significant impact on the scenario description. We conclude that a meaningful set of shared climate policy assumptions is useful for grouping individual climate policy analyses and facilitating their comparison. Shared climate policy assumptions should be designed to be policy relevant, and as a set to be broad enough to allow a comprehensive exploration of the climate change scenario space.

  12. Simplified scheme for entanglement preparation with Rydberg pumping via dissipation

    NASA Astrophysics Data System (ADS)

    Su, Shi-Lei; Guo, Qi; Wang, Hong-Fu; Zhang, Shou

    2015-08-01

    Inspired by recent work [Carr and Saffman, Phys. Rev. Lett. 111, 033607 (2013), 10.1103/PhysRevLett.111.033607], we propose a simplified scheme to prepare the two-atom maximally entangled states via dissipative Rydberg pumping. Compared with the former scheme, the simplified one involves fewer classical laser fields and Rydberg interactions, and the asymmetric Rydberg interactions are avoided. Master equation simulations demonstrate that the fidelity and the Clauser-Horne-Shimony-Holt correlation of the maximally entangled state could reach up to 0.999 and 2.821, respectively, under certain conditions. Furthermore, we extend the physical thoughts to prepare the three-dimensional entangled state, and the numerical simulations show that, in theory, both the fidelity and the negativity of the desired entanglement could be very close to unity under certain conditions.

  13. A simplified hydrocarbon reaction mechanism for combustion applications

    NASA Technical Reports Server (NTRS)

    Jachimowski, C. J.

    1985-01-01

    A simplified chemical kinetic reaction mechanism for the combustion of a hydrocarbon fuel is presented and discussed. The observed kinetic behavior of propane combustion as determined in shock tube studies and the calculated kinetic behavior of propane/air mixtures as derived from a detailed propane combustion mechanism were used as a data base for constructing and refining the mechanism. The mechanism contains 13 chemical species and 26 chemical reactions. Numerical studies revealed that the simplified mechanism can reproduce the observed kinetic behavior of propane/air mixtures, including laminar flame speeds, over a wide range of temperatures and pressures. It is anticipated that the mechanism can be used in various combustion applications and provide a more realistic description of the combustion process than can be achieved using global reaction mechanisms.

  14. Simplified partial digest problem: enumerative and dynamic programming algorithms.

    PubMed

    Blazewicz, Jacek; Burke, Edmund; Kasprzak, Marta; Kovalev, Alexandr; Kovalyov, Mikhail

    2007-01-01

    We study the Simplified Partial Digest Problem (SPDP), which is a mathematical model for a new simplified partial digest method of genome mapping. This method is easy for laboratory implementation and robust with respect to the experimental errors. SPDP is NP-hard in the strong sense. We present an $O(n2;n)$ time enumerative algorithm and an O(n(2q)) time dynamic programming algorithm for the error-free SPDP, where $n$ is the number of restriction sites and n is the number of distinct intersite distances. We also give examples of the problem, in which there are 2(n+2)/(3)-1 non-congruent solutions. These examples partially answer a question recently posed in the literature about the number of solutions of SPDP. We adapt our enumerative algorithm for handling SPDP with imprecise input data. Finally, we describe and discuss the results of the computer experiments with our algorithms.

  15. A SIMPLIFIED APPROACH TO ANALYZE AND MODEL INDUCTIVE VOLTAGE ADDER

    SciTech Connect

    ZHANG,W.; ENG, W.; PAI, C.; SANDBERG, J.; TAN, Y.; TIAN, Y.

    2007-06-25

    We have recently developed a simplified model and a set of simple formulas for inductive voltage adder design. This model reveals the relationship of output waveform parameters and hardware designs. A computer simulation has demonstrated that parameter estimation based on this approach is accurate as compared to an actual circuit. This approach can be used in early stages of project development to assist feasibility study, geometry selection in engineering design, and parameter selection of critical components. In this paper, we give the deduction of a simplified model. Among the estimation formulas we present are those for pulse rise time, system impedance, and number of stages. Examples are used to illustrate the advantage of this approach. This approach is also applicable to induction LINAC design.

  16. Working Mechanism for Flexible Perovskite Solar Cells with Simplified Architecture.

    PubMed

    Xu, Xiaobao; Chen, Qi; Hong, Ziruo; Zhou, Huanping; Liu, Zonghao; Chang, Wei-Hsuan; Sun, Pengyu; Chen, Huajun; De Marco, Nicholas; Wang, Mingkui; Yang, Yang

    2015-10-14

    In this communication, we report an efficient and flexible perovskite solar cell based on formamidinium lead trihalide (FAPbI3) with simplified configuration. The device achieved a champion efficiency of 12.70%, utilizing direct contact between metallic indium tin oxide (ITO) electrode and perovskite absorber. The underlying working mechanism is proposed subsequently, via a systematic investigation focusing on the heterojunction within this device. A significant charge storage has been observed in the perovskite, which is believed to generate photovoltage and serves as the driving force for charge transferring from the absorber to ITO electrode as well. More importantly, this simplified device structure on flexible substrates suggests its compatibility for scale-up fabrication, which paves the way for commercialization of perovskite photovoltaic technology.

  17. Interpretation of searches for supersymmetry with simplified models

    NASA Astrophysics Data System (ADS)

    Chatrchyan, S.; Khachatryan, V.; Sirunyan, A. M.; Tumasyan, A.; Adam, W.; Aguilo, E.; Bergauer, T.; Dragicevic, M.; Erö, J.; Fabjan, C.; Friedl, M.; Frühwirth, R.; Ghete, V. M.; Hörmann, N.; Hrubec, J.; Jeitler, M.; Kiesenhofer, W.; Knünz, V.; Krammer, M.; Krätschmer, I.; Liko, D.; Mikulec, I.; Pernicka, M.; Rabady, D.; Rahbaran, B.; Rohringer, C.; Rohringer, H.; Schöfbeck, R.; Strauss, J.; Taurok, A.; Waltenberger, W.; Wulz, C.-E.; Mossolov, V.; Shumeiko, N.; Suarez Gonzalez, J.; Bansal, M.; Bansal, S.; Cornelis, T.; De Wolf, E. A.; Janssen, X.; Luyckx, S.; Mucibello, L.; Ochesanu, S.; Roland, B.; Rougny, R.; Selvaggi, M.; Van Haevermaet, H.; Van Mechelen, P.; Van Remortel, N.; Van Spilbeeck, A.; Blekman, F.; Blyweert, S.; D'Hondt, J.; Gonzalez Suarez, R.; Kalogeropoulos, A.; Maes, M.; Olbrechts, A.; Van Doninck, W.; Van Mulders, P.; Van Onsem, G. P.; Villella, I.; Clerbaux, B.; De Lentdecker, G.; Dero, V.; Gay, A. P. R.; Hreus, T.; Léonard, A.; Marage, P. E.; Mohammadi, A.; Reis, T.; Thomas, L.; Vander Velde, C.; Vanlaer, P.; Wang, J.; Adler, V.; Beernaert, K.; Cimmino, A.; Costantini, S.; Garcia, G.; Grunewald, M.; Klein, B.; Lellouch, J.; Marinov, A.; Mccartin, J.; Ocampo Rios, A. A.; Ryckbosch, D.; Strobbe, N.; Thyssen, F.; Tytgat, M.; Walsh, S.; Yazgan, E.; Zaganidis, N.; Basegmez, S.; Bruno, G.; Castello, R.; Ceard, L.; Delaere, C.; du Pree, T.; Favart, D.; Forthomme, L.; Giammanco, A.; Hollar, J.; Lemaitre, V.; Liao, J.; Militaru, O.; Nuttens, C.; Pagano, D.; Pin, A.; Piotrzkowski, K.; Vizan Garcia, J. M.; Beliy, N.; Caebergs, T.; Daubie, E.; Hammad, G. H.; Alves, G. A.; Correa Martins Junior, M.; Martins, T.; Pol, M. E.; Souza, M. H. G.; Aldá Júnior, W. L.; Carvalho, W.; Custódio, A.; Da Costa, E. M.; De Jesus Damiao, D.; De Oliveira Martins, C.; Fonseca De Souza, S.; Malbouisson, H.; Malek, M.; Matos Figueiredo, D.; Mundim, L.; Nogima, H.; Prado Da Silva, W. L.; Santoro, A.; Soares Jorge, L.; Sznajder, A.; Vilela Pereira, A.; Anjos, T. S.; Bernardes, C. A.; Dias, F. A.; Fernandez Perez Tomei, T. R.; Gregores, E. M.; Lagana, C.; Marinho, F.; Mercadante, P. G.; Novaes, S. F.; Padula, Sandra S.; Genchev, V.; Iaydjiev, P.; Piperov, S.; Rodozov, M.; Stoykova, S.; Sultanov, G.; Tcholakov, V.; Trayanov, R.; Vutova, M.; Dimitrov, A.; Hadjiiska, R.; Kozhuharov, V.; Litov, L.; Pavlov, B.; Petkov, P.; Bian, J. G.; Chen, G. M.; Chen, H. S.; Jiang, C. H.; Liang, D.; Liang, S.; Meng, X.; Tao, J.; Wang, J.; Wang, X.; Wang, Z.; Xiao, H.; Xu, M.; Zang, J.; Zhang, Z.; Asawatangtrakuldee, C.; Ban, Y.; Guo, Y.; Li, W.; Liu, S.; Mao, Y.; Qian, S. J.; Teng, H.; Wang, D.; Zhang, L.; Zou, W.; Avila, C.; Gomez, J. P.; Gomez Moreno, B.; Osorio Oliveros, A. F.; Sanabria, J. C.; Godinovic, N.; Lelas, D.; Plestina, R.; Polic, D.; Puljak, I.; Antunovic, Z.; Kovac, M.; Brigljevic, V.; Duric, S.; Kadija, K.; Luetic, J.; Mekterovic, D.; Morovic, S.; Attikis, A.; Galanti, M.; Mavromanolakis, G.; Mousa, J.; Nicolaou, C.; Ptochos, F.; Razis, P. A.; Finger, M.; Finger, M., Jr.; Assran, Y.; Elgammal, S.; Ellithi Kamel, A.; Mahmoud, M. A.; Mahrous, A.; Radi, A.; Kadastik, M.; Müntel, M.; Raidal, M.; Rebane, L.; Tiko, A.; Eerola, P.; Fedi, G.; Voutilainen, M.; Härkönen, J.; Heikkinen, A.; Karimäki, V.; Kinnunen, R.; Kortelainen, M. J.; Lampén, T.; Lassila-Perini, K.; Lehti, S.; Lindén, T.; Luukka, P.; Mäenpää, T.; Peltola, T.; Tuominen, E.; Tuominiemi, J.; Tuovinen, E.; Ungaro, D.; Wendland, L.; Banzuzi, K.; Karjalainen, A.; Korpela, A.; Tuuva, T.; Besancon, M.; Choudhury, S.; Dejardin, M.; Denegri, D.; Fabbro, B.; Faure, J. L.; Ferri, F.; Ganjour, S.; Givernaud, A.; Gras, P.; Hamel de Monchenault, G.; Jarry, P.; Locci, E.; Malcles, J.; Millischer, L.; Nayak, A.; Rander, J.; Rosowsky, A.; Titov, M.; Baffioni, S.; Beaudette, F.; Benhabib, L.; Bianchini, L.; Bluj, M.; Busson, P.; Charlot, C.; Daci, N.; Dahms, T.; Dalchenko, M.; Dobrzynski, L.; Florent, A.; Granier de Cassagnac, R.; Haguenauer, M.; Miné, P.; Mironov, C.; Naranjo, I. N.; Nguyen, M.; Ochando, C.; Paganini, P.; Sabes, D.; Salerno, R.; Sirois, Y.; Veelken, C.; Zabi, A.; Agram, J.-L.; Andrea, J.; Bloch, D.; Bodin, D.; Brom, J.-M.; Cardaci, M.; Chabert, E. C.; Collard, C.; Conte, E.; Drouhin, F.; Fontaine, J.-C.; Gelé, D.; Goerlach, U.; Juillot, P.; Le Bihan, A.-C.; Van Hove, P.; Fassi, F.; Mercier, D.; Beauceron, S.; Beaupere, N.; Bondu, O.; Boudoul, G.; Brochet, S.; Chasserat, J.; Chierici, R.; Contardo, D.; Depasse, P.; El Mamouni, H.; Fay, J.; Gascon, S.; Gouzevitch, M.; Ille, B.; Kurca, T.; Lethuillier, M.; Mirabito, L.; Perries, S.; Sgandurra, L.; Sordini, V.; Tschudi, Y.; Verdier, P.; Viret, S.; Tsamalaidze, Z.; Autermann, C.; Beranek, S.; Calpas, B.; Edelhoff, M.; Feld, L.; Heracleous, N.; Hindrichs, O.; Jussen, R.; Klein, K.; Merz, J.

    2013-09-01

    The results of searches for supersymmetry by the CMS experiment are interpreted in the framework of simplified models. The results are based on data corresponding to an integrated luminosity of 4.73 to 4.98fb-1. The data were collected at the LHC in proton-proton collisions at a center-of-mass energy of 7 TeV. This paper describes the method of interpretation and provides upper limits on the product of the production cross section and branching fraction as a function of new particle masses for a number of simplified models. These limits and the corresponding experimental acceptance calculations can be used to constrain other theoretical models and to compare different supersymmetry-inspired analyses.

  18. A simplified model of aerosol removal by containment sprays

    SciTech Connect

    Powers, D.A. ); Burson, S.B. . Div. of Safety Issue Resolution)

    1993-06-01

    Spray systems in nuclear reactor containments are described. The scrubbing of aerosols from containment atmospheres by spray droplets is discussed. Uncertainties are identified in the prediction of spray performance when the sprays are used as a means for decontaminating containment atmospheres. A mechanistic model based on current knowledge of the physical phenomena involved in spray performance is developed. With this model, a quantitative uncertainty analysis of spray performance is conducted using a Monte Carlo method to sample 20 uncertain quantities related to phenomena of spray droplet behavior as well as the initial and boundary conditions expected to be associated with severe reactor accidents. Results of the uncertainty analysis are used to construct simplified expressions for spray decontamination coefficients. Two variables that affect aerosol capture by water droplets are not treated as uncertain; they are (1) [open quote]Q[close quote], spray water flux into the containment, and (2) [open quote]H[close quote], the total fall distance of spray droplets. The choice of values of these variables is left to the user since they are plant and accident specific. Also, they can usually be ascertained with some degree of certainty. The spray decontamination coefficients are found to be sufficiently dependent on the extent of decontamination that the fraction of the initial aerosol remaining in the atmosphere, m[sub f], is explicitly treated in the simplified expressions. The simplified expressions for the spray decontamination coefficient are given. Parametric values for these expressions are found for median, 10 percentile, and 90 percentile values in the uncertainty distribution for the spray decontamination coefficient. Examples are given to illustrate the utility of the simplified expressions to predict spray decontamination of an aerosol-laden atmosphere.

  19. Adjuvant properties of a simplified C32 monomycolyl glycerol analogue.

    PubMed

    Bhowruth, Veemal; Minnikin, David E; Agger, Else Marie; Andersen, Peter; Bramwell, Vincent W; Perrie, Yvonne; Besra, Gurdyal S

    2009-04-01

    A simplified C(32) monomycolyl glycerol (MMG) analogue demonstrated enhanced immunostimulatory activity in a dioctadecyl ammonium bromide (DDA)/Ag85B-ESAT-6 formulation. Elevated levels of IFN-gamma and IL-6 were produced in spleen cells from mice immunised with a C(32) MMG analogue comparable activity to the potent Th1 adjuvant, trehalose 6,6'-di-behenate (TDB).

  20. Simplified Velocity Interferometer System for Any Reflector (VISAR) system

    SciTech Connect

    Sweatt, W.C.; Stanton, P.L.; Crump, O.B. Jr.

    1990-01-01

    A simplified, rugged VISAR (Velocity Interferometer System for Any Reflector) system has been developed using a non-removable delay element and an essentially non-adjustable interferometer cavity. In this system, the critical interference adjustments are performed during fabrication of the cavity, freeing the user from this task. Prototype systems are easy to use and give extremely high quality results. 6 refs., 7 figs.

  1. Surgical site infection after cardiac surgery: a simplified surveillance method.

    PubMed

    Lucet, Jean-Christophe

    2006-12-01

    We report the results of a 2-year, 7-center program of surveillance of deep sternal wound infection (DSWI) after cardiac surgery. DSWI was defined as the need for reoperation. Stratification data were abstracted from computerized files. The incidence of DSWI was 2.2% (198 of 8,816 cardiac surgery procedures). The risk factors identified were obesity, age, coronary artery bypass grafting, postoperative mechanical ventilation, and early surgical reexploration. The resource efficiency of this simplified surveillance method is discussed.

  2. Velocity profiles in a hot jet by simplified RELIEF

    NASA Technical Reports Server (NTRS)

    1991-01-01

    Raman Excitation + Laser Induced Electronic Fluorescence (RELIEF) is a double resonance velocimetry technique in which oxygen molecules are vibrationally excited via stimulated Raman scattering at a specific location within a flow field. After suitable time delay, typically 1-10 microseconds, the displacement of the tagged molecules is determined by laser induced fluorescence imaging. Providing support for the installation of simplified RELIEF flow tagging instrumentation at NASA LaRC was the principal goal of this research.

  3. A simplified approach for the computation of steady two-phase flow in inverted siphons.

    PubMed

    Diogo, A Freire; Oliveira, Maria C

    2016-01-15

    Hydraulic, sanitary, and sulfide control conditions of inverted siphons, particularly in large wastewater systems, can be substantially improved by continuous air injection in the base of the inclined rising branch. This paper presents a simplified approach that was developed for the two-phase flow of the rising branch using the energy equation for a steady pipe flow, based on the average fluid fraction, observed slippage between phases, and isothermal assumption. As in a conventional siphon design, open channel steady uniform flow is assumed in inlet and outlet chambers, corresponding to the wastewater hydraulic characteristics in the upstream and downstream sewers, and the descending branch operates in steady uniform single-phase pipe flow. The proposed approach is tested and compared with data obtained in an experimental siphon setup with two plastic barrels of different diameters operating separately as in a single-barrel siphon. Although the formulations developed are very simple, the results show a good adjustment for the set of the parameters used and conditions tested and are promising mainly for sanitary siphons with relatively moderate heights of the ascending branch.

  4. Simplified analytical model of penetration with lateral loading -- User`s guide

    SciTech Connect

    Young, C.W.

    1998-05-01

    The SAMPLL (Simplified Analytical Model of Penetration with Lateral Loading) computer code was originally developed in 1984 to realistically yet economically predict penetrator/target interactions. Since the code`s inception, its use has spread throughout the conventional and nuclear penetrating weapons community. During the penetrator/target interaction, the resistance of the material being penetrated imparts both lateral and axial loads on the penetrator. These loads cause changes to the penetrator`s motion (kinematics). SAMPLL uses empirically based algorithms, formulated from an extensive experimental data base, to replicate the loads the penetrator experiences during penetration. The lateral loads resulting from angle of attack and trajectory angle of the penetrator are explicitly treated in SAMPLL. The loads are summed and the kinematics calculated at each time step. SAMPLL has been continually improved, and the current version, Version 6.0, can handle cratering and spall effects, multiple target layers, penetrator damage/failure, and complex penetrator shapes. Version 6 uses the latest empirical penetration equations, and also automatically adjusts the penetrability index for certain target layers to account for layer thickness and confinement. This report describes the SAMPLL code, including assumptions and limitations, and includes a user`s guide.

  5. Simplified failure sequence evaluation of reactor pressure vessel head corroding in-core instrumentation assembly

    SciTech Connect

    McVicker, J.P.; Conner, J.T.; Hasrouni, P.N.; Reizman, A.

    1995-11-01

    In-Core Instrumentation (ICI) assemblies located on a Reactor Pressure Vessel Head have a history of boric acid leakage. The acid tends to corrode the nuts and studs which fasten the flanges of the assembly, thereby compromising the assembly`s structural integrity. This paper provides a simplified practical approach in determining the likelihood of an undetected progressing assembly stud deterioration, which would lead to a catastrophic loss of reactor coolant. The structural behavior of the In-Core Instrumentation flanged assembly is modeled using an elastic composite section assumption, with the studs transmitting tension and the pressure sealing gasket experiencing compression. Using the above technique, one can calculate the flange relative deflection and the consequential coolant loss flow rate, as well as the stress in any stud. A solved real life example develops the expected failure sequence and discusses the exigency of leak detection for safe shutdown. In the particular case of Calvert Cliffs Nuclear Power Plant (CCNPP) it is concluded that leak detection occurs before catastrophic failure of the ICI flange assembly.

  6. Simplified analysis of frame structures with viscoelastic dampers considering the effect of soil-structure interaction

    NASA Astrophysics Data System (ADS)

    Zhao, Xuefei; Wang, Shuguang; Du, Dongsheng; Liu, Weiqing

    2017-01-01

    In this study, simplified numerical models are developed to analyze the soil-structure interaction (SSI) effect on frame structures equipped with viscoelastic dampers (VEDs) based on pile group foundation. First, a single degree-of-freedom (SDOF) oscillator is successfully utilized to replace the SDOF energy dissipated structure considering the SSI effect. The equivalent period and damping ratio of the system are obtained through analogical analysis using the frequency transfer function with adoption of the modal strain energy (MSE) technique. A parametric analysis is carried out to study the SSI effect on the performance of VEDs. Then the equilibrium equations of the multi degree-of-freedom (MDOF) structure with VEDs considering SSI effect are established in the frequency domain. Based on the assumption that the superstructure of the coupled system possesses the classical normal mode, the MDOF superstructure is decoupled to a set of individual SDOF systems resting on a rigid foundation with adoption of the MSE technique through formula derivation. Numerical results demonstrate that the proposed methods have the advantage of reducing computational cost, however, retaining the satisfactory accuracy. The numerical method proposed herein can provide a fast evaluation of the efficiency of VEDs considering the SSI effect.

  7. Simplified tornado depressurization design methods for nuclear power plants

    SciTech Connect

    Howard, N.M.; Krasnopoler, M.I.

    1983-05-01

    A simplified approach for the calculation of tornado depressurization effects on nuclear power plant structures and components is based on a generic computer depressurization analysis for an arbitrary single volume V connected to the atmosphere by an effective vent area A. For a given tornado depressurization transient, the maximum depressurization ..delta..P of the volume was found to depend on the parameter V/A. The relation between ..delta..P and V/A can be represented by a single monotonically increasing curve for each of the three design-basis tornadoes described in the U.S. Nuclear Regulatory Commission's Regulatory Guide 1.76. These curves can be applied to most multiple-volume nuclear power plant structures by considering each volume and its controlling vent area. Where several possible flow areas could be controlling, the maximum value of V/A can be used to estimate a conservative value for ..delta..P. This simplified approach was shown to yield reasonably conservative results when compared to detailed computer calculations of moderately complex geometries. Treatment of severely complicated geometries, heating and ventilation systems, and multiple blowout panel arrangements were found to be beyond the limitations of the simplified analysis.

  8. Phase transitions in simplified models with long-range interactions

    NASA Astrophysics Data System (ADS)

    Rocha Filho, T. M.; Amato, M. A.; Mello, B. A.; Figueiredo, A.

    2011-10-01

    We study the origin of phase transitions in several simplified models with long-range interactions. For the self-gravitating ring model, we are unable to observe a possible phase transition predicted by Nardini and Casetti [Phys. Rev. EPLEEE81539-375510.1103/PhysRevE.80.060103 80, 060103R (2009).] from an energy landscape analysis. Instead we observe a sharp, although without any nonanalyticity, change from a core-halo to a core-only configuration in the spatial distribution functions for low energies. By introducing a different class of solvable simplified models without any critical points in the potential energy we show that a behavior similar to the thermodynamics of the ring model is obtained, with a first-order phase transition from an almost homogeneous high-energy phase to a clustered phase and the same core-halo to core configuration transition at lower energies. We discuss the origin of these features for the simplified models and show that the first-order phase transition comes from the maximization of the entropy of the system as a function of energy and an order parameter, as previously discussed by Hahn and Kastner [Phys. Rev. EPLEEE81539-375510.1103/PhysRevE.72.056134 72, 056134 (2005); Eur. Phys. J. BEPJBFY1434-602810.1140/epjb/e2006-00100-7 50, 311 (2006)], which seems to be the main mechanism causing phase transitions in long-range interacting systems.

  9. The current theoretical assumptions of the Bobath concept as determined by the members of BBTA.

    PubMed

    Raine, Sue

    2007-01-01

    The Bobath concept is a problem-solving approach to the assessment and treatment of individuals following a lesion of the central nervous system that offers therapists a framework for their clinical practice. The aim of this study was to facilitate a group of experts in determining the current theoretical assumptions underpinning the Bobath concept.A four-round Delphi study was used. The expert sample included all 15 members of the British Bobath Tutors Association. Initial statements were identified from the literature with respondents generating additional statements. Level of agreement was determined by using a five-point Likert scale. Level of consensus was set at 80%. Eighty-five statements were rated from the literature along with 115 generated by the group. Ninety-three statements were identified as representing the theoretical underpinning of the Bobath concept. The Bobath experts agreed that therapists need to be aware of the principles of motor learning such as active participation, opportunities for practice and meaningful goals. They emphasized that therapy is an interactive process between individual, therapist, and the environment and aims to promote efficiency of movement to the individual's maximum potential rather than normal movement. Treatment was identified by the experts as having "change of functional outcome" at its center.

  10. Testing of some assumptions about biodegradability in soil as measured by carbon dioxide evolution

    SciTech Connect

    El-Din Sharabi, N.; Bartha, R. )

    1993-04-01

    The Toxic Substance Control Act calls for a premanufacturing review of novel chemical substances including their biodegradability. Carbon dioxide evolution, using non-labeled carbon or [sup 14]C, is a common method of testing. This study examines assumptions of carbon dioxide evolution testing. Test substances used included: glucose, adipic acid, benzoic acid, and n-hexadecane. Chemical composition other than carbon content appears to influence minimally the percentages conversion to CO[sub 2]. However, that although CO[sub 2] evolution seemed proportional to the carbon content and concentration of the test substance, at least one-half of the evolved net CO[sub 2] did not come directly from the test substance. Conversion to CO[sub 2] in the soil appeared to depend on carbon content only. In experiments of 1 month or longer, the net CO[sub 2] evolution in response to substrate may be above 100% of the added substrate carbon. Whether this applies to all substrate additions remains to be studied. The authors conclude that net CO[sub 2] and [sup 14]CO[sub 2] evolution measurements are useful as a first-tier tests for assessing biodegradability in soil. 11 refs., 6 figs.

  11. Relaxing the closure assumption in single-season occupancy models: staggered arrival and departure times

    USGS Publications Warehouse

    Kendall, William L.; Hines, James E.; Nichols, James D.; Grant, Evan H. Campbell

    2013-01-01

    Occupancy statistical models that account for imperfect detection have proved very useful in several areas of ecology, including species distribution and spatial dynamics, disease ecology, and ecological responses to climate change. These models are based on the collection of multiple samples at each of a number of sites within a given season, during which it is assumed the species is either absent or present and available for detection while each sample is taken. However, for some species, individuals are only present or available for detection seasonally. We present a statistical model that relaxes the closure assumption within a season by permitting staggered entry and exit times for the species of interest at each site. Based on simulation, our open model eliminates bias in occupancy estimators and in some cases increases precision. The power to detect the violation of closure is high if detection probability is reasonably high. In addition to providing more robust estimation of occupancy, this model permits comparison of phenology across sites, species, or years, by modeling variation in arrival or departure probabilities. In a comparison of four species of amphibians in Maryland we found that two toad species arrived at breeding sites later in the season than a salamander and frog species, and departed from sites earlier.

  12. Appropriate technology for planning hydroelectric power projects in Nepal: the need for assumption analysis

    SciTech Connect

    Chandler, C.G.

    1981-06-01

    The study focuses on the project development process for hydroelectric project planning in Nepal. Chapter I describes the contrast between the vast potential for hydroelectric power development in Nepal and the current energy shortage within the country, not only for electricity, but for firewood and other fuel sources as well. Chapter II explores some of the unknown factors facing hydropower project planners in Nepal, where data for hydrologic, geologic, environmental, and sociological project components are lacking. The chapter also examines institutional and fiscal factors which constrain the planning process. Chapter III describes the critical role of assumptions in the project development process, and details the stages that a project goes through as it is planned. The chapter introduces the concept of assumption analysis as a technique for project planning, listing the potential conflict between the assumptions of foreign consultants and the host-country users of project outputs as an ingredient in the project's success or failure. Chapter IV demonstrates the mechanics and usefulness of assumption analysis through an Assumption Analysis Chart, which shows the interaction among project objectives, project alternatives, project assumptions, and the project development process. Assumption analysis techniques are expected to be useful among bilateral and multilateral aid donors servicing less developed countries.

  13. The effect of simplified adhesives on the bond strength to dentin of dual-cure resin cements.

    PubMed

    Shade, A M; Wajdowicz, M N; Bailey, C W; Vandewalle, K S

    2014-01-01

    The purpose of this study was to compare the shear bond strengths to dentin of two dual-cure resin cements, one with a unique initiator, NX3 (Kerr Corp), and the other with a traditional redox-initiator system, Calibra (Dentsply), when used in combination with simplified or nonsimplified adhesive agents. The two dual-cure resin cements, in either self- or dual-cure activation modes, were bonded to human dentin with four dental adhesives to create 16 subgroups of 10 specimens each. After 24 hours of storage in distilled water at 37°C, the specimens were tested in shear in a universal testing machine. With both NX3 and Calibra, bond strengths significantly increased when the specimens were dual cured. In addition, with either cement in either mode, the nonsimplified adhesives performed significantly better than did the simplified adhesive bonding agents. When used specifically with simplified adhesives in either cure mode, NX3 did not produce significantly higher bond strengths than did Calibra. In general, lower dentin bond strengths were found with simplified adhesives or self-cure activation with either resin cement.

  14. An Exploration of Dental Students' Assumptions About Community-Based Clinical Experiences.

    PubMed

    Major, Nicole; McQuistan, Michelle R

    2016-03-01

    The aim of this study was to ascertain which assumptions dental students recalled feeling prior to beginning community-based clinical experiences and whether those assumptions were fulfilled or challenged. All fourth-year students at the University of Iowa College of Dentistry & Dental Clinics participate in community-based clinical experiences. At the completion of their rotations, they write a guided reflection paper detailing the assumptions they had prior to beginning their rotations and assessing the accuracy of their assumptions. For this qualitative descriptive study, the 218 papers from three classes (2011-13) were analyzed for common themes. The results showed that the students had a variety of assumptions about their rotations. They were apprehensive about working with challenging patients, performing procedures for which they had minimal experience, and working too slowly. In contrast, they looked forward to improving their clinical and patient management skills and knowledge. Other assumptions involved the site (e.g., the equipment/facility would be outdated; protocols/procedures would be similar to the dental school's). Upon reflection, students reported experiences that both fulfilled and challenged their assumptions. Some continued to feel apprehensive about treating certain patient populations, while others found it easier than anticipated. Students were able to treat multiple patients per day, which led to increased speed and patient management skills. However, some reported challenges with time management. Similarly, students were surprised to discover some clinics were new/updated although some had limited instruments and materials. Based on this study's findings about students' recalled assumptions and reflective experiences, educators should consider assessing and addressing their students' assumptions prior to beginning community-based dental education experiences.

  15. A simplified analytical solution for thermal response of a one-dimensional, steady state transpiration cooling system in radiative and convective environment

    NASA Technical Reports Server (NTRS)

    Kubota, H.

    1976-01-01

    A simplified analytical method for calculation of thermal response within a transpiration-cooled porous heat shield material in an intense radiative-convective heating environment is presented. The essential assumptions of the radiative and convective transfer processes in the heat shield matrix are the two-temperature approximation and the specified radiative-convective heatings of the front surface. Sample calculations for porous silica with CO2 injection are presented for some typical parameters of mass injection rate, porosity, and material thickness. The effect of these parameters on the cooling system is discussed.

  16. Troubling 'lived experience': a post-structural critique of mental health nursing qualitative research assumptions.

    PubMed

    Grant, A

    2014-08-01

    Qualitative studies in mental health nursing research deploying the 'lived experience' construct are often written on the basis of conventional qualitative inquiry assumptions. These include the presentation of the 'authentic voice' of research participants, related to their 'lived experience' and underpinned by a meta-assumption of the 'metaphysics of presence'. This set of assumptions is critiqued on the basis of contemporary post-structural qualitative scholarship. Implications for mental health nursing qualitative research emerging from this critique are described in relation to illustrative published work, and some benefits and challenges for researchers embracing post-structural sensibilities are outlined.

  17. The crux of the method: assumptions in ordinary least squares and logistic regression.

    PubMed

    Long, Rebecca G

    2008-10-01

    Logistic regression has increasingly become the tool of choice when analyzing data with a binary dependent variable. While resources relating to the technique are widely available, clear discussions of why logistic regression should be used in place of ordinary least squares regression are difficult to find. The current paper compares and contrasts the assumptions of ordinary least squares with those of logistic regression and explains why logistic regression's looser assumptions make it adept at handling violations of the more important assumptions in ordinary least squares.

  18. The Intensity-dependence of Tropical Cyclone Intensification in a Simplified Energetic System and a Full Physics Model

    NASA Astrophysics Data System (ADS)

    Wang, Y.

    2015-12-01

    Several recent studies based on best track data for tropical cyclones (TCs) over the North Atlantic revealed a strong dependence of the TC intensification rate (IR) on TC intensity. The TC IR shows a maximum (i.e., most rapid intensification) when the TC intensity reaches about 35 - 40 m s-1. Physically, this intensity-dependence has been explained in terms of the inner-core inertial stability and the decreased potential for a TC to intensify as it approaches its maximum potential intensity (MPI). Here, we propose a new simplified dynamical system based on TC energetics. For this purpose, the TC system is considered as a Carnot heat engine as in earlier theoretical studies, and formulated on the grounds of major energy production and dissipation processes. This offers a way to explain the IR behavior of TCs in terms of its sensitivity to the physical processes involved, and to derive a physically-based relationship between TC intensification and intensity. This relationship is consistent with observations, and in quantitative agreement with that gained from empirical relations in an alternative dynamical system proposed by DeMaria(2009), which has been formulated analogously to the logistic growth equation (LGE). Results from idealized full-physics model simulations confirm the validity of several key assumptions implicitly involved in the simplified dynamical system and the intensity-dependence of the TC IR as well.

  19. Validations and psychological properties of a simplified Chinese version of pain anxiety symptoms scale (SC-PASS).

    PubMed

    Zhou, Xiao-Yi; Xu, Xi-Ming; Wang, Fei; Wu, Sui-Yi; Yang, Yi-Lin; Li, Ming; Huang, Jian-Ming; Wei, Xian-Zhao

    2017-03-01

    The Pain Anxiety Symptoms Scale (PASS) has been developed to evaluate pain anxiety, which leads to avoidance of daily activities and normal movements. However, a simplified Chinese version of PASS is still not available. Physicians are not aware of which patients are prone to anxiety, and what the risk factors are.To cross-culturally adapt the PASS into a simplified Chinese version and test the reliability and validity. Factors affecting pain anxiety were also explored.The PASS was first translated into a simplified Chinese version according to a forward-backward method. Then, validations were tested including content validity, construct validity, and reliability. Content validity was analyzed by response trend. Construct validity was analyzed by confirmatory factor analysis (CFA), exploratory factor analysis, and priori hypotheses testing. Reliability was analyzed by internal consistency and test-retest reliability. Risk factors of catastrophizing were analyzed by performing multivariate liner regression.A total of 219 patients were included in the study. The scores of items were well distributed. Both CFA and exploratory factor analysis suggested a 2nd-order, 4-factor model, accounting for 65.42% of the total variance according to principle component analysis. SC-PASS obtained good reliability with a Cronbach α = 0.92 and ICC = 0.90. College education, long pain duration, and both married and divorced status were risk factors. Factors reduced pain-related anxiety were no medication assumption, female sex, widowed status, non-Han ethnicity, and having no religious belief.The SC-PASS was applicable in Chinese patients and it was suitable for the clinical uses in mainland China.

  20. Validations and psychological properties of a simplified Chinese version of pain anxiety symptoms scale (SC-PASS)

    PubMed Central

    Zhou, Xiao-Yi; Xu, Xi-Ming; Wang, Fei; Wu, Sui-Yi; Yang, Yi-Lin; Li, Ming; Huang, Jian-Ming; Wei, Xian-Zhao

    2017-01-01

    Abstract The Pain Anxiety Symptoms Scale (PASS) has been developed to evaluate pain anxiety, which leads to avoidance of daily activities and normal movements. However, a simplified Chinese version of PASS is still not available. Physicians are not aware of which patients are prone to anxiety, and what the risk factors are. To cross-culturally adapt the PASS into a simplified Chinese version and test the reliability and validity. Factors affecting pain anxiety were also explored. The PASS was first translated into a simplified Chinese version according to a forward-backward method. Then, validations were tested including content validity, construct validity, and reliability. Content validity was analyzed by response trend. Construct validity was analyzed by confirmatory factor analysis (CFA), exploratory factor analysis, and priori hypotheses testing. Reliability was analyzed by internal consistency and test–retest reliability. Risk factors of catastrophizing were analyzed by performing multivariate liner regression. A total of 219 patients were included in the study. The scores of items were well distributed. Both CFA and exploratory factor analysis suggested a 2nd-order, 4-factor model, accounting for 65.42% of the total variance according to principle component analysis. SC-PASS obtained good reliability with a Cronbach α = 0.92 and ICC = 0.90. College education, long pain duration, and both married and divorced status were risk factors. Factors reduced pain-related anxiety were no medication assumption, female sex, widowed status, non-Han ethnicity, and having no religious belief. The SC-PASS was applicable in Chinese patients and it was suitable for the clinical uses in mainland China. PMID:28272194

  1. 76 FR 14437 - Economic Simplified Boiling Water Reactor Standard Design: GE Hitachi Nuclear Energy; Issuance of...

    Federal Register 2010, 2011, 2012, 2013, 2014

    2011-03-16

    ... From the Federal Register Online via the Government Publishing Office ] NUCLEAR REGULATORY COMMISSION Economic Simplified Boiling Water Reactor Standard Design: GE Hitachi Nuclear Energy; Issuance of... GE Hitachi Nuclear Energy (GEH) for the economic simplified boiling water reactor (ESBWR)...

  2. Health Information in Chinese - Simplified (简体中文): MedlinePlus

    MedlinePlus

    ... Simplified) Bilingual PDF Health Information Translations Cancer Treatment Side Effects 癌症治疗的副作用 - 简体中文 (Chinese - Simplified) Bilingual PDF Health Information Translations Changes in Taste and Smell ...

  3. Evaluating the random representation assumption of lexical semantics in cognitive models.

    PubMed

    Johns, Brendan T; Jones, Michael N

    2010-10-01

    A common assumption implicit in cognitive models is that lexical semantics can be approximated by using randomly generated representations to stand in for word meaning. However, the use of random representations contains the hidden assumption that semantic similarity is symmetrically distributed across randomly selected words or between instances within a semantic category. We evaluated this assumption by computing similarity distributions for randomly selected words from a number of well-known semantic measures and comparing them with the distributions from random representations commonly used in cognitive models. The similarity distributions from all semantic measures were positively skewed compared with the symmetric normal distributions assumed by random representations. We discuss potential consequences that this false assumption may have for conclusions drawn from process models that use random representations.

  4. Learning disabilities theory and Soviet psychology: a comparison of basic assumptions.

    PubMed

    Coles, G S

    1982-09-01

    Critics both within and outside the Learning Disabilities (LD) field have pointed to the weaknesses of LD theory. Beginning with the premise that a significant problem of LD theory has been its failure to explore fully its fundamental assumptions, this paper examines a number of these assumptions about individual and social development, cognition, and learning. These assumptions are compared with a contrasting body of premises found in Soviet psychology, particularly in the works of Vygotsky, Leontiev, and Luria. An examination of the basic assumptions of LD theory and Soviet psychology shows that a major difference lies in their respective nondialectical and dialectical interpretation of the relationship of social factors and cognition, learning, and neurological development.

  5. Washington International Renewable Energy Conference (WIREC) 2008 Pledges. Methodology and Assumptions Summary

    SciTech Connect

    Babiuch, Bill; Bilello, Daniel E.; Cowlin, Shannon C.; Mann, Margaret; Wise, Alison

    2008-08-01

    This report describes the methodology and assumptions used by NREL in quantifying the potential CO2 reductions resulting from more than 140 governments, international organizations, and private-sector representatives pledging to advance the uptake of renewable energy.

  6. Model complexity and performance: How far can we simplify?

    NASA Astrophysics Data System (ADS)

    Raick, C.; Soetaert, K.; Grégoire, M.

    2006-07-01

    Handling model complexity and reliability is a key area of research today. While complex models containing sufficient detail have become possible due to increased computing power, they often lead to too much uncertainty. On the other hand, very simple models often crudely oversimplify the real ecosystem and can not be used for management purposes. Starting from a complex and validated 1D pelagic ecosystem model of the Ligurian Sea (NW Mediterranean Sea), we derived simplified aggregated models in which either the unbalanced algal growth, the functional group diversity or the explicit description of the microbial loop was sacrificed. To overcome the problem of data availability with adequate spatial and temporal resolution, the outputs of the complex model are used as the baseline of perfect knowledge to calibrate the simplified models. Objective criteria of model performance were used to compare the simplified models’ results to the complex model output and to the available data at the DYFAMED station in the central Ligurian Sea. We show that even the simplest (NPZD) model is able to represent the global ecosystem features described by the complex model (e.g. primary and secondary productions, particulate organic matter export flux, etc.). However, a certain degree of sophistication in the formulation of some biogeochemical processes is required to produce realistic behaviors (e.g. the phytoplankton competition, the potential carbon or nitrogen limitation of the zooplankton ingestion, the model trophic closure, etc.). In general, a 9 state-variable model that has the functional group diversity removed, but which retains the bacterial loop and the unbalanced algal growth, performs best.

  7. 76 FR 82315 - Agency Information Collection Activities: Entry/Immediate Delivery Application and Simplified Entry

    Federal Register 2010, 2011, 2012, 2013, 2014

    2011-12-30

    ... Delivery Application and Simplified Entry AGENCY: U.S. Customs and Border Protection, Department of... Delivery Application (Forms 3461 and 3461 ALT) and Simplified Entry. This is a proposed revision and... forms of information. Title: Entry/Immediate Delivery Application and Simplified Entry. OMB Number:...

  8. Comparison of Performance Effectiveness of Linear Control Algorithms Developed for a Simplified Ground Vehicle Suspension System

    DTIC Science & Technology

    2011-04-01

    Comparison of Performance Effectiveness of Linear Control Algorithms Developed for a Simplified Ground Vehicle Suspension System by Ross... Linear Control Algorithms Developed for a Simplified Ground Vehicle Suspension System Ross Brown Motile Robotics, Inc, research contractor at U.S... Linear Control Algorithms Developed for a Simplified Ground Vehicle Suspension System 5a. CONTRACT NUMBER 5b. GRANT NUMBER 5c. PROGRAM ELEMENT

  9. Legal and ethical implications of health care provider insurance risk assumption.

    PubMed

    Cox, Thomas

    2010-01-01

    From bedside to boardroom, nurses deal with the consequences of health care provider insurance risk assumption. Professional caregiver insurance risk refers to insurance risks assumed through contracts with third parties, federal and state Medicare and Medicaid program mandates, and the diagnosis-related groups and Prospective Payment Systems. This article analyzes the financial, legal, and ethical implications of provider insurance risk assumption by focusing on the degree to which patient benefits are reduced.

  10. NGNP: High Temperature Gas-Cooled Reactor Key Definitions, Plant Capabilities, and Assumptions

    SciTech Connect

    Wayne Moe

    2013-05-01

    This document provides key definitions, plant capabilities, and inputs and assumptions related to the Next Generation Nuclear Plant to be used in ongoing efforts related to the licensing and deployment of a high temperature gas-cooled reactor. These definitions, capabilities, and assumptions were extracted from a number of NGNP Project sources such as licensing related white papers, previously issued requirement documents, and preapplication interactions with the Nuclear Regulatory Commission (NRC).

  11. A statistical analysis of the dependency of closure assumptions in cumulus parameterization on the horizontal resolution

    SciTech Connect

    Xu, K.M.

    1994-12-01

    Simulated data from the UCLA cumulus ensemble model are used to investigate the quasi-universal validity of closure assumptions used in existing cumulus parameterizations. A closure assumption is quasi-universally valid if it is sensitive neither to convective cloud regimes nor to horizontal resolutions of large-scale/mesoscale models. The dependency of three types of closure assumptions, as classified by Arakawa and Chen, on the horizontal resolution is addressed in this study. Type I is the constraint on the coupling of the time tendencies of large-scale temperature and water vapor mixing ratio. Type II is the constraint on the coupling of cumulus heating and cumulus drying. Type III is a direct constraint on the intensity of a cumulus ensemble. The macroscopic behavior of simulated cumulus convection is first compared with the observed behavior in view of Type I and Type II closure assumptions using `quick-look` and canonical correlation analyses. It is found that they are statistically similar to each other. The three types of closure assumptions are further examined with simulated data averaged over selected subdomain sizes ranging from 64 to 512 km. It is found that the dependency of Type I and Type II closure assumptions on the horizontal resolution is very weak and that Type III closure assumption is somewhat dependent upon the horizontal resolution. The influences of convective and mesoscale processes on the closure assumptions are also addressed by comparing the structures of canonical components with the corresponding vertical profiles in the convective and stratiform regions of cumulus ensembles analyzed directly from simulated data. The implication of these results for cumulus parameterization is discussed.

  12. Crack in the Foundation Defense Transformation and the Underlying Assumption of Dominant Knowledge in Future War

    DTIC Science & Technology

    2003-04-07

    technologies seemed to regard the lack of evidence for their claims as an indicator of their powers of imagination rather than a deficiency. Millenialism ...tomorrow’s military force rests on the unfounded assumption that technologies emerging from the “information revolution” will lift the fog of war and permit...conventional war. The intellectual foundation for building tomorrow’s military force rests on the unfounded assumption that technologies emerging from

  13. The occlusal guard: a simplified technique for fabrication and equilibration.

    PubMed

    Antonelli, John; Hottel, Timothy L; Siegel, Sharon C; Brandt, Robert; Silva, Gladston

    2013-01-01

    Hard occlusal guards have been used effectively to treat myofacial pain originating from parafunctional activities. Also, they can protect the natural dentition when it opposes porcelain restorations, help to evaluate changes in occlusal vertical dimension during full mouth rehabilitation, minimize further tooth loss in patients with abfraction lesions, and redirect occlusal loads more favorably onto dental implant-supported prostheses. A simplified technique is described to fabricate a properly designed wax model of an occlusal guard that can be processed in acrylic in the same manner used to construct a complete denture.

  14. A simplified analytic form for generation of axisymmetric plasma boundaries

    NASA Astrophysics Data System (ADS)

    Luce, T. C.

    2017-04-01

    An improved method has been formulated for generating analytic boundary shapes as input for axisymmetric MHD equilibria. This method uses the family of superellipses as the basis function, as previously introduced. The improvements are a simplified notation, reduction of the number of simultaneous nonlinear equations to be solved, and the realization that not all combinations of input parameters admit a solution to the nonlinear constraint equations. The method tests for the existence of a self-consistent solution and, when no solution exists, it uses a deterministic method to find a nearby solution. Examples of generation of boundaries, including tests with an equilibrium solver, are given.

  15. Simplified quantum bit commitment using single photon nonlocality

    NASA Astrophysics Data System (ADS)

    He, Guang Ping

    2014-10-01

    We simplified our previously proposed quantum bit commitment (QBC) protocol based on the Mach-Zehnder interferometer, by replacing symmetric beam splitters with asymmetric ones. It eliminates the need for random sending time of the photons; thus, the feasibility and efficiency are both improved. The protocol is immune to the cheating strategy in the Mayers-Lo-Chau no-go theorem of unconditionally secure QBC, because the density matrices of the committed states do not satisfy a crucial condition on which the no-go theorem holds.

  16. A simplified estimate of seismic soil strain for lifeline systems

    SciTech Connect

    Kamiyama, Makoto; O`Rourke, M.J.

    1995-12-31

    A simplified method for estimating seismic soil strains is presented. It is based on a semi-empirical attenuation law for peak ground velocity as well as on the theory of one dimensional wave propagation. The method requires information on the earthquake magnitude, epicentral distance, focal depth and the N-value distribution for the soil profile in question. The soil strains predicted by the method are compared with observed soil strains from three observation sites in Japan. It is shown that the predicted strains agree reasonably well with observed values. In areas away from ground failure such as liquefaction, the seismic behavior of buried pipelines is closely related to soil strains.

  17. Technology Solutions Case Study: Combustion Safety Simplified Test Protocol

    SciTech Connect

    L. Brand, D. Cautley, D. Bohac, P. Francisco, L. Shen, and S. Gloss

    2015-12-01

    Combustions safety is an important step in the process of upgrading homes for energy efficiency. There are several approaches used by field practitioners, but researchers have indicated that the test procedures in use are complex to implement and provide too many false positives. Field failures often mean that the house is not upgraded until after remediation or not at all, if not include in the program. In this report the PARR and NorthernSTAR DOE Building America Teams provide a simplified test procedure that is easier to implement and should produce fewer false positives.

  18. Parameter adaptation in a simplified pulse-coupled neural network

    NASA Astrophysics Data System (ADS)

    Szekely, Geza; Lindblad, Thomas

    1999-03-01

    In a general purpose pulse coupled neural network (PCNN) algorithm the following parameters are used: 2 weight matrices, 3 time constants, 3 normalization factors and 2 further parameters. In a given application, one has to determine the near optimal parameter set to achieve the desired goal. Here a simplified PCNN is described which contains a parameter fitting part, in the least squares sense. Given input and a desired output image, the program is able to determine the optimal value of a selected PCNN parameter. This method can be extended to more general PCNN algorithms, because partial derivatives are not required for the fitting. Only the sum of squares of the differences is used.

  19. A simplified approach to calibrating [sup 14]C dates

    SciTech Connect

    Talma, A.S.; Vogel, J.C. )

    1993-01-01

    The authors propose a simplified approach to the calibration of radiocarbon dates. They use splines through the tree-ring data as calibration curves, thereby eliminating a large part of the statistical scatter of the actual data points. To express the age range, they transform the [plus minus]1 [sigma] and [plus minus]2 [sigma] values of the BP age to calendar dates and interpret them as the 68% and 95% confidence intervals. This approach by-passes the conceptual problems of the transfer of individual probability values from the radiocarbon to the calendar age. They have adapted software to make this calibration possible.

  20. Simplified design procedures for fiber composite structural components/joints

    NASA Technical Reports Server (NTRS)

    Murthy, P. L. N.; Chamis, Christos C.

    1990-01-01

    Simplified step-by-step design procedures are summarized, which are suitable for the preliminary design of composite structural components such as panels (laminates) and composite built-up structures (box beams). Similar procedures are also summarized for the preliminary design of composite bolted and adhesively bonded joints. The summary is presented in terms of sample design cases complemented with typical results. Guidelines are provided which can be used in the design selection process of composite structural components/joints. Also, procedures to account for cyclic loads, hygrothermal effects and lamination residual stresses are included.