Finite elements and finite differences for transonic flow calculations
NASA Technical Reports Server (NTRS)
Hafez, M. M.; Murman, E. M.; Wellford, L. C.
1978-01-01
The paper reviews the chief finite difference and finite element techniques used for numerical solution of nonlinear mixed elliptic-hyperbolic equations governing transonic flow. The forms of the governing equations for unsteady two-dimensional transonic flow considered are the Euler equation, the full potential equation in both conservative and nonconservative form, the transonic small-disturbance equation in both conservative and nonconservative form, and the hodograph equations for the small-disturbance case and the full-potential case. Finite difference methods considered include time-dependent methods, relaxation methods, semidirect methods, and hybrid methods. Finite element methods include finite element Lax-Wendroff schemes, implicit Galerkin method, mixed variational principles, dual iterative procedures, optimal control methods and least squares.
Research and development of LANDSAT-based crop inventory techniques
NASA Technical Reports Server (NTRS)
Horvath, R.; Cicone, R. C.; Malila, W. A. (Principal Investigator)
1982-01-01
A wide spectrum of technology pertaining to the inventory of crops using LANDSAT without in situ training data is addressed. Methods considered include Bayesian based through-the-season methods, estimation technology based on analytical profile fitting methods, and expert-based computer aided methods. Although the research was conducted using U.S. data, the adaptation of the technology to the Southern Hemisphere, especially Argentina was considered.
Preinvestigation evaluation of corrective measure technologies for the Badger Army Ammunition Plant
DOE Office of Scientific and Technical Information (OSTI.GOV)
Benioff, P.A.; Tsai, S.Y.
1989-02-01
This report briefly describes and evaluates the suitability of corrective measure technologies for possible use at the solid waste management units (SWMUs) at the Badger Army Ammunition Plant (BAAP), near Baraboo, Wisconsin. Corrective measure technologies considered for contaminated soils include excavation plus on- or off-site disposal in landfills or by incineration, use of solidification or stabilization methods, and in-situ methods such as bioreclamation and chemical or physical methods. Technologies considered for treatment of contaminated groundwater include groundwater pumping followed by discharge or treatment by air stripping and use of subsurface barriers. 5 refs., 1 tab.
An empirical investigation of methods for nonsymmetric linear systems
NASA Technical Reports Server (NTRS)
Sherman, A. H.
1981-01-01
The present investigation is concerned with a comparison of methods for solving linear algebraic systems which arise from finite difference discretizations of the elliptic convection-diffusion equation in a planar region Omega with Dirichlet boundary conditions. Such linear systems are typically of the form Ax = b where A is an N x N sparse nonsymmetric matrix. In a discussion of discretizations, it is assumed that a regular rectilinear mesh of width h has been imposed on Omega. The discretizations considered include central differences, upstream differences, and modified upstream differences. Six methods for solving Ax = b are considered. Three variants of Gaussian elimination have been chosen as representatives of state-of-the-art software for direct methods under different assumptions about pivoting. Three iterative methods are also included.
Recent developments of the NESSUS probabilistic structural analysis computer program
NASA Technical Reports Server (NTRS)
Millwater, H.; Wu, Y.-T.; Torng, T.; Thacker, B.; Riha, D.; Leung, C. P.
1992-01-01
The NESSUS probabilistic structural analysis computer program combines state-of-the-art probabilistic algorithms with general purpose structural analysis methods to compute the probabilistic response and the reliability of engineering structures. Uncertainty in loading, material properties, geometry, boundary conditions and initial conditions can be simulated. The structural analysis methods include nonlinear finite element and boundary element methods. Several probabilistic algorithms are available such as the advanced mean value method and the adaptive importance sampling method. The scope of the code has recently been expanded to include probabilistic life and fatigue prediction of structures in terms of component and system reliability and risk analysis of structures considering cost of failure. The code is currently being extended to structural reliability considering progressive crack propagation. Several examples are presented to demonstrate the new capabilities.
43 CFR 426.13 - Excess land appraisals.
Code of Federal Regulations, 2011 CFR
2011-10-01
... cost methods, as applicable. Reclamation will consider nonproject water supply factors as provided in... standard appraisal procedures. (c) Appraisals of nonproject water supplies. (1) The appraiser will consider nonproject water supply factors, where appropriate, including: (i) Ground water pumping lift; (ii) Surface...
Responses of a constructed plant community to simulated glyphosate and dicamba drift
Background/Questions/Methods As part of its regulation of pesticides, the US Environmental Protection Agency must consider environmental risks, including impacts to nontarget plants exposed to pesticide drift. Normally these risk assessments consider impacts to individual spec...
Instructional Basics: Oppelt Standard Method of Therapeutic and Recreational Ice Skating.
ERIC Educational Resources Information Center
Oppelt, Kurt
Detailed in the booklet is the standard ice skating method and considered are the benefits of therapeutic ice skating for the handicapped and aged. Values for the mentally retarded and physically handicapped are seen to include physiological (such as increased flexibility and improved posture), psychological (including satifaction and enhanced…
NASA Technical Reports Server (NTRS)
Ehret, R. M.
1974-01-01
The concepts explored in a state of the art review of those engineering fracture mechanics considered most applicable to the space shuttle vehicle include fracture toughness, precritical flaw growth, failure mechanisms, inspection methods (including proof test logic), and crack growth predictive analysis techniques.
Riley, Emily; Harris, Patrick; Kent, Jennifer; Sainsbury, Peter; Lane, Anna; Baum, Fran
2018-01-01
Background: Transport policy and practice impacts health. Environmental Impact Assessments (EIAs) are regulated public policy mechanisms that can be used to consider the health impacts of major transport projects before they are approved. The way health is considered in these environmental assessments (EAs) is not well known. This research asked: How and to what extent was human health considered in EAs of four major transport projects in Australia. Methods: We developed a comprehensive coding framework to analyse the Environmental Impact Statements (EISs) of four transport infrastructure projects: three road and one light rail. The coding framework was designed to capture how health was directly and indirectly included. Results: We found that health was partially considered in all four EISs. In the three New South Wales (NSW) projects, but not the one South Australian project, this was influenced by the requirements issued to proponents by the government which directed the content of the EIS. Health was assessed using human health risk assessment (HHRA). We found this to be narrow in focus and revealed a need for a broader social determinants of health approach, using multiple methods. The road assessments emphasised air quality and noise risks, concluding these were minimal or predicted to improve. The South Australian project was the only road project not to include health data explicitly. The light rail EIS considered the health benefits of the project whereas the others focused on risk. Only one project considered mental health, although in less detail than air quality or noise. Conclusion: Our findings suggest EIAs lag behind the known evidence linking transport infrastructure to health. If health is to be comprehensively included, a more complete model of health is required, as well as a shift away from health risk assessment as the main method used. This needs to be mandatory for all significant developments. We also found that considering health only at the EIA stage may be a significant limitation, and there is a need for health issues to be considered when earlier, fundamental decisions about the project are being made. PMID:29524938
How Dangerous Can Localized Corrosion Be? An Experiment that Studies Its Effects.
ERIC Educational Resources Information Center
Celdran, R.; Gonzalo, P.
1988-01-01
Considers three common cases of localized corrosion of metals: pitting, crevice, and stress corrosion. Provides experimental methods for studying all three methods. Includes a discussion of expected results. (ML)
Evaluation and Education of Children with Brain Damage.
ERIC Educational Resources Information Center
Bortner, Morton, Ed.
Ten papers consider brain damaged children. Brain damage is considered as an educational category, and the following aspects of evaluation are treated: disorders of oral communication, hearing impairment, psychological deficit, psychiatric factors, and neurological considerations. Educational strategies discussed include the educational methods of…
Entropic Lattice Boltzmann Methods
2001-12-10
model of fluid dynamics in one dimension, first considered by Renda et al. in 1997 [14]. Here the geometric picture involves a four dimensional polytope...convention of including constant terms in an extra column of the matrix, using the device of appending 1 to the column vector of unknowns. In general, there...we apply the entropic lattice Boltzmann method to a simple five-velocity model of fluid dynamics in one dimension, first considered by Renda et al
42 CFR 431.424 - Evaluation requirements.
Code of Federal Regulations, 2012 CFR
2012-10-01
... evaluations. Demonstration evaluations will include the following: (1) Quantitative research methods. (i... of appropriate evaluation strategies (including experimental and other quantitative and qualitative... demonstration. (ii) CMS will consider alternative evaluation designs when quantitative designs are technically...
42 CFR 431.424 - Evaluation requirements.
Code of Federal Regulations, 2013 CFR
2013-10-01
... evaluations. Demonstration evaluations will include the following: (1) Quantitative research methods. (i... of appropriate evaluation strategies (including experimental and other quantitative and qualitative... demonstration. (ii) CMS will consider alternative evaluation designs when quantitative designs are technically...
42 CFR 431.424 - Evaluation requirements.
Code of Federal Regulations, 2014 CFR
2014-10-01
... evaluations. Demonstration evaluations will include the following: (1) Quantitative research methods. (i... of appropriate evaluation strategies (including experimental and other quantitative and qualitative... demonstration. (ii) CMS will consider alternative evaluation designs when quantitative designs are technically...
Handicapped Workers Provisions in Union-Management Agreements.
ERIC Educational Resources Information Center
Schwartz, Benjamin
Methods and procedures provided by 1,614 collective bargaining agreements to accommodate handicapped workers in New York State are described. Handicapped veteran clauses are considered; also considered are the prevalence and content of handicapped worker clauses, including transfers provided for in general terms, transfers subject to seniority…
Bayesian Computation for Log-Gaussian Cox Processes: A Comparative Analysis of Methods
Teng, Ming; Nathoo, Farouk S.; Johnson, Timothy D.
2017-01-01
The Log-Gaussian Cox Process is a commonly used model for the analysis of spatial point pattern data. Fitting this model is difficult because of its doubly-stochastic property, i.e., it is an hierarchical combination of a Poisson process at the first level and a Gaussian Process at the second level. Various methods have been proposed to estimate such a process, including traditional likelihood-based approaches as well as Bayesian methods. We focus here on Bayesian methods and several approaches that have been considered for model fitting within this framework, including Hamiltonian Monte Carlo, the Integrated nested Laplace approximation, and Variational Bayes. We consider these approaches and make comparisons with respect to statistical and computational efficiency. These comparisons are made through several simulation studies as well as through two applications, the first examining ecological data and the second involving neuroimaging data. PMID:29200537
ERIC Educational Resources Information Center
Wankat, Phillip C.
1984-01-01
Discusses a simple method for following the movement of a solute in an adsorption or ion exchange system. This movement is used to study a variety of operational methods, including continuous flow and pulsed flow counter-current operations and simulated counter-current systems. Effect of changing thermodynamic variables is also considered. (JM)
Traditional and Modern Cell Culture in Virus Diagnosis.
Hematian, Ali; Sadeghifard, Nourkhoda; Mohebi, Reza; Taherikalani, Morovat; Nasrolahi, Abbas; Amraei, Mansour; Ghafourian, Sobhan
2016-04-01
Cell cultures are developed from tissue samples and then disaggregated by mechanical, chemical, and enzymatic methods to extract cells suitable for isolation of viruses. With the recent advances in technology, cell culture is considered a gold standard for virus isolation. This paper reviews the evolution of cell culture methods and demonstrates why cell culture is a preferred method for identification of viruses. In addition, the advantages and disadvantages of both traditional and modern cell culture methods for diagnosis of each type of virus are discussed. Detection of viruses by the novel cell culture methods is considered more accurate and sensitive. However, there is a need to include some more accurate methods such as molecular methods in cell culture for precise identification of viruses.
Using case studies to teach an engineering technology technical writing class
NASA Technical Reports Server (NTRS)
Green, M. M.
1981-01-01
The use of the case method in teaching various technical communication skills is described. Features of the method considered include: solving communication problems, identifying an audience, planning written communications, presenting written communications, and using visual aids.
Design sensitivity analysis with Applicon IFAD using the adjoint variable method
NASA Technical Reports Server (NTRS)
Frederick, Marjorie C.; Choi, Kyung K.
1984-01-01
A numerical method is presented to implement structural design sensitivity analysis using the versatility and convenience of existing finite element structural analysis program and the theoretical foundation in structural design sensitivity analysis. Conventional design variables, such as thickness and cross-sectional areas, are considered. Structural performance functionals considered include compliance, displacement, and stress. It is shown that calculations can be carried out outside existing finite element codes, using postprocessing data only. That is, design sensitivity analysis software does not have to be imbedded in an existing finite element code. The finite element structural analysis program used in the implementation presented is IFAD. Feasibility of the method is shown through analysis of several problems, including built-up structures. Accurate design sensitivity results are obtained without the uncertainty of numerical accuracy associated with selection of a finite difference perturbation.
Rethinking knowledge and pedagogy in dental education.
Whipp, J L; Ferguson, D J; Wells, L M; Iacopino, A M
2000-12-01
Dentistry as a profession has often been considered both art and science. Traditional dental education has attempted to address both; however, in many places only the science of dentistry is emphasized. The move toward competency-based curricula in dental education requires an expansion of what constitutes meaningful knowledge in the curriculum and what pedagogies best support that curriculum. The scientific and technical knowledge considered foundational to clinical practice is not sufficient to teach competencies associated with the art of dentistry. Habermas, a social scientist, offers a way of looking beyond technical knowledge to consider two other forms of knowledge: practical and emancipatory. Pedagogy that supports development of practical and emancipatory knowledge includes problem-based learning and case methods, heuristics, reflective practica, journals, storytelling, and performance-based assessment methods. These important teaching strategies are being integrated into various dental curricula including a new competency-based dental curriculum at Marquette University's School of Dentistry. It will be critical for dental educators to continue developing these methods to provide efficient and effective education for future practitioners in both the art and science of dentistry.
Goffin, Dorothee; Delzenne, Nathalie; Blecker, Christophe; Hanon, Emilien; Deroanne, Claude; Paquot, Michel
2011-05-01
This critical review article presents the current state of knowledge on isomalto-oligosaccharides, some well known functional oligosaccharides in Asia, to evaluate their potential as emergent prebiotics in the American and European functional food market. It includes first a unique inventory of the different families of compounds which have been considered as IMOs and their specific structure. A description has been given of the different production methods including the involved enzymes and their specific activities, the substrates, and the types of IMOs produced. Considering the structural complexity of IMO products, specific characterization methods are described, as well as purification methods which enable the body to get rid of digestible oligosaccharides. Finally, an extensive review of their techno-functional and nutritional properties enables placing IMOs inside the growing prebiotic market. This review is of particular interest considering that IMO commercialization in America and Europe is a topical subject due to the recent submission by Bioneutra Inc. (Canada) of a novel food file to the UK Food Standards Agency, as well as several patents for IMO production.
Uncertainties in predicting solar panel power output
NASA Technical Reports Server (NTRS)
Anspaugh, B.
1974-01-01
The problem of calculating solar panel power output at launch and during a space mission is considered. The major sources of uncertainty and error in predicting the post launch electrical performance of the panel are considered. A general discussion of error analysis is given. Examples of uncertainty calculations are included. A general method of calculating the effect on the panel of various degrading environments is presented, with references supplied for specific methods. A technique for sizing a solar panel for a required mission power profile is developed.
A Nonpharmacologic Method for Enhancing Sleep in PTSD
2015-10-01
medications include: Alcohol (during intoxication or withdrawal); cannabis (during intoxication); hallucinogens (during intoxication), phencyclidine... medications are taken solely under appropriate medical supervision, this criterion is not considered to be met. SEDATIVE/ HYPNOTIC/ANX CANNABIS ...are taken solely under appropriate medical supervision, this criterion is not considered to be met. SEDATIVE/ HYPNOTIC/ANX CANNABIS STIMULANTS
Postsecondary Education: Opportunities and Challenges for Students Who Are Deaf-Blind.
ERIC Educational Resources Information Center
Everson, Jane, Ed.; Enos, JoAnn, Ed.
1995-01-01
Some of the opportunities and challenges associated with postsecondary education for young adults who are deaf-blind are considered. Suggestions are offered for transition planning teams to consider, including analysis of students' most and least favorite classes, vocational goals, housing goals, and methods of paying for postsecondary education.…
ERIC Educational Resources Information Center
GüvendIr, Emre
2013-01-01
Considering the significance of taking student preferences into account while organizing teaching practices, the current study explores which teaching method prospective foreign language teachers mostly prefer their teacher to use in the language acquisition course. A teaching methods evaluation form that includes six commonly used teaching…
ERIC Educational Resources Information Center
Thatcher, L. L.; And Others
Analytical methods for determining important components of fission and natural radioactivity found in water are reported. The discussion of each method includes conditions for application of the method, a summary of the method, interferences, required apparatus, procedures, calculations and estimation of precision. Isotopes considered are…
Is Self-Paced Instruction Really Worth It?
ERIC Educational Resources Information Center
Roberson, J. A.; Crowe, C. T.
1975-01-01
Describes a self-paced, learning-for-mastery course in undergraduate fluid mechanics. Includes the method of course assessment, method of student evaluation, and a description of the instructor's role and work load. Summarizes aspects of self-paced instruction considered favorable and unfavorable. (GS)
Teaching Qualitative Methods: A Face-to-Face Encounter.
ERIC Educational Resources Information Center
Keen, Mike F.
1996-01-01
Considers the complete ethnographic project as a strategy for teaching qualitative methods. Describes an undergraduate class where students chose an ethnographic setting, gathered and analyzed data, and wrote a final report. Settings included Laundromats, bingo halls, auctions, karaoke clubs, and bowling leagues. (MJP)
Propellant Mass Fraction Calculation Methodology for Launch Vehicles
NASA Technical Reports Server (NTRS)
Holt, James B.; Monk, Timothy S.
2009-01-01
Propellant Mass Fraction (pmf) calculation methods vary throughout the aerospace industry. While typically used as a means of comparison between competing launch vehicle designs, the actual pmf calculation method varies slightly from one entity to another. It is the purpose of this paper to present various methods used to calculate the pmf of a generic launch vehicle. This includes fundamental methods of pmf calculation which consider only the loaded propellant and the inert mass of the vehicle, more involved methods which consider the residuals and any other unusable propellant remaining in the vehicle, and other calculations which exclude large mass quantities such as the installed engine mass. Finally, a historic comparison is made between launch vehicles on the basis of the differing calculation methodologies.
Comparison of Anaerobic Susceptibility Results Obtained by Different Methods
Rosenblatt, J. E.; Murray, P. R.; Sonnenwirth, A. C.; Joyce, J. L.
1979-01-01
Susceptibility tests using 7 antimicrobial agents (carbenicillin, chloramphenicol, clindamycin, penicillin, cephalothin, metronidazole, and tetracycline) were run against 35 anaerobes including Bacteroides fragilis (17), other gram-negative bacilli (7), clostridia (5), peptococci (4), and eubacteria (2). Results in triplicate obtained by the microbroth dilution method and the aerobic modification of the broth disk method were compared with those obtained with an agar dilution method using Wilkins-Chalgren agar. Media used in the microbroth dilution method included Wilkins-Chalgren broth, brain heart infusion broth, brucella broth, tryptic soy broth, thioglycolate broth, and Schaedler's broth. A result differing by more than one dilution from the Wilkins-Chalgren agar result was considered a discrepancy, and when there was a change in susceptibility status this was termed a significant discrepancy. The microbroth dilution method using Wilkins-Chalgren broth and thioglycolate broth produced the fewest total discrepancies (22 and 24, respectively), and Wilkins-Chalgren broth, thioglycolate, and Schaedler's broth had the fewest significant discrepancies (6, 5, and 5, respectively). With the broth disk method, there were 15 significant discrepancies, although half of these were with tetracycline, which was the antimicrobial agent associated with the highest number of significant discrepancies (33), considering all of the test methods and media. PMID:464560
Ageing airplane repair assessment program for Airbus A300
NASA Technical Reports Server (NTRS)
Gaillardon, J. M.; Schmidt, HANS-J.; Brandecker, B.
1992-01-01
This paper describes the current status of the repair categorization activities and includes all details about the methodologies developed for determination of the inspection program for the skin on pressurized fuselages. For inspection threshold determination two methods are defined based on fatigue life approach, a simplified and detailed method. The detailed method considers 15 different parameters to assess the influences of material, geometry, size location, aircraft usage, and workmanship on the fatigue life of the repair and the original structure. For definition of the inspection intervals a general method is developed which applies to all concerned repairs. For this the initial flaw concept is used by considering 6 parameters and the detectable flaw sizes depending on proposed nondestructive inspection methods. An alternative method is provided for small repairs allowing visual inspection with shorter intervals.
Effect of processing on Polymer/Composite structure and properties
NASA Technical Reports Server (NTRS)
1982-01-01
Advances in the vitality and economic health of the field of polymer forecasting are discussed. A consistent and rational point of view which considers processing as a participant in the underlying triad of relationships which comprise materials science and engineering is outlined. This triad includes processing as it influences material structure, and ultimately properties. Methods in processing structure properties, polymer science and engineering, polymer chemistry and synthesis, structure and modification and optimization through processing, and methods of melt flow modeling in processing structure property relations of polymer were developed. Mechanical properties of composites are considered, and biomedical materials research to include polymer processing effects are studied. An analysis of the design technology of advances graphite/epoxy composites is also reported.
Fourth NASA Langley Formal Methods Workshop
NASA Technical Reports Server (NTRS)
Holloway, C. Michael (Compiler); Hayhurst, Kelly J. (Compiler)
1997-01-01
This publication consists of papers presented at NASA Langley Research Center's fourth workshop on the application of formal methods to the design and verification of life-critical systems. Topic considered include: Proving properties of accident; modeling and validating SAFER in VDM-SL; requirement analysis of real-time control systems using PVS; a tabular language for system design; automated deductive verification of parallel systems. Also included is a fundamental hardware design in PVS.
Prediction of protein post-translational modifications: main trends and methods
NASA Astrophysics Data System (ADS)
Sobolev, B. N.; Veselovsky, A. V.; Poroikov, V. V.
2014-02-01
The review summarizes main trends in the development of methods for the prediction of protein post-translational modifications (PTMs) by considering the three most common types of PTMs — phosphorylation, acetylation and glycosylation. Considerable attention is given to general characteristics of regulatory interactions associated with PTMs. Different approaches to the prediction of PTMs are analyzed. Most of the methods are based only on the analysis of the neighbouring environment of modification sites. The related software is characterized by relatively low accuracy of PTM predictions, which may be due both to the incompleteness of training data and the features of PTM regulation. Advantages and limitations of the phylogenetic approach are considered. The prediction of PTMs using data on regulatory interactions, including the modular organization of interacting proteins, is a promising field, provided that a more carefully selected training data will be used. The bibliography includes 145 references.
Methods for determination of radioactive substances in water and fluvial sediments
Thatcher, Leland Lincoln; Janzer, Victor J.; Edwards, Kenneth W.
1977-01-01
Analytical methods for the determination of some of the more important components of fission or neutron activation product radioactivity and of natural radioactivity found in water are reported. The report for each analytical method includes conditions for application of the method, a summary of the method, interferences, required apparatus and reagents, analytical procedures, calculations, reporting of results, and estimation of precision. The fission product isotopes considered are cesium-137, strontium-90, and ruthenium-106. The natural radioelements and isotopes considered are uranium, lead-210, radium-226, radium-228, tritium, and carbon-14. A gross radioactivity survey method and a uranium isotope ratio method are given. When two analytical methods are in routine use for an individual isotope, both methods are reported with identification of the specific areas of application of each. Techniques for the collection and preservation of water samples to be analyzed for radioactivity are discussed.
Calculation methods for compressible turbulent boundary layers, 1976
NASA Technical Reports Server (NTRS)
Bushnell, D. M.; Cary, A. M., Jr.; Harris, J. E.
1977-01-01
Equations and closure methods for compressible turbulent boundary layers are discussed. Flow phenomena peculiar to calculation of these boundary layers were considered, along with calculations of three dimensional compressible turbulent boundary layers. Procedures for ascertaining nonsimilar two and three dimensional compressible turbulent boundary layers were appended, including finite difference, finite element, and mass-weighted residual methods.
Simulation of the optical coating deposition
NASA Astrophysics Data System (ADS)
Grigoriev, Fedor; Sulimov, Vladimir; Tikhonravov, Alexander
2018-04-01
A brief review of the mathematical methods of thin-film growth simulation and results of their applications is presented. Both full-atomistic and multi-scale approaches that were used in the studies of thin-film deposition are considered. The results of the structural parameter simulation including density profiles, roughness, porosity, point defect concentration, and others are discussed. The application of the quantum level methods to the simulation of the thin-film electronic and optical properties is considered. Special attention is paid to the simulation of the silicon dioxide thin films.
NASA Technical Reports Server (NTRS)
Gray, D. J.
1978-01-01
Cryogenic transportation methods for providing liquid hydrogen requirements are examined in support of shuttle transportation system launch operations at Kennedy Space Center, Florida, during the time frames 1982-1991 in terms of cost and operational effectiveness. Transportation methods considered included sixteen different options employing mobile semi-trailer tankers, railcars, barges and combinations of each method. The study concludes that the most effective method of delivering liquid hydrogen from the vendor production facility in New Orleans to Kennedy Space Center includes maximum utilization of existing mobile tankers and railcars supplemented by maximum capacity mobile tankers procured incrementally in accordance with shuttle launch rates actually achieved.
THE CRITICAL-PATH METHOD OF CONSTRUCTION CONTROL.
ERIC Educational Resources Information Center
DOMBROW, RODGER T.; MAUCHLY, JOHN
THIS DISCUSSION PRESENTS A DEFINITION AND BRIEF DESCRIPTION OF THE CRITICAL-PATH METHOD AS APPLIED TO BUILDING CONSTRUCTION. INTRODUCING REMARKS CONSIDER THE MOST PERTINENT QUESTIONS PERTAINING TO CPM AND THE NEEDS ASSOCIATED WITH MINIMIZING TIME AND COST ON CONSTRUCTION PROJECTS. SPECIFIC DISCUSSION INCLUDES--(1) ADVANTAGES OF NETWORK TECHNIQUES,…
Parental Depressive Symptoms and Children's Sleep: The Role of Family Conflict
ERIC Educational Resources Information Center
El-Sheikh, Mona; Kelly, Ryan J.; Bagley, Erika J.; Wetter, Emily K.
2012-01-01
Background: We used a multi-method and multi-informant design to identify developmental pathways through which parental depressive symptoms contribute to children's sleep problems. Environmental factors including adult inter-partner conflict and parent-child conflict were considered as process variables of this relation. Methods: An ethnically and…
Frailty and Organization of Health and Social Care.
Clegg, Andrew; Young, John
2015-01-01
In this chapter, we consider how health and social care can best be organized for older people with frailty. We will consider the merits of routine frailty identification, including risk stratification methods, to inform the provision of evidence-based treatment and holistic, goal-oriented care. We will also consider how best to place older people with frailty at the heart of health and social care systems so that the complex challenges associated with this vulnerable group are addressed. 2015 S. Karger AG, Basel.
MaCH-Admix: Genotype Imputation for Admixed Populations
Liu, Eric Yi; Li, Mingyao; Wang, Wei; Li, Yun
2012-01-01
Imputation in admixed populations is an important problem but challenging due to the complex linkage disequilibrium (LD) pattern. The emergence of large reference panels such as that from the 1,000 Genomes Project enables more accurate imputation in general, and in particular for admixed populations and for uncommon variants. To efficiently benefit from these large reference panels, one key issue to consider in modern genotype imputation framework is the selection of effective reference panels. In this work, we consider a number of methods for effective reference panel construction inside a hidden Markov model and specific to each target individual. These methods fall into two categories: identity-by-state (IBS) based and ancestry-weighted approach. We evaluated the performance on individuals from recently admixed populations. Our target samples include 8,421 African Americans and 3,587 Hispanic Americans from the Women’s Health Initiative, which allow assessment of imputation quality for uncommon variants. Our experiments include both large and small reference panels; large, medium, and small target samples; and in genome regions of varying levels of LD. We also include BEAGLE and IMPUTE2 for comparison. Experiment results with large reference panel suggest that our novel piecewise IBS method yields consistently higher imputation quality than other methods/software. The advantage is particularly noteworthy among uncommon variants where we observe up to 5.1% information gain with the difference being highly significant (Wilcoxon signed rank test P-value < 0.0001). Our work is the first that considers various sensible approaches for imputation in admixed populations and presents a comprehensive comparison. PMID:23074066
A Multiscale Software Tool for Field/Circuit Co-Simulation
2011-12-15
technology fields: Number of graduating undergraduates who achieved a 3.5 GPA to 4.0 (4.0 max scale): Number of graduating undergraduates funded by a...times more efficient than FDTD for such a problem in 3D . The techniques in class (c) above include the discontinuous Galerkin method and multidomain...implements a finite-differential-time-domain method on single field propagation in a 3D space. We consider a cavity model which includes two electric
NASA Astrophysics Data System (ADS)
Bulatov, S. V.
2018-05-01
The article considers the method of short-term combined forecasting, which includes theoretical and experimental estimates of the need for details of units and assemblies, which allows obtaining the optimum number of spare parts necessary for rolling stock operation without downtime in repair areas.
Autism Diagnosis and Screening: Factors to Consider in Differential Diagnosis
ERIC Educational Resources Information Center
Matson, Johnny L.; Beighley, Jennifer; Turygin, Nicole
2012-01-01
There has been an exponential growth in assessment methods to diagnose disorders on the autism spectrum. Many reasons for this trend exist and include advancing knowledge on how to make a diagnosis, the heterogeneity of the spectrum, the realization that different methods may be needed based on age and intellectual disability. Other factors…
ERIC Educational Resources Information Center
Hagedorn, Linda Serra
1998-01-01
A study explored two distinct methods of calculating a precise measure of gender-based wage differentials among college faculty. The first estimation considered wage differences using a formula based on human capital; the second included compensation for past discriminatory practices. Both measures were used to predict three specific aspects of…
Reconstruction of phonon relaxation times from systems featuring interfaces with unknown properties
NASA Astrophysics Data System (ADS)
Forghani, Mojtaba; Hadjiconstantinou, Nicolas G.
2018-05-01
We present a method for reconstructing the phonon relaxation-time function τω=τ (ω ) (including polarization) and associated phonon free-path distribution from thermal spectroscopy data for systems featuring interfaces with unknown properties. Our method does not rely on the effective thermal-conductivity approximation or a particular physical model of the interface behavior. The reconstruction is formulated as an optimization problem in which the relaxation times are determined as functions of frequency by minimizing the discrepancy between the experimentally measured temperature profiles and solutions of the Boltzmann transport equation for the same system. Interface properties such as transmissivities are included as unknowns in the optimization; however, because for the thermal spectroscopy problems considered here the reconstruction is not very sensitive to the interface properties, the transmissivities are only approximately reconstructed and can be considered as byproducts of the calculation whose primary objective is the accurate determination of the relaxation times. The proposed method is validated using synthetic experimental data obtained from Monte Carlo solutions of the Boltzmann transport equation. The method is shown to remain robust in the presence of uncertainty (noise) in the measurement.
Airbreathing hypersonic vehicle design and analysis methods
NASA Technical Reports Server (NTRS)
Lockwood, Mary Kae; Petley, Dennis H.; Hunt, James L.; Martin, John G.
1996-01-01
The design, analysis, and optimization of airbreathing hypersonic vehicles requires analyses involving many highly coupled disciplines at levels of accuracy exceeding those traditionally considered in a conceptual or preliminary-level design. Discipline analysis methods including propulsion, structures, thermal management, geometry, aerodynamics, performance, synthesis, sizing, closure, and cost are discussed. Also, the on-going integration of these methods into a working environment, known as HOLIST, is described.
Algorithms for Learning Preferences for Sets of Objects
NASA Technical Reports Server (NTRS)
Wagstaff, Kiri L.; desJardins, Marie; Eaton, Eric
2010-01-01
A method is being developed that provides for an artificial-intelligence system to learn a user's preferences for sets of objects and to thereafter automatically select subsets of objects according to those preferences. The method was originally intended to enable automated selection, from among large sets of images acquired by instruments aboard spacecraft, of image subsets considered to be scientifically valuable enough to justify use of limited communication resources for transmission to Earth. The method is also applicable to other sets of objects: examples of sets of objects considered in the development of the method include food menus, radio-station music playlists, and assortments of colored blocks for creating mosaics. The method does not require the user to perform the often-difficult task of quantitatively specifying preferences; instead, the user provides examples of preferred sets of objects. This method goes beyond related prior artificial-intelligence methods for learning which individual items are preferred by the user: this method supports a concept of setbased preferences, which include not only preferences for individual items but also preferences regarding types and degrees of diversity of items in a set. Consideration of diversity in this method involves recognition that members of a set may interact with each other in the sense that when considered together, they may be regarded as being complementary, redundant, or incompatible to various degrees. The effects of such interactions are loosely summarized in the term portfolio effect. The learning method relies on a preference representation language, denoted DD-PREF, to express set-based preferences. In DD-PREF, a preference is represented by a tuple that includes quality (depth) functions to estimate how desired a specific value is, weights for each feature preference, the desired diversity of feature values, and the relative importance of diversity versus depth. The system applies statistical concepts to estimate quantitative measures of the user s preferences from training examples (preferred subsets) specified by the user. Once preferences have been learned, the system uses those preferences to select preferred subsets from new sets. The method was found to be viable when tested in computational experiments on menus, music playlists, and rover images. Contemplated future development efforts include further tests on more diverse sets and development of a sub-method for (a) estimating the parameter that represents the relative importance of diversity versus depth, and (b) incorporating background knowledge about the nature of quality functions, which are special functions that specify depth preferences for features.
Finite element methods and Navier-Stokes equations
NASA Astrophysics Data System (ADS)
Cuvelier, C.; Segal, A.; van Steenhoven, A. A.
This book is devoted to two and three-dimensional FEM analysis of the Navier-Stokes (NS) equations describing one flow of a viscous incompressible fluid. Three different approaches to the NS equations are described: a direct method, a penalty method, and a method that constructs discrete solenoidal vector fields. Subjects of current research which are important from the industrial/technological viewpoint are considered, including capillary-free boundaries, nonisothermal flows, turbulence, and non-Newtonian fluids.
Prediction of forces and moments for hypersonic flight vehicle control effectors
NASA Technical Reports Server (NTRS)
Maughmer, Mark D.; Long, Lyle N.; Guilmette, Neal; Pagano, Peter
1993-01-01
This research project includes three distinct phases. For completeness, all three phases of the work are briefly described in this report. The goal was to develop methods of predicting flight control forces and moments for hypersonic vehicles which could be used in a preliminary design environment. The first phase included a preliminary assessment of subsonic/supersonic panel methods and hypersonic local flow inclination methods for such predictions. While these findings clearly indicated the usefulness of such methods for conceptual design activities, deficiencies exist in some areas. Thus, a second phase of research was conducted in which a better understanding was sought for the reasons behind the successes and failures of the methods considered, particularly for the cases at hypersonic Mach numbers. This second phase involved using computational fluid dynamics methods to examine the flow fields in detail. Through these detailed predictions, the deficiencies in the simple surface inclination methods were determined. In the third phase of this work, an improvement to the surface inclination methods was developed. This used a novel method for including viscous effects by modifying the geometry to include the viscous/shock layer.
A comprehensive probabilistic analysis model of oil pipelines network based on Bayesian network
NASA Astrophysics Data System (ADS)
Zhang, C.; Qin, T. X.; Jiang, B.; Huang, C.
2018-02-01
Oil pipelines network is one of the most important facilities of energy transportation. But oil pipelines network accident may result in serious disasters. Some analysis models for these accidents have been established mainly based on three methods, including event-tree, accident simulation and Bayesian network. Among these methods, Bayesian network is suitable for probabilistic analysis. But not all the important influencing factors are considered and the deployment rule of the factors has not been established. This paper proposed a probabilistic analysis model of oil pipelines network based on Bayesian network. Most of the important influencing factors, including the key environment condition and emergency response are considered in this model. Moreover, the paper also introduces a deployment rule for these factors. The model can be used in probabilistic analysis and sensitive analysis of oil pipelines network accident.
Numerical methods for the stochastic Landau-Lifshitz Navier-Stokes equations.
Bell, John B; Garcia, Alejandro L; Williams, Sarah A
2007-07-01
The Landau-Lifshitz Navier-Stokes (LLNS) equations incorporate thermal fluctuations into macroscopic hydrodynamics by using stochastic fluxes. This paper examines explicit Eulerian discretizations of the full LLNS equations. Several computational fluid dynamics approaches are considered (including MacCormack's two-step Lax-Wendroff scheme and the piecewise parabolic method) and are found to give good results for the variance of momentum fluctuations. However, neither of these schemes accurately reproduces the fluctuations in energy or density. We introduce a conservative centered scheme with a third-order Runge-Kutta temporal integrator that does accurately produce fluctuations in density, energy, and momentum. A variety of numerical tests, including the random walk of a standing shock wave, are considered and results from the stochastic LLNS solver are compared with theory, when available, and with molecular simulations using a direct simulation Monte Carlo algorithm.
Methods for recording and measuring tonic GABAA receptor-mediated inhibition
Bright, Damian P.; Smart, Trevor G.
2013-01-01
Tonic inhibitory conductances mediated by GABAA receptors have now been identified and characterized in many different brain regions. Most experimental studies of tonic GABAergic inhibition have been carried out using acute brain slice preparations but tonic currents have been recorded under a variety of different conditions. This diversity of recording conditions is likely to impact upon many of the factors responsible for controlling tonic inhibition and can make comparison between different studies difficult. In this review, we will firstly consider how various experimental conditions, including age of animal, recording temperature and solution composition, are likely to influence tonic GABAA conductances. We will then consider some technical considerations related to how the tonic conductance is measured and subsequently analyzed, including how the use of current noise may provide a complementary and reliable method for quantifying changes in tonic current. PMID:24367296
Community Detection in Complex Networks via Clique Conductance.
Lu, Zhenqi; Wahlström, Johan; Nehorai, Arye
2018-04-13
Network science plays a central role in understanding and modeling complex systems in many areas including physics, sociology, biology, computer science, economics, politics, and neuroscience. One of the most important features of networks is community structure, i.e., clustering of nodes that are locally densely interconnected. Communities reveal the hierarchical organization of nodes, and detecting communities is of great importance in the study of complex systems. Most existing community-detection methods consider low-order connection patterns at the level of individual links. But high-order connection patterns, at the level of small subnetworks, are generally not considered. In this paper, we develop a novel community-detection method based on cliques, i.e., local complete subnetworks. The proposed method overcomes the deficiencies of previous similar community-detection methods by considering the mathematical properties of cliques. We apply the proposed method to computer-generated graphs and real-world network datasets. When applied to networks with known community structure, the proposed method detects the structure with high fidelity and sensitivity. When applied to networks with no a priori information regarding community structure, the proposed method yields insightful results revealing the organization of these complex networks. We also show that the proposed method is guaranteed to detect near-optimal clusters in the bipartition case.
Method for making carbon super capacitor electrode materials
Firsich, D.W.; Ingersoll, D.; Delnick, F.M.
1998-07-07
A method is described for making near-net-shape, monolithic carbon electrodes for energy storage devices. The method includes the controlled pyrolysis and activation of a pressed shape of methyl cellulose powder with pyrolysis being carried out in two stages; pre-oxidation, preferably in air at a temperature between 200--250 C, followed by carbonization under an inert atmosphere. An activation step to adjust the surface area of the carbon shape to a value desirable for the application being considered, including heating the carbon shape in an oxidizing atmosphere to a temperature of at least 300 C, follows carbonization. 1 fig.
Method for making carbon super capacitor electrode materials
Firsich, David W.; Ingersoll, David; Delnick, Frank M.
1998-01-01
A method for making near-net-shape, monolithic carbon electrodes for energy storage devices. The method includes the controlled pyrolysis and activation of a pressed shape of methyl cellulose powder with pyrolysis being carried out in two stages; pre-oxidation, preferably in air at a temperature between 200.degree.-250.degree. C., followed by carbonization under an inert atmosphere. An activation step to adjust the surface area of the carbon shape to a value desirable for the application being considered, including heating the carbon shape in an oxidizing atmosphere to a temperature of at least 300.degree. C., follows carbonization.
Seismic Risk Studies in the United States.
ERIC Educational Resources Information Center
Algermissen, S.T.
A new seismic risk map of the United States is presented, along with strain release and maximum Modified Mercalli intesity maps of the country. Frequency of occurrence of damaging earthquakes was not considered in zone ratings, but included frequency studies may aid interpretation. Discussion of methods is included with review of calculations. (MH)
Examining Menstrual Tracking to Inform the Design of Personal Informatics Tools
Epstein, Daniel A.; Lee, Nicole B.; Kang, Jennifer H.; Agapie, Elena; Schroeder, Jessica; Pina, Laura R.; Fogarty, James; Kientz, Julie A.; Munson, Sean A.
2017-01-01
We consider why and how women track their menstrual cycles, examining their experiences to uncover design opportunities and extend the field's understanding of personal informatics tools. To understand menstrual cycle tracking practices, we collected and analyzed data from three sources: 2,000 reviews of popular menstrual tracking apps, a survey of 687 people, and follow-up interviews with 12 survey respondents. We find that women track their menstrual cycle for varied reasons that include remembering and predicting their period as well as informing conversations with healthcare providers. Participants described six methods of tracking their menstrual cycles, including use of technology, awareness of their premenstrual physiological states, and simply remembering. Although women find apps and calendars helpful, these methods are ineffective when predictions of future menstrual cycles are inaccurate. Designs can create feelings of exclusion for gender and sexual minorities. Existing apps also generally fail to consider life stages that women experience, including young adulthood, pregnancy, and menopause. Our findings encourage expanding the field's conceptions of personal informatics. PMID:28516176
2010-01-01
Background Cluster analysis, and in particular hierarchical clustering, is widely used to extract information from gene expression data. The aim is to discover new classes, or sub-classes, of either individuals or genes. Performing a cluster analysis commonly involve decisions on how to; handle missing values, standardize the data and select genes. In addition, pre-processing, involving various types of filtration and normalization procedures, can have an effect on the ability to discover biologically relevant classes. Here we consider cluster analysis in a broad sense and perform a comprehensive evaluation that covers several aspects of cluster analyses, including normalization. Result We evaluated 2780 cluster analysis methods on seven publicly available 2-channel microarray data sets with common reference designs. Each cluster analysis method differed in data normalization (5 normalizations were considered), missing value imputation (2), standardization of data (2), gene selection (19) or clustering method (11). The cluster analyses are evaluated using known classes, such as cancer types, and the adjusted Rand index. The performances of the different analyses vary between the data sets and it is difficult to give general recommendations. However, normalization, gene selection and clustering method are all variables that have a significant impact on the performance. In particular, gene selection is important and it is generally necessary to include a relatively large number of genes in order to get good performance. Selecting genes with high standard deviation or using principal component analysis are shown to be the preferred gene selection methods. Hierarchical clustering using Ward's method, k-means clustering and Mclust are the clustering methods considered in this paper that achieves the highest adjusted Rand. Normalization can have a significant positive impact on the ability to cluster individuals, and there are indications that background correction is preferable, in particular if the gene selection is successful. However, this is an area that needs to be studied further in order to draw any general conclusions. Conclusions The choice of cluster analysis, and in particular gene selection, has a large impact on the ability to cluster individuals correctly based on expression profiles. Normalization has a positive effect, but the relative performance of different normalizations is an area that needs more research. In summary, although clustering, gene selection and normalization are considered standard methods in bioinformatics, our comprehensive analysis shows that selecting the right methods, and the right combinations of methods, is far from trivial and that much is still unexplored in what is considered to be the most basic analysis of genomic data. PMID:20937082
Using the entire history in the analysis of nested case cohort samples.
Rivera, C L; Lumley, T
2016-08-15
Countermatching designs can provide more efficient estimates than simple matching or case-cohort designs in certain situations such as when good surrogate variables for an exposure of interest are available. We extend pseudolikelihood estimation for the Cox model under countermatching designs to models where time-varying covariates are considered. We also implement pseudolikelihood with calibrated weights to improve efficiency in nested case-control designs in the presence of time-varying variables. A simulation study is carried out, which considers four different scenarios including a binary time-dependent variable, a continuous time-dependent variable, and the case including interactions in each. Simulation results show that pseudolikelihood with calibrated weights under countermatching offers large gains in efficiency if compared to case-cohort. Pseudolikelihood with calibrated weights yielded more efficient estimators than pseudolikelihood estimators. Additionally, estimators were more efficient under countermatching than under case-cohort for the situations considered. The methods are illustrated using the Colorado Plateau uranium miners cohort. Furthermore, we present a general method to generate survival times with time-varying covariates. Copyright © 2016 John Wiley & Sons, Ltd. Copyright © 2016 John Wiley & Sons, Ltd.
de Irala, Jokin; Lopez del Burgo, Cristina; Lopez de Fez, Carmen M; Arredondo, Jorge; Mikolajczyk, Rafael T; Stanford, Joseph B
2007-06-27
Informed consent in family planning includes knowledge of mechanism of action. Some methods of family planning occasionally work after fertilization. Knowing about postfertilization effects may be important to some women before choosing a certain family planning method. The objective of this survey is to explore women's attitudes towards postfertilization effects of family planning methods, and beliefs and characteristics possibly associated with those attitudes. Cross-sectional survey in a sample of 755 potentially fertile women, aged 18-49, from Primary Care Health Centres in Pamplona, Spain. Participants were given a 30-item, self-administered, anonymous questionnaire about family planning methods and medical and surgical abortion. Logistic regression was used to identify variables associated with women's attitudes towards postfertilization effects. The response rate was 80%. The majority of women were married, held an academic degree and had no children. Forty percent of women would not consider using a method that may work after fertilization but before implantation and 57% would not consider using one that may work after implantation. While 35.3% of the sample would stop using a method if they learned that it sometimes works after fertilization, this percentage increased to 56.3% when referring to a method that sometimes works after implantation. Women who believe that human life begins at fertilization and those who consider it is important to distinguish between natural and induced embryo loss were less likely to consider the use of a method with postfertilization effects. Information about potential postfertilization effects of family planning methods may influence women's acceptance and choice of a particular family planning method. Additional studies in other populations are necessary to evaluate whether these beliefs are important to those populations.
Acyl Meldrum's acid derivatives: application in organic synthesis
NASA Astrophysics Data System (ADS)
Janikowska, K.; Rachoń, J.; Makowiec, S.
2014-07-01
This review is focused on an important class of Meldrum's acid derivatives commonly known as acyl Meldrum's acids. The preparation methods of these compounds are considered including the recently proposed and rather rarely used ones. The chemical properties of acyl Meldrum's acids are described in detail, including thermal stability and reactions with various nucleophiles. The possible mechanisms of these transformations are analyzed. The bibliography includes 134 references.
Friend suggestion in social network based on user log
NASA Astrophysics Data System (ADS)
Kaviya, R.; Vanitha, M.; Sumaiya Thaseen, I.; Mangaiyarkarasi, R.
2017-11-01
Simple friend recommendation algorithms such as similarity, popularity and social aspects is the basic requirement to be explored to methodically form high-performance social friend recommendation. Suggestion of friends is followed. No tags of character were followed. In the proposed system, we use an algorithm for network correlation-based social friend recommendation (NC-based SFR).It includes user activities like where one lives and works. A new friend recommendation method, based on network correlation, by considering the effect of different social roles. To model the correlation between different networks, we develop a method that aligns these networks through important feature selection. We consider by preserving the network structure for a more better recommendations so that it significantly improves the accuracy for better friend-recommendation.
Computer vs. Typewriter: Changes in Teaching Methods.
ERIC Educational Resources Information Center
Frankeberger, Lynda
1990-01-01
Factors to consider in making a decision whether to convert traditional typewriting classrooms to microcomputer classrooms include effects on oral instruction, ethical issues in file transfer, and use of keyboarding software and timed writing software. (JOW)
Reflective cracking control : interim report - fourth year.
DOT National Transportation Integrated Search
1999-07-01
Reflective cracking has long been considered a major problem associated with asphalt : pavements. Several methods including milling, crack sealing and fabric membranes have : been used in an attempt to eliminate or delay the reflective cracking proce...
Code of Federal Regulations, 2010 CFR
2010-04-01
... (IND) Application, any information obtained during the inspection of an extramural facility having a... Administration does not consider results of validation studies of analytical and assay methods and control...
ERIC Educational Resources Information Center
Harris, Carolyn
1979-01-01
Reviews methods being developed for mass deacidification of books to prevent deterioration of paper. The use of diethyl zinc, liquified gas, and morpholine, and the advantages, disadvantages, and cost of each are considered. A 26-item bibliography is included. (JD)
NASA Technical Reports Server (NTRS)
Hansen, J. E. (Editor)
1975-01-01
Topics considered at the conference included the dynamics, structure, chemistry, and evolution of the Venus atmosphere, as well as cloud physics and motion. Infrared, ultraviolet, and radio occultation methods of analysis are discussed, and atmospheric models are described.
The challenges of including impacts on biodiversity in agricultural life cycle assessments.
Gabel, Vanessa M; Meier, Matthias S; Köpke, Ulrich; Stolze, Matthias
2016-10-01
Agriculture is considered to be one of the main drivers for worldwide biodiversity loss but the impacts of agricultural production on biodiversity have not been extensively considered in Life Cycle Assessments (LCAs). Recent realisation that biodiversity impact should be included in comprehensive LCAs has led to attempts to develop and implement methods for biodiversity impact assessment. In this review, twenty-two different biodiversity impact assessment methods have been analysed to identify their strengths and weaknesses in terms of their comprehensiveness in the evaluation of agricultural products. Different criteria, which had to meet the specific requirements of biodiversity research, life cycle assessment methodology, and the evaluation of agricultural products, were selected to investigate the identified methods. Very few of the methods were developed with the specific intention of being used for agricultural LCAs. Furthermore, none of the methods can be applied globally while at the same time being able to differentiate between various agricultural intensities. Global value chains and the increasing awareness of different biodiversity impacts of agricultural production systems demand the development of evaluation methods that are able to overcome these shortcomings. Despite the progress that has already been achieved, there are still unresolved difficulties which need further research and improvement. Copyright © 2016 Elsevier Ltd. All rights reserved.
An annular superposition integral for axisymmetric radiators.
Kelly, James F; McGough, Robert J
2007-02-01
A fast integral expression for computing the nearfield pressure is derived for axisymmetric radiators. This method replaces the sum of contributions from concentric annuli with an exact double integral that converges much faster than methods that evaluate the Rayleigh-Sommerfeld integral or the generalized King integral. Expressions are derived for plane circular pistons using both continuous wave and pulsed excitations. Several commonly used apodization schemes for the surface velocity distribution are considered, including polynomial functions and a "smooth piston" function. The effect of different apodization functions on the spectral content of the wave field is explored. Quantitative error and time comparisons between the new method, the Rayleigh-Sommerfeld integral, and the generalized King integral are discussed. At all error levels considered, the annular superposition method achieves a speed-up of at least a factor of 4 relative to the point-source method and a factor of 3 relative to the generalized King integral without increasing the computational complexity.
NASA Technical Reports Server (NTRS)
Holt, James B.; Monk, Timothy S.
2009-01-01
Propellant Mass Fraction (pmf) calculation methods vary throughout the aerospace industry. While typically used as a means of comparison between candidate launch vehicle designs, the actual pmf calculation method varies slightly from one entity to another. It is the purpose of this paper to present various methods used to calculate the pmf of launch vehicles. This includes fundamental methods of pmf calculation that consider only the total propellant mass and the dry mass of the vehicle; more involved methods that consider the residuals, reserves and any other unusable propellant remaining in the vehicle; and calculations excluding large mass quantities such as the installed engine mass. Finally, a historical comparison is made between launch vehicles on the basis of the differing calculation methodologies, while the unique mission and design requirements of the Ares V Earth Departure Stage (EDS) are examined in terms of impact to pmf.
Riley, Emily; Harris, Patrick; Kent, Jennifer; Sainsbury, Peter; Lane, Anna; Baum, Fran
2018-05-10
Transport policy and practice impacts health. Environmental Impact Assessments (EIAs) are regulated public policy mechanisms that can be used to consider the health impacts of major transport projects before they are approved. The way health is considered in these environmental assessments (EAs) is not well known. This research asked: How and to what extent was human health considered in EAs of four major transport projects in Australia. We developed a comprehensive coding framework to analyse the Environmental Impact Statements (EISs) of four transport infrastructure projects: three road and one light rail. The coding framework was designed to capture how health was directly and indirectly included. We found that health was partially considered in all four EISs. In the three New South Wales (NSW) projects, but not the one South Australian project, this was influenced by the requirements issued to proponents by the government which directed the content of the EIS. Health was assessed using human health risk assessment (HHRA). We found this to be narrow in focus and revealed a need for a broader social determinants of health approach, using multiple methods. The road assessments emphasised air quality and noise risks, concluding these were minimal or predicted to improve. The South Australian project was the only road project not to include health data explicitly. The light rail EIS considered the health benefits of the project whereas the others focused on risk. Only one project considered mental health, although in less detail than air quality or noise. Our findings suggest EIAs lag behind the known evidence linking transport infrastructure to health. If health is to be comprehensively included, a more complete model of health is required, as well as a shift away from health risk assessment as the main method used. This needs to be mandatory for all significant developments. We also found that considering health only at the EIA stage may be a significant limitation, and there is a need for health issues to be considered when earlier, fundamental decisions about the project are being made. © 2018 The Author(s); Published by Kerman University of Medical Sciences. This is an open-access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
A Brief Review of the Current Status of Alternatives of Chlorine Disinfection of Water
Anderson, A. C.; Reimers, R. S.; DeKernion, P.
1982-01-01
This paper briefly outlines some of the alternative disinfectants being considered in lieu of chlorination. Methods currently in use as well as those in the research stage are included. Each method is assessed with respect to disinfection efficiency and environmental impact. (Am J Public Health 1982; 72:1290-1293.) PMID:7125035
Fruit and vegetable intake of U.S. adults by two methods: What We Eat in America, NHANES 2009-2012
USDA-ARS?s Scientific Manuscript database
Estimates of fruit and vegetable (FV) consumption vary depending on how intake is defined and the research purpose. Two methods used by researchers to evaluate intake include epidemiological (EPI), which considers all FV defined as cup equivalents (CE) after disaggregating foods and beverages, and b...
Code of Federal Regulations, 2010 CFR
2010-04-01
... this paragraph (b)(2) include the following— (i) A mathematical error; (ii) An entry on a document that... errors or omissions that occurred before the publication of these regulations. Any reasonable method used... February 24, 1994, will be considered proper, provided that the method is consistent with the rules of...
The Minerva Multi-Microprocessor.
A multiprocessor system is described which is an experiment in low cost, extensible, multiprocessor architectures. Global issues such as inclusion of a central bus, design of the bus arbiter, and methods of interrupt handling are considered. The system initially includes two processor types, based on microprocessors, and these are discussed. Methods for reducing processor demand for the central bus are described.
Tiedeman, C.R.; Hill, M.C.; D'Agnese, F. A.; Faunt, C.C.
2003-01-01
Calibrated models of groundwater systems can provide substantial information for guiding data collection. This work considers using such models to guide hydrogeologic data collection for improving model predictions by identifying model parameters that are most important to the predictions. Identification of these important parameters can help guide collection of field data about parameter values and associated flow system features and can lead to improved predictions. Methods for identifying parameters important to predictions include prediction scaled sensitivities (PSS), which account for uncertainty on individual parameters as well as prediction sensitivity to parameters, and a new "value of improved information" (VOII) method presented here, which includes the effects of parameter correlation in addition to individual parameter uncertainty and prediction sensitivity. In this work, the PSS and VOII methods are demonstrated and evaluated using a model of the Death Valley regional groundwater flow system. The predictions of interest are advective transport paths originating at sites of past underground nuclear testing. Results show that for two paths evaluated the most important parameters include a subset of five or six of the 23 defined model parameters. Some of the parameters identified as most important are associated with flow system attributes that do not lie in the immediate vicinity of the paths. Results also indicate that the PSS and VOII methods can identify different important parameters. Because the methods emphasize somewhat different criteria for parameter importance, it is suggested that parameters identified by both methods be carefully considered in subsequent data collection efforts aimed at improving model predictions.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kerner, Ryan; Mann, R.B.
We investigate quantum tunnelling methods for calculating black hole temperature, specifically the null-geodesic method of Parikh and Wilczek and the Hamilton-Jacobi Ansatz method of Angheben et al. We consider application of these methods to a broad class of spacetimes with event horizons, including Rindler and nonstatic spacetimes such as Kerr-Newman and Taub-NUT. We obtain a general form for the temperature of Taub-NUT-AdS black holes that is commensurate with other methods. We examine the limitations of these methods for extremal black holes, taking the extremal Reissner-Nordstrom spacetime as a case in point.
A discontinuous Galerkin method for poroelastic wave propagation: The two-dimensional case
NASA Astrophysics Data System (ADS)
Dudley Ward, N. F.; Lähivaara, T.; Eveson, S.
2017-12-01
In this paper, we consider a high-order discontinuous Galerkin (DG) method for modelling wave propagation in coupled poroelastic-elastic media. The upwind numerical flux is derived as an exact solution for the Riemann problem including the poroelastic-elastic interface. Attenuation mechanisms in both Biot's low- and high-frequency regimes are considered. The current implementation supports non-uniform basis orders which can be used to control the numerical accuracy element by element. In the numerical examples, we study the convergence properties of the proposed DG scheme and provide experiments where the numerical accuracy of the scheme under consideration is compared to analytic and other numerical solutions.
Exposure assessment for endocrine disruptors: some considerations in the design of studies.
Rice, Carol; Birnbaum, Linda S; Cogliano, James; Mahaffey, Kathryn; Needham, Larry; Rogan, Walter J; vom Saal, Frederick S
2003-01-01
In studies designed to evaluate exposure-response relationships in children's development from conception through puberty, multiple factors that affect the generation of meaningful exposure metrics must be considered. These factors include multiple routes of exposure; the timing, frequency, and duration of exposure; need for qualitative and quantitative data; sample collection and storage protocols; and the selection and documentation of analytic methods. The methods for exposure data collection and analysis must be sufficiently robust to accommodate the a priori hypotheses to be tested, as well as hypotheses generated from the data. A number of issues that must be considered in study design are summarized here. PMID:14527851
Voltera's Solution of the Wave Equation as Applied to Three-Dimensional Supersonic Airfoil Problems
NASA Technical Reports Server (NTRS)
Heslet, Max A; Lomax, Harvard; Jones, Arthur L
1947-01-01
A surface integral is developed which yields solutions of the linearized partial differential equation for supersonic flow. These solutions satisfy boundary conditions arising in wing theory. Particular applications of this general method are made, using acceleration potentials, to flat surfaces and to uniformly loaded lifting surfaces. Rectangular and trapezoidal plan forms are considered along with triangular forms adaptable to swept-forward and swept-back wings. The case of the triangular plan form in sideslip is also included. Emphasis is placed on the systematic application of the method to the lifting surfaces considered and on the possibility of further application.
ERIC Educational Resources Information Center
Fisher, William; Koue, Glen
1991-01-01
Discusses general issues involved in conflict management and provides more specific examples of conflict management in libraries. Causes of conflict are considered, including organizational structure, departmentalization, performance appraisal, poor communication, and technological change; and methods of dealing with conflict are described,…
NASA Astrophysics Data System (ADS)
Rozanov, Vladimir V.; Vountas, Marco
2014-01-01
Rotational Raman scattering of solar light in Earth's atmosphere leads to the filling-in of Fraunhofer and telluric lines observed in the reflected spectrum. The phenomenological derivation of the inelastic radiative transfer equation including rotational Raman scattering is presented. The different forms of the approximate radiative transfer equation with first-order rotational Raman scattering terms are obtained employing the Cabannes, Rayleigh, and Cabannes-Rayleigh scattering models. The solution of these equations is considered in the framework of the discrete-ordinates method using rigorous and approximate approaches to derive particular integrals. An alternative forward-adjoint technique is suggested as well. A detailed description of the model including the exact spectral matching and a binning scheme that significantly speeds up the calculations is given. The considered solution techniques are implemented in the radiative transfer software package SCIATRAN and a specified benchmark setup is presented to enable readers to compare with own results transparently.
Procedures for making gaseous industrial waste safe
NASA Astrophysics Data System (ADS)
Matros, Yu Sh; Noskov, Aleksandr S.
1990-10-01
The application of various methods (adsorption, absorption, thermal afterburning, catalytic purification, and others) for the removal of sulphur and nitrogen oxides, toxic organic compounds, hydrogen sulphide, and carbon monoxide from industrial waste gases is described. Much attention is devoted to the catalytic procedure for making the gases safe using an energy collecting non-stationary method (reversible process). The advantages and limitations of various gas purification methods are considered. The bibliography includes 279 references.
Accounting for ecosystem services in life cycle assessment, Part I: a critical review.
Zhang, Yi; Singh, Shweta; Bakshi, Bhavik R
2010-04-01
If life cycle oriented methods are to encourage sustainable development, they must account for the role of ecosystem goods and services, since these form the basis of planetary activities and human well-being. This article reviews methods that are relevant to accounting for the role of nature and that could be integrated into life cycle oriented approaches. These include methods developed by ecologists for quantifying ecosystem services, by ecological economists for monetary valuation, and life cycle methods such as conventional life cycle assessment, thermodynamic methods for resource accounting such as exergy and emergy analysis, variations of the ecological footprint approach, and human appropriation of net primary productivity. Each approach has its strengths: economic methods are able to quantify the value of cultural services; LCA considers emissions and assesses their impact; emergy accounts for supporting services in terms of cumulative exergy; and ecological footprint is intuitively appealing and considers biocapacity. However, no method is able to consider all the ecosystem services, often due to the desire to aggregate all resources in terms of a single unit. This review shows that comprehensive accounting for ecosystem services in LCA requires greater integration among existing methods, hierarchical schemes for interpreting results via multiple levels of aggregation, and greater understanding of the role of ecosystems in supporting human activities. These present many research opportunities that must be addressed to meet the challenges of sustainability.
X-ray and synchrotron methods in studies of cultural heritage sites
DOE Office of Scientific and Technical Information (OSTI.GOV)
Koval’chuk, M. V.; Yatsishina, E. B.; Blagov, A. E.
2016-09-15
X-ray and synchrotron methods that are most widely used in studies of cultural heritage objects (including archaeological sites)—X-ray diffraction analysis, X-ray spectroscopy, and visualization techniques— have been considered. The reported examples show high efficiency and informativeness of natural science studies when solving most diverse problems of archaeology, history, the study of art, museology, etc.
ERIC Educational Resources Information Center
Bird, Jo; Colliver, Yeshe; Edwards, Susan
2014-01-01
Participatory research methods argue that young children should be enabled to contribute their perspectives on research seeking to understand their worldviews. Visual research methods, including the use of still and video cameras with young children have been viewed as particularly suited to this aim because cameras have been considered easy and…
Hage, David S.; Anguizola, Jeanethe A.; Bi, Cong; Li, Rong; Matsuda, Ryan; Papastavros, Efthimia; Pfaunmiller, Erika; Vargas, John; Zheng, Xiwei
2012-01-01
Affinity chromatography is a separation technique that has become increasingly important in work with biological samples and pharmaceutical agents. This method is based on the use of a biologically-related agent as a stationary phase to selectively retain analytes or to study biological interactions. This review discusses the basic principles behind affinity chromatography and examines recent developments that have occurred in the use of this method for biomedical and pharmaceutical analysis. Techniques based on traditional affinity supports are discussed, but an emphasis is placed on methods in which affinity columns are used as part of HPLC systems or in combination with other analytical methods. General formats for affinity chromatography that are considered include step elution schemes, weak affinity chromatography, affinity extraction and affinity depletion. Specific separation techniques that are examined include lectin affinity chromatography, boronate affinity chromatography, immunoaffinity chromatography, and immobilized metal ion affinity chromatography. Approaches for the study of biological interactions by affinity chromatography are also presented, such as the measurement of equilibrium constants, rate constants, or competition and displacement effects. In addition, related developments in the use of immobilized enzyme reactors, molecularly imprinted polymers, dye ligands and aptamers are briefly considered. PMID:22305083
NASA Astrophysics Data System (ADS)
Yun, Dong-Un; Lee, Sang-Kwon
2017-06-01
In this paper, we present a novel method for an objective evaluation of knocking noise emitted by diesel engines based on the temporal and frequency masking theory. The knocking sound of a diesel engine is a vibro-acoustic sound correlated with the high-frequency resonances of the engine structure and a periodic impulsive sound with amplitude modulation. Its period is related to the engine speed and includes specific frequency bands related to the resonances of the engine structure. A knocking sound with the characteristics of a high-frequency impulsive wave can be masked by low-frequency sounds correlated with the harmonics of the firing frequency and broadband noise. The degree of modulation of the knocking sound signal was used for such objective evaluations in previous studies, without considering the masking effect. However, the frequency masking effect must be considered for the objective evaluation of the knocking sound. In addition to the frequency masking effect, the temporal masking effect occurs because the period of the knocking sound changes according to the engine speed. Therefore, an evaluation method considering the temporal and frequency masking effect is required to analyze the knocking sound objectively. In this study, an objective evaluation method considering the masking effect was developed based on the masking theory of sound and signal processing techniques. The method was applied successfully for the objective evaluation of the knocking sound of a diesel engine.
Horton, Bethany Jablonski; Wages, Nolan A.; Conaway, Mark R.
2016-01-01
Toxicity probability interval designs have received increasing attention as a dose-finding method in recent years. In this study, we compared the two-stage, likelihood-based continual reassessment method (CRM), modified toxicity probability interval (mTPI), and the Bayesian optimal interval design (BOIN) in order to evaluate each method's performance in dose selection for Phase I trials. We use several summary measures to compare the performance of these methods, including percentage of correct selection (PCS) of the true maximum tolerable dose (MTD), allocation of patients to doses at and around the true MTD, and an accuracy index. This index is an efficiency measure that describes the entire distribution of MTD selection and patient allocation by taking into account the distance between the true probability of toxicity at each dose level and the target toxicity rate. The simulation study considered a broad range of toxicity curves and various sample sizes. When considering PCS, we found that CRM outperformed the two competing methods in most scenarios, followed by BOIN, then mTPI. We observed a similar trend when considering the accuracy index for dose allocation, where CRM most often outperformed both the mTPI and BOIN. These trends were more pronounced with increasing number of dose levels. PMID:27435150
A study of two statistical methods as applied to shuttle solid rocket booster expenditures
NASA Technical Reports Server (NTRS)
Perlmutter, M.; Huang, Y.; Graves, M.
1974-01-01
The state probability technique and the Monte Carlo technique are applied to finding shuttle solid rocket booster expenditure statistics. For a given attrition rate per launch, the probable number of boosters needed for a given mission of 440 launches is calculated. Several cases are considered, including the elimination of the booster after a maximum of 20 consecutive launches. Also considered is the case where the booster is composed of replaceable components with independent attrition rates. A simple cost analysis is carried out to indicate the number of boosters to build initially, depending on booster costs. Two statistical methods were applied in the analysis: (1) state probability method which consists of defining an appropriate state space for the outcome of the random trials, and (2) model simulation method or the Monte Carlo technique. It was found that the model simulation method was easier to formulate while the state probability method required less computing time and was more accurate.
Performing skin microbiome research: A method to the madness
Kong, Heidi H.; Andersson, Björn; Clavel, Thomas; Common, John E.; Jackson, Scott A.; Olson, Nathan D.; Segre, Julia A.; Traidl-Hoffmann, Claudia
2017-01-01
Growing interest in microbial contributions to human health and disease has increasingly led investigators to examine the microbiome in both healthy skin and cutaneous disorders, including acne, psoriasis and atopic dermatitis. The need for common language, effective study design, and validated methods are critical for high-quality, standardized research. Features, unique to skin, pose particular challenges when conducting microbiome research. This review discusses microbiome research standards and highlights important factors to consider, including clinical study design, skin sampling, sample processing, DNA sequencing, control inclusion, and data analysis. PMID:28063650
Application of Density Functional Theory to Systems Containing Metal Atoms
NASA Technical Reports Server (NTRS)
Bauschlicher, Charles W., Jr.
2006-01-01
The accuracy of density functional theory (DFT) for problems involving metal atoms is considered. The DFT results are compared with experiment as well as results obtained using the coupled cluster approach. The comparisons include geometries, frequencies, and bond energies. The systems considered include MO2, M(OH)+n, MNO+, and MCO+2. The DFT works well for frequencies and geometries, even in case with symmetry breaking; however, some examples have been found where the symmetry breaking is quite severe and the DFT methods do not work well. The calculation of bond energies is more difficult and examples of successes as well as failures of DFT will be given.
Defining Happiness for Young Adults with Schizophrenia: A Building Block for Recovery
Buckland, Helen T.; Schepp, Karen G.; Crusoe, Kristen
2013-01-01
Purpose Findings from this mixed methods descriptive study include a definition of happiness for young adults with schizophrenia (SCZ). Methods Thirteen men and women, ages 23 to 35, completed a series of three individual one-hour interviews over six weeks. Results This definition included themes of material happiness, relational happiness, and health happiness. Although these themes correspond to those for young adults without SCZ, four barriers to happiness were identified for this vulnerable population: 1) fear, 2) isolation, 3) medication, 4) not being considered “normal.” Conclusion Implications of these findings include designing interventions to support recovery by overcoming barriers to happiness. PMID:24070992
Equipment Decisions: Micros, Terminals, or Typewriters.
ERIC Educational Resources Information Center
Swanson, Jean C.
1986-01-01
Discusses factors to be considered when deciding whether to buy typewriters, terminals, or microcomputers for the school typing room. Factors include (1) objectives of the typewriting program, (2) market needs, (3) student learning methods, (4) costs, and (5) instructional materials. (CH)
NASA Technical Reports Server (NTRS)
Anderton, D. A.
1982-01-01
Aeronautical research programs are discussed in relation to research methods and the status of the programs. The energy efficient aircraft, STOL aircraft and general aviation aircraft are considered. Aerodynamic concepts, rotary wing aircraft, aircraft safety, noise reduction, and aircraft configurations are among the topics included.
Challenges in recovering resources from acid mine drainage
Nordstrom, D. Kirk; Bowell, Robert J.; Campbell, Kate M.; Alpers, Charles N.
2017-01-01
Metal recovery from mine waters and effluents is not a new approach but one that has occurred largely opportunistically over the last four millennia. Due to the need for low-cost resources and increasingly stringent environmental conditions, mine waters are being considered in a fresh light with a designed, deliberate approach to resource recovery often as part of a larger water treatment evaluation. Mine water chemistry is highly dependent on many factors including geology, ore deposit composition and mineralogy, mining methods, climate, site hydrology, and others. Mine waters are typically Ca-Mg-SO4±Al±Fe with a broad range in pH and metal content. The main issue in recovering components of these waters having potential economic value, such as base metals or rare earth elements, is the separation of these from more reactive metals such as Fe and Al. Broad categories of methods for separating and extracting substances from acidic mine drainage are chemical and biological. Chemical methods include solution, physicochemical, and electrochemical technologies. Advances in membrane techniques such as reverse osmosis have been substantial and the technique is both physical and chemical. Biological methods may be further divided into microbiological and macrobiological, but only the former is considered here as a recovery method, as the latter is typically used as a passive form of water treatment.
Quantifying Groundwater Model Uncertainty
NASA Astrophysics Data System (ADS)
Hill, M. C.; Poeter, E.; Foglia, L.
2007-12-01
Groundwater models are characterized by the (a) processes simulated, (b) boundary conditions, (c) initial conditions, (d) method of solving the equation, (e) parameterization, and (f) parameter values. Models are related to the system of concern using data, some of which form the basis of observations used most directly, through objective functions, to estimate parameter values. Here we consider situations in which parameter values are determined by minimizing an objective function. Other methods of model development are not considered because their ad hoc nature generally prohibits clear quantification of uncertainty. Quantifying prediction uncertainty ideally includes contributions from (a) to (f). The parameter values of (f) tend to be continuous with respect to both the simulated equivalents of the observations and the predictions, while many aspects of (a) through (e) are discrete. This fundamental difference means that there are options for evaluating the uncertainty related to parameter values that generally do not exist for other aspects of a model. While the methods available for (a) to (e) can be used for the parameter values (f), the inferential methods uniquely available for (f) generally are less computationally intensive and often can be used to considerable advantage. However, inferential approaches require calculation of sensitivities. Whether the numerical accuracy and stability of the model solution required for accurate sensitivities is more broadly important to other model uses is an issue that needs to be addressed. Alternative global methods can require 100 or even 1,000 times the number of runs needed by inferential methods, though methods of reducing the number of needed runs are being developed and tested. Here we present three approaches for quantifying model uncertainty and investigate their strengths and weaknesses. (1) Represent more aspects as parameters so that the computationally efficient methods can be broadly applied. This approach is attainable through universal model analysis software such as UCODE-2005, PEST, and joint use of these programs, which allow many aspects of a model to be defined as parameters. (2) Use highly parameterized models to quantify aspects of (e). While promising, this approach implicitly includes parameterizations that may be considered unreasonable if investigated explicitly, so that resulting measures of uncertainty may be too large. (3) Use a combination of inferential and global methods that can be facilitated using the new software MMA (Multi-Model Analysis), which is constructed using the JUPITER API. Here we consider issues related to the model discrimination criteria calculated by MMA.
Lagishetty, Chakradhar V; Duffull, Stephen B
2015-11-01
Clinical studies include occurrences of rare variables, like genotypes, which due to their frequency and strength render their effects difficult to estimate from a dataset. Variables that influence the estimated value of a model-based parameter are termed covariates. It is often difficult to determine if such an effect is significant, since type I error can be inflated when the covariate is rare. Their presence may have either an insubstantial effect on the parameters of interest, hence are ignorable, or conversely they may be influential and therefore non-ignorable. In the case that these covariate effects cannot be estimated due to power and are non-ignorable, then these are considered nuisance, in that they have to be considered but due to type 1 error are of limited interest. This study assesses methods of handling nuisance covariate effects. The specific objectives include (1) calibrating the frequency of a covariate that is associated with type 1 error inflation, (2) calibrating its strength that renders it non-ignorable and (3) evaluating methods for handling these non-ignorable covariates in a nonlinear mixed effects model setting. Type 1 error was determined for the Wald test. Methods considered for handling the nuisance covariate effects were case deletion, Box-Cox transformation and inclusion of a specific fixed effects parameter. Non-ignorable nuisance covariates were found to be effectively handled through addition of a fixed effect parameter.
NASA Astrophysics Data System (ADS)
Botti, Lorenzo; Di Pietro, Daniele A.
2018-10-01
We propose and validate a novel extension of Hybrid High-Order (HHO) methods to meshes featuring curved elements. HHO methods are based on discrete unknowns that are broken polynomials on the mesh and its skeleton. We propose here the use of physical frame polynomials over mesh elements and reference frame polynomials over mesh faces. With this choice, the degree of face unknowns must be suitably selected in order to recover on curved meshes the same convergence rates as on straight meshes. We provide an estimate of the optimal face polynomial degree depending on the element polynomial degree and on the so-called effective mapping order. The estimate is numerically validated through specifically crafted numerical tests. All test cases are conducted considering two- and three-dimensional pure diffusion problems, and include comparisons with discontinuous Galerkin discretizations. The extension to agglomerated meshes with curved boundaries is also considered.
NASA Astrophysics Data System (ADS)
Pianosi, Francesca; Lal Shrestha, Durga; Solomatine, Dimitri
2010-05-01
This research presents an extension of UNEEC (Uncertainty Estimation based on Local Errors and Clustering, Shrestha and Solomatine, 2006, 2008 & Solomatine and Shrestha, 2009) method in the direction of explicit inclusion of parameter uncertainty. UNEEC method assumes that there is an optimal model and the residuals of the model can be used to assess the uncertainty of the model prediction. It is assumed that all sources of uncertainty including input, parameter and model structure uncertainty are explicitly manifested in the model residuals. In this research, theses assumptions are relaxed, and the UNEEC method is extended to consider parameter uncertainty as well (abbreviated as UNEEC-P). In UNEEC-P, first we use Monte Carlo (MC) sampling in parameter space to generate N model realizations (each of which is a time series), estimate the prediction quantiles based on the empirical distribution functions of the model residuals considering all the residual realizations, and only then apply the standard UNEEC method that encapsulates the uncertainty of a hydrologic model (expressed by quantiles of the error distribution) in a machine learning model (e.g., ANN). UNEEC-P is applied first to a linear regression model of synthetic data, and then to a real case study of forecasting inflow to lake Lugano in northern Italy. The inflow forecasting model is a stochastic heteroscedastic model (Pianosi and Soncini-Sessa, 2009). The preliminary results show that the UNEEC-P method produces wider uncertainty bounds, which is consistent with the fact that the method considers also parameter uncertainty of the optimal model. In the future UNEEC method will be further extended to consider input and structure uncertainty which will provide more realistic estimation of model predictions.
NASA Technical Reports Server (NTRS)
Oliger, Joseph
1997-01-01
Topics considered include: high-performance computing; cognitive and perceptual prostheses (computational aids designed to leverage human abilities); autonomous systems. Also included: development of a 3D unstructured grid code based on a finite volume formulation and applied to the Navier-stokes equations; Cartesian grid methods for complex geometry; multigrid methods for solving elliptic problems on unstructured grids; algebraic non-overlapping domain decomposition methods for compressible fluid flow problems on unstructured meshes; numerical methods for the compressible navier-stokes equations with application to aerodynamic flows; research in aerodynamic shape optimization; S-HARP: a parallel dynamic spectral partitioner; numerical schemes for the Hamilton-Jacobi and level set equations on triangulated domains; application of high-order shock capturing schemes to direct simulation of turbulence; multicast technology; network testbeds; supercomputer consolidation project.
Interactive Performance and Focus Groups with Adolescents: The Power of Play
Norris, Anne E.; Aroian, Karen J.; Warren, Stefanie
2012-01-01
Conducting focus groups with adolescents can be challenging given their developmental needs, particularly with sensitive topics. These challenges include intense need for peer approval, declining social trust, short attention span, and reliance on concrete operations thinking. In this article we describe an adaptation of interactive performance as an alternative to traditional focus group method. We used this method in a study of discrimination experienced by Muslims (ages 13-17) and of peer pressure to engage in sexual behavior experienced by Hispanic girls (ages 10-14). Recommendations for use of this method include using an interdisciplinary team, planning for large amounts of disclosure towards the end of the focus group, and considering the fit of this method to the study topic. PMID:22949032
Code of Federal Regulations, 2011 CFR
2011-10-01
... of public lands under this section shall be conducted under competitive bidding procedures... competitive bidding or without competitive bidding. In recognizing public policies, the Secretary shall give... considered in determining the method of sale. These factors include, but are not limited to: Competitive...
Holographic analysis as an inspection method for welded thin-wall tubing
NASA Technical Reports Server (NTRS)
Brooks, Lawrence; Mulholland, John; Genin, Joseph; Matthews, Larryl
1990-01-01
The feasibility of using holographic interferometry for locating flaws in welded tubing is explored. Two holographic techniques are considered: traditional holographic interferometry and electronic speckle pattern interferometry. Several flaws including cold laps, discontinuities, and tube misalignments are detected.
Reviews the issues and approaches involved in considering and adopting cost-effectiveness tests for energy efficiency, including discussing each perspective represented by the five standard cost-effectiveness tests and clarifying key terms.
Developments in Sampling and Analysis Instrumentation for Stationary Sources
ERIC Educational Resources Information Center
Nader, John S.
1973-01-01
Instrumentation for the measurement of pollutant emissions is considered including sample-site selection, sample transport, sample treatment, sample analysis, and data reduction, display, and interpretation. Measurement approaches discussed involve sample extraction from within the stack and electro-optical methods. (BL)
Reduction of chemical reaction models
NASA Technical Reports Server (NTRS)
Frenklach, Michael
1991-01-01
An attempt is made to reconcile the different terminologies pertaining to reduction of chemical reaction models. The approaches considered include global modeling, response modeling, detailed reduction, chemical lumping, and statistical lumping. The advantages and drawbacks of each of these methods are pointed out.
Alternatives to Piloting Textbooks.
ERIC Educational Resources Information Center
Muther, Connie
1985-01-01
Using short-term pilot programs to evaluate textbooks can lead to unreliable results and interfere with effective education. Alternative methods for evaluating textbook-based programs include obtaining documented analyses of competitors' products from sales agents, visiting districts using programs being considered, and examining publishers' own…
A framework for the social valuation of ecosystem services.
Felipe-Lucia, María R; Comín, Francisco A; Escalera-Reyes, Javier
2015-05-01
Methods to assess ecosystem services using ecological or economic approaches are considerably better defined than methods for the social approach. To identify why the social approach remains unclear, we reviewed current trends in the literature. We found two main reasons: (i) the cultural ecosystem services are usually used to represent the whole social approach, and (ii) the economic valuation based on social preferences is typically included in the social approach. Next, we proposed a framework for the social valuation of ecosystem services that provides alternatives to economics methods, enables comparison across studies, and supports decision-making in land planning and management. The framework includes the agreements emerged from the review, such as considering spatial-temporal flows, including stakeholders from all social ranges, and using two complementary methods to value ecosystem services. Finally, we provided practical recommendations learned from the application of the proposed framework in a case study.
NASA Astrophysics Data System (ADS)
Begunov, Roman S.; Ryzvanovich, Galina A.
2013-01-01
Main methods for the synthesis of fused imidazole derivatives with a bridgehead nitrogen atom are systematically considered and summarized. The reaction mechanisms that underlie the methods for the synthesis of pyrido[1,2-a]benzimidazoles and related compounds are described. Biological properties and mechanisms of the biological activity of fused azaheterocycles are discussed. The bibliography includes 152 references.
Hong, Kyong Hee
2014-01-01
Colonoscopy is a widely accepted method for the evaluation of the colon and terminal ileum. Its diagnostic accuracy and therapeutic safety are influenced by prerequisites, including modulation of medication and bowel cleansing. Appropriate choices of sedative medication and bowel-cleansing regimen, together with diet modification, should be made based on the patient's underlying disease, age, and medication intake. Moreover, effective methods for patient education regarding bowel preparation should be considered. PMID:25133119
Modification of polymers by polymeric additives
NASA Astrophysics Data System (ADS)
Nesterov, A. E.; Lebedev, E. V.
1989-08-01
The conditions for the thermodynamic compatibility of polymers and methods for its enhancement are examined. The study of the influence of various factors on the concentration-temperature limits of compatibility, dispersion stabilisation processes, and methods for the improvement of adhesion between phases in mixtures of thermodynamically incompatible polymers is described. Questions concerning the improvement of the physicomechanical characteristics of polymer dispersions are considered. The bibliography includes 200 references.
ERIC Educational Resources Information Center
Philippines Univ., Quezon City. Science Education Center.
This module discusses methods of obtaining table salt from seawater. Topic areas considered include: (1) obtaining salt by solar evaporation of seawater in holes; (2) obtaining salt by boiling seawater in pots; (3) how table salt is obtained from seawater in the Philippines; and (4) methods of making salt by solar evaporation of seawater in the…
A text zero-watermarking method based on keyword dense interval
NASA Astrophysics Data System (ADS)
Yang, Fan; Zhu, Yuesheng; Jiang, Yifeng; Qing, Yin
2017-07-01
Digital watermarking has been recognized as a useful technology for the copyright protection and authentication of digital information. However, rarely did the former methods focus on the key content of digital carrier. The idea based on the protection of key content is more targeted and can be considered in different digital information, including text, image and video. In this paper, we use text as research object and a text zero-watermarking method which uses keyword dense interval (KDI) as the key content is proposed. First, we construct zero-watermarking model by introducing the concept of KDI and giving the method of KDI extraction. Second, we design detection model which includes secondary generation of zero-watermark and the similarity computing method of keyword distribution. Besides, experiments are carried out, and the results show that the proposed method gives better performance than other available methods especially in the attacks of sentence transformation and synonyms substitution.
Saito, Maiko; Kurosawa, Yae; Okuyama, Tsuneo
2012-02-01
Antibody purification using proteins A and G has been a standard method for research and industrial processes. The conventional method, however, includes a three-step process, including buffer exchange, before chromatography. In addition, proteins A and G require low pH elution, which causes antibody aggregation and inactivates the antibody's immunity. This report proposes a two-step method using hydroxyapatite chromatography and membrane filtration, without proteins A and G. This novel method shortens the running time to one-third the conventional method for each cycle. Using our two-step method, 90.2% of the monoclonal antibodies purified were recovered in the elution fraction, the purity achieved was >90%, and most of the antigen-specific activity was retained. This report suggests that the two-step method using hydroxyapatite chromatography and membrane filtration should be considered as an alternative to purification using proteins A and G.
Statistical methods used in articles published by the Journal of Periodontal and Implant Science.
Choi, Eunsil; Lyu, Jiyoung; Park, Jinyoung; Kim, Hae-Young
2014-12-01
The purposes of this study were to assess the trend of use of statistical methods including parametric and nonparametric methods and to evaluate the use of complex statistical methodology in recent periodontal studies. This study analyzed 123 articles published in the Journal of Periodontal & Implant Science (JPIS) between 2010 and 2014. Frequencies and percentages were calculated according to the number of statistical methods used, the type of statistical method applied, and the type of statistical software used. Most of the published articles considered (64.4%) used statistical methods. Since 2011, the percentage of JPIS articles using statistics has increased. On the basis of multiple counting, we found that the percentage of studies in JPIS using parametric methods was 61.1%. Further, complex statistical methods were applied in only 6 of the published studies (5.0%), and nonparametric statistical methods were applied in 77 of the published studies (38.9% of a total of 198 studies considered). We found an increasing trend towards the application of statistical methods and nonparametric methods in recent periodontal studies and thus, concluded that increased use of complex statistical methodology might be preferred by the researchers in the fields of study covered by JPIS.
Data mining for signals in spontaneous reporting databases: proceed with caution.
Stephenson, Wendy P; Hauben, Manfred
2007-04-01
To provide commentary and points of caution to consider before incorporating data mining as a routine component of any Pharmacovigilance program, and to stimulate further research aimed at better defining the predictive value of these new tools as well as their incremental value as an adjunct to traditional methods of post-marketing surveillance. Commentary includes review of current data mining methodologies employed and their limitations, caveats to consider in the use of spontaneous reporting databases and caution against over-confidence in the results of data mining. Future research should focus on more clearly delineating the limitations of the various quantitative approaches as well as the incremental value that they bring to traditional methods of pharmacovigilance.
The History of a Decision: A Standard Vibration Test Method for Qualification
Rizzo, Davinia; Blackburn, Mark
2017-01-01
As Mil-Std-810G and subsequent versions have included multiple degree of freedom vibration test methodologies, it is important to understand the history and factors that drove the original decision in Mil-Std-810 to focus on single degree of freedom (SDOF) vibration testing. By assessing the factors and thought process of early Mil-Std-810 vibration test methods, it enables one to better consider the use of multiple degree of freedom testing now that it is feasible with today’s technology and documented in Mil-Std-810. This paper delves into the details of the decision made in the 1960s for the SDOF vibration testing standards in Mil-Std-810more » beyond the limitations of technology at the time. We also consider the implications for effective test planning today considering the advances in test capabilities and improvements in understanding of the operational environment.« less
The History of a Decision: A Standard Vibration Test Method for Qualification
DOE Office of Scientific and Technical Information (OSTI.GOV)
Rizzo, Davinia; Blackburn, Mark
As Mil-Std-810G and subsequent versions have included multiple degree of freedom vibration test methodologies, it is important to understand the history and factors that drove the original decision in Mil-Std-810 to focus on single degree of freedom (SDOF) vibration testing. By assessing the factors and thought process of early Mil-Std-810 vibration test methods, it enables one to better consider the use of multiple degree of freedom testing now that it is feasible with today’s technology and documented in Mil-Std-810. This paper delves into the details of the decision made in the 1960s for the SDOF vibration testing standards in Mil-Std-810more » beyond the limitations of technology at the time. We also consider the implications for effective test planning today considering the advances in test capabilities and improvements in understanding of the operational environment.« less
A Comparison of Trajectory Optimization Methods for the Impulsive Minimum Fuel Rendezvous Problem
NASA Technical Reports Server (NTRS)
Hughes, Steven P.; Mailhe, Laurie M.; Guzman, Jose J.
2003-01-01
In this paper we present, a comparison of trajectory optimization approaches for the minimum fuel rendezvous problem. Both indirect and direct methods are compared for a variety of test cases. The indirect approach is based on primer vector theory. The direct approaches are implemented numerically and include Sequential Quadratic Programming (SQP). Quasi- Newton and Nelder-Meade Simplex. Several cost function parameterizations are considered for the direct approach. We choose one direct approach that appears to be the most flexible. Both the direct and indirect methods are applied to a variety of test cases which are chosen to demonstrate the performance of each method in different flight regimes. The first test case is a simple circular-to-circular coplanar rendezvous. The second test case is an elliptic-to-elliptic line of apsides rotation. The final test case is an orbit phasing maneuver sequence in a highly elliptic orbit. For each test case we present a comparison of the performance of all methods we consider in this paper.
An annular superposition integral for axisymmetric radiators
Kelly, James F.; McGough, Robert J.
2007-01-01
A fast integral expression for computing the nearfield pressure is derived for axisymmetric radiators. This method replaces the sum of contributions from concentric annuli with an exact double integral that converges much faster than methods that evaluate the Rayleigh-Sommerfeld integral or the generalized King integral. Expressions are derived for plane circular pistons using both continuous wave and pulsed excitations. Several commonly used apodization schemes for the surface velocity distribution are considered, including polynomial functions and a “smooth piston” function. The effect of different apodization functions on the spectral content of the wave field is explored. Quantitative error and time comparisons between the new method, the Rayleigh-Sommerfeld integral, and the generalized King integral are discussed. At all error levels considered, the annular superposition method achieves a speed-up of at least a factor of 4 relative to the point-source method and a factor of 3 relative to the generalized King integral without increasing the computational complexity. PMID:17348500
DOE Office of Scientific and Technical Information (OSTI.GOV)
Groth, R.H.; Calabro, D.S.
1969-11-01
The two methods normally used for the analysis of NO/sub x/ are the Saltzman and the phenoldisulfonic acid technique. This paper describes an evaluation of these wet chemical methods to determine their practical application to engine exhaust gas analysis. Parameters considered for the Saltzman method included bubbler collection efficiency, NO to NO/sub 2/ conversion efficiency, masking effect of other contaminants usually present in exhaust gases and the time-temperature effect of these contaminants on store developed solutions. Collection efficiency and the effects of contaminants were also considered for the phenoldisulfonic acid method. Test results indicated satisfactory collection and conversion efficiencies formore » the Saltzman method, but contaminants seriously affected the measurement accuracy particularly if the developed solution was stored for a number of hours at room temperature before analysis. Storage at 32/sup 0/F minimized effect. The standard procedure for the phenoldisulfonic acid method gave good results, but the process was found to be too time consuming for routine analysis and measured only total NO/sub x/. 3 references, 9 tables.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Marxen, Olaf, E-mail: olaf.marxen@vki.ac.be; Aeronautics and Aerospace Department, von Karman Institute for Fluid Dynamics, Chaussée de Waterloo, 72, 1640 Rhode-St-Genèse; Magin, Thierry E.
2013-12-15
A new numerical method is presented here that allows to consider chemically reacting gases during the direct numerical simulation of a hypersonic fluid flow. The method comprises the direct coupling of a solver for the fluid mechanical model and a library providing the physio-chemical model. The numerical method for the fluid mechanical model integrates the compressible Navier–Stokes equations using an explicit time advancement scheme and high-order finite differences. This Navier–Stokes code can be applied to the investigation of laminar-turbulent transition and boundary-layer instability. The numerical method for the physio-chemical model provides thermodynamic and transport properties for different gases as wellmore » as chemical production rates, while here we exclusively consider a five species air mixture. The new method is verified for a number of test cases at Mach 10, including the one-dimensional high-temperature flow downstream of a normal shock, a hypersonic chemical reacting boundary layer in local thermodynamic equilibrium and a hypersonic reacting boundary layer with finite-rate chemistry. We are able to confirm that the diffusion flux plays an important role for a high-temperature boundary layer in local thermodynamic equilibrium. Moreover, we demonstrate that the flow for a case previously considered as a benchmark for the investigation of non-equilibrium chemistry can be regarded as frozen. Finally, the new method is applied to investigate the effect of finite-rate chemistry on boundary layer instability by considering the downstream evolution of a small-amplitude wave and comparing results with those obtained for a frozen gas as well as a gas in local thermodynamic equilibrium.« less
Benchmarking, Total Quality Management, and Libraries.
ERIC Educational Resources Information Center
Shaughnessy, Thomas W.
1993-01-01
Discussion of the use of Total Quality Management (TQM) in higher education and academic libraries focuses on the identification, collection, and use of reliable data. Methods for measuring quality, including benchmarking, are described; performance measures are considered; and benchmarking techniques are examined. (11 references) (MES)
78 FR 13252 - Pyroxasulfone; Pesticide Tolerances
Federal Register 2010, 2011, 2012, 2013, 2014
2013-02-27
... one of the following methods: Federal eRulemaking Portal: http://www.regulations.gov . Follow the... relationship of the results of the studies to human risk. EPA has also considered available information... adverse effects in several target organs. Effects seen in animal studies included cardiac toxicity...
ERIC Educational Resources Information Center
World Federation of the Deaf, Rome (Italy).
Rehabilitation of hearing is considered in five conference papers. Two papers come from Poland: "Rehabilitation of Hearing in Children 'Deaf' in First 5 Years of Age" by D. Borkowska-Gaertig and others and "Possibilities of Hearing Improvement in Adults with Conservative Methods" by T. Bystrzanowska. Also included are…
Application of Density Functional Theory to Systems Containing Metal Atoms
NASA Technical Reports Server (NTRS)
Bauschlicher, Charles W., Jr.; Arnold, James O. (Technical Monitor)
1997-01-01
The accuracy of density functional theory (DFT) for problems involving metal atoms is considered. The DFT results are compared with experiment as well as results obtained using the coupled cluster approach. The comparisons include geometries, frequencies, and bond energies. The systems considered include MO2, M(OH)+(sub n), MNO+, and MCO+(sub 2). The DFT works well for frequencies and geometries, even in cases with symmetry breaking; however, some examples have been found where the symmetry breaking is quite severe and the DFT methods do not work well. The calculation of bond energies is more difficult and examples of the successes as well as failures of DFT will be given.
The solution of the optimization problem of small energy complexes using linear programming methods
NASA Astrophysics Data System (ADS)
Ivanin, O. A.; Director, L. B.
2016-11-01
Linear programming methods were used for solving the optimization problem of schemes and operation modes of distributed generation energy complexes. Applicability conditions of simplex method, applied to energy complexes, including installations of renewable energy (solar, wind), diesel-generators and energy storage, considered. The analysis of decomposition algorithms for various schemes of energy complexes was made. The results of optimization calculations for energy complexes, operated autonomously and as a part of distribution grid, are presented.
Physical Processes in the MAGO/MFT Systems
DOE Office of Scientific and Technical Information (OSTI.GOV)
Garanin, Sergey F; Reinovsky, Robert E.
2015-03-23
The Monograph is devoted to theoretical discussion of the physical effects, which are most significant for the alternative approach to the problem of controlled thermonuclear fusion (CTF): the MAGO/MTF approach. The book includes the description of the approach, its difference from the major CTF systems—magnetic confinement and inertial confinement systems. General physical methods of the processes simulation in this approach are considered, including plasma transport phenomena and radiation, and the theory of transverse collisionless shock waves, the surface discharges theory, important for such kind of research. Different flows and magneto-hydrodynamic plasma instabilities occurring in the frames of this approach aremore » also considered. In virtue of the general physical essence of the considered phenomena the presented results are applicable to a wide range of plasma physics and hydrodynamics processes. The book is intended for the plasma physics and hydrodynamics specialists, post-graduate students, and senior students-physicists.« less
Spacelike matching to null infinity
DOE Office of Scientific and Technical Information (OSTI.GOV)
Zenginoglu, Anil; Tiglio, Manuel
2009-07-15
We present two methods to include the asymptotic domain of a background spacetime in null directions for numerical solutions of evolution equations so that both the radiation extraction problem and the outer boundary problem are solved. The first method is based on the geometric conformal approach, the second is a coordinate based approach. We apply these methods to the case of a massless scalar wave equation on a Kerr spacetime. Our methods are designed to allow existing codes to reach the radiative zone by including future null infinity in the computational domain with relatively minor modifications. We demonstrate the flexibilitymore » of the methods by considering both Boyer-Lindquist and ingoing Kerr coordinates near the black hole. We also confirm numerically predictions concerning tail decay rates for scalar fields at null infinity in Kerr spacetime due to Hod for the first time.« less
Measuring the quantum geometric tensor in two-dimensional photonic and exciton-polariton systems
NASA Astrophysics Data System (ADS)
Bleu, O.; Solnyshkov, D. D.; Malpuech, G.
2018-05-01
We propose theoretically a method that allows to measure all the components of the quantum geometric tensor (the metric tensor and the Berry curvature) in a photonic system. The method is based on standard optical measurements. It applies to two-band systems, which can be mapped to a pseudospin, and to four-band systems, which can be described by two entangled pseudospins. We apply this method to several specific cases. We consider a 2D planar cavity with two polarization eigenmodes, where the pseudospin measurement can be performed via polarization-resolved photoluminescence. We also consider the s band of a staggered honeycomb lattice with polarization-degenerate modes (scalar photons), where the sublattice pseudospin can be measured by performing spatially resolved interferometric measurements. We finally consider the s band of a honeycomb lattice with polarized (spinor) photons as an example of a four-band model. We simulate realistic experimental situations in all cases. We find the photon eigenstates by solving the Schrödinger equation including pumping and finite lifetime, and then simulate the measurements to finally extract realistic mappings of the k-dependent tensor components.
Nonlinear dynamics of cardiovascular ageing
Shiogai, Y.; Stefanovska, A.; McClintock, P.V.E.
2010-01-01
The application of methods drawn from nonlinear and stochastic dynamics to the analysis of cardiovascular time series is reviewed, with particular reference to the identification of changes associated with ageing. The natural variability of the heart rate (HRV) is considered in detail, including the respiratory sinus arrhythmia (RSA) corresponding to modulation of the instantaneous cardiac frequency by the rhythm of respiration. HRV has been intensively studied using traditional spectral analyses, e.g. by Fourier transform or autoregressive methods, and, because of its complexity, has been used as a paradigm for testing several proposed new methods of complexity analysis. These methods are reviewed. The application of time–frequency methods to HRV is considered, including in particular the wavelet transform which can resolve the time-dependent spectral content of HRV. Attention is focused on the cardio-respiratory interaction by introduction of the respiratory frequency variability signal (RFV), which can be acquired simultaneously with HRV by use of a respiratory effort transducer. Current methods for the analysis of interacting oscillators are reviewed and applied to cardio-respiratory data, including those for the quantification of synchronization and direction of coupling. These reveal the effect of ageing on the cardio-respiratory interaction through changes in the mutual modulation of the instantaneous cardiac and respiratory frequencies. Analyses of blood flow signals recorded with laser Doppler flowmetry are reviewed and related to the current understanding of how endothelial-dependent oscillations evolve with age: the inner lining of the vessels (the endothelium) is shown to be of crucial importance to the emerging picture. It is concluded that analyses of the complex and nonlinear dynamics of the cardiovascular system can illuminate the mechanisms of blood circulation, and that the heart, the lungs and the vascular system function as a single entity in dynamical terms. Clear evidence is found for dynamical ageing. PMID:20396667
Nonlinear dynamics of cardiovascular ageing
NASA Astrophysics Data System (ADS)
Shiogai, Y.; Stefanovska, A.; McClintock, P. V. E.
2010-03-01
The application of methods drawn from nonlinear and stochastic dynamics to the analysis of cardiovascular time series is reviewed, with particular reference to the identification of changes associated with ageing. The natural variability of the heart rate (HRV) is considered in detail, including the respiratory sinus arrhythmia (RSA) corresponding to modulation of the instantaneous cardiac frequency by the rhythm of respiration. HRV has been intensively studied using traditional spectral analyses, e.g. by Fourier transform or autoregressive methods, and, because of its complexity, has been used as a paradigm for testing several proposed new methods of complexity analysis. These methods are reviewed. The application of time-frequency methods to HRV is considered, including in particular the wavelet transform which can resolve the time-dependent spectral content of HRV. Attention is focused on the cardio-respiratory interaction by introduction of the respiratory frequency variability signal (RFV), which can be acquired simultaneously with HRV by use of a respiratory effort transducer. Current methods for the analysis of interacting oscillators are reviewed and applied to cardio-respiratory data, including those for the quantification of synchronization and direction of coupling. These reveal the effect of ageing on the cardio-respiratory interaction through changes in the mutual modulation of the instantaneous cardiac and respiratory frequencies. Analyses of blood flow signals recorded with laser Doppler flowmetry are reviewed and related to the current understanding of how endothelial-dependent oscillations evolve with age: the inner lining of the vessels (the endothelium) is shown to be of crucial importance to the emerging picture. It is concluded that analyses of the complex and nonlinear dynamics of the cardiovascular system can illuminate the mechanisms of blood circulation, and that the heart, the lungs and the vascular system function as a single entity in dynamical terms. Clear evidence is found for dynamical ageing.
NASA Technical Reports Server (NTRS)
Bok, L. D.
1973-01-01
The study included material selection and trade-off for the structural components of the wheel and brake optimizing weight vs cost and feasibility for the space shuttle type application. Analytical methods were used to determine section thickness for various materials, and a table was constructed showing weight vs. cost trade-off. The wheel and brake were further optimized by considering design philosophies that deviate from standard aircraft specifications, and designs that best utilize the materials being considered.
Determination of stresses in gas-turbine disks subjected to plastic flow and creep
NASA Technical Reports Server (NTRS)
Millenson, M B; Manson, S S
1948-01-01
A finite-difference method previously presented for computing elastic stresses in rotating disks is extended to include the computation of the disk stresses when plastic flow and creep are considered. A finite-difference method is employed to eliminate numerical integration and to permit nontechnical personnel to make the calculations with a minimum of engineering supervision. Illustrative examples are included to facilitate explanation of the procedure by carrying out the computations on a typical gas-turbine disk through a complete running cycle. The results of the numerical examples presented indicate that plastic flow markedly alters the elastic-stress distribution.
Zhu, Feng; Kalra, Anil; Saif, Tal; Yang, Zaihan; Yang, King H; King, Albert I
2016-01-01
Traumatic brain injury due to primary blast loading has become a signature injury in recent military conflicts and terrorist activities. Extensive experimental and computational investigations have been conducted to study the interrelationships between intracranial pressure response and intrinsic or 'input' parameters such as the head geometry and loading conditions. However, these relationships are very complicated and are usually implicit and 'hidden' in a large amount of simulation/test data. In this study, a data mining method is proposed to explore such underlying information from the numerical simulation results. The heads of different species are described as a highly simplified two-part (skull and brain) finite element model with varying geometric parameters. The parameters considered include peak incident pressure, skull thickness, brain radius and snout length. Their interrelationship and coupling effect are discovered by developing a decision tree based on the large simulation data-set. The results show that the proposed data-driven method is superior to the conventional linear regression method and is comparable to the nonlinear regression method. Considering its capability of exploring implicit information and the relatively simple relationships between response and input variables, the data mining method is considered to be a good tool for an in-depth understanding of the mechanisms of blast-induced brain injury. As a general method, this approach can also be applied to other nonlinear complex biomechanical systems.
Infrared Ship Target Segmentation Based on Spatial Information Improved FCM.
Bai, Xiangzhi; Chen, Zhiguo; Zhang, Yu; Liu, Zhaoying; Lu, Yi
2016-12-01
Segmentation of infrared (IR) ship images is always a challenging task, because of the intensity inhomogeneity and noise. The fuzzy C-means (FCM) clustering is a classical method widely used in image segmentation. However, it has some shortcomings, like not considering the spatial information or being sensitive to noise. In this paper, an improved FCM method based on the spatial information is proposed for IR ship target segmentation. The improvements include two parts: 1) adding the nonlocal spatial information based on the ship target and 2) using the spatial shape information of the contour of the ship target to refine the local spatial constraint by Markov random field. In addition, the results of K -means are used to initialize the improved FCM method. Experimental results show that the improved method is effective and performs better than the existing methods, including the existing FCM methods, for segmentation of the IR ship images.
Immunoaffinity chromatography: an introduction to applications and recent developments
Moser, Annette C
2010-01-01
Immunoaffinity chromatography (IAC) combines the use of LC with the specific binding of antibodies or related agents. The resulting method can be used in assays for a particular target or for purification and concentration of analytes prior to further examination by another technique. This review discusses the history and principles of IAC and the various formats that can be used with this method. An overview is given of the general properties of antibodies and of antibody-production methods. The supports and immobilization methods used with antibodies in IAC and the selection of application and elution conditions for IAC are also discussed. Several applications of IAC are considered, including its use in purification, immunodepletion, direct sample analysis, chromatographic immunoassays and combined analysis methods. Recent developments include the use of IAC with CE or MS, ultrafast immunoextraction methods and the use of immunoaffinity columns in microanalytical systems. PMID:20640220
Petticrew, Mark; Rehfuess, Eva; Noyes, Jane; Higgins, Julian P T; Mayhew, Alain; Pantoja, Tomas; Shemilt, Ian; Sowden, Amanda
2013-11-01
Although there is increasing interest in the evaluation of complex interventions, there is little guidance on how evidence from complex interventions may be reviewed and synthesized, and the relevance of the plethora of evidence synthesis methods to complexity is unclear. This article aims to explore how different meta-analytical approaches can be used to examine aspects of complexity; describe the contribution of various narrative, tabular, and graphical approaches to synthesis; and give an overview of the potential choice of selected qualitative and mixed-method evidence synthesis approaches. The methodological discussions presented here build on a 2-day workshop held in Montebello, Canada, in January 2012, involving methodological experts from the Campbell and Cochrane Collaborations and from other international review centers (Anderson L, Petticrew M, Chandler J, et al. systematic reviews of complex interventions. In press). These systematic review methodologists discussed the broad range of existing methods and considered the relevance of these methods to reviews of complex interventions. The evidence from primary studies of complex interventions may be qualitative or quantitative. There is a wide range of methodological options for reviewing and presenting this evidence. Specific contributions of statistical approaches include the use of meta-analysis, meta-regression, and Bayesian methods, whereas narrative summary approaches provide valuable precursors or alternatives to these. Qualitative and mixed-method approaches include thematic synthesis, framework synthesis, and realist synthesis. A suitable combination of these approaches allows synthesis of evidence for understanding complex interventions. Reviewers need to consider which aspects of complex interventions should be a focus of their review and what types of quantitative and/or qualitative studies they will be including, and this will inform their choice of review methods. These may range from standard meta-analysis through to more complex mixed-method synthesis and synthesis approaches that incorporate theory and/or user's perspectives. Copyright © 2013 Elsevier Inc. All rights reserved.
Interactive performance and focus groups with adolescents: the power of play.
Norris, Anne E; Aroian, Karen J; Warren, Stefanie; Wirth, Jeff
2012-12-01
Conducting focus groups with adolescents can be challenging given their developmental needs, particularly with sensitive topics. These challenges include intense need for peer approval, declining social trust, short attention span, and reliance on concrete operations thinking. In this article, we describe an adaptation of interactive performance as an alternative to traditional focus group method. We used this method in a study of discrimination experienced by Muslims (ages 13-17) and of peer pressure to engage in sexual behavior experienced by Hispanic girls (ages 10-14). Recommendations for use of this method include using an interdisciplinary team, planning for large amounts of disclosure towards the end of the focus group, and considering the fit of this method to the study topic. Copyright © 2012 Wiley Periodicals, Inc.
Stability condition for the drive bunch in a collinear wakefield accelerator
DOE Office of Scientific and Technical Information (OSTI.GOV)
Baturin, S. S.; Zholents, A.
The beam breakup instability of the drive bunch in the structure-based collinear wakefield accelerator is considered and a stabilizing method is proposed. The method includes using the specially designed beam focusing channel, applying the energy chirp along the electron bunch, and keeping energy chirp constant during the drive bunch deceleration. A stability condition is derived that defines the limit on the accelerating field for the witness bunch.
The Use of Land Power to Counter the Iranian Nuclear Proliferation Challenge
2007-06-15
US actions by regional and global opponents (al Qaeda, Russia, etc) to an invasion. Specifically, I will not address the potential for the con spread...facilities will not neutralize the facilities in the long-term, and air attacks will inspire Iranian nationalism against pro -US change. It may not be...thesis committee. Other evaluation methods considered included the strengths-weaknesses-opportunities-threats ( SWOT ) method, and the diplomacy
Stuart, Elizabeth A.; Lee, Brian K.; Leacy, Finbarr P.
2013-01-01
Objective Examining covariate balance is the prescribed method for determining when propensity score methods are successful at reducing bias. This study assessed the performance of various balance measures, including a proposed balance measure based on the prognostic score (also known as the disease-risk score), to determine which balance measures best correlate with bias in the treatment effect estimate. Study Design and Setting The correlations of multiple common balance measures with bias in the treatment effect estimate produced by weighting by the odds, subclassification on the propensity score, and full matching on the propensity score were calculated. Simulated data were used, based on realistic data settings. Settings included both continuous and binary covariates and continuous covariates only. Results The standardized mean difference in prognostic scores, the mean standardized mean difference, and the mean t-statistic all had high correlations with bias in the effect estimate. Overall, prognostic scores displayed the highest correlations of all the balance measures considered. Prognostic score measure performance was generally not affected by model misspecification and performed well under a variety of scenarios. Conclusion Researchers should consider using prognostic score–based balance measures for assessing the performance of propensity score methods for reducing bias in non-experimental studies. PMID:23849158
Brunovskis, Anette; Surtees, Rebecca
2010-01-01
Recent discussions of trafficking research have included calls for more innovative studies and new methodologies in order to move beyond the current trafficking narrative, which is often based on unrepresentative samples and overly simplified images. While new methods can potentially play a role in expanding the knowledge base on trafficking, this article argues that the solution is not entirely about applying new methods, but as much about using current methods to greater effect and with careful attention to their limitations and ethical constraints. Drawing on the authors' experience in researching trafficking issues in a number of projects over the past decade, the article outlines and exemplifies some of the methodological and ethical issues to be considered and accommodated when conducting research with trafficked persons -- including unrepresentative samples; access to respondents; selection biases by "gatekeepers" and self selection by potential respondents. Such considerations should inform not only how research is undertaken but also how this information is read and understood. Moreover, many of these considerations equally apply when considering the application of new methods within this field. The article maintains that a better understanding of how these issues come into play and inform trafficking research will translate into tools for conducting improved research in this field and, by implication, new perspectives on human trafficking.
Chronic hepatitis C and liver fibrosis
Sebastiani, Giada; Gkouvatsos, Konstantinos; Pantopoulos, Kostas
2014-01-01
Chronic infection with hepatitis C virus (HCV) is a leading cause of liver-related morbidity and mortality worldwide and predisposes to liver fibrosis and end-stage liver complications. Liver fibrosis is the excessive accumulation of extracellular matrix proteins, including collagen, and is considered as a wound healing response to chronic liver injury. Its staging is critical for the management and prognosis of chronic hepatitis C (CHC) patients, whose number is expected to rise over the next decades, posing a major health care challenge. This review provides a brief update on HCV epidemiology, summarizes basic mechanistic concepts of HCV-dependent liver fibrogenesis, and discusses methods for assessment of liver fibrosis that are routinely used in clinical practice. Liver biopsy was until recently considered as the gold standard to diagnose and stage liver fibrosis. However, its invasiveness and drawbacks led to the development of non-invasive methods, which include serum biomarkers, transient elastography and combination algorithms. Clinical studies with CHC patients demonstrated that non-invasive methods are in most cases accurate for diagnosis and for monitoring liver disease complications. Moreover, they have a high prognostic value and are cost-effective. Non-invasive methods for assessment of liver fibrosis are gradually being incorporated into new guidelines and are becoming standard of care, which significantly reduces the need for liver biopsy. PMID:25170193
Can Value Added Add Value to Teacher Evaluation?
ERIC Educational Resources Information Center
Darling-Hammond, Linda
2015-01-01
The five thoughtful papers included in this issue of "Educational Researcher" ("ER") raise new questions about the use of value-added methods (VAMs) to estimate teachers' contributions to students' learning as part of personnel evaluation. The papers address both technical and implementation concerns, considering potential…
Functional Multiple-Set Canonical Correlation Analysis
ERIC Educational Resources Information Center
Hwang, Heungsun; Jung, Kwanghee; Takane, Yoshio; Woodward, Todd S.
2012-01-01
We propose functional multiple-set canonical correlation analysis for exploring associations among multiple sets of functions. The proposed method includes functional canonical correlation analysis as a special case when only two sets of functions are considered. As in classical multiple-set canonical correlation analysis, computationally, the…
Local algebraic analysis of differential systems
NASA Astrophysics Data System (ADS)
Kaptsov, O. V.
2015-06-01
We propose a new approach for studying the compatibility of partial differential equations. This approach is a synthesis of the Riquier method, Gröbner basis theory, and elements of algebraic geometry. As applications, we consider systems including the wave equation and the sine-Gordon equation.
ERIC Educational Resources Information Center
GIBSON, ARTHUR R.; STEPHANS, THOMAS M.
ACCELERATION OF PUPILS AND SUBJECTS IS CONSIDERED A MEANS OF EDUCATING THE ACADEMICALLY GIFTED STUDENT. FIVE INTRODUCTORY ARTICLES PROVIDE A FRAMEWORK FOR THINKING ABOUT ACCELERATION. FIVE PROJECT REPORTS OF ACCELERATED PROGRAMS IN OHIO ARE INCLUDED. ACCELERATION IS NOW BEING REGARDED MORE FAVORABLY THAN FORMERLY, BECAUSE METHODS HAVE BEEN…
Teaching Intracultural and Intercultural Communication: A Critique and Suggested Method.
ERIC Educational Resources Information Center
DeVoss, Danielle; Jasken, Julia; Hayden, Dawn
2002-01-01
Summarizes recent literature about the importance of paying attention to intercultural communication. Analyzes the productive approaches in popular business and technical communication textbooks. Presents five challenges for business and technical communication teachers to consider. Includes teaching modules that address these challenges. Notes…
ECOSYSTEM IMPACTS OF GEOENGINEERING: A Review for Developing a Science Plan
DOE Office of Scientific and Technical Information (OSTI.GOV)
Russell, Lynn M; Jackson, Robert B; Norby, Richard J
2012-01-01
Geoengineering methods are intended to reduce the magnitude of climate change, which is already having demonstrable effects on ecosystem structure and functioning. Two different types of activities have been proposed: solar radiation management (SRM), or sunlight reflection methods, which involves reflecting a small percentage of solar light back into space to offset the warming due to greenhouse gases, and carbon dioxide removal (CDR), which includes a range of engineered and biological processes to remove carbon dioxide (CO2) from the atmosphere. This report evaluates some of the possible impacts of CDR and SRM on the physical climate and their subsequent influencemore » on ecosystems, which include the risks and uncertainties associated with new kinds of purposeful perturbations to the Earth. Therefore, the question considered in this review is whether CDR and SRM methods would exacerbate or alleviate the deleterious impacts on ecosystems associated with climate changes that might occur in the foreseeable future.Geoengineering methods are intended to reduce the magnitude of climate change, which is already having demonstrable effects on ecosystem structure and functioning. Two different types of activities have been proposed: solar radiation management (SRM), or sunlight reflection methods, which involves reflecting a small percentage of solar light back into space to offset the warming due to greenhouse gases, and carbon dioxide removal (CDR), which includes a range of engineered and biological processes to remove carbon dioxide (CO2) from the atmosphere. This report evaluates some of the possible impacts of CDR and SRM on the physical climate and their subsequent influence on ecosystems, which include the risks and uncertainties associated with new kinds of purposeful perturbations to the Earth. Therefore, the question considered in this review is whether CDR and SRM methods would exacerbate or alleviate the deleterious impacts on ecosystems associated with climate changes that might occur in the foreseeable future.« less
NASA Astrophysics Data System (ADS)
Ghanbari, Behzad; Inc, Mustafa
2018-04-01
The present paper suggests a novel technique to acquire exact solutions of nonlinear partial differential equations. The main idea of the method is to generalize the exponential rational function method. In order to examine the ability of the method, we consider the resonant nonlinear Schrödinger equation (R-NLSE). Many variants of exact soliton solutions for the equation are derived by the proposed method. Physical interpretations of some obtained solutions is also included. One can easily conclude that the new proposed method is very efficient and finds the exact solutions of the equation in a relatively easy way.
Cost approach of health care entity intangible asset valuation.
Reilly, Robert F
2012-01-01
In the valuation synthesis and conclusion process, the analyst should consider the following question: Does the selected valuation approach(es) and method(s) accomplish the analyst's assignment? Also, does the selected valuation approach and method actually quantify the desired objective of the intangible asset analysis? The analyst should also consider if the selected valuation approach and method analyzes the appropriate bundle of legal rights. The analyst should consider if there were sufficient empirical data available to perform the selected valuation approach and method. The valuation synthesis should consider if there were sufficient data available to make the analyst comfortable with the value conclusion. The valuation analyst should consider if the selected approach and method will be understandable to the intended audience. In the valuation synthesis and conclusion, the analyst should also consider which approaches and methods deserve the greatest consideration with respect to the intangible asset's RUL. The intangible asset RUL is a consideration of each valuation approach. In the income approach, the RUL may affect the projection period for the intangible asset income subject to either yield capitalization or direct capitalization. In the cost approach, the RUL may affect the total amount of obsolescence, if any, from the estimate cost measure (that is, the intangible reproduction cost new or replacement cost new). In the market approach, the RUL may effect the selection, rejection, and/or adjustment of the comparable or guideline intangible asset sale and license transactional data. The experienced valuation analyst will use professional judgment to weight the various value indications to conclude a final intangible asset value, based on: The analyst's confidence in the quantity and quality of available data; The analyst's level of due diligence performed on that data; The relevance of the valuation method to the intangible asset life cycle stage and degree of marketability; and The degree of variation in the range of value indications. Valuation analysts value health care intangible assets for a number of reasons. In addition to regulatory compliance reasons, these reasons include various transaction, taxation, financing, litigation, accounting, bankruptcy, and planning purposes. The valuation analyst should consider all generally accepted intangible asset valuation approaches, methods, and procedures. Many valuation analysts are more familiar with market approach and income approach valuation methods. However, there are numerous instances when cost approach valuation methods are also applicable to the health care intangible asset valuation. This discussion summarized the analyst's procedures and considerations with regard to the cost approach. The cost approach is often applicable to the valuation of intangible assets in the health care industry. However, the cost approach is only applicable if the valuation analyst (1) appropriately considers all of the cost components and (2) appropriately identifies and quantifies all obsolescence allowances. Regardless of the health care intangible asset or the reason for the valuation, the analyst should be familiar with all generally accepted valuation approaches and methods. And, the valuation analyst should have a clear, convincing, and cogent rationale for (1) accepting each approach and method applied and (2) rejecting each approach and method not applied. That way, the valuation analyst will best achieve the purpose and objective of the health care intangible asset valuation.
Chung, Eun-Sung; Kim, Yeonjoo
2014-12-15
This study proposed a robust prioritization framework to identify the priorities of treated wastewater (TWW) use locations with consideration of various uncertainties inherent in the climate change scenarios and the decision-making process. First, a fuzzy concept was applied because future forecast precipitation and their hydrological impact analysis results displayed significant variances when considering various climate change scenarios and long periods (e.g., 2010-2099). Second, various multi-criteria decision making (MCDM) techniques including weighted sum method (WSM), Technique for Order of Preference by Similarity to Ideal Solution (TOPSIS) and fuzzy TOPSIS were introduced to robust prioritization because different MCDM methods use different decision philosophies. Third, decision making method under complete uncertainty (DMCU) including maximin, maximax, minimax regret, Hurwicz, and equal likelihood were used to find robust final rankings. This framework is then applied to a Korean urban watershed. As a result, different rankings were obviously appeared between fuzzy TOPSIS and non-fuzzy MCDMs (e.g., WSM and TOPSIS) because the inter-annual variability in effectiveness was considered only with fuzzy TOPSIS. Then, robust prioritizations were derived based on 18 rankings from nine decadal periods of RCP4.5 and RCP8.5. For more robust rankings, five DMCU approaches using the rankings from fuzzy TOPSIS were derived. This framework combining fuzzy TOPSIS with DMCU approaches can be rendered less controversial among stakeholders under complete uncertainty of changing environments. Copyright © 2014 Elsevier Ltd. All rights reserved.
Research data collection methods: from paper to tablet computers.
Wilcox, Adam B; Gallagher, Kathleen D; Boden-Albala, Bernadette; Bakken, Suzanne R
2012-07-01
Primary data collection is a critical activity in clinical research. Even with significant advances in technical capabilities, clear benefits of use, and even user preferences for using electronic systems for collecting primary data, paper-based data collection is still common in clinical research settings. However, with recent developments in both clinical research and tablet computer technology, the comparative advantages and disadvantages of data collection methods should be determined. To describe case studies using multiple methods of data collection, including next-generation tablets, and consider their various advantages and disadvantages. We reviewed 5 modern case studies using primary data collection, using methods ranging from paper to next-generation tablet computers. We performed semistructured telephone interviews with each project, which considered factors relevant to data collection. We address specific issues with workflow, implementation and security for these different methods, and identify differences in implementation that led to different technology considerations for each case study. There remain multiple methods for primary data collection, each with its own strengths and weaknesses. Two recent methods are electronic health record templates and next-generation tablet computers. Electronic health record templates can link data directly to medical records, but are notably difficult to use. Current tablet computers are substantially different from previous technologies with regard to user familiarity and software cost. The use of cloud-based storage for tablet computers, however, creates a specific challenge for clinical research that must be considered but can be overcome.
Constrained CVT meshes and a comparison of triangular mesh generators
DOE Office of Scientific and Technical Information (OSTI.GOV)
Nguyen, Hoa; Burkardt, John; Gunzburger, Max
2009-01-01
Mesh generation in regions in Euclidean space is a central task in computational science, and especially for commonly used numerical methods for the solution of partial differential equations, e.g., finite element and finite volume methods. We focus on the uniform Delaunay triangulation of planar regions and, in particular, on how one selects the positions of the vertices of the triangulation. We discuss a recently developed method, based on the centroidal Voronoi tessellation (CVT) concept, for effecting such triangulations and present two algorithms, including one new one, for CVT-based grid generation. We also compare several methods, including CVT-based methods, for triangulatingmore » planar domains. To this end, we define several quantitative measures of the quality of uniform grids. We then generate triangulations of several planar regions, including some having complexities that are representative of what one may encounter in practice. We subject the resulting grids to visual and quantitative comparisons and conclude that all the methods considered produce high-quality uniform grids and that the CVT-based grids are at least as good as any of the others.« less
Recent developments in skin mimic systems to predict transdermal permeation.
Waters, Laura J
2015-01-01
In recent years there has been a drive to create experimental techniques that can facilitate the accurate and precise prediction of transdermal permeation without the use of in vivo studies. This review considers why permeation data is essential, provides a brief summary as to how skin acts as a natural barrier to permeation and discusses why in vivo studies are undesirable. This is followed by an in-depth discussion on the extensive range of alternative methods that have been developed in recent years. All of the major 'skin mimic systems' are considered including: in vitro models using synthetic membranes, mathematical models including quantitative structure-permeability relationships (QSPRs), human skin equivalents and chromatographic based methods. All of these model based systems are ideally trying to achieve the same end-point, namely a reliable in vitro-in vivo correlation, i.e. matching non-in vivo obtained data with that from human clinical trials. It is only by achieving this aim, that any new method of obtaining permeation data can be acknowledged as a potential replacement for animal studies, for the determination of transdermal permeation. In this review, the relevance and potential applicability of the various models systems will also be discussed.
Silva, Nuno Miguel; Rio, Jeremy; Currat, Mathias
2017-12-15
Recent advances in sequencing technologies have allowed for the retrieval of ancient DNA data (aDNA) from skeletal remains, providing direct genetic snapshots from diverse periods of human prehistory. Comparing samples taken in the same region but at different times, hereafter called "serial samples", may indicate whether there is continuity in the peopling history of that area or whether an immigration of a genetically different population has occurred between the two sampling times. However, the exploration of genetic relationships between serial samples generally ignores their geographical locations and the spatiotemporal dynamics of populations. Here, we present a new coalescent-based, spatially explicit modelling approach to investigate population continuity using aDNA, which includes two fundamental elements neglected in previous methods: population structure and migration. The approach also considers the extensive temporal and geographical variance that is commonly found in aDNA population samples. We first showed that our spatially explicit approach is more conservative than the previous (panmictic) approach and should be preferred to test for population continuity, especially when small and isolated populations are considered. We then applied our method to two mitochondrial datasets from Germany and France, both including modern and ancient lineages dating from the early Neolithic. The results clearly reject population continuity for the maternal line over the last 7500 years for the German dataset but not for the French dataset, suggesting regional heterogeneity in post-Neolithic migratory processes. Here, we demonstrate the benefits of using a spatially explicit method when investigating population continuity with aDNA. It constitutes an improvement over panmictic methods by considering the spatiotemporal dynamics of genetic lineages and the precise location of ancient samples. The method can be used to investigate population continuity between any pair of serial samples (ancient-ancient or ancient-modern) and to investigate more complex evolutionary scenarios. Although we based our study on mitochondrial DNA sequences, diploid molecular markers of different types (DNA, SNP, STR) can also be simulated with our approach. It thus constitutes a promising tool for the analysis of the numerous aDNA datasets being produced, including genome wide data, in humans but also in many other species.
The Large Introductory Class as an Exercise in Organization Design.
ERIC Educational Resources Information Center
Wagner, John A., III; Van Dyne, Linn
1999-01-01
Four methods for large group instruction differ in control and coordination dimensions: (1) centralization with mutual adjustment; (2) centralization with standardization; (3) decentralization with standardization; and (4) decentralization with mutual adjustment. Other factors to consider include class size and interests of various constituencies:…
Effective Retention Strategies for Diverse Employees.
ERIC Educational Resources Information Center
Musser, Linda R.
2001-01-01
Discusses methods to determine why employees leave or stay, based on experiences at Pennsylvania State University libraries. Considers retention tools that work best to retain diverse employees, including mentoring, networking, career and learning opportunities, balance between work and home life, a welcoming climate, and support for research.…
Isothermal Titration Calorimetry Can Provide Critical Thinking Opportunities
ERIC Educational Resources Information Center
Moore, Dale E.; Goode, David R.; Seney, Caryn S.; Boatwright, Jennifer M.
2016-01-01
College chemistry faculties might not have considered including isothermal titration calorimetry (ITC) in their majors' curriculum because experimental data from this instrumental method are often analyzed via automation (software). However, the software-based data analysis can be replaced with a spreadsheet-based analysis that is readily…
Considering Professional Identity to Enhance Agriculture Teacher Development
ERIC Educational Resources Information Center
Shoulders, Catherine W.; Myers, Brian E.
2011-01-01
The professional identity secondary agriculture teachers display can affect their receptiveness and interest in different professional development events, yet is often overlooked when designing professional development because it is not included in the consensus of proven methods of professional development design and delivery (Desimone, 2009).…
The parts of a research paper? What your readers expect
USDA-ARS?s Scientific Manuscript database
Scientific papers are organized into sections that are easy for scientific readers to follow. This second part of a three part series, summarizes the points that should be considered when writing the main sections of a research report. These sections typically include Introduction Methods, Results,...
Teaching Cases on Family Involvement
ERIC Educational Resources Information Center
Harvard Family Research Project, 2010
2010-01-01
Teaching cases are a valuable tool in preparing teachers and school administrators to engage effectively with families. Because the case method presents a story in practice, it offers students an active learning opportunity. Teaching cases involve real world situations and consider the perspectives of various stakeholders, including teachers,…
Financial. Training Module 4.310.3.77.
ERIC Educational Resources Information Center
Miller, Pat; Bonnstetter, Ron
This document is an instructional module package prepared in objective form for use by an instructor familiar with inventory records, purchasing and budgeting for water and wastewater treatment plants. Included are objectives, instructor guides, student handouts, and transparency masters. The module considers methods of inventory control, proper…
Propeller dynamic and aeroelastic effects
NASA Technical Reports Server (NTRS)
Mccormick, B. W.
1980-01-01
Various aspects of propeller blade dynamics are considered including those factors which are exciting the blades and the dynamic response of the blades to the excitations. Methods for treating this dynamic system are described and problems are discussed which may arise with advanced turboprop designs employing thin, swept blades.
The 1986 advances in bioengineering
DOE Office of Scientific and Technical Information (OSTI.GOV)
Lantz, S.A.; King, A.I.
1986-01-01
This book presents the papers given at a conference on biomedicine. Topics considered at the conference included a mathematical method for obtaining three-dimensional information from standard two-dimensional radiographs, the human lumbar spine, scoliosis and instrumentation, vehicle crashworthiness, lung mechanics, physiological fluid mechanics, microgravity, cardiovascular mechanics, and soft tissue.
Numerical Differentiation of Noisy, Nonsmooth Data
Chartrand, Rick
2011-01-01
We consider the problem of differentiating a function specified by noisy data. Regularizing the differentiation process avoids the noise amplification of finite-difference methods. We use total-variation regularization, which allows for discontinuous solutions. The resulting simple algorithm accurately differentiates noisy functions, including those which have a discontinuous derivative.
Solving Differential Equations Using Modified Picard Iteration
ERIC Educational Resources Information Center
Robin, W. A.
2010-01-01
Many classes of differential equations are shown to be open to solution through a method involving a combination of a direct integration approach with suitably modified Picard iterative procedures. The classes of differential equations considered include typical initial value, boundary value and eigenvalue problems arising in physics and…
Knowledge about idiopathic scoliosis among students of physiotherapy.
Ciazynski, D; Czernicki, K; Durmala, J
2008-01-01
The aim of the study was to determine the level of basic knowledge about idiopathic scoliosis (IS) among students of physiotherapy. The study included 37 students of Medical University of Silesia (17F and 20M aged 22-25, mean 22.6), attending the 3(rd) year of a 1(st) degree of physiotherapy. All students had credits in kinesiotherapy, including methods of conservative treatment of IS. Students were examined using a questionnaire, comprising general knowledge of IS, questions related to sagittal plane correction, influence of various physical activities on IS and known methods of conservative treatment. 81 students considered IS as 3-D deformity. 62.2% of those questioned would diagnose IS when the Cobb angle reaches 10 degrees . All students agreed that the aetiology of IS remains unknown. 54.1% considered forcible extensory exercises of back as favourable in IS. Questioned students mostly preferred swimming (94.6%), yoga (73.0%) and martial arts (32.4%) as beneficial to IS. The methods of conservative treatment which were known best were: Lehnert-Schroth-Weiss (94.6%), Klapp (91.9%), Majoch (89.2%) and Dobosiewicz (78.4%). The conclusions indicate that the average level of knowledge of idiopathic scoliosis among students of physiotherapy is unsatisfactory, despite the education programme including the SOSORT guidelines. Education in the field of scoliosis should be comprehensive and meet contemporary guidelines and standards.
Analysis of digital communication signals and extraction of parameters
NASA Astrophysics Data System (ADS)
Al-Jowder, Anwar
1994-12-01
The signal classification performance of four types of electronics support measure (ESM) communications detection systems is compared from the standpoint of the unintended receiver (interceptor). Typical digital communication signals considered include binary phase shift keying (BPSK), quadrature phase shift keying (QPSK), frequency shift keying (FSK), and on-off keying (OOK). The analysis emphasizes the use of available signal processing software. Detection methods compared include broadband energy detection, FFT-based narrowband energy detection, and two correlation methods which employ the fast Fourier transform (FFT). The correlation methods utilize modified time-frequency distributions, where one of these is based on the Wigner-Ville distribution (WVD). Gaussian white noise is added to the signal to simulate various signal-to-noise ratios (SNR's).
Marion, Bill
2017-03-27
Here, a numerical method is provided for solving the integral equation for the angle-of-incidence (AOI) correction factor for diffuse radiation incident photovoltaic (PV) modules. The types of diffuse radiation considered include sky, circumsolar, horizon, and ground-reflected. The method permits PV module AOI characteristics to be addressed when calculating AOI losses associated with diffuse radiation. Pseudo code is provided to aid users in the implementation, and results are shown for PV modules with tilt angles from 0° to 90°. Diffuse AOI losses are greatest for small PV module tilt angles. Including AOI losses associated with the diffuse irradiance will improve predictionsmore » of PV system performance.« less
Methodological Aspects of Focus Groups in Health Research
Tausch, Anja P.; Menold, Natalja
2016-01-01
Although focus groups are commonly used in health research to explore the perspectives of patients or health care professionals, few studies consider methodological aspects in this specific context. For this reason, we interviewed nine researchers who had conducted focus groups in the context of a project devoted to the development of an electronic personal health record. We performed qualitative content analysis on the interview data relating to recruitment, communication between the focus group participants, and appraisal of the focus group method. The interview data revealed aspects of the focus group method that are particularly relevant for health research and that should be considered in that context. They include, for example, the preferability of face-to-face recruitment, the necessity to allow participants in patient groups sufficient time to introduce themselves, and the use of methods such as participant-generated cards and prioritization. PMID:28462326
Tausch, Anja P; Menold, Natalja
2016-01-01
Although focus groups are commonly used in health research to explore the perspectives of patients or health care professionals, few studies consider methodological aspects in this specific context. For this reason, we interviewed nine researchers who had conducted focus groups in the context of a project devoted to the development of an electronic personal health record. We performed qualitative content analysis on the interview data relating to recruitment, communication between the focus group participants, and appraisal of the focus group method. The interview data revealed aspects of the focus group method that are particularly relevant for health research and that should be considered in that context. They include, for example, the preferability of face-to-face recruitment, the necessity to allow participants in patient groups sufficient time to introduce themselves, and the use of methods such as participant-generated cards and prioritization.
Fuzzy set methods for object recognition in space applications
NASA Technical Reports Server (NTRS)
Keller, James M.
1991-01-01
During the reporting period, the development of the theory and application of methodologies for decision making under uncertainty was addressed. Two subreports are included; the first on properties of general hybrid operators, while the second considers some new research on generalized threshold logic units. In the first part, the properties of the additive gamma-model, where the intersection part is first considered to be the product of the input values and the union part is obtained by an extension of De Morgan's law to fuzzy sets, is explored. Then the Yager's class of union and intersection is used in the additive gamma-model. The inputs are weighted to some power that represents their importance and thus their contribution to the compensation process. In the second part, the extension of binary logic synthesis methods to multiple valued logic synthesis methods to enable the synthesis of decision networks when the input/output variables are not binary is discussed.
Bourasseau, Emeric; Maillet, Jean-Bernard
2011-04-21
This paper presents a new method to obtain chemical equilibrium properties of detonation products mixtures including a solid carbon phase. In this work, the solid phase is modelled through a mesoparticle immersed in the fluid, such that the heterogeneous character of the mixture is explicitly taken into account. Inner properties of the clusters are taken from an equation of state obtained in a previous work, and interaction potential between the nanocluster and the fluid particles is derived from all-atoms simulations using the LCBOPII potential (Long range Carbon Bond Order Potential II). It appears that differences in chemical equilibrium results obtained with this method and the "composite ensemble method" (A. Hervouet et al., J. Phys. Chem. B, 2008, 112.), where fluid and solid phases are considered as non-interacting, are not significant, underlining the fact that considering the inhomogeneity of such system is crucial.
Multiscale Structure of UXO Site Characterization: Spatial Estimation and Uncertainty Quantification
DOE Office of Scientific and Technical Information (OSTI.GOV)
Ostrouchov, George; Doll, William E.; Beard, Les P.
2009-01-01
Unexploded ordnance (UXO) site characterization must consider both how the contamination is generated and how we observe that contamination. Within the generation and observation processes, dependence structures can be exploited at multiple scales. We describe a conceptual site characterization process, the dependence structures available at several scales, and consider their statistical estimation aspects. It is evident that most of the statistical methods that are needed to address the estimation problems are known but their application-specific implementation may not be available. We demonstrate estimation at one scale and propose a representation for site contamination intensity that takes full account of uncertainty,more » is flexible enough to answer regulatory requirements, and is a practical tool for managing detailed spatial site characterization and remediation. The representation is based on point process spatial estimation methods that require modern computational resources for practical application. These methods have provisions for including prior and covariate information.« less
a New Method for Fmeca Based on Fuzzy Theory and Expert System
NASA Astrophysics Data System (ADS)
Byeon, Yoong-Tae; Kim, Dong-Jin; Kim, Jin-O.
2008-10-01
Failure Mode Effects and Criticality Analysis (FMECA) is one of most widely used methods in modern engineering system to investigate potential failure modes and its severity upon the system. FMECA evaluates criticality and severity of each failure mode and visualize the risk level matrix putting those indices to column and row variable respectively. Generally, those indices are determined subjectively by experts and operators. However, this process has no choice but to include uncertainty. In this paper, a method for eliciting expert opinions considering its uncertainty is proposed to evaluate the criticality and severity. In addition, a fuzzy expert system is constructed in order to determine the crisp value of risk level for each failure mode. Finally, an illustrative example system is analyzed in the case study. The results are worth considering in deciding the proper policies for each component of the system.
NASA Astrophysics Data System (ADS)
Talaghat, M. R.; Jokar, S. M.; Modarres, E.
2017-10-01
The reduction of fossil fuel resources and environmental issues made researchers find alternative fuels include biodiesels. One of the most widely used methods for production of biodiesel on a commercial scale is transesterification method. In this work, the biodiesel production by a transesterification method was modeled. Sodium hydroxide was considered as a catalyst to produce biodiesel from canola oil and methanol in a continuous tubular ceramic membranes reactor. As the Biodiesel production reaction from triglycerides is an equilibrium reaction, the reaction rate constants depend on temperature and related linearly to catalyst concentration. By using the mass balance for a membrane tubular reactor and considering the variation of raw materials and products concentration with time, the set of governing equations were solved by numerical methods. The results clearly show the superiority of membrane reactor than conventional tubular reactors. Afterward, the influences of molar ratio of alcohol to oil, weight percentage of the catalyst, and residence time on the performance of biodiesel production reactor were investigated.
Shareef, Hussain; Mohamed, Azah
2017-01-01
The electric vehicle (EV) is considered a premium solution to global warming and various types of pollution. Nonetheless, a key concern is the recharging of EV batteries. Therefore, this study proposes a novel approach that considers the costs of transportation loss, buildup, and substation energy loss and that incorporates harmonic power loss into optimal rapid charging station (RCS) planning. A novel optimization technique, called binary lightning search algorithm (BLSA), is proposed to solve the optimization problem. BLSA is also applied to a conventional RCS planning method. A comprehensive analysis is conducted to assess the performance of the two RCS planning methods by using the IEEE 34-bus test system as the power grid. The comparative studies show that the proposed BLSA is better than other optimization techniques. The daily total cost in RCS planning of the proposed method, including harmonic power loss, decreases by 10% compared with that of the conventional method. PMID:29220396
Islam, Md Mainul; Shareef, Hussain; Mohamed, Azah
2017-01-01
The electric vehicle (EV) is considered a premium solution to global warming and various types of pollution. Nonetheless, a key concern is the recharging of EV batteries. Therefore, this study proposes a novel approach that considers the costs of transportation loss, buildup, and substation energy loss and that incorporates harmonic power loss into optimal rapid charging station (RCS) planning. A novel optimization technique, called binary lightning search algorithm (BLSA), is proposed to solve the optimization problem. BLSA is also applied to a conventional RCS planning method. A comprehensive analysis is conducted to assess the performance of the two RCS planning methods by using the IEEE 34-bus test system as the power grid. The comparative studies show that the proposed BLSA is better than other optimization techniques. The daily total cost in RCS planning of the proposed method, including harmonic power loss, decreases by 10% compared with that of the conventional method.
Strategy on energy saving reconstruction of distribution networks based on life cycle cost
NASA Astrophysics Data System (ADS)
Chen, Xiaofei; Qiu, Zejing; Xu, Zhaoyang; Xiao, Chupeng
2017-08-01
Because the actual distribution network reconstruction project funds are often limited, the cost-benefit model and the decision-making method are crucial for distribution network energy saving reconstruction project. From the perspective of life cycle cost (LCC), firstly the research life cycle is determined for the energy saving reconstruction of distribution networks with multi-devices. Then, a new life cycle cost-benefit model for energy-saving reconstruction of distribution network is developed, in which the modification schemes include distribution transformers replacement, lines replacement and reactive power compensation. In the operation loss cost and maintenance cost area, the operation cost model considering the influence of load season characteristics and the maintenance cost segmental model of transformers are proposed. Finally, aiming at the highest energy saving profit per LCC, a decision-making method is developed while considering financial and technical constraints as well. The model and method are applied to a real distribution network reconstruction, and the results prove that the model and method are effective.
The Facebook Influence Model: A Concept Mapping Approach
Kota, Rajitha; Schoohs, Shari; Whitehill, Jennifer M.
2013-01-01
Abstract Facebook is a popular social media Web site that has been hypothesized to exert potential influence over users' attitudes, intentions, or behaviors. The purpose of this study was to develop a conceptual framework to explain influential aspects of Facebook. This mixed methods study applied concept mapping methodology, a validated five-step method to visually represent complex topics. The five steps comprise preparation, brainstorming, sort and rank, analysis, and interpretation. College student participants were identified using purposeful sampling. The 80 participants had a mean age of 20.5 years, and included 36% males. A total of 169 statements were generated during brainstorming, and sorted into between 6 and 22 groups. The final concept map included 13 clusters. Interpretation data led to grouping of clusters into four final domains, including connection, comparison, identification, and Facebook as an experience. The Facebook Influence Concept Map illustrates key constructs that contribute to influence, incorporating perspectives of older adolescent Facebook users. While Facebook provides a novel lens through which to consider behavioral influence, it can best be considered in the context of existing behavioral theory. The concept map may be used toward development of potential future intervention efforts. PMID:23621717
The Facebook influence model: a concept mapping approach.
Moreno, Megan A; Kota, Rajitha; Schoohs, Shari; Whitehill, Jennifer M
2013-07-01
Facebook is a popular social media Web site that has been hypothesized to exert potential influence over users' attitudes, intentions, or behaviors. The purpose of this study was to develop a conceptual framework to explain influential aspects of Facebook. This mixed methods study applied concept mapping methodology, a validated five-step method to visually represent complex topics. The five steps comprise preparation, brainstorming, sort and rank, analysis, and interpretation. College student participants were identified using purposeful sampling. The 80 participants had a mean age of 20.5 years, and included 36% males. A total of 169 statements were generated during brainstorming, and sorted into between 6 and 22 groups. The final concept map included 13 clusters. Interpretation data led to grouping of clusters into four final domains, including connection, comparison, identification, and Facebook as an experience. The Facebook Influence Concept Map illustrates key constructs that contribute to influence, incorporating perspectives of older adolescent Facebook users. While Facebook provides a novel lens through which to consider behavioral influence, it can best be considered in the context of existing behavioral theory. The concept map may be used toward development of potential future intervention efforts.
NASA Astrophysics Data System (ADS)
Szafranko, Elżbieta
2017-10-01
Assessment of variant solutions developed for a building investment project needs to be made at the stage of planning. While considering alternative solutions, the investor defines various criteria, but a direct evaluation of the degree of their fulfilment by developed variant solutions can be very difficult. In practice, there are different methods which enable the user to include a large number of parameters into an analysis, but their implementation can be challenging. Some methods require advanced mathematical computations, preceded by complicating input data processing, and the generated results may not lend themselves easily to interpretation. Hence, during her research, the author has developed a systemic approach, which involves several methods and whose goal is to compare their outcome. The final stage of the proposed method consists of graphic interpretation of results. The method has been tested on a variety of building and development projects.
On the Use of Accelerated Aging Methods for Screening High Temperature Polymeric Composite Materials
NASA Technical Reports Server (NTRS)
Gates, Thomas S.; Grayson, Michael A.
1999-01-01
A rational approach to the problem of accelerated testing of high temperature polymeric composites is discussed. The methods provided are considered tools useful in the screening of new materials systems for long-term application to extreme environments that include elevated temperature, moisture, oxygen, and mechanical load. The need for reproducible mechanisms, indicator properties, and real-time data are outlined as well as the methodologies for specific aging mechanisms.
NASA Astrophysics Data System (ADS)
Saide, P. E.; Steinhoff, D.; Kosovic, B.; Weil, J.; Smith, N.; Blewitt, D.; Delle Monache, L.
2017-12-01
There are a wide variety of methods that have been proposed and used to estimate methane emissions from oil and gas production by using air composition and meteorology observations in conjunction with dispersion models. Although there has been some verification of these methodologies using controlled releases and concurrent atmospheric measurements, it is difficult to assess the accuracy of these methods for more realistic scenarios considering factors such as terrain, emissions from multiple components within a well pad, and time-varying emissions representative of typical operations. In this work we use a large-eddy simulation (LES) to generate controlled but realistic synthetic observations, which can be used to test multiple source term estimation methods, also known as an Observing System Simulation Experiment (OSSE). The LES is based on idealized simulations of the Weather Research & Forecasting (WRF) model at 10 m horizontal grid-spacing covering an 8 km by 7 km domain with terrain representative of a region located in the Barnett shale. Well pads are setup in the domain following a realistic distribution and emissions are prescribed every second for the components of each well pad (e.g., chemical injection pump, pneumatics, compressor, tanks, and dehydrator) using a simulator driven by oil and gas production volume, composition and realistic operational conditions. The system is setup to allow assessments under different scenarios such as normal operations, during liquids unloading events, or during other prescribed operational upset events. Methane and meteorology model output are sampled following the specifications of the emission estimation methodologies and considering typical instrument uncertainties, resulting in realistic observations (see Figure 1). We will show the evaluation of several emission estimation methods including the EPA Other Test Method 33A and estimates using the EPA AERMOD regulatory model. We will also show source estimation results from advanced methods such as variational inverse modeling, and Bayesian inference and stochastic sampling techniques. Future directions including other types of observations, other hydrocarbons being considered, and assessment of additional emission estimation methods will be discussed.
NASA Astrophysics Data System (ADS)
Kees, C. E.; Miller, C. T.; Dimakopoulos, A.; Farthing, M.
2016-12-01
The last decade has seen an expansion in the development and application of 3D free surface flow models in the context of environmental simulation. These models are based primarily on the combination of effective algorithms, namely level set and volume-of-fluid methods, with high-performance, parallel computing. These models are still computationally expensive and suitable primarily when high-fidelity modeling near structures is required. While most research on algorithms and implementations has been conducted in the context of finite volume methods, recent work has extended a class of level set schemes to finite element methods on unstructured methods. This work considers models of three-phase flow in domains containing air, water, and granular phases. These multi-phase continuum mechanical formulations show great promise for applications such as analysis of coastal and riverine structures. This work will consider formulations proposed in the literature over the last decade as well as new formulations derived using the thermodynamically constrained averaging theory, an approach to deriving and closing macroscale continuum models for multi-phase and multi-component processes. The target applications require the ability to simulate wave breaking and structure over-topping, particularly fully three-dimensional, non-hydrostatic flows that drive these phenomena. A conservative level set scheme suitable for higher-order finite element methods is used to describe the air/water phase interaction. The interaction of these air/water flows with granular materials, such as sand and rubble, must also be modeled. The range of granular media dynamics targeted including flow and wave transmision through the solid media as well as erosion and deposition of granular media and moving bed dynamics. For the granular phase we consider volume- and time-averaged continuum mechanical formulations that are discretized with the finite element method and coupled to the underlying air/water flow via operator splitting (fractional step) schemes. Particular attention will be given to verification and validation of the numerical model and important qualitative features of the numerical methods including phase conservation, wave energy dissipation, and computational efficiency in regimes of interest.
The method of generating functions in exact scalar field inflationary cosmology
NASA Astrophysics Data System (ADS)
Chervon, Sergey V.; Fomin, Igor V.; Beesham, Aroonkumar
2018-04-01
The construction of exact solutions in scalar field inflationary cosmology is of growing interest. In this work, we review the results which have been obtained with the help of one of the most effective methods, viz., the method of generating functions for the construction of exact solutions in scalar field cosmology. We also include in the debate the superpotential method, which may be considered as the bridge to the slow roll approximation equations. Based on the review, we suggest a classification for the generating functions, and find a connection for all of them with the superpotential.
Applications of fuzzy ranking methods to risk-management decisions
NASA Astrophysics Data System (ADS)
Mitchell, Harold A.; Carter, James C., III
1993-12-01
The Department of Energy is making significant improvements to its nuclear facilities as a result of more stringent regulation, internal audits, and recommendations from external review groups. A large backlog of upgrades has resulted. Currently, a prioritization method is being utilized which relies on a matrix of potential consequence and probability of occurrence. The attributes of the potential consequences considered include likelihood, exposure, public health and safety, environmental impact, site personnel safety, public relations, legal liability, and business loss. This paper describes an improved method which utilizes fuzzy multiple attribute decision methods to rank proposed improvement projects.
Generalizations of Tikhonov's regularized method of least squares to non-Euclidean vector norms
NASA Astrophysics Data System (ADS)
Volkov, V. V.; Erokhin, V. I.; Kakaev, V. V.; Onufrei, A. Yu.
2017-09-01
Tikhonov's regularized method of least squares and its generalizations to non-Euclidean norms, including polyhedral, are considered. The regularized method of least squares is reduced to mathematical programming problems obtained by "instrumental" generalizations of the Tikhonov lemma on the minimal (in a certain norm) solution of a system of linear algebraic equations with respect to an unknown matrix. Further studies are needed for problems concerning the development of methods and algorithms for solving reduced mathematical programming problems in which the objective functions and admissible domains are constructed using polyhedral vector norms.
Prinsen, Cecilia A C; Vohra, Sunita; Rose, Michael R; King-Jones, Susanne; Ishaque, Sana; Bhaloo, Zafira; Adams, Denise; Terwee, Caroline B
2014-06-25
The Core Outcome Measures in Effectiveness Trials (COMET) initiative aims to facilitate the development and application of 'core outcome sets' (COS). A COS is an agreed minimum set of outcomes that should be measured and reported in all clinical trials of a specific disease or trial population. The overall aim of the Core Outcome Measurement Instrument Selection (COMIS) project is to develop a guideline on how to select outcome measurement instruments for outcomes included in a COS. As part of this project, we describe our current efforts to achieve a consensus on the methods for selecting outcome measurement instruments for outcomes to be included in a COS. A Delphi study is being performed by a panel of international experts representing diverse stakeholders with the intention that this will result in a guideline for outcome measurement instrument selection. Informed by a literature review, a Delphi questionnaire was developed to identify potentially relevant tasks on instrument selection. The Delphi study takes place in a series of rounds. In the first round, panelists were asked to rate the importance of different tasks in the selection of outcome measurement instruments. They were encouraged to justify their choices and to add other relevant tasks. Consensus was reached if at least 70% of the panelists considered a task 'highly recommended' or 'desirable' and if no opposing arguments were provided. These tasks will be included in the guideline. Tasks that at least 50% of the panelists considered 'not relevant' will be excluded from the guideline. Tasks that were indeterminate will be taken to the second round. All responses of the first round are currently being aggregated and will be fed back to panelists in the second round. A third round will only be performed if the results of the second round require it. Since the Delphi method allows a large group of international experts to participate, we consider it to be the preferred consensus-based method for our study. Based upon this consultation process, a guideline will be developed on instrument selection for outcomes to be included in a COS.
NASA Astrophysics Data System (ADS)
Del Carpio R., Maikol; Hashemi, M. Javad; Mosqueda, Gilberto
2017-10-01
This study examines the performance of integration methods for hybrid simulation of large and complex structural systems in the context of structural collapse due to seismic excitations. The target application is not necessarily for real-time testing, but rather for models that involve large-scale physical sub-structures and highly nonlinear numerical models. Four case studies are presented and discussed. In the first case study, the accuracy of integration schemes including two widely used methods, namely, modified version of the implicit Newmark with fixed-number of iteration (iterative) and the operator-splitting (non-iterative) is examined through pure numerical simulations. The second case study presents the results of 10 hybrid simulations repeated with the two aforementioned integration methods considering various time steps and fixed-number of iterations for the iterative integration method. The physical sub-structure in these tests consists of a single-degree-of-freedom (SDOF) cantilever column with replaceable steel coupons that provides repeatable highlynonlinear behavior including fracture-type strength and stiffness degradations. In case study three, the implicit Newmark with fixed-number of iterations is applied for hybrid simulations of a 1:2 scale steel moment frame that includes a relatively complex nonlinear numerical substructure. Lastly, a more complex numerical substructure is considered by constructing a nonlinear computational model of a moment frame coupled to a hybrid model of a 1:2 scale steel gravity frame. The last two case studies are conducted on the same porotype structure and the selection of time steps and fixed number of iterations are closely examined in pre-test simulations. The generated unbalance forces is used as an index to track the equilibrium error and predict the accuracy and stability of the simulations.
Analytical procedures for water-soluble vitamins in foods and dietary supplements: a review.
Blake, Christopher J
2007-09-01
Water-soluble vitamins include the B-group vitamins and vitamin C. In order to correctly monitor water-soluble vitamin content in fortified foods for compliance monitoring as well as to establish accurate data banks, an accurate and precise analytical method is a prerequisite. For many years microbiological assays have been used for analysis of B vitamins. However they are no longer considered to be the gold standard in vitamins analysis as many studies have shown up their deficiencies. This review describes the current status of analytical methods, including microbiological assays and spectrophotometric, biosensor and chromatographic techniques. In particular it describes the current status of the official methods and highlights some new developments in chromatographic procedures and detection methods. An overview is made of multivitamin extractions and analyses for foods and supplements.
Infinitely many symmetries and conservation laws for quad-graph equations via the Gardner method
NASA Astrophysics Data System (ADS)
Rasin, Alexander G.
2010-06-01
The application of the Gardner method for the generation of conservation laws to all the ABS equations is considered. It is shown that all the necessary information for the application of the Gardner method, namely Bäcklund transformations and initial conservation laws, follows from the multidimensional consistency of ABS equations. We also apply the Gardner method to an asymmetric equation which is not included in the ABS classification. An analog of the Gardner method for the generation of symmetries is developed and applied to the discrete Korteweg-de Vries equation. It can also be applied to all the other ABS equations.
Difference method to search for the anisotropy of primary cosmic radiation
NASA Astrophysics Data System (ADS)
Pavlyuchenko, V. P.; Martirosov, R. M.; Nikolskaya, N. M.; Erlykin, A. D.
2018-01-01
The original difference method used in the search for an anisotropy of primary cosmic radiation at the knee region of its energy spectrum is considered. Its methodical features and properties are analyzed. It is shown that this method, in which properties of particle fluxes (rather than an intensity) are investigated, is stable against random experimental errors and allows one to separate anomalies connected with the laboratory coordinate system from anomalies in the celestial coordinate system. The method uses the multiple scattering of charged particles in the magnetic fields of the Galaxy to study the whole celestial sphere, including the regions outside the line of sight of the installation.
Prediction of intestinal absorption and blood-brain barrier penetration by computational methods.
Clark, D E
2001-09-01
This review surveys the computational methods that have been developed with the aim of identifying drug candidates likely to fail later on the road to market. The specifications for such computational methods are outlined, including factors such as speed, interpretability, robustness and accuracy. Then, computational filters aimed at predicting "drug-likeness" in a general sense are discussed before methods for the prediction of more specific properties--intestinal absorption and blood-brain barrier penetration--are reviewed. Directions for future research are discussed and, in concluding, the impact of these methods on the drug discovery process, both now and in the future, is briefly considered.
A Survey of Rabbit Handling Methods Within the United Kingdom and the Republic of Ireland.
Oxley, James Andrew; Ellis, Clare Frances; McBride, E Anne; McCormick, Wanda Denise
2018-04-25
Rabbits are commonly kept in a variety of settings, including homes, laboratories, and veterinary clinics. Despite the popularity of keeping this prey species, little research has investigated current methods of handling. The aim of this study was to examine the experience of caregivers (owners and keepers) in using five handling methods commonly referred to in books written for companion animal (pet) owners and veterinary and/or laboratory personnel. An online survey was completed by 2644 respondents, representing all three of these groups, and breeders. Data were acquired to determine sources that participants used to gain knowledge of different handling methods, the methods they used and for what purposes they used them, and their perceptions of any associated difficulties or welfare concerns. Results indicated that participants most frequently used the method of supporting a rabbit's body against a person's chest, which was considered the easiest and most welfare-friendly method of the handling methods explored. "Scruffing with rear support" was the least used method and was considered to be distressing and painful for the rabbit. As rabbits are a terrestrial prey species, being picked up is likely an innately stressful experience. Additional research is encouraged to explore the experience of rabbits during handling to identify methods that can be easily used with the fewest welfare compromises.
DOT National Transportation Integrated Search
1985-09-01
Tile objective of the research was to determine or develop effective methods of s1owing traffic to an acceptable speed in work zones. Factors considered in the study included cost, motorist and worker safety, institutional constraints, and probabilit...
Code of Federal Regulations, 2014 CFR
2014-04-01
... Cosmetic Act specifically authorizes inspection of consulting laboratories as well as any factory... Federal Food, Drug, and Cosmetic Act. The Food and Drug Administration's position is that by the... Administration does not consider results of validation studies of analytical and assay methods and control...
Code of Federal Regulations, 2011 CFR
2011-04-01
... Cosmetic Act specifically authorizes inspection of consulting laboratories as well as any factory... Federal Food, Drug, and Cosmetic Act. The Food and Drug Administration's position is that by the... Administration does not consider results of validation studies of analytical and assay methods and control...
Code of Federal Regulations, 2013 CFR
2013-04-01
... Cosmetic Act specifically authorizes inspection of consulting laboratories as well as any factory... Federal Food, Drug, and Cosmetic Act. The Food and Drug Administration's position is that by the... Administration does not consider results of validation studies of analytical and assay methods and control...
Code of Federal Regulations, 2012 CFR
2012-04-01
... Cosmetic Act specifically authorizes inspection of consulting laboratories as well as any factory... Federal Food, Drug, and Cosmetic Act. The Food and Drug Administration's position is that by the... Administration does not consider results of validation studies of analytical and assay methods and control...
CD-ROM Integration Peaks Student Interest in Inquiry.
ERIC Educational Resources Information Center
O'Bannon, Blanche
1997-01-01
Discussion of learning processes examines past educational practices and considers how CD-ROM technology can impact teaching and learning. A lesson plan for elementary school science that uses a CD-ROM encyclopedia is presented that includes instructional goals, performance objectives, teaching and learning activities, and assessment methods.…
The application of elastohydrodynamic lubrication in gear tooth contacts
NASA Technical Reports Server (NTRS)
Townsend, D. P.
1972-01-01
An analytical method is presented for determining elastohydrodynamic film thickness in gears from theory and how the film affects gear failure and life. The practical aspects of gear lubrication are presented, including mechanical and service variables which must be considered to obtain optimum gear performance under severe operating conditions.
48 CFR 9904.406-61 - Interpretation.
Code of Federal Regulations, 2010 CFR
2010-10-01
... ACCOUNTING STANDARDS COST ACCOUNTING STANDARDS 9904.406-61 Interpretation. (a) Questions have arisen as to... categories of costs that have been included in the past and may be considered in the future as restructuring... restructuring costs shall not exceed five years. The straight-line method of amortization should normally be...
Working towards Skills: Perspectives on Workforce Development in SMEs. Research Report.
ERIC Educational Resources Information Center
Hughes, Maria; Keddie, Vince; Webb, Peter; Corney, Mark
Research into workforce development (WD) considered the relationship between corporate assessments of workers' development needs and WD strategies; how learning at work takes place; and what learning methods are used and their effectiveness. Focus was on practice in small and medium-sized enterprises (SMEs). Methodology included a literature…
USDA-ARS?s Scientific Manuscript database
Probiotics are considered an alternative to other treatments enhancing immune response, growth, or feed efficiency in animal rearing systems. While probiotics are extensively used in livestock practices, their use in aquaculture has not been as prevalent due to many factors, including a knowledge ga...
NASA Technical Reports Server (NTRS)
Aguilera, Frank J.
2015-01-01
A guiding principle for conducting research in technology, science, and engineering, leading to innovation is based on our use of research methodology (both qualitative and qualitative). A brief review of research methodology will be presented with an overview of NASA process in developing aeronautics technologies and other things to consider in research including what is innovation.
23 CFR 645.117 - Cost development and reimbursement.
Code of Federal Regulations, 2010 CFR
2010-04-01
... of any alternate costing method should consider the factors listed in paragraphs (b) through (g) of... worked on the project are reimbursable when supported by adequate records. This includes labor associated... the utility may be reimbursed for the time worked directly on the project when supported by adequate...
ERIC Educational Resources Information Center
Prevost, A. Toby; Mason, Dan; Griffin, Simon; Kinmonth, Ann-Louise; Sutton, Stephen; Spiegelhalter, David
2007-01-01
Practical meta-analysis of correlation matrices generally ignores covariances (and hence correlations) between correlation estimates. The authors consider various methods for allowing for covariances, including generalized least squares, maximum marginal likelihood, and Bayesian approaches, illustrated using a 6-dimensional response in a series of…
9 CFR 101.2 - Administrative terminology.
Code of Federal Regulations, 2011 CFR
2011-01-01
... Supplemental Assay Methods. Herd. Any group of animals, including birds, fish, and reptiles, maintained at a... subsequently housed at the common location. If the principal animals of a group are moved to a different location, the group is still considered the same herd. Herd of origin. The herd from which the...
9 CFR 101.2 - Administrative terminology.
Code of Federal Regulations, 2012 CFR
2012-01-01
... Supplemental Assay Methods. Herd. Any group of animals, including birds, fish, and reptiles, maintained at a... subsequently housed at the common location. If the principal animals of a group are moved to a different location, the group is still considered the same herd. Herd of origin. The herd from which the...
9 CFR 101.2 - Administrative terminology.
Code of Federal Regulations, 2014 CFR
2014-01-01
... Supplemental Assay Methods. Herd. Any group of animals, including birds, fish, and reptiles, maintained at a... subsequently housed at the common location. If the principal animals of a group are moved to a different location, the group is still considered the same herd. Herd of origin. The herd from which the...
9 CFR 101.2 - Administrative terminology.
Code of Federal Regulations, 2013 CFR
2013-01-01
... Supplemental Assay Methods. Herd. Any group of animals, including birds, fish, and reptiles, maintained at a... subsequently housed at the common location. If the principal animals of a group are moved to a different location, the group is still considered the same herd. Herd of origin. The herd from which the...
The Words in My Pencil: Considering Children's Writing.
ERIC Educational Resources Information Center
Martin, Anne
Methods of eliciting writing from children are explored in this monograph, which includes samples of nine- and ten-year-old students' writing. The monograph offers four recommendations for improving writing skills: make writing a daily activity; take children's writing seriously; maintain flexible standards of form and language; and identify group…
Turbidity. Training Module 5.240.2.77.
ERIC Educational Resources Information Center
Bonte, John L.; Davidson, Arnold C.
This document is an instructional module package prepared in objective form for use by an instructor familiar with candle turbidimeter and the nephelometric method of turbidity analysis. Included are objectives, an instructor guide, student handout, and transparency masters. A video tape is also available from the author. This module considers use…
NASA Technical Reports Server (NTRS)
Aguilera, Frank J.
2015-01-01
A guiding principle for conducting research in technology, science, and engineering, leading to innovation is based on our use of research methodology (both qualitative and quantitative). A brief review of research methodology will be presented with an overview of NASA process in developing aeronautics technologies and other things to consider in research including what is innovation.
Pre-Treatment. Training Module 2.102.2.77.
ERIC Educational Resources Information Center
Kirkwood Community Coll., Cedar Rapids, IA.
This document is an instructional module package designed in the objective format for use by an instructor familiar with pre-treatment unit operation. Included are objectives, instructor guide, student handouts and transparency masters. The module considers design, operation, maintenance, and safety of common methods of grit removal, screening,…
The Feasibility of Establishing Satellite Campuses for Georgia State University.
ERIC Educational Resources Information Center
Strickland, Wayne G.
Georgia State University, one of the southeast's major urban universities, is considering new methods of delivering its educational services. This report addresses the concept of satellite campuses for Georgia State University by examining those factors affecting their development. Topics included are: the market potential for educational services…
A Practical Guide for Teaching the Mentally Retarded to Swim.
ERIC Educational Resources Information Center
American Association for Health, Physical Education, and Recreation, Washington, DC.
A guide for teaching the retarded to swim begins with a general discussion of retardation, the need for individualization, and staff qualifications. Factors discussed in program organization and administration include community agencies, staff training, examples of records and forms, and first aid procedures. Suggested methods consider perceptual…
The reliability of clinical decisions based on the cervical vertebrae maturation staging method.
Sohrabi, Aydin; Babay Ahari, Sahar; Moslemzadeh, Hossein; Rafighi, Ali; Aghazadeh, Zahra
2016-02-01
Of the various techniques used to determine the optimum timing for growth modification treatments, the cervical vertebrae maturation method has great advantages, including validity and no need for extra X-ray exposure. Recently, the reproducibility of this method has been questioned. The aim of this study was to investigate the cause of poor reproducibility of this method and to assess the reproducibility of the clinical decisions made based on it. Seventy lateral cephalograms of Iranian patients aged 9‒15 years were observed twice by five experienced orthodontists. In addition to determining the developmental stage, each single parameter involved in this method was assessed in terms of inter- and intra-observer reproducibility. In order to evaluate the reproducibility of clinical decisions based on this method, cervical vertebrae maturation staging (CVMS) I and II were considered as phase 1 and CVMS IV and V were considered as phase 3. By considering the clinical approach of the CVMS method, inter-observer reproducibility of this method increased from 0.48 to 0.61 (moderate to substantial) and intra-observer reproducibility enhanced from 0.72 to 0.74. 1. Complete visualization of the first four cervical vertebrae was an inclusion criterion, which also limits the clinical application of CVMS method. 2. These results can be generalized when determining growth modification treatments solely for Class II patients. Difficulty in determining the morphology of C3 and C4 leads to poor reproducibility of the CVMS method. Despite this, it has acceptable reproducibility in determining the timing of functional treatment for Class II patients. © The Author 2015. Published by Oxford University Press on behalf of the European Orthodontic Society. All rights reserved. For permissions, please email: journals.permissions@oup.com.
Constrained simultaneous multi-state reconfigurable wing structure configuration optimization
NASA Astrophysics Data System (ADS)
Snyder, Matthew
A reconfigurable aircraft is capable of in-flight shape change to increase mission performance or provide multi-mission capability. Reconfigurability has always been a consideration in aircraft design, from the Wright Flyer, to the F-14, and most recently the Lockheed-Martin folding wing concept. The Wright Flyer used wing-warping for roll control, the F-14 had a variable-sweep wing to improve supersonic flight capabilities, and the Lockheed-Martin folding wing demonstrated radical in-flight shape change. This dissertation will examine two questions that aircraft reconfigurability raises, especially as reconfiguration increases in complexity. First, is there an efficient method to develop a light weight structure which supports all the loads generated by each configuration? Second, can this method include the capability to propose a sub-structure topology that weighs less than other considered designs? The first question requires a method that will design and optimize multiple configurations of a reconfigurable aerostructure. Three options exist, this dissertation will show one is better than the others. Simultaneous optimization considers all configurations and their respective load cases and constraints at the same time. Another method is sequential optimization which considers each configuration of the vehicle one after the other - with the optimum design variable values from the first configuration becoming the lower bounds for subsequent configurations. This process repeats for each considered configuration and the lower bounds update as necessary. The third approach is aggregate combination — this method keeps the thickness or area of each member for the most critical configuration, the configuration that requires the largest cross-section. This research will show that simultaneous optimization produces a lower weight and different topology for the considered structures when compared to the sequential and aggregate techniques. To answer the second question, the developed optimization algorithm combines simultaneous optimization with a new method for determining the optimum location of the structural members of the sub-structure. The method proposed here considers an over-populated structural model, one in which there are initially more members than necessary. Using a unique iterative process, the optimization algorithm removes members from the design if they do not carry enough load to justify their presence. The initial set of members includes ribs, spars and a series of cross-members that diagonally connect the ribs and spars. The final result is a different structure, which is lower weight than one developed from sequential optimization or aggregate combination, and suggests the primary load paths. Chapter 1 contains background information on reconfigurable aircraft and a description of the new reconfigurable air vehicle being considered by the Air Vehicles Directorate of the Air Force Research Laboratory. This vehicle serves as a platform to test the proposed optimization process. Chapters 2 and 3 overview the optimization method and Chapter 4 provides some background analysis which is unique to this particular reconfigurable air vehicle. Chapter 5 contains the results of the optimizations and demonstrates how changing constraints or initial configuration impacts the final weight and topology of the wing structure. The final chapter contains conclusions and comments on some future work which would further enhance the effectiveness of the simultaneous reconfigurable structural topology optimization process developed and used in this dissertation.
NASA Astrophysics Data System (ADS)
Grubov, V. V.; Runnova, A. E.; Hramov, A. E.
2018-05-01
A new method for adaptive filtration of experimental EEG signals in humans and for removal of different physiological artifacts has been proposed. The algorithm of the method includes empirical mode decomposition of EEG, determination of the number of empirical modes that are considered, analysis of the empirical modes and search for modes that contains artifacts, removal of these modes, and reconstruction of the EEG signal. The method was tested on experimental human EEG signals and demonstrated high efficiency in the removal of different types of physiological EEG artifacts.
Proximate Composition Analysis.
2016-01-01
The proximate composition of foods includes moisture, ash, lipid, protein and carbohydrate contents. These food components may be of interest in the food industry for product development, quality control (QC) or regulatory purposes. Analyses used may be rapid methods for QC or more accurate but time-consuming official methods. Sample collection and preparation must be considered carefully to ensure analysis of a homogeneous and representative sample, and to obtain accurate results. Estimation methods of moisture content, ash value, crude lipid, total carbohydrates, starch, total free amino acids and total proteins are put together in a lucid manner.
Mergers and acquisitions: new arrangements in health care. Part 1.
Grant, E A
1988-02-01
Mergers and acquisitions are assuming a more important role in the healthcare industry today. These transactions require various issues be considered, such as valuation, capital planning, and so forth. In this article, the first in a five-part series on mergers and acquisitions, the fundamental methods and techniques of valuation are discussed. Some of these valuation methods, including comparative market transactions and free cash flow, are explained and examples are used to help potential purchasers and sellers to determine an organization's true value. Other articles in this series will include legal issues, tax implications, purchase investigations, and capital planning for mergers and acquisitions.
Statistical methods for studying disease subtype heterogeneity.
Wang, Molin; Spiegelman, Donna; Kuchiba, Aya; Lochhead, Paul; Kim, Sehee; Chan, Andrew T; Poole, Elizabeth M; Tamimi, Rulla; Tworoger, Shelley S; Giovannucci, Edward; Rosner, Bernard; Ogino, Shuji
2016-02-28
A fundamental goal of epidemiologic research is to investigate the relationship between exposures and disease risk. Cases of the disease are often considered a single outcome and assumed to share a common etiology. However, evidence indicates that many human diseases arise and evolve through a range of heterogeneous molecular pathologic processes, influenced by diverse exposures. Pathogenic heterogeneity has been considered in various neoplasms such as colorectal, lung, prostate, and breast cancers, leukemia and lymphoma, and non-neoplastic diseases, including obesity, type II diabetes, glaucoma, stroke, cardiovascular disease, autism, and autoimmune disease. In this article, we discuss analytic options for studying disease subtype heterogeneity, emphasizing methods for evaluating whether the association of a potential risk factor with disease varies by disease subtype. Methods are described for scenarios where disease subtypes are categorical and ordinal and for cohort studies, matched and unmatched case-control studies, and case-case study designs. For illustration, we apply the methods to a molecular pathological epidemiology study of alcohol intake and colon cancer risk by tumor LINE-1 methylation subtypes. User-friendly software to implement the methods is publicly available. Copyright © 2015 John Wiley & Sons, Ltd.
Yan, Yifei; Zhang, Lisong; Yan, Xiangzhen
2016-01-01
In this paper, a single-slope tunnel pipeline was analysed considering the effects of vertical earth pressure, horizontal soil pressure, inner pressure, thermal expansion force and pipeline—soil friction. The concept of stagnation point for the pipeline was proposed. Considering the deformation compatibility condition of the pipeline elbow, the push force of anchor blocks of a single-slope tunnel pipeline was derived based on an energy method. Then, the theoretical formula for this force is thus generated. Using the analytical equation, the push force of the anchor block of an X80 large-diameter pipeline from the West—East Gas Transmission Project was determined. Meanwhile, to verify the results of the analytical method, and the finite element method, four categories of finite element codes were introduced to calculate the push force, including CAESARII, ANSYS, AutoPIPE and ALGOR. The results show that the analytical results agree well with the numerical results, and the maximum relative error is only 4.1%. Therefore, the results obtained with the analytical method can satisfy engineering requirements. PMID:26963097
NASA Astrophysics Data System (ADS)
Tian, C.; Weng, J.; Liu, Y.
2017-11-01
The convection heat transfer coefficient is one of the evaluation indexes of the brake disc performance. The method used in this paper to calculate the convection heat transfer coefficient is a fluid-solid coupling simulation method, because the calculation results through the empirical formula method have great differences. The model, including a brake disc, a car body, a bogie and flow field, was built, meshed and simulated in the software FLUENT. The calculation models were K-epsilon Standard model and Energy model. The working condition of the brake disc was considered. The coefficient of various parts can be obtained through the method in this paper. The simulation result shows that, under 160 km/h speed, the radiating ribs have the maximum convection heat transfer coefficient and the value is 129.6W/(m2·K), the average coefficient of the whole disc is 100.4W/(m2·K), the windward of ribs is positive-pressure area and the leeward of ribs is negative-pressure area, the maximum pressure is 2663.53Pa.
Rollier, Patricia; Lombard, Bertrand; Guillier, Laurent; François, Danièle; Romero, Karol; Pierru, Sylvie; Bouhier, Laurence; Gnanou Besse, Nathalie
2018-05-01
The reference method for the detection and enumeration of L. monocytogenes in food (Standards EN ISO 11290-1&2) have been validated by inter-laboratory studies in the frame of the Mandate M381 from European Commission to CEN. In this paper, the inter-laboratory studies led in 2013 on 5 matrices (cold-smoked salmon, milk powdered infant food formula, vegetables, environment, and cheese) to validate Standard EN ISO 11290-2 are reported. According to the results obtained, the method of the revised Standard EN ISO 11290-2 can be considered as a good method for the enumeration of L. monocytogenes in foods and food processing environment, in particular for the matrices included in the study. Values of repeatability and reproducibility standard deviations can be considered satisfactory for this type of method with a confirmation stage, since most of them were below 0.3 log 10 , also at low levels, close to the regulatory limit of 100 CFU/g. Copyright © 2018 Elsevier B.V. All rights reserved.
Eigenstates and dynamics of Hooke's atom: Exact results and path integral simulations
NASA Astrophysics Data System (ADS)
Gholizadehkalkhoran, Hossein; Ruokosenmäki, Ilkka; Rantala, Tapio T.
2018-05-01
The system of two interacting electrons in one-dimensional harmonic potential or Hooke's atom is considered, again. On one hand, it appears as a model for quantum dots in a strong confinement regime, and on the other hand, it provides us with a hard test bench for new methods with the "space splitting" arising from the one-dimensional Coulomb potential. Here, we complete the numerous previous studies of the ground state of Hooke's atom by including the excited states and dynamics, not considered earlier. With the perturbation theory, we reach essentially exact eigenstate energies and wave functions for the strong confinement regime as novel results. We also consider external perturbation induced quantum dynamics in a simple separable case. Finally, we test our novel numerical approach based on real-time path integrals (RTPIs) in reproducing the above. The RTPI turns out to be a straightforward approach with exact account of electronic correlations for solving the eigenstates and dynamics without the conventional restrictions of electronic structure methods.
Suárez, Inmaculada; Coto, Baudilio
2015-08-14
Average molecular weights and polydispersity indexes are some of the most important parameters considered in the polymer characterization. Usually, gel permeation chromatography (GPC) and multi angle light scattering (MALS) are used for this determination, but GPC values are overestimated due to the dispersion introduced by the column separation. Several procedures were proposed to correct such effect usually involving more complex calibration processes. In this work, a new method of calculation has been considered including diffusion effects. An equation for the concentration profile due to diffusion effects along the GPC column was considered to be a Fickian function and polystyrene narrow standards were used to determine effective diffusion coefficients. The molecular weight distribution function of mono and poly disperse polymers was interpreted as a sum of several Fickian functions representing a sample formed by only few kind of polymer chains with specific molecular weight and diffusion coefficient. Proposed model accurately fit the concentration profile along the whole elution time range as checked by the computed standard deviation. Molecular weights obtained by this new method are similar to those obtained by MALS or traditional GPC while polydispersity index values are intermediate between those obtained by the traditional GPC combined to Universal Calibration method and the MALS method. Values for Pearson and Lin coefficients shows improvement in the correlation of polydispersity index values determined by GPC and MALS methods when diffusion coefficients and new methods are used. Copyright © 2015 Elsevier B.V. All rights reserved.
LASERS IN MEDICINE: Laser diagnostics of biofractals
NASA Astrophysics Data System (ADS)
Ushenko, A. G.
1999-12-01
An optical approach to the problem of modelling and diagnostics of the structures of biofractal formations was considered in relation to human bone tissue. A model was proposed for the optical properties of this tissue, including three levels of fractal organisation: microcrystalline, macrocrystalline, and architectural. The studies were based on laser coherent polarimetry ensuring the retrieval of the fullest information about the optical and polarisation properties of bone tissue. A method was developed for contactless noninvasive diagnostics of the orientational and mineralogical structure of bone tissue considered as a biofractal.
Andrusyszyn, M A; Cragg, C E; Humbert, J
2001-04-01
The relationships among multiple distance delivery methods, preferred learning style, content, and achievement was sought for primary care nurse practitioner students. A researcher-designed questionnaire was completed by 86 (71%) participants, while 6 engaged in follow-up interviews. The results of the study included: participants preferred learning by "considering the big picture"; "setting own learning plans"; and "focusing on concrete examples." Several positive associations were found: learning on own with learning by reading, and setting own learning plans; small group with learning through discussion; large group with learning new things through hearing and with having learning plans set by others. The most preferred method was print-based material and the least preferred method was audio tape. The most suited method for content included video teleconferencing for counseling, political action, and transcultural issues; and video tape for physical assessment. Convenience, self-direction, and timing of learning were more important than delivery method or learning style. Preferred order of learning was reading, discussing, observing, doing, and reflecting. Recommended considerations when designing distance courses include a mix of delivery methods, specific content, outcomes, learner characteristics, and state of technology.
Raebel, Marsha A; Shetterly, Susan; Lu, Christine Y; Flory, James; Gagne, Joshua J; Harrell, Frank E; Haynes, Kevin; Herrinton, Lisa J; Patorno, Elisabetta; Popovic, Jennifer; Selvan, Mano; Shoaibi, Azadeh; Wang, Xingmei; Roy, Jason
2016-07-01
Our purpose was to quantify missing baseline laboratory results, assess predictors of missingness, and examine performance of missing data methods. Using the Mini-Sentinel Distributed Database from three sites, we selected three exposure-outcome scenarios with laboratory results as baseline confounders. We compared hazard ratios (HRs) or risk differences (RDs) and 95% confidence intervals (CIs) from models that omitted laboratory results, included only available results (complete cases), and included results after applying missing data methods (multiple imputation [MI] regression, MI predictive mean matching [PMM] indicator). Scenario 1 considered glucose among second-generation antipsychotic users and diabetes. Across sites, glucose was available for 27.7-58.9%. Results differed between complete case and missing data models (e.g., olanzapine: HR 0.92 [CI 0.73, 1.12] vs 1.02 [0.90, 1.16]). Across-site models employing different MI approaches provided similar HR and CI; site-specific models provided differing estimates. Scenario 2 evaluated creatinine among individuals starting high versus low dose lisinopril and hyperkalemia. Creatinine availability: 44.5-79.0%. Results differed between complete case and missing data models (e.g., HR 0.84 [CI 0.77, 0.92] vs. 0.88 [0.83, 0.94]). HR and CI were identical across MI methods. Scenario 3 examined international normalized ratio (INR) among warfarin users starting interacting versus noninteracting antimicrobials and bleeding. INR availability: 20.0-92.9%. Results differed between ignoring INR versus including INR using missing data methods (e.g., RD 0.05 [CI -0.03, 0.13] vs 0.09 [0.00, 0.18]). Indicator and PMM methods gave similar estimates. Multi-site studies must consider site variability in missing data. Different missing data methods performed similarly. Copyright © 2016 John Wiley & Sons, Ltd. Copyright © 2016 John Wiley & Sons, Ltd.
"The Scientific Method" as Myth and Ideal
NASA Astrophysics Data System (ADS)
Woodcock, Brian A.
2014-10-01
"The Scientific Method" as it has been portrayed in popular and introductory contexts has been declared a myth. The variation that one finds in introductory presentations of "The Scientific Method" is explained by the fact that there is no canonical account among historians and philosophers of science. What, in particular, is wrong with "The Scientific Method"? This essay provides a fairly comprehensive survey of shortcomings of "The Scientific Method". Included are corrections to several misconceptions that often accompany such presentations. Rather than treating "The Scientific Method" as a useful approximation or an ideal, the myth should be discarded. Lessons can be learned for introductory pedagogical contexts from considering the shortcomings of the myth.
DIAGNOSIS AND TREATMENT OF POSTERIOR INTEROSSEOUS NERVE ENTRAPMENT: SYSTEMATIC REVIEW
MORAES, MARCO AURÉLIO DE; GONÇALVES, RUBENS GUILHERME; SANTOS, JOÃO BAPTISTA GOMES DOS; BELLOTI, JOÃO CARLOS; FALOPPA, FLÁVIO; MORAES, VINÍCIUS YNOE DE
2017-01-01
ABSTRACT Compressive syndromes of the radial nerve have different presentations. There is no consensus on diagnostic and therapeutic methods. The aim of this review is to summarize such methods. Eletronic searches related terms, held in databases (1980-2016): Pubmed (via Medline), Lilacs (via Scielo) and Google Scholar. Through pre-defined protocol, we identified relevant studies. We excluded case reports. Aspects of diagnosis and treatment were synthesized for analysis and tables. Quantitative analyzes were followed by their dispersion variables. Fourteen studies were included. All studies were considered as level IV evidence. Most studies consider aspects of clinical history and provocative maneuvers. There is no consensus on the use of electromyography, and methods are heterogeneous. Studies have shown that surgical treatment (muscle release and neurolysis) has variable success rate, ranging from 20 to 96.5%. Some studies applied self reported scores, though the heterogeneity of the population does not allow inferential analyzes on the subject. few complications reported. Most studies consider the diagnosis of compressive radial nerve syndromes essentially clinical. The most common treatment was combined muscle release and neurolysis, with heterogeneous results. There is a need for comparative studies. Level of Evidence III, Systematic Review. PMID:28642652
Considering the Lives of Microbes in Microbial Communities.
Shank, Elizabeth A
2018-01-01
Over the last decades, sequencing technologies have transformed our ability to investigate the composition and functional capacity of microbial communities. Even so, critical questions remain about these complex systems that cannot be addressed by the bulk, community-averaged data typically provided by sequencing methods. In this Perspective, I propose that future advances in microbiome research will emerge from considering "the lives of microbes": we need to create methods to explicitly interrogate how microbes exist and interact in native-setting-like microenvironments. This approach includes developing approaches that expose the phenotypic heterogeneity of microbes; exploring the effects of coculture cues on cellular differentiation and metabolite production; and designing visualization systems that capture features of native microbial environments while permitting the nondestructive observation of microbial interactions over space and time with single-cell resolution.
Conditioning 3D object-based models to dense well data
NASA Astrophysics Data System (ADS)
Wang, Yimin C.; Pyrcz, Michael J.; Catuneanu, Octavian; Boisvert, Jeff B.
2018-06-01
Object-based stochastic simulation models are used to generate categorical variable models with a realistic representation of complicated reservoir heterogeneity. A limitation of object-based modeling is the difficulty of conditioning to dense data. One method to achieve data conditioning is to apply optimization techniques. Optimization algorithms can utilize an objective function measuring the conditioning level of each object while also considering the geological realism of the object. Here, an objective function is optimized with implicit filtering which considers constraints on object parameters. Thousands of objects conditioned to data are generated and stored in a database. A set of objects are selected with linear integer programming to generate the final realization and honor all well data, proportions and other desirable geological features. Although any parameterizable object can be considered, objects from fluvial reservoirs are used to illustrate the ability to simultaneously condition multiple types of geologic features. Channels, levees, crevasse splays and oxbow lakes are parameterized based on location, path, orientation and profile shapes. Functions mimicking natural river sinuosity are used for the centerline model. Channel stacking pattern constraints are also included to enhance the geological realism of object interactions. Spatial layout correlations between different types of objects are modeled. Three case studies demonstrate the flexibility of the proposed optimization-simulation method. These examples include multiple channels with high sinuosity, as well as fragmented channels affected by limited preservation. In all cases the proposed method reproduces input parameters for the object geometries and matches the dense well constraints. The proposed methodology expands the applicability of object-based simulation to complex and heterogeneous geological environments with dense sampling.
A Review of High-Order and Optimized Finite-Difference Methods for Simulating Linear Wave Phenomena
NASA Technical Reports Server (NTRS)
Zingg, David W.
1996-01-01
This paper presents a review of high-order and optimized finite-difference methods for numerically simulating the propagation and scattering of linear waves, such as electromagnetic, acoustic, or elastic waves. The spatial operators reviewed include compact schemes, non-compact schemes, schemes on staggered grids, and schemes which are optimized to produce specific characteristics. The time-marching methods discussed include Runge-Kutta methods, Adams-Bashforth methods, and the leapfrog method. In addition, the following fourth-order fully-discrete finite-difference methods are considered: a one-step implicit scheme with a three-point spatial stencil, a one-step explicit scheme with a five-point spatial stencil, and a two-step explicit scheme with a five-point spatial stencil. For each method studied, the number of grid points per wavelength required for accurate simulation of wave propagation over large distances is presented. Recommendations are made with respect to the suitability of the methods for specific problems and practical aspects of their use, such as appropriate Courant numbers and grid densities. Avenues for future research are suggested.
Noncovalent Interactions of DNA Bases with Naphthalene and Graphene.
Cho, Yeonchoo; Min, Seung Kyu; Yun, Jeonghun; Kim, Woo Youn; Tkatchenko, Alexandre; Kim, Kwang S
2013-04-09
The complexes of a DNA base bound to graphitic systems are studied. Considering naphthalene as the simplest graphitic system, DNA base-naphthalene complexes are scrutinized at high levels of ab initio theory including coupled cluster theory with singles, doubles, and perturbative triples excitations [CCSD(T)] at the complete basis set (CBS) limit. The stacked configurations are the most stable, where the CCSD(T)/CBS binding energies of guanine, adenine, thymine, and cytosine are 9.31, 8.48, 8.53, 7.30 kcal/mol, respectively. The energy components are investigated using symmetry-adapted perturbation theory based on density functional theory including the dispersion energy. We compared the CCSD(T)/CBS results with several density functional methods applicable to periodic systems. Considering accuracy and availability, the optB86b nonlocal functional and the Tkatchenko-Scheffler functional are used to study the binding energies of nucleobases on graphene. The predicted values are 18-24 kcal/mol, though many-body effects on screening and energy need to be further considered.
NASA Technical Reports Server (NTRS)
Bond, Thomas H. (Technical Monitor); Anderson, David N.
2004-01-01
This manual reviews the derivation of the similitude relationships believed to be important to ice accretion and examines ice-accretion data to evaluate their importance. Both size scaling and test-condition scaling methods employing the resulting similarity parameters are described, and experimental icing tests performed to evaluate scaling methods are reviewed with results. The material included applies primarily to unprotected, unswept geometries, but some discussion of how to approach other situations is included as well. The studies given here and scaling methods considered are applicable only to Appendix-C icing conditions. Nearly all of the experimental results presented have been obtained in sea-level tunnels. Recommendations are given regarding which scaling methods to use for both size scaling and test-condition scaling, and icing test results are described to support those recommendations. Facility limitations and size-scaling restrictions are discussed. Finally, appendices summarize the air, water and ice properties used in NASA scaling studies, give expressions for each of the similarity parameters used and provide sample calculations for the size-scaling and test-condition scaling methods advocated.
Lucas, Patricia J; Baird, Janis; Arai, Lisa; Law, Catherine; Roberts, Helen M
2007-01-01
Background The inclusion of qualitative studies in systematic reviews poses methodological challenges. This paper presents worked examples of two methods of data synthesis (textual narrative and thematic), used in relation to one review, with the aim of enabling researchers to consider the strength of different approaches. Methods A systematic review of lay perspectives of infant size and growth was conducted, locating 19 studies (including both qualitative and quantitative). The data extracted from these were synthesised using both a textual narrative and a thematic synthesis. Results The processes of both methods are presented, showing a stepwise progression to the final synthesis. Both methods led us to similar conclusions about lay views toward infant size and growth. Differences between methods lie in the way they dealt with study quality and heterogeneity. Conclusion On the basis of the work reported here, we consider textual narrative and thematic synthesis have strengths and weaknesses in relation to different research questions. Thematic synthesis holds most potential for hypothesis generation, but may obscure heterogeneity and quality appraisal. Textual narrative synthesis is better able to describe the scope of existing research and account for the strength of evidence, but is less good at identifying commonality. PMID:17224044
Intercomparison of methods for image quality characterization. II. Noise power spectrum
DOE Office of Scientific and Technical Information (OSTI.GOV)
Dobbins, James T. III; Samei, Ehsan; Ranger, Nicole T.
Second in a two-part series comparing measurement techniques for the assessment of basic image quality metrics in digital radiography, in this paper we focus on the measurement of the image noise power spectrum (NPS). Three methods were considered: (1) a method published by Dobbins et al. [Med. Phys. 22, 1581-1593 (1995)] (2) a method published by Samei et al. [Med. Phys. 30, 608-622 (2003)], and (3) a new method sanctioned by the International Electrotechnical Commission (IEC 62220-1, 2003), developed as part of an international standard for the measurement of detective quantum efficiency. In addition to an overall comparison of themore » estimated NPS between the three techniques, the following factors were also evaluated for their effect on the measured NPS: horizontal versus vertical directional dependence, the use of beam-limiting apertures, beam spectrum, and computational methods of NPS analysis, including the region-of-interest (ROI) size and the method of ROI normalization. Of these factors, none was found to demonstrate a substantial impact on the amplitude of the NPS estimates ({<=}3.1% relative difference in NPS averaged over frequency, for each factor considered separately). Overall, the three methods agreed to within 1.6%{+-}0.8% when averaged over frequencies >0.15 mm{sup -1}.« less
Statistical Approaches Used to Assess the Equity of Access to Food Outlets: A Systematic Review
Lamb, Karen E.; Thornton, Lukar E.; Cerin, Ester; Ball, Kylie
2015-01-01
Background Inequalities in eating behaviours are often linked to the types of food retailers accessible in neighbourhood environments. Numerous studies have aimed to identify if access to healthy and unhealthy food retailers is socioeconomically patterned across neighbourhoods, and thus a potential risk factor for dietary inequalities. Existing reviews have examined differences between methodologies, particularly focussing on neighbourhood and food outlet access measure definitions. However, no review has informatively discussed the suitability of the statistical methodologies employed; a key issue determining the validity of study findings. Our aim was to examine the suitability of statistical approaches adopted in these analyses. Methods Searches were conducted for articles published from 2000–2014. Eligible studies included objective measures of the neighbourhood food environment and neighbourhood-level socio-economic status, with a statistical analysis of the association between food outlet access and socio-economic status. Results Fifty-four papers were included. Outlet accessibility was typically defined as the distance to the nearest outlet from the neighbourhood centroid, or as the number of food outlets within a neighbourhood (or buffer). To assess if these measures were linked to neighbourhood disadvantage, common statistical methods included ANOVA, correlation, and Poisson or negative binomial regression. Although all studies involved spatial data, few considered spatial analysis techniques or spatial autocorrelation. Conclusions With advances in GIS software, sophisticated measures of neighbourhood outlet accessibility can be considered. However, approaches to statistical analysis often appear less sophisticated. Care should be taken to consider assumptions underlying the analysis and the possibility of spatially correlated residuals which could affect the results. PMID:29546115
Hu, Xiao; Maffucci, Irene; Contini, Alessandro
2018-05-13
The inclusion of direct effects mediated by water during the ligand-receptor recognition is a hot-topic of modern computational chemistry applied to drug discovery and development. Docking or virtual screening with explicit hydration is still debatable, despite the successful cases that have been presented in the last years. Indeed, how to select the water molecules that will be included in the docking process or how the included waters should be treated remain open questions. In this review, we will discuss some of the most recent methods that can be used in computational drug discovery and drug development when the effect of a single water, or of a small network of interacting waters, needs to be explicitly considered. Here, we analyse software to aid the selection, or to predict the position, of water molecules that are going to be explicitly considered in later docking studies. We also present software and protocols able to efficiently treat flexible water molecules during docking, including examples of applications. Finally, we discuss methods based on molecular dynamics simulations that can be used to integrate docking studies or to reliably and efficiently compute binding energies of ligands in presence of interfacial or bridging water molecules. Software applications aiding the design of new drugs that exploit water molecules, either as displaceable residues or as bridges to the receptor, are constantly being developed. Although further validation is needed, workflows that explicitly consider water will probably become a standard for computational drug discovery soon. Copyright© Bentham Science Publishers; For any queries, please email at epub@benthamscience.org.
USDA-ARS?s Scientific Manuscript database
Abstract Due to its association with Raffaelea lauricola, a pathogen that causes laurel wilt, the redbay ambrosia beetle, Xyleborus glabratus is considered one of the most damaging pests of Persea species including avocado. Currently there is no satisfactory method to control this pest. Biological c...
ERIC Educational Resources Information Center
Wimberly, Anne E. Streaty
2004-01-01
Using an autobiographical approach for pedagogical reflection, the author raises questions about how to include "hospitable kinship" and "gift exchange" in teaching and learning. Her experience with a Zimbabwean community circle of hospitable kinship has prompted her to consider how this method of community formation might be…
Charles W. Stuber: Maize geneticist and pioneer of marker-assisted selection
USDA-ARS?s Scientific Manuscript database
Charles W. Stuber is considered a pioneer of quantitative genetic mapping and marker-assisted selection in maize. The achievements of his four decade career in research include the development of genetic marker systems used in maize and adapted in many other crops, the first methods and studies to i...
Robert S. Motroni; Daniel A. Airola; Robin K. Ma rose; Nancy D. Tosta
1991-01-01
A geographic information system was used to assess wildlife species richness (number of species) in valley-foothill hardwood habitats throughout California to set priorities for conservation attention. Species richness values were assessed and compared using three methods: one that included all species without considering canopy cover conditions and species preferences...
Cuban Voices: A Case Study of English Language Teacher Education
ERIC Educational Resources Information Center
Smith, Steven John
2016-01-01
This case study uses qualitative research methods and a postcolonial paradigm to listen to the voices of Cuban teacher educators describing how they educate and prepare English language teachers in Cuba. English language teacher education in Cuba includes features that are considered innovative, contemporary and good practice in the Western world.…
Measures for improving the zeppelin airships for long distance transportation
NASA Technical Reports Server (NTRS)
Duerr, L. F.
1980-01-01
Factors to be considered in the construction of dirigibles include the design and weight of support structures, static and aerodynamic loads on the main ring, the annealing of support materials, and the dynamic gas pressure. Adaptations made for using helium as the lifting gas, and a method for extracting ballast are described.
Spreading Chaos: The Role of Popularizations in the Diffusion of Scientific Ideas
ERIC Educational Resources Information Center
Paul, Danette
2004-01-01
Scientific popularizations are generally considered translations (often dubious ones) of scientific research for a lay audience. This study explores the role popularizations play within scientific discourse, specifically in the development of chaos theory. The methods included a review of the popular and the semipopular books on chaos theory from…
Beating liquid helium: the technologies of cryogen-free superconducting magnets
NASA Astrophysics Data System (ADS)
Burgoyne, John
2015-03-01
Cryogen-free superconducting magnets have been available now for almost 15 years, but have only become standard commercial products in more recent years. In this review we will consider the pros and cons of ``dry'' design including superconducting wire development and selection, thermal budgeting, and the alternative methods for achieving magnet cooling.
Educating the Gifted Child: Challenge and Response in the U.S.A.
ERIC Educational Resources Information Center
Khatena, Joe
Considered in the paper are past and current developments in the education of gifted and talented students in the United States. Discussed are the following aspects: identification (including specific evaluations of giftedness, and distinctions between creativity and I.Q.); programing methods and approaches (such as creative problem solving…
Social and Individual Frame Factors in L2 Learning: Comparative Aspects.
ERIC Educational Resources Information Center
Ekstrand, Lars H.
A large number of factors are considered in their role in second language learning. Individual factors include language aptitude, personality, attitudes and motivation, and the role of the speaker's native language. Teacher factors involve the method of instruction, the sex of the teacher, and a teacher's training and competence, while…
10 Suggestions for Enhancing Lecturing
ERIC Educational Resources Information Center
Heitzmann, Ray
2010-01-01
Criticism of the lecture method remains a staple of discussion and writing in academia--and most of the time it's deserved! Those interested in improving this aspect of their teaching might wish to consider some or all of the following suggestions for enhancing lectures. These include: (1) Lectures must start with a "grabber"; (2)…
Academic Quality Rankings of American Colleges and Universities.
ERIC Educational Resources Information Center
Webster, David S.
Past and current methods used in academic quality rankings of U.S. colleges and universities are discussed. In addition to a literature and historical review, modern quality rankings are compared with early (pre-1959) rankings, including past rankings of medical, dental, legal and black education. Also considered are the exemplary 1982 evaluation…
Evaluation Method for an App Involving Kitchen Activities.
Burzagli, Laura; Gaggioli, Andrea
2017-01-01
An app, developed for the national Italian project "Design for all", is evaluated here. This app supports the user in most of his/her cooking activities. The evaluation process integrates different technologies, such as expert interview, cognitive walkthrough and focus group, to consider different aspects that include, but are not limited to, usability and accessibility.
When Academics Integrate Research Skill Development in the Curriculum
ERIC Educational Resources Information Center
Willison, J. W.
2012-01-01
This study considered outcomes when 27 academics explicitly developed and assessed student research skills in 28 regular (non-research methods) semester-length courses. These courses ranged from small (n = 17) to medium-large (n = 222) and included those from first year to masters in business, engineering, health science, humanities and science,…
The Assayer's Scale: Was Intelligence the Ultimate Currency of the Information Age?
ERIC Educational Resources Information Center
Fuller, Renee
1992-01-01
This article considers the role of the basic cognitive unit, called the "story engram," in young children's learning to read, including children ranging in ability from severe mental retardation to giftedness. It illustrates how the "Ball-Stick-Bird" method of beginning reading can facilitate this process because of the…
Huang, Jian; Zhang, Cun-Hui
2013-01-01
The ℓ1-penalized method, or the Lasso, has emerged as an important tool for the analysis of large data sets. Many important results have been obtained for the Lasso in linear regression which have led to a deeper understanding of high-dimensional statistical problems. In this article, we consider a class of weighted ℓ1-penalized estimators for convex loss functions of a general form, including the generalized linear models. We study the estimation, prediction, selection and sparsity properties of the weighted ℓ1-penalized estimator in sparse, high-dimensional settings where the number of predictors p can be much larger than the sample size n. Adaptive Lasso is considered as a special case. A multistage method is developed to approximate concave regularized estimation by applying an adaptive Lasso recursively. We provide prediction and estimation oracle inequalities for single- and multi-stage estimators, a general selection consistency theorem, and an upper bound for the dimension of the Lasso estimator. Important models including the linear regression, logistic regression and log-linear models are used throughout to illustrate the applications of the general results. PMID:24348100
Comparison of heaving buoy and oscillating flap wave energy converters
NASA Astrophysics Data System (ADS)
Abu Bakar, Mohd Aftar; Green, David A.; Metcalfe, Andrew V.; Najafian, G.
2013-04-01
Waves offer an attractive source of renewable energy, with relatively low environmental impact, for communities reasonably close to the sea. Two types of simple wave energy converters (WEC), the heaving buoy WEC and the oscillating flap WEC, are studied. Both WECs are considered as simple energy converters because they can be modelled, to a first approximation, as single degree of freedom linear dynamic systems. In this study, we estimate the response of both WECs to typical wave inputs; wave height for the buoy and corresponding wave surge for the flap, using spectral methods. A nonlinear model of the oscillating flap WEC that includes the drag force, modelled by the Morison equation is also considered. The response to a surge input is estimated by discrete time simulation (DTS), using central difference approximations to derivatives. This is compared with the response of the linear model obtained by DTS and also validated using the spectral method. Bendat's nonlinear system identification (BNLSI) technique was used to analyze the nonlinear dynamic system since the spectral analysis was only suitable for linear dynamic system. The effects of including the nonlinear term are quantified.
A Comprehensive Comparison of Relativistic Particle Integrators
NASA Astrophysics Data System (ADS)
Ripperda, B.; Bacchini, F.; Teunissen, J.; Xia, C.; Porth, O.; Sironi, L.; Lapenta, G.; Keppens, R.
2018-03-01
We compare relativistic particle integrators commonly used in plasma physics, showing several test cases relevant for astrophysics. Three explicit particle pushers are considered, namely, the Boris, Vay, and Higuera–Cary schemes. We also present a new relativistic fully implicit particle integrator that is energy conserving. Furthermore, a method based on the relativistic guiding center approximation is included. The algorithms are described such that they can be readily implemented in magnetohydrodynamics codes or Particle-in-Cell codes. Our comparison focuses on the strengths and key features of the particle integrators. We test the conservation of invariants of motion and the accuracy of particle drift dynamics in highly relativistic, mildly relativistic, and non-relativistic settings. The methods are compared in idealized test cases, i.e., without considering feedback onto the electrodynamic fields, collisions, pair creation, or radiation. The test cases include uniform electric and magnetic fields, {\\boldsymbol{E}}× {\\boldsymbol{B}} fields, force-free fields, and setups relevant for high-energy astrophysics, e.g., a magnetic mirror, a magnetic dipole, and a magnetic null. These tests have direct relevance for particle acceleration in shocks and in magnetic reconnection.
Shilton, Katie
2015-02-01
The technical details of Internet architecture affect social debates about privacy and autonomy, intellectual property, cybersecurity, and the basic performance and reliability of Internet services. This paper explores one method for practicing anticipatory ethics in order to understand how a new infrastructure for the Internet might impact these social debates. This paper systematically examines values expressed by an Internet architecture engineering team-the Named Data Networking project-based on data gathered from publications and internal documents. Networking engineers making technical choices also weigh non-technical values when working on Internet infrastructure. Analysis of the team's documents reveals both values invoked in response to technical constraints and possibilities, such as efficiency and dynamism, as well as values, including privacy, security and anonymity, which stem from a concern for personal liberties. More peripheral communitarian values espoused by the engineers include democratization and trust. The paper considers the contextual and social origins of these values, and then uses them as a method of practicing anticipatory ethics: considering the impact such priorities may have on a future Internet.
Structural Design Considerations for an 8-m Space Telescope
NASA Technical Reports Server (NTRS)
Arnold, William R. Sr.; Stahl, H. Philip
2009-01-01
NASA's upcoming ARES V launch vehicle, with its' immerse payload capacities (both volume and mass) has opened the possibilities for a whole new paradigm of space observatories. It becomes practical to consider a monolith mirror of sufficient size to permit significant scientific advantages, both in collection area and smoothness or figure at a reasonable price. The technologies and engineering to manufacture and test 8 meter class monoliths is mature, with nearly a dozen of such mirrors already in operation around the world. This paper will discuss the design requirements to adapt an 8m meniscus mirror into a Space Telescope System, both launch and operational considerations are included. With objects this massive and structurally sensitive, the mirror design must include all stages of the process. Based upon the experiences of the Hubble Space Telescope, testing and verification at both component and integrated system levels are considered vital to mission success. To this end, two different component level test methods for gravity sag (the so call zero- gravity simulation or test mount) are proposed, with one of these methods suitable for the full up system level testing as well.
Structural design considerations for an 8-m space telescope
NASA Astrophysics Data System (ADS)
Arnold, William r., Sr.; Stahl, H. Philip
2009-08-01
NASA's upcoming ARES V launch vehicle, with its' immense payload capacities (both volume and mass) has opened the possibilities for a whole new paradigm of space observatories. It becomes practical to consider a monolith mirror of sufficient size to permit significant scientific advantages, both in collection area and smoothness or figure at a reasonable price. The technologies and engineering to manufacture and test 8 meter class monoliths is mature, with nearly a dozen of such mirrors already in operation around the world. This paper will discuss the design requirements to adapt an 8m meniscus mirror into a Space Telescope System, both launch and operational considerations are included. With objects this massive and structurally sensitive, the mirror design must include all stages of the process. Based upon the experiences of the Hubble Space Telescope, testing and verification at both component and integrated system levels are considered vital to mission success. To this end, two different component level test methods for gravity sag (the so call zero- gravity simulation or test mount) are proposed, with one of these methods suitable for the full up system level testing as well.
Belaid, L; Dumont, A; Chaillet, N; Zertal, A; De Brouwere, V; Hounton, S; Ridde, V
2016-10-01
To synthesise evidence on the implementation, costs and cost-effectiveness of demand generation interventions and their effectiveness in improving uptake of modern contraception methods. A Cochrane systematic review was conducted. Searches were performed in electronic databases (MEDLINE, EMBASE) and the grey literature. Randomised controlled trials, cluster randomised trials and quasi-experimental studies, including controlled before-after studies (CBAs) and cost and cost-effectiveness studies that aimed to assess demand interventions (including community- and facility-based interventions, financial mechanisms and mass media campaigns) in low- and middle-income countries were considered. Meta-analyses and narrative synthesis were conducted. In total, 20 papers meeting the inclusion criteria were included in this review. Of those, 13 were used for meta-analysis. Few data were available on implementation and on the influence of context on demand interventions. Involving family members during counselling, providing education activities and increasing exposure to those activities could enhance the success of demand interventions. Demand generation interventions were positively associated with increases in current use (pooled OR 1.57; 95% CI: 1.46-1.69, P < 0.01). Financial mechanism interventions (vouchers) appeared effective to increase use of modern contraceptive methods (pooled OR 2.16; 95% CI: 1.91-2.45, P < 0.01; I 2 = 0%). Demand interventions improved knowledge (pooled OR 1.02; 95% CI 0.63-1.64, P = 0.93) and attitudes towards family planning and improved discussion with partners/husbands around modern contraceptive methods. However, given the limited number of studies included in each category of demand generation interventions, the dates of publication of the studies and their low quality, caution is advised in considering the results. Very limited evidence was available on costs; studies including data on costs were old and inconsistent. Demand generation interventions contribute to increases in modern contraceptive methods use. However, more studies with robust designs are needed to identify the most effective demand generation intervention to increase uptake of modern contraceptive methods. More evidence is also needed about implementation, costs and cost-effectiveness to inform decisions on sustainability and scaling-up. © 2016 The Authors. Tropical Medicine & International Health Published by John Wiley & Sons Ltd.
Pollitzer, R.
1956-01-01
The first portion of this study describes in detail the different aspects of stool examinations, including the collection, preservation, and pooling of specimens, macroscopic and bacterioscopic examination, enrichment methods, and cultivation on a variety of solid media. The author also deals with the examination of vomits and of water. The performance and value of different identification tests (agglutination, haemolysis, and bacteriophage) and confirmatory tests are then considered. An annex is included on bacteriological procedures in the laboratory diagnosis of cholera. PMID:13356145
Nanostructured silver sulfide: synthesis of various forms and their application
NASA Astrophysics Data System (ADS)
Sadovnikov, S. I.; Rempel, A. A.; Gusev, A. I.
2018-04-01
The results of experimental studies on nanostructured silver sulfide are analyzed and generalized. The influence of small particle size on nonstoichiometry of silver sulfide is discussed. Methods for the synthesis of various forms of nanostructured Ag2S including nanopowders, stable colloidal solutions, quantum dots, core–shell nanoparticles and heteronanostructures are described. The advantages and drawbacks of different synthetic procedures are analyzed. Main fields of application of nanostructured silver sulfide are considered. The bibliography includes 184 references.
Direct conversion of nuclear radiation energy
DOE Office of Scientific and Technical Information (OSTI.GOV)
Miley, George H.
1970-01-01
This book presents a comprehensive study of methods for converting nuclear radiationi directly without resorting to a heat cycle. The concepts discussed primarily involve direct collection of charged particles released by radioisotopes and by nuclear and thermonuclear reactors. Areas considered include basic energy conversion, charged-particle transport theory, secondary-electron emission, and leakage currents and associated problems. Applications to both nuclear instrumentaion and power sources are discussed. Problems are also included as an aid to the reader or for classroom use.
Optimal solar sail planetocentric trajectories
NASA Technical Reports Server (NTRS)
Sackett, L. L.
1977-01-01
The analysis of solar sail planetocentric optimal trajectory problem is described. A computer program was produced to calculate optimal trajectories for a limited performance analysis. A square sail model is included and some consideration is given to a heliogyro sail model. Orbit to a subescape point and orbit to orbit transfer are considered. Trajectories about the four inner planets can be calculated and shadowing, oblateness, and solar motion may be included. Equinoctial orbital elements are used to avoid the classical singularities, and the method of averaging is applied to increase computational speed. Solution of the two-point boundary value problem which arises from the application of optimization theory is accomplished with a Newton procedure. Time optimal trajectories are emphasized, but a penalty function has been considered to prevent trajectories which intersect a planet's surface.
NASA Astrophysics Data System (ADS)
Salman Shahid, Syed; Bikson, Marom; Salman, Humaira; Wen, Peng; Ahfock, Tony
2014-06-01
Objectives. Computational methods are increasingly used to optimize transcranial direct current stimulation (tDCS) dose strategies and yet complexities of existing approaches limit their clinical access. Since predictive modelling indicates the relevance of subject/pathology based data and hence the need for subject specific modelling, the incremental clinical value of increasingly complex modelling methods must be balanced against the computational and clinical time and costs. For example, the incorporation of multiple tissue layers and measured diffusion tensor (DTI) based conductivity estimates increase model precision but at the cost of clinical and computational resources. Costs related to such complexities aggregate when considering individual optimization and the myriad of potential montages. Here, rather than considering if additional details change current-flow prediction, we consider when added complexities influence clinical decisions. Approach. Towards developing quantitative and qualitative metrics of value/cost associated with computational model complexity, we considered field distributions generated by two 4 × 1 high-definition montages (m1 = 4 × 1 HD montage with anode at C3 and m2 = 4 × 1 HD montage with anode at C1) and a single conventional (m3 = C3-Fp2) tDCS electrode montage. We evaluated statistical methods, including residual error (RE) and relative difference measure (RDM), to consider the clinical impact and utility of increased complexities, namely the influence of skull, muscle and brain anisotropic conductivities in a volume conductor model. Main results. Anisotropy modulated current-flow in a montage and region dependent manner. However, significant statistical changes, produced within montage by anisotropy, did not change qualitative peak and topographic comparisons across montages. Thus for the examples analysed, clinical decision on which dose to select would not be altered by the omission of anisotropic brain conductivity. Significance. Results illustrate the need to rationally balance the role of model complexity, such as anisotropy in detailed current flow analysis versus value in clinical dose design. However, when extending our analysis to include axonal polarization, the results provide presumably clinically meaningful information. Hence the importance of model complexity may be more relevant with cellular level predictions of neuromodulation.
Kim, Soo-Yeon; Kim, Eun-Kyung; Kwak, Jin Young; Moon, Hee Jung; Yoon, Jung Hyun
2015-02-01
BRAF(V600E) mutation analysis has been used as a complementary diagnostic tool to ultrasonography-guided, fine-needle aspiration (US-FNA) in the diagnosis of thyroid nodule with high specificity reported up to 100%. When highly sensitive analytic methods are used, however, false-positive results of BRAF(V600E) mutation analysis have been reported. In this study, we investigated the clinical, US features, and outcome of patients with thyroid nodules with benign cytology but positive BRAF(V600E) mutation using highly sensitive analytic methods from US-FNA. This study included 22 nodules in 22 patients (3 men, 19 women; mean age, 53 years) with benign cytology but positive BRAF(V600E) mutation from US-FNA. US features were categorized according to the internal components, echogenicity, margin, calcifications, and shape. Suspicious US features included markedly hypoechogenicity, noncircumscribed margins, micro or mixed calcifications, and nonparallel shape. Nodules were considered to have either concordant or discordant US features to benign cytology. Medical records and imaging studies were reviewed for final cytopathology results and outcomes during follow-up. Among the 22 nodules, 17 nodules were reviewed. Fifteen of 17 nodules were malignant, and 2 were benign. The benign nodules were confirmed as adenomatous hyperplasia with underlying lymphocytic thyroiditis and a fibrotic nodule with dense calcification. Thirteen of the 15 malignant nodules had 2 or more suspicious US features, and all 15 nodules were considered to have discordant cytology considering suspicious US features. Five nodules had been followed with US or US-FNA without resection, and did not show change in size or US features on follow-up US examinations. BRAF(V600E) mutation analysis is a highly sensitive diagnostic tool in the diagnosis of papillary thyroid carcinomas. In the management of thyroid nodules with benign cytology but positive BRAF(V600E) mutation, thyroidectomy should be considered in nodules which have 2 or more suspicious US features and are considered discordant on image-cytology correlation. Copyright © 2015 Elsevier Inc. All rights reserved.
Nanomaterials as stationary phases and supports in liquid chromatography.
Beeram, Sandya R; Rodriguez, Elliott; Doddavenkatanna, Suresh; Li, Zhao; Pekarek, Allegra; Peev, Darin; Goerl, Kathryn; Trovato, Gianfranco; Hofmann, Tino; Hage, David S
2017-10-01
The development of various nanomaterials over the last few decades has led to many applications for these materials in liquid chromatography (LC). This review will look at the types of nanomaterials that have been incorporated into LC systems and the applications that have been explored for such systems. A number of carbon-based nanomaterials and inorganic nanomaterials have been considered for use in LC, ranging from carbon nanotubes, fullerenes and nanodiamonds to metal nanoparticles and nanostructures based on silica, alumina, zirconia and titanium dioxide. Many ways have been described for incorporating these nanomaterials into LC systems. These methods have included covalent immobilization, adsorption, entrapment, and the synthesis or direct development of nanomaterials as part of a chromatographic support. Nanomaterials have been used in many types of LC. These applications have included the reversed-phase, normal-phase, ion-exchange, and affinity modes of LC, as well as related methods such as chiral separations, ion-pair chromatography and hydrophilic interaction liquid chromatography. Both small and large analytes (e.g., dyes, drugs, amino acids, peptides and proteins) have been used to evaluate possible applications for these nanomaterial-based methods. The use of nanomaterials in columns, capillaries and planar chromatography has been considered as part of these efforts. Potential advantages of nanomaterials in these applications have included their good chemical and physical stabilities, the variety of interactions many nanomaterials can have with analytes, and their unique retention properties in some separation formats. © 2017 WILEY-VCH Verlag GmbH & Co. KGaA, Weinheim.
Water demand forecasting: review of soft computing methods.
Ghalehkhondabi, Iman; Ardjmand, Ehsan; Young, William A; Weckman, Gary R
2017-07-01
Demand forecasting plays a vital role in resource management for governments and private companies. Considering the scarcity of water and its inherent constraints, demand management and forecasting in this domain are critically important. Several soft computing techniques have been developed over the last few decades for water demand forecasting. This study focuses on soft computing methods of water consumption forecasting published between 2005 and 2015. These methods include artificial neural networks (ANNs), fuzzy and neuro-fuzzy models, support vector machines, metaheuristics, and system dynamics. Furthermore, it was discussed that while in short-term forecasting, ANNs have been superior in many cases, but it is still very difficult to pick a single method as the overall best. According to the literature, various methods and their hybrids are applied to water demand forecasting. However, it seems soft computing has a lot more to contribute to water demand forecasting. These contribution areas include, but are not limited, to various ANN architectures, unsupervised methods, deep learning, various metaheuristics, and ensemble methods. Moreover, it is found that soft computing methods are mainly used for short-term demand forecasting.
NASA Astrophysics Data System (ADS)
Wang, Lingquan; Zeng, Zhong; Zhang, Liangqi; Qiao, Long; Zhang, Yi; Lu, Yiyu
2018-04-01
Navier-Stokes (NS) equations with no-slip boundary conditions fail to realistically describe micro-flows with considering nanoscale phenomena. Particularly, in kerogen pores, slip-flow and surface diffusion are important. In this study, we propose a new slip boundary scheme for the lattice Boltzmann (LB) method through the non-equilibrium extrapolation scheme to simulate the slip-flow considering surface diffusion effect. Meanwhile, the second-order slip velocity can be taken into account. The predicted characteristics in a two-dimensional micro-flow, including slip-velocity, velocity distribution along the flow direction with/without surface diffusion are present. The results in this study are compared with available analytical and reference results, and good agreements are achieved.
Vasilyev, K N
2013-01-01
When developing new software products and adapting existing software, project leaders have to decide which functionalities to keep, adapt or develop. They have to consider that the cost of making errors during the specification phase is extremely high. In this paper a formalised approach is proposed that considers the main criteria for selecting new software functions. The application of this approach minimises the chances of making errors in selecting the functions to apply. Based on the work on software development and support projects in the area of water resources and flood damage evaluation in economic terms at CH2M HILL (the developers of the flood modelling package ISIS), the author has defined seven criteria for selecting functions to be included in a software product. The approach is based on the evaluation of the relative significance of the functions to be included into the software product. Evaluation is achieved by considering each criterion and the weighting coefficients of each criterion in turn and applying the method of normalisation. This paper includes a description of this new approach and examples of its application in the development of new software products in the are of the water resources management.
Methods Used in Economic Evaluations of Chronic Kidney Disease Testing — A Systematic Review
Sutton, Andrew J.; Breheny, Katie; Deeks, Jon; Khunti, Kamlesh; Sharpe, Claire; Ottridge, Ryan S.; Stevens, Paul E.; Cockwell, Paul; Kalra, Philp A.; Lamb, Edmund J.
2015-01-01
Background The prevalence of chronic kidney disease (CKD) is high in general populations around the world. Targeted testing and screening for CKD are often conducted to help identify individuals that may benefit from treatment to ameliorate or prevent their disease progression. Aims This systematic review examines the methods used in economic evaluations of testing and screening in CKD, with a particular focus on whether test accuracy has been considered, and how analysis has incorporated issues that may be important to the patient, such as the impact of testing on quality of life and the costs they incur. Methods Articles that described model-based economic evaluations of patient testing interventions focused on CKD were identified through the searching of electronic databases and the hand searching of the bibliographies of the included studies. Results The initial electronic searches identified 2,671 papers of which 21 were included in the final review. Eighteen studies focused on proteinuria, three evaluated glomerular filtration rate testing and one included both tests. The full impact of inaccurate test results was frequently not considered in economic evaluations in this setting as a societal perspective was rarely adopted. The impact of false positive tests on patients in terms of the costs incurred in re-attending for repeat testing, and the anxiety associated with a positive test was almost always overlooked. In one study where the impact of a false positive test on patient quality of life was examined in sensitivity analysis, it had a significant impact on the conclusions drawn from the model. Conclusion Future economic evaluations of kidney function testing should examine testing and monitoring pathways from the perspective of patients, to ensure that issues that are important to patients, such as the possibility of inaccurate test results, are properly considered in the analysis. PMID:26465773
A systematic review of pseudophakic monovision for presbyopia correction
Labiris, Georgios; Toli, Aspa; Perente, Aslin; Ntonti, Panagiota; Kozobolis, Vassilios P.
2017-01-01
A systematic review of the recent literature regarding pseudophakic monovision as a reliable methods for presbyopia correction was performed based on the PubMed, MEDLINE, Nature and the American Academy of Ophthalmology databases in July 2015 and data from 18 descriptive and 12 comparative studies were included in this narrative review. Pseudophakic monosvision seems to be an effective method for presbyopia with high rates of spectacles independence and minimal dysphotopsia side-effects, that should be considered by the modern cataract surgeons. PMID:28730093
NASA Astrophysics Data System (ADS)
Alekseev, É. I.; Bazarov, E. N.
1992-09-01
A theoretical justification is given of the widely used method of stabilization of the output signal from a fiber-optic gyroscope with a broad-band radiation source by a Lyot depolarizer. Different variants of including a depolarizer in such a gyroscope are considered and the role of the dichroism and birefringence induced in the gyroscope system is discussed.
Study of alternate methods of disposal of propellants and gases at KSC
NASA Technical Reports Server (NTRS)
Moore, W. I.
1970-01-01
A comprehensive study was conducted at KSC launch support facilities to determine the nature and extent of potential hazards from propellant and gas releases to the environment. The results of the study, alternate methods for reducing or eliminating the hazards, and recommendations pertaining to these alternatives are presented. The operational modes of the propellant or hazardous gas systems considered include: system charging, system standby, system operation, and post-test operations. The results are outlined on an area-by-area basis.
On Solutions for the Transient Response of Beams
NASA Technical Reports Server (NTRS)
Leonard, Robert W.
1959-01-01
Williams type modal solutions of the elementary and Timoshenko beam equations are presented for the response of several uniform beams to a general applied load. Example computations are shown for a free-free beam subject to various concentrated loads at its center. Discussion includes factors influencing the convergence of modal solutions and factors to be considered in a choice of beam theory. Results obtained by two numerical procedures, the traveling-wave method and Houbolt's method, are also presented and discussed.
Dolansky, Mary A; Schexnayder, Julie; Patrician, Patricia A; Sales, Anne
Although quality and safety competencies were developed and disseminated nearly a decade ago by the Quality and Safety Education for Nurses (QSEN) project, the uptake in schools of nursing has been slow. The use of implementation science methods may be useful to accelerate quality and safety competency integration in nursing education. The article includes a definition and description of implementation science methods and practical implementation strategies for nurse educators to consider when integrating the QSEN competencies into nursing curriculum.
Identification of cutting force coefficients in machining process considering cutter vibration
NASA Astrophysics Data System (ADS)
Yao, Qi; Luo, Ming; Zhang, Dinghua; Wu, Baohai
2018-03-01
Among current cutting force models, cutting force coefficients still are the foundation of predicting calculation combined with consideration of geometry engagement variation, equipment characteristics, material properties and so on. Attached with unimpeachable significance, the traditional and some novel identification methods of cutting force coefficient are still faced with trouble, including repeated onerous work, over ideal measuring condition, variation of value due to material divergence, interference from measuring units. To utilize the large amount of data from real manufacturing section, enlarge data sources and enrich cutting data base for former prediction task, a novel identification method is proposed by considering stiffness properties of the cutter-holder-spindle system in this paper. According to previously proposed studies, the direct result of cutter vibration is the form of dynamic undeformed chip thickness. This fluctuation is considered in two stages of this investigation. Firstly, a cutting force model combined with cutter vibration is established in detailed way. Then, on the foundation of modeling, a novel identification method is developed, in which the dynamic undeformed chip thickness could be obtained by using collected data. In a carefully designed experiment procedure, the reliability of model is validated by comparing predicted and measured results. Under different cutting condition and cutter stiffness, data is collected for the justification of identification method. The results showed divergence in calculated coefficients is acceptable confirming the possibility of accomplishing targets by applying this new method. In discussion, the potential directions of improvement are proposed.
Analysis of survival data from telemetry projects
Bunck, C.M.; Winterstein, S.R.; Pollock, K.H.
1985-01-01
Telemetry techniques can be used to study the survival rates of animal populations and are particularly suitable for species or settings for which band recovery models are not. Statistical methods for estimating survival rates and parameters of survival distributions from observations of radio-tagged animals will be described. These methods have been applied to medical and engineering studies and to the study of nest success. Estimates and tests based on discrete models, originally introduced by Mayfield, and on continuous models, both parametric and nonparametric, will be described. Generalizations, including staggered entry of subjects into the study and identification of mortality factors will be considered. Additional discussion topics will include sample size considerations, relocation frequency for subjects, and use of covariates.
NASA Astrophysics Data System (ADS)
Bernardes, S.; Cotten, D. L.
2016-12-01
University-based satellite programs have been successfully used as a platform for teaching STEM related fields, bringing tremendous benefits to graduate and undergraduate education. Considering their infrastructure and curricula, tech schools have traditionally been considered logical candidates for hosting such programs. More recently, with the dissemination of small satellites initiatives, non-tech schools have been presented the opportunity of developing satellite design and implementation programs. This work reports on the experiences and challenges associated with implementing a satellite program at the University of Georgia (UGA), a non-tech university. With funding from the Air Force Research Laboratory's (AFRL) University Nanosat Program (UNP) and NASA's Undergraduate Student Instrument Project (USIP) a team of undergraduates at UGA has recently been tasked with building two small satellites and helping to create a Small Satellite Research Laboratory (SSRL) at the university. Unique features of the satellite program at UGA include its team of students from a broad range of backgrounds and departments (Engineering, Computer Science, Art, Business, and Geography) and the previous exposure of many of these students to synergistic technologies, including arduino and unmanned aerial systems. We show how informal exposure to those technologies and willingness of students to focus on areas outside of their field of study can benefit from the implementation of satellite programs. In this regard, we report on methods and techniques used to find and recruit driven and knowledgeable students to work in a high paced field such as satellite system integration. We show how students and faculty from multiple departments have collaborated to reach a common, far reaching goal and describe our proposed methods to evaluate and measure educational goals based around SSRL and its projects. We also present the challenges associated with the lack of a developed engineering program, including our solutions to a shortage of equipment and expertise regarding building satellite systems and a satellite laboratory. Finally, we our outreach methods, including K-12, and share our experience and successes finding industry partners, considering an absence of background in the field and prior collaborations.
Methods of geometrical integration in accelerator physics
NASA Astrophysics Data System (ADS)
Andrianov, S. N.
2016-12-01
In the paper we consider a method of geometric integration for a long evolution of the particle beam in cyclic accelerators, based on the matrix representation of the operator of particles evolution. This method allows us to calculate the corresponding beam evolution in terms of two-dimensional matrices including for nonlinear effects. The ideology of the geometric integration introduces in appropriate computational algorithms amendments which are necessary for preserving the qualitative properties of maps presented in the form of the truncated series generated by the operator of evolution. This formalism extends both on polarized and intense beams. Examples of practical applications are described.
Unified Description of Inelastic Propensity Rules for Electron Transport through Nanoscale Junctions
NASA Astrophysics Data System (ADS)
Paulsson, Magnus; Frederiksen, Thomas; Ueba, Hiromu; Lorente, Nicolás; Brandbyge, Mads
2008-06-01
We present a method to analyze the results of first-principles based calculations of electronic currents including inelastic electron-phonon effects. This method allows us to determine the electronic and vibrational symmetries in play, and hence to obtain the so-called propensity rules for the studied systems. We show that only a few scattering states—namely those belonging to the most transmitting eigenchannels—need to be considered for a complete description of the electron transport. We apply the method on first-principles calculations of four different systems and obtain the propensity rules in each case.
NASA Technical Reports Server (NTRS)
Nakagawa, Y.
1981-01-01
The method described as the method of nearcharacteristics by Nakagawa (1980) is renamed the method of projected characteristics. Making full use of properties of the projected characteristics, a new and simpler formulation is developed. As a result, the formulation for the examination of the general three-dimensional problems is presented. It is noted that since in practice numerical solutions must be obtained, the final formulation is given in the form of difference equations. The possibility of including effects of viscous and ohmic dissipations in the formulation is considered, and the physical interpretation is discussed. A systematic manner is then presented for deriving physically self-consistent, time-dependent boundary equations for MHD initial boundary problems. It is demonstrated that the full use of the compatibility equations (differential equations relating variations at two spatial locations and times) is required in determining the time-dependent boundary conditions. In order to provide a clear physical picture as an example, the evolution of axisymmetric global magnetic field by photospheric differential rotation is considered.
Greenwald, Ralf R.; Quitadamo, Ian J.
2014-01-01
A changing undergraduate demographic and the need to help students develop advanced critical thinking skills in neuroanatomy courses has prompted many faculty to consider new teaching methods including clinical case studies. This study compared primarily conventional and inquiry-based clinical case (IBCC) teaching methods to determine which would produce greater gains in critical thinking and content knowledge. Results showed students in the conventional neuroanatomy course gained less than 3 national percentile ranks while IBCC students gained over 7.5 within one academic term using the valid and reliable California Critical Thinking Skills Test. In addition to 2.5 times greater gains in critical thinking, IBCC teaching methods also produced 12% greater final exam performance and 11% higher grades using common grade performance benchmarks. Classroom observations also indicated that IBCC students were more intellectually engaged and participated to a greater extent in classroom discussions. Through the results of this study, it is hoped that faculty who teach neuroanatomy and desire greater critical thinking and content student learning outcomes will consider using the IBCC method. PMID:24693256
Greenwald, Ralf R; Quitadamo, Ian J
2014-01-01
A changing undergraduate demographic and the need to help students develop advanced critical thinking skills in neuroanatomy courses has prompted many faculty to consider new teaching methods including clinical case studies. This study compared primarily conventional and inquiry-based clinical case (IBCC) teaching methods to determine which would produce greater gains in critical thinking and content knowledge. Results showed students in the conventional neuroanatomy course gained less than 3 national percentile ranks while IBCC students gained over 7.5 within one academic term using the valid and reliable California Critical Thinking Skills Test. In addition to 2.5 times greater gains in critical thinking, IBCC teaching methods also produced 12% greater final exam performance and 11% higher grades using common grade performance benchmarks. Classroom observations also indicated that IBCC students were more intellectually engaged and participated to a greater extent in classroom discussions. Through the results of this study, it is hoped that faculty who teach neuroanatomy and desire greater critical thinking and content student learning outcomes will consider using the IBCC method.
Advances in the microrheology of complex fluids
NASA Astrophysics Data System (ADS)
Waigh, Thomas Andrew
2016-07-01
New developments in the microrheology of complex fluids are considered. Firstly the requirements for a simple modern particle tracking microrheology experiment are introduced, the error analysis methods associated with it and the mathematical techniques required to calculate the linear viscoelasticity. Progress in microrheology instrumentation is then described with respect to detectors, light sources, colloidal probes, magnetic tweezers, optical tweezers, diffusing wave spectroscopy, optical coherence tomography, fluorescence correlation spectroscopy, elastic- and quasi-elastic scattering techniques, 3D tracking, single molecule methods, modern microscopy methods and microfluidics. New theoretical techniques are also reviewed such as Bayesian analysis, oversampling, inversion techniques, alternative statistical tools for tracks (angular correlations, first passage probabilities, the kurtosis, motor protein step segmentation etc), issues in micro/macro rheological agreement and two particle methodologies. Applications where microrheology has begun to make some impact are also considered including semi-flexible polymers, gels, microorganism biofilms, intracellular methods, high frequency viscoelasticity, comb polymers, active motile fluids, blood clots, colloids, granular materials, polymers, liquid crystals and foods. Two large emergent areas of microrheology, non-linear microrheology and surface microrheology are also discussed.
2015-01-01
Background In recent years, with advances in techniques for protein structure analysis, the knowledge about protein structure and function has been published in a vast number of articles. A method to search for specific publications from such a large pool of articles is needed. In this paper, we propose a method to search for related articles on protein structure analysis by using an article itself as a query. Results Each article is represented as a set of concepts in the proposed method. Then, by using similarities among concepts formulated from databases such as Gene Ontology, similarities between articles are evaluated. In this framework, the desired search results vary depending on the user's search intention because a variety of information is included in a single article. Therefore, the proposed method provides not only one input article (primary article) but also additional articles related to it as an input query to determine the search intention of the user, based on the relationship between two query articles. In other words, based on the concepts contained in the input article and additional articles, we actualize a relevant literature search that considers user intention by varying the degree of attention given to each concept and modifying the concept hierarchy graph. Conclusions We performed an experiment to retrieve relevant papers from articles on protein structure analysis registered in the Protein Data Bank by using three query datasets. The experimental results yielded search results with better accuracy than when user intention was not considered, confirming the effectiveness of the proposed method. PMID:25952498
Living systematic reviews: 3. Statistical methods for updating meta-analyses.
Simmonds, Mark; Salanti, Georgia; McKenzie, Joanne; Elliott, Julian
2017-11-01
A living systematic review (LSR) should keep the review current as new research evidence emerges. Any meta-analyses included in the review will also need updating as new material is identified. If the aim of the review is solely to present the best current evidence standard meta-analysis may be sufficient, provided reviewers are aware that results may change at later updates. If the review is used in a decision-making context, more caution may be needed. When using standard meta-analysis methods, the chance of incorrectly concluding that any updated meta-analysis is statistically significant when there is no effect (the type I error) increases rapidly as more updates are performed. Inaccurate estimation of any heterogeneity across studies may also lead to inappropriate conclusions. This paper considers four methods to avoid some of these statistical problems when updating meta-analyses: two methods, that is, law of the iterated logarithm and the Shuster method control primarily for inflation of type I error and two other methods, that is, trial sequential analysis and sequential meta-analysis control for type I and II errors (failing to detect a genuine effect) and take account of heterogeneity. This paper compares the methods and considers how they could be applied to LSRs. Copyright © 2017 Elsevier Inc. All rights reserved.
a Comparison of Empirical and Inteligent Methods for Dust Detection Using Modis Satellite Data
NASA Astrophysics Data System (ADS)
Shahrisvand, M.; Akhoondzadeh, M.
2013-09-01
Nowadays, dust storm in one of the most important natural hazards which is considered as a national concern in scientific communities. This paper considers the capabilities of some classical and intelligent methods for dust detection from satellite imagery around the Middle East region. In the study of dust detection, MODIS images have been a good candidate due to their suitable spectral and temporal resolution. In this study, physical-based and intelligent methods including decision tree, ANN (Artificial Neural Network) and SVM (Support Vector Machine) have been applied to detect dust storms. Among the mentioned approaches, in this paper, SVM method has been implemented for the first time in domain of dust detection studies. Finally, AOD (Aerosol Optical Depth) images, which are one the referenced standard products of OMI (Ozone Monitoring Instrument) sensor, have been used to assess the accuracy of all the implemented methods. Since the SVM method can distinguish dust storm over lands and oceans simultaneously, therefore the accuracy of SVM method is achieved better than the other applied approaches. As a conclusion, this paper shows that SVM can be a powerful tool for production of dust images with remarkable accuracy in comparison with AOT (Aerosol Optical Thickness) product of NASA.
A decade of individual participant data meta-analyses: A review of current practice.
Simmonds, Mark; Stewart, Gavin; Stewart, Lesley
2015-11-01
Individual participant data (IPD) systematic reviews and meta-analyses are often considered to be the gold standard for meta-analysis. In the ten years since the first review into the methodology and reporting practice of IPD reviews was published much has changed in the field. This paper investigates current reporting and statistical practice in IPD systematic reviews. A systematic review was performed to identify systematic reviews that collected and analysed IPD. Data were extracted from each included publication on a variety of issues related to the reporting of IPD review process, and the statistical methods used. There has been considerable growth in the use of "one-stage" methods to perform IPD meta-analyses. The majority of reviews consider at least one covariate other than the primary intervention, either using subgroup analysis or including covariates in one-stage regression models. Random-effects analyses, however, are not often used. Reporting of review methods was often limited, with few reviews presenting a risk-of-bias assessment. Details on issues specific to the use of IPD were little reported, including how IPD were obtained; how data was managed and checked for consistency and errors; and for how many studies and participants IPD were sought and obtained. While the last ten years have seen substantial changes in how IPD meta-analyses are performed there remains considerable scope for improving the quality of reporting for both the process of IPD systematic reviews, and the statistical methods employed in them. It is to be hoped that the publication of the PRISMA-IPD guidelines specific to IPD reviews will improve reporting in this area. Copyright © 2015 Elsevier Inc. All rights reserved.
Hall, Leslie; Jude, Kurt P; Clark, Shirley L; Wengenack, Nancy L
2011-06-24
The rapid detection of antimicrobial resistance is important in the effort to control the increase in resistant Mycobacterium tuberculosis (Mtb). Antimicrobial susceptibility testing (AST) of Mtb has traditionally been performed by the agar method of proportion or by macrobroth testing on an instrument such as the BACTEC (Becton Dickinson, Sparks, MD), VersaTREK (TREK Diagnostics, Cleveland, OH) or BacT/ALERT (bioMérieux, Hazelwood, MO). The agar proportion method, while considered the "gold" standard of AST, is labor intensive and requires calculation of resistance by performing colony counts on drug-containing agar as compared to drug-free agar. If there is ≥1% growth on the drug-containing medium as compared to drug-free medium, the organism is considered resistant to that drug. The macrobroth methods require instrumentation and test break point ("critical") drug concentrations for the first line drugs (isoniazid, ethambutol, rifampin, and pyrazinamide). The method described here is commercially available in a 96 well microtiter plate format [MYCOTB (TREK Diagnostics)] and contains increasing concentrations of 12 antimicrobials used for treatment of tuberculosis including both first (isoniazid, rifampin, ethambutol) and second line drugs (amikacin, cycloserine, ethionamide, kanamycin, moxifloxacin, ofloxacin, para-aminosalicylic acid, rifabutin, and streptomycin). Pyrazinamide, a first line drug, is not included in the microtiter plate due to its need for acidic test conditions. Advantages of the microtiter system include both ease of set up and faster turn around time (14 days) compared with traditional agar proportion (21 days). In addition, the plate can be set up from inoculum prepared using either broth or solid medium. Since the microtiter plate format is new and since Mtb presents unique safety challenges in the laboratory, this protocol will describe how to safely setup, incubate and read the microtiter plate.
NASA Astrophysics Data System (ADS)
Wang, Yunong; Cheng, Rongjun; Ge, Hongxia
2017-08-01
In this paper, a lattice hydrodynamic model is derived considering not only the effect of flow rate difference but also the delayed feedback control signal which including more comprehensive information. The control method is used to analyze the stability of the model. Furthermore, the critical condition for the linear steady traffic flow is deduced and the numerical simulation is carried out to investigate the advantage of the proposed model with and without the effect of flow rate difference and the control signal. The results are consistent with the theoretical analysis correspondingly.
Supplier selection based on complex indicator of finished products quality
NASA Astrophysics Data System (ADS)
Chernikova, Anna; Golovkina, Svetlana; Kuzmina, Svetlana; Demenchenok, Tatiana
2017-10-01
In the article the authors consider possible directions of solving problems when selecting a supplier for deliveries of raw materials and materials of an industrial enterprise, possible difficulties are analyzed and ways of their solution are suggested. Various methods are considered to improve the efficiency of the supplier selection process based on the analysis of the paper bags supplier selection process for the needs of the construction company. In the article the calculation of generalized indicators and complex indicator, which should include single indicators, formed in groups that reflect different aspects of quality, is presented.
Methods for converting industrial zones
NASA Astrophysics Data System (ADS)
Talipova, L.; Kosyakov, E.; Polyakova, Irina
2017-10-01
In this article, industrial zones of Saint Petersburg and Hong Kong were considered. Competitive projects aimed at developing the grey belt of Saint Petersburg were considered. The methodology of the survey of reconstruction of the industrial zone of Hong Kong is also analyzed. The potential of the city’s grey belt lies in its location on the border of the city’s historical centre. Rational use of this potential will make it possible to achieve numerous objectives, including development of the city’s transport infrastructure, positioning of business functions, and organization of housing and the city’s system of green public spaces.
Symmetrical Josephson vortex interferometer as an advanced ballistic single-shot detector
DOE Office of Scientific and Technical Information (OSTI.GOV)
Soloviev, I. I., E-mail: isol@phys.msu.ru; Lukin Scientific Research Institute of Physical Problems, 124460 Zelenograd, Moscow; Laboratory of Cryogenic Nanoelectronics, Nizhny Novgorod State Technical University n.a. R.E. Alekseev, 603950 Nizhny Novgorod
2014-11-17
We consider a ballistic detector formed in an interferometer manner which operational principle relies on Josephson vortex scattering at a measurement potential. We propose an approach to symmetrize the detector scheme and explore arising advantages in the signal-to-noise ratio and in the back-action on a measured object by means of recently presented numerical and analytical methods for modeling of a soliton scattering dynamics in the presence of thermal fluctuations. The obtained characteristics for experimentally relevant parameters reveal practical applicability of the considered schemes including possibility of coupling with standard digital rapid single flux quantum circuits.
NASA Technical Reports Server (NTRS)
Poultney, S. K.
1971-01-01
The behavior of the photomultiplier is considered, as well as the method of derivation of the photomultiplier output pulse and its relation to the reflected light pulse width and amplitude, and the calibration of range precision and accuracy. Pulsed laser radars with light pulse widths of 30, 3, and 0.1 nanosec a considered, with the 0.1 nanosec system capable of highest precision in several modes of operation, including a high repetition rate, single photoelectron reception mode. An alternate calibration scheme using a fast, triggerable light pulser is described in detail.
NASA Astrophysics Data System (ADS)
Parkin, G.; O'Donnell, G.; Ewen, J.; Bathurst, J. C.; O'Connell, P. E.; Lavabre, J.
1996-02-01
Validation methods commonly used to test catchment models are not capable of demonstrating a model's fitness for making predictions for catchments where the catchment response is not known (including hypothetical catchments, and future conditions of existing catchments which are subject to land-use or climate change). This paper describes the first use of a new method of validation (Ewen and Parkin, 1996. J. Hydrol., 175: 583-594) designed to address these types of application; the method involves making 'blind' predictions of selected hydrological responses which are considered important for a particular application. SHETRAN (a physically based, distributed catchment modelling system) is tested on a small Mediterranean catchment. The test involves quantification of the uncertainty in four predicted features of the catchment response (continuous hydrograph, peak discharge rates, monthly runoff, and total runoff), and comparison of observations with the predicted ranges for these features. The results of this test are considered encouraging.
Dissemination of psychosocial treatments for anxiety: the importance of taking a broad perspective.
Taylor, Steven; Abramowitz, Jonathan S
2013-12-01
Dissemination methods are used to increase the likelihood that a given treatment or form of clinical practice is implemented by clinicians in the community. Therapist training in treatment methods is an important component of dissemination. Successful dissemination also requires that roadblocks to treatment implementation are identified and circumvented, such as misconceptions that clinicians might hold about a given treatment. The present article offers a commentary on the papers included in the special issue on treatment dissemination for anxiety disorders. Most papers focus on issues concerning the training and education of clinicians with regard to exposure therapy. Training and education is an important step but should be part of a broad, multifaceted approach. There are several other important methods of treatment dissemination, including methods developed and implemented with success by the pharmaceutical industry, might also be used to disseminate psychosocial therapies. Optimal dissemination likely requires a broad perspective in which multiple dissemination methods are considered for implementation. Copyright © 2013 Elsevier Ltd. All rights reserved.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Mcwilliams, A. J.
2015-09-08
This report reviews literature on reprocessing high temperature gas-cooled reactor graphite fuel components. A basic review of the various fuel components used in the pebble bed type reactors is provided along with a survey of synthesis methods for the fabrication of the fuel components. Several disposal options are considered for the graphite pebble fuel elements including the storage of intact pebbles, volume reduction by separating the graphite from fuel kernels, and complete processing of the pebbles for waste storage. Existing methods for graphite removal are presented and generally consist of mechanical separation techniques such as crushing and grinding chemical techniquesmore » through the use of acid digestion and oxidation. Potential methods for reprocessing the graphite pebbles include improvements to existing methods and novel technologies that have not previously been investigated for nuclear graphite waste applications. The best overall method will be dependent on the desired final waste form and needs to factor in the technical efficiency, political concerns, cost, and implementation.« less
The assessment of cognitive errors using an observer-rated method.
Drapeau, Martin
2014-01-01
Cognitive Errors (CEs) are a key construct in cognitive behavioral therapy (CBT). Integral to CBT is that individuals with depression process information in an overly negative or biased way, and that this bias is reflected in specific depressotypic CEs which are distinct from normal information processing. Despite the importance of this construct in CBT theory, practice, and research, few methods are available to researchers and clinicians to reliably identify CEs as they occur. In this paper, the author presents a rating system, the Cognitive Error Rating Scale, which can be used by trained observers to identify and assess the cognitive errors of patients or research participants in vivo, i.e., as they are used or reported by the patients or participants. The method is described, including some of the more important rating conventions to be considered when using the method. This paper also describes the 15 cognitive errors assessed, and the different summary scores, including valence of the CEs, that can be derived from the method.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Miller, D. C.; Gu, X.; Haldenman, S.
The curing of cross-linkable encapsulation is a critical consideration for photovoltaic (PV) modules manufactured using a lamination process. Concerns related to ethylene-co-vinyl acetate (EVA) include the quality (e.g., expiration and uniformity) of the films or completion (duration) of the cross-linking of the EVA within a laminator. Because these issues are important to both EVA and module manufacturers, an international standard has recently been proposed by the Encapsulation Task-Group within the Working Group 2 (WG2) of the International Electrotechnical Commission (IEC) Technical Committee 82 (TC82) for the quantification of the degree of cure for EVA encapsulation. The present draft of themore » standard calls for the use of differential scanning calorimetry (DSC) as the rapid, enabling secondary (test) method. Both the residual enthalpy- and melt/freeze-DSC methods are identified. The DSC methods are calibrated against the gel content test, the primary (reference) method. Aspects of other established methods, including indentation and rotor cure metering, were considered by the group. Key details of the test procedure will be described.« less
Human versus automation in responding to failures: an expected-value analysis
NASA Technical Reports Server (NTRS)
Sheridan, T. B.; Parasuraman, R.
2000-01-01
A simple analytical criterion is provided for deciding whether a human or automation is best for a failure detection task. The method is based on expected-value decision theory in much the same way as is signal detection. It requires specification of the probabilities of misses (false negatives) and false alarms (false positives) for both human and automation being considered, as well as factors independent of the choice--namely, costs and benefits of incorrect and correct decisions as well as the prior probability of failure. The method can also serve as a basis for comparing different modes of automation. Some limiting cases of application are discussed, as are some decision criteria other than expected value. Actual or potential applications include the design and evaluation of any system in which either humans or automation are being considered.
A mass-energy preserving Galerkin FEM for the coupled nonlinear fractional Schrödinger equations
NASA Astrophysics Data System (ADS)
Zhang, Guoyu; Huang, Chengming; Li, Meng
2018-04-01
We consider the numerical simulation of the coupled nonlinear space fractional Schrödinger equations. Based on the Galerkin finite element method in space and the Crank-Nicolson (CN) difference method in time, a fully discrete scheme is constructed. Firstly, we focus on a rigorous analysis of conservation laws for the discrete system. The definitions of discrete mass and energy here correspond with the original ones in physics. Then, we prove that the fully discrete system is uniquely solvable. Moreover, we consider the unconditionally convergent properties (that is to say, we complete the error estimates without any mesh ratio restriction). We derive L2-norm error estimates for the nonlinear equations and L^{∞}-norm error estimates for the linear equations. Finally, some numerical experiments are included showing results in agreement with the theoretical predictions.
NASA Technical Reports Server (NTRS)
Riley, Donald R.
2016-01-01
Calculated numerical values for some aerodynamic terms and stability Derivatives for several different wings in unseparated inviscid incompressible flow were made using a discrete vortex method involving a limited number of horseshoe vortices. Both longitudinal and lateral-directional derivatives were calculated for steady conditions as well as for sinusoidal oscillatory motions. Variables included the number of vortices used and the rotation axis/moment center chordwise location. Frequencies considered were limited to the range of interest to vehicle dynamic stability (kb <.24 ). Comparisons of some calculated numerical results with experimental wind-tunnel measurements were in reasonable agreement in the low angle-of-attack range considering the differences existing between the mathematical representation and experimental wind-tunnel models tested. Of particular interest was the presence of induced drag for the oscillatory condition.
Emerging surface characterization techniques for carbon steel corrosion: a critical brief review.
Dwivedi, D; Lepkova, K; Becker, T
2017-03-01
Carbon steel is a preferred construction material in many industrial and domestic applications, including oil and gas pipelines, where corrosion mitigation using film-forming corrosion inhibitor formulations is a widely accepted method. This review identifies surface analytical techniques that are considered suitable for analysis of thin films at metallic substrates, but are yet to be applied to analysis of carbon steel surfaces in corrosive media or treated with corrosion inhibitors. The reviewed methods include time of flight-secondary ion mass spectrometry, X-ray absorption spectroscopy methods, particle-induced X-ray emission, Rutherford backscatter spectroscopy, Auger electron spectroscopy, electron probe microanalysis, near-edge X-ray absorption fine structure spectroscopy, X-ray photoemission electron microscopy, low-energy electron diffraction, small-angle neutron scattering and neutron reflectometry, and conversion electron Moessbauer spectrometry. Advantages and limitations of the analytical methods in thin-film surface investigations are discussed. Technical parameters of nominated analytical methods are provided to assist in the selection of suitable methods for analysis of metallic substrates deposited with surface films. The challenges associated with the applications of the emerging analytical methods in corrosion science are also addressed.
Emerging surface characterization techniques for carbon steel corrosion: a critical brief review
NASA Astrophysics Data System (ADS)
Dwivedi, D.; Lepkova, K.; Becker, T.
2017-03-01
Carbon steel is a preferred construction material in many industrial and domestic applications, including oil and gas pipelines, where corrosion mitigation using film-forming corrosion inhibitor formulations is a widely accepted method. This review identifies surface analytical techniques that are considered suitable for analysis of thin films at metallic substrates, but are yet to be applied to analysis of carbon steel surfaces in corrosive media or treated with corrosion inhibitors. The reviewed methods include time of flight-secondary ion mass spectrometry, X-ray absorption spectroscopy methods, particle-induced X-ray emission, Rutherford backscatter spectroscopy, Auger electron spectroscopy, electron probe microanalysis, near-edge X-ray absorption fine structure spectroscopy, X-ray photoemission electron microscopy, low-energy electron diffraction, small-angle neutron scattering and neutron reflectometry, and conversion electron Moessbauer spectrometry. Advantages and limitations of the analytical methods in thin-film surface investigations are discussed. Technical parameters of nominated analytical methods are provided to assist in the selection of suitable methods for analysis of metallic substrates deposited with surface films. The challenges associated with the applications of the emerging analytical methods in corrosion science are also addressed.
Emerging surface characterization techniques for carbon steel corrosion: a critical brief review
Dwivedi, D.; Becker, T.
2017-01-01
Carbon steel is a preferred construction material in many industrial and domestic applications, including oil and gas pipelines, where corrosion mitigation using film-forming corrosion inhibitor formulations is a widely accepted method. This review identifies surface analytical techniques that are considered suitable for analysis of thin films at metallic substrates, but are yet to be applied to analysis of carbon steel surfaces in corrosive media or treated with corrosion inhibitors. The reviewed methods include time of flight-secondary ion mass spectrometry, X-ray absorption spectroscopy methods, particle-induced X-ray emission, Rutherford backscatter spectroscopy, Auger electron spectroscopy, electron probe microanalysis, near-edge X-ray absorption fine structure spectroscopy, X-ray photoemission electron microscopy, low-energy electron diffraction, small-angle neutron scattering and neutron reflectometry, and conversion electron Moessbauer spectrometry. Advantages and limitations of the analytical methods in thin-film surface investigations are discussed. Technical parameters of nominated analytical methods are provided to assist in the selection of suitable methods for analysis of metallic substrates deposited with surface films. The challenges associated with the applications of the emerging analytical methods in corrosion science are also addressed. PMID:28413351
DOE Office of Scientific and Technical Information (OSTI.GOV)
Seth, Arpan; Klise, Katherine A.; Siirola, John D.
In the event of contamination in a water distribution network (WDN), source identification (SI) methods that analyze sensor data can be used to identify the source location(s). Knowledge of the source location and characteristics are important to inform contamination control and cleanup operations. Various SI strategies that have been developed by researchers differ in their underlying assumptions and solution techniques. The following manuscript presents a systematic procedure for testing and evaluating SI methods. The performance of these SI methods is affected by various factors including the size of WDN model, measurement error, modeling error, time and number of contaminant injections,more » and time and number of measurements. This paper includes test cases that vary these factors and evaluates three SI methods on the basis of accuracy and specificity. The tests are used to review and compare these different SI methods, highlighting their strengths in handling various identification scenarios. These SI methods and a testing framework that includes the test cases and analysis tools presented in this paper have been integrated into EPA’s Water Security Toolkit (WST), a suite of software tools to help researchers and others in the water industry evaluate and plan various response strategies in case of a contamination incident. Lastly, a set of recommendations are made for users to consider when working with different categories of SI methods.« less
Project FIRES. Volume 4: Prototype Protective Ensemble Qualification Test Report, Phase 1B
NASA Technical Reports Server (NTRS)
Abeles, F. J.
1980-01-01
The qualification testing of a prototype firefighter's protective ensemble is documented. Included are descriptions of the design requirements, the testing methods, and the test apparatus. The tests include measurements of individual subsystem characteristics in areas relating to both physical testing, such as heat, flame, impact penetration and human factors testing, such as dexterity, grip, and mobility. Also, measurements related to both physical and human factors testing of the complete ensemble, such as water protection, metabolic expenditures, and compatibility are considered.
Robust Programming Problems Based on the Mean-Variance Model Including Uncertainty Factors
NASA Astrophysics Data System (ADS)
Hasuike, Takashi; Ishii, Hiroaki
2009-01-01
This paper considers robust programming problems based on the mean-variance model including uncertainty sets and fuzzy factors. Since these problems are not well-defined problems due to fuzzy factors, it is hard to solve them directly. Therefore, introducing chance constraints, fuzzy goals and possibility measures, the proposed models are transformed into the deterministic equivalent problems. Furthermore, in order to solve these equivalent problems efficiently, the solution method is constructed introducing the mean-absolute deviation and doing the equivalent transformations.
Modelling of capillary-driven flow for closed paper-based microfluidic channels
NASA Astrophysics Data System (ADS)
Songok, Joel; Toivakka, Martti
2017-06-01
Paper-based microfluidics is an emerging field focused on creating inexpensive devices, with simple fabrication methods for applications in various fields including healthcare, environmental monitoring and veterinary medicine. Understanding the flow of liquid is important in achieving consistent operation of the devices. This paper proposes capillary models to predict flow in paper-based microfluidic channels, which include a flow accelerating hydrophobic top cover. The models, which consider both non-absorbing and absorbing substrates, are in good agreement with the experimental results.
Pölkki, Tarja; Kanste, Outi; Kääriäinen, Maria; Elo, Satu; Kyngäs, Helvi
2014-02-01
To analyse systematic review articles published in the top 10 nursing journals to determine the quality of the methods employed within them. Systematic review is defined as a scientific research method that synthesises high-quality scientific knowledge on a given topic. The number of such reviews in nursing science has increased dramatically during recent years, but their methodological quality has not previously been assessed. A review of the literature using a narrative approach. Ranked impact factor scores for nursing journals were obtained from the Journal Citation Report database of the Institute of Scientific Information (ISI Web of Knowledge). All issues from the years 2009 and 2010 of the top 10 ranked journals were included. CINAHL and MEDLINE databases were searched to locate studies using the search terms 'systematic review' and 'systematic literature review'. A total of 39 eligible studies were identified. Their methodological quality was evaluated through the specific criteria of quality assessment, description of synthesis and strengths and weaknesses reported in the included studies. Most of the eligible systematic reviews included several different designs or types of quantitative study. The majority included a quality assessment, and a total of 17 different criteria were identified. The method of synthesis was mentioned in about half of the reviews, the most common being narrative synthesis. The weaknesses of reviews were discussed, while strengths were rarely highlighted. The methodological quality of the systematic reviews examined varied considerably, although they were all published in nursing journals with a high-impact factor. Despite the fact that systematic reviews are considered the most robust source of research evidence, they vary in methodological quality. This point is important to consider in clinical practice when applying the results to patient care. © 2013 Blackwell Publishing Ltd.
Lucas, Patricia J; Baird, Janis; Arai, Lisa; Law, Catherine; Roberts, Helen M
2007-01-15
The inclusion of qualitative studies in systematic reviews poses methodological challenges. This paper presents worked examples of two methods of data synthesis (textual narrative and thematic), used in relation to one review, with the aim of enabling researchers to consider the strength of different approaches. A systematic review of lay perspectives of infant size and growth was conducted, locating 19 studies (including both qualitative and quantitative). The data extracted from these were synthesised using both a textual narrative and a thematic synthesis. The processes of both methods are presented, showing a stepwise progression to the final synthesis. Both methods led us to similar conclusions about lay views toward infant size and growth. Differences between methods lie in the way they dealt with study quality and heterogeneity. On the basis of the work reported here, we consider textual narrative and thematic synthesis have strengths and weaknesses in relation to different research questions. Thematic synthesis holds most potential for hypothesis generation, but may obscure heterogeneity and quality appraisal. Textual narrative synthesis is better able to describe the scope of existing research and account for the strength of evidence, but is less good at identifying commonality.
Classification of ROTSE Variable Stars using Machine Learning
NASA Astrophysics Data System (ADS)
Wozniak, P. R.; Akerlof, C.; Amrose, S.; Brumby, S.; Casperson, D.; Gisler, G.; Kehoe, R.; Lee, B.; Marshall, S.; McGowan, K. E.; McKay, T.; Perkins, S.; Priedhorsky, W.; Rykoff, E.; Smith, D. A.; Theiler, J.; Vestrand, W. T.; Wren, J.; ROTSE Collaboration
2001-12-01
We evaluate several Machine Learning algorithms as potential tools for automated classification of variable stars. Using the ROTSE sample of ~1800 variables from a pilot study of 5% of the whole sky, we compare the effectiveness of a supervised technique (Support Vector Machines, SVM) versus unsupervised methods (K-means and Autoclass). There are 8 types of variables in the sample: RR Lyr AB, RR Lyr C, Delta Scuti, Cepheids, detached eclipsing binaries, contact binaries, Miras and LPVs. Preliminary results suggest a very high ( ~95%) efficiency of SVM in isolating a few best defined classes against the rest of the sample, and good accuracy ( ~70-75%) for all classes considered simultaneously. This includes some degeneracies, irreducible with the information at hand. Supervised methods naturally outperform unsupervised methods, in terms of final error rate, but unsupervised methods offer many advantages for large sets of unlabeled data. Therefore, both types of methods should be considered as promising tools for mining vast variability surveys. We project that there are more than 30,000 periodic variables in the ROTSE-I data base covering the entire local sky between V=10 and 15.5 mag. This sample size is already stretching the time capabilities of human analysts.
Mobile Robot and Mobile Manipulator Research Towards ASTM Standards Development.
Bostelman, Roger; Hong, Tsai; Legowik, Steven
2016-01-01
Performance standards for industrial mobile robots and mobile manipulators (robot arms onboard mobile robots) have only recently begun development. Low cost and standardized measurement techniques are needed to characterize system performance, compare different systems, and to determine if recalibration is required. This paper discusses work at the National Institute of Standards and Technology (NIST) and within the ASTM Committee F45 on Driverless Automatic Guided Industrial Vehicles. This includes standards for both terminology, F45.91, and for navigation performance test methods, F45.02. The paper defines terms that are being considered. Additionally, the paper describes navigation test methods that are near ballot and docking test methods being designed for consideration within F45.02. This includes the use of low cost artifacts that can provide alternatives to using relatively expensive measurement systems.
Mobile Robot and Mobile Manipulator Research Towards ASTM Standards Development
Bostelman, Roger; Hong, Tsai; Legowik, Steven
2017-01-01
Performance standards for industrial mobile robots and mobile manipulators (robot arms onboard mobile robots) have only recently begun development. Low cost and standardized measurement techniques are needed to characterize system performance, compare different systems, and to determine if recalibration is required. This paper discusses work at the National Institute of Standards and Technology (NIST) and within the ASTM Committee F45 on Driverless Automatic Guided Industrial Vehicles. This includes standards for both terminology, F45.91, and for navigation performance test methods, F45.02. The paper defines terms that are being considered. Additionally, the paper describes navigation test methods that are near ballot and docking test methods being designed for consideration within F45.02. This includes the use of low cost artifacts that can provide alternatives to using relatively expensive measurement systems. PMID:28690359
Soulakova, Julia N; Bright, Brianna C
2013-01-01
A large-sample problem of illustrating noninferiority of an experimental treatment over a referent treatment for binary outcomes is considered. The methods of illustrating noninferiority involve constructing the lower two-sided confidence bound for the difference between binomial proportions corresponding to the experimental and referent treatments and comparing it with the negative value of the noninferiority margin. The three considered methods, Anbar, Falk-Koch, and Reduced Falk-Koch, handle the comparison in an asymmetric way, that is, only the referent proportion out of the two, experimental and referent, is directly involved in the expression for the variance of the difference between two sample proportions. Five continuity corrections (including zero) are considered with respect to each approach. The key properties of the corresponding methods are evaluated via simulations. First, the uncorrected two-sided confidence intervals can, potentially, have smaller coverage probability than the nominal level even for moderately large sample sizes, for example, 150 per group. Next, the 15 testing methods are discussed in terms of their Type I error rate and power. In the settings with a relatively small referent proportion (about 0.4 or smaller), the Anbar approach with Yates' continuity correction is recommended for balanced designs and the Falk-Koch method with Yates' correction is recommended for unbalanced designs. For relatively moderate (about 0.6) and large (about 0.8 or greater) referent proportion, the uncorrected Reduced Falk-Koch method is recommended, although in this case, all methods tend to be over-conservative. These results are expected to be used in the design stage of a noninferiority study when asymmetric comparisons are envisioned. Copyright © 2013 John Wiley & Sons, Ltd.
NASA Astrophysics Data System (ADS)
Ghaderi, F.; Pahlavani, P.
2015-12-01
A multimodal multi-criteria route planning (MMRP) system provides an optimal multimodal route from an origin point to a destination point considering two or more criteria in a way this route can be a combination of public and private transportation modes. In this paper, the simulate annealing (SA) and the fuzzy analytical hierarchy process (fuzzy AHP) were combined in order to find this route. In this regard, firstly, the effective criteria that are significant for users in their trip were determined. Then the weight of each criterion was calculated using the fuzzy AHP weighting method. The most important characteristic of this weighting method is the use of fuzzy numbers that aids the users to consider their uncertainty in pairwise comparison of criteria. After determining the criteria weights, the proposed SA algorithm were used for determining an optimal route from an origin to a destination. One of the most important problems in a meta-heuristic algorithm is trapping in local minima. In this study, five transportation modes, including subway, bus rapid transit (BRT), taxi, walking, and bus were considered for moving between nodes. Also, the fare, the time, the user's bother, and the length of the path were considered as effective criteria for solving the problem. The proposed model was implemented in an area in centre of Tehran in a GUI MATLAB programming language. The results showed a high efficiency and speed of the proposed algorithm that support our analyses.
Helicopter cabin noise: Methods of source and path identification and characterization
NASA Technical Reports Server (NTRS)
Murray, B. S.; Wilby, J. F.
1978-01-01
Internal noise sources in a helicopter are considered. These include propulsion machinery, comprising engine and transmission, and turbulent boundary layer effects. It is shown that by using relatively simple concepts together with careful experimental work it is possible to generate reliable data on which to base the design of high performance noise control treatments.
ERIC Educational Resources Information Center
Zholdasbekova, S.; Karataev, G.; Yskak, A.; Zholdasbekov, A.; Nurzhanbaeva, J.
2015-01-01
This article describes the major components of required technological skills (TS) for future designers taught during the academic process of a college. It considers the choices in terms of the various logical operations required by the fashion industry including fabric processing, assembly charts, performing work operations, etc. The article…
Teacher Candidates' Perception Level of Environmental Pollutant and Their Risk Factors
ERIC Educational Resources Information Center
Oztas, Fulya; Kalipci, Erkan
2009-01-01
It is generally accepted that the environmental education deals with a wide range of environmental experiences, methods and processes. Teaching the subject of the environment should not be considered as an easy task. It should not only cover pure ecology education; but also include the citizenship responsibilities and the problems that are sourced…
A Uniform Identity: Schoolgirl Snapshots and the Spoken Visual
ERIC Educational Resources Information Center
Spencer, Stephanie
2007-01-01
This article discusses the possibility for expanding our understanding of the visual to include the "spoken visual" within oral history analysis. It suggests that adding a further reading, that of the visualized body, to the voice-centred relational method we can consider the meaning of the uniformed body for the individual. It uses as a…
ERIC Educational Resources Information Center
Massachusetts Inst. of Tech., Cambridge. Sensory Aids Evaluation and Development Center.
Proceedings from a conference on braille production and services are summarized. Only equipment which is ready for use is considered. Specific methods of producing braille discussed include the following: use of plates from a stereograph, computer line printer, braille embossers or braille writers, continuous strip embossed tape from a teletype…
ERIC Educational Resources Information Center
Cohen, Erik H.
2016-01-01
This article considers the role of teacher agency and curricular flexibility as pedagogic features of Shoah education in Israeli state schools. The analysis is based on a recent national study which included a quantitative survey (questionnaires), qualitative methods (focus groups, interviews, observations) and a socio-historical review. As…
The Effects of Including a Callous-Unemotional Specifier for the Diagnosis of Conduct Disorder
ERIC Educational Resources Information Center
Kahn, Rachel E.; Frick, Paul J.; Youngstrom, Eric; Findling, Robert L.; Youngstrom, Jennifer Kogos
2012-01-01
Background: "With Significant Callous-Unemotional Traits" has been proposed as a specifier for conduct disorder (CD) in the upcoming revision of the Diagnostic and Statistical Manual of Mental Disorders (DSM-V). The impact of this specifier on children diagnosed with CD should be considered. Methods: A multi-site cross-sectional design with…
ERIC Educational Resources Information Center
D'Haem, Jeanne; Griswold, Peter
2017-01-01
This mixed-methods study examined teacher preparation for developing family partnerships. The attitudes and practices of teacher educators and the attitudes and experiences of student teachers were explored in focus groups, documents, and a survey instrument. Results indicated that although partnerships were considered important by faculty and…
Costs of Storing and Transporting Hydrogen
DOE Office of Scientific and Technical Information (OSTI.GOV)
Amos, W. A.
An analysis was performed to estimate the costs associated with storing and transporting hydrogen. These costs can be added to a hydrogen production cost to determine the total delivered cost of hydrogen. Storage methods analyzed included compressed gas, liquid hydrogen, metal hydride, and underground storage. Major capital and operating costs were considered over a range of production rates and storage times.
Antianaerobic Antimicrobials: Spectrum and Susceptibility Testing
Wexler, Hannah M.; Goldstein, Ellie J. C.
2013-01-01
SUMMARY Susceptibility testing of anaerobic bacteria recovered from selected cases can influence the choice of antimicrobial therapy. The Clinical and Laboratory Standards Institute (CLSI) has standardized many laboratory procedures, including anaerobic susceptibility testing (AST), and has published documents for AST. The standardization of testing methods by the CLSI allows comparisons of resistance trends among various laboratories. Susceptibility testing should be performed on organisms recovered from sterile body sites, those that are isolated in pure culture, or those that are clinically important and have variable or unique susceptibility patterns. Organisms that should be considered for individual isolate testing include highly virulent pathogens for which susceptibility cannot be predicted, such as Bacteroides, Prevotella, Fusobacterium, and Clostridium spp.; Bilophila wadsworthia; and Sutterella wadsworthensis. This review describes the current methods for AST in research and reference laboratories. These methods include the use of agar dilution, broth microdilution, Etest, and the spiral gradient endpoint system. The antimicrobials potentially effective against anaerobic bacteria include beta-lactams, combinations of beta-lactams and beta-lactamase inhibitors, metronidazole, chloramphenicol, clindamycin, macrolides, tetracyclines, and fluoroquinolones. The spectrum of efficacy, antimicrobial resistance mechanisms, and resistance patterns against these agents are described. PMID:23824372
James, Andrew J. A.; Konik, Robert M.; Lecheminant, Philippe; ...
2018-02-26
We review two important non-perturbative approaches for extracting the physics of low-dimensional strongly correlated quantum systems. Firstly, we start by providing a comprehensive review of non-Abelian bosonization. This includes an introduction to the basic elements of conformal field theory as applied to systems with a current algebra, and we orient the reader by presenting a number of applications of non-Abelian bosonization to models with large symmetries. We then tie this technique into recent advances in the ability of cold atomic systems to realize complex symme-tries. Secondly, we discuss truncated spectrum methods for the numerical study of systems in one andmore » two dimensions. For one-dimensional systems we provide the reader with considerable insight into the methodology by reviewing canonical applications of the technique to the Ising model (and its variants) and the sine-Gordon model. Following this we review recent work on the development of renormalization groups, both numerical and analytical, that alleviate the effects of truncating the spectrum. Using these technologies, we consider a number of applications to one-dimensional systems: properties of carbon nanotubes, quenches in the Lieb-Liniger model, 1+1D quantum chro-modynamics, as well as Landau-Ginzburg theories. In the final part we move our attention to consider truncated spectrum methods applied to two-dimensional systems. This involves combining truncated spectrum methods with matrix product state algorithms. Lastly, we describe applications of this method to two-dimensional systems of free fermions and the quantum Ising model, including their non-equilibrium dynamics.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
James, Andrew J. A.; Konik, Robert M.; Lecheminant, Philippe
We review two important non-perturbative approaches for extracting the physics of low-dimensional strongly correlated quantum systems. Firstly, we start by providing a comprehensive review of non-Abelian bosonization. This includes an introduction to the basic elements of conformal field theory as applied to systems with a current algebra, and we orient the reader by presenting a number of applications of non-Abelian bosonization to models with large symmetries. We then tie this technique into recent advances in the ability of cold atomic systems to realize complex symme-tries. Secondly, we discuss truncated spectrum methods for the numerical study of systems in one andmore » two dimensions. For one-dimensional systems we provide the reader with considerable insight into the methodology by reviewing canonical applications of the technique to the Ising model (and its variants) and the sine-Gordon model. Following this we review recent work on the development of renormalization groups, both numerical and analytical, that alleviate the effects of truncating the spectrum. Using these technologies, we consider a number of applications to one-dimensional systems: properties of carbon nanotubes, quenches in the Lieb-Liniger model, 1+1D quantum chro-modynamics, as well as Landau-Ginzburg theories. In the final part we move our attention to consider truncated spectrum methods applied to two-dimensional systems. This involves combining truncated spectrum methods with matrix product state algorithms. Lastly, we describe applications of this method to two-dimensional systems of free fermions and the quantum Ising model, including their non-equilibrium dynamics.« less
NASA Astrophysics Data System (ADS)
James, Andrew J. A.; Konik, Robert M.; Lecheminant, Philippe; Robinson, Neil J.; Tsvelik, Alexei M.
2018-04-01
We review two important non-perturbative approaches for extracting the physics of low-dimensional strongly correlated quantum systems. Firstly, we start by providing a comprehensive review of non-Abelian bosonization. This includes an introduction to the basic elements of conformal field theory as applied to systems with a current algebra, and we orient the reader by presenting a number of applications of non-Abelian bosonization to models with large symmetries. We then tie this technique into recent advances in the ability of cold atomic systems to realize complex symmetries. Secondly, we discuss truncated spectrum methods for the numerical study of systems in one and two dimensions. For one-dimensional systems we provide the reader with considerable insight into the methodology by reviewing canonical applications of the technique to the Ising model (and its variants) and the sine-Gordon model. Following this we review recent work on the development of renormalization groups, both numerical and analytical, that alleviate the effects of truncating the spectrum. Using these technologies, we consider a number of applications to one-dimensional systems: properties of carbon nanotubes, quenches in the Lieb–Liniger model, 1 + 1D quantum chromodynamics, as well as Landau–Ginzburg theories. In the final part we move our attention to consider truncated spectrum methods applied to two-dimensional systems. This involves combining truncated spectrum methods with matrix product state algorithms. We describe applications of this method to two-dimensional systems of free fermions and the quantum Ising model, including their non-equilibrium dynamics.
James, Andrew J A; Konik, Robert M; Lecheminant, Philippe; Robinson, Neil J; Tsvelik, Alexei M
2018-02-26
We review two important non-perturbative approaches for extracting the physics of low-dimensional strongly correlated quantum systems. Firstly, we start by providing a comprehensive review of non-Abelian bosonization. This includes an introduction to the basic elements of conformal field theory as applied to systems with a current algebra, and we orient the reader by presenting a number of applications of non-Abelian bosonization to models with large symmetries. We then tie this technique into recent advances in the ability of cold atomic systems to realize complex symmetries. Secondly, we discuss truncated spectrum methods for the numerical study of systems in one and two dimensions. For one-dimensional systems we provide the reader with considerable insight into the methodology by reviewing canonical applications of the technique to the Ising model (and its variants) and the sine-Gordon model. Following this we review recent work on the development of renormalization groups, both numerical and analytical, that alleviate the effects of truncating the spectrum. Using these technologies, we consider a number of applications to one-dimensional systems: properties of carbon nanotubes, quenches in the Lieb-Liniger model, 1 + 1D quantum chromodynamics, as well as Landau-Ginzburg theories. In the final part we move our attention to consider truncated spectrum methods applied to two-dimensional systems. This involves combining truncated spectrum methods with matrix product state algorithms. We describe applications of this method to two-dimensional systems of free fermions and the quantum Ising model, including their non-equilibrium dynamics.
NASA Astrophysics Data System (ADS)
Pereira, A. A.; Gironas, J. A.; Passalacqua, P.; Mejia, A.; Niemann, J. D.
2017-12-01
Previous work has shown that lithological, tectonic and climatic processes have a major influence in shaping the geomorphology of river networks. Accordingly, quantitative classification methods have been developed to identify and characterize network types (dendritic, parallel, pinnate, rectangular and trellis) based solely on the self-affinity of their planform properties, computed from available Digital Elevation Model (DEM) data. In contrast, this research aim is to include both horizontal and vertical properties to evaluate a quantitative classification method for river networks. We include vertical properties to consider the unique surficial conditions (e.g., large and steep height drops, volcanic activity, and complexity of stream networks) of the Andes Mountains. Furthermore, the goal of the research is also to explain the implications and possible relations between the hydro-geomorphological properties and climatic conditions. The classification method is applied to 42 basins in the southern Andes in Chile, ranging in size from 208 Km2 to 8,000 Km2. The planform metrics include the incremental drainage area, stream course irregularity and junction angles, while the vertical metrics include the hypsometric curve and the slope-area relationship. We introduce new network structures (Brush, Funnel and Low Sinuosity Rectangular), possibly unique to the Andes, that can be quantitatively differentiated from previous networks identified in other geographic regions. Then, this research evaluates the effect that excluding different Strahler order streams has on the horizontal properties and therefore in the classification. We found that climatic conditions are not only linked to horizontal parameters, but also to vertical ones, finding significant correlation between climatic variables (average near-surface temperature and rainfall) and vertical measures (parameters associated with the hypsometric curve and slope-area relation). The proposed classification shows differences among basins previously classified as the same type, which are not noticeable in their horizontal properties and helps reduce misclassifications within the old clusters. Additional hydro-geomorphological metrics are to be considered in the classification method to improve the effectiveness of it.
Chasey, K L; McKee, R H
1993-01-01
Lubricant base oils are petroleum products that are predominantly derived from the vacuum distillation of crude oil. Various types of refinement can be employed during the manufacturing process, and evidence suggests that certain of the associated process streams produce skin cancer. Polycyclic aromatic compounds (PACs), some of which are considered as the causative agents, are removed, concentrated or chemically converted during the refinement process. In order to understand the effects of various types of refinement processes on carcinogenic potential, 94 oils were evaluated in the mouse epidermal cancer bioassay. This Exxon database is unique, because of the wide range of crude oils and processing histories represented. Seven processing history classifications are described, and conclusions concerning the impacts of each refinement process on dermal carcinogenicity are discussed. This research also included an evaluation of selected biological and chemical test methods for predicting carcinogenic potential. These included a modified version of the Ames test for mutagenicity, as well as analytical characterizations of the polycyclic aromatic structures in the oils. For classification purposes, a sample was considered to be carcinogenic if it resulted in the production of two or more tumor-bearing animals (in test groups of either 40 or 50 animals). The modified Ames test was considered to be positive if the mutagenicity index was > or = 2.0, and PAC analyses were similarly designated as positive or negative according to proposed guidelines. All of the alternative test methods showed similar agreement with dermal carcinogenicity bioassay data; concordance values were > or = 80%. However, each test was incorrect in ca. 10%-20% of the cases evaluated.(ABSTRACT TRUNCATED AT 250 WORDS)
Effect of nonlinearity in hybrid kinetic Monte Carlo-continuum models.
Balter, Ariel; Lin, Guang; Tartakovsky, Alexandre M
2012-01-01
Recently there has been interest in developing efficient ways to model heterogeneous surface reactions with hybrid computational models that couple a kinetic Monte Carlo (KMC) model for a surface to a finite-difference model for bulk diffusion in a continuous domain. We consider two representative problems that validate a hybrid method and show that this method captures the combined effects of nonlinearity and stochasticity. We first validate a simple deposition-dissolution model with a linear rate showing that the KMC-continuum hybrid agrees with both a fully deterministic model and its analytical solution. We then study a deposition-dissolution model including competitive adsorption, which leads to a nonlinear rate, and show that in this case the KMC-continuum hybrid and fully deterministic simulations do not agree. However, we are able to identify the difference as a natural result of the stochasticity coming from the KMC surface process. Because KMC captures inherent fluctuations, we consider it to be more realistic than a purely deterministic model. Therefore, we consider the KMC-continuum hybrid to be more representative of a real system.
Effect of Nonlinearity in Hybrid Kinetic Monte Carlo-Continuum Models
DOE Office of Scientific and Technical Information (OSTI.GOV)
Balter, Ariel I.; Lin, Guang; Tartakovsky, Alexandre M.
2012-04-23
Recently there has been interest in developing efficient ways to model heterogeneous surface reactions with hybrid computational models that couple a KMC model for a surface to a finite difference model for bulk diffusion in a continuous domain. We consider two representative problems that validate a hybrid method and also show that this method captures the combined effects of nonlinearity and stochasticity. We first validate a simple deposition/dissolution model with a linear rate showing that the KMC-continuum hybrid agrees with both a fully deterministic model and its analytical solution. We then study a deposition/dissolution model including competitive adsorption, which leadsmore » to a nonlinear rate, and show that, in this case, the KMC-continuum hybrid and fully deterministic simulations do not agree. However, we are able to identify the difference as a natural result of the stochasticity coming from the KMC surface process. Because KMC captures inherent fluctuations, we consider it to be more realistic than a purely deterministic model. Therefore, we consider the KMC-continuum hybrid to be more representative of a real system.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Jeffery, D.J.
This report considers a wide range of possible methods for improving the route guidance information available to drivers, and for saving up to L600M per year of national resource costs which are effectively wasted by drivers who fail to select optimum routes for unfamiliar journeys. The methods considered include improvements to maps and roadsigns, both of which could prove highly cost effective although they would leave some 80 per cent of the wastage unrecovered. A substantial proportion of this remaining wastage might be avoided with a Viewdata route planning and guidance scheme which could be achieved at relatively low costmore » to public funds, or with automatic systems. Of a range of automatic electronic guidance systems considered the most cost effective solution would be provided by a system which used buried loops to provide a two-way communication link between a roadside and a vehicle unit. Such a scheme would necessarily involve a considerable investment from public funds, but would offer the greatest potential for further development and, in particular, could provide the basis of a comprehensive traffic control tool.« less
Hey, Jody; Nielsen, Rasmus
2007-01-01
In 1988, Felsenstein described a framework for assessing the likelihood of a genetic data set in which all of the possible genealogical histories of the data are considered, each in proportion to their probability. Although not analytically solvable, several approaches, including Markov chain Monte Carlo methods, have been developed to find approximate solutions. Here, we describe an approach in which Markov chain Monte Carlo simulations are used to integrate over the space of genealogies, whereas other parameters are integrated out analytically. The result is an approximation to the full joint posterior density of the model parameters. For many purposes, this function can be treated as a likelihood, thereby permitting likelihood-based analyses, including likelihood ratio tests of nested models. Several examples, including an application to the divergence of chimpanzee subspecies, are provided. PMID:17301231
NASA Technical Reports Server (NTRS)
Saleeb, A. F.; Chang, T. Y. P.; Wilt, T.; Iskovitz, I.
1989-01-01
The research work performed during the past year on finite element implementation and computational techniques pertaining to high temperature composites is outlined. In the present research, two main issues are addressed: efficient geometric modeling of composite structures and expedient numerical integration techniques dealing with constitutive rate equations. In the first issue, mixed finite elements for modeling laminated plates and shells were examined in terms of numerical accuracy, locking property and computational efficiency. Element applications include (currently available) linearly elastic analysis and future extension to material nonlinearity for damage predictions and large deformations. On the material level, various integration methods to integrate nonlinear constitutive rate equations for finite element implementation were studied. These include explicit, implicit and automatic subincrementing schemes. In all cases, examples are included to illustrate the numerical characteristics of various methods that were considered.
Host-microbe interactions in distal airways: relevance to chronic airway diseases.
Martin, Clémence; Burgel, Pierre-Régis; Lepage, Patricia; Andréjak, Claire; de Blic, Jacques; Bourdin, Arnaud; Brouard, Jacques; Chanez, Pascal; Dalphin, Jean-Charles; Deslée, Gaetan; Deschildre, Antoine; Gosset, Philippe; Touqui, Lhousseine; Dusser, Daniel
2015-03-01
This article is the summary of a workshop, which took place in November 2013, on the roles of microorganisms in chronic respiratory diseases. Until recently, it was assumed that lower airways were sterile in healthy individuals. However, it has long been acknowledged that microorganisms could be identified in distal airway secretions from patients with various respiratory diseases, including cystic fibrosis (CF) and non-CF bronchiectasis, chronic obstructive pulmonary disease, asthma and other chronic airway diseases (e.g. post-transplantation bronchiolitis obliterans). These microorganisms were sometimes considered as infectious agents that triggered host immune responses and contributed to disease onset and/or progression; alternatively, microorganisms were often considered as colonisers, which were considered unlikely to play roles in disease pathophysiology. These concepts were developed at a time when the identification of microorganisms relied on culture-based methods. Importantly, the majority of microorganisms cannot be cultured using conventional methods, and the use of novel culture-independent methods that rely on the identification of microorganism genomes has revealed that healthy distal airways display a complex flora called the airway microbiota. The present article reviews some aspects of current literature on host-microbe (mostly bacteria and viruses) interactions in healthy and diseased airways, with a special focus on distal airways. Copyright ©ERS 2015.
Dexter, Franklin; Ledolter, Johannes; Hindman, Bradley J
2016-01-01
In this Statistical Grand Rounds, we review methods for the analysis of the diversity of procedures among hospitals, the activities among anesthesia providers, etc. We apply multiple methods and consider their relative reliability and usefulness for perioperative applications, including calculations of SEs. We also review methods for comparing the similarity of procedures among hospitals, activities among anesthesia providers, etc. We again apply multiple methods and consider their relative reliability and usefulness for perioperative applications. The applications include strategic analyses (e.g., hospital marketing) and human resource analytics (e.g., comparisons among providers). Measures of diversity of procedures and activities (e.g., Herfindahl and Gini-Simpson index) are used for quantification of each facility (hospital) or anesthesia provider, one at a time. Diversity can be thought of as a summary measure. Thus, if the diversity of procedures for 48 hospitals is studied, the diversity (and its SE) is being calculated for each hospital. Likewise, the effective numbers of common procedures at each hospital can be calculated (e.g., by using the exponential of the Shannon index). Measures of similarity are pairwise assessments. Thus, if quantifying the similarity of procedures among cases with a break or handoff versus cases without a break or handoff, a similarity index represents a correlation coefficient. There are several different measures of similarity, and we compare their features and applicability for perioperative data. We rely extensively on sensitivity analyses to interpret observed values of the similarity index.
An Update on Drugs Used for Lumbosacral Epidural Anesthesia and Analgesia in Dogs
Steagall, Paulo V. M.; Simon, Bradley T.; Teixeira Neto, Francisco J.; Luna, Stelio P. L.
2017-01-01
This review aims to report an update on drugs administered into the epidural space for anesthesia and analgesia in dogs, describing their potential advantages and disadvantages in the clinical setting. Databases searched include Pubmed, Google scholar, and CAB abstracts. Benefits of administering local anesthetics, opioids, and alpha2 agonists into the epidural space include the use of lower doses of general anesthetics (anesthetic “sparing” effect), perioperative analgesia, and reduced side effects associated with systemic administration of drugs. However, the potential for cardiorespiratory compromise, neurotoxicity, and other adverse effects should be considered when using the epidural route of administration. When these variables are considered, the epidural technique is useful as a complementary method of anesthesia for preventive and postoperative analgesia and/or as part of a balanced anesthesia technique. PMID:28553642
The integrated manual and automatic control of complex flight systems
NASA Technical Reports Server (NTRS)
Schmidt, David K.
1991-01-01
Research dealt with the general area of optimal flight control synthesis for manned flight vehicles. The work was generic; no specific vehicle was the focus of study. However, the class of vehicles generally considered were those for which high authority, multivariable control systems might be considered, for the purpose of stabilization and the achievement of optimal handling characteristics. Within this scope, the topics of study included several optimal control synthesis techniques, control-theoretic modeling of the human operator in flight control tasks, and the development of possible handling qualities metrics and/or measures of merit. Basic contributions were made in all these topics, including human operator (pilot) models for multi-loop tasks, optimal output feedback flight control synthesis techniques; experimental validations of the methods developed, and fundamental modeling studies of the air-to-air tracking and flared landing tasks.
Guidance for using mixed methods design in nursing practice research.
Chiang-Hanisko, Lenny; Newman, David; Dyess, Susan; Piyakong, Duangporn; Liehr, Patricia
2016-08-01
The mixed methods approach purposefully combines both quantitative and qualitative techniques, enabling a multi-faceted understanding of nursing phenomena. The purpose of this article is to introduce three mixed methods designs (parallel; sequential; conversion) and highlight interpretive processes that occur with the synthesis of qualitative and quantitative findings. Real world examples of research studies conducted by the authors will demonstrate the processes leading to the merger of data. The examples include: research questions; data collection procedures and analysis with a focus on synthesizing findings. Based on experience with mixed methods studied, the authors introduce two synthesis patterns (complementary; contrasting), considering application for practice and implications for research. Copyright © 2015 Elsevier Inc. All rights reserved.
Dealing with noise and physiological artifacts in human EEG recordings: empirical mode methods
NASA Astrophysics Data System (ADS)
Runnova, Anastasiya E.; Grubov, Vadim V.; Khramova, Marina V.; Hramov, Alexander E.
2017-04-01
In the paper we propose the new method for removing noise and physiological artifacts in human EEG recordings based on empirical mode decomposition (Hilbert-Huang transform). As physiological artifacts we consider specific oscillatory patterns that cause problems during EEG analysis and can be detected with additional signals recorded simultaneously with EEG (ECG, EMG, EOG, etc.) We introduce the algorithm of the proposed method with steps including empirical mode decomposition of EEG signal, choosing of empirical modes with artifacts, removing these empirical modes and reconstructing of initial EEG signal. We show the efficiency of the method on the example of filtration of human EEG signal from eye-moving artifacts.
Ground Vibration Test Planning and Pre-Test Analysis for the X-33 Vehicle
NASA Technical Reports Server (NTRS)
Bedrossian, Herand; Tinker, Michael L.; Hidalgo, Homero
2000-01-01
This paper describes the results of the modal test planning and the pre-test analysis for the X-33 vehicle. The pre-test analysis included the selection of the target modes, selection of the sensor and shaker locations and the development of an accurate Test Analysis Model (TAM). For target mode selection, four techniques were considered, one based on the Modal Cost technique, one based on Balanced Singular Value technique, a technique known as the Root Sum Squared (RSS) method, and a Modal Kinetic Energy (MKE) approach. For selecting sensor locations, four techniques were also considered; one based on the Weighted Average Kinetic Energy (WAKE), one based on Guyan Reduction (GR), one emphasizing engineering judgment, and one based on an optimum sensor selection technique using Genetic Algorithm (GA) search technique combined with a criteria based on Hankel Singular Values (HSV's). For selecting shaker locations, four techniques were also considered; one based on the Weighted Average Driving Point Residue (WADPR), one based on engineering judgment and accessibility considerations, a frequency response method, and an optimum shaker location selection based on a GA search technique combined with a criteria based on HSV's. To evaluate the effectiveness of the proposed sensor and shaker locations for exciting the target modes, extensive numerical simulations were performed. Multivariate Mode Indicator Function (MMIF) was used to evaluate the effectiveness of each sensor & shaker set with respect to modal parameter identification. Several TAM reduction techniques were considered including, Guyan, IRS, Modal, and Hybrid. Based on a pre-test cross-orthogonality checks using various reduction techniques, a Hybrid TAM reduction technique was selected and was used for all three vehicle fuel level configurations.
Monte Carlo kinetics simulations of ice-mantle formation on interstellar grains
NASA Astrophysics Data System (ADS)
Garrod, Robin
2015-08-01
The majority of interstellar dust-grain chemical kinetics models use rate equations, or alternative population-based simulation methods, to trace the time-dependent formation of grain-surface molecules and ice mantles. Such methods are efficient, but are incapable of considering explicitly the morphologies of the dust grains, the structure of the ices formed thereon, or the influence of local surface composition on the chemistry.A new Monte Carlo chemical kinetics model, MIMICK, is presented here, whose prototype results were published recently (Garrod 2013, ApJ, 778, 158). The model calculates the strengths and positions of the potential mimima on the surface, on the fly, according to the individual pair-wise (van der Waals) bonds between surface species, allowing the structure of the ice to build up naturally as surface diffusion and chemistry occur. The prototype model considered contributions to a surface particle's potential only from contiguous (or "bonded") neighbors; the full model considers contributions from surface constituents from short to long range. Simulations are conducted on a fully 3-D user-generated dust-grain with amorphous surface characteristics. The chemical network has also been extended from the simple water system previously published, and now includes 33 chemical species and 55 reactions. This allows the major interstellar ice components to be simulated, such as water, methane, ammonia and methanol, as well as a small selection of more complex molecules, including methyl formate (HCOOCH3).The new model results indicate that the porosity of interstellar ices are dependent on multiple variables, including gas density, the dust temperature, and the relative accretion rates of key gas-phase species. The results presented also have implications for the formation of complex organic molecules on dust-grain surfaces at very low temperatures.
A Method of DTM Construction Based on Quadrangular Irregular Networks and Related Error Analysis
Kang, Mengjun
2015-01-01
A new method of DTM construction based on quadrangular irregular networks (QINs) that considers all the original data points and has a topological matrix is presented. A numerical test and a real-world example are used to comparatively analyse the accuracy of QINs against classical interpolation methods and other DTM representation methods, including SPLINE, KRIGING and triangulated irregular networks (TINs). The numerical test finds that the QIN method is the second-most accurate of the four methods. In the real-world example, DTMs are constructed using QINs and the three classical interpolation methods. The results indicate that the QIN method is the most accurate method tested. The difference in accuracy rank seems to be caused by the locations of the data points sampled. Although the QIN method has drawbacks, it is an alternative method for DTM construction. PMID:25996691
Accurate D-bar Reconstructions of Conductivity Images Based on a Method of Moment with Sinc Basis.
Abbasi, Mahdi
2014-01-01
Planar D-bar integral equation is one of the inverse scattering solution methods for complex problems including inverse conductivity considered in applications such as Electrical impedance tomography (EIT). Recently two different methodologies are considered for the numerical solution of D-bar integrals equation, namely product integrals and multigrid. The first one involves high computational burden and the other one suffers from low convergence rate (CR). In this paper, a novel high speed moment method based using the sinc basis is introduced to solve the two-dimensional D-bar integral equation. In this method, all functions within D-bar integral equation are first expanded using the sinc basis functions. Then, the orthogonal properties of their products dissolve the integral operator of the D-bar equation and results a discrete convolution equation. That is, the new moment method leads to the equation solution without direct computation of the D-bar integral. The resulted discrete convolution equation maybe adapted to a suitable structure to be solved using fast Fourier transform. This allows us to reduce the order of computational complexity to as low as O (N (2)log N). Simulation results on solving D-bar equations arising in EIT problem show that the proposed method is accurate with an ultra-linear CR.
Salari, Fatemeh; Shahjahani, Mohammad; Shahrabi, Saeid; Saki, Najmaldin
2014-11-01
After advances in experimental and clinical testing, minimal residual disease (MRD) assay results are considered a determining factor in treatment of acute lymphoblastic leukemia patients. According to MRD assay results, bone marrow (BM) leukemic burden and the rate of its decline after treatment can be directly evaluated. Detailed knowledge of the leukemic burden in BM can minimize toxicity and treatment complications in patients by tailoring the therapeutic dose based on patients' conditions. In addition, reduction of MRD before allo-HSCT is an important prerequisite for reception of transplant by the patient. In direct examination of MRD by morphological methods (even by a professional hematologist), leukemic cells can be under- or over-estimated due to similarity with hematopoietic precursor cells. As a result, considering the importance of MRD, it is necessary to use other methods including flow cytometry, polymerase chain reaction (PCR) amplification and RQ-PCR to detect MRD. Each of these methods has its own advantages and disadvantages in terms of accuracy and sensitivity. In this review article, different MRD assay methods and their sensitivity, correlation of MRD assay results with clinical symptoms of the patient as well as pitfalls in results of these methods are evaluated. In the final section, recent advances in MRD have been addressed.
Using GIS and Ahp for Planning Primer Transportation of Forest Products
NASA Astrophysics Data System (ADS)
Akay, A. E.; Yılmaz, B.
2017-11-01
Primer transportation is one of the most costly and time consuming forestry activities in extraction of timber from forest lands. Transportation methods are essentially determined based on terrain characteristics, especially ground slope. Besides, unsuitable machine selection and unplanned operations may cause ecological damages such as soil disturbance. Soil damage can lead to long term impacts on forest ecosystem. Thus, the optimum transportation methods should be determined by considering not only economic factors but also topographical factors and soil conditions. In recent decades, some of the advanced features of Geographical Information System (GIS) assist decision makers to solve such complex transportation problems with various constraints. In this study, it was aimed to plan forest transportation operation by using GIS integrated Analytical Hierarchy Process (AHP) method, considering ground slope, soil type, and available transportation equipment in the region. This method was implemented within the border of İnegöl Forest Enterprise Chief in the city of Bursa in Turkey. Alternative transportation method included cable system, chute system, skidder, and farm tractor. GIS-based method integrated with AHP found that skidder was the optimal transportation method for about 60% of the study area, while farm tractor was the second most suitable method with 25% ground cover. The results indicated that GIS-based decision support systems can be effectively used as rational, quick, and economic tool for forest transportation planning.
NASA Astrophysics Data System (ADS)
Gorash, Yevgen; Comlekci, Tugrul; MacKenzie, Donald
2017-05-01
This study investigates the effects of fatigue material data and finite element types on accuracy of residual life assessments under high cycle fatigue. The bending of cross-beam connections is simulated in ANSYS Workbench for different combinations of structural member shapes made of a typical structural steel. The stress analysis of weldments with specific dimensions and loading applied is implemented using solid and shell elements. The stress results are transferred to the fatigue code nCode DesignLife for the residual life prediction. Considering the effects of mean stress using FKM approach, bending and thickness according to BS 7608:2014, fatigue life is predicted using the Volvo method and stress integration rules from ASME Boiler & Pressure Vessel Code. Three different pairs of S-N curves are considered in this work including generic seam weld curves and curves for the equivalent Japanese steel JIS G3106-SM490B. The S-N curve parameters for the steel are identified using the experimental data available from NIMS fatigue data sheets employing least square method and considering thickness and mean stress corrections. The numerical predictions are compared to the available experimental results indicating the most preferable fatigue data input, range of applicability and FE-model formulation to achieve the best accuracy.
Recruiting Adolescent Research Participants: In-Person Compared to Social Media Approaches.
Moreno, Megan A; Waite, Alan; Pumper, Megan; Colburn, Trina; Holm, Matt; Mendoza, Jason
2017-01-01
Recruiting adolescent participants for research is challenging. The purpose of this study was to compare traditional in-person recruitment methods to social media recruitment. We recruited adolescents aged 14-18 years for a pilot physical activity intervention study, including a wearable physical activity tracking device and a Facebook group. Participants were recruited (a) in person from a local high school and an adolescent medicine clinic and (b) through social media, including Facebook targeted ads, sponsored tweets on Twitter, and a blog post. Data collected included total exposure (i.e., reach), engagement (i.e., interaction), and effectiveness. Effectiveness included screening and enrollment for each recruitment method, as well as time and resources spent on each recruitment method. In-person recruitment reached a total of 297 potential participants of which 37 enrolled in the study. Social media recruitment reached a total of 34,272 potential participants of which 8 enrolled in the study. Social media recruitment methods utilized an average of 1.6 hours of staff time and cost an average of $40.99 per participant enrolled, while in-person recruitment methods utilized an average of 0.75 hours of staff time and cost an average of $19.09 per participant enrolled. Social media recruitment reached more potential participants, but the cost per participant enrolled was higher compared to traditional methods. Studies need to consider benefits and downsides of traditional and social media recruitment methods based on study goals and population.
Khreis, Haneen; Nieuwenhuijsen, Mark J
2017-03-17
Background : Current levels of traffic-related air pollution (TRAP) are associated with the development of childhood asthma, although some inconsistencies and heterogeneity remain. An important part of the uncertainty in studies of TRAP-associated asthma originates from uncertainties in the TRAP exposure assessment and assignment methods. In this work, we aim to systematically review the exposure assessment methods used in the epidemiology of TRAP and childhood asthma, highlight recent advances, remaining research gaps and make suggestions for further research. Methods : We systematically reviewed epidemiological studies published up until 8 September 2016 and available in Embase, Ovid MEDLINE (R), and "Transport database". We included studies which examined the association between children's exposure to TRAP metrics and their risk of "asthma" incidence or lifetime prevalence, from birth to the age of 18 years old. Results : We found 42 studies which examined the associations between TRAP and subsequent childhood asthma incidence or lifetime prevalence, published since 1999. Land-use regression modelling was the most commonly used method and nitrogen dioxide (NO₂) was the most commonly used pollutant in the exposure assessments. Most studies estimated TRAP exposure at the residential address and only a few considered the participants' mobility. TRAP exposure was mostly assessed at the birth year and only a few studies considered different and/or multiple exposure time windows. We recommend that further work is needed including e.g., the use of new exposure metrics such as the composition of particulate matter, oxidative potential and ultra-fine particles, improved modelling e.g., by combining different exposure assessment models, including mobility of the participants, and systematically investigating different exposure time windows. Conclusions : Although our previous meta-analysis found statistically significant associations for various TRAP exposures and subsequent childhood asthma, further refinement of the exposure assessment may improve the risk estimates, and shed light on critical exposure time windows, putative agents, underlying mechanisms and drivers of heterogeneity.
Khreis, Haneen; Nieuwenhuijsen, Mark J.
2017-01-01
Background: Current levels of traffic-related air pollution (TRAP) are associated with the development of childhood asthma, although some inconsistencies and heterogeneity remain. An important part of the uncertainty in studies of TRAP-associated asthma originates from uncertainties in the TRAP exposure assessment and assignment methods. In this work, we aim to systematically review the exposure assessment methods used in the epidemiology of TRAP and childhood asthma, highlight recent advances, remaining research gaps and make suggestions for further research. Methods: We systematically reviewed epidemiological studies published up until 8 September 2016 and available in Embase, Ovid MEDLINE (R), and “Transport database”. We included studies which examined the association between children’s exposure to TRAP metrics and their risk of “asthma” incidence or lifetime prevalence, from birth to the age of 18 years old. Results: We found 42 studies which examined the associations between TRAP and subsequent childhood asthma incidence or lifetime prevalence, published since 1999. Land-use regression modelling was the most commonly used method and nitrogen dioxide (NO2) was the most commonly used pollutant in the exposure assessments. Most studies estimated TRAP exposure at the residential address and only a few considered the participants’ mobility. TRAP exposure was mostly assessed at the birth year and only a few studies considered different and/or multiple exposure time windows. We recommend that further work is needed including e.g., the use of new exposure metrics such as the composition of particulate matter, oxidative potential and ultra-fine particles, improved modelling e.g., by combining different exposure assessment models, including mobility of the participants, and systematically investigating different exposure time windows. Conclusions: Although our previous meta-analysis found statistically significant associations for various TRAP exposures and subsequent childhood asthma, further refinement of the exposure assessment may improve the risk estimates, and shed light on critical exposure time windows, putative agents, underlying mechanisms and drivers of heterogeneity. PMID:28304360
Testing contamination source identification methods for water distribution networks
Seth, Arpan; Klise, Katherine A.; Siirola, John D.; ...
2016-04-01
In the event of contamination in a water distribution network (WDN), source identification (SI) methods that analyze sensor data can be used to identify the source location(s). Knowledge of the source location and characteristics are important to inform contamination control and cleanup operations. Various SI strategies that have been developed by researchers differ in their underlying assumptions and solution techniques. The following manuscript presents a systematic procedure for testing and evaluating SI methods. The performance of these SI methods is affected by various factors including the size of WDN model, measurement error, modeling error, time and number of contaminant injections,more » and time and number of measurements. This paper includes test cases that vary these factors and evaluates three SI methods on the basis of accuracy and specificity. The tests are used to review and compare these different SI methods, highlighting their strengths in handling various identification scenarios. These SI methods and a testing framework that includes the test cases and analysis tools presented in this paper have been integrated into EPA’s Water Security Toolkit (WST), a suite of software tools to help researchers and others in the water industry evaluate and plan various response strategies in case of a contamination incident. Lastly, a set of recommendations are made for users to consider when working with different categories of SI methods.« less
Evaluation of esophageal motility utilizing the functional lumen imaging probe (FLIP)
Carlson, Dustin A.; Kahrilas, Peter J.; Lin, Zhiyue; Hirano, Ikuo; Gonsalves, Nirmala; Listernick, Zoe; Ritter, Katherine; Tye, Michael; Ponds, Fraukje A.; Wong, Ian; Pandolfino, John E.
2016-01-01
Background Esophagogastric junction (EGJ) distensibility and distension-mediated peristalsis can be assessed with the functional lumen imaging probe (FLIP) during a sedated upper endoscopy. We aimed to describe esophageal motility assessment using FLIP topography in patients presenting with dysphagia. Methods 145 patients (ages 18 – 85, 54% female) with dysphagia that completed upper endoscopy with a 16-cm FLIP assembly and high-resolution manometry (HRM) were included. HRM was analyzed according to the Chicago Classification of esophageal motility disorders; major esophageal motility disorders were considered ‘abnormal’. FLIP studies were analyzed using a customized program to calculate the EGJ-distensibility index (DI) and generate FLIP topography plots to identify esophageal contractility patterns. FLIP topography was considered ‘abnormal’ if EGJ-DI was < 2.8 mm2/mmHg or contractility pattern demonstrated absent contractility or repetitive, retrograde contractions. Results HRM was abnormal in 111 (77%) patients: 70 achalasia (19 type I, 39 type II, 12 type III), 38 EGJ outflow obstruction, and three jackhammer esophagus. FLIP topography was abnormal in 106 (95%) of these patients, including all 70 achalasia patients. HRM was ‘normal’ in 34 (23%) patients: five ineffective esophageal motility and 29 normal motility. 17 (50%) had abnormal FLIP topography including 13 (37%) with abnormal EGJ-DI. Conclusions FLIP topography provides a well-tolerated method for esophageal motility assessment (especially to identify achalasia) at the time of upper endoscopy. FLIP topography findings that are discordant with HRM may indicate otherwise undetected abnormalities of esophageal function, thus FLIP provides an alternative and complementary method to HRM for evaluation of non-obstructive dysphagia. PMID:27725650
NASA Astrophysics Data System (ADS)
Elder, Delwin L.; Johnson, Lewis E.; Tillack, Andreas F.; Robinson, Bruce H.; Haffner, Christian; Heni, Wolfgang; Hoessbacher, Claudia; Fedoryshyn, Yuriy; Salamin, Yannick; Baeuerle, Benedikt; Josten, Arne; Ayata, Masafumi; Koch, Ueli; Leuthold, Juerg; Dalton, Larry R.
2018-02-01
Multi-scale (correlated quantum and statistical mechanics) modeling methods have been advanced and employed to guide the improvement of organic electro-optic (OEO) materials, including by analyzing electric field poling induced electro-optic activity in nanoscopic plasmonic-organic hybrid (POH) waveguide devices. The analysis of in-device electro-optic activity emphasizes the importance of considering both the details of intermolecular interactions within organic electro-optic materials and interactions at interfaces between OEO materials and device architectures. Dramatic improvement in electro-optic device performance-including voltage-length performance, bandwidth, energy efficiency, and lower optical losses have been realized. These improvements are critical to applications in telecommunications, computing, sensor technology, and metrology. Multi-scale modeling methods illustrate the complexity of improving the electro-optic activity of organic materials, including the necessity of considering the trade-off between improving poling-induced acentric order through chromophore modification and the reduction of chromophore number density associated with such modification. Computational simulations also emphasize the importance of developing chromophore modifications that serve multiple purposes including matrix hardening for enhanced thermal and photochemical stability, control of matrix dimensionality, influence on material viscoelasticity, improvement of chromophore molecular hyperpolarizability, control of material dielectric permittivity and index of refraction properties, and control of material conductance. Consideration of new device architectures is critical to the implementation of chipscale integration of electronics and photonics and achieving the high bandwidths for applications such as next generation (e.g., 5G) telecommunications.
NASA Astrophysics Data System (ADS)
Ma, Hélène; Gronchi, Giovanni F.
2014-07-01
We advertise a new method of preliminary orbit determination for space debris using radar observations, which we call Infang †. We can perform a linkage of two sets of four observations collected at close times. The context is characterized by the accuracy of the range ρ, whereas the right ascension α and the declination δ are much more inaccurate due to observational errors. This method can correct α, δ, assuming the exact knowledge of the range ρ. Considering no perturbations from the J 2 effect, but including errors in the observations, we can compare the new method, the classical method of Gibbs, and the more recent Keplerian integrals method. The development of Infang is still on-going and will be further improved and tested.
NASA Astrophysics Data System (ADS)
Leiserson, Mark D. M.; Tatar, Diana; Cowen, Lenore J.; Hescott, Benjamin J.
A new method based on a mathematically natural local search framework for max cut is developed to uncover functionally coherent module and BPM motifs in high-throughput genetic interaction data. Unlike previous methods which also consider physical protein-protein interaction data, our method utilizes genetic interaction data only; this becomes increasingly important as high-throughput genetic interaction data is becoming available in settings where less is known about physical interaction data. We compare modules and BPMs obtained to previous methods and across different datasets. Despite needing no physical interaction information, the BPMs produced by our method are competitive with previous methods. Biological findings include a suggested global role for the prefoldin complex and a SWR subcomplex in pathway buffering in the budding yeast interactome.
Leiserson, Mark D M; Tatar, Diana; Cowen, Lenore J; Hescott, Benjamin J
2011-11-01
A new method based on a mathematically natural local search framework for max cut is developed to uncover functionally coherent module and BPM motifs in high-throughput genetic interaction data. Unlike previous methods, which also consider physical protein-protein interaction data, our method utilizes genetic interaction data only; this becomes increasingly important as high-throughput genetic interaction data is becoming available in settings where less is known about physical interaction data. We compare modules and BPMs obtained to previous methods and across different datasets. Despite needing no physical interaction information, the BPMs produced by our method are competitive with previous methods. Biological findings include a suggested global role for the prefoldin complex and a SWR subcomplex in pathway buffering in the budding yeast interactome.
Acoustic fatigue: Overview of activities at NASA Langley
NASA Technical Reports Server (NTRS)
Mixson, John S.; Roussos, Louis A.
1987-01-01
A number of aircraft and spacecraft configurations are being considered for future development. These include high-speed turboprop aircraft, advanced vertical take-off and landing fighter aircraft, and aerospace planes for hypersonic intercontinental cruise or flight to orbit and return. Review of the acoustic environment expected for these vehicles indicates levels high enough that acoustic fatigue must be considered. Unfortunately, the sonic fatique design technology used for current aircraft may not be adequate for these future vehicles. This has resulted in renewed emphasis on acoustic fatigue research at the NASA Langley Research Center. The overall objective of the Langley program is to develop methods and information for design of aerospace vehicles that will resist acoustic fatigue. The program includes definition of the acoustic loads acting on structures due to exhaust jets of boundary layers, and subsequent determination of the stresses within the structure due to these acoustic loads. Material fatigue associated with the high frequency structural stress reversal patterns resulting from acoustic loadings is considered to be an area requiring study, but no activity is currently underway.
Transition operators in electromagnetic-wave diffraction theory. II - Applications to optics
NASA Technical Reports Server (NTRS)
Hahne, G. E.
1993-01-01
The theory developed by Hahne (1992) for the diffraction of time-harmonic electromagnetic waves from fixed obstacles is briefly summarized and extended. Applications of the theory are considered which comprise, first, a spherical harmonic expansion of the so-called radiation impedance operator in the theory, for a spherical surface, and second, a reconsideration of familiar short-wavelength approximation from the new standpoint, including a derivation of the so-called physical optics method on the basis of quasi-planar approximation to the radiation impedance operator, augmented by the method of stationary phase. The latter includes a rederivation of the geometrical optics approximation for the complete Green's function for the electromagnetic field in the presence of a smooth- and a convex-surfaced perfectly electrically conductive obstacle.
Marom, Gil; Bluestein, Danny
2016-01-01
This paper evaluated the influence of various numerical implementation assumptions on predicting blood damage in cardiovascular devices using Lagrangian methods with Eulerian computational fluid dynamics. The implementation assumptions that were tested included various seeding patterns, stochastic walk model, and simplified trajectory calculations with pathlines. Post processing implementation options that were evaluated included single passage and repeated passages stress accumulation and time averaging. This study demonstrated that the implementation assumptions can significantly affect the resulting stress accumulation, i.e., the blood damage model predictions. Careful considerations should be taken in the use of Lagrangian models. Ultimately, the appropriate assumptions should be considered based the physics of the specific case and sensitivity analysis, similar to the ones presented here, should be employed.
Spin formalism and applications to new physics searches
DOE Office of Scientific and Technical Information (OSTI.GOV)
Haber, H.E.
1994-12-01
An introduction to spin techniques in particle physics is given. Among the topics covered are: helicity formalism and its applications to the decay and scattering of spin-1/2 and spin-1 particles, techniques for evaluating helicity amplitudes (including projection operator methods and the spinor helicity method), and density matrix techniques. The utility of polarization and spin correlations for untangling new physics beyond the Standard Model at future colliders such as the LHC and a high energy e{sup +}e{sup {minus}} linear collider is then considered. A number of detailed examples are explored including the search for low-energy supersymmetry, a non-minimal Higgs boson sector,more » and new gauge bosons beyond the W{sup {+-}} and Z.« less
NASA Astrophysics Data System (ADS)
Rakotomanga, Prisca; Soussen, Charles; Blondel, Walter C. P. M.
2017-03-01
Diffuse reflectance spectroscopy (DRS) has been acknowledged as a valuable optical biopsy tool for in vivo characterizing pathological modifications in epithelial tissues such as cancer. In spatially resolved DRS, accurate and robust estimation of the optical parameters (OP) of biological tissues is a major challenge due to the complexity of the physical models. Solving this inverse problem requires to consider 3 components: the forward model, the cost function, and the optimization algorithm. This paper presents a comparative numerical study of the performances in estimating OP depending on the choice made for each of the latter components. Mono- and bi-layer tissue models are considered. Monowavelength (scalar) absorption and scattering coefficients are estimated. As a forward model, diffusion approximation analytical solutions with and without noise are implemented. Several cost functions are evaluated possibly including normalized data terms. Two local optimization methods, Levenberg-Marquardt and TrustRegion-Reflective, are considered. Because they may be sensitive to the initial setting, a global optimization approach is proposed to improve the estimation accuracy. This algorithm is based on repeated calls to the above-mentioned local methods, with initial parameters randomly sampled. Two global optimization methods, Genetic Algorithm (GA) and Particle Swarm Optimization (PSO), are also implemented. Estimation performances are evaluated in terms of relative errors between the ground truth and the estimated values for each set of unknown OP. The combination between the number of variables to be estimated, the nature of the forward model, the cost function to be minimized and the optimization method are discussed.
Galeotti, Francesco; Barile, Elisa; Lanzotti, Virginia; Dolci, Marcello; Curir, Paolo
2008-01-01
One flavone-C-glycoside and two flavonol-O-glycosides were recognized and isolated as the main flavonoidal components in nine different carnation cultivars, and their chemical structures have been determined by spectroscopic methods, including UV detection, MS and NMR. The distribution of these three compounds in flowers, leaves, stems, young sprouts, and roots of each cultivar was evaluated by a simple HPLC-UV method: the graphic representation of their content in the different tissues allows to identify and characterize unambiguously each considered carnation cultivar. The presented method could be an easy, inexpensive and reliable tool for carnation cultivar discrimination.
An approach to operating system testing
NASA Technical Reports Server (NTRS)
Sum, R. N., Jr.; Campbell, R. H.; Kubitz, W. J.
1984-01-01
To ensure the reliability and performance of a new system, it must be verified or validated in some manner. Currently, testing is the only resonable technique available for doing this. Part of this testing process is the high level system test. System testing is considered with respect to operating systems and in particular UNIX. This consideration results in the development and presentation of a good method for performing the system test. The method includes derivations from the system specifications and ideas for management of the system testing project. Results of applying the method to the IBM System/9000 XENIX operating system test and the development of a UNIX test suite are presented.
NASA Astrophysics Data System (ADS)
Antsiferov, SV; Sammal, AS; Deev, PV
2018-03-01
To determine the stress-strain state of multilayer support of vertical shafts, including cross-sectional deformation of the tubing rings as against the design, the authors propose an analytical method based on the provision of the mechanics of underground structures and surrounding rock mass as the elements of an integrated deformable system. The method involves a rigorous solution of the corresponding problem of elasticity, obtained using the mathematical apparatus of the theory of analytic functions of a complex variable. The design method is implemented as a software program allowing multivariate applied computation. Examples of the calculation are given.
Generalized Pattern Search methods for a class of nonsmooth optimization problems with structure
NASA Astrophysics Data System (ADS)
Bogani, C.; Gasparo, M. G.; Papini, A.
2009-07-01
We propose a Generalized Pattern Search (GPS) method to solve a class of nonsmooth minimization problems, where the set of nondifferentiability is included in the union of known hyperplanes and, therefore, is highly structured. Both unconstrained and linearly constrained problems are considered. At each iteration the set of poll directions is enforced to conform to the geometry of both the nondifferentiability set and the boundary of the feasible region, near the current iterate. This is the key issue to guarantee the convergence of certain subsequences of iterates to points which satisfy first-order optimality conditions. Numerical experiments on some classical problems validate the method.
Rationale for windshield glass system specification requirements for shuttle orbiter
NASA Technical Reports Server (NTRS)
Hayashida, K.; King, G. L.; Tesinsiky, J.; Wittenburg, D. R.
1972-01-01
A preliminary procurement specification for the space shuttle orbiter windshield pane, and some of the design considerations and rationale leading to its development are presented. The windshield designer is given the necessary methods and procedures for assuring glass pane structural integrity by proof test. These methods and procedures are fully developed for annealed and thermally tempered aluminosilicate, borosilicate, and soda lime glass and for annealed fused silica. Application of the method to chemically tempered glass is considered. Other considerations are vision requirements, protection against bird impact, hail, frost, rain, and meteoroids. The functional requirements of the windshield system during landing, ferrying, boost, space flight, and entry are included.
Contraception Update: Sterilization.
Antell, Karen; Deshmukh, Prium; Brown, Elizabeth J
2017-11-01
Female sterilization procedures include postpartum partial salpingectomy via cesarean or minilaparotomy incision, interval laparoscopic procedures, or hysteroscopic placement of microinserts. Rates of failure and serious complications are low and comparable among the various methods. A hysteroscopic procedure requires a 3-month confirmatory hysterosalpingogram before it is considered effective for contraception. Hysteroscopic sterilization has been shown to be associated with a higher reoperation rate than laparoscopic procedures. For male sterilization, vasectomy is a noninvasive and highly effective method. Vasectomy is an outpatient procedure performed under local anesthesia. The procedure requires confirmation of azoospermia with a semen analysis 8 to 16 weeks after the procedure. Patients who are considering sterilization should be counseled about all the available options and the permanent nature of such procedures. Written permission from the American Academy of Family Physicians is required for reproduction of this material in whole or in part in any form or medium.
NASA Astrophysics Data System (ADS)
Farrokhabadi, A.; Mokhtari, J.; Koochi, A.; Abadyan, M.
2015-06-01
In this paper, the impact of the Casimir attraction on the electromechanical stability of nanowire-fabricated nanotweezers is investigated using a theoretical continuum mechanics model. The Dirichlet mode is considered and an asymptotic solution, based on path integral approach, is applied to consider the effect of vacuum fluctuations in the model. The Euler-Bernoulli beam theory is employed to derive the nonlinear governing equation of the nanotweezers. The governing equations are solved by three different approaches, i.e. the modified variation iteration method, generalized differential quadrature method and using a lumped parameter model. Various perspectives of the problem, including the comparison with the van der Waals force regime, the variation of instability parameters and effects of geometry are addressed in present paper. The proposed approach is beneficial for the precise determination of the electrostatic response of the nanotweezers in the presence of Casimir force.
The spatial pattern of suicide in the US in relation to deprivation, fragmentation and rurality.
Congdon, Peter
2011-01-01
Analysis of geographical patterns of suicide and psychiatric morbidity has demonstrated the impact of latent ecological variables (such as deprivation, rurality). Such latent variables may be derived by conventional multivariate techniques from sets of observed indices (for example, by principal components), by composite variable methods or by methods which explicitly consider the spatial framework of areas and, in particular, the spatial clustering of latent risks and outcomes. This article considers a latent random variable approach to explaining geographical contrasts in suicide in the US; and it develops a spatial structural equation model incorporating deprivation, social fragmentation and rurality. The approach allows for such latent spatial constructs to be correlated both within and between areas. Potential effects of area ethnic mix are also included. The model is applied to male and female suicide deaths over 2002–06 in 3142 US counties.
Formal Methods for Autonomic and Swarm-based Systems
NASA Technical Reports Server (NTRS)
Rouff, Christopher; Vanderbilt, Amy; Hinchey, Mike; Truszkowski, Walt; Rash, James
2004-01-01
Swarms of intelligent rovers and spacecraft are being considered for a number of future NASA missions. These missions will provide MSA scientist and explorers greater flexibility and the chance to gather more science than traditional single spacecraft missions. These swarms of spacecraft are intended to operate for large periods of time without contact with the Earth. To do this, they must be highly autonomous, have autonomic properties and utilize sophisticated artificial intelligence. The Autonomous Nano Technology Swarm (ANTS) mission is an example of one of the swarm type of missions NASA is considering. This mission will explore the asteroid belt using an insect colony analogy cataloging the mass, density, morphology, and chemical composition of the asteroids, including any anomalous concentrations of specific minerals. Verifying such a system would be a huge task. This paper discusses ongoing work to develop a formal method for verifying swarm and autonomic systems.
Deloria-Knoll, Maria; Feikin, Daniel R; Scott, J Anthony G; O'Brien, Katherine L; DeLuca, Andrea N; Driscoll, Amanda J; Levine, Orin S
2012-04-01
Methods for the identification and selection of patients (cases) with severe or very severe pneumonia and controls for the Pneumonia Etiology Research for Child Health (PERCH) project were needed. Issues considered include eligibility criteria and sampling strategies, whether to enroll hospital or community controls, whether to exclude controls with upper respiratory tract infection (URTI) or nonsevere pneumonia, and matching criteria, among others. PERCH ultimately decided to enroll community controls and an additional human immunodeficiency virus (HIV)-infected control group at high HIV-prevalence sites matched on age and enrollment date of cases; controls with symptoms of URTI or nonsevere pneumonia will not be excluded. Systematic sampling of cases (when necessary) and random sampling of controls will be implemented. For each issue, we present the options that were considered, the advantages and disadvantages of each, the rationale for the methods selected for PERCH, and remaining implications and limitations.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Li, Kenan; Jacobsen, Chris
Fresnel zone plates used for X-ray nanofocusing face high-aspect-ratio nanofabrication challenges in combining narrow transverse features (for high spatial resolution) along with extended optical modulation along the X-ray beam direction (to improve efficiency). The stacking of multiple Fresnel zone plates along the beam direction has already been shown to offer improved characteristics of resolution and efficiency when compared with thin single zone plates. Using multislice wave propagation simulation methods, here a number of new schemes for the stacking of multiple Fresnel zone plates are considered. These include consideration of optimal thickness and spacing in the axial direction, and methods tomore » capture a fraction of the light otherwise diffracted into unwanted orders, and instead bring it into the desired first-order focus. In conclusion, the alignment tolerances for stacking multiple Fresnel zone plates are also considered.« less
National Ecosystem Assessments in Europe: A Review
Schröter, Matthias; Albert, Christian; Marques, Alexandra; Tobon, Wolke; Lavorel, Sandra; Maes, Joachim; Brown, Claire; Klotz, Stefan; Bonn, Aletta
2016-01-01
Abstract National ecosystem assessments form an essential knowledge base for safeguarding biodiversity and ecosystem services. We analyze eight European (sub-)national ecosystem assessments (Portugal, United Kingdom, Spain, Norway, Flanders, Netherlands, Finland, and Germany) and compare their objectives, political context, methods, and operationalization. We observed remarkable differences in breadth of the assessment, methods employed, variety of services considered, policy mandates, and funding mechanisms. Biodiversity and ecosystem services are mainly assessed independently, with biodiversity conceptualized as underpinning services, as a source of conflict with services, or as a service in itself. Recommendations derived from our analysis for future ecosystem assessments include the needs to improve the common evidence base, to advance the mapping of services, to consider international flows of services, and to connect more strongly to policy questions. Although the context specificity of national ecosystem assessments is acknowledged as important, a greater harmonization across assessments could help to better inform common European policies and future pan-regional assessments. PMID:28533561
Low-Cost Approaches to III–V Semiconductor Growth for Photovoltaic Applications
Greenaway, Ann L.; Boucher, Jason W.; Oener, Sebastian Z.; ...
2017-08-31
III–V semiconductors form the most efficient single- and multijunction photovoltaics. Metal–organic vapor-phase epitaxy, which uses toxic and pyrophoric gas-phase precursors, is the primary commercial growth method for these materials. In order for the use of highly efficient III–V-based devices to be expanded as the demand for renewable electricity grows, a lower-cost approach to the growth of these materials is needed. This Review focuses on three deposition techniques compatible with current device architectures: hydride vapor-phase epitaxy, close-spaced vapor transport, and thin-film vapor–liquid–solid growth. Here, we consider recent advances in each technique, including the available materials space, before providing an in-depth comparisonmore » of growth technology advantages and limitations and considering the impact of modifications to the method of production on the cost of the final photovoltaics.« less
Precision digital control systems
NASA Astrophysics Data System (ADS)
Vyskub, V. G.; Rozov, B. S.; Savelev, V. I.
This book is concerned with the characteristics of digital control systems of great accuracy. A classification of such systems is considered along with aspects of stabilization, programmable control applications, digital tracking systems and servomechanisms, and precision systems for the control of a scanning laser beam. Other topics explored are related to systems of proportional control, linear devices and methods for increasing precision, approaches for further decreasing the response time in the case of high-speed operation, possibilities for the implementation of a logical control law, and methods for the study of precision digital control systems. A description is presented of precision automatic control systems which make use of electronic computers, taking into account the existing possibilities for an employment of computers in automatic control systems, approaches and studies required for including a computer in such control systems, and an analysis of the structure of automatic control systems with computers. Attention is also given to functional blocks in the considered systems.
Improvement of the System of Training of Specialists by University for Coal Mining Enterprises
NASA Astrophysics Data System (ADS)
Mikhalchenko, Vadim; Seredkina, Irina
2017-11-01
In the article the ingenious technique of the Quality Function Deployment with reference to the process of training of specialists with higher education by university is considered. The method is based on the step-by-step conversion of customer requirements into specific organizational, meaningful and functional transformations of the technological process of the university. A fully deployed quality function includes four stages of tracking customer requirements while creating a product: product planning and design, process design, production design. The Quality Function Deployment can be considered as one of the methods for optimizing the technological processes of training of specialists with higher education in the current economic conditions. Implemented at the initial stages of the life cycle of the technological process, it ensures not only the high quality of the "product" of graduate school, but also the fullest possible satisfaction of consumer's requests and expectations.
Multiscale spatial and temporal estimation of the b-value
NASA Astrophysics Data System (ADS)
García-Hernández, R.; D'Auria, L.; Barrancos, J.; Padilla, G.
2017-12-01
The estimation of the spatial and temporal variations of the Gutenberg-Richter b-value is of great importance in different seismological applications. One of the problems affecting its estimation is the heterogeneous distribution of the seismicity which makes its estimate strongly dependent upon the selected spatial and/or temporal scale. This is especially important in volcanoes where dense clusters of earthquakes often overlap the background seismicity. Proposed solutions for estimating temporal variations of the b-value include considering equally spaced time intervals or variable intervals having an equal number of earthquakes. Similar approaches have been proposed to image the spatial variations of this parameter as well.We propose a novel multiscale approach, based on the method of Ogata and Katsura (1993), allowing a consistent estimation of the b-value regardless of the considered spatial and/or temporal scales. Our method, named MUST-B (MUltiscale Spatial and Temporal characterization of the B-value), basically consists in computing estimates of the b-value at multiple temporal and spatial scales, extracting for a give spatio-temporal point a statistical estimator of the value, as well as and indication of the characteristic spatio-temporal scale. This approach includes also a consistent estimation of the completeness magnitude (Mc) and of the uncertainties over both b and Mc.We applied this method to example datasets for volcanic (Tenerife, El Hierro) and tectonic areas (Central Italy) as well as an example application at global scale.
Reinstating the 'Queen': understanding philosophical inquiry in nursing.
Pesut, Barbara; Johnson, Joy
2008-01-01
This paper is an introduction to the characteristics of philosophical inquiry. Despite over a century of philosophical thinking in nursing, philosophical inquiry has yet to be positioned as contributing substantially to the field of nursing's inquiry. There is a need to articulate the nature and characteristics of philosophical inquiry for researchers new to this perspective. We begin by addressing a common question that surfaces when one begins a work that is philosophical in nature, how does one differentiate between nursing philosophy and nursing theory? We then address the nature and characteristics of philosophical inquiry. We conclude by considering the question of whether philosophical inquiry might be considered a form of qualitative inquiry. Unlike science, which relies upon investigative methods, philosophical inquiry relies upon the capacities to think and reason. Problems characteristic of philosophical inquiry include conceptual clarification, analysis of arguments and problems related to the ontology, epistemology and ethics of nursing. Although methodological approaches to philosophical inquiry are diverse, common tools include assumptions and the intellectual processes of conceptualizing, judging and reasoning within a context of wonder. Some have argued that to neglect philosophy in nursing is to place the discipline at risk. However, there is little guidance available to researchers new to this method of inquiry. By providing a beginning roadmap, our hope is that philosophical inquiry will take its place alongside scientific methods of inquiry with the goal of constructing robust knowledge for the discipline of nursing.
ERIC Educational Resources Information Center
Kliewer, Christopher; Fitzgerald, Linda May; Meyer-Mork, Jodi; Hartman, Patresa; English-Sand, Pat; Raschke, Donna
2004-01-01
In this study, Christopher Kliewer, Linda Fitzgerald, Jodi Meyer-Mork, Patresa Hartman, Pat English-Sand, and Donna Raschke use ethnographic methods to explore literacy development in young children considered to have significant disabilities. The study settings included nine preschool and kindergarten classrooms across five programs, all of which…
Methotrexate for rheumatoid arthritis patients who are on hemodialysis.
Al-Hasani, Hasanein; Roussou, Euthalia
2011-12-01
Methotrexate (MTX) can be toxic to patients suffering from end stage renal disease (ESRD) on hemodialysis even at low doses. This increase in toxicity is more notable in terms of bone marrow suppression in the form of pancytopenia. Many methods of elimination including dialysis itself have been proven ineffective, and alternate treatments with anti-TNF alpha blockers can be considered.
NASA Technical Reports Server (NTRS)
Bains, R. W.; Herwig, H. A.; Luedeman, J. K.; Torina, E. M.
1974-01-01
The Shuttle Electric Power System (SEPS) computer program is considered in terms of the program manual, programmer guide, and program utilization. The main objective is to provide the information necessary to interpret and use the routines comprising the SEPS program. Subroutine descriptions including the name, purpose, method, variable definitions, and logic flow are presented.
ERIC Educational Resources Information Center
Brown, David R.; Brewster, Cheryl D.; Karides, Marina; Lukas, Lou A.
2011-01-01
Collaboration is essential to manage complex real world problems. We used phenomenologic methods to elaborate a description of collaboration between two departments at an academic medical center who considered their relationship to represent a model of effective collaboration. Key collaborative structures included a shared vision and commitment by…
Dynamic target ionization using an ultrashort pulse of a laser field
NASA Astrophysics Data System (ADS)
Makarov, D. N.; Matveev, V. I.; Makarova, K. A.
2014-09-01
Ionization processes under the interaction of an ultrashort pulse of an electromagnetic field with atoms in nonstationary states are considered. As an example, the ionization probability of the hydrogen-like atom upon the decay of quasi-stationary state is calculated. The method developed can be applied to complex systems, including targets in collisional states and various chemical reactions.
Kinematic support using elastic elements
NASA Technical Reports Server (NTRS)
Geirsson, Arni; Debra, Daniel B.
1988-01-01
The design of kinematic supports using elastic elements is reviewed. The two standard methods (cone, Vee and flat and three Vees) are presented and a design example involving a machine tool metrology bench is given. Design goals included thousandfold strain attenuation in the bench relative to the base when the base strains due to temperature variations and shifting loads. Space applications are also considered.
ERIC Educational Resources Information Center
Lester, Pamela Denise
2017-01-01
A qualitative method of research was chosen for this study. This ethnographic case study examined school psychologists' and the referral process for special education services. The participants included school psychologists in a specific county in the state of Maryland. School psychologists are considered crucial members of an Individualized…
Michael R. Vanderberg; Kevin Boston; John Bailey
2011-01-01
Accounting for the probability of loss due to disturbance events can influence the prediction of carbon flux over a planning horizon, and can affect the determination of optimal silvicultural regimes to maximize terrestrial carbon storage. A preliminary model that includes forest disturbance-related carbon loss was developed to maximize expected values of carbon stocks...
ERIC Educational Resources Information Center
Crossley, Michael, Ed.; Vulliamy, Graham, Ed.
This book contains 11 essays that offer in-depth accounts of qualitative research in developing countries. Each chapter focuses upon a specific method and considers related theoretical and practical issues with reference to recent experiences in selected developing countries. Key issues addressed include: (1) the identification of appropriate…
Factors associated with fertility moderation in India.
Sharma, S; Singhal, D S; Sharma, B B; Gupta, Y P
1991-04-01
The authors analyze intermediate variables associated with fertility decline in India from the 1960s to 1988. The focus is on comparisons among states as revealed primarily by data on couples protected from unwanted pregnancies by family planning methods. Variables considered include female age at marriage, female literacy, infant mortality, poverty, expenditure on health and family welfare, and income. Data are from official sources.
The rise and fall of dissociative identity disorder.
Paris, Joel
2012-12-01
Dissociative identity disorder (DID), once considered rare, was frequently diagnosed during the 1980s and 1990s, after which interest declined. This is the trajectory of a medical fad. DID was based on poorly conceived theories and used potentially damaging treatment methods. The problem continues, given that the DSM-5 includes DID and accords dissociative disorders a separate chapter in its manual.
Adaptive coding of MSS imagery. [Multi Spectral band Scanners
NASA Technical Reports Server (NTRS)
Habibi, A.; Samulon, A. S.; Fultz, G. L.; Lumb, D.
1977-01-01
A number of adaptive data compression techniques are considered for reducing the bandwidth of multispectral data. They include adaptive transform coding, adaptive DPCM, adaptive cluster coding, and a hybrid method. The techniques are simulated and their performance in compressing the bandwidth of Landsat multispectral images is evaluated and compared using signal-to-noise ratio and classification consistency as fidelity criteria.
ERIC Educational Resources Information Center
Dean, Charlotte
2007-01-01
The paper reflects on how photography was used during a small piece of mixed methods qualitative research with some young Travellers and their families. The context for the study was "Every Child Matters" and the accompanying push aimed to include groups considered as marginalised, such as young Travellers, in mainstream educational provision, as…
NASA Technical Reports Server (NTRS)
El-Sum, H. M. A.
1976-01-01
The international status of the art of acousto optical imaging techniques adaptable to nondestructive testing and, interfacing methods for acoustical and optical holography in nondestructive testing research are studied. Evaluation of 20 different techniques encompassed investigation of varieties of detectors and detection schemes, all of which are described and summarized. Related investigation is reported in an Appendix. Important remarks on image quality, factors to be considered in designing a particular system, and conclusions and recommendations are presented. Three bibliographies are included.
Jump phenomena. [large amplitude responses of nonlinear systems
NASA Technical Reports Server (NTRS)
Reiss, E. L.
1980-01-01
The paper considers jump phenomena composed of large amplitude responses of nonlinear systems caused by small amplitude disturbances. Physical problems where large jumps in the solution amplitude are important features of the response are described, including snap buckling of elastic shells, chemical reactions leading to combustion and explosion, and long-term climatic changes of the earth's atmosphere. A new method of rational functions was then developed which consists of representing the solutions of the jump problems as rational functions of the small disturbance parameter; this method can solve jump problems explicitly.
NASA Astrophysics Data System (ADS)
Fain, M. K.; Starinova, O. L.
2016-04-01
The paper outlines the method for determination of the locally optimal stepwise control structure in the problem of the low thrust spacecraft transfer optimization in the Earth-Moon system, including the L1-L2 transfer. The total flight time as an optimization criterion is considered. The optimal control programs were obtained by using the Pontryagin's maximum principle. As a result of optimization, optimal control programs, corresponding trajectories, and minimal total flight times were determined.
NASA Astrophysics Data System (ADS)
Sawicki, J.; Siedlaczek, P.; Staszczyk, A.
2018-03-01
A numerical three-dimensional model for computing residual stresses generated in cross section of steel 42CrMo4 after nitriding is presented. The diffusion process is analyzed by the finite-element method. The internal stresses are computed using the obtained profile of the distribution of the nitrogen concentration. The special features of the intricate geometry of the treated articles including edges and angles are considered. Comparative analysis of the results of the simulation and of the experimental measurement of residual stresses is performed by the Waisman-Philips method.
How economics can further the success of ecological restoration.
Iftekhar, Md Sayed; Polyakov, Maksym; Ansell, Dean; Gibson, Fiona; Kay, Geoffrey M
2017-04-01
Restoration scientists and practitioners have recently begun to include economic and social aspects in the design and investment decisions for restoration projects. With few exceptions, ecological restoration studies that include economics focus solely on evaluating costs of restoration projects. However, economic principles, tools, and instruments can be applied to a range of other factors that affect project success. We considered the relevance of applying economics to address 4 key challenges of ecological restoration: assessing social and economic benefits, estimating overall costs, project prioritization and selection, and long-term financing of restoration programs. We found it is uncommon to consider all types of benefits (such as nonmarket values) and costs (such as transaction costs) in restoration programs. Total benefit of a restoration project can be estimated using market prices and various nonmarket valuation techniques. Total cost of a project can be estimated using methods based on property or land-sale prices, such as hedonic pricing method and organizational surveys. Securing continuous (or long-term) funding is also vital to accomplishing restoration goals and can be achieved by establishing synergy with existing programs, public-private partnerships, and financing through taxation. © 2016 Society for Conservation Biology.
A systematic review of dynamics in climate risk and vulnerability assessments
NASA Astrophysics Data System (ADS)
Jurgilevich, Alexandra; Räsänen, Aleksi; Groundstroem, Fanny; Juhola, Sirkku
2017-01-01
Understanding climate risk is crucial for effective adaptation action, and a number of assessment methodologies have emerged. We argue that the dynamics of the individual components in climate risk and vulnerability assessments has received little attention. In order to highlight this, we systematically reviewed 42 sub-national climate risk and vulnerability assessments. We analysed the assessments using an analytical framework with which we evaluated (1) the conceptual approaches to vulnerability and exposure used, (2) if current or future risks were assessed, and (3) if and how changes over time (i.e. dynamics) were considered. Of the reviewed assessments, over half addressed future risks or vulnerability; and of these future-oriented studies, less than 1/3 considered both vulnerability and exposure dynamics. While the number of studies that include dynamics is growing, and while all studies included socio-economic aspects, often only biophysical dynamics was taken into account. We discuss the challenges of assessing socio-economic and spatial dynamics, particularly the poor availability of data and methods. We suggest that future-oriented studies assessing risk dynamics would benefit from larger stakeholder involvement, discussion of the assessment purpose, the use of multiple methods, inclusion of uncertainty/sensitivity analyses and pathway approaches.
Multiple disturbances classifier for electric signals using adaptive structuring neural networks
NASA Astrophysics Data System (ADS)
Lu, Yen-Ling; Chuang, Cheng-Long; Fahn, Chin-Shyurng; Jiang, Joe-Air
2008-07-01
This work proposes a novel classifier to recognize multiple disturbances for electric signals of power systems. The proposed classifier consists of a series of pipeline-based processing components, including amplitude estimator, transient disturbance detector, transient impulsive detector, wavelet transform and a brand-new neural network for recognizing multiple disturbances in a power quality (PQ) event. Most of the previously proposed methods usually treated a PQ event as a single disturbance at a time. In practice, however, a PQ event often consists of various types of disturbances at the same time. Therefore, the performances of those methods might be limited in real power systems. This work considers the PQ event as a combination of several disturbances, including steady-state and transient disturbances, which is more analogous to the real status of a power system. Six types of commonly encountered power quality disturbances are considered for training and testing the proposed classifier. The proposed classifier has been tested on electric signals that contain single disturbance or several disturbances at a time. Experimental results indicate that the proposed PQ disturbance classification algorithm can achieve a high accuracy of more than 97% in various complex testing cases.
NASA Astrophysics Data System (ADS)
Hosseini, Seyed Farhad; Hashemian, Ali; Moetakef-Imani, Behnam; Hadidimoud, Saied
2018-03-01
In the present paper, the isogeometric analysis (IGA) of free-form planar curved beams is formulated based on the nonlinear Timoshenko beam theory to investigate the large deformation of beams with variable curvature. Based on the isoparametric concept, the shape functions of the field variables (displacement and rotation) in a finite element analysis are considered to be the same as the non-uniform rational basis spline (NURBS) basis functions defining the geometry. The validity of the presented formulation is tested in five case studies covering a wide range of engineering curved structures including from straight and constant curvature to variable curvature beams. The nonlinear deformation results obtained by the presented method are compared to well-established benchmark examples and also compared to the results of linear and nonlinear finite element analyses. As the nonlinear load-deflection behavior of Timoshenko beams is the main topic of this article, the results strongly show the applicability of the IGA method to the large deformation analysis of free-form curved beams. Finally, it is interesting to notice that, until very recently, the large deformations analysis of free-form Timoshenko curved beams has not been considered in IGA by researchers.
Integrated nonthermal treatment system study
DOE Office of Scientific and Technical Information (OSTI.GOV)
Biagi, C.; Bahar, D.; Teheranian, B.
1997-01-01
This report presents the results of a study of nonthermal treatment technologies. The study consisted of a systematic assessment of five nonthermal treatment alternatives. The treatment alternatives consist of widely varying technologies for safely destroying the hazardous organic components, reducing the volume, and preparing for final disposal of the contact-handled mixed low-level waste (MLLW) currently stored in the US Department of Energy complex. The alternatives considered were innovative nonthermal treatments for organic liquids and sludges, process residue, soil and debris. Vacuum desorption or various washing approaches are considered for treatment of soil, residue and debris. Organic destruction methods include mediatedmore » electrochemical oxidation, catalytic wet oxidation, and acid digestion. Other methods studied included stabilization technologies and mercury separation of treatment residues. This study is a companion to the integrated thermal treatment study which examined 19 alternatives for thermal treatment of MLLW waste. The quantities and physical and chemical compositions of the input waste are based on the inventory database developed by the US Department of Energy. The Integrated Nonthermal Treatment Systems (INTS) systems were evaluated using the same waste input (2,927 pounds per hour) as the Integrated Thermal Treatment Systems (ITTS). 48 refs., 68 figs., 37 tabs.« less
Ethical Perspectives on RNA Interference Therapeutics
Ebbesen, Mette; Jensen, Thomas G.; Andersen, Svend; Pedersen, Finn Skou
2008-01-01
RNA interference is a mechanism for controlling normal gene expression which has recently begun to be employed as a potential therapeutic agent for a wide range of disorders, including cancer, infectious diseases and metabolic disorders. Clinical trials with RNA interference have begun. However, challenges such as off-target effects, toxicity and safe delivery methods have to be overcome before RNA interference can be considered as a conventional drug. So, if RNA interference is to be used therapeutically, we should perform a risk-benefit analysis. It is ethically relevant to perform a risk-benefit analysis since ethical obligations about not inflicting harm and promoting good are generally accepted. But the ethical issues in RNA interference therapeutics not only include a risk-benefit analysis, but also considerations about respecting the autonomy of the patient and considerations about justice with regard to the inclusion criteria for participation in clinical trials and health care allocation. RNA interference is considered a new and promising therapeutic approach, but the ethical issues of this method have not been greatly discussed, so this article analyses these issues using the bioethical theory of principles of the American bioethicists, Tom L. Beauchamp and James F. Childress. PMID:18612370
Methods for implementing a medicine outlet survey: lessons from the anti-malarial market.
O'Connell, Kathryn A; Poyer, Stephen; Solomon, Tsione; Munroe, Erik; Patouillard, Edith; Njogu, Julius; Evance, Illah; Hanson, Kara; Shewchuk, Tanya; Goodman, Catherine
2013-02-05
In recent years an increasing number of public investments and policy changes have been made to improve the availability, affordability and quality of medicines available to consumers in developing countries, including anti-malarials. It is important to monitor the extent to which these interventions are successful in achieving their aims using quantitative data on the supply side of the market. There are a number of challenges related to studying supply, including outlet sampling, gaining provider cooperation and collecting accurate data on medicines. This paper provides guidance on key steps to address these issues when conducting a medicine outlet survey in a developing country context. While the basic principles of good survey design and implementation are important for all surveys, there are a set of specific issues that should be considered when conducting a medicine outlet survey. This paper draws on the authors' experience of designing and implementing outlet surveys, including the lessons learnt from ACTwatch outlet surveys on anti-malarial retail supply, and other key studies in the field. Key lessons and points of debate are distilled around the following areas: selecting a sample of outlets; techniques for collecting and analysing data on medicine availability, price and sales volumes; and methods for ensuring high quality data in general. The authors first consider the inclusion criteria for outlets, contrasting comprehensive versus more focused approaches. Methods for developing a reliable sampling frame of outlets are then presented, including use of existing lists, key informants and an outlet census. Specific issues in the collection of data on medicine prices and sales volumes are discussed; and approaches for generating comparable price and sales volume data across products using the adult equivalent treatment dose (AETD) are explored. The paper concludes with advice on practical considerations, including questionnaire design, field worker training, and data collection. Survey materials developed by ACTwatch for investigating anti-malarial markets in sub-Saharan Africa and Asia provide a helpful resource for future studies in this area.
Type-2 fuzzy set extension of DEMATEL method combined with perceptual computing for decision making
NASA Astrophysics Data System (ADS)
Hosseini, Mitra Bokaei; Tarokh, Mohammad Jafar
2013-05-01
Most decision making methods used to evaluate a system or demonstrate the weak and strength points are based on fuzzy sets and evaluate the criteria with words that are modeled with fuzzy sets. The ambiguity and vagueness of the words and different perceptions of a word are not considered in these methods. For this reason, the decision making methods that consider the perceptions of decision makers are desirable. Perceptual computing is a subjective judgment method that considers that words mean different things to different people. This method models words with interval type-2 fuzzy sets that consider the uncertainty of the words. Also, there are interrelations and dependency between the decision making criteria in the real world; therefore, using decision making methods that cannot consider these relations is not feasible in some situations. The Decision-Making Trail and Evaluation Laboratory (DEMATEL) method considers the interrelations between decision making criteria. The current study used the combination of DEMATEL and perceptual computing in order to improve the decision making methods. For this reason, the fuzzy DEMATEL method was extended into type-2 fuzzy sets in order to obtain the weights of dependent criteria based on the words. The application of the proposed method is presented for knowledge management evaluation criteria.
Electron beam directed energy device and methods of using same
Retsky, Michael W.
2007-10-16
A method and apparatus is disclosed for an electron beam directed energy device. The device consists of an electron gun with one or more electron beams. The device includes one or more accelerating plates with holes aligned for beam passage. The plates may be flat or preferably shaped to direct each electron beam to exit the electron gun at a predetermined orientation. In one preferred application, the device is located in outer space with individual beams that are directed to focus at a distant target to be used to impact and destroy missiles. The aimings of the separate beams are designed to overcome Coulomb repulsion. A method is also presented for directing the beams to a target considering the variable terrestrial magnetic field. In another preferred application, the electron beam is directed into the ground to produce a subsurface x-ray source to locate and/or destroy buried or otherwise hidden objects including explosive devices.
Physiologic measures of sexual function in women: a review.
Woodard, Terri L; Diamond, Michael P
2009-07-01
To review and describe physiologic measures of assessing sexual function in women. Literature review. Studies that use instruments designed to measure female sexual function. Women participating in studies of female sexual function. Various instruments that measure physiologic features of female sexual function. Appraisal of the various instruments, including their advantages and disadvantages. Many unique physiologic methods of evaluating female sexual function have been developed during the past four decades. Each method has its benefits and limitations. Many physiologic methods exist, but most are not well-validated. In addition there has been an inability to correlate most physiologic measures with subjective measures of sexual arousal. Furthermore, given the complex nature of the sexual response in women, physiologic measures should be considered in context of other data, including the history, physical examination, and validated questionnaires. Nonetheless, the existence of appropriate physiologic measures is vital to our understanding of female sexual function and dysfunction.
Stopping power of dense plasmas: The collisional method and limitations of the dielectric formalism.
Clauser, C F; Arista, N R
2018-02-01
We present a study of the stopping power of plasmas using two main approaches: the collisional (scattering theory) and the dielectric formalisms. In the former case, we use a semiclassical method based on quantum scattering theory. In the latter case, we use the full description given by the extension of the Lindhard dielectric function for plasmas of all degeneracies. We compare these two theories and show that the dielectric formalism has limitations when it is used for slow heavy ions or atoms in dense plasmas. We present a study of these limitations and show the regimes where the dielectric formalism can be used, with appropriate corrections to include the usual quantum and classical limits. On the other hand, the semiclassical method shows the correct behavior for all plasma conditions and projectile velocity and charge. We consider different models for the ion charge distributions, including bare and dressed ions as well as neutral atoms.
Taser and Conducted Energy Weapons.
LeClair, Thomas G; Meriano, Tony
2015-01-01
It is clear that CEWs are an increasingly prevalent law enforcement tool, adopted to address a complex and challenging problem. The potential for serious injury from a single deployment of a CEW is extremely low. The debate regarding the link between these electrical weapons and sudden in-custody death is likely to continue because their use is often in complex and volatile situations. Any consideration of injuries has to be put into that context. One must also consider what injuries to a subject would result if an alternative force method was used. Furthermore, the potential benefits of CEWs, including reduction in injuries to the public and law-enforcement officers, need to be considered.
Synthesis and antitumour activity of 4-aminoquinazoline derivatives
NASA Astrophysics Data System (ADS)
Lipunova, G. N.; Nosova, E. V.; Charushin, V. N.; Chupakhin, O. N.
2016-07-01
Pieces of data on the synthesis and antitumour activity of 4-aminoquinazolines are summarized and analyzed. Key methods for the synthesis of these compounds are considered, primarily cyclocondensation of carboxylic acid derivatives, as well as the oxidation of quinazolines and the cyclization of disubstituted thioureas. Improvements of synthetic schemes for erlotinib, gefitinib and lapatinib, which are the best-known pharmaceuticals based on compounds of the title class, are also considered. Synthetic strategies and biological activities for new 4-aminoquinazoline derivatives that are EGFR-tyrosine kinase inhibitors, multiactive compounds, and labelled compounds for use as positron emission tomography (PET) imaging agents are discussed. The bibliography includes 263 references.
Effects of forming history on crash simulation of a vehicle
NASA Astrophysics Data System (ADS)
Gökler, M. İ.; Doğan, U. Ç.; Darendeliler, H.
2016-08-01
The effects of forming on the crash simulation of a vehicle have been investigated by considering the load paths produced by sheet metal forming process. The frontal crash analysis has been performed by the finite element method, firstly without considering the forming history, to find out the load paths that absorb the highest energy. The sheet metal forming simulations have been realized for each structural component of the load paths and the frontal crash analysis has been repeated by including forming history. The results of the simulations with and without forming effects have been compared with the physical crash test results available in literature.
A generalized vortex lattice method for subsonic and supersonic flow applications
NASA Technical Reports Server (NTRS)
Miranda, L. R.; Elliot, R. D.; Baker, W. M.
1977-01-01
If the discrete vortex lattice is considered as an approximation to the surface-distributed vorticity, then the concept of the generalized principal part of an integral yields a residual term to the vorticity-induced velocity field. The proper incorporation of this term to the velocity field generated by the discrete vortex lines renders the present vortex lattice method valid for supersonic flow. Special techniques for simulating nonzero thickness lifting surfaces and fusiform bodies with vortex lattice elements are included. Thickness effects of wing-like components are simulated by a double (biplanar) vortex lattice layer, and fusiform bodies are represented by a vortex grid arranged on a series of concentrical cylindrical surfaces. The analysis of sideslip effects by the subject method is described. Numerical considerations peculiar to the application of these techniques are also discussed. The method has been implemented in a digital computer code. A users manual is included along with a complete FORTRAN compilation, an executed case, and conversion programs for transforming input for the NASA wave drag program.
Sukegawa, Shintaro; Kanno, Takahiro; Shibata, Akane; Matsumoto, Kenichi; Sukegawa-Takahashi, Yuka; Sakaida, Kyousuke; Furuki, Yoshihiko
2017-07-01
As a surgical approach targeting the pterygopalatine fossa following maxillary cancer due to tumor invasion, Crockett's method is conventional and useful. However, if the tumor is confined to the area between the maxilla and pterygopalatine fossa, it is not necessary to include the zygomatico-orbital in the access osteotomy, and the orbital floor may be preserved. Depending on the range of tumor invasion, the current study reports a more minimally invasive, modified Crockett's surgery that may be considered, which includes resection with modified osteotomy lines and repositioning with fixation of the zygoma and zygomatic arch following maxillary cancer ablation. In addition, the majority of patients with advanced maxillary cancer may require postoperative radiotherapy or chemoradiotherapy following maxillectomy according to several guidelines. Therefore, using a low-profile bioactive resorbable plate system as a method of repositioning and fixing the resected and preserved zygoma and zygomatic arch may be more effective in this modified Crockett's method for maxillectomy.
Methods for structural design at elevated temperatures
NASA Technical Reports Server (NTRS)
Ellison, A. M.; Jones, W. E., Jr.; Leimbach, K. R.
1973-01-01
A procedure which can be used to design elevated temperature structures is discussed. The desired goal is to have the same confidence in the structural integrity at elevated temperature as the factor of safety gives on mechanical loads at room temperature. Methods of design and analysis for creep, creep rupture, and creep buckling are presented. Example problems are included to illustrate the analytical methods. Creep data for some common structural materials are presented. Appendix B is description, user's manual, and listing for the creep analysis program. The program predicts time to a given creep or to creep rupture for a material subjected to a specified stress-temperature-time spectrum. Fatigue at elevated temperature is discussed. Methods of analysis for high stress-low cycle fatigue, fatigue below the creep range, and fatigue in the creep range are included. The interaction of thermal fatigue and mechanical loads is considered, and a detailed approach to fatigue analysis is given for structures operating below the creep range.
Cederkvist, Karin; Jensen, Marina B; Holm, Peter E
2017-08-01
Stormwater treatment facilities (STFs) are becoming increasingly widespread but knowledge on their performance is limited. This is due to difficulties in obtaining representative samples during storm events and documenting removal of the broad range of contaminants found in stormwater runoff. This paper presents a method to evaluate STFs by addition of synthetic runoff with representative concentrations of contaminant species, including the use of tracer for correction of removal rates for losses not caused by the STF. A list of organic and inorganic contaminant species, including trace elements representative of runoff from roads is suggested, as well as relevant concentration ranges. The method was used for adding contaminants to three different STFs including a curbstone extension with filter soil, a dual porosity filter, and six different permeable pavements. Evaluation of the method showed that it is possible to add a well-defined mixture of contaminants despite different field conditions by having a flexibly system, mixing different stock-solutions on site, and use bromide tracer for correction of outlet concentrations. Bromide recovery ranged from only 12% in one of the permeable pavements to 97% in the dual porosity filter, stressing the importance of including a conservative tracer for correction of contaminant retention values. The method is considered useful in future treatment performance testing of STFs. The observed performance of the STFs is presented in coming papers. Copyright © 2017 Elsevier Ltd. All rights reserved.
Marembo, Joan; Zvinavashe, Mathilda; Nyamakura, Rudo; Shaibu, Sheila; Mogobe, Keitshokile Dintle
2014-10-01
To assess factors influencing infant-feeding methods selected by HIV-infected mothers. A descriptive quantitative study was conducted among 80 mothers with babies aged 0-6 months who were randomly selected and interviewed. Descriptive statistics were used to summarize the findings. Factors considered by women in choosing the infant-feeding methods included sociocultural acceptability (58.8%), feasibility and support from significant others (35%), knowledge of the selected method (55%), affordability (61.2%), implementation of the infant-feeding method without interference (62.5%), and safety (47.5%). Exclusive breast-feeding was the most preferred method of infant feeding. Disclosure of HIV status by a woman to her partner is a major condition for successful replacement feeding method, especially within the African cultural context. However, disclosure of HIV status to the partner was feared by most women as only 16.2% of the women disclosed their HIV status to partners. The factors considered by women in choosing the infant-feeding option were ability to implement the options without interference from significant others, affordability, and sociocultural acceptability. Knowledge of the selected option, its advantages and disadvantages, safety, and feasibility were also important factors. Nurses and midwives have to educate clients and support them in their choice of infant-feeding methods. © 2013 The Authors. Japan Journal of Nursing Science © 2013 Japan Academy of Nursing Science.
The hominin fossil record: taxa, grades and clades
Wood, Bernard; Lonergan, Nicholas
2008-01-01
This paper begins by reviewing the fossil evidence for human evolution. It presents summaries of each of the taxa recognized in a relatively speciose hominin taxonomy. These taxa are grouped in grades, namely possible and probable hominins, archaic hominins, megadont archaic hominins, transitional hominins, pre-modern Homo and anatomically modern Homo. The second part of this contribution considers some of the controversies that surround hominin taxonomy and systematics. The first is the vexed question of how you tell an early hominin from an early panin, or from taxa belonging to an extinct clade closely related to the Pan-Homo clade. Secondly, we consider how many species should be recognized within the hominin fossil record, and review the philosophies and methods used to identify taxa within the hominin fossil record. Thirdly, we examine how relationships within the hominin clade are investigated, including descriptions of the methods used to break down an integrated structure into tractable analytical units, and then how cladograms are generated and compared. We then review the internal structure of the hominin clade, including the problem of how many subclades should be recognized within the hominin clade, and we examine the reliability of hominin cladistic hypotheses. The last part of the paper reviews the concepts of a genus, including the criteria that should be used for recognizing genera within the hominin clade. PMID:18380861
Tolhurst, Rachel; Amekudzi, Yaa Peprah; Nyonator, Frank K; Bertel Squire, S; Theobald, Sally
2008-03-01
This paper explores the gendered dynamics of intra-household bargaining around treatment seeking for children with fever revealed through two qualitative research studies in the Volta Region of Ghana, and discusses the influence of different gender and health discourses on the likely policy implications drawn from such findings. Methods used included focus group discussions, in-depth and critical incidence interviews, and Participatory Learning and Action methods. We found that treatment seeking behaviour for children was influenced by norms of decision-making power and 'ownership' of children, access to and control over resources to pay for treatment, norms of responsibility for payment, marital status, household living arrangements, and the quality of relationships between mothers, fathers and elders. However, the implications of these findings may be interpreted from different perspectives. Most studies that have considered gender in relation to malaria have done so within a narrow biomedical approach to health that focuses only on the outcomes of gender relations in terms of the (non-)utilisation of allopathic healthcare. However, we argue that a 'gender transformatory' approach, which aims to promote women's empowerment, needs to include but go beyond this model, to consider broader potential outcomes of intra-household bargaining for women's and men's interests, including their livelihoods and 'bargaining positions'.
Quantifying phalangeal curvature: an empirical comparison of alternative methods.
Stern, J T; Jungers, W L; Susman, R L
1995-05-01
It has been generally assumed and theoretically argued that the curvature of finger and toe bones seen in some nonhuman primates is associated with cheiridial use in an arboreal setting. Assessment of such curvature in fossil primates has been used to infer the positional behavior of these animals. Several methods of quantifying curvature of bones have been proposed. The measure most commonly applied to phalanges is that of included angle, but this has come under some criticism. We consider various other approaches for quantifying phalangeal curvature, demonstrating that some are equivalent to use of included angle, but that one--normalized curvature moment arm (NCMA)--represents a true alternative. A comparison of NCMA to included angle, both calculated on manual and pedal proximal phalanges of humans, apes, some monkeys, and the Hadar fossils, revealed that these two different measures of curvature are highly correlated and result in very similar distributional patterns.
Efficient graph-cut tattoo segmentation
NASA Astrophysics Data System (ADS)
Kim, Joonsoo; Parra, Albert; Li, He; Delp, Edward J.
2015-03-01
Law enforcement is interested in exploiting tattoos as an information source to identify, track and prevent gang-related crimes. Many tattoo image retrieval systems have been described. In a retrieval system tattoo segmentation is an important step for retrieval accuracy since segmentation removes background information in a tattoo image. Existing segmentation methods do not extract the tattoo very well when the background includes textures and color similar to skin tones. In this paper we describe a tattoo segmentation approach by determining skin pixels in regions near the tattoo. In these regions graph-cut segmentation using a skin color model and a visual saliency map is used to find skin pixels. After segmentation we determine which set of skin pixels are connected with each other that form a closed contour including a tattoo. The regions surrounded by the closed contours are considered tattoo regions. Our method segments tattoos well when the background includes textures and color similar to skin.
Use and misuse of mixed methods in population oral health research: A scoping review.
Gupta, A; Keuskamp, D
2018-05-30
Despite the known benefits of a mixed methods approach in health research, little is known of its use in the field of population oral health. To map the extent of literature using a mixed methods approach to examine population oral health outcomes. For a comprehensive search of all the available literature published in the English language, databases including PubMed, Dentistry and Oral Sciences Source (DOSS), CINAHL, Web of Science and EMBASE (including Medline) were searched using a range of keywords from inception to October 2017. Only peer-reviewed, population-based studies of oral health outcomes conducted among non-institutionalised participants and using mixed methods were considered eligible for inclusion. Only nine studies met the inclusion criteria and were included in the review. The most frequent oral health outcome investigated was caries experience. However, most studies lacked a theoretical rationale or framework for using mixed methods, or supporting the use of qualitative data. Concurrent triangulation with a convergent design was the most commonly used mixed methods typology for integrating quantitative and qualitative data. The tools used to collect quantitative and qualitative data were mostly limited to surveys and interviews. With growing complexity recognised in the determinants of oral disease, future studies addressing population oral health outcomes are likely to benefit from the use of mixed methods. Explicit consideration of theoretical framework and methodology will strengthen those investigations. Copyright© 2018 Dennis Barber Ltd.
Beers, Erna; Egberts, Toine C. G.; Leufkens, Hubert G. M.; Jansen, Paul A. F.
2013-01-01
Background The ICH E7 guideline intends to improve the knowledge about medicines in geriatric patients. As a legislative document, it might not reflect the needs of healthcare professionals. This study investigated what information healthcare professionals, regulatory agencies and pharmaceutical industries consider necessary for rational drug prescribing to older individuals. Methods and Findings A 29-item-questionnaire was composed, considering the representation in trials, pharmacokinetics, efficacy, safety, and convenience of use in older individuals, with space for additions. Forty-three European professionals with an interest in medication for older individuals were included. In order to investigate their relevance, five items were included in a second questionnaire, with 11 control items. Median scores, differences between clinical and non-clinical respondents and response consistency were analysed. Consistency was present in 10 control items. Therefore, all items of the first questionnaire and the five additional items were analysed. Thirty-seven (86%) respondents returned the first questionnaire; 31/37 (84%) the second. Information about age-related differences in adverse events, locomotor effects, drug-disease interactions, dosing instructions, and information about the proportion of included 65+ patients was considered necessary by most respondents. Clinicians considered information significantly more important than the non-clinical respondents about the inclusion of 75+, time-until-benefit in older people, anticholinergic effects, drug-disease interactions, and convenience of use. Main study limitations are the focus on information for daily practice, while the ICH E7 guideline is a legislative document focused on market approval of a new medicine. Also, a questionnaire with a Likert scale has its limitations; this was addressed by providing space for comments. Conclusions This study reveals that items considered necessary are currently not included in the ICH E7 guideline. Also, clinicians’ and non-clinicians’ opinions differed significantly in 15% of the items. Therefore, all stakeholders should collaborate to improve the availability of information for the rational prescribing to older individuals. PMID:23977208
How to conduct a qualitative meta-analysis: Tailoring methods to enhance methodological integrity.
Levitt, Heidi M
2018-05-01
Although qualitative research has long been of interest in the field of psychology, meta-analyses of qualitative literatures (sometimes called meta-syntheses) are still quite rare. Like quantitative meta-analyses, these methods function to aggregate findings and identify patterns across primary studies, but their aims, procedures, and methodological considerations may vary. This paper explains the function of qualitative meta-analyses and their methodological development. Recommendations have broad relevance but are framed with an eye toward their use in psychotherapy research. Rather than arguing for the adoption of any single meta-method, this paper advocates for considering how procedures can best be selected and adapted to enhance a meta-study's methodological integrity. Through the paper, recommendations are provided to help researchers identify procedures that can best serve their studies' specific goals. Meta-analysts are encouraged to consider the methodological integrity of their studies in relation to central research processes, including identifying a set of primary research studies, transforming primary findings into initial units of data for a meta-analysis, developing categories or themes, and communicating findings. The paper provides guidance for researchers who desire to tailor meta-analytic methods to meet their particular goals while enhancing the rigor of their research.
Wang, Mo; Ling, Jie; Chen, Ying; Song, Jie; Sun, E; Shi, Zi-Qi; Feng, Liang; Jia, Xiao-Bin; Wei, Ying-Jie
2017-11-01
The increasingly apparent liver injury problems of bone strengthening Chinese medicines have brought challenges for clinical application, and it is necessary to consider both effectiveness and safety in screening anti-osteoporosis Chinese medicines. Metabolic transformation is closely related to drug efficacy and toxicity, so it is significant to comprehensively consider metabolism-action/toxicity(M-Act/Tox) for screening anti-osteoporosis Chinese medicines. The current evaluation models and the number of compounds(including metabolites) severely restrict efficient screening in vivo. By referring to previous relevant research and domestic and abroad literature, zebrafish M-Act/Tox integrative method was put forward for efficiently screening anti-osteoporosis herb medicines, which has organically integrated zebrafish metabolism model, osteoporosis model and toxicity evaluation method. This method can break through the bottleneck and blind spots that trace compositions can't achieve efficient and integrated in vivo evaluation, and realize both efficient and comprehensive screening on anti-osteoporosis traditional medicines based on in vivo process taking both safety and effectiveness into account, which is significant to accelerate discovery of effective and safe innovative traditional Chinese medicines for osteoporosis. Copyright© by the Chinese Pharmaceutical Association.
Status of turbulence modeling for hypersonic propulsion flowpaths
NASA Astrophysics Data System (ADS)
Georgiadis, Nicholas J.; Yoder, Dennis A.; Vyas, Manan A.; Engblom, William A.
2014-06-01
This report provides an assessment of current turbulent flow calculation methods for hypersonic propulsion flowpaths, particularly the scramjet engine. Emphasis is placed on Reynolds-averaged Navier-Stokes (RANS) methods, but some discussion of newer methods such as large eddy simulation (LES) is also provided. The report is organized by considering technical issues throughout the scramjet-powered vehicle flowpath, including laminar-to-turbulent boundary layer transition, shock wave/turbulent boundary layer interactions, scalar transport modeling (specifically the significance of turbulent Prandtl and Schmidt numbers), and compressible mixing. Unit problems are primarily used to conduct the assessment. In the combustor, results from calculations of a direct connect supersonic combustion experiment are also used to address the effects of turbulence model selection and in particular settings for the turbulent Prandtl and Schmidt numbers. It is concluded that RANS turbulence modeling shortfalls are still a major limitation to the accuracy of hypersonic propulsion simulations, whether considering individual components or an overall system. Newer methods such as LES-based techniques may be promising, but are not yet at a maturity to be used routinely by the hypersonic propulsion community. The need for fundamental experiments to provide data for turbulence model development and validation is discussed.
Cottenet, Geoffrey; Blancpain, Carine; Sonnard, Véronique; Chuah, Poh Fong
2013-08-01
Considering the increase of the total cultivated land area dedicated to genetically modified organisms (GMO), the consumers' perception toward GMO and the need to comply with various local GMO legislations, efficient and accurate analytical methods are needed for their detection and identification. Considered as the gold standard for GMO analysis, the real-time polymerase chain reaction (RTi-PCR) technology was optimised to produce a high-throughput GMO screening method. Based on simultaneous 24 multiplex RTi-PCR running on a ready-to-use 384-well plate, this new procedure allows the detection and identification of 47 targets on seven samples in duplicate. To comply with GMO analytical quality requirements, a negative and a positive control were analysed in parallel. In addition, an internal positive control was also included in each reaction well for the detection of potential PCR inhibition. Tested on non-GM materials, on different GM events and on proficiency test samples, the method offered high specificity and sensitivity with an absolute limit of detection between 1 and 16 copies depending on the target. Easy to use, fast and cost efficient, this multiplex approach fits the purpose of GMO testing laboratories.
Barker, F. Keith; Oyler-McCance, Sara; Tomback, Diana F.
2015-01-01
Next generation sequencing methods allow rapid, economical accumulation of data that have many applications, even at relatively low levels of genome coverage. However, the utility of shotgun sequencing data sets for specific goals may vary depending on the biological nature of the samples sequenced. We show that the ability to assemble mitogenomes from three avian samples of two different tissue types varies widely. In particular, data with coverage typical of microsatellite development efforts (∼1×) from DNA extracted from avian blood failed to cover even 50% of the mitogenome, relative to at least 500-fold coverage from muscle-derived data. Researchers should consider possible applications of their data and select the tissue source for their work accordingly. Practitioners analyzing low-coverage shotgun sequencing data (including for microsatellite locus development) should consider the potential benefits of mitogenome assembly, including internal barcode verification of species identity, mitochondrial primer development, and phylogenetics.
Causes and Recommendations for Unanticipated Ink Retention Following Tattoo Removal Treatment
Chen, Cynthia L.; Desai, Alpesh; Desai, Tejas
2013-01-01
While placement of ink into the skin is a long-standing tradition, patients are now seeking tattoo removal on a more frequent basis. Once considered acceptable removal options, tattoo ink removal via physical destruction included dermabrasion, chemical destruction, salabrasion, thermal destruction, and cryotherapy. Now these options are used extremely infrequently. These modalities provided unpredictable results and often required prolonged healing times and left patients with skin discoloration, pain, scarring, and ink retention. Even the widely adopted use of lasers, now considered the gold standard method, offers some level of unpredictability surrounding the natural progression of ink resolution. Multiple factors need to be taken into consideration when successfully removing tattoo pigment including the modalities used, number and frequency of treatments, proper device technique, and physiological barriers to tattoo removal. This paper serves to elucidate the common causes of ink retention following tattoo removal treatment with recommendations on how best to address this relatively common occurrence. PMID:23882312
Segregation Phenomena on the Crystal Surface of Chemical Compounds
NASA Astrophysics Data System (ADS)
Tomashpol'skii, Yu. Ya.
2018-06-01
The current state of the theoretical and experimental studies of changes in the chemical structure and composition caused by segregation phenomena on the surface of chemical compounds was reviewed. The review considers the experimental data obtained exclusively on single crystals, which were studied by modern instrumental methods, including in situ Auger electron spectrometry, X-ray spectral microanalysis, high-resolution scanning and transmission electron microscopy, secondary electron emission, and atomic force microscopy. The models that suggest the crystal-chemical diffusion and liquid-phase mechanisms of segregation were described. The parameters of the theory include the type of chemical bond, elastic constants, and crystal-chemical characteristics of substances. The models make it possible to predict the nature of changes in the surface composition: segregation tendency, segregant type, and degree of nonstoichiometry. A new direction in surface segregation was considered, which is promising for nanoelectronics and emission electronics.
Healthcare managers' decision making: findings of a small scale exploratory study.
Macdonald, Jackie; Bath, Peter A; Booth, Andrew
2008-12-01
Managers who work in publicly funded healthcare organizations are an understudied group. Some of the influences on their decisions may be unique to healthcare. This study considers how to integrate research knowledge effectively into healthcare managers' decision making, and how to manage and integrate information that will include community data. This first phase in a two-phase mixed methods research study used a qualitative, multiple case studies design. Nineteen semi-structured interviews were undertaken using the critical incident technique. Interview transcripts were analysed using the NatCen Framework. One theme represented ;information and decisions'. Cases were determined to involve complex multi-level, multi-situational decisions with participants in practical rather than ceremonial work roles. Most considered organizational knowledge in the first two decision phases and external knowledge, including research, in the third phase. All participants engaged in satisficing to some degree.
On The Calculation Of Derivatives From Digital Information
NASA Astrophysics Data System (ADS)
Pettett, Christopher G.; Budney, David R.
1982-02-01
Biomechanics analysis frequently requires cinematographic studies as a first step toward understanding the essential mechanics of a sport or exercise. In order to understand the exertion by the athlete, cinematography is used to establish the kinematics from which the energy exchanges can be considered and the equilibrium equations can be studied. Errors in the raw digital information necessitate smoothing of the data before derivatives can be obtained. Researchers employ a variety of curve-smoothing techniques including filtering and polynomial spline methods. It is essential that the researcher understands the accuracy which can be expected in velocities and accelerations obtained from smoothed digital information. This paper considers particular types of data inherent in athletic motion and the expected accuracy of calculated velocities and accelerations using typical error distributions in the raw digital information. Included in this paper are high acceleration, impact and smooth motion types of data.
Composite structures for magnetosphere imager spacecraft
NASA Technical Reports Server (NTRS)
Chu, Tsuchin
1994-01-01
Results of a trade study addressing the issues and benefits in using carbon fiber reinforced composites for the Magnetosphere Imager (MI) spacecraft are presented. The MI mission is now part of the Sun/Earth Connection Program. To qualify for this category, new technology and innovative methods to reduce the cost and size have to be considered. Topics addressed cover: (1) what is a composite, including advantages and disadvantages of composites and carbon/graphite fibers; and (2) structural design for MI, including composite design configuration, material selection, and analysis of composite structures.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Not Available
1987-12-01
This bibliography contains citations concerning the toxicity of food additives (excluding antioxidants) and their effects on the liver, kidneys, bladder, and other organs. The carcinogenic and teratogenic properties of these substances are also considered. The synthetic sweeteners, particularly the saccharins and cyclamates, and other additives, including nitrates and nitrites are discussed. Methods to detect and quantitate these additives are also included. (This updated bibliography contains 340 citations, 132 of which are new entries to the previous edition.)
Second-Order Moller-Plesset Perturbation Theory for Molecular Dirac-Hartree-Fock Wave Functions
NASA Technical Reports Server (NTRS)
Dyall, Kenneth G.; Arnold, James O. (Technical Monitor)
1994-01-01
Moller-Plesset perturbation theory is developed to second order for a selection of Kramers restricted Dirac-Hartree-Fock closed and open-shell reference wave functions. The open-shell wave functions considered are limited to those with no more than two electrons in open shells, but include the case of a two-configuration SCF reference. Denominator shifts are included in the style of Davidson's OPT2 method. An implementation which uses unordered integrals with labels is presented, and results are given for a few test cases.
Multifunctional biocompatible coatings on magnetic nanoparticles
NASA Astrophysics Data System (ADS)
Bychkova, A. V.; Sorokina, O. N.; Rosenfeld, M. A.; Kovarski, A. L.
2012-11-01
Methods for coating formation on magnetic nanoparticles used in biology and medicine are considered. Key requirements to the coatings are formulated, namely, biocompatibility, stability, the possibility of attachment of pharmaceutical agents, and the absence of toxicity. The behaviour of nanoparticle/coating nanosystems in the body including penetration through cellular membranes and the excretion rates and routes is analyzed. Parameters characterizing the magnetic properties of these systems and their magnetic controllability are described. Factors limiting the applications of magnetically controlled nanosystems for targeted drug delivery are discussed. The bibliography includes 405 references.
Group decision-making approach for flood vulnerability identification using the fuzzy VIKOR method
NASA Astrophysics Data System (ADS)
Lee, G.; Jun, K. S.; Cung, E. S.
2014-09-01
This study proposes an improved group decision making (GDM) framework that combines VIKOR method with fuzzified data to quantify the spatial flood vulnerability including multi-criteria evaluation indicators. In general, GDM method is an effective tool for formulating a compromise solution that involves various decision makers since various stakeholders may have different perspectives on their flood risk/vulnerability management responses. The GDM approach is designed to achieve consensus building that reflects the viewpoints of each participant. The fuzzy VIKOR method was developed to solve multi-criteria decision making (MCDM) problems with conflicting and noncommensurable criteria. This comprising method can be used to obtain a nearly ideal solution according to all established criteria. Triangular fuzzy numbers are used to consider the uncertainty of weights and the crisp data of proxy variables. This approach can effectively propose some compromising decisions by combining the GDM method and fuzzy VIKOR method. The spatial flood vulnerability of the south Han River using the GDM approach combined with the fuzzy VIKOR method was compared with the results from general MCDM methods, such as the fuzzy TOPSIS and classical GDM methods, such as those developed by Borda, Condorcet, and Copeland. The evaluated priorities were significantly dependent on the employed decision-making method. The proposed fuzzy GDM approach can reduce the uncertainty in the data confidence and weight derivation techniques. Thus, the combination of the GDM approach with the fuzzy VIKOR method can provide robust prioritization because it actively reflects the opinions of various groups and considers uncertainty in the input data.
2014-01-01
Background Review of theory is an area of growing methodological advancement. Theoretical reviews are particularly useful where the literature is complex, multi-discipline, or contested. It has been suggested that adopting methods from systematic reviews may help address these challenges. However, the methodological approaches to reviews of theory, including the degree to which systematic review methods can be incorporated, have received little discussion in the literature. We recently employed systematic review methods in a review of theories about the causal relationship between income and health. Methods This article discusses some of the methodological issues we considered in developing the review and offers lessons learnt from our experiences. It examines the stages of a systematic review in relation to how they could be adapted for a review of theory. The issues arising and the approaches taken in the review of theories in income and health are considered, drawing on the approaches of other reviews of theory. Results Different approaches to searching were required, including electronic and manual searches, and electronic citation tracking to follow the development of theories. Determining inclusion criteria was an iterative process to ensure that inclusion criteria were specific enough to make the review practical and focused, but not so narrow that key literature was excluded. Involving subject specialists was valuable in the literature searches to ensure principal papers were identified and during the inductive approaches used in synthesis of theories to provide detailed understanding of how theories related to another. Reviews of theory are likely to involve iterations and inductive processes throughout, and some of the concepts and techniques that have been developed for qualitative evidence synthesis can be usefully translated to theoretical reviews of this kind. Conclusions It may be useful at the outset of a review of theory to consider whether the key aim of the review is to scope out theories relating to a particular issue; to conduct in-depth analysis of key theoretical works with the aim of developing new, overarching theories and interpretations; or to combine both these processes in the review. This can help decide the most appropriate methodological approach to take at particular stages of the review. PMID:25312937
Eom, Youngsub; Ryu, Dongok; Kim, Dae Wook; Yang, Seul Ki; Song, Jong Suk; Kim, Sug-Whan; Kim, Hyo Myung
2016-10-01
To evaluate the toric intraocular lens (IOL) calculation considering posterior corneal astigmatism, incision-induced posterior corneal astigmatism, and effective lens position (ELP). Two thousand samples of corneal parameters with keratometric astigmatism ≥ 1.0 D were obtained using bootstrap methods. The probability distributions for incision-induced keratometric and posterior corneal astigmatisms, as well as ELP were estimated from the literature review. The predicted residual astigmatism error using method D with an IOL add power calculator (IAPC) was compared with those derived using methods A, B, and C through Monte-Carlo simulation. Method A considered the keratometric astigmatism and incision-induced keratometric astigmatism, method B considered posterior corneal astigmatism in addition to the A method, method C considered incision-induced posterior corneal astigmatism in addition to the B method, and method D considered ELP in addition to the C method. To verify the IAPC used in this study, the predicted toric IOL cylinder power and its axis using the IAPC were compared with ray-tracing simulation results. The median magnitude of the predicted residual astigmatism error using method D (0.25 diopters [D]) was smaller than that derived using methods A (0.42 D), B (0.38 D), and C (0.28 D) respectively. Linear regression analysis indicated that the predicted toric IOL cylinder power and its axis had excellent goodness-of-fit between the IAPC and ray-tracing simulation. The IAPC is a simple but accurate method for predicting the toric IOL cylinder power and its axis considering posterior corneal astigmatism, incision-induced posterior corneal astigmatism, and ELP.
NASA Astrophysics Data System (ADS)
Matrosov, E.; Padula, S.; Huskova, I.; Harou, J. J.
2012-12-01
Population growth and the threat of drier or changed climates are likely to increase water scarcity world-wide. A combination of demand management (water conservation) and new supply infrastructure is often needed to meet future projected demands. In this case system planners must decide what to implement, when and at what capacity. Choices can range from infrastructure to policies or a mix of the two, culminating in a complex planning problem. Decision making under uncertainty frameworks can be used to help planners with this planning problem. This presentation introduces, applies and compares four decision making under uncertainty frameworks. The application is to the Thames basin water resource system which includes the city of London. The approaches covered here include least-economic cost capacity expansion optimization (EO), Robust Decision Making (RDM), Info-Gap Decision Theory (Info-gap) and many-objective evolutionary optimization (MOEO). EO searches for the least-economic cost program, i.e. the timing, sizing, and choice of supply-demand management actions/upgrades which meet projected water demands. Instead of striving for optimality, the RDM and Info-gap approaches help build plans that are robust to 'deep' uncertainty in future conditions. The MOEO framework considers multiple performance criteria and uses water systems simulators as a function evaluator for the evolutionary algorithm. Visualizations show Pareto approximate tradeoffs between multiple objectives. In this presentation we detail the application of each framework to the Thames basin (including London) water resource planning problem. Supply and demand options are proposed by the major water companies in the basin. We apply the EO method using a 29 year time horizon and an annual time step considering capital, operating (fixed and variable), social and environmental costs. The method considers all plausible combinations of supply and conservation schemes and capacities proposed by water companies and generates the least-economic cost annual plan. The RDM application uses stochastic simulation under a weekly time-step and regret analysis to choose a candidate strategy. We then use a statistical cluster algorithm to identify future states of the world under which the strategy is vulnerable. The method explicitly considers the effects of uncertainty in supply, demands and energy price on multiple performance criteria. The Info-gap approach produces robustness and opportuneness plots that show the performance of different plans under the most dire and favorable sets of future conditions. The same simulator, supply and demand options and uncertainties are considered as in the RDM application. The MOEO application considers many more combinations of supply and demand options while still employing a simulator that enables a more realistic representation of the physical system and operating rules. A computer cluster is employed to ease the computational burden. Visualization software allows decision makers to interactively view tradeoffs in many dimensions. Benefits and limitations of each framework are discussed and recommendations for future planning in the basin are provided.
Acceleration methods for multi-physics compressible flow
NASA Astrophysics Data System (ADS)
Peles, Oren; Turkel, Eli
2018-04-01
In this work we investigate the Runge-Kutta (RK)/Implicit smoother scheme as a convergence accelerator for complex multi-physics flow problems including turbulent, reactive and also two-phase flows. The flows considered are subsonic, transonic and supersonic flows in complex geometries, and also can be either steady or unsteady flows. All of these problems are considered to be a very stiff. We then introduce an acceleration method for the compressible Navier-Stokes equations. We start with the multigrid method for pure subsonic flow, including reactive flows. We then add the Rossow-Swanson-Turkel RK/Implicit smoother that enables performing all these complex flow simulations with a reasonable CFL number. We next discuss the RK/Implicit smoother for time dependent problem and also for low Mach numbers. The preconditioner includes an intrinsic low Mach number treatment inside the smoother operator. We also develop a modified Roe scheme with a corresponding flux Jacobian matrix. We then give the extension of the method for real gas and reactive flow. Reactive flows are governed by a system of inhomogeneous Navier-Stokes equations with very stiff source terms. The extension of the RK/Implicit smoother requires an approximation of the source term Jacobian. The properties of the Jacobian are very important for the stability of the method. We discuss what the chemical physics theory of chemical kinetics tells about the mathematical properties of the Jacobian matrix. We focus on the implication of the Le-Chatelier's principle on the sign of the diagonal entries of the Jacobian. We present the implementation of the method for turbulent flow. We use a two RANS turbulent model - one equation model - Spalart-Allmaras and a two-equation model - k-ω SST model. The last extension is for two-phase flows with a gas as a main phase and Eulerian representation of a dispersed particles phase (EDP). We present some examples for such flow computations inside a ballistic evaluation rocket motor. The numerical examples in this work include transonic flow about a RAE2822 airfoil, about a M6 Onera wing, NACA0012 airfoil at very low Mach number, two-phase flow inside a Ballistic evaluation motor (BEM), a turbulent reactive shear layer and a time dependent Sod's tube problem.
Quality control for federal clean water act and safe drinking water act regulatory compliance.
Askew, Ed
2013-01-01
QC sample results are required in order to have confidence in the results from analytical tests. Some of the AOAC water methods include specific QC procedures, frequencies, and acceptance criteria. These are considered to be the minimum controls needed to perform the method successfully. Some regulatory programs, such as those in 40 CFR Part 136.7, require additional QC or have alternative acceptance limits. Essential QC measures include method calibration, reagent standardization, assessment of each analyst's capabilities, analysis of blind check samples, determination of the method's sensitivity (method detection level or quantification limit), and daily evaluation of bias, precision, and the presence of laboratory contamination or other analytical interference. The details of these procedures, their performance frequency, and expected ranges of results are set out in this manuscript. The specific regulatory requirements of 40 CFR Part 136.7 for the Clean Water Act, the laboratory certification requirements of 40 CFR Part 141 for the Safe Drinking Water Act, and the ISO 17025 accreditation requirements under The NELAC Institute are listed.
Crystallization mosaic effect generation by superpixels
NASA Astrophysics Data System (ADS)
Xie, Yuqi; Bo, Pengbo; Yuan, Ye; Wang, Kuanquan
2015-03-01
Art effect generation from digital images using computational tools has been a hot research topic in recent years. We propose a new method for generating crystallization mosaic effects from color images. Two key problems in generating pleasant mosaic effect are studied: grouping pixels into mosaic tiles and arrangement of mosaic tiles adapting to image features. To give visually pleasant mosaic effect, we propose to create mosaic tiles by pixel clustering in feature space of color information, taking compactness of tiles into consideration as well. Moreover, we propose a method for processing feature boundaries in images which gives guidance for arranging mosaic tiles near image features. This method gives nearly uniform shape of mosaic tiles, adapting to feature lines in an esthetic way. The new approach considers both color distance and Euclidean distance of pixels, and thus is capable of giving mosaic tiles in a more pleasing manner. Some experiments are included to demonstrate the computational efficiency of the present method and its capability of generating visually pleasant mosaic tiles. Comparisons with existing approaches are also included to show the superiority of the new method.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Miller, D.; Wohlgemuth, J.; Gu, X.
2013-11-01
The curing of cross-linkable encapsulation is a critical consideration for photovoltaic (PV) modules manufactured using a lamination process. Concerns related to ethylene-co-vinyl acetate (EVA) include the quality (e.g., expiration and uniformity) of the films or completion (duration) of the cross-linking of the EVA within a laminator. Because these issues are important to both EVA and module manufacturers, an international standard has recently been proposed by the Encapsulation Task-Group within the Working Group 2 (WG2) of the International Electrotechnical Commission (IEC) Technical Committee 82 (TC82) for the quantification of the degree of cure for EVA encapsulation. The present draft of themore » standard calls for the use of differential scanning calorimetry (DSC) as the rapid, enabling secondary (test) method. Both the residual enthalpy- and melt/freeze-DSC methods are identified. The DSC methods are calibrated against the gel content test, the primary (reference) method. Aspects of other established methods, including indentation and rotor cure metering, were considered by the group. Key details of the test procedure will be described.« less
An ab initio study of the C3(+) cation using multireference methods
NASA Technical Reports Server (NTRS)
Taylor, Peter R.; Martin, J. M. L.; Francois, J. P.; Gijbels, R.
1991-01-01
The energy difference between the linear 2 sigma(sup +, sub u) and cyclic 2B(sub 2) structures of C3(+) has been investigated using large (5s3p2d1f) basis sets and multireference electron correlation treatments, including complete active space self consistent fields (CASSCF), multireference configuration interaction (MRCI), and averaged coupled-pair functional (ACPF) methods, as well as the single-reference quadratic configuration interaction (QCISD(T)) method. Our best estimate, including a correction for basis set incompleteness, is that the linear form lies above the cyclic from by 5.2(+1.5 to -1.0) kcal/mol. The 2 sigma(sup +, sub u) state is probably not a transition state, but a local minimum. Reliable computation of the cyclic/linear energy difference in C3(+) is extremely demanding of the electron correlation treatment used: of the single-reference methods previously considered, CCSD(T) and QCISD(T) perform best. The MRCI + Q(0.01)/(4s2p1d) energy separation of 1.68 kcal/mol should provide a comparison standard for other electron correlation methods applied to this system.
Sampling Methods in Cardiovascular Nursing Research: An Overview.
Kandola, Damanpreet; Banner, Davina; O'Keefe-McCarthy, Sheila; Jassal, Debbie
2014-01-01
Cardiovascular nursing research covers a wide array of topics from health services to psychosocial patient experiences. The selection of specific participant samples is an important part of the research design and process. The sampling strategy employed is of utmost importance to ensure that a representative sample of participants is chosen. There are two main categories of sampling methods: probability and non-probability. Probability sampling is the random selection of elements from the population, where each element of the population has an equal and independent chance of being included in the sample. There are five main types of probability sampling including simple random sampling, systematic sampling, stratified sampling, cluster sampling, and multi-stage sampling. Non-probability sampling methods are those in which elements are chosen through non-random methods for inclusion into the research study and include convenience sampling, purposive sampling, and snowball sampling. Each approach offers distinct advantages and disadvantages and must be considered critically. In this research column, we provide an introduction to these key sampling techniques and draw on examples from the cardiovascular research. Understanding the differences in sampling techniques may aid nurses in effective appraisal of research literature and provide a reference pointfor nurses who engage in cardiovascular research.
Education in the responsible conduct of research in psychology: methods and scope.
DiLorenzo, Terry A; Becker-Fiegeles, Jill; Gibelman, Margaret
2014-01-01
In this mixed-method study of education in the responsible conduct of research (RCR) in psychology, phase one survey respondents (n = 141) reported that faculty and students were familiar with RCR standards and procedures to educate them were believed to be adequate. However, educational methods varied widely. In phase two, seven survey respondents completed in-depth interviews assessing RCR training and education and research review procedures. Educational methods through which RCR content was presented included the following ones: traditional (lectures), technical (web-based), and experiential (internships), but RCR was often minimally considered in the formal curriculum. Our results suggest that psychology training programs might benefit from more formal consideration of RCR education and training in the curriculum.
Cultural psychiatry: research strategies and future directions.
Kirmayer, Laurence J; Ban, Lauren
2013-01-01
This chapter reviews some key aspects of current research in cultural psychiatry and explores future prospects. The first section discusses the multiple meanings of culture in the contemporary world and their relevance for understanding mental health and illness. The next section considers methodological strategies for unpacking the concept of culture and studying the impact of cultural variables, processes and contexts. Multiple methods are needed to address the many different components or dimensions of cultural identity and experience that constitute local worlds, ways of life or systems of knowledge. Quantitative and observational methods of clinical epidemiology and experimental science as well as qualitative ethnographic methods are needed to capture crucial aspects of culture as systems of meaning and practice. Emerging issues in cultural psychiatric research include: cultural variations in illness experience and expression; the situated nature of cognition and emotion; cultural configurations of self and personhood; concepts of mental disorder and mental health literacy; and the prospect of ecosocial models of health and culturally based interventions. The conclusion considers the implications of the emerging perspectives from cultural neuroscience for psychiatric theory and practice. Copyright © 2013 S. Karger AG, Basel.
NASA Technical Reports Server (NTRS)
Jaeckel, Louis A.
1989-01-01
To study the problems of encoding visual images for use with a Sparse Distributed Memory (SDM), I consider a specific class of images- those that consist of several pieces, each of which is a line segment or an arc of a circle. This class includes line drawings of characters such as letters of the alphabet. I give a method of representing a segment of an arc by five numbers in a continuous way; that is, similar arcs have similar representations. I also give methods for encoding these numbers as bit strings in an approximately continuous way. The set of possible segments and arcs may be viewed as a five-dimensional manifold M, whose structure is like a Mobious strip. An image, considered to be an unordered set of segments and arcs, is therefore represented by a set of points in M - one for each piece. I then discuss the problem of constructing a preprocessor to find the segments and arcs in these images, although a preprocessor has not been developed. I also describe a possible extension of the representation.
Cui, Shan-shan; Zou, Yan-hong; Li, Qian; Li, Li-na; Zhang, Ning; Liu, Xi-pu
2014-12-01
To assess the agreement between gonioscopy and ultrasound biomicroscopy (UBM) in detecting angle closure in Chinese patients with shallow anterior chamber. An observational comparative study of the two different examination methods was conducted. Patients with normal intraocular pressure and temporal peripheral anterior chamber depth less than a quarter of corneal thickness based on slit lamp examination were included in this study from December 2007 to May 2009 in the outpatient clinic of First Hospital of Tsinghua University. Gonioscopy was performed with a Goldman goniolens in dark room first and followed by full beam light and indentation. If the filtering trabecular meshwork was invisible or any peripheral anterior synechia was found, that quadrant of the angle was considered closed. UBM was first undertaken in a darkened room then repeated with normal room lighting. If iridotrabecular apposition was showed, that quadrant of the angle was considered closed. The status of angle closure of each quadrant with different methods was recorded. 85 eyes of 46 patients were included in this study. The agreement between gonioscopy and UBM was poor (Κ<0.4) with Kappa analysis in both dark and light conditions in each quadrant. The accordance of agreement between gonioscopy and UBM was hardly affected by age or sex, while in dark condition, eyes with deeper anterior chamber (P=0.005) or plateau iris configuration tended to produce different results (P=0.075) in the 2 methods. Gonioscopy and UBM are both indispensable methods for detecting angle closure, neither can completely replace the other.
New developments in the theoretical treatment of low dimensional strongly correlated systems.
James, Andrew J A; Konik, Robert M; Lecheminant, Philippe; Robinson, Neil; Tsvelik, Alexei M
2017-10-09
We review two important non-perturbative approaches for extracting the physics of low- dimensional strongly correlated quantum systems. Firstly, we start by providing a comprehensive review of non-Abelian bosonization. This includes an introduction to the basic elements of confor- mal field theory as applied to systems with a current algebra, and we orient the reader by presenting a number of applications of non-Abelian bosonization to models with large symmetries. We then tie this technique into recent advances in the ability of cold atomic systems to realize complex symme- tries. Secondly, we discuss truncated spectrum methods for the numerical study of systems in one and two dimensions. For one-dimensional systems we provide the reader with considerable insight into the methodology by reviewing canonical applications of the technique to the Ising model (and its variants) and the sine-Gordon model. Following this we review recent work on the development of renormalization groups, both numerical and analytical, that alleviate the effects of truncating the spectrum. Using these technologies, we consider a number of applications to one-dimensional systems: properties of carbon nanotubes, quenches in the Lieb-Liniger model, 1+1D quantum chro- modynamics, as well as Landau-Ginzburg theories. In the final part we move our attention to consider truncated spectrum methods applied to two-dimensional systems. This involves combining truncated spectrum methods with matrix product state algorithms. We describe applications of this method to two-dimensional systems of free fermions and the quantum Ising model, including their non-equilibrium dynamics. © 2017 IOP Publishing Ltd.
NASA Astrophysics Data System (ADS)
Dey, Seonaid R. A.; Moore, Robert J.; Cole, Steven J.; Wells, Steven C.
2017-04-01
In many regions of high annual snowfall, snowmelt modelling can prove to be a vital component of operational flood forecasting and warning systems. Although Britain as a whole does not experience prolonged periods of lying snow, with the exception of the Scottish Highlands, the inclusion of snowmelt modelling can still have a significant impact on the skill of flood forecasts. Countrywide operational flood forecasts over Britain are produced using the national Grid-to-Grid (G2G) distributed hydrological model. For Scotland, snowmelt is included in these forecasts through a G2G snow hydrology module involving temperature-based snowfall/rainfall partitioning and functions for temperature-excess snowmelt, snowpack storage and drainage. Over England and Wales, the contribution of snowmelt is included by pre-processing the precipitation prior to input into G2G. This removes snowfall diagnosed from weather model outputs and adds snowmelt from an energy budget land surface scheme to form an effective liquid water gridded input to G2G. To review the operational options for including snowmelt modelling in G2G over Britain, a project was commissioned by the Environment Agency through the Flood Forecasting Centre (FFC) for England and Wales and in partnership with the Scottish Environment Protection Agency (SEPA) and Natural Resources Wales (NRW). Results obtained from this snowmelt review project will be reported on here. The operational methods used by the FFC and SEPA are compared on past snowmelt floods, alongside new alternative methods of treating snowmelt. Both case study and longer-term analyses are considered, covering periods selected from the winters 2009-2010, 2012-2013, 2013-2014 and 2014-2015. Over Scotland, both of the snowmelt methods used operationally by FFC and SEPA provided a clear improvement to the river flow simulations. Over England and Wales, fewer and less significant snowfall events occurred, leading to less distinction in the results between the methods. It is noted that, for all methods considered, large uncertainties remain in flood forecasts influenced by snowmelt. Understanding and quantifying these uncertainties should lead to more informed flood forecasts and associated guidance information.
Campbell, Mhairi; Egan, Matt; Lorenc, Theo; Bond, Lyndal; Popham, Frank; Fenton, Candida; Benzeval, Michaela
2014-10-13
Review of theory is an area of growing methodological advancement. Theoretical reviews are particularly useful where the literature is complex, multi-discipline, or contested. It has been suggested that adopting methods from systematic reviews may help address these challenges. However, the methodological approaches to reviews of theory, including the degree to which systematic review methods can be incorporated, have received little discussion in the literature. We recently employed systematic review methods in a review of theories about the causal relationship between income and health. This article discusses some of the methodological issues we considered in developing the review and offers lessons learnt from our experiences. It examines the stages of a systematic review in relation to how they could be adapted for a review of theory. The issues arising and the approaches taken in the review of theories in income and health are considered, drawing on the approaches of other reviews of theory. Different approaches to searching were required, including electronic and manual searches, and electronic citation tracking to follow the development of theories. Determining inclusion criteria was an iterative process to ensure that inclusion criteria were specific enough to make the review practical and focused, but not so narrow that key literature was excluded. Involving subject specialists was valuable in the literature searches to ensure principal papers were identified and during the inductive approaches used in synthesis of theories to provide detailed understanding of how theories related to another. Reviews of theory are likely to involve iterations and inductive processes throughout, and some of the concepts and techniques that have been developed for qualitative evidence synthesis can be usefully translated to theoretical reviews of this kind. It may be useful at the outset of a review of theory to consider whether the key aim of the review is to scope out theories relating to a particular issue; to conduct in-depth analysis of key theoretical works with the aim of developing new, overarching theories and interpretations; or to combine both these processes in the review. This can help decide the most appropriate methodological approach to take at particular stages of the review.
Lee, Eun Gyung; Kim, Seung Won; Feigley, Charles E.; Harper, Martin
2015-01-01
This study introduces two semi-quantitative methods, Structured Subjective Assessment (SSA) and Control of Substances Hazardous to Health (COSHH) Essentials, in conjunction with two-dimensional Monte Carlo simulations for determining prior probabilities. Prior distribution using expert judgment was included for comparison. Practical applications of the proposed methods were demonstrated using personal exposure measurements of isoamyl acetate in an electronics manufacturing facility and of isopropanol in a printing shop. Applicability of these methods in real workplaces was discussed based on the advantages and disadvantages of each method. Although these methods could not be completely independent of expert judgments, this study demonstrated a methodological improvement in the estimation of the prior distribution for the Bayesian decision analysis tool. The proposed methods provide a logical basis for the decision process by considering determinants of worker exposure. PMID:23252451
Leiserson, Mark D.M.; Tatar, Diana; Cowen, Lenore J.
2011-01-01
Abstract A new method based on a mathematically natural local search framework for max cut is developed to uncover functionally coherent module and BPM motifs in high-throughput genetic interaction data. Unlike previous methods, which also consider physical protein-protein interaction data, our method utilizes genetic interaction data only; this becomes increasingly important as high-throughput genetic interaction data is becoming available in settings where less is known about physical interaction data. We compare modules and BPMs obtained to previous methods and across different datasets. Despite needing no physical interaction information, the BPMs produced by our method are competitive with previous methods. Biological findings include a suggested global role for the prefoldin complex and a SWR subcomplex in pathway buffering in the budding yeast interactome. PMID:21882903
Pistollato, Francesca; Ohayon, Elan L; Lam, Ann; Langley, Gillian R; Novak, Thomas J; Pamies, David; Perry, George; Trushina, Eugenia; Williams, Robin S B; Roher, Alex E; Hartung, Thomas; Harnad, Stevan; Barnard, Neal; Morris, Martha Clare; Lai, Mei-Chun; Merkley, Ryan; Chandrasekera, P Charukeshi
2016-06-28
Much of Alzheimer disease (AD) research has been traditionally based on the use of animals, which have been extensively applied in an effort to both improve our understanding of the pathophysiological mechanisms of the disease and to test novel therapeutic approaches. However, decades of such research have not effectively translated into substantial therapeutic success for human patients. Here we critically discuss these issues in order to determine how existing human-based methods can be applied to study AD pathology and develop novel therapeutics. These methods, which include patient-derived cells, computational analysis and models, together with large-scale epidemiological studies represent novel and exciting tools to enhance and forward AD research. In particular, these methods are helping advance AD research by contributing multifactorial and multidimensional perspectives, especially considering the crucial role played by lifestyle risk factors in the determination of AD risk. In addition to research techniques, we also consider related pitfalls and flaws in the current research funding system. Conversely, we identify encouraging new trends in research and government policy. In light of these new research directions, we provide recommendations regarding prioritization of research funding. The goal of this document is to stimulate scientific and public discussion on the need to explore new avenues in AD research, considering outcome and ethics as core principles to reliably judge traditional research efforts and eventually undertake new research strategies.
Pistollato, Francesca; Ohayon, Elan L.; Lam, Ann; Langley, Gillian R.; Novak, Thomas J.; Pamies, David; Perry, George; Trushina, Eugenia; Williams, Robin S.B.; Roher, Alex E.; Hartung, Thomas; Harnad, Stevan; Barnard, Neal; Morris, Martha Clare; Lai, Mei-Chun; Merkley, Ryan; Chandrasekera, P. Charukeshi
2016-01-01
Much of Alzheimer disease (AD) research has been traditionally based on the use of animals, which have been extensively applied in an effort to both improve our understanding of the pathophysiological mechanisms of the disease and to test novel therapeutic approaches. However, decades of such research have not effectively translated into substantial therapeutic success for human patients. Here we critically discuss these issues in order to determine how existing human-based methods can be applied to study AD pathology and develop novel therapeutics. These methods, which include patient-derived cells, computational analysis and models, together with large-scale epidemiological studies represent novel and exciting tools to enhance and forward AD research. In particular, these methods are helping advance AD research by contributing multifactorial and multidimensional perspectives, especially considering the crucial role played by lifestyle risk factors in the determination of AD risk. In addition to research techniques, we also consider related pitfalls and flaws in the current research funding system. Conversely, we identify encouraging new trends in research and government policy. In light of these new research directions, we provide recommendations regarding prioritization of research funding. The goal of this document is to stimulate scientific and public discussion on the need to explore new avenues in AD research, considering outcome and ethics as core principles to reliably judge traditional research efforts and eventually undertake new research strategies. PMID:27229915
Klugarova, Jitka; Klugar, Miloslav; Mareckova, Jana; Gallo, Jiri; Kelnarova, Zuzana
2016-01-01
Total hip replacement is the most effective and safest method for treating severe degenerative, traumatic and other diseases of the hip joint. Total hip replacement can reliably relieve pain and improve function in the majority of patients for a period of 15 to 20 years or more postoperatively. Physical therapy follows each total hip replacement surgery. Physical therapy protocols after total hip replacement in the post-discharge period vary widely in terms of setting (inpatient, outpatient), content (the particular set of exercises used), and frequency (e.g. daily versus twice a week). In current literature, there is no systematic review which has compared the effectiveness of inpatient and outpatient physical therapy in patients after total hip replacement in the post-discharge period. The objective of this systematic review was to compare the effectiveness of inpatient physical therapy with outpatient physical therapy on the quality of life and gait measures in older adults after total hip replacement in the post-discharge period. This review considered studies that include older adults (over 65 years) who have had total hip replacement and are in the post-discharge period. Adults with bilateral or multiple simultaneous surgeries and also patients who have had hemiarthroplasty of the hip joint were excluded.This review considered studies that included any type of physical therapy delivered in inpatient settings provided by professionals with education in physical therapy. Inpatient physical therapy delivered at any frequency and over any duration was included.This review considered studies that included as a comparator any type of physical therapy delivered in outpatient settings provided by professionals with education in physical therapy or no physical therapy.This review considered studies that included the following primary and secondary outcomes. The primary outcome was quality of life, assessed by any validated assessment tool. The secondary outcome was measures of gait assessed by any valid methods.This review considered both experimental and observational study designs including randomized controlled trials, non-randomized controlled trials, quasi-experimental, before and after studies, prospective and retrospective cohort studies, case control studies and analytical cross sectional studies for inclusion. The search strategy aimed to find both published and unpublished studies. A three-step search strategy was utilized in 12 databases. Studies published in all languages and any date were considered for inclusion in this review. Assessment of methodological quality was not conducted as no studies were identified that met the inclusion criteria. Data extraction and synthesis was not performed because no studies were included in this systematic review. During to the three-step search strategy 4330 papers were identified. The primary and secondary reviewer independently retrieved 42 potentially relevant papers according to the inclusion criteria by title and abstract screening. Following assessment of full text all of the retrieved papers were excluded based on the inclusion criteria. There is no scientific evidence comparing the effectiveness of inpatient physical therapy with outpatient physical therapy in older patients after total hip replacement in the post-discharge period. This systematic review has identified gaps in the literature for comparing the effectiveness of inpatient physical therapy with and outpatient physical therapy on the quality of life and gait measures in older adults after total hip replacement in the post-discharge period. Prospective randomized double blind multicenter controlled trials are needed to answer this important clinical question.
Density Deconvolution With EPI Splines
2015-09-01
effects of various substances on test subjects [11], [12]. Whereas in geophysics, a shot may be fired into the ground, in pharmacokinetics, a signal is...be significant, including medicine, bioinformatics, chemistry, as- tronomy, and econometrics , as well as an extensive review of kernel based methods...demonstrate the effectiveness of our model in simulations motivated by test instances in [32]. We consider an additive measurement model scenario where
Daryamide Analogues from a Marine-Derived Streptomyces species.
Fu, Peng; La, Scott; MacMillan, John B
2017-04-28
Three new cyclohexene amine derivatives, daryamides D-F (1-3), a new arylamine derivative, carpatamide D (4), and a new ornithine lactamization derivative, ornilactam A (5), were isolated from the marine-derived Streptomyces strain SNE-011. Their structures, including absolute configurations, were elucidated on the basis of spectroscopic analysis and chemical methods. The carpatamide skeleton could be considered as the biosynthetic precursor of the daryamides.
ERIC Educational Resources Information Center
Cartwright, T. J.; Hallar, B.
2018-01-01
In this study, we present the long-term influence of an after school science practicum associated with an elementary science methods course. The practicum or field experience could be considered a community-based service learning programme as it is situated both within and for the community. Study participants included eight third- and fifth-grade…
Economic evaluation: Concepts, selected studies, system costs, and a proposed program
NASA Technical Reports Server (NTRS)
Osterhoudt, F. H. (Principal Investigator)
1979-01-01
The more usual approaches to valuing crop information are reviewed and an integrated approach is recommended. Problems associated with implementation are examined. What has already been accomplished in the economic evaluation of LACIE-type information is reported including various studies of benefits. The costs of the existing and proposed systems are considered. A method and approach is proposed for further studies.
ERIC Educational Resources Information Center
Lueder, E. J.
The rationale, design, and data collection methods of a study of factors that may influence the effectiveness of adult instructional groups that included interaction are discussed. Two aspects to be considered when studying instructional groups are called work and emotionality. The Work-Emotionality Theory is discussed. Six types of…
Preliminary candidate advanced avionics system for general aviation
NASA Technical Reports Server (NTRS)
Mccalla, T. M.; Grismore, F. L.; Greatline, S. E.; Birkhead, L. M.
1977-01-01
An integrated avionics system design was carried out to the level which indicates subsystem function, and the methods of overall system integration. Sufficient detail was included to allow identification of possible system component technologies, and to perform reliability, modularity, maintainability, cost, and risk analysis upon the system design. Retrofit to older aircraft, availability of this system to the single engine two place aircraft, was considered.
Green chemistry: development trajectory
NASA Astrophysics Data System (ADS)
Moiseev, I. I.
2013-07-01
Examples of applications of green chemistry methods in heavy organic synthesis are analyzed. Compounds, which can be produced by the processing of the biomass, and the criteria for the selection of the most promising products are summarized. The current status of the ethanol production and processing is considered. The possibilities of the use of high fatty acid triglycerides, glycerol, succinic acid, and isoprene are briefly discussed. The bibliography includes 67 references.
Statistics of rain-rate estimates for a single attenuating radar
NASA Technical Reports Server (NTRS)
Meneghini, R.
1976-01-01
The effects of fluctuations in return power and the rain-rate/reflectivity relationship, are included in the estimates, as well as errors introduced in the attempt to recover the unattenuated return power. In addition to the Hitschfeld-Bordan correction, two alternative techniques are considered. The performance of the radar is shown to be dependent on the method by which attenuation correction is made.
A Nonpharmacologic Method for Enhancing Sleep in PTSD
2016-10-01
are taken solely under appropriate medical supervision, this criterion is not considered to be met. SEDATIVE/ HYPNOTIC/ANX CANNABIS STIMULANTS...OF “1” OR “3.” Etiological substances/ medications include: alcohol (I/W); caffeine (I/W); cannabis (I/W); opioids (I/W); sedatives, hypnotics, or...U.S. Army Medical Research and Materiel Command Fort Detrick, Maryland 21702-5012 DISTRIBUTION STATEMENT: Approved for Public Release
Quantum-chemical Calculations in the Study of Antitumour Compounds
NASA Astrophysics Data System (ADS)
Luzhkov, V. B.; Bogdanov, G. N.
1986-01-01
The results of quantum-chemical calculations on antitumour preparations concerning the mechanism of their action at the electronic and molecular levels and structure-activity correlations are discussed in this review. Preparations whose action involves alkylating and free-radial mechanisms, complex-forming agents, and antimetabolites are considered. Modern quantum-chemical methods for calculations on biologically active substances are described. The bibliography includes 106 references.
30 CFR 206.174 - How do I value gas production when an index-based method cannot be used?
Code of Federal Regulations, 2010 CFR
2010-07-01
... to consider include prices received in spot sales of gas, residue gas or gas plant products, other... part, or timely, for a quantity of gas, residue gas, or gas plant product. (j) Non-binding MMS reviews..., DEPARTMENT OF THE INTERIOR MINERALS REVENUE MANAGEMENT PRODUCT VALUATION Indian Gas § 206.174 How do I value...
Carbocyclic nucleoside analogues: classification, target enzymes, mechanisms of action and synthesis
NASA Astrophysics Data System (ADS)
Matyugina, E. S.; Khandazhinskaya, A. P.; Kochetkov, Sergei N.
2012-08-01
Key biological targets (S-adenosyl-L-homocysteine hydrolase, telomerase, human immunodeficiency virus reverse transcriptase, herpes virus DNA polymerase and hepatitis B virus DNA polymerase) and the mechanisms of action of carbocyclic nucleoside analogues are considered. Structural types of analogues are discussed. Methods of synthesis for the most promising compounds and the spectrum of their biological activities are described. The bibliography includes 126 references.
NASA Astrophysics Data System (ADS)
Sahoo, Madhumita; Sahoo, Satiprasad; Dhar, Anirban; Pradhan, Biswajeet
2016-10-01
Groundwater vulnerability assessment has been an accepted practice to identify the zones with relatively increased potential for groundwater contamination. DRASTIC is the most popular secondary information-based vulnerability assessment approach. Original DRASTIC approach considers relative importance of features/sub-features based on subjective weighting/rating values. However variability of features at a smaller scale is not reflected in this subjective vulnerability assessment process. In contrast to the subjective approach, the objective weighting-based methods provide flexibility in weight assignment depending on the variation of the local system. However experts' opinion is not directly considered in the objective weighting-based methods. Thus effectiveness of both subjective and objective weighting-based approaches needs to be evaluated. In the present study, three methods - Entropy information method (E-DRASTIC), Fuzzy pattern recognition method (F-DRASTIC) and Single parameter sensitivity analysis (SA-DRASTIC), were used to modify the weights of the original DRASTIC features to include local variability. Moreover, a grey incidence analysis was used to evaluate the relative performance of subjective (DRASTIC and SA-DRASTIC) and objective (E-DRASTIC and F-DRASTIC) weighting-based methods. The performance of the developed methodology was tested in an urban area of Kanpur City, India. Relative performance of the subjective and objective methods varies with the choice of water quality parameters. This methodology can be applied without/with suitable modification. These evaluations establish the potential applicability of the methodology for general vulnerability assessment in urban context.
Selection of a method to rate the strength of scientific evidence for AORN recommendations.
Steelman, Victoria M; Pape, Theresa; King, Cecil A; Graling, Paula; Gaberson, Kathleen B
2011-04-01
The use of scientific evidence to support national recommendations about clinical decisions has become an expectation of multidisciplinary health care organizations. The objectives of this project were to identify the most applicable evidence-rating method for perioperative nursing practice, evaluate the reliability of this method for perioperative nursing recommendations, and identify barriers and facilitators to adoption of this method for AORN recommendations. A panel of perioperative nurse experts evaluated 46 evidence-rating systems for quality, quantity, and consistency. We rated the methods that fully covered all three domains on five aspects of applicability to perioperative nursing practice recommendations. The Oncology Nursing Society's method was rated highest for all five aspects of applicability, and interrater reliability of this method for perioperative recommendations was 100%. Potential barriers to implementation of the rating method include knowledge deficit, staff resources, resistance to change, and fear of showing that lower levels of evidence support some recommendations. Facilitators included education, resource allocation, and starting small. Barriers and facilitators will be considered by the implementation team that will develop a plan to achieve integration of evidence rating into AORN documents. The AORN Board of Directors approved adoption of this method in June 2010. Copyright © 2011 AORN, Inc. Published by Elsevier Inc. All rights reserved.
NASA Technical Reports Server (NTRS)
Maskew, B.
1983-01-01
A general low-order surface-singularity panel method is used to predict the aerodynamic characteristics of a problem where a wing-tip vortex from one wing closely interacts with an aft mounted wing in a low Reynolds Number flow; i.e., 125,000. Nonlinear effects due to wake roll-up and the influence of the wings on the vortex path are included in the calculation by using a coupled iterative wake relaxation scheme. The interaction also affects the wing pressures and boundary layer characteristics: these effects are also considered using coupled integral boundary layer codes and preliminary calculations using free vortex sheet separation modelling are included. Calculated results are compared with water tunnel experimental data with generally remarkably good agreement.
Effects of Earth's curvature in full-wave modeling of VLF propagation
NASA Astrophysics Data System (ADS)
Qiu, L.; Lehtinen, N. G.; Inan, U. S.; Stanford VLF Group
2011-12-01
We show how to include curvature in the full-wave finite element approach to calculate ELF/VLF wave propagation in horizontally stratified earth-ionosphere waveguide. A general curvilinear stratified system is considered, and the numerical solutions of full-wave method in curvilinear system are compared with the analytic solutions in the cylindrical and spherical waveguides filled with an isotropic medium. We calculate the attenuation and height gain for modes in the Earth-ionosphere waveguide, taking into account the anisotropicity of ionospheric plasma, for different assumptions about the Earth's curvature, and quantify the corrections due to the curvature. The results are compared with the results of previous models, such as LWPC, as well as with ground and satellite observations, and show improved accuracy compared with full-wave method without including the curvature effect.
Marom, Gil; Bluestein, Danny
2016-01-01
Summary This paper evaluated the influence of various numerical implementation assumptions on predicting blood damage in cardiovascular devices using Lagrangian methods with Eulerian computational fluid dynamics. The implementation assumptions that were tested included various seeding patterns, stochastic walk model, and simplified trajectory calculations with pathlines. Post processing implementation options that were evaluated included single passage and repeated passages stress accumulation and time averaging. This study demonstrated that the implementation assumptions can significantly affect the resulting stress accumulation, i.e., the blood damage model predictions. Careful considerations should be taken in the use of Lagrangian models. Ultimately, the appropriate assumptions should be considered based the physics of the specific case and sensitivity analysis, similar to the ones presented here, should be employed. PMID:26679833
Medical Management of Oral Lichen Planus: A Systematic Review
Chokshi, Krunal; Desai, Sachin; Malu, Rahul; Chokshi, Achala
2016-01-01
Introduction Oral Lichen Planus (OLP) is a chronic inflammatory, T-cell-mediated autoimmune oral mucosal disease with unclear aetiology. The clinical management of OLP poses considerable difficulties to the oral physician. Aim The aim was to assess the efficacy of any form of intervention used to medically manage OLP. Materials and Methods We searched and analysed the following databases (from January 1990 to December 2014):- Cochrane Oral Health Group Trials Register, Cochrane Central Register of Controlled Trials (CENTRAL), MEDLINE and EMBASE. All Randomised Controlled Trials (RCTs) for the medical management of OLP which compared active treatment with placebo or between active treatments were considered in this systematic review. Participants of any age, gender or race having symptomatic OLP (including mixed forms), unconnected to any identifiable cause (e.g. lichenoid drug reactions) and confirmed by histopathology have been included. Interventions of all types, including topical treatments or systemic drugs of variable dosage, duration & frequency of delivery have been considered. All the trials identified were appraised by five review authors and the data for all the trials were synthesised using specifically designed data extraction form. Binary data has been presented as risk ratios (RR) with 95% confidence intervals (CI) and continuous data as mean differences (MD) with 95% CIs. Results A total of 35 RCTs were included in this systematic review on medical management of OLP. No strong evidence suggesting superiority of any specific intervention in reducing pain and clinical signs of OLP were shown by the RCTs included here. Conclusion Future RCTs on a larger scale, adopting standardized outcome assessing parameters should be considered. PMID:27042598
Garau, Martina; Shah, Koonal Kirit; Sharma, Priya; Towse, Adrian
2015-01-01
The aim of this study was to explore whether wealth effects of health interventions, including productivity gains and savings in other sectors, are considered in resource allocations by health technology assessment (HTA) agencies and government departments. To analyze reasons for including, or not including, wealth effects. Semi-structured interviews with decision makers and academic experts in eight countries (Australia, France, Germany, Italy, Poland, South Korea, Sweden, and the United Kingdom). There is evidence suggesting that health interventions can produce economic gains for patients and national economies. However, we found that the link between health and wealth does not influence decision making in any country with the exception of Sweden. This is due to a combination of factors, including system fragmentation, methodological issues, and the economic recession forcing national governments to focus on short-term measures. In countries with established HTA processes and methods allowing, in principle, the inclusion of wider effects in exceptional cases or secondary analyses, it might be possible to overcome the methodological and practical barriers and see a more systematic consideration of wealth effect in decision making. This would be consistent with principles of efficient priority setting. Barriers for the consideration of wealth effects in government decision making are more fundamental, due to an enduring separation of budgets within the public sector and current financial pressures. However, governments should consider all relevant effects from public investments, including healthcare, even when benefits can only be captured in the medium- and long-term. This will ensure that resources are allocated where they bring the best returns.
NASA Astrophysics Data System (ADS)
Sylus, K. J.; H., Ramesh
2018-04-01
In the coastal aquifer, seawater intrusion considered the major problem which contaminates freshwater and reduces its quality for domestic use. In order to find seawater intrusion, the groundwater quality analysis for the different chemical parameter was considered as the basic method to find out contamination. This analysis was carried out as per Bureau of Indian standards (2012) and World Health Organisations (1996). In this study, Bicarbonate parameter was considered for groundwater quality analysis which ranges the permissible limit in between 200-600 mg/l. The groundwater system was modelled using Groundwater modelling software (GMS) in which the FEMWATER package used for flow and transport. The FEMWATER package works in the principle of finite element method. The base input data of model include elevation, Groundwater head, First bottom and second bottom of the study area. The modelling results show the spatial occurrence of contamination in the study area of Netravathi and Gurpur river confluence at the various time period. Further, the results of the modelling also show that the contamination occurs up to a distance of 519m towards the freshwater zone of the study area.
Physical, Mental, and Financial Impacts From Drought in Two California Counties, 2015
Barreau, Tracy; Conway, David; Haught, Karen; Jackson, Rebecca; Kreutzer, Richard; Lockman, Andrew; Minnick, Sharon; Roisman, Rachel; Rozell, David; Smorodinsky, Svetlana; Tafoya, Dana
2017-01-01
Objectives. To evaluate health impacts of drought during the most severe drought in California’s recorded history with a rapid assessment method. Methods. We conducted Community Assessments for Public Health Emergency Response during October through November 2015 in Tulare County and Mariposa County to evaluate household water access, acute stressors, exacerbations of chronic diseases and behavioral health issues, and financial impacts. We evaluated pairwise associations by logistic regression with pooled data. Results. By assessment area, households reported not having running water (3%–12%); impacts on finances (25%–39%), property (39%–54%), health (10%–20%), and peace of mind (33%–61%); worsening of a chronic disease (16%–46%); acute stress (8%–26%); and considering moving (14%–34%). Impacts on finances or property were each associated with impacts on health and peace of mind, and acute stress. Conclusions. Drought-impacted households might perceive physical and mental health effects and might experience financial or property impacts related to the drought. Public Health Implications. Local jurisdictions should consider implementing drought assistance programs, including behavioral health, and consider rapid assessments to inform public health action. PMID:28323464
A New Model for Simulating Gas Metal Arc Welding based on Phase Field Model
NASA Astrophysics Data System (ADS)
Jiang, Yongyue; Li, Li; Zhao, Zhijiang
2017-11-01
Lots of physical process, such as metal melting, multiphase fluids flow, heat and mass transfer and thermocapillary effect (Marangoni) and so on, will occur in gas metal arc welding (GMAW) which should be considered as a mixture system. In this paper, based on the previous work, we propose a new model to simulate GMAW including Navier-Stokes equation, the phase field model and energy equation. Unlike most previous work, we take the thermocapillary effect into the phase field model considering mixture energy which is different of volume of fluid method (VOF) widely used in GMAW before. We also consider gravity, electromagnetic force, surface tension, buoyancy effect and arc pressure in momentum equation. The spray transfer especially the projected transfer in GMAW is computed as numerical examples with a continuous finite element method and a modified midpoint scheme. Pulse current is set as welding current as the numerical example to show the numerical simulation of metal transfer which fits the theory of GMAW well. From the result compared with the data of high-speed photography and VOF model, the accuracy and stability of the model and scheme are easily validated and also the new model has the higher precieion.
Zonta, Zivko J; Flotats, Xavier; Magrí, Albert
2014-08-01
The procedure commonly used for the assessment of the parameters included in activated sludge models (ASMs) relies on the estimation of their optimal value within a confidence region (i.e. frequentist inference). Once optimal values are estimated, parameter uncertainty is computed through the covariance matrix. However, alternative approaches based on the consideration of the model parameters as probability distributions (i.e. Bayesian inference), may be of interest. The aim of this work is to apply (and compare) both Bayesian and frequentist inference methods when assessing uncertainty for an ASM-type model, which considers intracellular storage and biomass growth, simultaneously. Practical identifiability was addressed exclusively considering respirometric profiles based on the oxygen uptake rate and with the aid of probabilistic global sensitivity analysis. Parameter uncertainty was thus estimated according to both the Bayesian and frequentist inferential procedures. Results were compared in order to evidence the strengths and weaknesses of both approaches. Since it was demonstrated that Bayesian inference could be reduced to a frequentist approach under particular hypotheses, the former can be considered as a more generalist methodology. Hence, the use of Bayesian inference is encouraged for tackling inferential issues in ASM environments.
NASA Astrophysics Data System (ADS)
Serra, Reviewed By Martin J.
2000-01-01
Genomics is one of the most rapidly expanding areas of science. This book is an outgrowth of a series of lectures given by one of the former heads (CRC) of the Human Genome Initiative. The book is designed to reach a wide audience, from biologists with little chemical or physical science background through engineers, computer scientists, and physicists with little current exposure to the chemical or biological principles of genetics. The text starts with a basic review of the chemical and biological properties of DNA. However, without either a biochemistry background or a supplemental biochemistry text, this chapter and much of the rest of the text would be difficult to digest. The second chapter is designed to put DNA into the context of the larger chromosomal unit. Specialized chromosomal structures and sequences (centromeres, telomeres) are introduced, leading to a section on chromosome organization and purification. The next 4 chapters cover the physical (hybridization, electrophoresis), chemical (polymerase chain reaction), and biological (genetic) techniques that provide the backbone of genomic analysis. These chapters cover in significant detail the fundamental principles underlying each technique and provide a firm background for the remainder of the text. Chapters 79 consider the need and methods for the development of physical maps. Chapter 7 primarily discusses chromosomal localization techniques, including in situ hybridization, FISH, and chromosome paintings. The next two chapters focus on the development of libraries and clones. In particular, Chapter 9 considers the limitations of current mapping and clone production. The current state and future of DNA sequencing is covered in the next three chapters. The first considers the current methods of DNA sequencing - especially gel-based methods of analysis, although other possible approaches (mass spectrometry) are introduced. Much of the chapter addresses the limitations of current methods, including analysis of error in sequencing and current bottlenecks in the sequencing effort. The next chapter describes the steps necessary to scale current technologies for the sequencing of entire genomes. Chapter 12 examines alternate methods for DNA sequencing. Initially, methods of single-molecule sequencing and sequencing by microscopy are introduced; the majority of the chapter is devoted to the development of DNA sequencing methods using chip microarrays and hybridization. The remaining chapters (13-15) consider the uses and analysis of DNA sequence information. The initial focus is on the identification of genes. Several examples are given of the use of DNA sequence information for diagnosis of inherited or infectious diseases. The sequence-specific manipulation of DNA is discussed in Chapter 14. The final chapter deals with the implications of large-scale sequencing, including methods for identifying genes and finding errors in DNA sequences, to the development of computer algorithms for the interpretation of DNA sequence information. The text figures are black and white line drawings that, although clearly done, seem a bit primitive for 1999. While I appreciated the simplicity of the drawings, many students accustomed to more colorful presentations will find them wanting. The four color figures in the center of the text seem an afterthought and add little to the text's clarity. Each chapter has a set of additional reading sources, mostly primary sources. Often, specialized topics are offset into boxes that provide clarification and amplification without cluttering the text. An appendix includes a list of the Web-based database resources. As an undergraduate instructor who has previously taught biochemistry, molecular biology, and a course on the human genome, I found many interesting tidbits and amplifications throughout the text. I would recommend this book as a text for an advanced undergraduate or beginning graduate course in genomics. Although the text works though several examples of genetic and genome analysis, additional problem/homework sets would need to be developed to ensure student comprehension. The text steers clear of the ethical implications of the Human Genome Initiative and remains true to its subtitle The Science and Technology .
Sensitivity recovery for the AX-PET prototype using inter-crystal scattering events
NASA Astrophysics Data System (ADS)
Gillam, John E.; Solevi, Paola; Oliver, Josep F.; Casella, Chiara; Heller, Matthieu; Joram, Christian; Rafecas, Magdalena
2014-08-01
The development of novel detection devices and systems such as the AX-positron emission tomography (PET) demonstrator often introduce or increase the measurement of atypical coincidence events such as inter-crystal scattering (ICS). In more standard systems, ICS events often go undetected and the small measured fraction may be ignored. As the measured quantity of such events in the data increases, so too does the importance of considering them during image reconstruction. Generally, treatment of ICS events will attempt to determine which of the possible candidate lines of response (LoRs) correctly determine the annihilation photon trajectory. However, methods of assessment often have low success rates or are computationally demanding. In this investigation alternative approaches are considered. Experimental data was taken using the AX-PET prototype and a NEMA phantom. Three methods of ICS treatment were assessed—each of which considered all possible candidate LoRs during image reconstruction. Maximum likelihood expectation maximization was used in conjunction with both standard (line-like) and novel (V-like in this investigation) detection responses modeled within the system matrix. The investigation assumed that no information other than interaction locations was available to distinguish between candidates, yet the methods assessed all provided means by which such information could be included. In all cases it was shown that the signal to noise ratio is increased using ICS events. However, only one method, which used full modeling of the ICS response in the system matrix—the V-like model—provided enhancement in all figures of merit assessed in this investigation. Finally, the optimal method of ICS incorporation was demonstrated using data from two small animals measured using the AX-PET demonstrator.
Doble, Brett; Wordsworth, Sarah; Rogers, Chris A; Welbourn, Richard; Byrne, James; Blazeby, Jane M
2017-08-01
This review aims to evaluate the current literature on the procedural costs of bariatric surgery for the treatment of severe obesity. Using a published framework for the conduct of micro-costing studies for surgical interventions, existing cost estimates from the literature are assessed for their accuracy, reliability and comprehensiveness based on their consideration of seven 'important' cost components. MEDLINE, PubMed, key journals and reference lists of included studies were searched up to January 2017. Eligible studies had to report per-case, total procedural costs for any type of bariatric surgery broken down into two or more individual cost components. A total of 998 citations were screened, of which 13 studies were included for analysis. Included studies were mainly conducted from a US hospital perspective, assessed either gastric bypass or adjustable gastric banding procedures and considered a range of different cost components. The mean total procedural costs for all included studies was US$14,389 (range, US$7423 to US$33,541). No study considered all of the recommended 'important' cost components and estimation methods were poorly reported. The accuracy, reliability and comprehensiveness of the existing cost estimates are, therefore, questionable. There is a need for a comparative cost analysis of the different approaches to bariatric surgery, with the most appropriate costing approach identified to be micro-costing methods. Such an analysis will not only be useful in estimating the relative cost-effectiveness of different surgeries but will also ensure appropriate reimbursement and budgeting by healthcare payers to ensure barriers to access this effective treatment by severely obese patients are minimised.
Haby, Michelle M; Chapman, Evelina; Clark, Rachel; Galvão, Luiz A C
2016-04-01
Objective To inform policy by providing an overview of systematic reviews on interventions that facilitate sustainable energy use and have a positive impact on health. Methods Systematic review methods were used to synthesize evidence from multiple systematic reviews and economic evaluations through a comprehensive search of 13 databases and nine websites based on a pre-defined protocol, including clear inclusion criteria. Both grey and peer-reviewed literature published in English, Spanish, and Portuguese during the 17 years from January 1997 - January 2014 was included. To classify as "sustainable," interventions needed to aim to positively impact at least two dimensions of the integrated framework for sustainable development and include measures of health impact. Results Five systematic reviews and one economic evaluation met the inclusion criteria. The most promising interventions that impacted health were electricity for lighting and other uses (developing countries); improved stoves for cooking and health and/or cleaner fuels for cooking (developing countries); and household energy efficiency measures (developed countries). These interventions also had potential environmental and economic impacts. Their cost-effectiveness is not known, nor is their impact on health inequalities. Conclusions What is needed now is careful implementation of interventions where the impacts are likely to be positive but their implementation needs to be rigorously evaluated, including possible adverse impacts. Care needs to be taken not to exacerbate health inequalities and to consider context, human behavior and cultural factors so that the potential health benefits are realized in real-life implementation. Possible impact on health inequalities needs to be considered and measured in future primary studies and systematic reviews.
Benzi, Michele; Evans, Thomas M.; Hamilton, Steven P.; ...
2017-03-05
Here, we consider hybrid deterministic-stochastic iterative algorithms for the solution of large, sparse linear systems. Starting from a convergent splitting of the coefficient matrix, we analyze various types of Monte Carlo acceleration schemes applied to the original preconditioned Richardson (stationary) iteration. We expect that these methods will have considerable potential for resiliency to faults when implemented on massively parallel machines. We also establish sufficient conditions for the convergence of the hybrid schemes, and we investigate different types of preconditioners including sparse approximate inverses. Numerical experiments on linear systems arising from the discretization of partial differential equations are presented.