Science.gov

Sample records for applied systems analysis

  1. Sneak analysis applied to process systems

    NASA Astrophysics Data System (ADS)

    Whetton, Cris

    Traditional safety analyses, such as HAZOP, FMEA, FTA, and MORT, are less than effective at identifying hazards resulting from incorrect 'flow' - whether this be flow of information, actions, electric current, or even the literal flow of process fluids. Sneak Analysis (SA) has existed since the mid nineteen-seventies as a means of identifying such conditions in electric circuits; in which area, it is usually known as Sneak Circuit Analysis (SCA). This paper extends the ideas of Sneak Circuit Analysis to a general method of Sneak Analysis applied to process plant. The methods of SA attempt to capitalize on previous work in the electrical field by first producing a pseudo-electrical analog of the process and then analyzing the analog by the existing techniques of SCA, supplemented by some additional rules and clues specific to processes. The SA method is not intended to replace any existing method of safety analysis; instead, it is intended to supplement such techniques as HAZOP and FMEA by providing systematic procedures for the identification of a class of potential problems which are not well covered by any other method.

  2. Applying Systems Analysis to Program Failure in Organizations.

    ERIC Educational Resources Information Center

    Holt, Margaret E.; And Others

    1986-01-01

    Certain systems analysis techniques can be applied to examinations of program failure in continuing education to locate weaknesses in planning and implementing stages. Questions to guide an analysis and various procedures are recommended. Twelve issues that contribute to failures or discontinuations are identified. (Author/MLW)

  3. Probabilistic Analysis Techniques Applied to Complex Spacecraft Power System Modeling

    NASA Technical Reports Server (NTRS)

    Hojnicki, Jeffrey S.; Rusick, Jeffrey J.

    2005-01-01

    Electric power system performance predictions are critical to spacecraft, such as the International Space Station (ISS), to ensure that sufficient power is available to support all the spacecraft s power needs. In the case of the ISS power system, analyses to date have been deterministic, meaning that each analysis produces a single-valued result for power capability because of the complexity and large size of the model. As a result, the deterministic ISS analyses did not account for the sensitivity of the power capability to uncertainties in model input variables. Over the last 10 years, the NASA Glenn Research Center has developed advanced, computationally fast, probabilistic analysis techniques and successfully applied them to large (thousands of nodes) complex structural analysis models. These same techniques were recently applied to large, complex ISS power system models. This new application enables probabilistic power analyses that account for input uncertainties and produce results that include variations caused by these uncertainties. Specifically, N&R Engineering, under contract to NASA, integrated these advanced probabilistic techniques with Glenn s internationally recognized ISS power system model, System Power Analysis for Capability Evaluation (SPACE).

  4. Probabilistic Analysis Techniques Applied to Complex Spacecraft Power System Modeling

    NASA Technical Reports Server (NTRS)

    Hojnicki, Jeffrey S.; Rusick, Jeffrey J.

    2005-01-01

    Electric power system performance predictions are critical to spacecraft, such as the International Space Station (ISS), to ensure that sufficient power is available to support all the spacecraft s power needs. In the case of the ISS power system, analyses to date have been deterministic, meaning that each analysis produces a single-valued result for power capability because of the complexity and large size of the model. As a result, the deterministic ISS analyses did not account for the sensitivity of the power capability to uncertainties in model input variables. Over the last 10 years, the NASA Glenn Research Center has developed advanced, computationally fast, probabilistic analysis techniques and successfully applied them to large (thousands of nodes) complex structural analysis models. These same techniques were recently applied to large, complex ISS power system models. This new application enables probabilistic power analyses that account for input uncertainties and produce results that include variations caused by these uncertainties. Specifically, N&R Engineering, under contract to NASA, integrated these advanced probabilistic techniques with Glenn s internationally recognized ISS power system model, System Power Analysis for Capability Evaluation (SPACE).

  5. Thermographic techniques applied to solar collector systems analysis

    SciTech Connect

    Eden, A.

    1980-02-01

    The use of thermography to analyze large solar collector array systems under dynamic operating conditions is discussed. The research at the Solar Energy Research Institute (SERI) in this area has focused on thermographic techniques and equipment to determine temperature distributions, flow patterns, and air blockages in solar collectors. The results of this extensive study, covering many sites and types of collectors, illustrate the capabilities of infrared (IR) analysis as a qualitative analysis tool and operation and maintenance procedure when applied to large arrays. Thermographic analysis of most collector systems qualitatively showed relative temperature distributions that indicated balanced flow patterns. In three significant cases, blocked or broken collector arrays, which previously had gone undetected, were discovered. Using this analysis, validation studies of large computer codes could examine collector arrays for flow patterns or blockages that could cause disagreement between actual and predicted performance. Initial operation and balancing of large systems could be accomplished without complicated sensor systems not needed for normal operations. Maintenance personnel could quickly check their systems without climbing onto the roof and without complicated sensor systems.

  6. Systems design analysis applied to launch vehicle configuration

    NASA Astrophysics Data System (ADS)

    Ryan, R.; Verderaime, V.

    1993-01-01

    As emphasis shifts from optimum-performance aerospace systems to least lift-cycle costs, systems designs must seek, adapt, and innovate cost improvement techniques in design through operations. The systems design process of concept, definition, and design was assessed for the types and flow of total quality management techniques that may be applicable in a launch vehicle systems design analysis. Techniques discussed are task ordering, quality leverage, concurrent engineering, Pareto's principle, robustness, quality function deployment, criteria, and others. These cost oriented techniques are as applicable to aerospace systems design analysis as to any large commercial system.

  7. Systems design analysis applied to launch vehicle configuration

    NASA Technical Reports Server (NTRS)

    Ryan, R.; Verderaime, V.

    1993-01-01

    As emphasis shifts from optimum-performance aerospace systems to least lift-cycle costs, systems designs must seek, adapt, and innovate cost improvement techniques in design through operations. The systems design process of concept, definition, and design was assessed for the types and flow of total quality management techniques that may be applicable in a launch vehicle systems design analysis. Techniques discussed are task ordering, quality leverage, concurrent engineering, Pareto's principle, robustness, quality function deployment, criteria, and others. These cost oriented techniques are as applicable to aerospace systems design analysis as to any large commercial system.

  8. Applying an Activity System to Online Collaborative Group Work Analysis

    ERIC Educational Resources Information Center

    Choi, Hyungshin; Kang, Myunghee

    2010-01-01

    This study determines whether an activity system provides a systematic framework to analyse collaborative group work. Using an activity system as a unit of analysis, the research examined learner behaviours, conflicting factors and facilitating factors while students engaged in collaborative work via asynchronous computer-mediated communication.…

  9. Dynamical systems analysis applied to working memory data.

    PubMed

    Gasimova, Fidan; Robitzsch, Alexander; Wilhelm, Oliver; Boker, Steven M; Hu, Yueqin; Hülür, Gizem

    2014-01-01

    In the present paper we investigate weekly fluctuations in the working memory capacity (WMC) assessed over a period of 2 years. We use dynamical system analysis, specifically a second order linear differential equation, to model weekly variability in WMC in a sample of 112 9th graders. In our longitudinal data we use a B-spline imputation method to deal with missing data. The results show a significant negative frequency parameter in the data, indicating a cyclical pattern in weekly memory updating performance across time. We use a multilevel modeling approach to capture individual differences in model parameters and find that a higher initial performance level and a slower improvement at the MU task is associated with a slower frequency of oscillation. Additionally, we conduct a simulation study examining the analysis procedure's performance using different numbers of B-spline knots and values of time delay embedding dimensions. Results show that the number of knots in the B-spline imputation influence accuracy more than the number of embedding dimensions.

  10. Grounding system analysis in transients programs applying electromagnetic field approach

    SciTech Connect

    Heimbach, M.; Grcev, L.D.

    1997-01-01

    Lightning protection studies of substations and power systems require knowledge of the dynamic behavior of large grounding grids during electromagnetic transients. This paper presents strategies which allow to incorporate complex grounding structures computed using a rigorous electromagnetic model in transients programs. A novel technique for rational function representation of frequency-dependent grounding system impedances in the EMTP is described. An arbitrary number of feeding points can be modeled as mutual coupling is taken into account. Overvoltages throughout electrical power systems and the transient ground potential rise in the surroundings of grounding structures can be computed.

  11. Experiences with Probabilistic Analysis Applied to Controlled Systems

    NASA Technical Reports Server (NTRS)

    Kenny, Sean P.; Giesy, Daniel P.

    2004-01-01

    This paper presents a semi-analytic method for computing frequency dependent means, variances, and failure probabilities for arbitrarily large-order closed-loop dynamical systems possessing a single uncertain parameter or with multiple highly correlated uncertain parameters. The approach will be shown to not suffer from the same computational challenges associated with computing failure probabilities using conventional FORM/SORM techniques. The approach is demonstrated by computing the probabilistic frequency domain performance of an optimal feed-forward disturbance rejection scheme.

  12. System Analysis Applied to Autonomy: Application to High-Altitude Long-Endurance Remotely Operated Aircraft

    NASA Technical Reports Server (NTRS)

    Young, Larry A.; Yetter, Jeffrey A.; Guynn, Mark D.

    2006-01-01

    Maturation of intelligent systems technologies and their incorporation into aerial platforms are dictating the development of new analysis tools and incorporation of such tools into existing system analysis methodologies in order to fully capture the trade-offs of autonomy on vehicle and mission success. A first-order "system analysis of autonomy" methodology is outlined in this paper. Further, this analysis methodology is subsequently applied to notional high-altitude long-endurance (HALE) aerial vehicle missions.

  13. Applying reliability analysis to design electric power systems for More-electric aircraft

    NASA Astrophysics Data System (ADS)

    Zhang, Baozhu

    The More-Electric Aircraft (MEA) is a type of aircraft that replaces conventional hydraulic and pneumatic systems with electrically powered components. These changes have significantly challenged the aircraft electric power system design. This thesis investigates how reliability analysis can be applied to automatically generate system topologies for the MEA electric power system. We first use a traditional method of reliability block diagrams to analyze the reliability level on different system topologies. We next propose a new methodology in which system topologies, constrained by a set reliability level, are automatically generated. The path-set method is used for analysis. Finally, we interface these sets of system topologies with control synthesis tools to automatically create correct-by-construction control logic for the electric power system.

  14. System Analysis Applied to Autonomy: Application to Human-Rated Lunar/Mars Landers

    NASA Technical Reports Server (NTRS)

    Young, Larry A.

    2006-01-01

    System analysis is an essential technical discipline for the modern design of spacecraft and their associated missions. Specifically, system analysis is a powerful aid in identifying and prioritizing the required technologies needed for mission and/or vehicle development efforts. Maturation of intelligent systems technologies, and their incorporation into spacecraft systems, are dictating the development of new analysis tools, and incorporation of such tools into existing system analysis methodologies, in order to fully capture the trade-offs of autonomy on vehicle and mission success. A "system analysis of autonomy" methodology will be outlined and applied to a set of notional human-rated lunar/Mars lander missions toward answering these questions: 1. what is the optimum level of vehicle autonomy and intelligence required? and 2. what are the specific attributes of an autonomous system implementation essential for a given surface lander mission/application in order to maximize mission success? Future human-rated lunar/Mars landers, though nominally under the control of their crew, will, nonetheless, be highly automated systems. These automated systems will range from mission/flight control functions, to vehicle health monitoring and prognostication, to life-support and other "housekeeping" functions. The optimum degree of autonomy afforded to these spacecraft systems/functions has profound implications from an exploration system architecture standpoint.

  15. System Sensitivity Analysis Applied to the Conceptual Design of a Dual-Fuel Rocket SSTO

    NASA Technical Reports Server (NTRS)

    Olds, John R.

    1994-01-01

    This paper reports the results of initial efforts to apply the System Sensitivity Analysis (SSA) optimization method to the conceptual design of a single-stage-to-orbit (SSTO) launch vehicle. SSA is an efficient, calculus-based MDO technique for generating sensitivity derivatives in a highly multidisciplinary design environment. The method has been successfully applied to conceptual aircraft design and has been proven to have advantages over traditional direct optimization methods. The method is applied to the optimization of an advanced, piloted SSTO design similar to vehicles currently being analyzed by NASA as possible replacements for the Space Shuttle. Powered by a derivative of the Russian RD-701 rocket engine, the vehicle employs a combination of hydrocarbon, hydrogen, and oxygen propellants. Three primary disciplines are included in the design - propulsion, performance, and weights & sizing. A complete, converged vehicle analysis depends on the use of three standalone conceptual analysis computer codes. Efforts to minimize vehicle dry (empty) weight are reported in this paper. The problem consists of six system-level design variables and one system-level constraint. Using SSA in a 'manual' fashion to generate gradient information, six system-level iterations were performed from each of two different starting points. The results showed a good pattern of convergence for both starting points. A discussion of the advantages and disadvantages of the method, possible areas of improvement, and future work is included.

  16. Success Factors of European Syndromic Surveillance Systems: A Worked Example of Applying Qualitative Comparative Analysis

    PubMed Central

    Ziemann, Alexandra; Fouillet, Anne; Brand, Helmut; Krafft, Thomas

    2016-01-01

    Introduction Syndromic surveillance aims at augmenting traditional public health surveillance with timely information. To gain a head start, it mainly analyses existing data such as from web searches or patient records. Despite the setup of many syndromic surveillance systems, there is still much doubt about the benefit of the approach. There are diverse interactions between performance indicators such as timeliness and various system characteristics. This makes the performance assessment of syndromic surveillance systems a complex endeavour. We assessed if the comparison of several syndromic surveillance systems through Qualitative Comparative Analysis helps to evaluate performance and identify key success factors. Materials and Methods We compiled case-based, mixed data on performance and characteristics of 19 syndromic surveillance systems in Europe from scientific and grey literature and from site visits. We identified success factors by applying crisp-set Qualitative Comparative Analysis. We focused on two main areas of syndromic surveillance application: seasonal influenza surveillance and situational awareness during different types of potentially health threatening events. Results We found that syndromic surveillance systems might detect the onset or peak of seasonal influenza earlier if they analyse non-clinical data sources. Timely situational awareness during different types of events is supported by an automated syndromic surveillance system capable of analysing multiple syndromes. To our surprise, the analysis of multiple data sources was no key success factor for situational awareness. Conclusions We suggest to consider these key success factors when designing or further developing syndromic surveillance systems. Qualitative Comparative Analysis helped interpreting complex, mixed data on small-N cases and resulted in concrete and practically relevant findings. PMID:27182731

  17. Biomolecular filters for improved separation of output signals in enzyme logic systems applied to biomedical analysis.

    PubMed

    Halámek, Jan; Zhou, Jian; Halámková, Lenka; Bocharova, Vera; Privman, Vladimir; Wang, Joseph; Katz, Evgeny

    2011-11-15

    Biomolecular logic systems processing biochemical input signals and producing "digital" outputs in the form of YES/NO were developed for analysis of physiological conditions characteristic of liver injury, soft tissue injury, and abdominal trauma. Injury biomarkers were used as input signals for activating the logic systems. Their normal physiological concentrations were defined as logic-0 level, while their pathologically elevated concentrations were defined as logic-1 values. Since the input concentrations applied as logic 0 and 1 values were not sufficiently different, the output signals being at low and high values (0, 1 outputs) were separated with a short gap making their discrimination difficult. Coupled enzymatic reactions functioning as a biomolecular signal processing system with a built-in filter property were developed. The filter process involves a partial back-conversion of the optical-output-signal-yielding product, but only at its low concentrations, thus allowing the proper discrimination between 0 and 1 output values.

  18. Applied Behavior Analysis

    ERIC Educational Resources Information Center

    Szapacs, Cindy

    2006-01-01

    Teaching strategies that work for typically developing children often do not work for those diagnosed with an autism spectrum disorder. However, teaching strategies that work for children with autism do work for typically developing children. In this article, the author explains how the principles and concepts of Applied Behavior Analysis can be…

  19. Applying Monte-Carlo simulations to optimize an inelastic neutron scattering system for soil carbon analysis

    USDA-ARS?s Scientific Manuscript database

    Computer Monte-Carlo (MC) simulations (Geant4) of neutron propagation and acquisition of gamma response from soil samples was applied to evaluate INS system performance characteristic [sensitivity, minimal detectable level (MDL)] for soil carbon measurement. The INS system model with best performanc...

  20. RFI analysis applied to the TDRSS system. [Tracking and Data Relay Satellite System performance

    NASA Technical Reports Server (NTRS)

    Jenny, J. A.

    1973-01-01

    The effect of radio frequency interference (RFI) on the proposed Tracking and Data Relay Satellite System (TDRSS) was assessed. The method of assessing RFI was to create a discrete emitter listing containing all the required parameters of transmitters in the applicable VHF and UHF frequency bands. The transmitter and spacecraft receiver characteristics were used to calculate the RFI contribution due to each emitter. The individual contributions were summed to obtain the total impact in the operational bandwidth. Using an as yet incomplete emitter base, it is concluded that the 136- to 137-MHz band should be used by TDRSS rather than the whole 136- to 138-MHz band because of the higher interference levels in the 137- to 138 MHz band. Even when restricting the link to 136 to 137 MHz, the existing link design is marginal, and it is recommended that interference reduction units, such as the adaptive digital filter, be incorporated in the TDRSS ground station.

  1. Human factors analysis and classification system applied to civil aircraft accidents in India.

    PubMed

    Gaur, Deepak

    2005-05-01

    The Human Factors Analysis and Classification System (HFACS) has gained wide acceptance as a tool to classify human factors in aircraft accidents and incidents. This study on application of HFACS to civil aircraft accident reports at Directorate General Civil of Aviation (DGCA), India, was conducted to ascertain the practicability of applying HFACS to existing investigation reports and to analyze the trends of human factor causes of civil aircraft accidents. Accident investigation reports held at DGCA, New Delhi, for the period 1990--99 were scrutinized. In all, 83 accidents occurred during this period, of which 48 accident reports were evaluated in this study. One or more human factors contributed to 37 of the 48 (77.1%) accidents. The commonest unsafe act was 'skill based errors' followed by 'decision errors.' Violations of laid down rules were contributory in 16 cases (33.3%). 'Preconditions for unsafe acts' were seen in 23 of the 48 cases (47.9%). A fairly large number (52.1%) had 'organizational influences' contributing to the accident. These results are in consonance with larger studies of accidents in the U.S. Navy and general aviation. Such a high percentage of 'organizational influences' has not been reported in other studies. This is a healthy sign for Indian civil aviation, provided effective remedial action for the same is undertaken.

  2. Applying Monte-Carlo simulations to optimize an inelastic neutron scattering system for soil carbon analysis.

    PubMed

    Yakubova, Galina; Kavetskiy, Aleksandr; Prior, Stephen A; Torbert, H Allen

    2017-10-01

    Computer Monte-Carlo (MC) simulations (Geant4) of neutron propagation and acquisition of gamma response from soil samples was applied to evaluate INS system performance characteristic [minimal detectible level (MDL), sensitivity] for soil carbon measurement. The INS system model with best performance characteristics was determined based on MC simulation results. Measurements of MDL using an experimental prototype based on this model demonstrated good agreement with simulated data. This prototype will be used as the base engineering design for a new INS system. Copyright © 2017. Published by Elsevier Ltd.

  3. Model-Based Systems Engineering With the Architecture Analysis and Design Language (AADL) Applied to NASA Mission Operations

    NASA Technical Reports Server (NTRS)

    Munoz Fernandez, Michela Miche

    2014-01-01

    The potential of Model Model Systems Engineering (MBSE) using the Architecture Analysis and Design Language (AADL) applied to space systems will be described. AADL modeling is applicable to real-time embedded systems- the types of systems NASA builds. A case study with the Juno mission to Jupiter showcases how this work would enable future missions to benefit from using these models throughout their life cycle from design to flight operations.

  4. Model-Based Systems Engineering With the Architecture Analysis and Design Language (AADL) Applied to NASA Mission Operations

    NASA Technical Reports Server (NTRS)

    Munoz Fernandez, Michela Miche

    2014-01-01

    The potential of Model Model Systems Engineering (MBSE) using the Architecture Analysis and Design Language (AADL) applied to space systems will be described. AADL modeling is applicable to real-time embedded systems- the types of systems NASA builds. A case study with the Juno mission to Jupiter showcases how this work would enable future missions to benefit from using these models throughout their life cycle from design to flight operations.

  5. Morphological analysis of galvanized coating applied under vibrowave process system conditions

    NASA Astrophysics Data System (ADS)

    Lebedev, V. A.; Ivanov, V. V.; Fedorov, V. P.

    2016-04-01

    The article presents the morphological research results of galvanized coating applied to the metal surface in the course of mechanical and chemical synthesis realized under vibrowave process system conditions. The paper reveals the specifics of the coating morphology, its activating role in free-moving indentors formed under the impact of low-frequency vibrations and its positive influence on the operational performance of the part surface layer. The advantages of this galvanized coating application method are presented in comparison with conventional methods.

  6. Applying STAMP in Accident Analysis

    NASA Technical Reports Server (NTRS)

    Leveson, Nancy; Daouk, Mirna; Dulac, Nicolas; Marais, Karen

    2003-01-01

    Accident models play a critical role in accident investigation and analysis. Most traditional models are based on an underlying chain of events. These models, however, have serious limitations when used for complex, socio-technical systems. Previously, Leveson proposed a new accident model (STAMP) based on system theory. In STAMP, the basic concept is not an event but a constraint. This paper shows how STAMP can be applied to accident analysis using three different views or models of the accident process and proposes a notation for describing this process.

  7. Evaluation Applied to Reliability Analysis of Reconfigurable, Highly Reliable, Fault-Tolerant, Computing Systems for Avionics

    NASA Technical Reports Server (NTRS)

    Migneault, G. E.

    1979-01-01

    Emulation techniques are proposed as a solution to a difficulty arising in the analysis of the reliability of highly reliable computer systems for future commercial aircraft. The difficulty, viz., the lack of credible precision in reliability estimates obtained by analytical modeling techniques are established. The difficulty is shown to be an unavoidable consequence of: (1) a high reliability requirement so demanding as to make system evaluation by use testing infeasible, (2) a complex system design technique, fault tolerance, (3) system reliability dominated by errors due to flaws in the system definition, and (4) elaborate analytical modeling techniques whose precision outputs are quite sensitive to errors of approximation in their input data. The technique of emulation is described, indicating how its input is a simple description of the logical structure of a system and its output is the consequent behavior. The use of emulation techniques is discussed for pseudo-testing systems to evaluate bounds on the parameter values needed for the analytical techniques.

  8. Identifying training deficiencies in military pilots by applying the human factors analysis and classification system.

    PubMed

    Li, Wen-Chin; Harris, Don

    2013-01-01

    Without accurate analysis, it is difficult to identify training needs and develop the content of training programs required for preventing aviation accidents. The human factors analysis and classification system (HFACS) is based on Reason's system-wide model of human error. In this study, 523 accidents from the Republic of China Air Force were analyzed in which 1762 human errors were categorized. The results of the analysis showed that errors of judgment and poor decision-making were commonly reported amongst pilots. As a result, it was concluded that there was a need for military pilots to be trained specifically in making decisions in tactical environments. However, application of HFACS also allowed the identification of systemic training deficiencies within the organization further contributing to the accidents observed.

  9. Applied Behavior Analysis in Education.

    ERIC Educational Resources Information Center

    Cooper, John O.

    1982-01-01

    Applied behavioral analysis in education is expanding rapidly. This article describes the dimensions of applied behavior analysis and the contributions this technology offers teachers in the area of systematic applications, direct and daily measurement, and experimental methodology. (CJ)

  10. SYSTEMS ANALYSIS AND SYNTHESIS APPLIED TO OCCUPATIONAL INSTRUCTION IN SECONDARY SCHOOLS.

    ERIC Educational Resources Information Center

    SILVERN, LEONARD C.

    GOAL OF THIS STUDY WAS TO DETERMINE THE FEASIBILITY OF DEVELOPING A MODEL TO DESCRIBE "REAL-LIFE" FEEDBACK SIGNAL PATHS FROM OUTSIDE THE SECONDARY SCHOOL OR SCHOOL DISTRICT TO AN OCCUPATIONAL TEACHER. AFTER A REVIEW AND EVALUATION OF THE LITERATURE IN SYSTEMS ANALYSIS AND SYNTHESIS SINCE 1914, IT WAS CONCLUDED THAT ATTENTION SHOULD BE DIRECTED TO…

  11. A new multi criteria classification approach in a multi agent system applied to SEEG analysis.

    PubMed

    Kinié, A; Ndiaye, M; Montois, J J; Jacquelet, Y

    2007-01-01

    This work is focused on the study of the organization of the SEEG signals during epileptic seizures with multi-agent system approach. This approach is based on cooperative mechanisms of auto-organization at the micro level and of emergence of a global function at the macro level. In order to evaluate this approach we propose a distributed collaborative approach for the classification of the interesting signals. This new multi-criteria classification method is able to provide a relevant brain area structures organisation and to bring out epileptogenic networks elements. The method is compared to another classification approach a fuzzy classification and gives better results when applied to SEEG signals.

  12. Topological data analysis (TDA) applied to reveal pedogenetic principles of European topsoil system.

    PubMed

    Savic, Aleksandar; Toth, Gergely; Duponchel, Ludovic

    2017-05-15

    Recent developments in applied mathematics are bringing new tools that are capable to synthesize knowledge in various disciplines, and help in finding hidden relationships between variables. One such technique is topological data analysis (TDA), a fusion of classical exploration techniques such as principal component analysis (PCA), and a topological point of view applied to clustering of results. Various phenomena have already received new interpretations thanks to TDA, from the proper choice of sport teams to cancer treatments. For the first time, this technique has been applied in soil science, to show the interaction between physical and chemical soil attributes and main soil-forming factors, such as climate and land use. The topsoil data set of the Land Use/Land Cover Area Frame survey (LUCAS) was used as a comprehensive database that consists of approximately 20,000 samples, each described by 12 physical and chemical parameters. After the application of TDA, results obtained were cross-checked against known grouping parameters including five types of land cover, nine types of climate and the organic carbon content of soil. Some of the grouping characteristics observed using standard approaches were confirmed by TDA (e.g., organic carbon content) but novel subtle relationships (e.g., magnitude of anthropogenic effect in soil formation), were discovered as well. The importance of this finding is that TDA is a unique mathematical technique capable of extracting complex relations hidden in soil science data sets, giving the opportunity to see the influence of physicochemical, biotic and abiotic factors on topsoil formation through fresh eyes. Copyright © 2017 Elsevier B.V. All rights reserved.

  13. Energy analysis of facade-integrated photovoltaic systems applied to UAE commercial buildings

    SciTech Connect

    Radhi, Hassan

    2010-12-15

    Developments in the design and manufacture of photovoltaic cells have recently been a growing concern in the UAE. At present, the embodied energy pay-back time (EPBT) is the criterion used for comparing the viability of such technology against other forms. However, the impact of PV technology on the thermal performance of buildings is not considered at the time of EPBT estimation. If additional energy savings gained over the PV system life are also included, the total EPBT could be shorter. This paper explores the variation of the total energy of building integrated photovoltaic systems (BiPV) as a wall cladding system applied to the UAE commercial sector and shows that the ratio between PV output and saving in energy due to PV panels is within the range of 1:3-1:4. The result indicates that for the southern and western facades in the UAE, the embodied energy pay-back time for photovoltaic system is within the range of 12-13 years. When reductions in operational energy are considered, the pay-back time is reduced to 3.0-3.2 years. This study comes to the conclusion that the reduction in operational energy due to PV panels represents an important factor in the estimation of EPBT. (author)

  14. Multi-attribute criteria applied to electric generation energy system analysis LDRD.

    SciTech Connect

    Kuswa, Glenn W.; Tsao, Jeffrey Yeenien; Drennen, Thomas E.; Zuffranieri, Jason V.; Paananen, Orman Henrie; Jones, Scott A.; Ortner, Juergen G.; Brewer, Jeffrey D.; Valdez, Maximo M.

    2005-10-01

    This report began with a Laboratory-Directed Research and Development (LDRD) project to improve Sandia National Laboratories multidisciplinary capabilities in energy systems analysis. The aim is to understand how various electricity generating options can best serve needs in the United States. The initial product is documented in a series of white papers that span a broad range of topics, including the successes and failures of past modeling studies, sustainability, oil dependence, energy security, and nuclear power. Summaries of these projects are included here. These projects have provided a background and discussion framework for the Energy Systems Analysis LDRD team to carry out an inter-comparison of many of the commonly available electric power sources in present use, comparisons of those options, and efforts needed to realize progress towards those options. A computer aid has been developed to compare various options based on cost and other attributes such as technological, social, and policy constraints. The Energy Systems Analysis team has developed a multi-criteria framework that will allow comparison of energy options with a set of metrics that can be used across all technologies. This report discusses several evaluation techniques and introduces the set of criteria developed for this LDRD.

  15. A new multi criteria classification approach in a multi agent system applied to SEEG analysis

    PubMed Central

    Kinie, Abel; Ndiaye, Mamadou Lamine L.; Montois, Jean-Jacques; Jacquelet, Yann

    2007-01-01

    This work is focused on the study of the organization of the SEEG signals during epileptic seizures with multi-agent system approach. This approach is based on cooperative mechanisms of auto-organization at the micro level and of emergence of a global function at the macro level. In order to evaluate this approach we propose a distributed collaborative approach for the classification of the interesting signals. This new multi-criteria classification method is able to provide a relevant brain area structures organisation and to bring out epileptogenic networks elements. The method is compared to another classification approach a fuzzy classification and gives better results when applied to SEEG signals. PMID:18002381

  16. SU-F-P-07: Applying Failure Modes and Effects Analysis to Treatment Planning System QA

    SciTech Connect

    Mathew, D; Alaei, P

    2016-06-15

    Purpose: A small-scale implementation of Failure Modes and Effects Analysis (FMEA) for treatment planning system QA by utilizing methodology of AAPM TG-100 report. Methods: FMEA requires numerical values for severity (S), occurrence (O) and detectability (D) of each mode of failure. The product of these three values gives a risk priority number (RPN). We have implemented FMEA for the treatment planning system (TPS) QA for two clinics which use Pinnacle and Eclipse TPS. Quantitative monthly QA data dating back to 4 years for Pinnacle and 1 year for Eclipse have been used to determine values for severity (deviations from predetermined doses at points or volumes), and occurrence of such deviations. The TPS QA protocol includes a phantom containing solid water and lung- and bone-equivalent heterogeneities. Photon and electron plans have been evaluated in both systems. The dose values at multiple distinct points of interest (POI) within the solid water, lung, and bone-equivalent slabs, as well as mean doses to several volumes of interest (VOI), have been re-calculated monthly using the available algorithms. Results: The computed doses vary slightly month-over-month. There have been more significant deviations following software upgrades, especially if the upgrade involved re-modeling of the beams. TG-100 guidance and the data presented here suggest an occurrence (O) of 2 depending on the frequency of re-commissioning the beams, severity (S) of 3, and detectability (D) of 2, giving an RPN of 12. Conclusion: Computerized treatment planning systems could pose a risk due to dosimetric errors and suboptimal treatment plans. The FMEA analysis presented here suggests that TPS QA should immediately follow software upgrades, but does not need to be performed every month.

  17. Applied Surface Analysis Workshop.

    DTIC Science & Technology

    1979-10-01

    field of surface analysis attended the Workshop. The list of participants follows. 5! A, I Charles Anderson Albert L. Botkin Case Western Reserve...Louis, MO 63166 University of Dayton 300 College Park Richard Chase Dayton, OH 45469 Case Western Reserve University University Circle Brian E. P...Dayton, OH 45469 300 College Park Dayton, OH 45469 Richard W. Hoffman Case Western Reserve University Martin Kordesch Cleveland, OH 44106 Case Western

  18. Technical, economic and environmental analysis of a MSW kerbside separate collection system applied to small communities.

    PubMed

    De Feo, G; Malvano, C

    2012-10-01

    The main aim of this study was to evaluate the costs and environmental impacts induced by a fixed model of MSW kerbside separate collection system for communities up to 10,000 inhabitants, in order to evaluate the convenience for the smaller municipalities to unite and form more economically and environmentally sound systems. This topic is important not only due to the large number of small municipalities (e.g. in Italy 72% of the municipalities has less than 5000 inhabitants) but also to the fact that separate collection systems are typically designed to take into account only the technical and economic aspects, which is a practice but not acceptable in the light of the sustainable development paradigm. In economic terms, between 1000 and 4000 inhabitants, the annual per capita cost for vehicles and personnel decreased, with a maximum at approximately 180€/inhabitants/year; while, from 5000 up to 10,000 inhabitants, the annual per capita cost was practically constant and equal to about 80€/inhabitants/year. For the municipalities of less than 5000 inhabitants, from an economic point of view the aggregation is always advantageous. The environmental impacts were calculated by means of the Life Cycle Assessment tool SimaPro 7.1, while the economic-environmental convenience was evaluated by combining in a simple multicriteria analysis, the annual total per capita cost (€/inhabitants/year) and the annual total per capita environmental impact (kEco-indicator point/inhabitants/year), giving the same importance to each criteria. The analysis was performed by means of the Paired Comparison Technique using the Simple Additive Weighting method. The economic and environmental convenience of the aggregation diminishes with the size of the municipalities: for less than 4000 inhabitants, the aggregation was almost always advantageous (91.7%); while, for more than or equal to 5000 inhabitants, the aggregation was convenient only in 33.3% of the cases. On the whole, out of

  19. Applying systems-level spectral imaging and analysis to reveal the organelle interactome.

    PubMed

    Valm, Alex M; Cohen, Sarah; Legant, Wesley R; Melunis, Justin; Hershberg, Uri; Wait, Eric; Cohen, Andrew R; Davidson, Michael W; Betzig, Eric; Lippincott-Schwartz, Jennifer

    2017-06-01

    The organization of the eukaryotic cell into discrete membrane-bound organelles allows for the separation of incompatible biochemical processes, but the activities of these organelles must be coordinated. For example, lipid metabolism is distributed between the endoplasmic reticulum for lipid synthesis, lipid droplets for storage and transport, mitochondria and peroxisomes for β-oxidation, and lysosomes for lipid hydrolysis and recycling. It is increasingly recognized that organelle contacts have a vital role in diverse cellular functions. However, the spatial and temporal organization of organelles within the cell remains poorly characterized, as fluorescence imaging approaches are limited in the number of different labels that can be distinguished in a single image. Here we present a systems-level analysis of the organelle interactome using a multispectral image acquisition method that overcomes the challenge of spectral overlap in the fluorescent protein palette. We used confocal and lattice light sheet instrumentation and an imaging informatics pipeline of five steps to achieve mapping of organelle numbers, volumes, speeds, positions and dynamic inter-organelle contacts in live cells from a monkey fibroblast cell line. We describe the frequency and locality of two-, three-, four- and five-way interactions among six different membrane-bound organelles (endoplasmic reticulum, Golgi, lysosome, peroxisome, mitochondria and lipid droplet) and show how these relationships change over time. We demonstrate that each organelle has a characteristic distribution and dispersion pattern in three-dimensional space and that there is a reproducible pattern of contacts among the six organelles, that is affected by microtubule and cell nutrient status. These live-cell confocal and lattice light sheet spectral imaging approaches are applicable to any cell system expressing multiple fluorescent probes, whether in normal conditions or when cells are exposed to disturbances such as

  20. Analysis and Design of International Emission Trading Markets Applying System Dynamics Techniques

    NASA Astrophysics Data System (ADS)

    Hu, Bo; Pickl, Stefan

    2010-11-01

    The design and analysis of international emission trading markets is an important actual challenge. Time-discrete models are needed to understand and optimize these procedures. We give an introduction into this scientific area and present actual modeling approaches. Furthermore, we develop a model which is embedded in a holistic problem solution. Measures for energy efficiency are characterized. The economic time-discrete "cap-and-trade" mechanism is influenced by various underlying anticipatory effects. With a systematic dynamic approach the effects can be examined. First numerical results show that fair international emissions trading can only be conducted with the use of protective export duties. Furthermore a comparatively high price which evokes emission reduction inevitably has an inhibiting effect on economic growth according to our model. As it always has been expected it is not without difficulty to find a balance between economic growth and emission reduction. It can be anticipated using our System Dynamics model simulation that substantial changes must be taken place before international emissions trading markets can contribute to global GHG emissions mitigation.

  1. Hydrogeochemistry and statistical analysis applied to understand fluoride provenance in the Guarani Aquifer System, Southern Brazil.

    PubMed

    Marimon, Maria Paula C; Roisenberg, Ari; Suhogusoff, Alexandra V; Viero, Antonio Pedro

    2013-06-01

    High fluoride concentrations (up to 11 mg/L) have been reported in the groundwater of the Guarani Aquifer System (Santa Maria Formation) in the central region of the state of Rio Grande do Sul, Southern Brazil. In this area, dental fluorosis is an endemic disease. This paper presents the geochemical data and the combination of statistical analysis (Principal components and cluster analyses) and geochemical modeling to achieve the hydrogeochemistry of the groundwater and discusses the possible fluoride origin. The groundwater from the Santa Maria Formation is comprised of four different geochemical groups. The first group corresponds to a sodium chloride groundwater which evolves to sodium bicarbonate, the second one, both containing fluoride anomalies. The third group is represented by calcium bicarbonate groundwater, and in the fourth, magnesium is the distinctive parameter. The statistical and geochemical analyses supported by isotopic measurements indicated that groundwater may have originated from mixtures of deeper aquifers and the fluoride concentrations could be derived from rock/water interactions (e.g., desorption from clay minerals).

  2. Enhancing the Effectiveness of Significant Event Analysis: Exploring Personal Impact and Applying Systems Thinking in Primary Care

    PubMed Central

    McNaughton, Elaine; Bruce, David; Holly, Deirdre; Forrest, Eleanor; Macleod, Marion; Kennedy, Susan; Power, Ailsa; Toppin, Denis; Black, Irene; Pooley, Janet; Taylor, Audrey; Swanson, Vivien; Kelly, Moya; Ferguson, Julie; Stirling, Suzanne; Wakeling, Judy; Inglis, Angela; McKay, John; Sargeant, Joan

    2016-01-01

    Introduction: Significant event analysis (SEA) is well established in many primary care settings but can be poorly implemented. Reasons include the emotional impact on clinicians and limited knowledge of systems thinking in establishing why events happen and formulating improvements. To enhance SEA effectiveness, we developed and tested “guiding tools” based on human factors principles. Methods: Mixed-methods development of guiding tools (Personal Booklet—to help with emotional demands and apply a human factors analysis at the individual level; Desk Pad—to guide a team-based systems analysis; and a written Report Format) by a multiprofessional “expert” group and testing with Scottish primary care practitioners who submitted completed enhanced SEA reports. Evaluation data were collected through questionnaire, telephone interviews, and thematic analysis of SEA reports. Results: Overall, 149/240 care practitioners tested the guiding tools and submitted completed SEA reports (62.1%). Reported understanding of how to undertake SEA improved postintervention (P < .001), while most agreed that the Personal Booklet was practical (88/123, 71.5%) and relevant to dealing with related emotions (93/123, 75.6%). The Desk Pad tool helped focus the SEA on systems issues (85/123, 69.1%), while most found the Report Format clear (94/123, 76.4%) and would recommend it (88/123, 71.5%). Most SEA reports adopted a systems approach to analyses (125/149, 83.9%), care improvement (74/149, 49.7), or planned actions (42/149, 28.2%). Discussion: Applying human factors principles to SEA potentially enables care teams to gain a systems-based understanding of why things go wrong, which may help with related emotional demands and with more effective learning and improvement. PMID:27583996

  3. Enhancing the Effectiveness of Significant Event Analysis: Exploring Personal Impact and Applying Systems Thinking in Primary Care.

    PubMed

    Bowie, Paul; McNaughton, Elaine; Bruce, David; Holly, Deirdre; Forrest, Eleanor; Macleod, Marion; Kennedy, Susan; Power, Ailsa; Toppin, Denis; Black, Irene; Pooley, Janet; Taylor, Audrey; Swanson, Vivien; Kelly, Moya; Ferguson, Julie; Stirling, Suzanne; Wakeling, Judy; Inglis, Angela; McKay, John; Sargeant, Joan

    2016-01-01

    Significant event analysis (SEA) is well established in many primary care settings but can be poorly implemented. Reasons include the emotional impact on clinicians and limited knowledge of systems thinking in establishing why events happen and formulating improvements. To enhance SEA effectiveness, we developed and tested "guiding tools" based on human factors principles. Mixed-methods development of guiding tools (Personal Booklet-to help with emotional demands and apply a human factors analysis at the individual level; Desk Pad-to guide a team-based systems analysis; and a written Report Format) by a multiprofessional "expert" group and testing with Scottish primary care practitioners who submitted completed enhanced SEA reports. Evaluation data were collected through questionnaire, telephone interviews, and thematic analysis of SEA reports. Overall, 149/240 care practitioners tested the guiding tools and submitted completed SEA reports (62.1%). Reported understanding of how to undertake SEA improved postintervention (P < .001), while most agreed that the Personal Booklet was practical (88/123, 71.5%) and relevant to dealing with related emotions (93/123, 75.6%). The Desk Pad tool helped focus the SEA on systems issues (85/123, 69.1%), while most found the Report Format clear (94/123, 76.4%) and would recommend it (88/123, 71.5%). Most SEA reports adopted a systems approach to analyses (125/149, 83.9%), care improvement (74/149, 49.7), or planned actions (42/149, 28.2%). Applying human factors principles to SEA potentially enables care teams to gain a systems-based understanding of why things go wrong, which may help with related emotional demands and with more effective learning and improvement.

  4. A Sensitivity Analysis of the Kalman Filter as Applied to an Inertial Navigation System

    DTIC Science & Technology

    1982-06-01

    4 _ * i i i i i i ........ ..T TABLE OF CONTENTS Page I. INTRODUCTION ------------------------------------------- 10 II. MODELS AND ESTIMATION...OF STABILITY --------------------------------------- 19 III. DYNAMIC AND MEASUREMENT SYSTEM MODELS ----------------------- 20 A. REFERENCE AXIS SYSTEM...21 2. Lateral Motion ---------------------------------- 21 C. MODEL DYNAMICS ------------------------------------- 22 1. Longitudinal Motion

  5. Hazard analysis and critical control point systems applied to public health risks: the example of seafood.

    PubMed

    Williams, R A; Zorn, D J

    1997-08-01

    The authors describe the way in which the two components of risk analysis--risk assessment and risk management--can be used in conjunction with the hazard analysis and critical control points concept to determine the allocation of resources at potential critical control points. This approach is examined in the context of risks to human health associated with seafood, and in particular with regard to ciguatera poisoning.

  6. Accuracy Analysis of Precise Point Positioning of Compass Navigation System Applied to Crustal Motion Monitoring

    NASA Astrophysics Data System (ADS)

    Wang, Yuebing

    2017-04-01

    Based on the observation data of Compass/GPSobserved at five stations, time span from July 1, 2014 to June 30, 2016. UsingPPP positioning model of the PANDA software developed by Wuhan University,Analyzedthe positioning accuracy of single system and Compass/GPS integrated resolving, and discussed the capability of Compass navigation system in crustal motion monitoring. The results showed that the positioning accuracy in the east-west directionof the Compass navigation system is lower than the north-south direction (the positioning accuracy de 3 times RMS), in general, the positioning accuracyin the horizontal direction is about 1 2cm and the vertical direction is about 5 6cm. The GPS positioning accuracy in the horizontal direction is better than 1cm and the vertical direction is about 1 2cm. The accuracy of Compass/GPS integrated resolving is quite to GPS. It is worth mentioning that although Compass navigation system precision point positioning accuracy is lower than GPS, two sets of velocity fields obtained by using the Nikolaidis (2002) model to analyze the Compass and GPS time series results respectively, the results showed that the maximum difference of the two sets of velocity field in horizontal directions is 1.8mm/a. The Compass navigation system can now be used to monitor the crustal movement of the large deformation area, based on the velocity field in horizontal direction.

  7. Matrix effects in applying mono- and polyclonal ELISA systems to the analysis of weathered oils in contaminated soil.

    PubMed

    Pollard, S J T; Farmer, J G; Knight, D M; Young, P J

    2002-01-01

    Commercial mono- and polyclonal enzyme-linked immunosorbent assay (ELISA) systems were applied to the on-site analysis of weathered hydrocarbon-contaminated soils at a former integrated steelworks. Comparisons were made between concentrations of solvent extractable matter (SEM) determined gravimetrically by Soxhlet (dichloromethane) extraction and those estimated immunologically by ELISA determination over a concentration range of 2000-330,000 mg SEM/kg soil dry weight. Both ELISA systems tinder-reported for the more weathered soil samples. Results suggest this is due to matrix effects in the sample rather than any inherent bias in the ELISA systems and it is concluded that, for weathered hydrocarbons typical of steelworks and coke production sites, the use of ELISA requires careful consideration as a field technique. Consideration of the target analyte relative to the composition of the hydrocarbon waste encountered appears critical.

  8. A novel image processing and measurement system applied to quantitative analysis of simulated tooth root canal shape

    NASA Astrophysics Data System (ADS)

    Yong, Tao; Yong, Wei; Jin, Guofan; Gao, Xuejun

    2005-02-01

    Dental pulp is located in root canal of tooth. To modern root canal therapy, "Root canal preparation" is the main means to debride dental pulp infection. The shape of root canal will be changed after preparation, so, when assessing the preparation instruments and techniques, the root canal shaping ability especially the apical offset is very important factor. In this paper, a novel digital image processing and measurement system is designed and applied to quantitative analysis of simulated canal shape. By image pretreatment, feature extraction, registration and fusion, the variation of the root canals' characteristics (before and after preparation) can be accurately compared and measured, so as to assess the shaping ability of instruments. When the scanning resolution is 1200dpi or higher, the registration and measurement precision of the system can achieve 0.021mm or higher. The performance of the system is tested by a series of simulated root canals and stainless steel K-files.

  9. [The collector system of the kidney. Applied anatomy based on the analysis of 3-dimensional casts].

    PubMed

    Sampaio, F J; Mandarim-De-Lacerda, C A; De Aragão, A H

    1987-01-01

    The kidney collector system were studied in 120 three-dimensional casts obtained by the injection-corrosion method. To the procurement of the first hundred casts, we made use of a polyester resin; however on the preparation of the last twenty we made use of a mixture containing latex and radiographic contrast as we intended to obtain, besides the casts, the roentgenograms. Two anatomic details, which have a great importance for endourology, were observed: a. the calices' cross draining the meso-renal region and the consequent inter-pelvis-calice (space (IPC) formation which is seen in the roentgenograms, b. a smaller calice which penetrates perpendicularly the surface of the renal pelvis or into the surface of a great calice. The urologists must appreciate these two anatomic structures, before the surgical procedures on the collector system.

  10. Core Logistics Capability Policy Applied to USAF Combat Aircraft Avionics Software: A Systems Engineering Analysis

    DTIC Science & Technology

    2010-06-01

    PROJECT Presented to the Faculty Department of Systems and Engineering Management Graduate School of Engineering and Management Air Force...a light grey box around the definitions. This table, a fit-for-purpose architecture product , is loosely based on the DoD Architecture Framework ...synopsis of ISO/IEC 12207, Raghu Singh of the Federal Aviation Administration states “Whenever a software product needs modifications, the development

  11. Applying ABC Analysis to the Navy’s Inventory Management System

    DTIC Science & Technology

    2014-09-01

    limited in scope, as it covers such a short time frame, but the results are in line with the theory behind using more than just requisition volume to...rate “ bang for the buck” resource allocation. The system looks solely at demand, casualty reports (CASREPs), and platform readiness drivers to group...at least some capacity, all other WSS performance metrics. The maritime forecasting difficulty is a big reason why WSS currently achieves supply

  12. Comparing Waste-to-Energy technologies by applying energy system analysis.

    PubMed

    Münster, Marie; Lund, Henrik

    2010-07-01

    Even when policies of waste prevention, re-use and recycling are prioritised, a fraction of waste will still be left which can be used for energy recovery. This article asks the question: How to utilise waste for energy in the best way seen from an energy system perspective? Eight different Waste-to-Energy technologies are compared with a focus on fuel efficiency, CO(2) reductions and costs. The comparison is carried out by conducting detailed energy system analyses of the present as well as a potential future Danish energy system with a large share of combined heat and power as well as wind power. The study shows potential of using waste for the production of transport fuels. Biogas and thermal gasification technologies are hence interesting alternatives to waste incineration and it is recommended to support the use of biogas based on manure and organic waste. It is also recommended to support research into gasification of waste without the addition of coal and biomass. Together the two solutions may contribute to alternate use of one third of the waste which is currently incinerated. The remaining fractions should still be incinerated with priority to combined heat and power plants with high electric efficiency.

  13. Output-Based Adaptive Meshing Applied to Space Launch System Booster Separation Analysis

    NASA Technical Reports Server (NTRS)

    Dalle, Derek J.; Rogers, Stuart E.

    2015-01-01

    This paper presents details of Computational Fluid Dynamic (CFD) simulations of the Space Launch System during solid-rocket booster separation using the Cart3D inviscid code with comparisons to Overflow viscous CFD results and a wind tunnel test performed at NASA Langley Research Center's Unitary PlanWind Tunnel. The Space Launch System (SLS) launch vehicle includes two solid-rocket boosters that burn out before the primary core stage and thus must be discarded during the ascent trajectory. The main challenges for creating an aerodynamic database for this separation event are the large number of basis variables (including orientation of the core, relative position and orientation of the boosters, and rocket thrust levels) and the complex flow caused by the booster separation motors. The solid-rocket boosters are modified from their form when used with the Space Shuttle Launch Vehicle, which has a rich flight history. However, the differences between the SLS core and the Space Shuttle External Tank result in the boosters separating with much narrower clearances, and so reducing aerodynamic uncertainty is necessary to clear the integrated system for flight. This paper discusses an approach that has been developed to analyze about 6000 wind tunnel simulations and 5000 flight vehicle simulations using Cart3D in adaptive-meshing mode. In addition, a discussion is presented of Overflow viscous CFD runs used for uncertainty quantification. Finally, the article presents lessons learned and improvements that will be implemented in future separation databases.

  14. Global positioning system network analysis with phase ambiguity resolution applied to crustal deformation studies in California

    NASA Technical Reports Server (NTRS)

    Dong, Da-Nan; Bock, Yehuda

    1989-01-01

    An efficient algorithm is developed for multisession adjustment of GPS data with simultaneous orbit determination and ambiguity resolution. Application of the algorithm to the analysis of data from a five-year campaign in progress in southern and central California to monitor tectonic motions using observations by GPS satellites, demonstrates improvements in estimates of station position and satellite orbits when the phase ambiguities are resolved. Most of the phase ambiguities in the GPS network were resolved, particularly for all the baselines of geophysical interest in California.

  15. Mechanism analysis of Magnetohydrodynamic heat shield system and optimization of externally applied magnetic field

    NASA Astrophysics Data System (ADS)

    Li, Kai; Liu, Jun; Liu, Weiqiang

    2017-04-01

    As a novel thermal protection technique for hypersonic vehicles, Magnetohydrodynamic (MHD) heat shield system has been proved to be of great intrinsic value in the hypersonic field. In order to analyze the thermal protection mechanisms of such a system, a physical model is constructed for analyzing the effect of the Lorentz force components in the counter and normal directions. With a series of numerical simulations, the dominating Lorentz force components are analyzed for the MHD heat flux mitigation in different regions of a typical reentry vehicle. Then, a novel magnetic field with variable included angle between magnetic induction line and streamline is designed, which significantly improves the performance of MHD thermal protection in the stagnation and shoulder areas. After that, the relationships between MHD shock control and MHD thermal protection are investigated, based on which the magnetic field above is secondarily optimized obtaining better performances of both shock control and thermal protection. Results show that the MHD thermal protection is mainly determined by the Lorentz force's effect on the boundary layer. From the stagnation to the shoulder region, the flow deceleration effect of the counter-flow component is weakened while the flow deflection effect of the normal component is enhanced. Moreover, there is no obviously positive correlation between the MHD shock control and thermal protection. But once a good Lorentz force's effect on the boundary layer is guaranteed, the thermal protection performance can be further improved with an enlarged shock stand-off distance by strengthening the counter-flow Lorentz force right after shock.

  16. Advanced Behavioral Applications in Schools: A Review of R. Douglas Greer's "Designing Teaching Strategies: An Applied Behavior Analysis Systems Approach"

    ERIC Educational Resources Information Center

    Moxley, Roy A.

    2004-01-01

    R. Douglas Greer's "Designing Teaching Strategies" is an important book directed to advanced students in applied behavior analysis for classrooms. This review presents some of the striking features of the Comprehensive Applied Behavior Analysis to Schooling (CABAS[R]) program and the individualized instruction that the book advances. These include…

  17. Advanced Behavioral Applications in Schools: A Review of R. Douglas Greer's "Designing Teaching Strategies: An Applied Behavior Analysis Systems Approach"

    ERIC Educational Resources Information Center

    Moxley, Roy A.

    2004-01-01

    R. Douglas Greer's "Designing Teaching Strategies" is an important book directed to advanced students in applied behavior analysis for classrooms. This review presents some of the striking features of the Comprehensive Applied Behavior Analysis to Schooling (CABAS[R]) program and the individualized instruction that the book advances. These include…

  18. Women in applied behavior analysis

    PubMed Central

    McSweeney, Frances K.; Donahoe, Patricia; Swindell, Samantha

    2000-01-01

    The status of women in applied behavior analysis was examined by comparing the participation of women in the Journal of Applied Behavior Analysis (JABA) to their participation in three similar journals. For all journals, the percentage of articles with at least one female author, the percentage of authors who are female, and the percentage of articles with a female first author increased from 1978 to 1997. Participation by women in JABA was equal to or greater than participation by women in the comparison journals. However, women appeared as authors on papers in special sections of Behavior Modification substantially more often when the editor was female than when the editor was male. In addition, female membership on the editorial boards of JABA, Behavior Modification, and Behaviour Research and Therapy failed to increase from 1978 to 1997. We conclude that a “glass ceiling” reduces the participation of women at the highest levels of applied behavior analysis and related fields. PMID:22478351

  19. NextGen Brain Microdialysis: Applying Modern Metabolomics Technology to the Analysis of Extracellular Fluid in the Central Nervous System

    PubMed Central

    Kao, Chi-Ya; Anderzhanova, Elmira; Asara, John M.; Wotjak, Carsten T.; Turck, Christoph W.

    2015-01-01

    Microdialysis is a powerful method for in vivo neurochemical analyses. It allows fluid sampling in a dynamic manner in specific brain regions over an extended period of time. A particular focus has been the neurochemical analysis of extracellular fluids to explore central nervous system functions. Brain microdialysis recovers neurotransmitters, low-molecular-weight neuromodulators and neuropeptides of special interest when studying behavior and drug effects. Other small molecules, such as central metabolites, are typically not assessed despite their potential to yield important information related to brain metabolism and activity in selected brain regions. We have implemented a liquid chromatography online mass spectrometry metabolomics platform for an expanded analysis of mouse brain microdialysates. The method is sensitive and delivers information for a far greater number of analytes than commonly used electrochemical and fluorescent detection or biochemical assays. The metabolomics platform was applied to the analysis of microdialysates in a foot shock-induced mouse model of posttraumatic stress disorder (PTSD). The rich metabolite data information was then used to delineate affected prefrontal molecular pathways that reflect individual susceptibility for developing PTSD-like symptoms. We demonstrate that hypothesis-free metabolomics can be adapted to the analysis of microdialysates for the discovery of small molecules with functional significance. PMID:27602357

  20. Frequency Domain Analysis of Beat-Less Control Method for Converter-Inverter Driving Systems Applied to AC Electric Cars

    NASA Astrophysics Data System (ADS)

    Kimura, Akira

    In inverter-converter driving systems for AC electric cars, the DC input voltage of an inverter contains a ripple component with a frequency that is twice as high as the line voltage frequency, because of a single-phase converter. The ripple component of the inverter input voltage causes pulsations on torques and currents of driving motors. To decrease the pulsations, a beat-less control method, which modifies a slip frequency depending on the ripple component, is applied to the inverter control. In the present paper, the beat-less control method was analyzed in the frequency domain. In the first step of the analysis, transfer functions, which revealed the relationship among the ripple component of the inverter input voltage, the slip frequency, the motor torque pulsation and the current pulsation, were derived with a synchronous rotating model of induction motors. An analysis model of the beat-less control method was then constructed using the transfer functions. The optimal setting of the control method was obtained according to the analysis model. The transfer functions and the analysis model were verified through simulations.

  1. Integration of health care process analysis in the design of a clinical information system: applying to the blood transfusion process.

    PubMed Central

    Staccini, P.; Joubert, M.; Quaranta, J. F.; Fieschi, D.; Fieschi, M.

    2000-01-01

    Hospital information systems have to support quality improvement objectives. The requirements of the system have to meet users' needs in relation to both the quality (efficacy, conformity, safety) and the monitoring of all health care activities (traceability). Information analysts need complementary methods to conceptualize clinical information systems that provide actors with immediate individual benefits and guide collective behavioral changes. A methodology is proposed to elicit users' needs using a process-oriented analysis, and it is applied to the field of blood transfusion. We defined a process data model, the main components of which are: activities, resources, constrains, guidelines and indicators. Although some aspects of activity, such as "where", "what else", and "why" are poorly represented by the data model alone, this method of requirement elicitation fits the dynamic of data input for the process to be traced. A hierarchical representation of hospital activities has to be found for this approach to be generalised within the organisation, for the processes to be interrelated, and for their characteristics to be assessed. PMID:11079999

  2. Integration of health care process analysis in the design of a clinical information system: applying to the blood transfusion process.

    PubMed

    Staccini, P; Joubert, M; Quaranta, J F; Fieschi, D; Fieschi, M

    2000-01-01

    Hospital information systems have to support quality improvement objectives. The requirements of the system have to meet users' needs in relation to both the quality (efficacy, conformity, safety) and the monitoring of all health care activities (traceability). Information analysts need complementary methods to conceptualize clinical information systems that provide actors with immediate individual benefits and guide collective behavioral changes. A methodology is proposed to elicit users' needs using a process-oriented analysis, and it is applied to the field of blood transfusion. We defined a process data model, the main components of which are: activities, resources, constrains, guidelines and indicators. Although some aspects of activity, such as "where", "what else", and "why" are poorly represented by the data model alone, this method of requirement elicitation fits the dynamic of data input for the process to be traced. A hierarchical representation of hospital activities has to be found for this approach to be generalised within the organisation, for the processes to be interrelated, and for their characteristics to be assessed.

  3. Simulating Nationwide Pandemics: Applying the Multi-scale Epidemiologic Simulation and Analysis System to Human Infectious Diseases

    SciTech Connect

    Dombroski, M; Melius, C; Edmunds, T; Banks, L E; Bates, T; Wheeler, R

    2008-09-24

    This study uses the Multi-scale Epidemiologic Simulation and Analysis (MESA) system developed for foreign animal diseases to assess consequences of nationwide human infectious disease outbreaks. A literature review identified the state of the art in both small-scale regional models and large-scale nationwide models and characterized key aspects of a nationwide epidemiological model. The MESA system offers computational advantages over existing epidemiological models and enables a broader array of stochastic analyses of model runs to be conducted because of those computational advantages. However, it has only been demonstrated on foreign animal diseases. This paper applied the MESA modeling methodology to human epidemiology. The methodology divided 2000 US Census data at the census tract level into school-bound children, work-bound workers, elderly, and stay at home individuals. The model simulated mixing among these groups by incorporating schools, workplaces, households, and long-distance travel via airports. A baseline scenario with fixed input parameters was run for a nationwide influenza outbreak using relatively simple social distancing countermeasures. Analysis from the baseline scenario showed one of three possible results: (1) the outbreak burned itself out before it had a chance to spread regionally, (2) the outbreak spread regionally and lasted a relatively long time, although constrained geography enabled it to eventually be contained without affecting a disproportionately large number of people, or (3) the outbreak spread through air travel and lasted a long time with unconstrained geography, becoming a nationwide pandemic. These results are consistent with empirical influenza outbreak data. The results showed that simply scaling up a regional small-scale model is unlikely to account for all the complex variables and their interactions involved in a nationwide outbreak. There are several limitations of the methodology that should be explored in future

  4. Advanced behavioral applications in schools: A review of R. Douglas Greer's designing teaching strategies: An applied behavior analysis systems approach

    PubMed Central

    Moxley, Roy A.

    2004-01-01

    R. Douglas Greer1s Designing Teaching Strategies is an important book directed to advanced students in applied behavior analysis for classrooms. This review presents some of the striking features of the Comprehensive Applied Behavior Analysis to Schooling (CABAS®) program and the individualized instruction that the book advances. These include its instruction in literacy, its use of graphing, and its flexibility in systematic organization. Although its readability could be improved, this book has much to recommend it in an approach that has acquired an international following. PMID:22477295

  5. Applied mathematics of chaotic systems

    SciTech Connect

    Jen, E.; Alber, M.; Camassa, R.; Choi, W.; Crutchfield, J.; Holm, D.; Kovacic, G.; Marsden, J.

    1996-07-01

    This is the final report of a three-year, Laboratory-Directed Research and Development (LDRD) project at the Los Alamos National Laboratory (LANL). The objectives of the project were to develop new mathematical techniques for describing chaotic systems and for reexpressing them in forms that can be solved analytically and computationally. The authors focused on global bifurcation analysis of rigid body motion in an ideal incompressible fluid and on an analytical technique for the exact solution of nonlinear cellular automata. For rigid-body motion, they investigated a new completely integrable partial differential equation (PDE) representing model motion of fronts in nematic crystals and studied perturbations of the integrable PDE. For cellular automata with multiple domain structures, the work has included: (1) identification of the associated set of conserved quantities for each type of domain; (2) use of the conserved quantities to construct isomorphism between the nonlinear system and a linear template; and (3) use of exact solvability methods to characterize detailed structure of equilibrium states and to derive bounds for maximal transience times.

  6. Functional analysis, a resilience improvement tool applied to a waste management system - application to the "household waste management chain"

    NASA Astrophysics Data System (ADS)

    Beraud, H.; Barroca, B.; Hubert, G.

    2012-12-01

    A waste management system plays a leading role in the capacity of an area to restart after flooding, as their impact on post-crisis management can be very considerable. Improving resilience, i.e. enabling it to maintain or recover acceptable operating levels after flooding is primordial. To achieve this, we must understand how the system works for bringing any potential dysfunctions to light and taking preventive measures. Functional analysis has been used for understanding the complexity of this type of system. The purpose of this article is to show the interest behind this type of method and the limits in its use for improving resilience of waste management system as well as other urban technical systems1, by means of theoretical modelling and its application on a study site. 1In a systemic vision of the city, urban technical systems combine all the user service systems that are essential for the city to operate (electricity, water supplies, transport, sewerage, etc.). These systems are generally organised in the form of networks (Coutard, 2010; CERTU, 2005).

  7. Network Analysis as a tool for assessing environmental sustainability: applying the ecosystem perspective to a Danish Water Management System.

    PubMed

    Pizzol, Massimo; Scotti, Marco; Thomsen, Marianne

    2013-03-30

    New insights into the sustainable use of natural resources in human systems can be gained through comparison with ecosystems via common indices. In both kinds of system, resources are processed by a number of users within a network, but we consider ecosystems as the only ones displaying sustainable patterns of growth and development. This study aims at using Network Analysis (NA) to move such "ecosystem perspective" from theory into practice. A Danish municipal Water Management System (WMS) is used as case study to test the NA methodology and to discuss its generic applicability. We identified water users within the WMS and represented their interactions as a network of water flows. We computed intensive and extensive indices of system-level performance for seven different network configurations illustrating past conditions (2004-2008) and future scenarios (2015 and 2020). We also computed the same indices for other 24 human systems and for 12 ecosystems, by using information from the existing scientific literature on NA. The comparison of these results reveals that the WMS is similar to the other human systems and that human systems generally differ from ecosystems. The WMS is highly efficient at processing the water resource, but the rigid and almost linear structure makes it vulnerable in situations of stress such as heavy rain events. The analysis of future scenarios showed a trend towards increased sustainability, but differences between past and expected future performance of the WMS are marginal. We argue that future interventions should create alternative pathways for reusing rainwater within the WMS, increasing its potential to withstand the occurrence of flooding. We discuss advantages, limitations, and general applicability of NA as a tool for assessing environmental sustainability in human systems. Copyright © 2013 Elsevier Ltd. All rights reserved.

  8. Discrete-event simulation applied to analysis of waiting lists. Evaluation of a prioritization system for cataract surgery.

    PubMed

    Comas, Mercè; Castells, Xavier; Hoffmeister, Lorena; Román, Rubén; Cots, Francesc; Mar, Javier; Gutiérrez-Moreno, Santiago; Espallargues, Mireia

    2008-12-01

    To outline the methods used to build a discrete-event simulation model for use in decision-making in the context of waiting list management strategies for cataract surgery by comparing a waiting list prioritization system with the routinely used first-in, first-out (FIFO) discipline. The setting was the Spanish health system. The model reproduced the process of cataract, from incidence of need of surgery (meeting indication criteria), through demand, inclusion on a waiting list, and surgery. "Nonexpressed Need" represented the population that, even with need, would not be included on a waiting list. Parameters were estimated from administrative data and research databases. The impact of introducing a prioritization system on the waiting list compared with the FIFO system was assessed. For all patients entering the waiting list, the main outcome variable was waiting time weighted by priority score. A sensitivity analysis with different scenarios of mean waiting time was used to compare the two alternatives. The prioritization system shortened waiting time (weighted by priority score) by 1.55 months (95% CI: 1.47 to 1.62) compared with the FIFO system. This difference was statistically significant for all scenarios (which were defined from a waiting time of 4 months to 24 months under the FIFO system). A tendency to greater time savings in scenarios with longer waiting times was observed. Discrete-event simulation is useful in decision-making when assessing health services. Introducing a waiting list prioritization system produced greater benefit than allocating surgery by waiting time only. Use of the simulation model would allow the impact of proposed policies to reduce waiting lists or assign resources more efficiently to be tested.

  9. Systems Intervention Applied to Counseling

    ERIC Educational Resources Information Center

    Hosford, Ray E.; Lecomte, Conrad

    1973-01-01

    A systems approach provides planning by which counselors can (a) specify what they are trying to accomplish in operational terms, (b) develop specific procedures to promote these objectives, (c) monitor the client's progress continuously, and (d) determine empirically whether or not the counseling goals have been accomplished. (Author)

  10. Probability methods applied to electric power systems

    SciTech Connect

    Not Available

    1989-11-01

    The roots of understanding probabilistic phenomena go back to antiquity. We have been willing to bet our money on the roll of dice or similar events for centuries. Yet, when it comes to betting our lives or livelihood, we have been slow to adapt probabilistic methods to describe uncertainty. As a matter of fact, we are loath to admit a probability of failure when it comes to such structures as bridges, buildings or airplanes. Electric utility engineers the world over realize that reliability of structures and systems can be improved and money can be saved if we use a more enlightened approach to uncertainty. The technology and analytical power that is now being made available to the working engineer make that possible. It is for this reason, the International Council on Probability Methods Applied to Power Systems (PMAPS) was formed in 1985. It is important that engineers have a forum for exchange of knowledge and ideas on subjects related to describing and coping with uncertainty. As the world becomes more complex and demands increase, it is the engineers lot to make the most efficient use possible of scarce resources. The papers contained within this document cover the design and analysis of transmission components, systems analysis and reliability assessment, testing of power components, systems operations planning and probabilistic analysis, power distribution systems, cost, and mathematical modeling. The individual papers have been individually cataloged and indexed.

  11. Electrochemical analysis of chloramphenicol using boron-doped diamond electrode applied to a flow-injection system.

    PubMed

    Chuanuwatanakul, Suchada; Chailapakul, Orawon; Motomizu, Shoji

    2008-04-01

    The electrochemical properties of chloramphenicol at a boron-doped diamond thin-film (BDD) electrode were studied using cyclic voltammetry. The highest current response of chloramphenicol was obtained with phosphate buffer, pH 6 (0.1 M) in 1% ethanol. The relationship between the concentration of chloramphenicol and the current response was linear over the range of 0.1-10 mM (R2=0.9990). The amount of chloramphenicol was analyzed by flow-injection analysis. A thin-layer flow cell equipped with a BDD electrode was used as an amperometric detector, and experiments were carried out at -0.7 V (vs. Ag/AgCl). The linear relationship between the current response and the concentration of chloramphenicol in the range of 0.1-50 microM (R2=0.9948) and the limit of detection of 0.03 microM (S/N=3) were obtained. This method has been successfully applied to the determination of chloramphenicol in sterile eye drops and milk sample by the standard addition method. The average recoveries of chloramphenicol in eye drops were 98.0%, and the average recoveries of chloramphenicol from spiked milk were 93.9-103%.

  12. A New Approach in Applying Systems Engineering Tools and Analysis to Determine Hepatocyte Toxicogenomics Risk Levels to Human Health.

    PubMed

    Gigrich, James; Sarkani, Shahryar; Holzer, Thomas

    2017-03-01

    There is an increasing backlog of potentially toxic compounds that cannot be evaluated with current animal-based approaches in a cost-effective and expeditious manner, thus putting human health at risk. Extrapolation of animal-based test results for human risk assessment often leads to different physiological outcomes. This article introduces the use of quantitative tools and methods from systems engineering to evaluate the risk of toxic compounds by the analysis of the amount of stress that human hepatocytes undergo in vitro when metabolizing GW7647 (1) over extended times and concentrations. Hepatocytes are exceedingly connected systems that make it challenging to understand the highly varied dimensional genomics data to determine risk of exposure. Gene expression data of peroxisome proliferator-activated receptor-α (PPARα) (2) binding was measured over multiple concentrations and varied times of GW7647 exposure and leveraging mahalanombis distance to establish toxicity threshold risk levels. The application of these novel systems engineering tools provides new insight into the intricate workings of human hepatocytes to determine risk threshold levels from exposure. This approach is beneficial to decision makers and scientists, and it can help reduce the backlog of untested chemical compounds due to the high cost and inefficiency of animal-based models.

  13. Economic impacts of bio-refinery and resource cascading systems: an applied general equilibrium analysis for Poland.

    PubMed

    Ignaciuk, Adriana M; Sanders, Johan

    2007-12-01

    Due to more stringent energy and climate policies, it is expected that many traditional chemicals will be replaced by their biomass-based substitutes, bio-chemicals. These innovations, however, can influence land allocation since the demand for land dedicated to specific crops might increase. Moreover, it can have an influence on traditional agricultural production. In this paper, we use an applied general equilibrium framework, in which we include two different bio-refinery processes and incorporate so-called cascading mechanisms. The bio-refinery processes use grass, as one of the major inputs, to produce bio-nylon and propane-diol (1,3PDO) to substitute currently produced fossil fuel-based nylon and ethane-diol. We examine the impact of specific climate policies on the bioelectricity share in total electricity production, land allocation, and production quantities and prices of selected commodities. The novel technologies become competitive, with an increased stringency of climate policies. This switch, however, does not induce a higher share of bioelectricity. The cascade does stimulate the production of bioelectricity, but it induces more of a shift in inputs in the bioelectricity sector (from biomass to the cascaded bio-nylon and 1, 3PDO) than an increase in production level of bioelectricity. We conclude that dedicated biomass crops will remain the main option for bioelectricity production: the contribution of the biomass systems remains limited. Moreover, the bioelectricity sector looses a competition for land for biomass production with bio-refineries.

  14. A mathematical method for the 3D analysis of rotating deformable systems applied on lumen-forming MDCK cell aggregates.

    PubMed

    Marmaras, Anastasios; Berge, Ulrich; Ferrari, Aldo; Kurtcuoglu, Vartan; Poulikakos, Dimos; Kroschewski, Ruth

    2010-04-01

    Cell motility contributes to the formation of organs and tissues, into which multiple cells self-organize. However such mammalian cellular motilities are not characterized in a quantitative manner and the systemic consequences are thus unknown. A mathematical tool to decipher cell motility, accounting for changes in cell shape, within a three-dimensional (3D) cell system was missing. We report here such a tool, usable on segmented images reporting the outline of clusters (cells) and allowing the time-resolved 3D analysis of circular motility of these as parts of a system (cell aggregate). Our method can analyze circular motility in sub-cellular, cellular, multi-cellular, and also non-cellular systems for which time-resolved segmented cluster outlines are available. To exemplify, we characterized the circular motility of lumen-initiating MDCK cell aggregates, embedded in extracellular matrix. We show that the organization of the major surrounding matrix fibers was not significantly affected during this cohort rotation. Using our developed tool, we discovered two classes of circular motion, rotation and random walk, organized in three behavior patterns during lumen initiation. As rotational movements were more rapid than random walk and as both could continue during lumen initiation, we conclude that neither the class nor the rate of motion regulates lumen initiation. We thus reveal a high degree of plasticity during a developmentally critical cell polarization step, indicating that lumen initiation is a robust process. However, motility rates decreased with increasing cell number, previously shown to correlate with epithelial polarization, suggesting that migratory polarization is converted into epithelial polarization during aggregate development. Copyright 2010 Wiley-Liss, Inc.

  15. Applied Information Systems Research Program Workshop

    NASA Technical Reports Server (NTRS)

    Bredekamp, Joe

    1991-01-01

    Viewgraphs on Applied Information Systems Research Program Workshop are presented. Topics covered include: the Earth Observing System Data and Information System; the planetary data system; Astrophysics Data System project review; OAET Computer Science and Data Systems Programs; the Center of Excellence in Space Data and Information Sciences; and CASIS background.

  16. Comparative analysis of the possibility of applying low-melting metals with the capillary-porous system in tokamak conditions

    NASA Astrophysics Data System (ADS)

    Lyublinski, I. E.; Vertkov, A. V.; Semenov, V. V.

    2016-12-01

    The use of capillary-porous systems (CPSs) with liquid Li, Ga, and Sn is considered as an alternative for solving the problem of creating plasma-facing elements (PFEs) of the fusion neutron source (FNS) and the DEMO-type reactor. The main advantages of CPSs with liquid metal compared with hard materials are their stability with respect to the degradation of properties in tokamak conditions and capability of surface self-restoration. The evaluation of applicability of liquid metals is performed on the basis of the analysis of their physical and chemical properties, the interaction with the tokamak plasma, and constructive and process features of in-vessel elements with CPSs implementing the application of these metals in a tokamak. It is shown that the upper limit of the PFE working temperature for all low-melting metals under consideration lies in the range of 550-600°C. The decisive factor for PFEs with Li is the limitation on the admissible atomic flux into plasma, while for those with Ga and Sn it is the corrosion resistance of construction materials. The upper limit of thermal loads in the steady-state operating mode for the considered promising PFE design with the use of Li, Ga, and Sn is close to 18-20 MW/m2. It is seen from the analysis that the use of metals with a low equilibrium vapor pressure of (Ga, Sn) gives no gain in extension of the region of admissible working temperatures of PFEs. However, with respect to the totality of properties, the possibility of implementing the self-restoration and stabilization effect of the liquid surface, the influence on the plasma discharge parameters, and the ability to protect the PFE surface in conditions of plasma perturbations and disruption, lithium is the most attractive liquid metal to create CPS-based PFEs for the tokamak.

  17. Comparative analysis of the possibility of applying low-melting metals with the capillary-porous system in tokamak conditions

    SciTech Connect

    Lyublinski, I. E. Vertkov, A. V. Semenov, V. V.

    2016-12-15

    The use of capillary-porous systems (CPSs) with liquid Li, Ga, and Sn is considered as an alternative for solving the problem of creating plasma-facing elements (PFEs) of the fusion neutron source (FNS) and the DEMO-type reactor. The main advantages of CPSs with liquid metal compared with hard materials are their stability with respect to the degradation of properties in tokamak conditions and capability of surface self-restoration. The evaluation of applicability of liquid metals is performed on the basis of the analysis of their physical and chemical properties, the interaction with the tokamak plasma, and constructive and process features of in-vessel elements with CPSs implementing the application of these metals in a tokamak. It is shown that the upper limit of the PFE working temperature for all low-melting metals under consideration lies in the range of 550–600°Ð¡. The decisive factor for PFEs with Li is the limitation on the admissible atomic flux into plasma, while for those with Ga and Sn it is the corrosion resistance of construction materials. The upper limit of thermal loads in the steady-state operating mode for the considered promising PFE design with the use of Li, Ga, and Sn is close to 18–20 MW/m{sup 2}. It is seen from the analysis that the use of metals with a low equilibrium vapor pressure of (Ga, Sn) gives no gain in extension of the region of admissible working temperatures of PFEs. However, with respect to the totality of properties, the possibility of implementing the self-restoration and stabilization effect of the liquid surface, the influence on the plasma discharge parameters, and the ability to protect the PFE surface in conditions of plasma perturbations and disruption, lithium is the most attractive liquid metal to create CPS-based PFEs for the tokamak.

  18. A Systematic Comprehensive Computational Model for Stake Estimation in Mission Assurance: Applying Cyber Security Econometrics System (CSES) to Mission Assurance Analysis Protocol (MAAP)

    SciTech Connect

    Abercrombie, Robert K; Sheldon, Frederick T; Grimaila, Michael R

    2010-01-01

    In earlier works, we presented a computational infrastructure that allows an analyst to estimate the security of a system in terms of the loss that each stakeholder stands to sustain as a result of security breakdowns. In this paper, we discuss how this infrastructure can be used in the subject domain of mission assurance as defined as the full life-cycle engineering process to identify and mitigate design, production, test, and field support deficiencies of mission success. We address the opportunity to apply the Cyberspace Security Econometrics System (CSES) to Carnegie Mellon University and Software Engineering Institute s Mission Assurance Analysis Protocol (MAAP) in this context.

  19. Reachability Analysis Applied to Space Situational Awareness

    NASA Astrophysics Data System (ADS)

    Holzinger, M.; Scheeres, D.

    Several existing and emerging applications of Space Situational Awareness (SSA) relate directly to spacecraft Rendezvous, Proximity Operations, and Docking (RPOD) and Formation / Cluster Flight (FCF). When multiple Resident Space Ob jects (RSOs) are in vicinity of one another with appreciable periods between observations, correlating new RSO tracks to previously known objects becomes a non-trivial problem. A particularly difficult sub-problem is seen when long breaks in observations are coupled with continuous, low- thrust maneuvers. Reachability theory, directly related to optimal control theory, can compute contiguous reachability sets for known or estimated control authority and can support such RSO search and correlation efforts in both ground and on-board settings. Reachability analysis can also directly estimate the minimum control authority of a given RSO. For RPOD and FCF applications, emerging mission concepts such as fractionation drastically increase system complexity of on-board autonomous fault management systems. Reachability theory, as applied to SSA in RPOD and FCF applications, can involve correlation of nearby RSO observations, control authority estimation, and sensor track re-acquisition. Additional uses of reachability analysis are formation reconfiguration, worst-case passive safety, and propulsion failure modes such as a "stuck" thruster. Existing reachability theory is applied to RPOD and FCF regimes. An optimal control policy is developed to maximize the reachability set and optimal control law discontinuities (switching) are examined. The Clohessy-Wiltshire linearized equations of motion are normalized to accentuate relative control authority for spacecraft propulsion systems at both Low Earth Orbit (LEO) and Geostationary Earth Orbit (GEO). Several examples with traditional and low thrust propulsion systems in LEO and GEO are explored to illustrate the effects of relative control authority on the time-varying reachability set surface. Both

  20. 2012 International Conference on Medical Physics and Biomedical Engineering Thermal Economic Analysis on LiBr Refrigeration -Heat Pump System Applied in CCHP System

    NASA Astrophysics Data System (ADS)

    Zhang, CuiZhen; Yang, Mo; Lu, Mei; Zhu, Jiaxian; Xu, Wendong

    LiBr refrigeration cooling water contains a lot of low-temperature heat source, can use this part of the heat source heat boiler feed water. This paper introduced LiBr refrigeration - heat pump system which recovery heat of the LiBr refrigeration cooling water by heat pump system to heat the feed water of boiler. Hot economic analysis on the system has been performed based on the experimental data. Results show that LiBr refrigeration-heat pump system brings 26.6 percent decrease in primary energy rate consumption comparing with the combined heat and power production system(CHP) and separate generation of cold;

  1. Stability Analysis of Finite Difference Approximations to Hyperbolic Systems,and Problems in Applied and Computational Matrix and Operator Theory

    DTIC Science & Technology

    1990-12-07

    Technology, Haifa, 1968. 45 10. Bhatia , Rajendra , Perturbation Bounds for Matrix Eigenvalues, Pitman Research Notes in Mathematics, H. Brezis, R. G. Douglas...is currently in press in the SIAM Journal of Matrix Analysis . The proof sheets have been corrected and returned to the editor during the period of this...appear. [3] R. A. Horn and C. R. Johnson, Matrix Analysis , Cambridge University Press, New York, 1985. 141 M. Marcus and N. Khan, A note on the

  2. Water-food-energy nexus index: analysis of water-energy-food nexus of crop's production system applying the indicators approach

    NASA Astrophysics Data System (ADS)

    El-Gafy, Inas

    2017-03-01

    Analysis the water-food-energy nexus is the first step to assess the decision maker in developing and evaluating national strategies that take into account the nexus. The main objective of the current research is providing a method for the decision makers to analysis the water-food-energy nexus of the crop production system at the national level and carrying out a quantitative assessment of it. Through the proposed method, indicators considering the water and energy consumption, mass productivity, and economic productivity were suggested. Based on these indicators a water-food-energy nexus index (WFENI) was performed. The study showed that the calculated WFENI of the Egyptian summer crops have scores that range from 0.21 to 0.79. Comparing to onion (the highest scoring WFENI,i.e., the best score), rice has the lowest WFENI among the summer food crops. Analysis of the water-food-energy nexus of forty-two Egyptian crops in year 2010 was caried out (energy consumed for irrigation represent 7.4% of the total energy footprint). WFENI can be applied to developed strategies for the optimal cropping pattern that minimizing the water and energy consumption and maximizing their productivity. It can be applied as a holistic tool to evaluate the progress in the water and agricultural national strategies. Moreover, WFENI could be applied yearly to evaluate the performance of the water-food-energy nexus managmant.

  3. The basic importance of applied behavior analysis

    PubMed Central

    Epling, W. Frank; Pierce, W. David

    1986-01-01

    We argue that applied behavior analysis is relevant to basic research. Modification studies, and a broad range of investigations that focus on the precipitating and maintaining conditions of socially significant human behavior, have basic importance. Applied behavior analysis may aid basic researchers in the design of externally valid experiments and thereby enhance the theoretical significance of basic research for understanding human behavior. Applied research with humans, directed at culturally-important problems, will help to propagate the science of human behavior. Such a science will also be furthered by analogue experiments that model socially important behavior. Analytical-applied studies and analogue experiments are forms of applied behavior analysis that could suggest new environment-behavior relationships. These relationships could lead to basic research and principles that further the prediction, control, and understanding of behavior. PMID:22478650

  4. System planning analysis applied to OTEC: initial cases by Florida Power Corporation. Task II report No. FC-5237-2

    SciTech Connect

    1980-03-01

    The objective of the task was to exercise the FPC system planning methodology on: (1) Base Case, 10 year generation expansion plan with coal plants providing base load expansion, and (2) same, but 400 MW of OTEC substituting for coal burning units with equal resultant system reliability. OTEC inputs were based on reasonable economic projections of direct capital cost and O and M costs for first-generation large commercial plants. OTEC inputs discussed in Section 2. The Base Case conditions for FPC system planning methodology involved base load coal fueled additions during the 1980's and early 1990's. The first trial runs of the PROMOD system planning model substituted OTEC for 400 MW purchases of coal generated power during 1988-1989 and then 400 MW coal capacity thereafter. Result showed higher system reliability than Base Case runs. Reruns with greater coal fueled capacity displacement showed that OTEC could substitute for 400 MW purchases in 1988-1989 and replace the 800 MW coal unit scheduled for 1990 to yield equivalent system reliability. However, a 1995 unit would need to be moved to 1994. Production costing computer model runs were used as input to Corporate Model to examine corporate financial impact. Present value of total revenue requirements were primary indication of relative competitiveness between Base Case and OTEC. Results show present value of total revenue requirements unfavorable to OTEC as compared to coal units. The disparity was in excess of the allowable range for possible consideration.

  5. Applied Behavior Analysis and Statistical Process Control?

    ERIC Educational Resources Information Center

    Hopkins, B. L.

    1995-01-01

    Incorporating statistical process control (SPC) methods into applied behavior analysis is discussed. It is claimed that SPC methods would likely reduce applied behavior analysts' intimate contacts with problems and would likely yield poor treatment and research decisions. Cases and data presented by Pfadt and Wheeler (1995) are cited as examples.…

  6. Applied Behavior Analysis and Statistical Process Control?

    ERIC Educational Resources Information Center

    Hopkins, B. L.

    1995-01-01

    Incorporating statistical process control (SPC) methods into applied behavior analysis is discussed. It is claimed that SPC methods would likely reduce applied behavior analysts' intimate contacts with problems and would likely yield poor treatment and research decisions. Cases and data presented by Pfadt and Wheeler (1995) are cited as examples.…

  7. On differentiation in applied behavior analysis

    PubMed Central

    Fawcett, Stephen B.

    1985-01-01

    Distinct types of activity in the field of applied behavior analysis are noted and discussed. Four metaphorical types of activity are considered: prospecting, farming, building, and guiding. Prospecting consists of time-limited exploration of a variety of beaviors, populations, or settings. Farming consists of producing new behaviors in the same setting using independent variables provided by the researchers or normally available in the setting. Building consists of combining procedural elements to create new programs or systems or to rehabilitate aspects of existing programs. Guiding involves pointing out connections between the principles of human behavior and the problems, populations, settings, and procedures with which researchers are (or could be) working. Advantages of each sphere are noted, and benefits of this division of labor to the field as a whole are discussed. PMID:22478631

  8. Applying Management Information Systems to Staffing.

    ERIC Educational Resources Information Center

    Hanson, Robert L.

    1982-01-01

    After reviewing some concepts and principles for effective data management, the author applies the concepts to nurse staffing systems for the management of human resources. He defines a seven-step process for establishing a management information system, from defining the management objective to implementing the system. (Author/CT)

  9. Applied Behavior Analysis in Flying Training Research.

    DTIC Science & Technology

    1980-01-01

    often referred to as behavior modification ) which promotes improvements in human learning through an analysis of the contingencies surrounding a...Company, in press. Bandura, A. Principles of behavior modification . New York: Holt, Rinehart & Winston, 1969. Bostow, D.E., & Bailey, J.S. Modification of...tutors for kindergarten children. Journal of Applied Behavior Analysis, 1974, 7, 223-232. Kazdin, A.E. Behavior modification in applied settings

  10. Numerical analysis of the shifting slabs applied in a wireless power transfer system to enhance magnetic coupling

    NASA Astrophysics Data System (ADS)

    Dong, Yayun; Yang, Xijun; Jin, Nan; Li, Wenwen; Yao, Chen; Tang, Houjun

    2017-05-01

    Shifting medium is a kind of metamaterial, which can optically shift a space or an object a certain distance away from its original position. Based on the shifting medium, we propose a concise pair of shifting slabs covering the transmitting or receiving coil in a two-coil wireless power transfer system to decrease the equivalent distance between the coils. The electromagnetic parameters of the shifting slabs are calculated by transformation optics. Numerical simulations validate that the shifting slabs can approximately shift the electromagnetic fields generated by the covered coil; thus, the magnetic coupling and the efficiency of the system are enhanced while remaining the physical transmission distance unchanged. We also verify the advantages of the shifting slabs over the magnetic superlens. Finally, we provide two methods to fabricate shifting slabs based on split-ring resonators.

  11. Hydrogeochemical variables regionalization--applying cluster analysis for a seasonal evolution model from an estuarine system affected by AMD.

    PubMed

    Grande, J A; Carro, B; Borrego, J; de la Torre, M L; Valente, T; Santisteban, M

    2013-04-15

    This study describes the spatial evolution of the hydrogeochemical parameters which characterise a strongly affected estuary by Acid Mine Drainage (AMD). The studied estuarine system receives AMD from the Iberian Pyrite Belt (SW Spain) and, simultaneously, is affected by the presence of an industrial chemical complex. Water sampling was performed in the year of 2008, comprising four sampling campaigns, in order to represent seasonality. The results show how the estuary can be divided into three areas of different behaviour in response to hydrogeochemical variables concentrations that define each sampling stations: on one hand, an area dominated by tidal influence; in the opposite end there is a second area including the points located in the two rivers headwaters that are not influenced by seawater; finally there is the area that can be defined as mixing zone. These areas are moved along the hydrological year due to seasonal chemical variations.

  12. Caldwell University's Department of Applied Behavior Analysis.

    PubMed

    Reeve, Kenneth F; Reeve, Sharon A

    2016-05-01

    Since 2004, faculty members at Caldwell University have developed three successful graduate programs in Applied Behavior Analysis (i.e., PhD, MA, non-degree programs), increased program faculty from two to six members, developed and operated an on-campus autism center, and begun a stand-alone Applied Behavior Analysis Department. This paper outlines a number of strategies used to advance these initiatives, including those associated with an extensive public relations campaign. We also outline challenges that have limited our programs' growth. These strategies, along with a consideration of potential challenges, might prove useful in guiding academicians who are interested in starting their own programs in behavior analysis.

  13. Thermodynamic Laws Applied to Economic Systems

    ERIC Educational Resources Information Center

    González, José Villacís

    2009-01-01

    Economic activity in its different manifestations--production, exchange, consumption and, particularly, information on quantities and prices--generates and transfers energy. As a result, we can apply to it the basic laws of thermodynamics. These laws are applicable within a system, i.e., in a country or between systems and countries. To these…

  14. System Applies Polymer Powder To Filament Tow

    NASA Technical Reports Server (NTRS)

    Baucom, Robert M.; Snoha, John J.; Marchello, Joseph M.

    1993-01-01

    Polymer powder applied uniformly and in continuous manner. Powder-coating system applies dry polymer powder to continuous fiber tow. Unique filament-spreading technique, combined with precise control of tension on fibers in system, ensures uniform application of polymer powder to web of spread filaments. Fiber tows impregnated with dry polymer powders ("towpregs") produced for preform-weaving and composite-material-molding applications. System and process valuable to prepreg industry, for production of flexible filament-windable tows and high-temperature polymer prepregs.

  15. Defining applied behavior analysis: An historical analogy

    PubMed Central

    Deitz, Samuel M.

    1982-01-01

    This article examines two criteria for a definition of applied behavior analysis. The criteria are derived from a 19th century attempt to establish medicine as a scientific field. The first criterion, experimental determinism, specifies the methodological boundaries of an experimental science. The second criterion, philosophic doubt, clarifies the tentative nature of facts and theories derived from those facts. Practices which will advance the science of behavior are commented upon within each criteria. To conclude, the problems of a 19th century form of empiricism in medicine are related to current practices in applied behavior analysis. PMID:22478557

  16. Goals Analysis Procedure Guidelines for Applying the Goals Analysis Process

    NASA Technical Reports Server (NTRS)

    Motley, Albert E., III

    2000-01-01

    One of the key elements to successful project management is the establishment of the "right set of requirements", requirements that reflect the true customer needs and are consistent with the strategic goals and objectives of the participating organizations. A viable set of requirements implies that each individual requirement is a necessary element in satisfying the stated goals and that the entire set of requirements, taken as a whole, is sufficient to satisfy the stated goals. Unfortunately, it is the author's experience that during project formulation phases' many of the Systems Engineering customers do not conduct a rigorous analysis of the goals and objectives that drive the system requirements. As a result, the Systems Engineer is often provided with requirements that are vague, incomplete, and internally inconsistent. To complicate matters, most systems development methodologies assume that the customer provides unambiguous, comprehensive and concise requirements. This paper describes the specific steps of a Goals Analysis process applied by Systems Engineers at the NASA Langley Research Center during the formulation of requirements for research projects. The objective of Goals Analysis is to identify and explore all of the influencing factors that ultimately drive the system's requirements.

  17. Applied Behavior Analysis as Technological Science.

    ERIC Educational Resources Information Center

    Iwata, Brian A.

    1991-01-01

    To the extent that applied behavior analysis represents a scientific and practical approach to the study of behavior, its technological character is essential. The most serious problem evident in the field is not that the research being done is too technical but that more good research of all types is needed. (JDD)

  18. Applied Behavior Analysis: Beyond Discrete Trial Teaching

    ERIC Educational Resources Information Center

    Steege, Mark W.; Mace, F. Charles; Perry, Lora; Longenecker, Harold

    2007-01-01

    We discuss the problem of autism-specific special education programs representing themselves as Applied Behavior Analysis (ABA) programs when the only ABA intervention employed is Discrete Trial Teaching (DTT), and often for limited portions of the school day. Although DTT has many advantages to recommend its use, it is not well suited to teach…

  19. Applied Behavior Analysis: Beyond Discrete Trial Teaching

    ERIC Educational Resources Information Center

    Steege, Mark W.; Mace, F. Charles; Perry, Lora; Longenecker, Harold

    2007-01-01

    We discuss the problem of autism-specific special education programs representing themselves as Applied Behavior Analysis (ABA) programs when the only ABA intervention employed is Discrete Trial Teaching (DTT), and often for limited portions of the school day. Although DTT has many advantages to recommend its use, it is not well suited to teach…

  20. Positive Behavior Support and Applied Behavior Analysis

    ERIC Educational Resources Information Center

    Johnston, J. M.; Foxx, R. M.; Jacobson, J. W.; Green, G.; Mulick, J. A.

    2006-01-01

    This article reviews the origins and characteristics of the positive behavior support (PBS) movement and examines those features in the context of the field of applied behavior analysis (ABA). We raise a number of concerns about PBS as an approach to delivery of behavioral services and its impact on how ABA is viewed by those in human services. We…

  1. Health system barriers and enablers to early access to breast cancer screening, detection, and diagnosis: a global analysis applied to the MENA region.

    PubMed

    Bowser, D; Marqusee, H; El Koussa, M; Atun, R

    2017-08-24

    To identify barriers and enablers that impact access to early screening, detection, and diagnosis of breast cancer both globally and more specifically in the Middle East and North Africa (MENA) region (with a specific focus on Egypt, Jordan, Oman, Saudi Arabia, United Arab Emirates [UAE], and Kuwait) with a specific focus on the health system. A systematic review of literature. We conducted a systematic reviewing using the PRISMA methodology. We searched PubMed, Global Index Medicus, and EMBASE for studies on 'breast cancer', 'breast neoplasm,' or 'screening, early detection, and early diagnosis' as well as key words related to the following barriers: religion, culture, health literacy, lack of knowledge/awareness/understanding, attitudes, fatalism/fear, shame/embarrassment, and physician gender from January 1, 2000 until September 1, 2016. Two independent reviewers screened both titles and abstracts. The application of inclusion and exclusion criteria yielded a final list of articles. A conceptual framework was used to guide the thematic analysis and examine health system barriers and enablers to breast cancer screening at the broader macro health system level, at the health provider level, and the individual level. The analysis was conducted globally and in the MENA region. A total of 11,936 references were identified through the initial search strategy, of which 55 were included in the final thematic analysis. The results found the following barriers and enablers to access to breast cancer screening at the health system level, the health provider level, and the individual level: health system structures such as health insurance and care coordination systems, costs, time concerns, provider characteristics including gender of the provider, quality of care issues, medical concerns, and fear. In addition, the following seven barriers and enablers were identified at the health system or provider level as significantly impacting screening for breast cancer: (1) access

  2. Applying WebMining on KM system

    NASA Astrophysics Data System (ADS)

    Shimazu, Keiko; Ozaki, Tomonobu; Furukawa, Koichi

    KM (Knowledge Management) systems have recently been adopted within the realm of enterprise management. On the other hand, data mining technology is widely acknowledged within Information systems' R&D Divisions. Specially, acquisition of meaningful information from Web usage data has become one of the most exciting eras. In this paper, we employ a Web based KM system and propose a framework for applying Web Usage Mining technology to KM data. As it turns out, task duration varies according to different user operations such as referencing a table-of-contents page, down-loading a target file, and writing to a bulletin board. This in turn makes it possible to easily predict the purpose of the user's task. By taking these observations into account, we segmented access log data manually. These results were compared with results abstained by applying the constant interval method. Next, we obtained a segmentation rule of Web access logs by applying a machine-learning algorithm to manually segmented access logs as training data. Then, the newly obtained segmentation rule was compared with other known methods including the time interval method by evaluating their segmentation results in terms of recall and precision rates and it was shown that our rule attained the best results in both measures. Furthermore, the segmented data were fed to an association rule miner and the obtained association rules were utilized to modify the Web structure.

  3. Reachability Analysis Applied to Space Situational Awareness

    DTIC Science & Technology

    2009-09-01

    12:207-242. [15] L. S. Breger, G. Inalhan, M. Tillerson, J. P. How, “Cooperative spacecraft Formation Flying: Model Predictive Control With Open And...applying them to the nonlinear relative orbit equations of motion, which are appropriate both for general SSA and spacecraft proximity operations... Nonlinear System (RNS). This assumption does not restrict the scope of these results in the context of SSA, as in orbital scenarios control and

  4. Spectral Selectivity Applied To Hybrid Concentration Systems

    NASA Astrophysics Data System (ADS)

    Hamdy, M. A.; Luttmann, F.; Osborn, D. E.; Jacobson, M. R.; MacLeod, H. A.

    1985-12-01

    The efficiency of conversion of concentrated solar energy can be improved by separating the solar spectrum into portions matched to specific photoquantum processes and the balance used for photothermal conversion. The basic approaches of spectrally selective beam splitters are presented. A detailed simulation analysis using TRNSYS is developed for a spectrally selective hybrid photovoltaic/photothermal concentrating system. The analysis shows definite benefits to a spectrally selective approach.

  5. Applied Information Systems Research Program Workshop

    NASA Technical Reports Server (NTRS)

    1991-01-01

    The first Applied Information Systems Research Program (AISRP) Workshop provided the impetus for several groups involved in information systems to review current activities. The objectives of the workshop included: (1) to provide an open forum for interaction and discussion of information systems; (2) to promote understanding by initiating a dialogue with the intended benefactors of the program, the scientific user community, and discuss options for improving their support; (3) create an advocacy in having science users and investigators of the program meet together and establish the basis for direction and growth; and (4) support the future of the program by building collaborations and interaction to encourage an investigator working group approach for conducting the program.

  6. The Applied Mathematics for Power Systems (AMPS)

    SciTech Connect

    Chertkov, Michael

    2012-07-24

    Increased deployment of new technologies, e.g., renewable generation and electric vehicles, is rapidly transforming electrical power networks by crossing previously distinct spatiotemporal scales and invalidating many traditional approaches for designing, analyzing, and operating power grids. This trend is expected to accelerate over the coming years, bringing the disruptive challenge of complexity, but also opportunities to deliver unprecedented efficiency and reliability. Our Applied Mathematics for Power Systems (AMPS) Center will discover, enable, and solve emerging mathematics challenges arising in power systems and, more generally, in complex engineered networks. We will develop foundational applied mathematics resulting in rigorous algorithms and simulation toolboxes for modern and future engineered networks. The AMPS Center deconstruction/reconstruction approach 'deconstructs' complex networks into sub-problems within non-separable spatiotemporal scales, a missing step in 20th century modeling of engineered networks. These sub-problems are addressed within the appropriate AMPS foundational pillar - complex systems, control theory, and optimization theory - and merged or 'reconstructed' at their boundaries into more general mathematical descriptions of complex engineered networks where important new questions are formulated and attacked. These two steps, iterated multiple times, will bridge the growing chasm between the legacy power grid and its future as a complex engineered network.

  7. Performance Monitoring Applied to System Supervision.

    PubMed

    Somon, Bertille; Campagne, Aurélie; Delorme, Arnaud; Berberian, Bruno

    2017-01-01

    Nowadays, automation is present in every aspect of our daily life and has some benefits. Nonetheless, empirical data suggest that traditional automation has many negative performance and safety consequences as it changed task performers into task supervisors. In this context, we propose to use recent insights into the anatomical and neurophysiological substrates of action monitoring in humans, to help further characterize performance monitoring during system supervision. Error monitoring is critical for humans to learn from the consequences of their actions. A wide variety of studies have shown that the error monitoring system is involved not only in our own errors, but also in the errors of others. We hypothesize that the neurobiological correlates of the self-performance monitoring activity can be applied to system supervision. At a larger scale, a better understanding of system supervision may allow its negative effects to be anticipated or even countered. This review is divided into three main parts. First, we assess the neurophysiological correlates of self-performance monitoring and their characteristics during error execution. Then, we extend these results to include performance monitoring and error observation of others or of systems. Finally, we provide further directions in the study of system supervision and assess the limits preventing us from studying a well-known phenomenon: the Out-Of-the-Loop (OOL) performance problem.

  8. Wavelet analysis applied to the IRAS cirrus

    NASA Technical Reports Server (NTRS)

    Langer, William D.; Wilson, Robert W.; Anderson, Charles H.

    1994-01-01

    The structure of infrared cirrus clouds is analyzed with Laplacian pyramid transforms, a form of non-orthogonal wavelets. Pyramid and wavelet transforms provide a means to decompose images into their spatial frequency components such that all spatial scales are treated in an equivalent manner. The multiscale transform analysis is applied to IRAS 100 micrometer maps of cirrus emission in the north Galactic pole region to extract features on different scales. In the maps we identify filaments, fragments and clumps by separating all connected regions. These structures are analyzed with respect to their Hausdorff dimension for evidence of the scaling relationships in the cirrus clouds.

  9. Positive Behavior Support and Applied Behavior Analysis

    PubMed Central

    Johnston, J.M; Foxx, Richard M; Jacobson, John W; Green, Gina; Mulick, James A

    2006-01-01

    This article reviews the origins and characteristics of the positive behavior support (PBS) movement and examines those features in the context of the field of applied behavior analysis (ABA). We raise a number of concerns about PBS as an approach to delivery of behavioral services and its impact on how ABA is viewed by those in human services. We also consider the features of PBS that have facilitated its broad dissemination and how ABA might benefit from emulating certain practices of the PBS movement. PMID:22478452

  10. A survey on fuzzy theory applied in geographic information system

    NASA Astrophysics Data System (ADS)

    Wu, Feiquan; Cui, Weihong; Chen, Houwu

    2009-10-01

    The real world is an infinite complex and very huge systems, the phenomenon and processes in this world have many complex relations among them. It is consecutive and cannot be treated as a determined one as the traditional geographic information system does; What's more, the uncertainty and fuzziness exist in every stage of data processing of GIS, from data collection, data storage to data analysis etc, so it is very meaningful to apply fuzzy theory in GIS for its ability to handle fuzziness and uncertainty of spatial data. The paper talks about the current situation of fuzzy theory applied in GIS, including the classification of application fields, its main methods, principles etc. The detailed fields we concerned include spatial object modeling, spatial reasoning, spatial analysis, spatial data mining, and reliability analysis of GIS data and so on. Furthermore, we put forward some development foregrounds and research orientations of fuzzy theory applied in GIS.

  11. A novel mating system analysis for modes of self-oriented mating applied to diploid and polyploid arctic Easter daisies (Townsendia hookeri).

    PubMed

    Thompson, S L; Ritland, K

    2006-08-01

    We have developed a new model for mating system analysis, which attempts to distinguish among alternative modes of self-oriented mating within populations. This model jointly estimates the rates of outcrossing, selfing, automixis and apomixis, through the use of information in the family structure given by dominant genetic marker data. The method is presented, its statistical properties evaluated, and is applied to three arctic Easter daisy populations, one consisting of diploids, the other two of tetraploids. The tetraploids are predominantly male sterile and reported to be apomictic while the diploids are male fertile. In each Easter daisy population, 10 maternal arrays of six progeny were assayed for amplified fragment length polymorphism markers. Estimates, confirmed with likelihood ratio tests of mating hypotheses, showed apomixis to be predominant in all populations (ca. 70%), but selfing or automixis was moderate (ca. 25%) in tetraploids. It was difficult to distinguish selfing from automixis, and simulations confirm that with even very large sample sizes, the estimates have a very strong negative statistical correlation, for example, they are not independent. No selfing or automixis was apparent in the diploid population, instead, moderate levels of outcrossing were detected (23%). Low but significant levels of outcrossing (2-4%) seemed to occur in the male-sterile tetraploid populations; this may be due to genotyping error of this level. Overall, this study shows apomixis can be partial, and provides evidence for higher levels of inbreeding in polyploids compared to diploids and for significant levels of apomixis in a diploid plant population.

  12. Artificial intelligence applied to process signal analysis

    NASA Technical Reports Server (NTRS)

    Corsberg, Dan

    1988-01-01

    Many space station processes are highly complex systems subject to sudden, major transients. In any complex process control system, a critical aspect of the human/machine interface is the analysis and display of process information. Human operators can be overwhelmed by large clusters of alarms that inhibit their ability to diagnose and respond to a disturbance. Using artificial intelligence techniques and a knowledge base approach to this problem, the power of the computer can be used to filter and analyze plant sensor data. This will provide operators with a better description of the process state. Once a process state is recognized, automatic action could be initiated and proper system response monitored.

  13. Artificial intelligence applied to process signal analysis

    NASA Technical Reports Server (NTRS)

    Corsberg, Dan

    1988-01-01

    Many space station processes are highly complex systems subject to sudden, major transients. In any complex process control system, a critical aspect of the human/machine interface is the analysis and display of process information. Human operators can be overwhelmed by large clusters of alarms that inhibit their ability to diagnose and respond to a disturbance. Using artificial intelligence techniques and a knowledge base approach to this problem, the power of the computer can be used to filter and analyze plant sensor data. This will provide operators with a better description of the process state. Once a process state is recognized, automatic action could be initiated and proper system response monitored.

  14. High Efficiency Micromachining System Applied in Nanolithography

    NASA Astrophysics Data System (ADS)

    Chen, Xing; Lee, Dong Weon; Choi, Young Soo

    Scanning probe lithography such as direct-writing lithographic processes and nanoscratching techniques based on scanning probe microscopy have presented new micromachining methods for microelectromechanical system (MEMS). In this paper, a micromachining system for thermal scanning probe lithography is introduced, which consists of the cantilever arrays and a big stroke micro XY-stage. A large machining area and high machining speed can be realized by combining arrays of cantilevers possessing sharp tips at their top with the novel micro XY-stage which can obtain big displacements under relatively low driving voltage and in a small size. According to the above configuration, this micromachining system is provided with high throughputs and suitable for industrialization due to its MEMS-based simple fabrication process. The novel micro XY-stage applied in this system is presented in detail including the unique structure and principles, which shows an obvious improvement and distinct advantages in comparison with traditional structures. It is analyzed by mathematical model and then simulated using finite element method (FEM), it is proved to be able to practically serve the micromachining system with high capability.

  15. Tribological systems as applied to aircraft engines

    NASA Technical Reports Server (NTRS)

    Buckley, D. H.

    1985-01-01

    Tribological systems as applied to aircraft are reviewed. The importance of understanding the fundamental concepts involved in such systems is discussed. Basic properties of materials which can be related to adhesion, friction and wear are presented and correlated with tribology. Surface processes including deposition and treatment are addressed in relation to their present and future application to aircraft components such as bearings, gears and seals. Lubrication of components with both liquids and solids is discussed. Advances in both new liquid molecular structures and additives for those structures are reviewed and related to the needs of advanced engines. Solids and polymer composites are suggested for increasing use and ceramic coatings containing fluoride compounds are offered for the extreme temperatures encountered in such components as advanced bearings and seals.

  16. Applied spectrophotometry: analysis of a biochemical mixture.

    PubMed

    Trumbo, Toni A; Schultz, Emeric; Borland, Michael G; Pugh, Michael Eugene

    2013-01-01

    Spectrophotometric analysis is essential for determining biomolecule concentration of a solution and is employed ubiquitously in biochemistry and molecular biology. The application of the Beer-Lambert-Bouguer Lawis routinely used to determine the concentration of DNA, RNA or protein. There is however a significant difference in determining the concentration of a given species (RNA, DNA, protein) in isolation (a contrived circumstance) as opposed to determining that concentration in the presence of other species (a more realistic situation). To present the student with a more realistic laboratory experience and also to fill a hole that we believe exists in student experience prior to reaching a biochemistry course, we have devised a three week laboratory experience designed so that students learn to: connect laboratory practice with theory, apply the Beer-Lambert-Bougert Law to biochemical analyses, demonstrate the utility and limitations of example quantitative colorimetric assays, demonstrate the utility and limitations of UV analyses for biomolecules, develop strategies for analysis of a solution of unknown biomolecular composition, use digital micropipettors to make accurate and precise measurements, and apply graphing software. Copyright © 2013 Wiley Periodicals, Inc.

  17. Applied surface analysis in magnetic storage technology

    NASA Astrophysics Data System (ADS)

    Windeln, Johannes; Bram, Christian; Eckes, Heinz-Ludwig; Hammel, Dirk; Huth, Johanna; Marien, Jan; Röhl, Holger; Schug, Christoph; Wahl, Michael; Wienss, Andreas

    2001-07-01

    This paper gives a synopsis of today's challenges and requirements for a surface analysis and materials science laboratory with a special focus on magnetic recording technology. The critical magnetic recording components, i.e. the protective carbon overcoat (COC), the disk layer structure, the read/write head including the giant-magnetoresistive (GMR) sensor, are described and options for their characterization with specific surface and structure analysis techniques are given. For COC investigations, applications of Raman spectroscopy to the structural analysis and determination of thickness, hydrogen and nitrogen content are discussed. Hardness measurements by atomic force microscopy (AFM) scratching techniques are presented. Surface adsorption phenomena on disk substrates or finished disks are characterized by contact angle analysis or so-called piezo-electric mass adsorption systems (PEMAS), also known as quartz crystal microbalance (QCM). A quickly growing field of applications is listed for various X-ray analysis techniques, such as disk magnetic layer texture analysis for X-ray diffraction, compositional characterization via X-ray fluorescence, compositional analysis with high lateral resolution via electron microprobe analysis. X-ray reflectometry (XRR) has become a standard method for the absolute measurement of individual layer thicknesses contained in multi-layer stacks and thus, is the successor of ellipsometry for this application. Due to the ongoing reduction of critical feature sizes, the analytical challenges in terms of lateral resolution, sensitivity limits and dedicated nano-preparation have been consistently growing and can only be met by state-of-the-art Auger electron spectrometers (AES), transmission electron microscopy (TEM) analysis, time-of-flight-secondary ion mass spectroscopy (ToF-SIMS) characterization, focused ion beam (FIB) sectioning and TEM lamella preparation via FIB. The depth profiling of GMR sensor full stacks was significantly

  18. Tissue Microarray Analysis Applied to Bone Diagenesis

    PubMed Central

    Mello, Rafael Barrios; Silva, Maria Regina Regis; Alves, Maria Teresa Seixas; Evison, Martin Paul; Guimarães, Marco Aurelio; Francisco, Rafaella Arrabaca; Astolphi, Rafael Dias; Iwamura, Edna Sadayo Miazato

    2017-01-01

    Taphonomic processes affecting bone post mortem are important in forensic, archaeological and palaeontological investigations. In this study, the application of tissue microarray (TMA) analysis to a sample of femoral bone specimens from 20 exhumed individuals of known period of burial and age at death is described. TMA allows multiplexing of subsamples, permitting standardized comparative analysis of adjacent sections in 3-D and of representative cross-sections of a large number of specimens. Standard hematoxylin and eosin, periodic acid-Schiff and silver methenamine, and picrosirius red staining, and CD31 and CD34 immunohistochemistry were applied to TMA sections. Osteocyte and osteocyte lacuna counts, percent bone matrix loss, and fungal spheroid element counts could be measured and collagen fibre bundles observed in all specimens. Decalcification with 7% nitric acid proceeded more rapidly than with 0.5 M EDTA and may offer better preservation of histological and cellular structure. No endothelial cells could be detected using CD31 and CD34 immunohistochemistry. Correlation between osteocytes per lacuna and age at death may reflect reported age-related responses to microdamage. Methodological limitations and caveats, and results of the TMA analysis of post mortem diagenesis in bone are discussed, and implications for DNA survival and recovery considered. PMID:28051148

  19. [Systemic reactions to orally applied metal alloys].

    PubMed

    Feilzer, A J; Kleverlaan, C J; Prahl, C; Muris, J

    2013-06-01

    Orally applied metal alloys can cause undesirable physical effects. A distinction needs to be made in this respect between local and systemic reactions and toxic and immunological reactions. A case is presented which illustrates this problem. In this case, the application of orthodontic appliances was probably the trigger for an exacerbation of nickel allergy. The oral exposure to nickel resulted in hand eczema. The patient was also exposed to nickel by single-unit fixed dental prostheses, a removable dental prosthesis, and food, as a result of which removal of the orthodontic appliances did not result in complete healing. Therefore, the single-unit fixed dental prostheses also had to be removed and food had to be prepared henceforward in nickel free pans.

  20. Computing and Systems Applied in Support of Coordinated ...

    EPA Pesticide Factsheets

    This talk focuses on how Dr. Loughlin is applying Computing and Systems models, tools and methods to more fully understand the linkages among energy systems, environmental quality, and climate change. Dr. Loughlin will highlight recent and ongoing research activities, including: applying sensitivity analysis to assess the impacts of clean energy technologies, conducting scenario analysis to explore the efficacy of environmental regulations under deep uncertainty, and developing decision support systems that allow analysts and decision-makers to examine state-level climate actions. Dr. Loughlin will conclude with a brief discussion of the lessons learned over the first half of his career. Dr. Loughlin has been invited to give the keynote talk at the 1st Annual Computing and Systems Graduate Research Symposium, sponsored by the Department of Civil, Construction and Environmental Engineering at North Carolina State University.

  1. Use of an automated chromium reduction system for hydrogen isotope ratio analysis of physiological fluids applied to doubly labeled water analysis.

    PubMed

    Schoeller, D A; Colligan, A S; Shriver, T; Avak, H; Bartok-Olson, C

    2000-09-01

    The doubly labeled water method is commonly used to measure total energy expenditure in free-living subjects. The method, however, requires accurate and precise deuterium abundance determinations, which can be laborious. The aim of this study was to evaluate a fully automated, high-throughput, chromium reduction technique for the measurement of deuterium abundances in physiological fluids. The chromium technique was compared with an off-line zinc bomb reduction technique and also subjected to test-retest analysis. Analysis of international water standards demonstrated that the chromium technique was accurate and had a within-day precision of <1 per thousand. Addition of organic matter to water samples demonstrated that the technique was sensitive to interference at levels between 2 and 5 g l(-1). Physiological samples could be analyzed without this interference, plasma by 10000 Da exclusion filtration, saliva by sedimentation and urine by decolorizing with carbon black. Chromium reduction of urine specimens from doubly labeled water studies indicated no bias relative to zinc reduction with a mean difference in calculated energy expenditure of -0.2 +/- 3.9%. Blinded reanalysis of urine specimens from a second doubly labeled water study demonstrated a test-retest coefficient of variation of 4%. The chromium reduction method was found to be a rapid, accurate and precise method for the analysis of urine specimens from doubly labeled water. Copyright 2000 John Wiley & Sons, Ltd.

  2. Overview of MSFC's Applied Fluid Dynamics Analysis Group Activities

    NASA Technical Reports Server (NTRS)

    Garcia, Roberto; Griffin, Lisa; Williams, Robert

    2002-01-01

    This viewgraph report presents an overview of activities and accomplishments of NASA's Marshall Space Flight Center's Applied Fluid Dynamics Analysis Group. Expertise in this group focuses on high-fidelity fluids design and analysis with application to space shuttle propulsion and next generation launch technologies. Topics covered include: computational fluid dynamics research and goals, turbomachinery research and activities, nozzle research and activities, combustion devices, engine systems, MDA development and CFD process improvements.

  3. Cluster analysis applied to multiparameter geophysical dataset

    NASA Astrophysics Data System (ADS)

    Di Giuseppe, M. G.; Troiano, A.; Troise, C.; De Natale, G.

    2012-04-01

    Multi-parameter acquisition is a common geophysical field practice nowadays. Regularly seismic velocity and attenuation, gravity and electromagnetic dataset are acquired in a certain area, to obtain a complete characterization of the some investigate feature of the subsoil. Such a richness of information is often underestimated, although an integration of the analysis could provide a notable improving in the imaging of the investigated structures, mostly because the handling of distinct parameters and their joint inversion still presents several and severe problems. Post-inversion statistical techniques represent a promising approach to these questions, providing a quick, simple and elegant way to obtain this advantageous but complex integration. We present an approach based on the partition of the analyzed multi parameter dataset in a number of different classes, identified as localized regions of high correlation. These classes, or 'Cluster', are structured in such a way that the observations pertaining to a certain group are more similar to each other than the observations belonging to a different one, according to an optimal logical criterion. Regions of the subsoil sharing the same physical characteristic are so identified, without a-priori or empirical relationship linking the distinct measured parameters. The retrieved imaging results highly affordable in a statistical sense, specifically due to this lack of external hypothesis that are, instead, indispensable in a full joint inversion, were works, as matter of fact, just a real constrain for the inversion process, not seldom of relative consistence. We apply our procedure to a certain number of experimental dataset, related to several structures at very different scales presents in the Campanian district (southern Italy). These structures goes from the shallows evidence of the active fault zone originating the M 7.9 Irpinia earthquake to the main feature characterizing the Campi Flegrei Caldera and the Mt

  4. System safety as applied to Skylab

    NASA Technical Reports Server (NTRS)

    Kleinknecht, K. S.; Miller, B. J.

    1974-01-01

    Procedural and organizational guidelines used in accordance with NASA safety policy for the Skylab missions are outlined. The basic areas examined in the safety program for Skylab were the crew interface, extra-vehicular activity (EVA), energy sources, spacecraft interface, and hardware complexity. Fire prevention was a primary goal, with firefighting as backup. Studies of the vectorcardiogram and sleep monitoring experiments exemplify special efforts to prevent fire and shock. The final fire control study included material review, fire detection capability, and fire extinguishing capability. Contractors had major responsibility for system safety. Failure mode and effects analysis (FMEA) and equipment criticality categories are outlined. Redundancy was provided on systems that were critical to crew survival (category I). The five key checkpoints in Skylab hardware development are explained. Skylab rescue capability was demonstrated by preparations to rescue the Skylab 3 crew after their spacecraft developed attitude control problems.

  5. Analysis of the interaction between experimental and applied behavior analysis.

    PubMed

    Virues-Ortega, Javier; Hurtado-Parrado, Camilo; Cox, Alison D; Pear, Joseph J

    2014-01-01

    To study the influences between basic and applied research in behavior analysis, we analyzed the coauthorship interactions of authors who published in JABA and JEAB from 1980 to 2010. We paid particular attention to authors who published in both JABA and JEAB (dual authors) as potential agents of cross-field interactions. We present a comprehensive analysis of dual authors' coauthorship interactions using social networks methodology and key word analysis. The number of dual authors more than doubled (26 to 67) and their productivity tripled (7% to 26% of JABA and JEAB articles) between 1980 and 2010. Dual authors stood out in terms of number of collaborators, number of publications, and ability to interact with multiple groups within the field. The steady increase in JEAB and JABA interactions through coauthors and the increasing range of topics covered by dual authors provide a basis for optimism regarding the progressive integration of basic and applied behavior analysis. © Society for the Experimental Analysis of Behavior.

  6. Finite element analysis applied to cornea reshaping.

    PubMed

    Cabrera Fernández, Delia; Niazy, A M; Kurtz, R M; Djotyan, G P; Juhasz, T

    2005-01-01

    A 2-D finite element model of the cornea is developed to simulate corneal reshaping and the resulting deformation induced by refractive surgery. In the numerical simulations, linear and nonlinear elastic models are applied when stiffness inhomogeneities varying with depth are considered. Multiple simulations are created that employ different geometric configurations for the removal of the corneal tissue. Side-by-side comparisons of the different constitutive laws are also performed. To facilitate the comparison, the material property constants are identified from the same experimental data, which are obtained from mechanical tests on corneal strips and membrane inflation experiments. We then validate the resulting models by comparing computed refractive power changes with clinical results. Tissue deformations created by simulated corneal tissue removal using finite elements are consistent with clinically observed postsurgical results. The model developed provides a much more predictable refractive outcome when the stiffness inhomogeneities of the cornea and nonlinearities of the deformations are included in the simulations. Finite element analysis is a useful tool for modeling surgical effects on the cornea and developing a better understanding of the biomechanics of the cornea. The creation of patient-specific simulations would allow surgical outcomes to be predicted based on individualized finite element models.

  7. Magnetic Analysis Techniques Applied to Desert Varnish

    NASA Technical Reports Server (NTRS)

    Schmidgall, E. R.; Moskowitz, B. M.; Dahlberg, E. D.; Kuhlman, K. R.

    2003-01-01

    Desert varnish is a black or reddish coating commonly found on rock samples from arid regions. Typically, the coating is very thin, less than half a millimeter thick. Previous research has shown that the primary components of desert varnish are silicon oxide clay minerals (60%), manganese and iron oxides (20-30%), and trace amounts of other compounds [1]. Desert varnish is thought to originate when windborne particles containing iron and manganese oxides are deposited onto rock surfaces where manganese oxidizing bacteria concentrate the manganese and form the varnish [4,5]. If desert varnish is indeed biogenic, then the presence of desert varnish on rock surfaces could serve as a biomarker, indicating the presence of microorganisms. This idea has considerable appeal, especially for Martian exploration [6]. Magnetic analysis techniques have not been extensively applied to desert varnish. The only previous magnetic study reported that based on room temperature demagnetization experiments, there were noticeable differences in magnetic properties between a sample of desert varnish and the substrate sandstone [7]. Based upon the results of the demagnetization experiments, the authors concluded that the primary magnetic component of desert varnish was either magnetite (Fe3O4) or maghemite ( Fe2O3).

  8. Research in applied mathematics, numerical analysis, and computer science

    NASA Technical Reports Server (NTRS)

    1984-01-01

    Research conducted at the Institute for Computer Applications in Science and Engineering (ICASE) in applied mathematics, numerical analysis, and computer science is summarized and abstracts of published reports are presented. The major categories of the ICASE research program are: (1) numerical methods, with particular emphasis on the development and analysis of basic numerical algorithms; (2) control and parameter identification; (3) computational problems in engineering and the physical sciences, particularly fluid dynamics, acoustics, and structural analysis; and (4) computer systems and software, especially vector and parallel computers.

  9. Toward applied behavior analysis of life aloft

    NASA Technical Reports Server (NTRS)

    Brady, J. V.

    1990-01-01

    This article deals with systems at multiple levels, at least from cell to organization. It also deals with learning, decision making, and other behavior at multiple levels. Technological development of a human behavioral ecosystem appropriate to space environments requires an analytic and synthetic orientation, explicitly experimental in nature, dictated by scientific and pragmatic considerations, and closely approximating procedures of established effectiveness in other areas of natural science. The conceptual basis of such an approach has its roots in environmentalism which has two main features: (1) knowledge comes from experience rather than from innate ideas, divine revelation, or other obscure sources; and (2) action is governed by consequences rather than by instinct, reason, will, beliefs, attitudes or even the currently fashionable cognitions. Without an experimentally derived data base founded upon such a functional analysis of human behavior, the overgenerality of "ecological systems" approaches render them incapable of ensuring the successful establishment of enduring space habitats. Without an experimentally derived function account of individual behavioral variability, a natural science of behavior cannot exist. And without a natural science of behavior, the social sciences will necessarily remain in their current status as disciplines of less than optimal precision or utility. Such a functional analysis of human performance should provide an operational account of behavior change in a manner similar to the way in which Darwin's approach to natural selection accounted for the evolution of phylogenetic lines (i.e., in descriptive, nonteleological terms). Similarly, as Darwin's account has subsequently been shown to be consonant with information obtained at the cellular level, so too should behavior principles ultimately prove to be in accord with an account of ontogenetic adaptation at a biochemical level. It would thus seem obvious that the most

  10. Toward applied behavior analysis of life aloft.

    PubMed

    Brady, J V

    1990-01-01

    This article deals with systems at multiple levels, at least from cell to organization. It also deals with learning, decision making, and other behavior at multiple levels. Technological development of a human behavioral ecosystem appropriate to space environments requires an analytic and synthetic orientation, explicitly experimental in nature, dictated by scientific and pragmatic considerations, and closely approximating procedures of established effectiveness in other areas of natural science. The conceptual basis of such an approach has its roots in environmentalism which has two main features: (1) knowledge comes from experience rather than from innate ideas, divine revelation, or other obscure sources; and (2) action is governed by consequences rather than by instinct, reason, will, beliefs, attitudes or even the currently fashionable cognitions. Without an experimentally derived data base founded upon such a functional analysis of human behavior, the overgenerality of "ecological systems" approaches render them incapable of ensuring the successful establishment of enduring space habitats. Without an experimentally derived function account of individual behavioral variability, a natural science of behavior cannot exist. And without a natural science of behavior, the social sciences will necessarily remain in their current status as disciplines of less than optimal precision or utility. Such a functional analysis of human performance should provide an operational account of behavior change in a manner similar to the way in which Darwin's approach to natural selection accounted for the evolution of phylogenetic lines (i.e., in descriptive, nonteleological terms). Similarly, as Darwin's account has subsequently been shown to be consonant with information obtained at the cellular level, so too should behavior principles ultimately prove to be in accord with an account of ontogenetic adaptation at a biochemical level. It would thus seem obvious that the most

  11. Toward applied behavior analysis of life aloft

    NASA Technical Reports Server (NTRS)

    Brady, J. V.

    1990-01-01

    This article deals with systems at multiple levels, at least from cell to organization. It also deals with learning, decision making, and other behavior at multiple levels. Technological development of a human behavioral ecosystem appropriate to space environments requires an analytic and synthetic orientation, explicitly experimental in nature, dictated by scientific and pragmatic considerations, and closely approximating procedures of established effectiveness in other areas of natural science. The conceptual basis of such an approach has its roots in environmentalism which has two main features: (1) knowledge comes from experience rather than from innate ideas, divine revelation, or other obscure sources; and (2) action is governed by consequences rather than by instinct, reason, will, beliefs, attitudes or even the currently fashionable cognitions. Without an experimentally derived data base founded upon such a functional analysis of human behavior, the overgenerality of "ecological systems" approaches render them incapable of ensuring the successful establishment of enduring space habitats. Without an experimentally derived function account of individual behavioral variability, a natural science of behavior cannot exist. And without a natural science of behavior, the social sciences will necessarily remain in their current status as disciplines of less than optimal precision or utility. Such a functional analysis of human performance should provide an operational account of behavior change in a manner similar to the way in which Darwin's approach to natural selection accounted for the evolution of phylogenetic lines (i.e., in descriptive, nonteleological terms). Similarly, as Darwin's account has subsequently been shown to be consonant with information obtained at the cellular level, so too should behavior principles ultimately prove to be in accord with an account of ontogenetic adaptation at a biochemical level. It would thus seem obvious that the most

  12. Image analysis applied to luminescence microscopy

    NASA Astrophysics Data System (ADS)

    Maire, Eric; Lelievre-Berna, Eddy; Fafeur, Veronique; Vandenbunder, Bernard

    1998-04-01

    We have developed a novel approach to study luminescent light emission during migration of living cells by low-light imaging techniques. The equipment consists in an anti-vibration table with a hole for a direct output under the frame of an inverted microscope. The image is directly captured by an ultra low- light level photon-counting camera equipped with an image intensifier coupled by an optical fiber to a CCD sensor. This installation is dedicated to measure in a dynamic manner the effect of SF/HGF (Scatter Factor/Hepatocyte Growth Factor) both on activation of gene promoter elements and on cell motility. Epithelial cells were stably transfected with promoter elements containing Ets transcription factor-binding sites driving a luciferase reporter gene. Luminescent light emitted by individual cells was measured by image analysis. Images of luminescent spots were acquired with a high aperture objective and time exposure of 10 - 30 min in photon-counting mode. The sensitivity of the camera was adjusted to a high value which required the use of a segmentation algorithm dedicated to eliminate the background noise. Hence, image segmentation and treatments by mathematical morphology were particularly indicated in these experimental conditions. In order to estimate the orientation of cells during their migration, we used a dedicated skeleton algorithm applied to the oblong spots of variable intensities emitted by the cells. Kinetic changes of luminescent sources, distance and speed of migration were recorded and then correlated with cellular morphological changes for each spot. Our results highlight the usefulness of the mathematical morphology to quantify kinetic changes in luminescence microscopy.

  13. Subpixels analysis model applied to floodplain monitoring

    NASA Astrophysics Data System (ADS)

    Giraldo Osorio, J. D.; García Galiano, S. G.

    2009-04-01

    The traditional techniques to gauge hydrological events often fail with the extreme events. A particular case is the floods spatial detection. In this work, the remote sensing techniques and Geographic Information Systems (GIS) have been merged to develop a key tool for monitoring of floods. The low density of gauge stations networks in the development countries becomes remote sensing techniques the most suitable and economic way to delimitate the flood area and compute the damages cost. The common classification techniques of satellite images use "hard methods" in the sense of a pixel is assigned to an unique land cover class. For coarse resolution, the pixels inevitably will be mixed, so "soft methods" can be used in order to assign several land cover classes according to the surface fractions covered by each one. The main objective of this work is the dynamic monitoring of floods in large areas, based on satellite images -with moderate spatial resolution but with high time resolution- and Digital Elevation Model (DEM). Classified maps with finer spatial resolution can be built through the methodology of Subpixels Analysis developed. The procedure is supported on both the Linear Mixture Model (LMM) and Spatial Coherence Analysis (SCA) hypothesis. The LMM builds the land cover fraction maps through an optimization procedure which uses Lagrange Multipliers, while the SCA defines the most likely place for the land cover fractions within the coarse pixel using linear programming. A subsequent procedure improves the flooded area identification using both the drainage direction and flow accumulation raster maps derived from DEM of the study zone. The Subpixels Analysis technique was validated using historical data of floods which were obtained from satellite images. The procedure improves the spatial resolution of classified maps from satellite images with coarse resolution, while the "hard methods" keep the spatial resolution from the input coarse satellite image.

  14. Digital photoelastic analysis applied to implant dentistry

    NASA Astrophysics Data System (ADS)

    Ramesh, K.; Hariprasad, M. P.; Bhuvanewari, S.

    2016-12-01

    Development of improved designs of implant systems in dentistry have necessitated the study of stress fields in the implant regions of the mandible/maxilla for better understanding of the biomechanics involved. Photoelasticity has been used for various studies related to dental implants in view of whole field visualization of maximum shear stress in the form of isochromatic contours. The potential of digital photoelasticity has not been fully exploited in the field of implant dentistry. In this paper, the fringe field in the vicinity of the connected implants (All-On-Four® concept) is analyzed using recent advances in digital photoelasticity. Initially, a novel 3-D photoelastic model making procedure, to closely mimic all the anatomical features of the human mandible is proposed. By choosing appropriate orientation of the model with respect to the light path, the essential region of interest were sought to be analysed while keeping the model under live loading conditions. Need for a sophisticated software module to carefully identify the model domain has been brought out. For data extraction, five-step method is used and isochromatics are evaluated by twelve fringe photoelasticity. In addition to the isochromatic fringe field, whole field isoclinic data is also obtained for the first time in implant dentistry, which could throw important information in improving the structural stability of the implant systems. Analysis is carried out for the implant in the molar as well as the incisor region. In addition, the interaction effects of loaded molar implant on the incisor area are also studied.

  15. Fluorescent protein biosensors applied to microphysiological systems

    PubMed Central

    Senutovitch, Nina; Boltz, Robert; DeBiasio, Richard; Gough, Albert; Taylor, D Lansing

    2015-01-01

    This mini-review discusses the evolution of fluorescence as a tool to study living cells and tissues in vitro and the present role of fluorescent protein biosensors (FPBs) in microphysiological systems (MPSs). FPBs allow the measurement of temporal and spatial dynamics of targeted cellular events involved in normal and perturbed cellular assay systems and MPSs in real time. FPBs evolved from fluorescent analog cytochemistry (FAC) that permitted the measurement of the dynamics of purified proteins covalently labeled with environmentally insensitive fluorescent dyes and then incorporated into living cells, as well as a large list of diffusible fluorescent probes engineered to measure environmental changes in living cells. In parallel, a wide range of fluorescence microscopy methods were developed to measure the chemical and molecular activities of the labeled cells, including ratio imaging, fluorescence lifetime, total internal reflection, 3D imaging, including super-resolution, as well as high-content screening. FPBs evolved from FAC by combining environmentally sensitive fluorescent dyes with proteins in order to monitor specific physiological events such as post-translational modifications, production of metabolites, changes in various ion concentrations, and the dynamic interaction of proteins with defined macromolecules in time and space within cells. Original FPBs involved the engineering of fluorescent dyes to sense specific activities when covalently attached to particular domains of the targeted protein. The subsequent development of fluorescent proteins (FPs), such as the green fluorescent protein, dramatically accelerated the adoption of studying living cells, since the genetic “labeling” of proteins became a relatively simple method that permitted the analysis of temporal–spatial dynamics of a wide range of proteins. Investigators subsequently engineered the fluorescence properties of the FPs for environmental sensitivity that, when combined with

  16. Liquid Chromatography Applied to Space System

    NASA Astrophysics Data System (ADS)

    Poinot, Pauline; Chazalnoel, Pascale; Geffroy, Claude; Sternberg, Robert; Carbonnier, Benjamin

    Searching for signs of past or present life in our Solar System is a real challenge that stirs up the curiosity of scientists. Until now, in situ instrumentation was designed to detect and determine concentrations of a wide number of organic biomarkers. The relevant method which was and still is employed in missions dedicated to the quest of life (from Viking to ExoMars) corresponds to the pyrolysis-GC-MS. Along the missions, this approach has been significantly improved in terms of extraction efficiency and detection with the use of chemical derivative agents (e.g. MTBSTFA, DMF-DMA, TMAH…), and in terms of analysis sensitivity and resolution with the development of in situ high-resolution mass spectrometer (e.g. TOF-MS). Thanks to such an approach, organic compounds such as amino acids, sugars, tholins or polycyclic aromatic hydrocarbons (PAHs) were expected to be found. However, while there’s a consensus that the GC-MS of Viking, Huygens, MSL and MOMA space missions worked the way they had been designed to, pyrolysis is much more in debate (Glavin et al. 2001; Navarro-González et al. 2006). Indeed, (1) it is thought to remove low levels of organics, (2) water and CO2 could interfere with the detection of likely organic pyrolysis products, and (3) only low to mid-molecular weight organic molecules can be detected by this technique. As a result, researchers are now focusing on other in situ techniques which are no longer based on the volatility of the organic matter, but on the liquid phase extraction and analysis. In this line, micro-fluidic systems involving sandwich and/or competitive immunoassays (e.g. LMC, SOLID; Parro et al. 2005; Sims et al. 2012), micro-chip capillary electrophoreses (e.g. MOA; Bada et al. 2008), or nanopore-based analysis (e.g. BOLD; Schulze-Makuch et al. 2012) have been conceived for in situ analysis. Thanks to such approaches, molecular biological polymers (polysaccharides, polypeptides, polynucleotides, phospholipids, glycolipids

  17. Moving Forward: Positive Behavior Support and Applied Behavior Analysis

    ERIC Educational Resources Information Center

    Tincani, Matt

    2007-01-01

    A controversy has emerged about the relationship between positive behavior support and applied behavior analysis. Some behavior analysts suggest that positive behavior support and applied behavior analysis are the same (e.g., Carr & Sidener, 2002). Others argue that positive behavior support is harmful to applied behavior analysis (e.g., Johnston,…

  18. Classical mechanics approach applied to analysis of genetic oscillators.

    PubMed

    Vasylchenkova, Anastasiia; Mraz, Miha; Zimic, Nikolaj; Moskon, Miha

    2016-04-05

    Biological oscillators present a fundamental part of several regulatory mechanisms that control the response of various biological systems. Several analytical approaches for their analysis have been reported recently. They are, however, limited to only specific oscillator topologies and/or to giving only qualitative answers, i.e., is the dynamics of an oscillator given the parameter space oscillatory or not. Here we present a general analytical approach that can be applied to the analysis of biological oscillators. It relies on the projection of biological systems to classical mechanics systems. The approach is able to provide us with relatively accurate results in the meaning of type of behaviour system reflects (i.e. oscillatory or not) and periods of potential oscillations without the necessity to conduct expensive numerical simulations. We demonstrate and verify the proposed approach on three different implementations of amplified negative feedback oscillator.

  19. The Evolution of Fungicide Resistance Resulting from Combinations of Foliar-Acting Systemic Seed Treatments and Foliar-Applied Fungicides: A Modeling Analysis.

    PubMed

    Kitchen, James L; van den Bosch, Frank; Paveley, Neil D; Helps, Joseph; van den Berg, Femke

    2016-01-01

    For the treatment of foliar diseases of cereals, fungicides may be applied as foliar sprays or systemic seed treatments which are translocated to leaves. Little research has been done to assess the resistance risks associated with foliar-acting systemic seed treatments when used alone or in combination with foliar sprays, even though both types of treatment may share the same mode of action. It is therefore unknown to what extent adding a systemic seed treatment to a foliar spray programme poses an additional resistance risk and whether in the presence of a seed treatment additional resistance management strategies (such as limiting the total number of treatments) are necessary to limit the evolution of fungicide-resistance. A mathematical model was developed to simulate an epidemic and the resistance evolution of Zymoseptoria tritici on winter wheat, which was used to compare different combinations of seed and foliar treatments by calculating the fungicide effective life, i.e. the number of years before effective disease control is lost to resistance. A range of parameterizations for the seed treatment fungicide and different fungicide uptake models were compared. Despite the different parameterizations, the model consistently predicted the same trends in that i) similar levels of efficacy delivered either by a foliar-acting seed treatment, or a foliar application, resulted in broadly similar resistance selection, ii) adding a foliar-acting seed treatment to a foliar spray programme increased resistance selection and usually decreased effective life, and iii) splitting a given total dose-by adding a seed treatment to foliar treatments, but decreasing dose per treatment-gave effective lives that were the same as, or shorter than those given by the spray programme alone. For our chosen plant-pathogen-fungicide system, the model results suggest that to effectively manage selection for fungicide-resistance, foliar acting systemic seed treatments should be included as

  20. Introduction: Conversation Analysis in Applied Linguistics

    ERIC Educational Resources Information Center

    Sert, Olcay; Seedhouse, Paul

    2011-01-01

    This short, introductory paper presents an up-to-date account of works within the field of Applied Linguistics which have been influenced by a Conversation Analytic paradigm. The article reviews recent studies in classroom interaction, materials development, proficiency assessment and language teacher education. We believe that the publication of…

  1. Applying Mechanics to Swimming Performance Analysis.

    ERIC Educational Resources Information Center

    Barthels, Katharine

    1989-01-01

    Swimming teachers and coaches can improve their feedback to swimmers, when correcting or refining swim movements, by applying some basic biomechanical concepts relevant to swimming. This article focuses on the biomechanical considerations used in analyzing swimming performance. Techniques for spotting and correcting problems that impede…

  2. Expert systems applied to spacecraft fire safety

    NASA Technical Reports Server (NTRS)

    Smith, Richard L.; Kashiwagi, Takashi

    1989-01-01

    Expert systems are problem-solving programs that combine a knowledge base and a reasoning mechanism to simulate a human expert. The development of an expert system to manage fire safety in spacecraft, in particular the NASA Space Station Freedom, is difficult but clearly advantageous in the long-term. Some needs in low-gravity flammability characteristics, ventilating-flow effects, fire detection, fire extinguishment, and decision models, all necessary to establish the knowledge base for an expert system, are discussed.

  3. Applying Clausewitz and Systems Thinking to Design

    DTIC Science & Technology

    2012-06-01

    Consequently, a new approach organized around the concept of systems took root. Ludwig von Bertalanffy , one of systems theory’s early proponents...so achieve political and military end states that truly matter in national security and international affairs. ENDNOTES - CHAPTER 18 1. Ludwig von... Bertalanffy , “An Outline of General System Theory,” British Journal for the Philosophy of Science, Vol. 1, 1950, pp. 134-165. 2. Gareth Morgan

  4. Superresolution applied to optical data storage systems

    NASA Astrophysics Data System (ADS)

    Walker, Edwin Parker

    1999-09-01

    This dissertation investigates superresolution applications in optical data storage systems. The performance of standard and superresolving magneto-optic data storage system are quantified by scalar diffraction modeling and experiments. Classical resolution measures are reviewed. Background on superresolution definitions and their conceptual development in scanning optical microscopes, optical data storage, and image processing is presented. Figures of merit for quantifying the performance of the systems are reviewed, such as system transfer function, two-point response, focused spot size, and signal-to-noise ratio. The description of the scalar diffraction modeling used to simulate an optical data storage system is reviewed. Operation of the magneto-optic data storage system and tradeoffs of superresolving techniques are discussed. The signal and noise spatial distribution in the pupil of an optical data storage system are shown to be different. For a particular spatial frequency bandwidth, the signal and noise are concentrated in different regions of the pupil. This understanding allows the use of optical filters that partially equalize the system transfer function and increase the signal-to-noise ratio. The main superresolution techniques investigated are those that increase the transmission of the higher spatial frequencies, or equalize the system transfer function, without changing the system cutoff frequency. The optical methods used to achieve superresolution are amplitude and phase filters placed in strategic system locations. One location influences the properties of the focused spot such as the irradiance distribution and width of the central core. Another location does not change the focused spot at all, but does change the signal and noise properties of the system. Electronic filtering techniques are also used to increase the transmission of the high spatial frequencies. The amplitude and phase filter sensitivities to aberration are also investigated

  5. Applying Modeling Tools to Ground System Procedures

    NASA Technical Reports Server (NTRS)

    Di Pasquale, Peter

    2012-01-01

    As part of a long-term effort to revitalize the Ground Systems (GS) Engineering Section practices, Systems Modeling Language (SysML) and Business Process Model and Notation (BPMN) have been used to model existing GS products and the procedures GS engineers use to produce them.

  6. Applying Modeling Tools to Ground System Procedures

    NASA Technical Reports Server (NTRS)

    Di Pasquale, Peter

    2012-01-01

    As part of a long-term effort to revitalize the Ground Systems (GS) Engineering Section practices, Systems Modeling Language (SysML) and Business Process Model and Notation (BPMN) have been used to model existing GS products and the procedures GS engineers use to produce them.

  7. Sensitivity and specificity analysis of fringing-field dielectric spectroscopy applied to a multi-layer system modelling the human skin.

    PubMed

    Huclova, Sonja; Baumann, Dirk; Talary, Mark S; Fröhlich, Jürg

    2011-12-21

    The sensitivity and specificity of dielectric spectroscopy for the detection of dielectric changes inside a multi-layered structure is investigated. We focus on providing a base for sensing physiological changes in the human skin, i.e. in the epidermal and dermal layers. The correlation between changes of the human skin's effective permittivity and changes of dielectric parameters and layer thickness of the epidermal and dermal layers is assessed using numerical simulations. Numerical models include fringing-field probes placed directly on a multi-layer model of the skin. The resulting dielectric spectra in the range from 100 kHz up to 100 MHz for different layer parameters and sensor geometries are used for a sensitivity and specificity analysis of this multi-layer system. First, employing a coaxial probe, a sensitivity analysis is performed for specific variations of the parameters of the epidermal and dermal layers. Second, the specificity of this system is analysed based on the roots and corresponding sign changes of the computed dielectric spectra and their first and second derivatives. The transferability of the derived results is shown by a comparison of the dielectric spectra of a coplanar probe and a scaled coaxial probe. Additionally, a comparison of the sensitivity of a coaxial probe and an interdigitated probe as a function of electrode distance is performed. It is found that the sensitivity for detecting changes of dielectric properties in the epidermal and dermal layers strongly depends on frequency. Based on an analysis of the dielectric spectra, changes in the effective dielectric parameters can theoretically be uniquely assigned to specific changes in permittivity and conductivity. However, in practice, measurement uncertainties may degrade the performance of the system.

  8. Sensitivity and specificity analysis of fringing-field dielectric spectroscopy applied to a multi-layer system modelling the human skin

    NASA Astrophysics Data System (ADS)

    Huclova, Sonja; Baumann, Dirk; Talary, Mark S.; Fröhlich, Jürg

    2011-12-01

    The sensitivity and specificity of dielectric spectroscopy for the detection of dielectric changes inside a multi-layered structure is investigated. We focus on providing a base for sensing physiological changes in the human skin, i.e. in the epidermal and dermal layers. The correlation between changes of the human skin's effective permittivity and changes of dielectric parameters and layer thickness of the epidermal and dermal layers is assessed using numerical simulations. Numerical models include fringing-field probes placed directly on a multi-layer model of the skin. The resulting dielectric spectra in the range from 100 kHz up to 100 MHz for different layer parameters and sensor geometries are used for a sensitivity and specificity analysis of this multi-layer system. First, employing a coaxial probe, a sensitivity analysis is performed for specific variations of the parameters of the epidermal and dermal layers. Second, the specificity of this system is analysed based on the roots and corresponding sign changes of the computed dielectric spectra and their first and second derivatives. The transferability of the derived results is shown by a comparison of the dielectric spectra of a coplanar probe and a scaled coaxial probe. Additionally, a comparison of the sensitivity of a coaxial probe and an interdigitated probe as a function of electrode distance is performed. It is found that the sensitivity for detecting changes of dielectric properties in the epidermal and dermal layers strongly depends on frequency. Based on an analysis of the dielectric spectra, changes in the effective dielectric parameters can theoretically be uniquely assigned to specific changes in permittivity and conductivity. However, in practice, measurement uncertainties may degrade the performance of the system.

  9. Analysis of pressure applied during microlaryngoscopy.

    PubMed

    Fang, Rui; Chen, Hao; Sun, Jingwu

    2012-05-01

    Direct laryngoscopy is the most predominantly used technique in laryngology, with the fulcrum-based laryngoscope serving as the most popular instrument. The purpose of this study was to accurately calculate and measure the pressure acting on the laryngopharynx and the tongue base during microlaryngoscopy. The relationship between postoperative throat pain and the time and pressure applied during microlaryngoscopy were also investigated. Fifty patients undergoing microlaryngeal surgery were included in this prospective study. Parameters that may help predict difficult laryngeal exposure were measured in the patients before microlaryngoscopy. Using static equilibrium and the law of the lever, the pressure acting on the laryngopharynx and the tongue base were calculated and related parameters were then tested for their influence on pressure. The time and pressure applied during microlaryngoscopy of each patient were compared with postoperative throat pain grade. The mean pressure was 292 ± 109 mmHg and was significantly influenced by BMI, neck circumference and full mouth opening, whereas no gender-based differences of any kind were found. The pressure applied during microlaryngoscopy was extremely high in patients with difficult laryngeal exposure (376 ± 62 mmHg), serving as a possible reason for the presence of throat pain or complications present following surgery. However, it was found that the duration of suspension laryngoscopy, not the pressure, had the most significant correlation with postoperative throat pain.

  10. A robust multisyringe system for process flow analysis. Part II. A multi-commuted injection system applied to the photometric determination of free acidity and iron(III) in metallurgical solutions.

    PubMed

    Albertús, F; Cladera, A; Cerda, V

    2000-12-01

    A new software-controlled volume-based system for sample introduction in process flow injection analysis was developed. By using a multi-syringe burette coupled with one or two additional commutation valves, the multi-commuted injection of precise sample volumes was accomplished. Characteristics and performance of the injection system were studied by injecting an indicator in a buffered carrier. Three configurations were implemented in order to achieve two different tasks: the single injection of a sample in a two- or three-channels manifold, and the dual injection into different streams. The two channel flow system using the single injection was applied to the determination of free acidity in diluted samples containing high levels of iron(III), by employing the single point titration methodology. The precipitation of ferric hydroxide was prevented using the ammonium and sodium salts of oxalate and acetate as buffer titrant. Methyl Red was employed as indicator. The procedure allows determination of acid concentration in solutions with a Fe(III)/H+ molar ratio up to 0.2. Samples with higher Fe(III)/H+ molar ratios were spiked with a known strong acid at dilution. The three-channel configuration was applied to the determination of ferric ions, using, as reagent, a merging mixture of sulfuric acid and potassium thiocyanate. The double injection system was implemented in series in a single (three-channel) manifold in such a way that a different injection volume and a changed reagent were used for each analyte. It was applied to the separated or sequential determination of free acidity and ferric ions. In this configuration, iron(III) was determined using 0.5-0.7% (w/v) sodium salicylate solution as reagent. The systems can operate at up to 100, 84 and 78 injections per hour, respectively. Determinations on synthetic and process samples compared well with the reference values and procedures. Recoveries of 95-102% with a maximum RSD value of 5.4% were found for acidity

  11. pH recycling aqueous two-phase systems applied in extraction of Maitake β-Glucan and mechanism analysis using low-field nuclear magnetic resonance.

    PubMed

    Hou, Huiyun; Cao, Xuejun

    2015-07-31

    In this paper, a recycling aqueous two-phase systems (ATPS) based on two pH-response copolymers PADB and PMDM were used in purification of β-Glucan from Grifola frondosa. The main parameters, such as polymer concentration, type and concentration of salt, extraction temperature and pH, were investigated to optimize partition conditions. The results demonstrated that β-Glucan was extracted into PADB-rich phase, while impurities were extracted into PMDM-rich phase. In this 2.5% PADB/2.5% PMDM ATPS, 7.489 partition coefficient and 96.92% extraction recovery for β-Glucan were obtained in the presence of 30mmol/L KBr, at pH 8.20, 30°C. The phase-forming copolymers could be recycled by adjusting pH, with recoveries of over 96.0%. Furthermore, the partition mechanism of Maitake β-Glucan in PADB/PMDM aqueous two-phase systems was studied. Fourier transform infrared spectra, ForteBio Octet system and low-field nuclear magnetic resonance (LF-NMR) were introduced for elucidating the partition mechanism of β-Glucan. Especially, LF-NMR was firstly used in the mechanism analysis in partition of aqueous two-phase systems. The change of transverse relaxation time (T2) in ATPS could reflect the interaction between polymers and β-Glucan. Copyright © 2015 Elsevier B.V. All rights reserved.

  12. The Analysis of Fixed Final State Optimal Control in Bilinear System Applied to Bone Marrow by Cell-Cycle Specific (CCS) Chemotherapy

    NASA Astrophysics Data System (ADS)

    Rainarli, E.; E Dewi, K.

    2017-04-01

    The research conducted by Fister & Panetta shown an optimal control model of bone marrow cells against Cell Cycle Specific chemotherapy drugs. The model used was a bilinear system model. Fister & Panetta research has proved existence, uniqueness, and characteristics of optimal control (the chemotherapy effect). However, by using this model, the amount of bone marrow at the final time could achieve less than 50 percent from the amount of bone marrow before given treatment. This could harm patients because the lack of bone marrow cells made the number of leukocytes declining and patients will experience leukemia. This research would examine the optimal control of a bilinear system that applied to fixed final state. It will be used to determine the length of optimal time in administering chemotherapy and kept bone marrow cells on the allowed level at the same time. Before simulation conducted, this paper shows that the system could be controlled by using a theory of Lie Algebra. Afterward, it shows the characteristics of optimal control. Based on the simulation, it indicates that strong chemotherapy drug given in a short time frame is the most optimal condition to keep bone marrow cells spine on the allowed level but still could put playing an effective treatment. It gives preference of the weight of treatment for keeping bone marrow cells. The result of chemotherapy’s effect (u) is not able to reach the maximum value. On the other words, it needs to make adjustments of medicine’s dosage to satisfy the final treatment condition e.g. the number of bone marrow cells should be at the allowed level.

  13. EG G Mound Applied Technologies payroll system

    SciTech Connect

    Not Available

    1992-02-07

    EG G Mound Applied Technologies, Inc., manages and operates the Mound Facility, Miamisburg, Ohio, under a cost-plus-award-fee contract administered by the Department of Energy's Albuquerque Field Office. The contractor's Payroll Department is responsible for prompt payment in the proper amount to all persons entitled to be paid, in compliance with applicable laws, regulations, and legal decisions. The objective was to determine whether controls were in place to avoid erroneous payroll payments. EG G Mound Applied Technologies, Inc., did not have all the internal controls required by General Accounting Office Title 6, Pay, Leave, and Allowances.'' Specifically, they did not have computerized edits, separation of duties and responsibilities, and restricted access to payroll data files. This condition occurred because its managers were not aware of Title 6 requirements. As a result, the contractor could not assure the Department of Energy that payroll costs were processes accurately; and fraud, waste, or abuse of Department of Energy funds could go undetected. Our sample of 212 payroll transactions from a population of 66,000 in FY 1991 disclosed only two minor processing errors and no instances of fraud, waste or abuse.

  14. Does terrestrial epidemiology apply to marine systems?

    USGS Publications Warehouse

    McCallum, Hamish I.; Kuris, Armand M.; Harvell, C. Drew; Lafferty, Kevin D.; Smith, Garriet W.; Porter, James

    2004-01-01

    Most of epidemiological theory has been developed for terrestrial systems, but the significance of disease in the ocean is now being recognized. However, the extent to which terrestrial epidemiology can be directly transferred to marine systems is uncertain. Many broad types of disease-causing organism occur both on land and in the sea, and it is clear that some emergent disease problems in marine environments are caused by pathogens moving from terrestrial to marine systems. However, marine systems are qualitatively different from terrestrial environments, and these differences affect the application of modelling and management approaches that have been developed for terrestrial systems. Phyla and body plans are more diverse in marine environments and marine organisms have different life histories and probably different disease transmission modes than many of their terrestrial counterparts. Marine populations are typically more open than terrestrial ones, with the potential for long-distance dispersal of larvae. Potentially, this might enable unusually rapid propagation of epidemics in marine systems, and there are several examples of this. Taken together, these differences will require the development of new approaches to modelling and control of infectious disease in the ocean.

  15. Meta-analysis in applied ecology.

    PubMed

    Stewart, Gavin

    2010-02-23

    This overview examines research synthesis in applied ecology and conservation. Vote counting and pooling unweighted averages are widespread despite the superiority of syntheses based on weighted combination of effects. Such analyses allow exploration of methodological uncertainty in addition to consistency of effects across species, space and time, but exploring heterogeneity remains controversial. Meta-analyses are required to generalize in ecology, and to inform evidence-based decision-making, but the more sophisticated statistical techniques and registers of research used in other disciplines must be employed in ecology to fully realize their benefits.

  16. Systems biology: the reincarnation of systems theory applied in biology?

    PubMed

    Wolkenhauer, O

    2001-09-01

    With the availability of quantitative data on the transcriptome and proteome level, there is an increasing interest in formal mathematical models of gene expression and regulation. International conferences, research institutes and research groups concerned with systems biology have appeared in recent years and systems theory, the study of organisation and behaviour per se, is indeed a natural conceptual framework for such a task. This is, however, not the first time that systems theory has been applied in modelling cellular processes. Notably in the 1960s systems theory and biology enjoyed considerable interest among eminent scientists, mathematicians and engineers. Why did these early attempts vanish from research agendas? Here we shall review the domain of systems theory, its application to biology and the lessons that can be learned from the work of Robert Rosen. Rosen emerged from the early developments in the 1960s as a main critic but also developed a new alternative perspective to living systems, a concept that deserves a fresh look in the post-genome era of bioinformatics.

  17. Sociotechnical Systems as Applied to Knowledge Work

    DTIC Science & Technology

    1998-06-01

    technology forums. The research will compare earlier studies of variations and forums to new studies involving intellectual capital , knowledge bases...knowledge development, intellectual capital and collaborative software. Chapter III details the development of the sociotechnical systems...development and intellectual capital to determine consistency among the variances and forums that contribute to barriers in the deliberation process. It

  18. Applied Spectrophotometry: Analysis of a Biochemical Mixture

    ERIC Educational Resources Information Center

    Trumbo, Toni A.; Schultz, Emeric; Borland, Michael G.; Pugh, Michael Eugene

    2013-01-01

    Spectrophotometric analysis is essential for determining biomolecule concentration of a solution and is employed ubiquitously in biochemistry and molecular biology. The application of the Beer-Lambert-Bouguer Lawis routinely used to determine the concentration of DNA, RNA or protein. There is however a significant difference in determining the…

  19. Science, skepticism, and applied behavior analysis.

    PubMed

    Normand, Matthew P

    2008-01-01

    Pseudoscientific claims concerning medical and psychological treatments of all varieties are commonplace. As behavior analysts, a sound skeptical approach to our science and practice is essential. The present paper offers an overview of science and skepticism and discusses the relationship of skepticism to behavior analysis, with an emphasis on the types of issues concerning behavior analysts in practice.

  20. Science, Skepticism, and Applied Behavior Analysis

    PubMed Central

    Normand, Matthew P

    2008-01-01

    Pseudoscientific claims concerning medical and psychological treatments of all varieties are commonplace. As behavior analysts, a sound skeptical approach to our science and practice is essential. The present paper offers an overview of science and skepticism and discusses the relationship of skepticism to behavior analysis, with an emphasis on the types of issues concerning behavior analysts in practice. PMID:22477687

  1. Applied Spectrophotometry: Analysis of a Biochemical Mixture

    ERIC Educational Resources Information Center

    Trumbo, Toni A.; Schultz, Emeric; Borland, Michael G.; Pugh, Michael Eugene

    2013-01-01

    Spectrophotometric analysis is essential for determining biomolecule concentration of a solution and is employed ubiquitously in biochemistry and molecular biology. The application of the Beer-Lambert-Bouguer Lawis routinely used to determine the concentration of DNA, RNA or protein. There is however a significant difference in determining the…

  2. Enrichment analysis applied to disease prognosis

    PubMed Central

    2013-01-01

    Enrichment analysis is well established in the field of transcriptomics, where it is used to identify relevant biological features that characterize a set of genes obtained in an experiment. This article proposes the application of enrichment analysis as a first step in a disease prognosis methodology, in particular of diseases with a strong genetic component. With this analysis the objective is to identify clinical and biological features that characterize groups of patients with a common disease, and that can be used to distinguish between groups of patients associated with disease-related events. Data mining methodologies can then be used to exploit those features, and assist medical doctors in the evaluation of the patients in respect to their predisposition for a specific event. In this work the disease hypertrophic cardiomyopathy (HCM) is used as a case-study, as a first test to assess the feasibility of the application of an enrichment analysis to disease prognosis. To perform this assessment, two groups of patients have been considered: patients that have suffered a sudden cardiac death episode and patients that have not. The results presented were obtained with genetic data and the Gene Ontology, in two enrichment analyses: an enrichment profiling aiming at characterizing a group of patients (e.g. that suffered a disease-related event) based on their mutations; and a differential enrichment aiming at identifying differentiating features between a sub-group of patients and all the patients with the disease. These analyses correspond to an adaptation of the standard enrichment analysis, since multiple sets of genes are being considered, one for each patient. The preliminary results are promising, as the sets of terms obtained reflect the current knowledge about the gene functions commonly altered in HCM patients, thus allowing their characterization. Nevertheless, some factors need to be taken into consideration before the full potential of the enrichment

  3. Applying a causal framework to system modeling.

    PubMed

    Lieu, C A; Elliston, K O

    2007-01-01

    The emerging field of systems biology represents a revolution in our ability to understand biology. Perhaps for the first time in history we have the capacity to pursue biological understanding using a computer-aided integrative approach in conjunction with classical reductionist approaches. Technology has given us not only the ability to identify and measure the individual molecules of life and the way they change, but also the power to study these molecules and their changes in the context of a big picture. It is through the creation of a computer-aided framework for human understanding that we can begin to comprehend how these collections of molecules act as integrated biological systems, and to utilize this knowledge to rationally engineer the future of science and medicine.

  4. Thermal analysis applied to irradiated propolis

    NASA Astrophysics Data System (ADS)

    Matsuda, Andrea Harumi; Machado, Luci Brocardo; del Mastro, Nélida Lucia

    2002-03-01

    Propolis is a resinous hive product, collected by bees. Raw propolis requires a decontamination procedure and irradiation appears as a promising technique for this purpose. The valuable properties of propolis for food and pharmaceutical industries have led to increasing interest in its technological behavior. Thermal analysis is a chemical analysis that gives information about changes on heating of great importance for technological applications. Ground propolis samples were 60Co gamma irradiated with 0 and 10 kGy. Thermogravimetry curves shown a similar multi-stage decomposition pattern for both irradiated and unirradiated samples up to 600°C. Similarly, through differential scanning calorimetry , a coincidence of melting point of irradiated and unirradiated samples was found. The results suggest that the irradiation process do not interfere on the thermal properties of propolis when irradiated up to 10 kGy.

  5. Biomechanics and motion analysis applied to sports.

    PubMed

    Zheng, N; Barrentine, S W

    2000-05-01

    The development of motion analysis and the application of biomechanical analysis techniques to sports has paralleled the exponential growth of computational and videographic technology. Technological developments have provided for advances in the investigation of the human body and the action of the human body during sports believed to be unobtainable a few years ago. Technological advancements have brought biomechanical applications into a wide range of fields from orthopedics to entertainment. An area that has made tremendous gains using biomechanics is sports science. Coaches, therapists, and physicians are using biomechanics to improve performance, rehabilitation, and the prevention of sports related injuries. Functional analyses of athletic movements that were impossible a few years ago are available and used today. With new advancements, the possibilities for investigating the way a human interacts and reacts to environmental conditions are ever expanding.

  6. APPLYING NEW METHODS TO RESEARCH REACTOR ANALYSIS.

    SciTech Connect

    DIAMOND,D.J.CHENG,L.HANSON,A.XU,J.CAREW,J.F.

    2004-02-05

    Detailed reactor physics and safety analyses are being performed for the 20 MW D{sub 2}O-moderated research reactor at the National Institute of Standards and Technology (NIST). The analyses employ state-of-the-art calculational methods and will contribute to an update to the Final Safety Analysis Report (FSAR). Three-dimensional MCNP Monte Carlo neutron and photon transport calculations are performed to determine power and reactivity parameters, including feedback coefficients and control element worths. The core depletion and determination of the fuel compositions are performed with MONTEBURNS to model the reactor at the beginning, middle, and end-of-cycle. The time-dependent analysis of the primary loop is determined with a RELAP5 transient analysis model that includes the pump, heat exchanger, fuel element geometry, and flow channels. A statistical analysis used to assure protection from critical heat flux (CHF) is performed using a Monte Carlo simulation of the uncertainties contributing to the CHF calculation. The power distributions used to determine the local fuel conditions and margin to CHF are determined with MCNP. Evaluations have been performed for the following accidents: (1) the control rod withdrawal startup accident, (2) the maximum reactivity insertion accident, (3) loss-of-flow resulting from loss of electrical power, (4) loss-of-flow resulting from a primary pump seizure, (5) loss-of-flow resulting from inadvertent throttling of a flow control valve, (6) loss-of-flow resulting from failure of both shutdown cooling pumps and (7) misloading of a fuel element. These analyses are significantly more rigorous than those performed previously. They have provided insights into reactor behavior and additional assurance that previous analyses were conservative and the reactor was being operated safely.

  7. Applying thiouracil tagging to mouse transcriptome analysis.

    PubMed

    Gay, Leslie; Karfilis, Kate V; Miller, Michael R; Doe, Chris Q; Stankunas, Kryn

    2014-02-01

    Transcriptional profiling is a powerful approach for studying mouse development, physiology and disease models. Here we describe a protocol for mouse thiouracil tagging (TU tagging), a transcriptome analysis technology that includes in vivo covalent labeling, purification and analysis of cell type-specific RNA. TU tagging enables the isolation of RNA from a given cell population of a complex tissue, avoiding transcriptional changes induced by cell isolation trauma, as well as the identification of actively transcribed RNAs and not preexisting transcripts. Therefore, in contrast to other cell-specific transcriptional profiling methods based on the purification of tagged ribosomes or nuclei, TU tagging provides a direct examination of transcriptional regulation. We describe how to (i) deliver 4-thiouracil to transgenic mice to thio-label cell lineage-specific transcripts, (ii) purify TU-tagged RNA and prepare libraries for Illumina sequencing and (iii) follow a straightforward bioinformatics workflow to identify cell type-enriched or differentially expressed genes. Tissue containing TU-tagged RNA can be obtained in 1 d, RNA-seq libraries can be generated within 2 d and, after sequencing, an initial bioinformatics analysis can be completed in 1 additional day.

  8. Hypercube Expert System Shell - Applying Production Parallelism.

    DTIC Science & Technology

    1989-12-01

    3 I~~i rst and foreino l AFA L un covered t h ough literatutre reviews that efforts to ap- plY traditional soft ware inienentation methods to...ogete via )ase o ie Aprac set o tAiN) fihe l ts. Tche k nownelg a.,; l Post Oce pfacitess foloe aei es o ilal fa on t ria hi i I sessions angnde...Systems." In Proceedings of the National Conftrcnc on Artificial lntdlligenc (To Appear), August 1988. 17. Gupta, Anoop, Milind Tambe, Dirk Kalp

  9. Reliability analysis applied to structural tests

    NASA Technical Reports Server (NTRS)

    Diamond, P.; Payne, A. O.

    1972-01-01

    The application of reliability theory to predict, from structural fatigue test data, the risk of failure of a structure under service conditions because its load-carrying capability is progressively reduced by the extension of a fatigue crack, is considered. The procedure is applicable to both safe-life and fail-safe structures and, for a prescribed safety level, it will enable an inspection procedure to be planned or, if inspection is not feasible, it will evaluate the life to replacement. The theory has been further developed to cope with the case of structures with initial cracks, such as can occur in modern high-strength materials which are susceptible to the formation of small flaws during the production process. The method has been applied to a structure of high-strength steel and the results are compared with those obtained by the current life estimation procedures. This has shown that the conventional methods can be unconservative in certain cases, depending on the characteristics of the structure and the design operating conditions. The suitability of the probabilistic approach to the interpretation of the results from full-scale fatigue testing of aircraft structures is discussed and the assumptions involved are examined.

  10. Nanosensor Technology Applied to Living Plant Systems.

    PubMed

    Kwak, Seon-Yeong; Wong, Min Hao; Lew, Tedrick Thomas Salim; Bisker, Gili; Lee, Michael A; Kaplan, Amir; Dong, Juyao; Liu, Albert Tianxiang; Koman, Volodymyr B; Sinclair, Rosalie; Hamann, Catherine; Strano, Michael S

    2017-06-12

    An understanding of plant biology is essential to solving many long-standing global challenges, including sustainable and secure food production and the generation of renewable fuel sources. Nanosensor platforms, sensors with a characteristic dimension that is nanometer in scale, have emerged as important tools for monitoring plant signaling pathways and metabolism that are nondestructive, minimally invasive, and capable of real-time analysis. This review outlines the recent advances in nanotechnology that enable these platforms, including the measurement of chemical fluxes even at the single-molecule level. Applications of nanosensors to plant biology are discussed in the context of nutrient management, disease assessment, food production, detection of DNA proteins, and the regulation of plant hormones. Current trends and future needs are discussed with respect to the emerging trends of precision agriculture, urban farming, and plant nanobionics.

  11. Scanning methods applied to bitemark analysis

    NASA Astrophysics Data System (ADS)

    Bush, Peter J.; Bush, Mary A.

    2010-06-01

    The 2009 National Academy of Sciences report on forensics focused criticism on pattern evidence subdisciplines in which statements of unique identity are utilized. One principle of bitemark analysis is that the human dentition is unique to the extent that a perpetrator may be identified based on dental traits in a bitemark. Optical and electron scanning methods were used to measure dental minutia and to investigate replication of detail in human skin. Results indicated that being a visco-elastic substrate, skin effectively reduces the resolution of measurement of dental detail. Conclusions indicate caution in individualization statements.

  12. Thermal diffusivity measurement system applied to polymers

    NASA Astrophysics Data System (ADS)

    Abad, B.; Díaz-Chao, P.; Almarza, A.; Amantia, D.; Vázquez-Campos, S.; Isoda, Y.; Shinohara, Y.; Briones, F.; Martín-González, M. S.

    2012-06-01

    In the search for cleaner energy sources, the improvement of the efficiency of the actual ones appears as a primary objective. In this way, thermoelectric materials, which are able to convert wasted heat into electricity, are reveal as an interesting way to improve efficiency of car engines, for example. Cost-effective energy harvesting from thermoelectric devices requires materials with high electrical conductivities and Seebeck coefficient, but low thermal conductivity. Conductive polymers can fulfil these conditions if they are doped appropriately. One of the most promising polymers is Polyaniline. In this work, the thermal conductivity of the polyaniline and mixtures of polyaniline with nanoclays has been studied, using a new experimental set-up developed in the lab. The novel system is based on the steady-state method and it is used to obtain the thermal diffusivity of the polymers and the nanocomposites.

  13. Multivariate analysis applied to tomato hybrid production.

    PubMed

    Balasch, S; Nuez, F; Palomares, G; Cuartero, J

    1984-11-01

    Twenty characters were measured on 60 tomato varieties cultivated in the open-air and in polyethylene plastic-house. Data were analyzed by means of principal components, factorial discriminant methods, Mahalanobis D(2) distances and principal coordinate techniques. Factorial discriminant and Mahalanobis D(2) distances methods, both of which require collecting data plant by plant, lead to similar conclusions as the principal components method that only requires taking data by plots. Characters that make up the principal components in both environments studied are the same, although the relative importance of each one of them varies within the principal components. By combining information supplied by multivariate analysis with the inheritance mode of characters, crossings among cultivars can be experimented with that will produce heterotic hybrids showing characters within previously established limits.

  14. Rotary spectra analysis applied to static stabilometry.

    PubMed

    Chiaramello, E; Knaflitz, M; Agostini, V

    2011-01-01

    Static stabilometry is a technique aimed at quantifying postural sway during quiet standing in the upright position. Many different models and many different techniques to analyze the trajectories of the Centre of Pressure (CoP) have been proposed. Most of the parameters calculated according to these different approaches are affected by a relevant intra- and inter-subject variability or do not have a clear physiological interpretation. In this study we hypothesize that CoP trajectories have rotational characteristics, therefore we decompose them in clockwise and counter-clockwise components, using the rotary spectra analysis. Rotary spectra obtained studying a population of healthy subjects are described through the group average of spectral parameters, i.e., 95% spectral bandwidth, mean frequency, median frequency, and skewness. Results are reported for the clockwise and the counter-clockwise components and refer to the upright position maintained with eyes open or closed. This study demonstrates that the approach is feasible and that some of the spectral parameters are statistically different between the open and closed eyes conditions. More research is needed to demonstrate the clinical applicability of this approach, but results so far obtained are promising.

  15. Digital Systems Analysis

    ERIC Educational Resources Information Center

    Martin, Vance S.

    2009-01-01

    There have been many attempts to understand how the Internet affects our modern world. There have also been numerous attempts to understand specific areas of the Internet. This article applies Immanuel Wallerstein's World Systems Analysis to our informationalist society. Understanding this world as divided among individual core, semi-periphery,…

  16. Digital Systems Analysis

    ERIC Educational Resources Information Center

    Martin, Vance S.

    2009-01-01

    There have been many attempts to understand how the Internet affects our modern world. There have also been numerous attempts to understand specific areas of the Internet. This article applies Immanuel Wallerstein's World Systems Analysis to our informationalist society. Understanding this world as divided among individual core, semi-periphery,…

  17. Design of the optocoupler applied to medical lighting systems.

    PubMed

    Yang, Xibin; Lit, Rui; Zhu, Jianfeng; Xiong, Daxi

    2012-12-01

    A new type of optocoupler applied to medical lighting system is proposed, and the principle, Etendue and design process is introduced. With the help of Tracrpro, modeling and simulation of the optocoupler is conducted and the parameters are optimized. Analysis of factors affecting the energy coupling efficiency is done. With a view towards the development of Ultra High Brightness Light Emitting Diodes (UHB-LEDs), which play an important role a new sources of lighting in various biomedical devices, including those used in diagnosis and treatment, a series of simulations are executed and a variety of solutions are achieved. According to simulation results, the design target of coupling efficiency is achieved and the optical uniformity is also significantly improved. According to the result of theoretical analysis, verification experiments are designed and simulation results are verified. The optocoupler, which has simple structure, compact size and low cost, is suitable for applications in the field of low-cost medical domain.

  18. Treatment Integrity in Applied Behavior Analysis with Children.

    ERIC Educational Resources Information Center

    Gresham, Frank M.; And Others

    1993-01-01

    A review of 158 applied behavior analysis studies with children as subjects, published in the "Journal of Applied Behavior Analysis" between 1980 and 1990, found that (1) 16% measured the accuracy of independent variable implementation, and (2) two-thirds did not operationally define components of the independent variable. Specific recommendations…

  19. Sensitivity analysis applied to stalled airfoil wake and steady control

    NASA Astrophysics Data System (ADS)

    Patino, Gustavo; Gioria, Rafael; Meneghini, Julio

    2014-11-01

    The sensitivity of an eigenvalue to base flow modifications induced by an external force is applied to the global unstable modes associated to the onset of vortex shedding in the wake of a stalled airfoil. In this work, the flow regime is close to the first instability of the system and its associated eigenvalue/eigenmode is determined. The sensitivity analysis to a general punctual external force allows establishing the regions where control devices must be in order to stabilize the global modes. Different types of steady control devices, passive and active, are used in the regions predicted by the sensitivity analysis to check the vortex shedding suppression, i.e. the primary instability bifurcation is delayed. The new eigenvalue, modified by the action of the device, is also calculated. Finally the spectral finite element method is employed to determine flow characteristics before and after of the bifurcation in order to cross check the results.

  20. Compatibility of person-centered planning and applied behavior analysis

    PubMed Central

    Holburn, Steve

    2001-01-01

    In response to Osborne (1999), the aims and practices of person-centered planning (PCP) are compared to the basic principles of applied behavior analysis set forth by Baer, Wolf, and Risley (1968, 1987). The principal goal of PCP is social integration of people with disabilities; it qualifies as a socially important behavior, and its problems have been displayed sufficiently. However, social integration is a complex social problem whose solution requires access to system contingencies that influence lifestyles. Nearly all of the component goals of PCP proposed by O'Brien (1987b) have been reliably quantified, although concurrent measurement of outcomes such as friendship, autonomy, and respect presents a formidable challenge. Behavioral principles such as contingency and contextual control are operative within PCP, but problems in achieving reliable implementation appear to impede an experimental analysis. PMID:22478371

  1. Applying machine learning techniques to DNA sequence analysis

    SciTech Connect

    Shavlik, J.W.

    1992-01-01

    We are developing a machine learning system that modifies existing knowledge about specific types of biological sequences. It does this by considering sample members and nonmembers of the sequence motif being learned. Using this information (which we call a domain theory''), our learning algorithm produces a more accurate representation of the knowledge needed to categorize future sequences. Specifically, the KBANN algorithm maps inference rules, such as consensus sequences, into a neural (connectionist) network. Neural network training techniques then use the training examples of refine these inference rules. We have been applying this approach to several problems in DNA sequence analysis and have also been extending the capabilities of our learning system along several dimensions.

  2. Applying machine learning techniques to DNA sequence analysis

    SciTech Connect

    Shavlik, J.W. . Dept. of Computer Sciences); Noordewier, M.O. . Dept. of Computer Science)

    1992-01-01

    We are primarily developing a machine teaming (ML) system that modifies existing knowledge about specific types of biological sequences. It does this by considering sample members and nonmembers of the sequence motif being teamed. Using this information, our teaming algorithm produces a more accurate representation of the knowledge needed to categorize future sequences. Specifically, our KBANN algorithm maps inference rules about a given recognition task into a neural network. Neural network training techniques then use the training examples to refine these inference rules. We call these rules a domain theory, following the convention in the machine teaming community. We have been applying this approach to several problems in DNA sequence analysis. In addition, we have been extending the capabilities of our teaming system along several dimensions. We have also been investigating parallel algorithms that perform sequence alignments in the presence of frameshift errors.

  3. Hyperspectral imaging applied to complex particulate solids systems

    NASA Astrophysics Data System (ADS)

    Bonifazi, Giuseppe; Serranti, Silvia

    2008-04-01

    HyperSpectral Imaging (HSI) is based on the utilization of an integrated hardware and software (HW&SW) platform embedding conventional imaging and spectroscopy to attain both spatial and spectral information from an object. Although HSI was originally developed for remote sensing, it has recently emerged as a powerful process analytical tool, for non-destructive analysis, in many research and industrial sectors. The possibility to apply on-line HSI based techniques in order to identify and quantify specific particulate solid systems characteristics is presented and critically evaluated. The originally developed HSI based logics can be profitably applied in order to develop fast, reliable and lowcost strategies for: i) quality control of particulate products that must comply with specific chemical, physical and biological constraints, ii) performance evaluation of manufacturing strategies related to processing chains and/or realtime tuning of operative variables and iii) classification-sorting actions addressed to recognize and separate different particulate solid products. Case studies, related to recent advances in the application of HSI to different industrial sectors, as agriculture, food, pharmaceuticals, solid waste handling and recycling, etc. and addressed to specific goals as contaminant detection, defect identification, constituent analysis and quality evaluation are described, according to authors' originally developed application.

  4. A Vision for Systems Engineering Applied to Wind Energy (Presentation)

    SciTech Connect

    Felker, F.; Dykes, K.

    2015-01-01

    This presentation was given at the Third Wind Energy Systems Engineering Workshop on January 14, 2015. Topics covered include the importance of systems engineering, a vision for systems engineering as applied to wind energy, and application of systems engineering approaches to wind energy research and development.

  5. Applying expertise to data in the Geologist's Assistant expert system

    SciTech Connect

    Berkbigler, K.P.; Papcun, G.J.; Marusak, N.L.; Hutson, J.E.

    1988-01-01

    The Geologist's Assistant combines expert system technology with numerical pattern-matching and online communication to a large database. This paper discusses the types of rules used for the expert system, the pattern-matching technique applied, and the implementation of the system using a commercial expert system development environment. 13 refs., 8 figs.

  6. Negative reinforcement in applied behavior analysis: an emerging technology.

    PubMed

    Iwata, B A

    1987-01-01

    Although the effects of negative reinforcement on human behavior have been studied for a number of years, a comprehensive body of applied research does not exist at this time. This article describes three aspects of negative reinforcement as it relates to applied behavior analysis: behavior acquired or maintained through negative reinforcement, the treatment of negatively reinforced behavior, and negative reinforcement as therapy. A consideration of research currently being done in these areas suggests the emergence of an applied technology on negative reinforcement.

  7. Setting events in applied behavior analysis: Toward a conceptual and methodological expansion

    PubMed Central

    Wahler, Robert G.; Fox, James J.

    1981-01-01

    The contributions of applied behavior analysis as a natural science approach to the study of human behavior are acknowledged. However, it is also argued that applied behavior analysis has provided limited access to the full range of environmental events that influence socially significant behavior. Recent changes in applied behavior analysis to include analysis of side effects and social validation represent ways in which the traditional applied behavior analysis conceptual and methodological model has been profitably expanded. A third area of expansion, the analysis of setting events, is proposed by the authors. The historical development of setting events as a behavior influence concept is traced. Modifications of the basic applied behavior analysis methodology and conceptual systems that seem necessary to setting event analysis are discussed and examples of descriptive and experimental setting event analyses are presented. PMID:16795646

  8. Animal Research in the "Journal of Applied Behavior Analysis"

    ERIC Educational Resources Information Center

    Edwards, Timothy L.; Poling, Alan

    2011-01-01

    This review summarizes the 6 studies with nonhuman animal subjects that have appeared in the "Journal of Applied Behavior Analysis" and offers suggestions for future research in this area. Two of the reviewed articles described translational research in which pigeons were used to illustrate and examine behavioral phenomena of applied significance…

  9. Negative Reinforcement in Applied Behavior Analysis: An Emerging Technology.

    ERIC Educational Resources Information Center

    Iwata, Brian A.

    1987-01-01

    The article describes three aspects of negative reinforcement as it relates to applied behavior analysis: behavior acquired or maintained through negative reinforcement, the treatment of negatively reinforced behavior, and negative reinforcement as therapy. Current research suggests the emergence of an applied technology on negative reinforcement.…

  10. B. F. Skinner's contributions to applied behavior analysis

    PubMed Central

    Morris, Edward K.; Smith, Nathaniel G.; Altus, Deborah E.

    2005-01-01

    Our paper reviews and analyzes B. F. Skinner's contributions to applied behavior analysis in order to assess his role as the field's originator and founder. We found, first, that his contributions fall into five categorizes: the style and content of his science, his interpretations of typical and atypical human behavior, the implications he drew from his science for application, his descriptions of possible applications, and his own applications to nonhuman and human behavior. Second, we found that he explicitly or implicitly addressed all seven dimensions of applied behavior analysis. These contributions and the dimensions notwithstanding, he neither incorporated the field's scientific (e.g., analytic) and social dimensions (e.g., applied) into any program of published research such that he was its originator, nor did he systematically integrate, advance, and promote the dimensions so to have been its founder. As the founder of behavior analysis, however, he was the father of applied behavior analysis. PMID:22478444

  11. Multidisciplinary System Reliability Analysis

    NASA Technical Reports Server (NTRS)

    Mahadevan, Sankaran; Han, Song; Chamis, Christos C. (Technical Monitor)

    2001-01-01

    The objective of this study is to develop a new methodology for estimating the reliability of engineering systems that encompass multiple disciplines. The methodology is formulated in the context of the NESSUS probabilistic structural analysis code, developed under the leadership of NASA Glenn Research Center. The NESSUS code has been successfully applied to the reliability estimation of a variety of structural engineering systems. This study examines whether the features of NESSUS could be used to investigate the reliability of systems in other disciplines such as heat transfer, fluid mechanics, electrical circuits etc., without considerable programming effort specific to each discipline. In this study, the mechanical equivalence between system behavior models in different disciplines are investigated to achieve this objective. A new methodology is presented for the analysis of heat transfer, fluid flow, and electrical circuit problems using the structural analysis routines within NESSUS, by utilizing the equivalence between the computational quantities in different disciplines. This technique is integrated with the fast probability integration and system reliability techniques within the NESSUS code, to successfully compute the system reliability of multidisciplinary systems. Traditional as well as progressive failure analysis methods for system reliability estimation are demonstrated, through a numerical example of a heat exchanger system involving failure modes in structural, heat transfer and fluid flow disciplines.

  12. A system of vegetation classification applied to Hawaii

    Treesearch

    Michael G. Buck; Timothy E. Paysen

    1984-01-01

    A classification system for use in describing vegetation has been developed for Hawaii. Physiognomic and taxonomic criteria are used for a hierarchical stratification of vegetation in which the system categories are Formation, Subformation, Series, Association, and Phase. The System applies to local resource management activities and serves as a framework for resource...

  13. Competing Uses of Underground Systems Related to Energy Supply: Applying Single- and Multiphase Simulations for Site Characterization and Risk-Analysis

    NASA Astrophysics Data System (ADS)

    Kissinger, A.; Walter, L.; Darcis, M.; Flemisch, B.; Class, H.

    2012-04-01

    Global climate change, shortage of resources and the resulting turn towards renewable sources of energy lead to a growing demand for the utilization of subsurface systems. Among these competing uses are Carbon Capture and Storage (CCS), geothermal energy, nuclear waste disposal, "renewable" methane or hydrogen storage as well as the ongoing production of fossil resources like oil, gas, and coal. Besides competing among themselves, these technologies may also create conflicts with essential public interests like water supply. For example, the injection of CO2 into the underground causes an increase in pressure reaching far beyond the actual radius of influence of the CO2 plume, potentially leading to large amounts of displaced salt water. Finding suitable sites is a demanding task for several reasons. Natural systems as opposed to technical systems are always characterized by heterogeneity. Therefore, parameter uncertainty impedes reliable predictions towards capacity and safety of a site. State of the art numerical simulations combined with stochastic approaches need to be used to obtain a more reliable assessment of the involved risks and the radii of influence of the different processes. These simulations may include the modeling of single- and multiphase non-isothermal flow, geo-chemical and geo-mechanical processes in order to describe all relevant physical processes adequately. Stochastic approaches have the aim to estimate a bandwidth of the key output parameters based on uncertain input parameters. Risks for these different underground uses can then be made comparable with each other. Along with the importance and the urgency of the competing processes this may lead to a more profound basis for a decision. Communicating risks to stake holders and a concerned public is crucial for the success of finding a suitable site for CCS (or other subsurface utilization). We present and discuss first steps towards an approach for addressing the issue of competitive

  14. Causal Claims and the Operational Environment: An Analysis of Conventional and Emergent Causality as Applied to the Systems in 2007-2008 Iraq

    DTIC Science & Technology

    2012-06-08

    possesses a sense of agency or free-will and is capable of changing or adapting. That is to say they that, “organizations are strongly influenced by...increasing cooperation throughout their time in that area. Both of these systems have a sense of agency , read free-will, that enables them to choose whether... sense of agency which allows it to operate on its own as would a living, breathing human being. Third, I give an explanation of each open system’s

  15. First Attempt of Applying Factor Analysis in Moving Base Gravimetry

    NASA Astrophysics Data System (ADS)

    Li, X.; Roman, D. R.

    2014-12-01

    For gravimetric observation systems on mobile platforms (land/sea/airborne), the Low Signal to Noise Ratio (SNR) issue is the main barrier to achieving an accurate, high resolution gravity signal. Normally, low-pass filters (Childers et al 1999, Forsberg et al 2000, Kwon and Jekeli 2000, Hwang et al 2006) are applied to smooth or remove the high frequency "noise" - even though some of the high frequency component is not necessarily noise. This is especially true for aerogravity surveys such as those from the Gravity for the Redefinition of the American Vertical Datum (GRAV-D) project. These gravity survey flights have a spatial resolution of 10 km between tracks but higher resolution along track. The along track resolution is improved due to the lower flight height (6.1 km), equipment sensitivity, and improved modeling of potential errors. Additionally, these surveys suffer from a loss of signal power due to the increased flight elevation. Hence, application of a low-pass filter removes possible signal sensed in the along-track direction that might otherwise prove useful for various geophysical and geodetic applications. Some cutting-edge developments in Wavelets and Artificial Neural Networks had been successfully applied for obtaining improved results (Li 2008 and 2011, Liang and Liu 2013). However, a clearer and fundamental understanding of the error characteristics will further improve the quality of the gravity estimates out of these gravimetric systems. Here, instead of using any predefined basis function or any a priori model, the idea of Factor Analysis is first employed to try to extract the underlying factors of the noises in the systems. Real data sets collected by both land vehicle and aircraft will be processed as the examples.

  16. The Significance of Regional Analysis in Applied Geography.

    ERIC Educational Resources Information Center

    Sommers, Lawrence M.

    Regional analysis is central to applied geographic research, contributing to better planning and policy development for a variety of societal problems facing the United States. The development of energy policy serves as an illustration of the capabilities of this type of analysis. The United States has had little success in formulating a national…

  17. Overview of MSFC's Applied Fluid Dynamics Analysis Group Activities

    NASA Technical Reports Server (NTRS)

    Garcia, Roberto; Griffin, Lisa; Williams, Robert

    2003-01-01

    TD64, the Applied Fluid Dynamics Analysis Group, is one of several groups with high-fidelity fluids design and analysis expertise in the Space Transportation Directorate at Marshall Space Flight Center (MSFC). TD64 assists personnel working on other programs. The group participates in projects in the following areas: turbomachinery activities, nozzle activities, combustion devices, and the Columbia accident investigation.

  18. Research in Applied Mathematics Related to Nonlinear System Theory.

    DTIC Science & Technology

    1981-09-01

    control systems", Journal of Mathematical Analysis and Applications , 58: 98-112. R. E. KALMAN [1976] "Realization theory of linear dynamical systems...Theory. [19801 "Some remarks on the reachability of infinite- dimensional linear systems", Journal of Mathematical Analysis and Applications , 74: 568-577

  19. Development of adaptive control applied to chaotic systems

    NASA Astrophysics Data System (ADS)

    Rhode, Martin Andreas

    1997-12-01

    Continuous-time derivative control and adaptive map-based recursive feedback control techniques are used to control chaos in a variety of systems and in situations that are of practical interest. The theoretical part of the research includes the review of fundamental concept of control theory in the context of its applications to deterministic chaotic systems, the development of a new adaptive algorithm to identify the linear system properties necessary for control, and the extension of the recursive proportional feedback control technique, RPF, to high dimensional systems. Chaos control was applied to models of a thermal pulsed combustor, electro-chemical dissolution and the hyperchaotic Rossler system. Important implications for combustion engineering were suggested by successful control of the model of the thermal pulsed combustor. The system was automatically tracked while maintaining control into regions of parameter and state space where no stable attractors exist. In a simulation of the electrochemical dissolution system, application of derivative control to stabilize a steady state, and adaptive RPF to stabilize a period one orbit, was demonstrated. The high dimensional adaptive control algorithm was applied in a simulation using the Rossler hyperchaotic system, where a period-two orbit with two unstable directions was stabilized and tracked over a wide range of a system parameter. In the experimental part, the electrochemical system was studied in parameter space, by scanning the applied potential and the frequency of the rotating copper disk. The automated control algorithm is demonstrated to be effective when applied to stabilize a period-one orbit in the experiment. We show the necessity of small random perturbations applied to the system in order to both learn the dynamics and control the system at the same time. The simultaneous learning and control capability is shown to be an important part of the active feedback control.

  20. System theory as applied differential geometry. [linear system

    NASA Technical Reports Server (NTRS)

    Hermann, R.

    1979-01-01

    The invariants of input-output systems under the action of the feedback group was examined. The approach used the theory of Lie groups and concepts of modern differential geometry, and illustrated how the latter provides a basis for the discussion of the analytic structure of systems. Finite dimensional linear systems in a single independent variable are considered. Lessons of more general situations (e.g., distributed parameter and multidimensional systems) which are increasingly encountered as technology advances are presented.

  1. Factorial switching linear dynamical systems applied to physiological condition monitoring.

    PubMed

    Quinn, John A; Williams, Christopher K I; McIntosh, Neil

    2009-09-01

    Condition monitoring often involves the analysis of systems with hidden factors that switch between different modes of operation in some way. Given a sequence of observations, the task is to infer the filtering distribution of the switch setting at each time step. In this paper, we present factorial switching linear dynamical systems as a general framework for handling such problems. We show how domain knowledge and learning can be successfully combined in this framework, and introduce a new factor (the "X-factor") for dealing with unmodeled variation. We demonstrate the flexibility of this type of model by applying it to the problem of monitoring the condition of a premature baby receiving intensive care. The state of health of a baby cannot be observed directly, but different underlying factors are associated with particular patterns of physiological measurements and artifacts. We have explicit knowledge of common factors and use the X-factor to model novel patterns which are clinically significant but have unknown cause. Experimental results are given which show the developed methods to be effective on typical intensive care unit monitoring data.

  2. Complex, Dynamic Systems: A New Transdisciplinary Theme for Applied Linguistics?

    ERIC Educational Resources Information Center

    Larsen-Freeman, Diane

    2012-01-01

    In this plenary address, I suggest that Complexity Theory has the potential to contribute a transdisciplinary theme to applied linguistics. Transdisciplinary themes supersede disciplines and spur new kinds of creative activity (Halliday 2001 [1990]). Investigating complex systems requires researchers to pay attention to system dynamics. Since…

  3. A vegetation classification system applied to southern California

    Treesearch

    Timothy E. Paysen; Jeanine A. Derby; Hugh Black; Vernon C. Bleich; John W. Mincks

    1980-01-01

    A classification system for use in describing vegetation has been developed and is being applied to southern California. It is based upon a hierarchical stratification of vegetation, using physiognomic and taxonomic criteria. The system categories are Formation, Subformation. Series, Association, and Phase. Formations, Subformations, and Series have been specified for...

  4. Complex, Dynamic Systems: A New Transdisciplinary Theme for Applied Linguistics?

    ERIC Educational Resources Information Center

    Larsen-Freeman, Diane

    2012-01-01

    In this plenary address, I suggest that Complexity Theory has the potential to contribute a transdisciplinary theme to applied linguistics. Transdisciplinary themes supersede disciplines and spur new kinds of creative activity (Halliday 2001 [1990]). Investigating complex systems requires researchers to pay attention to system dynamics. Since…

  5. A strategy to apply quantitative epistasis analysis on developmental traits.

    PubMed

    Labocha, Marta K; Yuan, Wang; Aleman-Meza, Boanerges; Zhong, Weiwei

    2017-05-15

    Genetic interactions are keys to understand complex traits and evolution. Epistasis analysis is an effective method to map genetic interactions. Large-scale quantitative epistasis analysis has been well established for single cells. However, there is a substantial lack of such studies in multicellular organisms and their complex phenotypes such as development. Here we present a method to extend quantitative epistasis analysis to developmental traits. In the nematode Caenorhabditis elegans, we applied RNA interference on mutants to inactivate two genes, used an imaging system to quantitatively measure phenotypes, and developed a set of statistical methods to extract genetic interactions from phenotypic measurement. Using two different C. elegans developmental phenotypes, body length and sex ratio, as examples, we showed that this method could accommodate various metazoan phenotypes with performances comparable to those methods in single cell growth studies. Comparing with qualitative observations, this method of quantitative epistasis enabled detection of new interactions involving subtle phenotypes. For example, several sex-ratio genes were found to interact with brc-1 and brd-1, the orthologs of the human breast cancer genes BRCA1 and BARD1, respectively. We confirmed the brc-1 interactions with the following genes in DNA damage response: C34F6.1, him-3 (ortholog of HORMAD1, HORMAD2), sdc-1, and set-2 (ortholog of SETD1A, SETD1B, KMT2C, KMT2D), validating the effectiveness of our method in detecting genetic interactions. We developed a reliable, high-throughput method for quantitative epistasis analysis of developmental phenotypes.

  6. Extended Kramers-Moyal analysis applied to optical trapping.

    PubMed

    Honisch, Christoph; Friedrich, Rudolf; Hörner, Florian; Denz, Cornelia

    2012-08-01

    The Kramers-Moyal analysis is a well-established approach to analyze stochastic time series from complex systems. If the sampling interval of a measured time series is too low, systematic errors occur in the analysis results. These errors are labeled as finite time effects in the literature. In the present article, we present some new insights about these effects and discuss the limitations of a previously published method to estimate Kramers-Moyal coefficients at the presence of finite time effects. To increase the reliability of this method and to avoid misinterpretations, we extend it by the computation of error estimates for estimated parameters using a Monte Carlo error propagation technique. Finally, the extended method is applied to a data set of an optical trapping experiment yielding estimations of the forces acting on a Brownian particle trapped by optical tweezers. We find an increased Markov-Einstein time scale of the order of the relaxation time of the process, which can be traced back to memory effects caused by the interaction of the particle and the fluid. Above the Markov-Einstein time scale, the process can be very well described by the classical overdamped Markov model for Brownian motion.

  7. Quantitative Analysis of the Interdisciplinarity of Applied Mathematics

    PubMed Central

    Zhang, Pengyuan

    2015-01-01

    The increasing use of mathematical techniques in scientific research leads to the interdisciplinarity of applied mathematics. This viewpoint is validated quantitatively here by statistical and network analysis on the corpus PNAS 1999–2013. A network describing the interdisciplinary relationships between disciplines in a panoramic view is built based on the corpus. Specific network indicators show the hub role of applied mathematics in interdisciplinary research. The statistical analysis on the corpus content finds that algorithms, a primary topic of applied mathematics, positively correlates, increasingly co-occurs, and has an equilibrium relationship in the long-run with certain typical research paradigms and methodologies. The finding can be understood as an intrinsic cause of the interdisciplinarity of applied mathematics. PMID:26352604

  8. Quantitative Analysis of the Interdisciplinarity of Applied Mathematics.

    PubMed

    Xie, Zheng; Duan, Xiaojun; Ouyang, Zhenzheng; Zhang, Pengyuan

    2015-01-01

    The increasing use of mathematical techniques in scientific research leads to the interdisciplinarity of applied mathematics. This viewpoint is validated quantitatively here by statistical and network analysis on the corpus PNAS 1999-2013. A network describing the interdisciplinary relationships between disciplines in a panoramic view is built based on the corpus. Specific network indicators show the hub role of applied mathematics in interdisciplinary research. The statistical analysis on the corpus content finds that algorithms, a primary topic of applied mathematics, positively correlates, increasingly co-occurs, and has an equilibrium relationship in the long-run with certain typical research paradigms and methodologies. The finding can be understood as an intrinsic cause of the interdisciplinarity of applied mathematics.

  9. Applied Information Systems. Course Five. Information Systems Curriculum.

    ERIC Educational Resources Information Center

    O'Neil, Sharon Lund; Everett, Donna R.

    This course is the fifth of seven in the Information Systems curriculum. The purpose of the course is to build on skills acquired in the earlier courses. It reviews the importance of information to management and the organization and information systems concepts within an office. These components are provided for each task area: behavioral…

  10. Applied intelligent systems: blending fuzzy logic with conventional control

    NASA Astrophysics Data System (ADS)

    Filev, Dimitar; Syed, Fazal U.

    2010-05-01

    The aim of this paper is to show that design of applied intelligent control systems requires different types of blending between fuzzy logic and conventional control systems. Two alternative automotive applications - a manufacturing process control problem and an advisory system for fuel efficient driving - that benefit from both fuzzy and control theories are reviewed and different levels of prioritisations of both approaches are discussed based on the specificity of the applications.

  11. Flight control system design factors for applying automated testing techniques

    NASA Technical Reports Server (NTRS)

    Sitz, Joel R.; Vernon, Todd H.

    1990-01-01

    Automated validation of flight-critical embedded systems is being done at ARC Dryden Flight Research Facility. The automated testing techniques are being used to perform closed-loop validation of man-rated flight control systems. The principal design features and operational experiences of the X-29 forward-swept-wing aircraft and F-18 High Alpha Research Vehicle (HARV) automated test systems are discussed. Operationally applying automated testing techniques has accentuated flight control system features that either help or hinder the application of these techniques. The paper also discusses flight control system features which foster the use of automated testing techniques.

  12. Applying systems engineering methodologies to the micro- and nanoscale realm

    NASA Astrophysics Data System (ADS)

    Garrison Darrin, M. Ann

    2012-06-01

    Micro scale and nano scale technology developments have the potential to revolutionize smart and small systems. The application of systems engineering methodologies that integrate standalone, small-scale technologies and interface them with macro technologies to build useful systems is critical to realizing the potential of these technologies. This paper covers the expanding knowledge base on systems engineering principles for micro and nano technology integration starting with a discussion of the drivers for applying a systems approach. Technology development on the micro and nano scale has transition from laboratory curiosity to the realization of products in the health, automotive, aerospace, communication, and numerous other arenas. This paper focuses on the maturity (or lack thereof) of the field of nanosystems which is emerging in a third generation having transitioned from completing active structures to creating systems. The emphasis of applying a systems approach focuses on successful technology development based on the lack of maturity of current nano scale systems. Therefore the discussion includes details relating to enabling roles such as product systems engineering and technology development. Classical roles such as acquisition systems engineering are not covered. The results are also targeted towards small-scale technology developers who need to take into account systems engineering processes such as requirements definition, verification, and validation interface management and risk management in the concept phase of technology development to maximize the likelihood of success, cost effective micro and nano technology to increase the capability of emerging deployed systems and long-term growth and profits.

  13. Applying Model Based Systems Engineering to NASA's Space Communications Networks

    NASA Technical Reports Server (NTRS)

    Bhasin, Kul; Barnes, Patrick; Reinert, Jessica; Golden, Bert

    2013-01-01

    System engineering practices for complex systems and networks now require that requirement, architecture, and concept of operations product development teams, simultaneously harmonize their activities to provide timely, useful and cost-effective products. When dealing with complex systems of systems, traditional systems engineering methodology quickly falls short of achieving project objectives. This approach is encumbered by the use of a number of disparate hardware and software tools, spreadsheets and documents to grasp the concept of the network design and operation. In case of NASA's space communication networks, since the networks are geographically distributed, and so are its subject matter experts, the team is challenged to create a common language and tools to produce its products. Using Model Based Systems Engineering methods and tools allows for a unified representation of the system in a model that enables a highly related level of detail. To date, Program System Engineering (PSE) team has been able to model each network from their top-level operational activities and system functions down to the atomic level through relational modeling decomposition. These models allow for a better understanding of the relationships between NASA's stakeholders, internal organizations, and impacts to all related entities due to integration and sustainment of existing systems. Understanding the existing systems is essential to accurate and detailed study of integration options being considered. In this paper, we identify the challenges the PSE team faced in its quest to unify complex legacy space communications networks and their operational processes. We describe the initial approaches undertaken and the evolution toward model based system engineering applied to produce Space Communication and Navigation (SCaN) PSE products. We will demonstrate the practice of Model Based System Engineering applied to integrating space communication networks and the summary of its

  14. Applied behavior analysis: New directions from the laboratory

    PubMed Central

    Epling, W. Frank; Pierce, W. David

    1983-01-01

    Applied behavior analysis began when laboratory based principles were extended to humans inorder to change socially significant behavior. Recent laboratory findings may have applied relevance; however, the majority of basic researchers have not clearly communicated the practical implications of their work. The present paper samples some of the new findings and attempts to demonstrate their applied importance. Schedule-induced behavior which occurs as a by-product of contingencies of reinforcement is discussed. Possible difficulties in treatment and management of induced behaviors are considered. Next, the correlation-based law of effect and the implications of relative reinforcement are explored in terms of applied examples. Relative rate of reinforcement is then extended to the literature dealing with concurrent operants. Concurrent operant models may describe human behavior of applied importance, and several techniques for modification of problem behavior are suggested. As a final concern, the paper discusses several new paradigms. While the practical importance of these models is not clear at the moment, it may be that new practical advantages will soon arise. Thus, it is argued that basic research continues to be of theoretical and practical importance to applied behavior analysis. PMID:22478574

  15. Applying Sustainable Systems Development Approach to Educational Technology Systems

    ERIC Educational Resources Information Center

    Huang, Albert

    2012-01-01

    Information technology (IT) is an essential part of modern education. The roles and contributions of technology to education have been thoroughly documented in academic and professional literature. Despite the benefits, the use of educational technology systems (ETS) also creates a significant impact on the environment, primarily due to energy…

  16. Applying Sustainable Systems Development Approach to Educational Technology Systems

    ERIC Educational Resources Information Center

    Huang, Albert

    2012-01-01

    Information technology (IT) is an essential part of modern education. The roles and contributions of technology to education have been thoroughly documented in academic and professional literature. Despite the benefits, the use of educational technology systems (ETS) also creates a significant impact on the environment, primarily due to energy…

  17. Overview of MSFC's Applied Fluid Dynamics Analysis Group Activities

    NASA Technical Reports Server (NTRS)

    Garcia, Roberto; Wang, Tee-See; Griffin, Lisa; Turner, James E. (Technical Monitor)

    2001-01-01

    This document is a presentation graphic which reviews the activities of the Applied Fluid Dynamics Analysis Group at Marshall Space Flight Center (i.e., Code TD64). The work of this group focused on supporting the space transportation programs. The work of the group is in Computational Fluid Dynamic tool development. This development is driven by hardware design needs. The major applications for the design and analysis tools are: turbines, pumps, propulsion-to-airframe integration, and combustion devices.

  18. Support systems design and analysis

    NASA Technical Reports Server (NTRS)

    Ferguson, R. M.

    1985-01-01

    The integration of Kennedy Space Center (KSC) ground support systems with the new launch processing system and new launch vehicle provided KSC with a unique challenge in system design and analysis for the Space Transportation System. Approximately 70 support systems are controlled and monitored by the launch processing system. Typical systems are main propulsion oxygen and hydrogen loading systems, environmental control life support system, hydraulics, etc. An End-to-End concept of documentation and analysis was chosen and applied to these systems. Unique problems were resolved in the areas of software analysis, safing under emergency conditions, sampling rates, and control loop analysis. New methods of performing End-to-End reliability analyses were implemented. The systems design approach selected and the resolution of major problem areas are discussed.

  19. Animal research in the Journal of Applied Behavior Analysis.

    PubMed

    Edwards, Timothy L; Poling, Alan

    2011-01-01

    This review summarizes the 6 studies with nonhuman animal subjects that have appeared in the Journal of Applied Behavior Analysis and offers suggestions for future research in this area. Two of the reviewed articles described translational research in which pigeons were used to illustrate and examine behavioral phenomena of applied significance (say-do correspondence and fluency), 3 described interventions that changed animals' behavior (self-injury by a baboon, feces throwing and spitting by a chimpanzee, and unsafe trailer entry by horses) in ways that benefited the animals and the people in charge of them, and 1 described the use of trained rats that performed a service to humans (land-mine detection). We suggest that each of these general research areas merits further attention and that the Journal of Applied Behavior Analysis is an appropriate outlet for some of these publications.

  20. Applied Behavior Analysis Is a Science And, Therefore, Progressive

    ERIC Educational Resources Information Center

    Leaf, Justin B.; Leaf, Ronald; McEachin, John; Taubman, Mitchell; Ala'i-Rosales, Shahla; Ross, Robert K.; Smith, Tristram; Weiss, Mary Jane

    2016-01-01

    Applied behavior analysis (ABA) is a science and, therefore, involves progressive approaches and outcomes. In this commentary we argue that the spirit and the method of science should be maintained in order to avoid reductionist procedures, stifled innovation, and rote, unresponsive protocols that become increasingly removed from meaningful…

  1. The spread of behavior analysis to the applied fields 1

    PubMed Central

    Fraley, Lawrence E.

    1981-01-01

    This paper reviews the status of applied behavioral science as it exists in the various behavioral fields and considers the role of the Association for Behavior Analysis in serving those fields. The confounding effects of the traditions of psychology are discussed. Relevant issues are exemplified in the fields of law, communications, psychology, and education, but broader generalization is implied. PMID:22478537

  2. Applied Behavior Analysis Is a Science And, Therefore, Progressive

    ERIC Educational Resources Information Center

    Leaf, Justin B.; Leaf, Ronald; McEachin, John; Taubman, Mitchell; Ala'i-Rosales, Shahla; Ross, Robert K.; Smith, Tristram; Weiss, Mary Jane

    2016-01-01

    Applied behavior analysis (ABA) is a science and, therefore, involves progressive approaches and outcomes. In this commentary we argue that the spirit and the method of science should be maintained in order to avoid reductionist procedures, stifled innovation, and rote, unresponsive protocols that become increasingly removed from meaningful…

  3. Three Years of Intensive Applied Behavior Analysis: A Case Study

    ERIC Educational Resources Information Center

    Healy, Olive; O'Connor, Jennifer; Leader, Geraldine; Kenny, Neil

    2008-01-01

    A 2 years and 10 months old child receiving an early intensive teaching program using the Comprehensive Application of Behavior Analysis to Schooling (CABAS[R]) was taught using evidence-based teaching strategies and curricula based on existing research in both the applied and basic sciences. Progress was measured using behavioral assessment tools…

  4. Overview af MSFC's Applied Fluid Dynamics Analysis Group Activities

    NASA Technical Reports Server (NTRS)

    Garcia, Roberto; Griffin, Lisa; Williams, Robert

    2004-01-01

    This paper presents viewgraphs on NASA Marshall Space Flight Center's Applied Fluid Dynamics Analysis Group Activities. The topics include: 1) Status of programs at MSFC; 2) Fluid Mechanics at MSFC; 3) Relevant Fluid Dynamics Activities at MSFC; and 4) Shuttle Return to Flight.

  5. Applied Behavior Analysis: Current Myths in Public Education

    ERIC Educational Resources Information Center

    Fielding, Cheryl; Lowdermilk, John; Lanier, Lauren L.; Fannin, Abigail G.; Schkade, Jennifer L.; Rose, Chad A.; Simpson, Cynthia G.

    2013-01-01

    The effective use of behavior management strategies and related policies continues to be a debated issue in public education. Despite overwhelming evidence espousing the benefits of the implementation of procedures derived from principles based on the science of applied behavior analysis (ABA), educators often indicate many common misconceptions…

  6. Positive Behavior Support and Applied Behavior Analysis: A Familial Alliance

    ERIC Educational Resources Information Center

    Dunlap, Glen; Carr, Edward G.; Horner, Robert H.; Zarcone, Jennifer R.; Schwartz, Ilene

    2008-01-01

    Positive behavior support (PBS) emerged in the mid-1980s as an approach for understanding and addressing problem behaviors. PBS was derived primarily from applied behavior analysis (ABA). Over time, however, PBS research and practice has incorporated evaluative methods, assessment and intervention procedures, and conceptual perspectives associated…

  7. Positive Behavior Support and Applied Behavior Analysis: A Familial Alliance

    ERIC Educational Resources Information Center

    Dunlap, Glen; Carr, Edward G.; Horner, Robert H.; Zarcone, Jennifer R.; Schwartz, Ilene

    2008-01-01

    Positive behavior support (PBS) emerged in the mid-1980s as an approach for understanding and addressing problem behaviors. PBS was derived primarily from applied behavior analysis (ABA). Over time, however, PBS research and practice has incorporated evaluative methods, assessment and intervention procedures, and conceptual perspectives associated…

  8. B. F. Skinner's Contributions to Applied Behavior Analysis

    ERIC Educational Resources Information Center

    Morris, Edward K.; Smith, Nathaniel G.; Altus, Deborah E.

    2005-01-01

    Our paper reviews and analyzes B. F. Skinner's contributions to applied behavior analysis in order to assess his role as the field's originator and founder. We found, first, that his contributions fall into five categorizes: the style and content of his science, his interpretations of typical and atypical human behavior, the implications he drew…

  9. Progressive-Ratio Schedules and Applied Behavior Analysis

    ERIC Educational Resources Information Center

    Poling, Alan

    2010-01-01

    Establishing appropriate relations between the basic and applied areas of behavior analysis has been of long and persistent interest to the author. In this article, the author illustrates that there is a direct relation between how hard an organism will work for access to an object or activity, as indexed by the largest ratio completed under a…

  10. We Are Family: Applying Family Systems Theory to Classrooms.

    ERIC Educational Resources Information Center

    Braun, Joseph A., Jr.; Garrett, Joyce Lynn

    1988-01-01

    Describes how counselors can apply the family systems model to classrooms in helping teachers create a more open and effective climate. Discusses these strategies for implementation: basic communication and interpersonal skills, fairness conferences, classroom meetings, magic circle and circle of warmth, and role playing. (Author/ABL)

  11. Applied Information Systems Research Program (AISRP) Workshop 3 meeting proceedings

    NASA Technical Reports Server (NTRS)

    1993-01-01

    The third Workshop of the Applied Laboratory Systems Research Program (AISRP) met at the Univeristy of Colorado's Laboratory for Atmospheric and Space Physics in August of 1993. The presentations were organized into four sessions: Artificial Intelligence Techniques; Scientific Visualization; Data Management and Archiving; and Research and Technology.

  12. Associate of Applied Science Degree in Office Systems. Proposal.

    ERIC Educational Resources Information Center

    Gallaudet Coll., Washington, DC. School of Preparatory Studies.

    This proposal culminates a 5-year study of the possibility of awarding associate degrees at Gallaudet College, a private, liberal arts college for hearing impaired adults. The proposal outlines an Associate of Applied Science degree (AAS) in Office Systems at the School of Preparatory Studies. First, introductory material provides a brief history…

  13. We Are Family: Applying Family Systems Theory to Classrooms.

    ERIC Educational Resources Information Center

    Braun, Joseph A., Jr.; Garrett, Joyce Lynn

    1988-01-01

    Describes how counselors can apply the family systems model to classrooms in helping teachers create a more open and effective climate. Discusses these strategies for implementation: basic communication and interpersonal skills, fairness conferences, classroom meetings, magic circle and circle of warmth, and role playing. (Author/ABL)

  14. Recent reinforcement-schedule research and applied behavior analysis

    PubMed Central

    Lattal, Kennon A.; Neef, Nancy A.

    1996-01-01

    Reinforcement schedules are considered in relation to applied behavior analysis by examining several recent laboratory experiments with humans and other animals. The experiments are drawn from three areas of contemporary schedule research: behavioral history effects on schedule performance, the role of instructions in schedule performance of humans, and dynamic schedules of reinforcement. All of the experiments are discussed in relation to the role of behavioral history in current schedule performance. The paper concludes by extracting from the experiments some more general issues concerning reinforcement schedules in applied research and practice. PMID:16795888

  15. Applying Technology Ranking and Systems Engineering in Advanced Life Support

    NASA Technical Reports Server (NTRS)

    Jones, Harry; Luna, Bernadette (Technical Monitor)

    2000-01-01

    According to the Advanced Life Support (ALS) Program Plan, the Systems Modeling and Analysis Project (SMAP) has two important tasks: 1) prioritizing investments in ALS Research and Technology Development (R&TD), and 2) guiding the evolution of ALS systems. Investments could be prioritized simply by independently ranking different technologies, but we should also consider a technology's impact on system design. Guiding future ALS systems will require SMAP to consider many aspects of systems engineering. R&TD investments can be prioritized using familiar methods for ranking technology. The first step is gathering data on technology performance, safety, readiness level, and cost. Then the technologies are ranked using metrics or by decision analysis using net present economic value. The R&TD portfolio can be optimized to provide the maximum expected payoff in the face of uncertain future events. But more is needed. The optimum ALS system can not be designed simply by selecting the best technology for each predefined subsystem. Incorporating a new technology, such as food plants, can change the specifications of other subsystems, such as air regeneration. Systems must be designed top-down starting from system objectives, not bottom-up from selected technologies. The familiar top-down systems engineering process includes defining mission objectives, mission design, system specification, technology analysis, preliminary design, and detail design. Technology selection is only one part of systems analysis and engineering, and it is strongly related to the subsystem definitions. ALS systems should be designed using top-down systems engineering. R&TD technology selection should consider how the technology affects ALS system design. Technology ranking is useful but it is only a small part of systems engineering.

  16. Applying Technology Ranking and Systems Engineering in Advanced Life Support

    NASA Technical Reports Server (NTRS)

    Jones, Harry; Luna, Bernadette (Technical Monitor)

    2000-01-01

    According to the Advanced Life Support (ALS) Program Plan, the Systems Modeling and Analysis Project (SMAP) has two important tasks: 1) prioritizing investments in ALS Research and Technology Development (R&TD), and 2) guiding the evolution of ALS systems. Investments could be prioritized simply by independently ranking different technologies, but we should also consider a technology's impact on system design. Guiding future ALS systems will require SMAP to consider many aspects of systems engineering. R&TD investments can be prioritized using familiar methods for ranking technology. The first step is gathering data on technology performance, safety, readiness level, and cost. Then the technologies are ranked using metrics or by decision analysis using net present economic value. The R&TD portfolio can be optimized to provide the maximum expected payoff in the face of uncertain future events. But more is needed. The optimum ALS system can not be designed simply by selecting the best technology for each predefined subsystem. Incorporating a new technology, such as food plants, can change the specifications of other subsystems, such as air regeneration. Systems must be designed top-down starting from system objectives, not bottom-up from selected technologies. The familiar top-down systems engineering process includes defining mission objectives, mission design, system specification, technology analysis, preliminary design, and detail design. Technology selection is only one part of systems analysis and engineering, and it is strongly related to the subsystem definitions. ALS systems should be designed using top-down systems engineering. R&TD technology selection should consider how the technology affects ALS system design. Technology ranking is useful but it is only a small part of systems engineering.

  17. Recent results of nonlinear estimators applied to hereditary systems.

    NASA Technical Reports Server (NTRS)

    Schiess, J. R.; Roland, V. R.; Wells, W. R.

    1972-01-01

    An application of the extended Kalman filter to delayed systems to estimate the state and time delay is presented. Two nonlinear estimators are discussed and the results compared with those of the Kalman filter. For all the filters considered, the hereditary system was treated with the delay in the pure form and by using Pade approximations of the delay. A summary of the convergence properties of the filters studied is given. The results indicate that the linear filter applied to the delayed system performs inadequately while the nonlinear filters provide reasonable estimates of both the state and the parameters.

  18. Recent results of nonlinear estimators applied to hereditary systems.

    NASA Technical Reports Server (NTRS)

    Schiess, J. R.; Roland, V. R.; Wells, W. R.

    1972-01-01

    An application of the extended Kalman filter to delayed systems to estimate the state and time delay is presented. Two nonlinear estimators are discussed and the results compared with those of the Kalman filter. For all the filters considered, the hereditary system was treated with the delay in the pure form and by using Pade approximations of the delay. A summary of the convergence properties of the filters studied is given. The results indicate that the linear filter applied to the delayed system performs inadequately while the nonlinear filters provide reasonable estimates of both the state and the parameters.

  19. Validation in principal components analysis applied to EEG data.

    PubMed

    Costa, João Carlos G D; Da-Silva, Paulo José G; Almeida, Renan Moritz V R; Infantosi, Antonio Fernando C

    2014-01-01

    The well-known multivariate technique Principal Components Analysis (PCA) is usually applied to a sample, and so component scores are subjected to sampling variability. However, few studies address their stability, an important topic when the sample size is small. This work presents three validation procedures applied to PCA, based on confidence regions generated by a variant of a nonparametric bootstrap called the partial bootstrap: (i) the assessment of PC scores variability by the spread and overlapping of "confidence regions" plotted around these scores; (ii) the use of the confidence regions centroids as a validation set; and (iii) the definition of the number of nontrivial axes to be retained for analysis. The methods were applied to EEG data collected during a postural control protocol with twenty-four volunteers. Two axes were retained for analysis, with 91.6% of explained variance. Results showed that the area of the confidence regions provided useful insights on the variability of scores and suggested that some subjects were not distinguishable from others, which was not evident from the principal planes. In addition, potential outliers, initially suggested by an analysis of the first principal plane, could not be confirmed by the confidence regions.

  20. Circulation control technology applied to propulsive high lift systems

    NASA Technical Reports Server (NTRS)

    Englar, R. J.; Nichols, J. H., Jr.; Harris, M. J.; Eppel, J. C.; Shovlin, M. D.

    1984-01-01

    Technology developed for the Circulation Control Wing high-lift system has been extended to augment lift by entraining and redirecting engine thrust. Ejecting a thin jet sheet tangentially over a small curved deflecting surface adjacent to the slipstream of a turbofan engine causes the slipstream to flow around that deflecting surface. The angle of deflection is controlled pneumatically by varying the momentum of the thin jet sheet. The downward momentum of the slipstream enhances wing lift. This concept of pneumatically deflecting the slipstream has been applied to an upper surface blowing high-lift system and to a thrust deflecting system. The capability of the pneumatic upper surface blowing system was demonstrated in a series of investigations using a wind tunnel model and the NASA Quiet Short-haul Research Aircraft (QSRA). Full-scale thrust deflections greater than 90 deg were achieved. This mechanically simple system can provide increased maneuverability, heavy lift or overload capability, or short takeoff and landing performance.

  1. Circulation control technology applied to propulsive high lift systems

    NASA Technical Reports Server (NTRS)

    Englar, R. J.; Nichols, J. H., Jr.; Harris, M. J.; Eppel, J. C.; Shovlin, M. D.

    1984-01-01

    Technology developed for the Circulation Control Wing high-lift system has been extended to augment lift by entraining and redirecting engine thrust. Ejecting a thin jet sheet tangentially over a small curved deflecting surface adjacent to the slipstream of a turbofan engine causes the slipstream to flow around that deflecting surface. The angle of deflection is controlled pneumatically by varying the momentum of the thin jet sheet. The downward momentum of the slipstream enhances wing lift. This concept of pneumatically deflecting the slipstream has been applied to an upper surface blowing high-lift system and to a thrust deflecting system. The capability of the pneumatic upper surface blowing system was demonstrated in a series of investigations using a wind tunnel model and the NASA Quiet Short-haul Research Aircraft (QSRA). Full-scale thrust deflections greater than 90 deg were achieved. This mechanically simple system can provide increased maneuverability, heavy lift or overload capability, or short takeoff and landing performance.

  2. Boston Society's 11th Annual Applied Pharmaceutical Analysis conference.

    PubMed

    Lee, Violet; Liu, Ang; Groeber, Elizabeth; Moghaddam, Mehran; Schiller, James; Tweed, Joseph A; Walker, Gregory S

    2016-02-01

    Boston Society's 11th Annual Applied Pharmaceutical Analysis conference, Hyatt Regency Hotel, Cambridge, MA, USA, 14-16 September 2015 The Boston Society's 11th Annual Applied Pharmaceutical Analysis (APA) conference took place at the Hyatt Regency hotel in Cambridge, MA, on 14-16 September 2015. The 3-day conference affords pharmaceutical professionals, academic researchers and industry regulators the opportunity to collectively participate in meaningful and relevant discussions impacting the areas of pharmaceutical drug development. The APA conference was organized in three workshops encompassing the disciplines of regulated bioanalysis, discovery bioanalysis (encompassing new and emerging technologies) and biotransformation. The conference included a short course titled 'Bioanalytical considerations for the clinical development of antibody-drug conjugates (ADCs)', an engaging poster session, several panel and round table discussions and over 50 diverse talks from leading industry and academic scientists.

  3. Stratospheric Data Analysis System (STRATAN)

    NASA Technical Reports Server (NTRS)

    Rood, Richard B.; Fox-Rabinovitz, Michael; Lamich, David J.; Newman, Paul A.; Pfaendtner, James W.

    1990-01-01

    A state of the art stratospheric analyses using a coupled stratosphere/troposphere data assimilation system is produced. These analyses can be applied to stratospheric studies of all types. Of importance to this effort is the application of the Stratospheric Data Analysis System (STRATAN) to constituent transport and chemistry problems.

  4. Developing an interdisciplinary master's program in applied behavior analysis

    PubMed Central

    Lowenkron, Barry; Mitchell, Lynda

    1995-01-01

    At many universities, faculty interested in behavior analysis are spread across disciplines. This makes difficult the development of behavior-analytically oriented programs, and impedes regular contact among colleagues who share common interests. However, this separation by disciplines can be a source of strength if it is used to develop interdisciplinary programs. In this article we describe how a bottom-up strategy was used to develop two complementary interdisciplinary MS programs in applied behavior analysis, and conclude with a description of the benefits—some obvious, some surprising—that can emerge from the development of such programs. PMID:22478230

  5. Nuclear safety as applied to space power reactor systems

    SciTech Connect

    Cummings, G.E.

    1987-01-01

    To develop a strategy for incorporating and demonstrating safety, it is necessary to enumerate the unique aspects of space power reactor systems from a safety standpoint. These features must be differentiated from terrestrial nuclear power plants so that our experience can be applied properly. Some ideas can then be developed on how safe designs can be achieved so that they are safe and perceived to be safe by the public. These ideas include operating only after achieving a stable orbit, developing an inherently safe design, ''designing'' in safety from the start and managing the system development (design) so that it is perceived safe. These and other ideas are explored further in this paper.

  6. Discrete Event Supervisory Control Applied to Propulsion Systems

    NASA Technical Reports Server (NTRS)

    Litt, Jonathan S.; Shah, Neerav

    2005-01-01

    The theory of discrete event supervisory (DES) control was applied to the optimal control of a twin-engine aircraft propulsion system and demonstrated in a simulation. The supervisory control, which is implemented as a finite-state automaton, oversees the behavior of a system and manages it in such a way that it maximizes a performance criterion, similar to a traditional optimal control problem. DES controllers can be nested such that a high-level controller supervises multiple lower level controllers. This structure can be expanded to control huge, complex systems, providing optimal performance and increasing autonomy with each additional level. The DES control strategy for propulsion systems was validated using a distributed testbed consisting of multiple computers--each representing a module of the overall propulsion system--to simulate real-time hardware-in-the-loop testing. In the first experiment, DES control was applied to the operation of a nonlinear simulation of a turbofan engine (running in closed loop using its own feedback controller) to minimize engine structural damage caused by a combination of thermal and structural loads. This enables increased on-wing time for the engine through better management of the engine-component life usage. Thus, the engine-level DES acts as a life-extending controller through its interaction with and manipulation of the engine s operation.

  7. Photoeletrolytic system applied to remazol red brilliant degradation.

    PubMed

    Sousa, Mariana Lopes; Moraes, Peterson Bueno; Bidoia, Ederio Dino

    2011-01-01

    Toxicity tests using Sacharomycces cerevisiae were made with simulated textile effluents containing reactive dye (remazol red brilliant) treated by photoeletrolytic process, varying treatment time and applied current. The treatment incorporated an electrolytic reactor with rectangular titanium anode coated with 70% TiO(2)/30% RuO(2) cathode and a rectangular stainless steel coupled with another photolytic reactor containing a high power UV lamp. The treatment system was used in batch recirculation, in other words, the simulated effluent was driven by the system through a helical pump. It was observed that the higher the value of current applied, the longer the treatment has greater color removal of textile effluent and higher mortality of S. cerevisiae, killing up to 100% of the cells at the end of the treatment. With a lower current applied and having the treatment time of 5 minutes, the effluent showed a color removal of 97% and a lower mortality of S. cerevisiae than the effluent simulated without any treatment.

  8. Differential item functioning analysis by applying multiple comparison procedures.

    PubMed

    Eusebi, Paolo; Kreiner, Svend

    2015-01-01

    Analysis within a Rasch measurement framework aims at development of valid and objective test score. One requirement of both validity and objectivity is that items do not show evidence of differential item functioning (DIF). A number of procedures exist for the assessment of DIF including those based on analysis of contingency tables by Mantel-Haenszel tests and partial gamma coefficients. The aim of this paper is to illustrate Multiple Comparison Procedures (MCP) for analysis of DIF relative to a variable defining a very large number of groups, with an unclear ordering with respect to the DIF effect. We propose a single step procedure controlling the false discovery rate for DIF detection. The procedure applies for both dichotomous and polytomous items. In addition to providing evidence against a hypothesis of no DIF, the procedure also provides information on subset of groups that are homogeneous with respect to the DIF effect. A stepwise MCP procedure for this purpose is also introduced.

  9. Applied research on extension element model in hydraulic system

    NASA Astrophysics Data System (ADS)

    Tian, Junying; Wei, Bingyang; Han, Jianhai

    2010-08-01

    This paper further researches the application of extension element model of hydraulic system. A extension element model of hydraulic system contains three parts: the prototype of a hydraulic component, the matter-element and relationelement model of hydraulic components, and the matter-element and relation-element model of hydraulic systems. The foundation building a extension element model of hydraulic system is to build its database of the prototypes of a hydraulic component. The matter-element and relation-element model of hydraulic components is a model for physical hydraulic components. The the matter-element and relation-element models of hydraulic components on a hydraulic system can describe the internal links among hydraulic components. Through analysis of the connections among the matter-element and relation-element models of hydraulic components, rings, power bond graphs, and dynamic equations will be found and gained. Thus, the performance and digital simulation for a hydraulic system will be realized and finished.

  10. Activity anorexia: An interplay between basic and applied behavior analysis

    PubMed Central

    Pierce, W. David; Epling, W. Frank; Dews, Peter B.; Estes, William K.; Morse, William H.; Van Orman, Willard; Herrnstein, Richard J.

    1994-01-01

    The relationship between basic research with nonhumans and applied behavior analysis is illustrated by our work on activity anorexia. When rats are fed one meal a day and allowed to run on an activity wheel, they run excessively, stop eating, and die of starvation. Convergent evidence, from several different research areas, indicates that the behavior of these animals and humans who self-starve is functionally similar. A biobehavioral theory of activity anorexia is presented that details the cultural contingencies, behavioral processes, and physiology of anorexia. Diagnostic criteria and a three-stage treatment program for activity-based anorexia are outlined. The animal model permits basic research on anorexia that for practical and ethical reasons cannot be conducted with humans. Thus, basic research can have applied importance. PMID:22478169

  11. Activity anorexia: An interplay between basic and applied behavior analysis.

    PubMed

    Pierce, W D; Epling, W F; Dews, P B; Estes, W K; Morse, W H; Van Orman, W; Herrnstein, R J

    1994-01-01

    The relationship between basic research with nonhumans and applied behavior analysis is illustrated by our work on activity anorexia. When rats are fed one meal a day and allowed to run on an activity wheel, they run excessively, stop eating, and die of starvation. Convergent evidence, from several different research areas, indicates that the behavior of these animals and humans who self-starve is functionally similar. A biobehavioral theory of activity anorexia is presented that details the cultural contingencies, behavioral processes, and physiology of anorexia. Diagnostic criteria and a three-stage treatment program for activity-based anorexia are outlined. The animal model permits basic research on anorexia that for practical and ethical reasons cannot be conducted with humans. Thus, basic research can have applied importance.

  12. Enhancing implementation science by applying best principles of systems science.

    PubMed

    Northridge, Mary E; Metcalf, Sara S

    2016-10-04

    Implementation science holds promise for better ensuring that research is translated into evidence-based policy and practice, but interventions often fail or even worsen the problems they are intended to solve due to a lack of understanding of real world structures and dynamic complexity. While systems science alone cannot possibly solve the major challenges in public health, systems-based approaches may contribute to changing the language and methods for conceptualising and acting within complex systems. The overarching goal of this paper is to improve the modelling used in dissemination and implementation research by applying best principles of systems science. Best principles, as distinct from the more customary term 'best practices', are used to underscore the need to extract the core issues from the context in which they are embedded in order to better ensure that they are transferable across settings. Toward meaningfully grappling with the complex and challenging problems faced in adopting and integrating evidence-based health interventions and changing practice patterns within specific settings, we propose and illustrate four best principles derived from our systems science experience: (1) model the problem, not the system; (2) pay attention to what is important, not just what is quantifiable; (3) leverage the utility of models as boundary objects; and (4) adopt a portfolio approach to model building. To improve our mental models of the real world, system scientists have created methodologies such as system dynamics, agent-based modelling, geographic information science and social network simulation. To understand dynamic complexity, we need the ability to simulate. Otherwise, our understanding will be limited. The practice of dynamic systems modelling, as discussed herein, is the art and science of linking system structure to behaviour for the purpose of changing structure to improve behaviour. A useful computer model creates a knowledge repository and a

  13. Systems Analysis and Education Research Literature Review.

    ERIC Educational Resources Information Center

    Geisinger, Robert W.

    Systems approaches, developed in World War II for military and business operations, have been applied increasingly to educational affairs. Educational systems analysis has received widespread usage in finance and accounting, and has also been successfully applied to information systems, instructional systems development, school design and…

  14. Expert systems applied to two problems in nuclear power plants

    SciTech Connect

    Kim, K.Y.

    1988-01-01

    This dissertation describes two prototype expert systems applied to two problems in nuclear power plants. One problem is spare parts inventory control, and the other one is radionuclide release from containment during severe accident. The expert system for spare parts inventory control can handle spare parts requirements not only in corrective, preventive, or predictive maintenance, but also when failure rates of components or parts are updated by new data. Costs and benefits of spare parts inventory acquisition are evaluated with qualitative attributes such as spare part availability to provide the inventory manager with an improved basis for decision making. The expert system is implemented with Intelligence/Compiler on an IBM-AT. The other expert system for radionuclide release from containment can estimate magnitude, type, location, and time of release of radioactive materials from containment during a severe accident nearly on line, based on the actual measured physical parameters such as temperature and pressure inside the containment. The expert system has a function to check the validation of sensor data. The expert system is implemented with KEE on a Symbolics LISP machine.

  15. Cladistic analysis applied to the classification of volcanoes

    NASA Astrophysics Data System (ADS)

    Hone, D. W. E.; Mahony, S. H.; Sparks, R. S. J.; Martin, K. T.

    2007-11-01

    Cladistics is a systematic method of classification that groups entities on the basis of sharing similar characteristics in the most parsimonious manner. Here cladistics is applied to the classification of volcanoes using a dataset of 59 Quaternary volcanoes and 129 volcanic edifices of the Tohoku region, Northeast Japan. Volcano and edifice characteristics recorded in the database include attributes of volcano size, chemical composition, dominant eruptive products, volcano morphology, dominant landforms, volcano age and eruptive history. Without characteristics related to time the volcanic edifices divide into two groups, with characters related to volcano size, dominant composition and edifice morphology being the most diagnostic. Analysis including time based characteristics yields four groups with a good correlation between these groups and the two groups from the analysis without time for 108 out of 129 volcanic edifices. Thus when characters are slightly changed the volcanoes still form similar groupings. Analysis of the volcanoes both with and without time yields three groups based on compositional, eruptive products and morphological characters. Spatial clusters of volcanic centres have been recognised in the Tohoku region by Tamura et al. ( Earth Planet Sci Lett 197:105 106, 2002). The groups identified by cladistic analysis are distributed unevenly between the clusters, indicating a tendency for individual clusters to form similar kinds of volcanoes with distinctive but coherent styles of volcanism. Uneven distribution of volcano types between clusters can be explained by variations in dominant magma compositions through time, which are reflected in eruption products and volcanic landforms. Cladistic analysis can be a useful tool for elucidating dynamic igneous processes that could be applied to other regions and globally. Our exploratory study indicates that cladistics has promise as a method for classifying volcanoes and potentially elucidating dynamic

  16. An overview of statistical decomposition techniques applied to complex systems

    PubMed Central

    Tuncer, Yalcin; Tanik, Murat M.; Allison, David B.

    2009-01-01

    The current state of the art in applied decomposition techniques is summarized within a comparative uniform framework. These techniques are classified by the parametric or information theoretic approaches they adopt. An underlying structural model common to all parametric approaches is outlined. The nature and premises of a typical information theoretic approach are stressed. Some possible application patterns for an information theoretic approach are illustrated. Composition is distinguished from decomposition by pointing out that the former is not a simple reversal of the latter. From the standpoint of application to complex systems, a general evaluation is provided. PMID:19724659

  17. Applied Information Systems Research Program (AISRP). Workshop 2: Meeting Proceedings

    NASA Technical Reports Server (NTRS)

    1992-01-01

    The Earth and space science participants were able to see where the current research can be applied in their disciplines and computer science participants could see potential areas for future application of computer and information systems research. The Earth and Space Science research proposals for the High Performance Computing and Communications (HPCC) program were under evaluation. Therefore, this effort was not discussed at the AISRP Workshop. OSSA's other high priority area in computer science is scientific visualization, with the entire second day of the workshop devoted to it.

  18. Applying AI tools to operational space environmental analysis

    NASA Technical Reports Server (NTRS)

    Krajnak, Mike; Jesse, Lisa; Mucks, John

    1995-01-01

    The U.S. Air Force and National Oceanic Atmospheric Agency (NOAA) space environmental operations centers are facing increasingly complex challenges meeting the needs of their growing user community. These centers provide current space environmental information and short term forecasts of geomagnetic activity. Recent advances in modeling and data access have provided sophisticated tools for making accurate and timely forecasts, but have introduced new problems associated with handling and analyzing large quantities of complex data. AI (Artificial Intelligence) techniques have been considered as potential solutions to some of these problems. Fielding AI systems has proven more difficult than expected, in part because of operational constraints. Using systems which have been demonstrated successfully in the operational environment will provide a basis for a useful data fusion and analysis capability. Our approach uses a general purpose AI system already in operational use within the military intelligence community, called the Temporal Analysis System (TAS). TAS is an operational suite of tools supporting data processing, data visualization, historical analysis, situation assessment and predictive analysis. TAS includes expert system tools to analyze incoming events for indications of particular situations and predicts future activity. The expert system operates on a knowledge base of temporal patterns encoded using a knowledge representation called Temporal Transition Models (TTM's) and an event database maintained by the other TAS tools. The system also includes a robust knowledge acquisition and maintenance tool for creating TTM's using a graphical specification language. The ability to manipulate TTM's in a graphical format gives non-computer specialists an intuitive way of accessing and editing the knowledge base. To support space environmental analyses, we used TAS's ability to define domain specific event analysis abstractions. The prototype system defines

  19. Clinical response of two brush-applied peroxide whitening systems.

    PubMed

    Barlow, Ashley; Gerlach, Robert W; Date, Robert F; Brennan, Kieran; Struzycka, Izabela; Kwiatkowska, Anna; Wierzbicka, Maria

    2003-01-01

    A randomized, examiner-blind parallel design study was conducted to evaluate the clinical efficacy and tolerability of two brush-applied, peroxide-based tooth whiteners. A total of 38 subjects were randomized to Colgate Simply White, an 18% carbamide peroxide paint-on liquid in an applicator bottle, or Crest Night Effects, a 19% sodium percarbonate system in unit dose sachets that dries to form an adherent film. Treatment was for 14 days. The 18% carbamide peroxide gel was applied twice daily, while the 19% sodium percarbonate film was applied for overnight use. Subjects were evaluated at baseline, and again after 7 and 14 days of treatment using digital images to assess tooth color, and oral examinations to establish safety. Whitening efficacy was determined by evaluating the reduction in tooth yellowness (delta b*), increase in tooth brightness (delta L*), reduction in tooth redness (delta a*) and overall tooth color change relative to pure white (delta W*). After 14 days, the mean change in yellowness (delta b*) was -1.43 +/- 0.20 for the 19% sodium percarbonate film, and -0.44 +/- 0.06 for the 18% carbamide peroxide gel. Mean change in brightness (delta L*) was 1.13 +/- 0.13 and 0.55 +/- 0.13 in the film and gel groups, respectively. Between-group comparisons showed highly significant (p < or = 0.003) whitening for the 19% sodium percarbonate film compared to the 18% carbamide peroxide gel for all 14-day color measurements (delta b*, delta L*, delta a* or delta W). Both products were well-tolerated, with no subjects discontinuing treatment early due to a causal adverse event. In a direct clinical comparison of two brush-applied products over 14 days, Crest Night Effects provided significant and meaningful improvement in tooth color, representing more than double the whitening measured for the Colgate Simply White.

  20. Research of the grid computing system applied in optical simulation

    NASA Astrophysics Data System (ADS)

    Jin, Wei-wei; Wang, Yu-dong; Liu, Qiangsheng; Cen, Zhao-feng; Li, Xiao-tong; Lin, Yi-qun

    2008-03-01

    A grid computing in the field of optics is presented in this paper. Firstly, the basic principles and research background of grid computing are outlined in this paper, along with the overview of its applications and the development status quo. The paper also discusses several typical tasks scheduling algorithms. Secondly, it focuses on describing a task scheduling of grid computing applied in optical computation. The paper gives details about the task scheduling system, including the task partition, granularity selection and tasks allocation, especially the structure of the system. In addition, some details of communication on grid computing are also illustrated. In this system, the "makespan" and "load balancing" are comprehensively considered. Finally, we build a grid model to test the task scheduling strategy, and the results are analyzed in detail. Compared to one isolated computer, a grid comprised of one server and four processors can shorten the "makespan" to 1/4. At the same time, the experimental results of the simulation also illustrate that the proposed scheduling system is able to balance loads of all processors. In short, the system performs scheduling well in the grid environment.

  1. Systems engineering and analysis

    SciTech Connect

    Blanchard, B.S.; Fabrycky, W.J.

    1981-01-01

    An introduction to systems is provided and tools for systems analysis are considered, taking into account system definitions and concepts, approaches for bringing systems into being, models in systems analysis, economic analysis techniques, mathematical modeling and optimization, probability and statistics, queuing theory and analysis, and control concepts and techniques. The system design process is discussed along with the design for operational feasibility, systems engineering management, and system design case studies. Attention is given to conceptual design, preliminary system design, detail design and development, system test and evaluation, design for reliability, design for maintainability, design for supportability, design for economic feasibility, communication system design, finite population system design, energy storage system design, and procurement-inventory system design.

  2. Resonant-state expansion applied to planar open optical systems

    NASA Astrophysics Data System (ADS)

    Doost, M. B.; Langbein, W.; Muljarov, E. A.

    2012-02-01

    The resonant-state expansion (RSE), a rigorous perturbation theory of the Brillouin-Wigner type recently developed in electrodynamics[E. A. Muljarov, W. Langbein, and R. Zimmermann, Europhys. Lett.EULEEJ0295-507510.1209/0295-5075/92/50010 92, 50010 (2010)], is applied to planar, effectively one-dimensional optical systems, such as layered dielectric slabs and Bragg reflector microcavities. It is demonstrated that the RSE converges with a power law in the basis size. Algorithms for error estimation and their reduction by extrapolation are presented and evaluated. Complex eigenfrequencies, electromagnetic fields, and the Green's function of a selection of optical systems are calculated, as well as the observable transmission spectra. In particular, we find that for a Bragg-mirror microcavity, which has sharp resonances in the spectrum, the transmission calculated using the RSE reproduces the result of the transfer- or scattering-matrix method.

  3. An improved AVC strategy applied in distributed wind power system

    NASA Astrophysics Data System (ADS)

    Zhao, Y. N.; Liu, Q. H.; Song, S. Y.; Mao, W.

    2016-08-01

    Traditional AVC strategy is mainly used in wind farm and only concerns about grid connection point, which is not suitable for distributed wind power system. Therefore, this paper comes up with an improved AVC strategy applied in distributed wind power system. The strategy takes all nodes of distribution network into consideration and chooses the node having the most serious voltage deviation as control point to calculate the reactive power reference. In addition, distribution principles can be divided into two conditions: when wind generators access to network on single node, the reactive power reference is distributed according to reactive power capacity; when wind generators access to network on multi-node, the reference is distributed according to sensitivity. Simulation results show the correctness and reliability of the strategy. Compared with traditional control strategy, the strategy described in this paper can make full use of generators reactive power output ability according to the distribution network voltage condition and improve the distribution network voltage level effectively.

  4. Systems Vaccinology Applied to DNA Vaccines: Perspective and Challenges.

    PubMed

    Lever, Melissa; Silveira, Eduardo L; Nakaya, Helder I

    2017-01-01

    DNA vaccination represents a new milestone in our technological efforts to avoid infectious diseases. Although this method of vaccination has had success in providing protection in animals, these vaccines suffer from low immunogenicity in humans. Questions remain over the molecular mechanism of DNA vaccination, the best ways in which to safely increase vaccine reactogenecity, and what biomarkers can be used as correlates of protection. Systems vaccinology, which utilizes modern experimental and computational approaches to provide an integrated view of the vaccination process, offers the potential to answer these questions. In this review we discuss the current tools utilized in systems vaccinology, the ways in which they have and can be applied to DNA vaccinology, and challenges faced in the field.

  5. Laser communication applied for EDRS, the European data relay system

    NASA Astrophysics Data System (ADS)

    Heine, F.; Kämpfner, H.; Lange, R.; Czichy, R.; Lutzer, M.; Meyer, R.

    2011-12-01

    Future earth observation satellites call for GEO relay links to make their data immediately available to the user. While RF communication limits the GEO relay's data rate to roughly 1 Gbps (Gigabit per second) laser communication will extend its capacity into the 10 Gbps range as is required for the future systems. Laser communication will be applied for EDRS, the European data GEO relay system. At first, it is foreseen to provide the RF LEO-to-GEO link with an additional laser communication channel, still limiting, of course, the GEO relay's performance to the RF bottleneck. However besides the operational service, the EDRS laser terminals shall also demonstrate the performance of high data rate links for both, the inter-satellite link and the GEO-to-ground link.

  6. Lamb eimeriosis: applied treatment protocols in dairy sheep production systems.

    PubMed

    Saratsis, Anastasios; Karagiannis, Isidoros; Brozos, Christos; Kiossis, Evagellos; Tzanidakis, Nikolaos; Joachim, Anja; Sotiraki, Smaragda

    2013-09-01

    The effect of different treatment protocols using the triazinone compounds diclazuril and toltrazuril on Eimeria oocyst excretion, diarrhoea and weight gain was evaluated in the present study. The protocols were designed in order to best fit to common management practices in dairy production systems as applied in Greece. During the first two trials comparative preventive (11 days after birth) and therapeutic (18 days after birth) single treatments using either drug were performed on an intensive farm. In Trial 3 the efficacy of a repeated diclazuril treatment after weaning (at the start and 4 weeks later) of the animals in a semi-intensive farm was tested. The last trial was performed in order to assess the effect of a single oral dose of toltrazuril after weaning of the animals on the same farm. During an observation period of 6-7 weeks after treatment animals in all trials were clinically examined for diarrhoea and faecal samples were regularly monitored for Eimeria oocysts at weekly intervals. Body weight was also determined at the start and end of each trial. A single treatment with toltrazuril resulted in a significant reduction of oocyst excretion over the study period compared to the control, with very high efficacy values during the first 2-3 weeks after treatment, irrespective of the treatment protocol and the management system applied. This in general could not be confirmed in the protocols using diclazuril, which showed a much lower and mostly variable efficacy pattern. Copyright © 2013 Elsevier B.V. All rights reserved.

  7. Shape analysis applied in heavy ion reactions near Fermi energy

    NASA Astrophysics Data System (ADS)

    Zhang, S.; Huang, M.; Wada, R.; Liu, X.; Lin, W.; Wang, J.

    2017-03-01

    A new method is proposed to perform shape analyses and to evaluate their validity in heavy ion collisions near the Fermi energy. In order to avoid erroneous values of shape parameters in the calculation, a test particle method is utilized in which each nucleon is represented by n test particles, similar to that used in the Boltzmann–Uehling–Uhlenbeck (BUU) calculations. The method is applied to the events simulated by an antisymmetrized molecular dynamics model. The geometrical shape of fragments is reasonably extracted when n = 100 is used. A significant deformation is observed for all fragments created in the multifragmentation process. The method is also applied to the shape of the momentum distribution for event classification. In the momentum case, the errors in the eigenvalue calculation become much smaller than those of the geometrical shape analysis and the results become similar between those with and without the test particle method, indicating that in intermediate heavy ion collisions the shape analysis of momentum distribution can be used for the event classification without the test particle method.

  8. Automated SEM Modal Analysis Applied to the Diogenites

    NASA Technical Reports Server (NTRS)

    Bowman, L. E.; Spilde, M. N.; Papike, James J.

    1996-01-01

    Analysis of volume proportions of minerals, or modal analysis, is routinely accomplished by point counting on an optical microscope, but the process, particularly on brecciated samples such as the diogenite meteorites, is tedious and prone to error by misidentification of very small fragments, which may make up a significant volume of the sample. Precise volume percentage data can be gathered on a scanning electron microscope (SEM) utilizing digital imaging and an energy dispersive spectrometer (EDS). This form of automated phase analysis reduces error, and at the same time provides more information than could be gathered using simple point counting alone, such as particle morphology statistics and chemical analyses. We have previously studied major, minor, and trace-element chemistry of orthopyroxene from a suite of diogenites. This abstract describes the method applied to determine the modes on this same suite of meteorites and the results of that research. The modal abundances thus determined add additional information on the petrogenesis of the diogenites. In addition, low-abundance phases such as spinels were located for further analysis by this method.

  9. Automated SEM Modal Analysis Applied to the Diogenites

    NASA Technical Reports Server (NTRS)

    Bowman, L. E.; Spilde, M. N.; Papike, James J.

    1996-01-01

    Analysis of volume proportions of minerals, or modal analysis, is routinely accomplished by point counting on an optical microscope, but the process, particularly on brecciated samples such as the diogenite meteorites, is tedious and prone to error by misidentification of very small fragments, which may make up a significant volume of the sample. Precise volume percentage data can be gathered on a scanning electron microscope (SEM) utilizing digital imaging and an energy dispersive spectrometer (EDS). This form of automated phase analysis reduces error, and at the same time provides more information than could be gathered using simple point counting alone, such as particle morphology statistics and chemical analyses. We have previously studied major, minor, and trace-element chemistry of orthopyroxene from a suite of diogenites. This abstract describes the method applied to determine the modes on this same suite of meteorites and the results of that research. The modal abundances thus determined add additional information on the petrogenesis of the diogenites. In addition, low-abundance phases such as spinels were located for further analysis by this method.

  10. Quantified MS analysis applied to combinatorial heterogeneous catalyst libraries.

    PubMed

    Wang, Hua; Liu, Zhongmin; Shen, Jianghan

    2003-01-01

    A high-throughput screening system for secondary catalyst libraries has been developed by incorporation of an 80-pass reactor and a quantified multistream mass spectrometer screening (MSMSS) technique. With a low-melting alloy as the heating medium, a uniform reaction temperature could be obtained in the multistream reactor (maximum temperature differences are less than 1 K at 673 K). Quantification of the results was realized by combination of a gas chromatogram with the MSMSS, which could provide the product selectivities of each catalyst in a heterogeneous catalyst library. Because the catalyst loading of each reaction tube is comparable to that of the conventional microreaction system and because the parallel reactions could be operated under identical conditions (homogeneous temperature, same pressure and WHSV), the reaction results of a promising catalyst selected from the library could be reasonably applied to the further scale-up of the system. The aldol condensation of acetone, with obvious differences in the product distribution over different kind of catalysts, was selected as a model reaction to validate the screening system.

  11. Hypercube expert system shell-applying production parallelism. Master's thesis

    SciTech Connect

    Harding, W.A.

    1989-12-01

    This research investigation proposes a hypercube design which supports efficient symbolic computing to permit real-time control of an air vehicle by an expert system. Design efforts are aimed at alleviating common expert system bottlenecks, such as the inefficiency of symbolic programming languages like Lisp and the disproportionate amount of computation time commonly spent in the match phase of the expert system match-select-act cycle. Faster processing of Robotic Air Vehicle (RAV) expert system software is approached through (1) fast production matching using the state-saving Rete match algorithm, (2) efficient shell implementation using the C-Programming Language and (3) parallel processing of the RAV using multiple copies of a serial expert system shell. In this investigation, the serial C-Language Integrated Production System shell is modified to execute in parallel on the iPSC/2 Hypercube. Speedups achieved using this architecture are quantified through theoretical timing analysis, and comparison with serial architecture performance results, with earlier designs performance results, with best case results and with goal performance.

  12. Applying of digital signal processing to optical equisignal zone system

    NASA Astrophysics Data System (ADS)

    Maraev, Anton A.; Timofeev, Aleksandr N.; Gusarov, Vadim F.

    2015-05-01

    In this work we are trying to assess the application of array detectors and digital information processing to the system with the optical equisignal zone as a new method of evaluating of optical equisignal zone position. Peculiarities of optical equisignal zone formation are described. The algorithm of evaluation of optical equisignal zone position is applied to processing on the array detector. This algorithm enables to evaluate as lateral displacement as turning angles of the receiver relative to the projector. Interrelation of parameters of the projector and the receiver is considered. According to described principles an experimental set was made and then characterized. The accuracy of position evaluation of the equisignal zone is shown dependent of the size of the equivalent entrance pupil at processing.

  13. Receiver function analysis applied to refraction survey data

    NASA Astrophysics Data System (ADS)

    Subaru, T.; Kyosuke, O.; Hitoshi, M.

    2008-12-01

    For the estimation of the thickness of oceanic crust or petrophysical investigation of subsurface material, refraction or reflection seismic exploration is one of the methods frequently practiced. These explorations use four-component (x,y,z component of acceleration and pressure) seismometer, but only compressional wave or vertical component of seismometers tends to be used in the analyses. Hence, it is needed to use shear wave or lateral component of seismograms for more precise investigation to estimate the thickness of oceanic crust. Receiver function is a function at a place that can be used to estimate the depth of velocity interfaces by receiving waves from teleseismic signal including shear wave. Receiver function analysis uses both vertical and horizontal components of seismograms and deconvolves the horizontal with the vertical to estimate the spectral difference of P-S converted waves arriving after the direct P wave. Once the phase information of the receiver function is obtained, then one can estimate the depth of the velocity interface. This analysis has advantage in the estimation of the depth of velocity interface including Mohorovicic discontinuity using two components of seismograms when P-to-S converted waves are generated at the interface. Our study presents results of the preliminary study using synthetic seismograms. First, we use three types of geological models that are composed of a single sediment layer, a crust layer, and a sloped Moho, respectively, for underground sources. The receiver function can estimate the depth and shape of Moho interface precisely for the three models. Second, We applied this method to synthetic refraction survey data generated not by earthquakes but by artificial sources on the ground or sea surface. Compressional seismic waves propagate under the velocity interface and radiate converted shear waves as well as at the other deep underground layer interfaces. However, the receiver function analysis applied to the

  14. Applied Behavior Analysis is a Science and, Therefore, Progressive.

    PubMed

    Leaf, Justin B; Leaf, Ronald; McEachin, John; Taubman, Mitchell; Ala'i-Rosales, Shahla; Ross, Robert K; Smith, Tristram; Weiss, Mary Jane

    2016-02-01

    Applied behavior analysis (ABA) is a science and, therefore, involves progressive approaches and outcomes. In this commentary we argue that the spirit and the method of science should be maintained in order to avoid reductionist procedures, stifled innovation, and rote, unresponsive protocols that become increasingly removed from meaningful progress for individuals diagnosed with autism spectrum disorder (ASD). We describe this approach as progressive. In a progressive approach to ABA, the therapist employs a structured yet flexible process, which is contingent upon and responsive to child progress. We will describe progressive ABA, contrast it to reductionist ABA, and provide rationales for both the substance and intent of ABA as a progressive scientific method for improving conditions of social relevance for individuals with ASD.

  15. Evaluating the effectiveness of teacher training in Applied Behaviour Analysis.

    PubMed

    Grey, Ian M; Honan, Rita; McClean, Brian; Daly, Michael

    2005-09-01

    Interventions for children with autism based upon Applied Behaviour Analysis (ABA) has been repeatedly shown to be related both to educational gains and to reductions in challenging behaviours. However, to date, comprehensive training in ABA for teachers and others have been limited. Over 7 months, 11 teachers undertook 90 hours of classroom instruction and supervision in ABA. Each teacher conducted a comprehensive functional assessment and designed a behaviour support plan targeting one behaviour for one child with an autistic disorder. Target behaviours included aggression, non-compliance and specific educational skills. Teachers recorded observational data for the target behaviour for both baseline and intervention sessions. Support plans produced an average 80 percent change in frequency of occurrence of target behaviours. Questionnaires completed by parents and teachers at the end of the course indicated a beneficial effect for the children and the educational environment. The potential benefits of teacher implemented behavioural intervention are discussed.

  16. Finite element analysis applied to dentoalveolar trauma: methodology description.

    PubMed

    da Silva, B R; Moreira Neto, J J S; da Silva, F I; de Aguiar, A S W

    2011-01-01

    Dentoalveolar traumatic injuries are among the clinical conditions most frequently treated in dental practice. However, few studies so far have addressed the biomechanical aspects of these events, probably as a result of difficulties in carrying out satisfactory experimental and clinical studies as well as the unavailability of truly scientific methodologies. The aim of this paper was to describe the use of finite element analysis applied to the biomechanical evaluation of dentoalveolar trauma. For didactic purposes, the methodological process was divided into steps that go from the creation of a geometric model to the evaluation of final results, always with a focus on methodological characteristics, advantages, and disadvantages, so as to allow the reader to customize the methodology according to specific needs. Our description shows that the finite element method can faithfully reproduce dentoalveolar trauma, provided the methodology is closely followed and thoroughly evaluated.

  17. Finite Element Analysis Applied to Dentoalveolar Trauma: Methodology Description

    PubMed Central

    da Silva, B. R.; Moreira Neto, J. J. S.; da Silva, F. I.; de Aguiar, A. S. W.

    2011-01-01

    Dentoalveolar traumatic injuries are among the clinical conditions most frequently treated in dental practice. However, few studies so far have addressed the biomechanical aspects of these events, probably as a result of difficulties in carrying out satisfactory experimental and clinical studies as well as the unavailability of truly scientific methodologies. The aim of this paper was to describe the use of finite element analysis applied to the biomechanical evaluation of dentoalveolar trauma. For didactic purposes, the methodological process was divided into steps that go from the creation of a geometric model to the evaluation of final results, always with a focus on methodological characteristics, advantages, and disadvantages, so as to allow the reader to customize the methodology according to specific needs. Our description shows that the finite element method can faithfully reproduce dentoalveolar trauma, provided the methodology is closely followed and thoroughly evaluated. PMID:21991463

  18. Statistical model applied to motor evoked potentials analysis.

    PubMed

    Ma, Ying; Thakor, Nitish V; Jia, Xiaofeng

    2011-01-01

    Motor evoked potentials (MEPs) convey information regarding the functional integrity of the descending motor pathways. Absence of the MEP has been used as a neurophysiological marker to suggest cortico-spinal abnormalities in the operating room. Due to their high variability and sensitivity, detailed quantitative studies of MEPs are lacking. This paper applies a statistical method to characterize MEPs by estimating the number of motor units and single motor unit potential amplitudes. A clearly increasing trend of single motor unit potential amplitudes in the MEPs after each pulse of the stimulation pulse train is revealed by this method. This statistical method eliminates the effects of anesthesia, and provides an objective assessment of MEPs. Consequently this statistical method has high potential to be useful in future quantitative MEPs analysis.

  19. Barometric altimetry system as virtual constellation applied in CAPS

    NASA Astrophysics Data System (ADS)

    Ai, Guoxiang; Sheng, Peixuan; Du, Jinlin; Zheng, Yongguang; Cai, Xiande; Wu, Haitao; Hu, Yonghui; Hua, Yu; Li, Xiaohui

    2009-03-01

    This work describes the barometric altimetry as virtual constellation applied to the Chinese Area Positioning System (CAPS), which uses the transponders of communication satellites to transfer navigation messages to users. Barometric altimetry depends on the relationship of air pressure varying with altitude in the Earth’s atmosphere. Once the air pressure at a location is measured the site altitude can be found. This method is able to enhance and improve the availability of three-dimensional positioning. The difficulty is that the relation between barometric pressure and altitude is variable in different areas and under various weather conditions. Hence, in order to obtain higher accuracy, we need to acquire the real-time air pressure corresponding to an altimetric region’s reference height. On the other hand, the altimetry method will be applied to satellite navigation system, but the greatest difficulty lies in how to get the real-time air pressure value at the reference height in the broad areas overlaid by satellite navigation. We propose an innovational method to solve this problem. It is to collect the real-time air pressures and temperatures of the 1860 known-altitude weather observatories over China and around via satellite communication and to carry out time extrapolation forecast uniformly. To reduce data quantity, we first partition the data and encode them and then broadcast these information via navigation message to CAPS users’ receivers. Upon the interpolations being done in receivers, the reference air pressure and temperature at the receiver’s nearby place is derived. Lastly, combing with the receiver-observed real air pressure and temperature, the site’s altitude can be determined. The work is presented in the following aspects: the calculation principle, formulae, data collection, encoding, prediction, interpolation method, navigation message transmission together with errors causes and analyses. The advantages and shortcomings of the

  20. Synchronisation and coupling analysis: applied cardiovascular physics in sleep medicine.

    PubMed

    Wessel, Niels; Riedl, Maik; Kramer, Jan; Muller, Andreas; Penzel, Thomas; Kurths, Jurgen

    2013-01-01

    Sleep is a physiological process with an internal program of a number of well defined sleep stages and intermediate wakefulness periods. The sleep stages modulate the autonomous nervous system and thereby the sleep stages are accompanied by different regulation regimes for the cardiovascular and respiratory system. The differences in regulation can be distinguished by new techniques of cardiovascular physics. The number of patients suffering from sleep disorders increases unproportionally with the increase of the human population and aging, leading to very high expenses in the public health system. Therefore, the challenge of cardiovascular physics is to develop highly-sophisticated methods which are able to, on the one hand, supplement and replace expensive medical devices and, on the other hand, improve the medical diagnostics with decreasing the patient's risk. Methods of cardiovascular physics are used to analyze heart rate, blood pressure and respiration to detect changes of the autonomous nervous system in different diseases. Data driven modeling analysis, synchronization and coupling analysis and their applications to biosignals in healthy subjects and patients with different sleep disorders are presented. Newly derived methods of cardiovascular physics can help to find indicators for these health risks.

  1. Robust regression applied to fractal/multifractal analysis.

    NASA Astrophysics Data System (ADS)

    Portilla, F.; Valencia, J. L.; Tarquis, A. M.; Saa-Requejo, A.

    2012-04-01

    Fractal and multifractal are concepts that have grown increasingly popular in recent years in the soil analysis, along with the development of fractal models. One of the common steps is to calculate the slope of a linear fit commonly using least squares method. This shouldn't be a special problem, however, in many situations using experimental data the researcher has to select the range of scales at which is going to work neglecting the rest of points to achieve the best linearity that in this type of analysis is necessary. Robust regression is a form of regression analysis designed to circumvent some limitations of traditional parametric and non-parametric methods. In this method we don't have to assume that the outlier point is simply an extreme observation drawn from the tail of a normal distribution not compromising the validity of the regression results. In this work we have evaluated the capacity of robust regression to select the points in the experimental data used trying to avoid subjective choices. Based on this analysis we have developed a new work methodology that implies two basic steps: • Evaluation of the improvement of linear fitting when consecutive points are eliminated based on R p-value. In this way we consider the implications of reducing the number of points. • Evaluation of the significance of slope difference between fitting with the two extremes points and fitted with the available points. We compare the results applying this methodology and the common used least squares one. The data selected for these comparisons are coming from experimental soil roughness transect and simulated based on middle point displacement method adding tendencies and noise. The results are discussed indicating the advantages and disadvantages of each methodology. Acknowledgements Funding provided by CEIGRAM (Research Centre for the Management of Agricultural and Environmental Risks) and by Spanish Ministerio de Ciencia e Innovación (MICINN) through project no

  2. Common cause evaluations in applied risk analysis of nuclear power plants. [PWR

    SciTech Connect

    Taniguchi, T.; Ligon, D.; Stamatelatos, M.

    1983-04-01

    Qualitative and quantitative approaches were developed for the evaluation of common cause failures (CCFs) in nuclear power plants and were applied to the analysis of the auxiliary feedwater systems of several pressurized water reactors (PWRs). Key CCF variables were identified through a survey of experts in the field and a review of failure experience in operating PWRs. These variables were classified into categories of high, medium, and low defense against a CCF. Based on the results, a checklist was developed for analyzing CCFs of systems. Several known techniques for quantifying CCFs were also reviewed. The information provided valuable insights in the development of a new model for estimating CCF probabilities, which is an extension of and improvement over the Beta Factor method. As applied to the analysis of the PWR auxiliary feedwater systems, the method yielded much more realistic values than the original Beta Factor method for a one-out-of-three system.

  3. Performance analysis of high quality parallel preconditioners applied to 3D finite element structural analysis

    SciTech Connect

    Kolotilina, L.; Nikishin, A.; Yeremin, A.

    1994-12-31

    The solution of large systems of linear equations is a crucial bottleneck when performing 3D finite element analysis of structures. Also, in many cases the reliability and robustness of iterative solution strategies, and their efficiency when exploiting hardware resources, fully determine the scope of industrial applications which can be solved on a particular computer platform. This is especially true for modern vector/parallel supercomputers with large vector length and for modern massively parallel supercomputers. Preconditioned iterative methods have been successfully applied to industrial class finite element analysis of structures. The construction and application of high quality preconditioners constitutes a high percentage of the total solution time. Parallel implementation of high quality preconditioners on such architectures is a formidable challenge. Two common types of existing preconditioners are the implicit preconditioners and the explicit preconditioners. The implicit preconditioners (e.g. incomplete factorizations of several types) are generally high quality but require solution of lower and upper triangular systems of equations per iteration which are difficult to parallelize without deteriorating the convergence rate. The explicit type of preconditionings (e.g. polynomial preconditioners or Jacobi-like preconditioners) require sparse matrix-vector multiplications and can be parallelized but their preconditioning qualities are less than desirable. The authors present results of numerical experiments with Factorized Sparse Approximate Inverses (FSAI) for symmetric positive definite linear systems. These are high quality preconditioners that possess a large resource of parallelism by construction without increasing the serial complexity.

  4. Advanced imaging systems for diagnostic investigations applied to Cultural Heritage

    NASA Astrophysics Data System (ADS)

    Peccenini, E.; Albertin, F.; Bettuzzi, M.; Brancaccio, R.; Casali, F.; Morigi, M. P.; Petrucci, F.

    2014-12-01

    The diagnostic investigations are an important resource in the studies on Cultural Heritage to enhance the knowledge on execution techniques, materials and conservation status of a work of art. In this field, due to the great historical and artistic value of the objects, preservation is the main concern; for this reason, new technological equipment has been designed and developed in the Physics Departments of the Universities of Ferrara and Bologna to enhance the non-invasive approach to the study of pictorial artworks and other objects of cultural interest. Infrared (IR) reflectography, X-ray radiography and computed tomography (CT), applied to works of art, are joined by the same goal: to get hidden information on execution techniques and inner structure pursuing the non-invasiveness of the methods, although using different setup and physical principles. In this work transportable imaging systems to investigate large objects in museums and galleries are presented. In particular, 2D scanning devices for IR reflectography and X-ray radiography, CT systems and some applications to the Cultural Heritage are described.

  5. A Hygrothermal Risk Analysis Applied to Residential Unvented Attics

    SciTech Connect

    Pallin, Simon B; Kehrer, Manfred

    2013-01-01

    Aresidential building, constructed with an unvented attic, is acommonroof assembly in the United States.The expected hygrothermal performance and service life of the roof are difficult to estimate due to a number of varying parameters.Typical parameters expected to vary are the climate, direction, and slope of the roof as well as the radiation properties of the surface material. Furthermore, influential parameters are indoor moisture excess, air leakages through the attic floor, and leakages from air-handling unit and ventilation ducts. In addition, the type of building materials such as the insulation material and closed or open cell spray polyurethane foam will influence the future performance of the roof. A development of a simulation model of the roof assembly will enable a risk and sensitivity analysis, in which the most important varying parameters on the hygrothermal performance can be determined. The model is designed to perform probabilistic simulations using mathematical and hygrothermal calculation tools. The varying input parameters can be chosen from existing measurements, simulations, or standards. An analysis is applied to determine the risk of consequences, such as mold growth, rot, or energy demand of the HVAC unit. Furthermore, the future performance of the roof can be simulated in different climates to facilitate the design of an efficient and reliable roof construction with the most suitable technical solution and to determine the most appropriate building materials for a given climate

  6. Applying Association Rule of the Data Mining Method for the Network Event Analysis

    NASA Astrophysics Data System (ADS)

    Kim, Wankyung; Soh, Wooyoung

    2007-12-01

    Network event analysis gives useful information on the network status that helps protect from attacks. It involves finding sets of frequently used packet information such as IP addresses and requires real-time processing by its nature. This paper applies association rules to network event analysis. Originally association rules used for data mining can be applied to find frequent item sets. So, if frequent items occur on networks, information system can guess that there is a threat. But existed association rules such as Apriori algorithm are not suitable for analyzing network events on real-time due to the high usage of CPU and memory and thus low processing speed. This paper develops a network event audit module by applying association rules to network events using a new algorithm instead of Apriori algorithm. Test results show that the application of the new algorithm gives drastically low usage of both CPU and memory for network event analysis compared with existing Apriori algorithm.

  7. The Evidence-Based Practice of Applied Behavior Analysis.

    PubMed

    Slocum, Timothy A; Detrich, Ronnie; Wilczynski, Susan M; Spencer, Trina D; Lewis, Teri; Wolfe, Katie

    2014-05-01

    Evidence-based practice (EBP) is a model of professional decision-making in which practitioners integrate the best available evidence with client values/context and clinical expertise in order to provide services for their clients. This framework provides behavior analysts with a structure for pervasive use of the best available evidence in the complex settings in which they work. This structure recognizes the need for clear and explicit understanding of the strength of evidence supporting intervention options, the important contextual factors including client values that contribute to decision making, and the key role of clinical expertise in the conceptualization, intervention, and evaluation of cases. Opening the discussion of EBP in this journal, Smith (The Behavior Analyst, 36, 7-33, 2013) raised several key issues related to EBP and applied behavior analysis (ABA). The purpose of this paper is to respond to Smith's arguments and extend the discussion of the relevant issues. Although we support many of Smith's (The Behavior Analyst, 36, 7-33, 2013) points, we contend that Smith's definition of EBP is significantly narrower than definitions that are used in professions with long histories of EBP and that this narrowness conflicts with the principles that drive applied behavior analytic practice. We offer a definition and framework for EBP that aligns with the foundations of ABA and is consistent with well-established definitions of EBP in medicine, psychology, and other professions. In addition to supporting the systematic use of research evidence in behavior analytic decision making, this definition can promote clear communication about treatment decisions across disciplines and with important outside institutions such as insurance companies and granting agencies.

  8. Multitaper Spectral Analysis and Wavelet Denoising Applied to Helioseismic Data

    NASA Technical Reports Server (NTRS)

    Komm, R. W.; Gu, Y.; Hill, F.; Stark, P. B.; Fodor, I. K.

    1999-01-01

    Estimates of solar normal mode frequencies from helioseismic observations can be improved by using Multitaper Spectral Analysis (MTSA) to estimate spectra from the time series, then using wavelet denoising of the log spectra. MTSA leads to a power spectrum estimate with reduced variance and better leakage properties than the conventional periodogram. Under the assumption of stationarity and mild regularity conditions, the log multitaper spectrum has a statistical distribution that is approximately Gaussian, so wavelet denoising is asymptotically an optimal method to reduce the noise in the estimated spectra. We find that a single m-upsilon spectrum benefits greatly from MTSA followed by wavelet denoising, and that wavelet denoising by itself can be used to improve m-averaged spectra. We compare estimates using two different 5-taper estimates (Stepian and sine tapers) and the periodogram estimate, for GONG time series at selected angular degrees l. We compare those three spectra with and without wavelet-denoising, both visually, and in terms of the mode parameters estimated from the pre-processed spectra using the GONG peak-fitting algorithm. The two multitaper estimates give equivalent results. The number of modes fitted well by the GONG algorithm is 20% to 60% larger (depending on l and the temporal frequency) when applied to the multitaper estimates than when applied to the periodogram. The estimated mode parameters (frequency, amplitude and width) are comparable for the three power spectrum estimates, except for modes with very small mode widths (a few frequency bins), where the multitaper spectra broadened the modest compared with the periodogram. We tested the influence of the number of tapers used and found that narrow modes at low n values are broadened to the extent that they can no longer be fit if the number of tapers is too large. For helioseismic time series of this length and temporal resolution, the optimal number of tapers is less than 10.

  9. Applying Lean: Implementation of a Rapid Triage and Treatment System

    PubMed Central

    Murrell, Karen L.; Offerman, Steven R.; Kauffman, Mark B.

    2011-01-01

    Objective: Emergency department (ED) crowding creates issues with patient satisfaction, long wait times and leaving the ED without being seen by a doctor (LWBS). Our objective was to evaluate how applying Lean principles to develop a Rapid Triage and Treatment (RTT) system affected ED metrics in our community hospital. Methods: Using Lean principles, we made ED process improvements that led to the RTT system. Using this system, patients undergo a rapid triage with low-acuity patients seen and treated by a physician in the triage area. No changes in staffing, physical space or hospital resources occurred during the study period. We then performed a retrospective, observational study comparing hospital electronic medical record data six months before and six months after implementation of the RTT system. Results: ED census was 30,981 in the six months prior to RTT and 33,926 after. Ambulance arrivals, ED patient acuity and hospital admission rates were unchanged throughout the study periods. Mean ED length of stay was longer in the period before RTT (4.2 hours, 95% confidence interval [CI] = 4.2–4.3; standard deviation [SD] = 3.9) than after (3.6 hours, 95% CI = 3.6–3.7; SD = 3.7). Mean ED arrival to physician start time was 62.2 minutes (95% CI = 61.5–63.0; SD = 58.9) prior to RTT and 41.9 minutes (95% CI = 41.5–42.4; SD = 30.9) after. The LWBS rate for the six months prior to RTT was 4.5% (95% CI = 3.1–5.5) and 1.5% (95% CI = 0.6–1.8) after RTT initiation. Conclusion: Our experience shows that changes in ED processes using Lean thinking and available resources can improve efficiency. In this community hospital ED, use of an RTT system decreased patient wait times and LWBS rates. PMID:21691524

  10. Data Systems Task Analysis.

    DTIC Science & Technology

    1979-08-01

    PAPER TAPES 63. NOTIFY CUSTOMER ENGINEERS (CE) OR TECHNICAL REPRESENTATIVES OF EQUIPMENT FAILURE 64. NOTIFY PRCGRAMMERS OR ANALYSIS CF PROCESSING...A08873 MARINE CORPS WASHINGTON DC F/B 5/9 DATA SYSTEMS TASK ANALYSIS . (U) "CLASSIFIEO. -%mm . LEVELIs DATA SYSTEMS O0 TASK ANALYSIS DCI) OO JF AUG 28...QUANTICO, VIRGINIA, 22134 "Ji (o t , - ’ i " DSYS)879 UNITED STATES MARINE CORPS TASK ANALYSIS PROGRAM QUESTICNNAIRE BOCKLET INTROCUCT ION YUU HAVE BEEN

  11. Nanostructured SBA-15 host applied in ketorolac tromethamine release system.

    PubMed

    Cussa, Jorgelina; Juárez, Juliana M; Gómez Costa, Marcos B; Anunziata, Oscar A

    2017-08-01

    The ordered mesoporous silica SBA-15 has been applied in studies of ketorolac tromethamine adsorption and release. The SBA-15 materials with hexagonal and regular structure were obtained using a triblock copolymer Pluronic P123 as a template and TEOS as a silica source. Ketorolac tromethamine was adsorbed into SBA-15 silica nanochannels using ethanol as solvent. The physicochemical and textural properties of SBA-15 and ketorolac tromethamine/SBA-15 were characterized by X-ray diffraction, thermogravimetric analysis, transmission electron microscopy, fourier transform infrared spectroscopy and BET surface studies. Drug release was evaluated by soaking the loaded silica mesoporous material into a solution of HCl (0.1 N) at initial time (0-2 h) and buffer pH 7 at high times at 37 °C under continuous stirring. Oral commercial Keto tablets (Dolten(®)) and Keto solution (Keto power) were study for the contrast. Release studies were performed in order to evaluate the required therapeutic efficacy. SBA-15 provides significant improvement in the controlled release of ketorolac tromethamine. Release profile of KETO from SBA-15/KETO and control releases.

  12. "Applying" Conversation Analysis in Applied Linguistics: Evaluating English as a Second Language Textbook Dialogue.

    ERIC Educational Resources Information Center

    Wong, Jean

    This article examines English-as-a-Second-Language (ESL) textbook telephone dialogues against the backdrop about what is reported about real telephone interaction based in research in conversation analysis (CA). An analysis of eight ESL textbooks reveals that the fit between what conversation analysts say about the nature of natural telephone…

  13. To apply or not to apply: a survey analysis of grant writing costs and benefits.

    PubMed

    von Hippel, Ted; von Hippel, Courtney

    2015-01-01

    We surveyed 113 astronomers and 82 psychologists active in applying for federally funded research on their grant-writing history between January, 2009 and November, 2012. We collected demographic data, effort levels, success rates, and perceived non-financial benefits from writing grant proposals. We find that the average proposal takes 116 PI hours and 55 CI hours to write; although time spent writing was not related to whether the grant was funded. Effort did translate into success, however, as academics who wrote more grants received more funding. Participants indicated modest non-monetary benefits from grant writing, with psychologists reporting a somewhat greater benefit overall than astronomers. These perceptions of non-financial benefits were unrelated to how many grants investigators applied for, the number of grants they received, or the amount of time they devoted to writing their proposals. We also explored the number of years an investigator can afford to apply unsuccessfully for research grants and our analyses suggest that funding rates below approximately 20%, commensurate with current NIH and NSF funding, are likely to drive at least half of the active researchers away from federally funded research. We conclude with recommendations and suggestions for individual investigators and for department heads.

  14. To Apply or Not to Apply: A Survey Analysis of Grant Writing Costs and Benefits

    PubMed Central

    von Hippel, Ted; von Hippel, Courtney

    2015-01-01

    We surveyed 113 astronomers and 82 psychologists active in applying for federally funded research on their grant-writing history between January, 2009 and November, 2012. We collected demographic data, effort levels, success rates, and perceived non-financial benefits from writing grant proposals. We find that the average proposal takes 116 PI hours and 55 CI hours to write; although time spent writing was not related to whether the grant was funded. Effort did translate into success, however, as academics who wrote more grants received more funding. Participants indicated modest non-monetary benefits from grant writing, with psychologists reporting a somewhat greater benefit overall than astronomers. These perceptions of non-financial benefits were unrelated to how many grants investigators applied for, the number of grants they received, or the amount of time they devoted to writing their proposals. We also explored the number of years an investigator can afford to apply unsuccessfully for research grants and our analyses suggest that funding rates below approximately 20%, commensurate with current NIH and NSF funding, are likely to drive at least half of the active researchers away from federally funded research. We conclude with recommendations and suggestions for individual investigators and for department heads. PMID:25738742

  15. Toward a Blended Ontology: Applying Knowledge Systems to ...

    EPA Pesticide Factsheets

    Bionanomedicine and environmental research share need common terms and ontologies. This study applied knowledge systems, data mining, and bibliometrics used in nano-scale ADME research from 1991 to 2011. The prominence of nano-ADME in environmental research began to exceed the publication rate in medical research in 2006. That trend appears to continue as a result of the growing products in commerce using nanotechnology, that is, 5-fold growth in number of countries with nanomaterials research centers. Funding for this research virtually did not exist prior to 2002, whereas today both medical and environmental research is funded globally. Key nanoparticle research began with pharmacology and therapeutic drug-delivery and contrasting agents, but the advances have found utility in the environmental research community. As evidence ultrafine aerosols and aquatic colloids research increased 6-fold, indicating a new emphasis on environmental nanotoxicology. User-directed expert elicitation from the engineering and chemical/ADME domains can be combined with appropriate Boolean logic and queries to define the corpus of nanoparticle interest. The study combined pharmacological expertise and informatics to identify the corpus by building logical conclusions and observations. Publication records informatics can lead to an enhanced understanding the connectivity between fields, as well as overcoming the differences in ontology between the fields. The National Exposure Resea

  16. Gastritis staging: interobserver agreement by applying OLGA and OLGIM systems.

    PubMed

    Isajevs, Sergejs; Liepniece-Karele, Inta; Janciauskas, Dainius; Moisejevs, Georgijs; Putnins, Viesturs; Funka, Konrads; Kikuste, Ilze; Vanags, Aigars; Tolmanis, Ivars; Leja, Marcis

    2014-04-01

    Atrophic gastritis remains a difficult histopathological diagnosis with low interobserver agreement. The aim of our study was to compare gastritis staging and interobserver agreement between general and expert gastrointestinal (GI) pathologists using Operative Link for Gastritis Assessment (OLGA) and Operative Link on Gastric Intestinal Metaplasia (OLGIM). We enrolled 835 patients undergoing upper endoscopy in the study. Two general and two expert gastrointestinal pathologists graded biopsy specimens according to the Sydney classification, and the stage of gastritis was assessed by OLGA and OLGIM system. Using OLGA, 280 (33.4 %) patients had gastritis (stage I-IV), whereas with OLGIM this was 167 (19.9 %). OLGA stage III- IV gastritis was observed in 25 patients, whereas by OLGIM stage III-IV was found in 23 patients. Interobserver agreement between expert GI pathologists for atrophy in the antrum, incisura angularis, and corpus was moderate (kappa = 0.53, 0.57 and 0.41, respectively, p < 0.0001), but almost perfect for intestinal metaplasia (kappa = 0.82, 0.80 and 0.81, respectively, p < 0.0001). However, interobserver agreement between general pathologists was poor for atrophy, but moderate for intestinal metaplasia. OLGIM staging provided the highest interobserver agreement, but a substantial proportion of potentially high-risk individuals would be missed if only OLGIM staging is applied. Therefore, we recommend to use a combination of OLGA and OLGIM for staging of chronic gastritis.

  17. Analysis of possibility of applying the PVDF foil in industrial vibration sensors

    NASA Astrophysics Data System (ADS)

    Wróbel, A.

    2015-11-01

    There are many machines using the piezoelectric effects. Systems with smart materials are often used because they have high potential applications for example transducers can be applied to receive required characteristic of projected system. Every engineer and designer know how important it is properly mathematical model and method of the analysis. Also it is important to consider all parameters of analyzed system for example glue layer between elements. Geometrical and material parameters has a significant impact on the characteristics of the all system's components because the omission of the influence of one of them results in inaccuracy in the analysis of the system. In article the modeling and testing of vibrating systems with piezoelectric ceramic materials transducers used as actuators and vibration dampers. The method of analysis of the vibrating sensor systems will be presented, mathematical model, and characteristics, to determine the influence of the system's properties on these characteristics. Main scientific point of the project is to analyze and demonstrate possibility of applying new construction with the PVDF foil or any other belonging to a group of smart materials in industrial sensors. Currently, the vibration level sensors are used by practically all manufacturers of piezoelectric ceramic plates to generate and detect the vibration of the fork.

  18. Applying microscopy to the analysis of nuclear structure and function.

    PubMed

    Iborra, Francisco; Cook, Peter R; Jackson, Dean A

    2003-02-01

    One of the ultimate goals of biological research is to understand mechanisms of cell function within living organisms. With this in mind, many sophisticated technologies that allow us to inspect macromolecular structure in exquisite detail have been developed. Although knowledge of structure derived from techniques such as X-ray crystallography and nuclear magnetic resonance is of vital importance, these approaches cannot reveal the remarkable complexity of molecular interactions that exists in vivo. With this in mind, this review focuses on the use of microscopy techniques to analyze cell structure and function. We describe the different basic microscopic methodologies and how the routine techniques are best applied to particular biological problems. We also emphasize the specific capabilities and uses of light and electron microscopy and highlight their individual advantages and disadvantages. For completion, we also comment on the alternative possibilities provided by a variety of advanced imaging technologies. We hope that this brief analysis of the undoubted power of microscopy techniques will be enough to stimulate a wider participation in this rapidly developing area of biological discovery.

  19. Applying importance-performance analysis to patient safety culture.

    PubMed

    Lee, Yii-Ching; Wu, Hsin-Hung; Hsieh, Wan-Lin; Weng, Shao-Jen; Hsieh, Liang-Po; Huang, Chih-Hsuan

    2015-01-01

    The Sexton et al.'s (2006) safety attitudes questionnaire (SAQ) has been widely used to assess staff's attitudes towards patient safety in healthcare organizations. However, to date there have been few studies that discuss the perceptions of patient safety both from hospital staff and upper management. The purpose of this paper is to improve and to develop better strategies regarding patient safety in healthcare organizations. The Chinese version of SAQ based on the Taiwan Joint Commission on Hospital Accreditation is used to evaluate the perceptions of hospital staff. The current study then lies in applying importance-performance analysis technique to identify the major strengths and weaknesses of the safety culture. The results show that teamwork climate, safety climate, job satisfaction, stress recognition and working conditions are major strengths and should be maintained in order to provide a better patient safety culture. On the contrary, perceptions of management and hospital handoffs and transitions are important weaknesses and should be improved immediately. Research limitations/implications - The research is restricted in generalizability. The assessment of hospital staff in patient safety culture is physicians and registered nurses. It would be interesting to further evaluate other staff's (e.g. technicians, pharmacists and others) opinions regarding patient safety culture in the hospital. Few studies have clearly evaluated the perceptions of healthcare organization management regarding patient safety culture. Healthcare managers enable to take more effective actions to improve the level of patient safety by investigating key characteristics (either strengths or weaknesses) that healthcare organizations should focus on.

  20. Applying DNA computation to intractable problems in social network analysis.

    PubMed

    Chen, Rick C S; Yang, Stephen J H

    2010-09-01

    From ancient times to the present day, social networks have played an important role in the formation of various organizations for a range of social behaviors. As such, social networks inherently describe the complicated relationships between elements around the world. Based on mathematical graph theory, social network analysis (SNA) has been developed in and applied to various fields such as Web 2.0 for Web applications and product developments in industries, etc. However, some definitions of SNA, such as finding a clique, N-clique, N-clan, N-club and K-plex, are NP-complete problems, which are not easily solved via traditional computer architecture. These challenges have restricted the uses of SNA. This paper provides DNA-computing-based approaches with inherently high information density and massive parallelism. Using these approaches, we aim to solve the three primary problems of social networks: N-clique, N-clan, and N-club. Their accuracy and feasible time complexities discussed in the paper will demonstrate that DNA computing can be used to facilitate the development of SNA. Copyright 2010 Elsevier Ireland Ltd. All rights reserved.

  1. Local Bifurcation and Instability Theory Applied to Formability Analysis

    NASA Astrophysics Data System (ADS)

    Buco, B. D.; Oliveira, M. C.; Alves, J. L.; Menezes, L. F.; Ito, K.; Mori, N.

    2010-06-01

    Sheet metal forming components are typically studied with the aid of finite element method based virtual tryout tools, since they allow to save money, time and effort in the design, production and process set-up of deep drawn parts. In all these development phases the analysis of defects is performed with the aid of the material forming limit diagram (FLD), since it allows defining a safe region that reduces: (i) necking; (ii) wrinkling and (iii) large deformation occurrence. It is known that the FLD represented in the strain space presents some disadvantages. The local bifurcation criterion proposed by Ito and Goya defines the critical state for a local bifurcation to set in, as a function of the stress level to work-hardening rate ratio. Thus, the main advantage is that the FLD represented in the stress plane is completely objective [1]. In this work the Ito and Goya model is used to evaluate formability, as well fracture mode and direction along different strain paths: (i) uniaxial tension; (ii) equibiaxial stretch; and (iii) plane-strain. All numerical simulations are performed with the in-house code DD3IMP [2, 3] and NXT [4] in which the Ito and Goya model is implemented, is applied to analyze the results.

  2. The Significant Structure Theory Applied to a Mesophase System

    PubMed Central

    Ma, Shao-mu; Eyring, H.

    1975-01-01

    The significant structure theory of liquids is extended to the mesophase system with p-azoxyanisole as an example. This compound has two different structures, a nematic phase and an isotropic phase, in its liquid state. In this study the nematic phase is treated as subject to a second volume and temperature-dependent degeneracy formally like that due to melting. The isotropic phase is treated as a normal liquid. The specific heat, thermal expansion coefficient, compressibility, volume, entropy of transitions, and heat of transitions are calculated and compared to the observed values. This analysis differs from previous ones in including the volume dependence as well as the temperature dependence in one explicit expression for the Helmholtz free energy. PMID:16592217

  3. BATSE spectroscopy analysis system

    NASA Technical Reports Server (NTRS)

    Schaefer, Bradley E.; Bansal, Sandhia; Basu, Anju; Brisco, Phil; Cline, Thomas L.; Friend, Elliott; Laubenthal, Nancy; Panduranga, E. S.; Parkar, Nuru; Rust, Brad

    1992-01-01

    The Burst and Transient Source Experiment (BATSE) Spectroscopy Analysis System (BSAS) is the software system which is the primary tool for the analysis of spectral data from BATSE. As such, Guest Investigators and the community as a whole need to know its basic properties and characteristics. Described here are the characteristics of the BATSE spectroscopy detectors and the BSAS.

  4. LIBRARY SYSTEMS ANALYSIS.

    ERIC Educational Resources Information Center

    HIEBER, CAROLINE E.; TAYLOR, ROBERT S.

    STEPS THAT THE LIBRARY CAN TAKE IN PREPARING FOR A CHANGE TO AN AUTOMATED SYSTEM WERE INVESTIGATED. THE METHOD DESCRIBED WAS BASED ON AN ANALYSIS OF CLERICAL AND OFFICE WORK CALLED INTEGRATED PROCEDURES OF CONTROL (IPC). THE APPLICATION OF LIBRARY SYSTEMS ANALYSIS (LISA) TO THE LEHIGH UNIVERSITY LIBRARY IS DESCRIBED. THIS PROCESS INVOLVED (1)…

  5. Experiments With Temporal Reasoning Applied To Analysis Of Telemetry Data

    NASA Astrophysics Data System (ADS)

    Perkins, W. A.; Austin, A.

    1987-10-01

    Many applications of expert systems to Space Station Automation, such as monitoring, planning, and scheduling will involve reasoning about attributes of objects at different times. For example, in monitoring, the system must reason about changes in signal parameters over time because causal relationships among events are important. In order to reason efficiently and concurrently about attributes with different values at different times, different time formats, and different time validity conditions requires more complex knowledge representations than are generally available in expert systems. Representation issues dealing with point times, intervals, and relative times must also be resolved. We have implemented a temporal reasoning capability in a generic expert system shell (LES) to address these issues and to increase the flexibility of the knowledge representation for a variety of applications. For its first application, we chose monitoring of telemetry data from a satellite (the Space Telescope). Our work involved just the RCE (Rotor Controlled Electronics) bearing, a component of the reaction-wheels subsystem which has attributes such as ACTUAL-TEMPERATURE of the bearing, WHEEL-SPEED, and MOTOR-CURRENT. This task consists of collecting one attribute value per sensor per cycle, checking each value to see if it is within the acceptable range, and storing the each value with a time tag in the database. Processing becomes more complex when one or more readings are out of their acceptable range. The analysis to discover the cause involves examining several cycles of readings, as well as comparing the readings of different sensors over time. The temporal reasoning capability in LES allowed us to compare the most recent readings of two sensors; or to compare one current reading with a value collected some time earlier; or to collect several consecutive readings which are analyzed for trends. In addition, having time tags associated with attribute values permitted us

  6. Fundamental and applied studies of organic photovoltaic systems

    NASA Astrophysics Data System (ADS)

    Hill, Caleb M.

    Presented here are applied and fundamental studies of model organic photovoltaic (OPV) systems. Graphene oxide (GO) nanosheets were investigated as a potential electron acceptor in bulk heterojunction organic solar cells which employed poly[3-hexylthiophene] (P3HT) as an electron donor. GO nanosheets were transferred into organic solution through a surfactant-assisted phase transfer method. Electron transfer from P3HT to GO in solutions and thin films was established through fluorescence spectroscopy. Bulk heterojunction solar cells containing P3HT, P3HT-GO, and P3HT-phenyl-C61-butyric acid methyl ester (PCBM, a prototypical elector acceptor employed in polymer solar cells) were constructed and evaluated. Single molecule fluorescence spectroscopy was employed to study charge transfer between conjugated polymers and TiO2 at the single molecule level. The fluorescence of individual chains of the conjugated polymer poly[2-methoxy-5-(2-ethylhexyloxy)-1,4-phenylenevinylene] (MEH-PPV) at TiO2 surfaces was shown to exhibit increased intermittent (on/off "blinking") behavior compared to molecules on glass substrates. Single molecule fluorescence excitation anisotropy measurements showed the conformation of the polymer molecules did not differ appreciably between glass and TiO 2 substrates. The similarities in molecular conformation suggest that the observed differences in blinking activity are due to charge transfer between MEH-PPV and TiO2, which provides additional pathways between states of high and low fluorescence quantum efficiency. The electrodeposition of individual Ag nanoparticles (NPs), which can be used to enhance light harvesting in organic photovoltaic systems, was studied in situ via dark field scattering (DFS) microscopy. The scattering at the surface of an indium tin oxide (ITO) working electrode was measured during a potential sweep. Utilizing Mie scattering theory and high resolution scanning electron microscopy (SEM), the scattering data were used to

  7. Principles of micellar electrokinetic capillary chromatography applied in pharmaceutical analysis.

    PubMed

    Hancu, Gabriel; Simon, Brigitta; Rusu, Aura; Mircia, Eleonora; Gyéresi, Arpád

    2013-01-01

    Since its introduction capillary electrophoresis has shown great potential in areas where electrophoretic techniques have rarely been used before, including here the analysis of pharmaceutical substances. The large majority of pharmaceutical substances are neutral from electrophoretic point of view, consequently separations by the classic capillary zone electrophoresis; where separation is based on the differences between the own electrophoretic mobilities of the analytes; are hard to achieve. Micellar electrokinetic capillary chromatography, a hybrid method that combines chromatographic and electrophoretic separation principles, extends the applicability of capillary electrophoretic methods to neutral analytes. In micellar electrokinetic capillary chromatography, surfactants are added to the buffer solution in concentration above their critical micellar concentrations, consequently micelles are formed; micelles that undergo electrophoretic migration like any other charged particle. The separation is based on the differential partitioning of an analyte between the two-phase system: the mobile aqueous phase and micellar pseudostationary phase. The present paper aims to summarize the basic aspects regarding separation principles and practical applications of micellar electrokinetic capillary chromatography, with particular attention to those relevant in pharmaceutical analysis.

  8. Operational modal analysis applied to the concert harp

    NASA Astrophysics Data System (ADS)

    Chomette, B.; Le Carrou, J.-L.

    2015-05-01

    Operational modal analysis (OMA) methods are useful to extract modal parameters of operating systems. These methods seem to be particularly interesting to investigate the modal basis of string instruments during operation to avoid certain disadvantages due to conventional methods. However, the excitation in the case of string instruments is not optimal for OMA due to the presence of damped harmonic components and low noise in the disturbance signal. Therefore, the present study investigates the least-square complex exponential (LSCE) and the modified least-square complex exponential methods in the case of a string instrument to identify modal parameters of the instrument when it is played. The efficiency of the approach is experimentally demonstrated on a concert harp excited by some of its strings and the two methods are compared to a conventional modal analysis. The results show that OMA allows us to identify modes particularly present in the instrument's response with a good estimation especially if they are close to the excitation frequency with the modified LSCE method.

  9. Differential Network Analysis Applied to Preoperative Breast Cancer Chemotherapy Response

    PubMed Central

    Warsow, Gregor; Struckmann, Stephan; Kerkhoff, Claus; Reimer, Toralf; Engel, Nadja; Fuellen, Georg

    2013-01-01

    In silico approaches are increasingly considered to improve breast cancer treatment. One of these treatments, neoadjuvant TFAC chemotherapy, is used in cases where application of preoperative systemic therapy is indicated. Estimating response to treatment allows or improves clinical decision-making and this, in turn, may be based on a good understanding of the underlying molecular mechanisms. Ever increasing amounts of high throughput data become available for integration into functional networks. In this study, we applied our software tool ExprEssence to identify specific mechanisms relevant for TFAC therapy response, from a gene/protein interaction network. We contrasted the resulting active subnetwork to the subnetworks of two other such methods, OptDis and KeyPathwayMiner. We could show that the ExprEssence subnetwork is more related to the mechanistic functional principles of TFAC therapy than the subnetworks of the other two methods despite the simplicity of ExprEssence. We were able to validate our method by recovering known mechanisms and as an application example of our method, we identified a mechanism that may further explain the synergism between paclitaxel and doxorubicin in TFAC treatment: Paclitaxel may attenuate MELK gene expression, resulting in lower levels of its target MYBL2, already associated with doxorubicin synergism in hepatocellular carcinoma cell lines. We tested our hypothesis in three breast cancer cell lines, confirming it in part. In particular, the predicted effect on MYBL2 could be validated, and a synergistic effect of paclitaxel and doxorubicin could be demonstrated in the breast cancer cell lines SKBR3 and MCF-7. PMID:24349128

  10. Information Management Principles Applied to the Ballistic Missile Defense System

    DTIC Science & Technology

    2007-03-01

    if they are being applied. iv Acknowledgments I would like to thank the members who participated in the Delphi Group. Their...52 Initial Delphi Results...55 Second Round Delphi Results .............................................................................57

  11. Evaluation of atrazine degradation applied to different energy systems.

    PubMed

    Moreira, Ailton J; Pinheiro, Bianca S; Araújo, André F; Freschi, Gian P G

    2017-03-01

    Atrazine is an herbicide widely used in crops and has drawn attention due to potential pollution present in soil, sediment, water, and food. Since conventional methods are not potentially efficient to persistent degradation of organic compounds, new technology has been developed to remove them, especially practices utilizing advanced oxidation processes (AOPs). This work aims to evaluate the use of different energies (ultraviolet (UV), microwaves (MW), and radiations (MW-UV)) to the herbicide atrazine through the process of photo-oxidation. These systems found degradation rates of around 12% (UV), 28% (MW), and 83% (MW-UV), respectively, with time intervals of 120 s. After the photolytic processes, the samples were analyzed at a wavelength scanning the range of 190 to 300 nm, where the spectral analysis of the signal was used to evaluate the degradation of atrazine and the appearance of some other peaks (degradation products). The spectrum evaluation resulting from photolytic processes gave rise to a new signal which was confirmed by chromatography. This spectrum indicated the possible pathway of atrazine degradation by the process of photolytic MW-UV, generating atrazine-2-hydroxy, atrazine-desethyl-2-hidroxy, and atrazine-desisopropyl-2-hydroxy. The process indicated that in all situations, chloride was present in the analytic structure and was substituted by a hydroxyl group, which lowered the toxicity of the compound through the photolytic processMW-UV. Chromatographic analysis ascertained these preliminary assessments using spectrophotometry. It was also significantly observed that the process can be optimized by adjusting the pH of the solution, which was evident by an improvement of 10% in the rate of degradation when subjected to a pH solution equal to 8.37.

  12. Evaluation of atrazine degradation applied to different energy systems.

    PubMed

    Moreira, Ailton J; Pinheiro, Bianca S; Araújo, André F; Freschi, Gian P G

    2016-09-01

    Atrazine is an herbicide widely used in crops and has drawn attention due to potential pollution present in soil, sediment, water, and food. Since conventional methods are not potentially efficient to persistent degradation of organic compounds, new technology has been developed to remove them, especially practices utilizing advanced oxidation processes (AOPs). This work aims to evaluate the use of different energies (ultraviolet (UV), microwaves (MW), and radiations (MW-UV)) to the herbicide atrazine through the process of photo-oxidation. These systems found degradation rates of around 12 % (UV), 28 % (MW), and 83 % (MW-UV), respectively, with time intervals of 120 s. After the photolytic processes, the samples were analyzed at a wavelength scanning the range of 190 to 300 nm, where the spectral analysis of the signal was used to evaluate the degradation of atrazine and the appearance of some other peaks (degradation products). The spectrum evaluation resulting from photolytic processes gave rise to a new signal which was confirmed by chromatography. This spectrum indicated the possible pathway of atrazine degradation by the process of photolytic MW-UV, generating atrazine-2-hydroxy, atrazine-desethyl-2-hidroxy, and atrazine-desisopropyl-2-hydroxy. The process indicated that in all situations, chloride was present in the analytic structure and was substituted by a hydroxyl group, which lowered the toxicity of the compound through the photolytic process MW-UV. Chromatographic analysis ascertained these preliminary assessments using spectrophotometry. It was also significantly observed that the process can be optimized by adjusting the pH of the solution, which was evident by an improvement of 10 % in the rate of degradation when subjected to a pH solution equal to 8.37.

  13. The Recursion Method Applied to One-Dimensional Spin Systems

    NASA Astrophysics Data System (ADS)

    Viswanath, V. S.

    The recursion method is used for the study of the dynamics of quantum spin models at zero and infinite temperatures. Two alternative formulations of the recursion method are described in detail. Application of either formulation to quantum many-body systems yields a set of continued-fraction coefficients. Several new calculational techniques for the analysis of these continued-fraction coefficients developed during the course of my research are presented. The efficacy and accuracy of these techniques are demonstrated by applications to the few situations were exact nontrivial results are available. For the s = 1/2 XXZ model on a linear chain, new and reliable quantitative information has been obtained on the type of ordering in the ground-state, on the size of gaps in the dynamically relevant excitation spectrum, on the bandwidths of dominant structures in spectral densities, on the exponents of infrared singularities in the same functions, and on the detailed shape of spectral-weight distributions. Zero temperature dynamic structure factors for the one-dimensional spin-s XYZ model in a magnetic field have been calculated for systems with s = 1/2, 1, 3/2. The line shapes and peak positions have been shown to differ considerably from the corresponding spin-wave results. Time-dependent spin autocorrelation functions and their spectral densities for the semi-infinite one -dimensional s = 1/2 XY model at infinite temperature have been determined in part by rigorous calculations in the fermion representation and in part by the recursion method in the spin representation. The study of boundary effects yields valuable new insight into the dynamical processes which govern the transport of spin fluctuations in that model. The exact results also provide a benchmark against which the results of the recursion method have been compared and calibrated.

  14. Security System Technologies Applied to Ambient Assisted Living

    NASA Astrophysics Data System (ADS)

    Villacorta, Juan J.; Del Val, Lara; Jimenez, Ma Isabel; Izquierdo, Alberto

    The increasing elderly population has increased interest in the Ambient Assisted Living systems. This article presents a system for monitoring the disabled or elderly developed from an existing surveillance system. The modularity and adaptability characteristics of the system allow an easy adaptation for a different purpose. The proposed system uses a network of sensors capable of motion detection that includes fall warning, identification of persons and a configurable control system which allows its use in different scenarios.

  15. Phase plane analysis: applying chaos theory in health care.

    PubMed

    Priesmeyer, H R; Sharp, L F

    1995-01-01

    This article applies the new science of nonlinearity to administrative issues and accounts receivable management in health care, and it provides a new perspective on common operating and quality control measures.

  16. Space elevator systems level analysis

    SciTech Connect

    Laubscher, B. E.

    2004-01-01

    The Space Elevator (SE) represents a major paradigm shift in space access. It involves new, untried technologies in most of its subsystems. Thus the successful construction of the SE requires a significant amount of development, This in turn implies a high level of risk for the SE. This paper will present a systems level analysis of the SE by subdividing its components into their subsystems to determine their level of technological maturity. such a high-risk endeavor is to follow a disciplined approach to the challenges. A systems level analysis informs this process and is the guide to where resources should be applied in the development processes. It is an efficient path that, if followed, minimizes the overall risk of the system's development. systems level analysis is that the overall system is divided naturally into its subsystems, and those subsystems are further subdivided as appropriate for the analysis. By dealing with the complex system in layers, the parameter space of decisions is kept manageable. Moreover, A rational way to manage One key aspect of a resources are not expended capriciously; rather, resources are put toward the biggest challenges and most promising solutions. This overall graded approach is a proven road to success. The analysis includes topics such as nanotube technology, deployment scenario, power beaming technology, ground-based hardware and operations, ribbon maintenance and repair and climber technology.

  17. Multisensor Image Analysis System

    DTIC Science & Technology

    1993-04-15

    AD-A263 679 II Uli! 91 Multisensor Image Analysis System Final Report Authors. Dr. G. M. Flachs Dr. Michael Giles Dr. Jay Jordan Dr. Eric...or decision, unless so designated by other documentation. 93-09739 *>ft s n~. now illlllM3lMVf Multisensor Image Analysis System Final...Multisensor Image Analysis System 3. REPORT TYPE AND DATES COVERED FINAL: LQj&tt-Z JZOfVL 5. FUNDING NUMBERS 93 > 6. AUTHOR(S) Drs. Gerald

  18. SUBSURFACE VISUAL ALARM SYSTEM ANALYSIS

    SciTech Connect

    D.W. Markman

    2001-08-06

    The ''Subsurface Fire Hazard Analysis'' (CRWMS M&O 1998, page 61), and the document, ''Title III Evaluation Report for the Surface and Subsurface Communication System'', (CRWMS M&O 1999a, pages 21 and 23), both indicate the installed communication system is adequate to support Exploratory Studies Facility (ESF) activities with the exception of the mine phone system for emergency notification purposes. They recommend the installation of a visual alarm system to supplement the page/party phone system The purpose of this analysis is to identify data communication highway design approaches, and provide justification for the selected or recommended alternatives for the data communication of the subsurface visual alarm system. This analysis is being prepared to document a basis for the design selection of the data communication method. This analysis will briefly describe existing data or voice communication or monitoring systems within the ESF, and look at how these may be revised or adapted to support the needed data highway of the subsurface visual alarm. system. The existing PLC communication system installed in subsurface is providing data communication for alcove No.5 ventilation fans, south portal ventilation fans, bulkhead doors and generator monitoring system. It is given that the data communication of the subsurface visual alarm system will be a digital based system. It is also given that it is most feasible to take advantage of existing systems and equipment and not consider an entirely new data communication system design and installation. The scope and primary objectives of this analysis are to: (1) Briefly review and describe existing available data communication highways or systems within the ESF. (2) Examine technical characteristics of an existing system to disqualify a design alternative is paramount in minimizing the number of and depth of a system review. (3) Apply general engineering design practices or criteria such as relative cost, and degree of

  19. Multi-Criteria Analysis for Biomass Utilization Applying Data Envelopment Analysis

    NASA Astrophysics Data System (ADS)

    Morimoto, Hidetsugu; Hoshino, Satoshi; Kuki, Yasuaki

    This paper aimed to consider about material-recycling, preventing global warming, and economic efficiency on preset and planed 195 Biomass Towns applying DEA (Data Envelopment Analysis), which can evaluate operational efficiency entities such as private companies or projects. In the results, although the Biomass Town can recycle material efficiently, it was clarified that preventing global warming and business profitability was brushed off like it in Biomass Town Design. Moreover, from the point of view of operational efficiency, we suggested an improvement of the Biomass Town scale for more efficiency-enhancing applying DEA. We found that applying DEA was able to catch more improvements or indicator as compared with cost-benefit analysis and cost-effectiveness analysis.

  20. Systems Analysis Sub site

    SciTech Connect

    EERE

    2012-03-16

    Systems analysis provides direction, focus, and support for the development and introduction of hydrogen production, storage, and end-use technologies, and provides a basis for recommendations on a balanced portfolio of activities.

  1. Primary expert system applied in design of electron-optical system

    NASA Astrophysics Data System (ADS)

    Lei, Wei; Tong, Linsu

    1995-09-01

    It is well known that the structure of electron optical system is complicated, so many factors, such as magnification, geometrical aberrations, and space charge effect must be taken into account in the design. At present, the main work of electron optical system CAD is solving equations and calculating numerical values. However, the designs perhaps need more inference and expertise than numerical calculations. In this paper, a primary expert system which is applied in design of electron optical system is established. This expert system is combined with the simulation software SEU-3D program to design some practical electron optical systems. Although the knowledge base is small and rules are not abundant, this paper has used this system to obtain some very useful results. The initial success with this system suggests that further work need to be done whether more rules and knowledge will be added to extend the ability of expert system.

  2. [Digital learning object for diagnostic reasoning in nursing applied to the integumentary system].

    PubMed

    da Costa, Cecília Passos Vaz; Luz, Maria Helena Barros Araújo

    2015-12-01

    To describe the creation of a digital learning object for diagnostic reasoning in nursing applied to the integumentary system at a public university of Piaui. A methodological study applied to technological production based on the pedagogical framework of problem-based learning. The methodology for creating the learning object observed the stages of analysis, design, development, implementation and evaluation recommended for contextualized instructional design. The revised taxonomy of Bloom was used to list the educational goals. The four modules of the developed learning object were inserted into the educational platform Moodle. The theoretical assumptions allowed the design of an important online resource that promotes effective learning in the scope of nursing education. This study should add value to nursing teaching practices through the use of digital learning objects for teaching diagnostic reasoning applied to skin and skin appendages.

  3. Applying thematic analysis theory to practice: a researcher's experience.

    PubMed

    Tuckett, Anthony G

    2005-01-01

    This article describes an experience of thematic analysis. In order to answer the question 'What does analysis look like in practice?' it describes in brief how the methodology of grounded theory, the epistemology of social constructionism, and the theoretical stance of symbolic interactionism inform analysis. Additionally, analysis is examined by evidencing the systematic processes--here termed organising, coding, writing, theorising, and reading--that led the researcher to develop a final thematic schema.

  4. Launch vehicle systems design analysis

    NASA Technical Reports Server (NTRS)

    Ryan, Robert; Verderaime, V.

    1993-01-01

    Current launch vehicle design emphasis is on low life-cycle cost. This paper applies total quality management (TQM) principles to a conventional systems design analysis process to provide low-cost, high-reliability designs. Suggested TQM techniques include Steward's systems information flow matrix method, quality leverage principle, quality through robustness and function deployment, Pareto's principle, Pugh's selection and enhancement criteria, and other design process procedures. TQM quality performance at least-cost can be realized through competent concurrent engineering teams and brilliance of their technical leadership.

  5. Launch vehicle systems design analysis

    NASA Technical Reports Server (NTRS)

    Ryan, Robert; Verderaime, V.

    1993-01-01

    Current launch vehicle design emphasis is on low life-cycle cost. This paper applies total quality management (TQM) principles to a conventional systems design analysis process to provide low-cost, high-reliability designs. Suggested TQM techniques include Steward's systems information flow matrix method, quality leverage principle, quality through robustness and function deployment, Pareto's principle, Pugh's selection and enhancement criteria, and other design process procedures. TQM quality performance at least-cost can be realized through competent concurrent engineering teams and brilliance of their technical leadership.

  6. Machine Learning Method Applied in Readout System of Superheated Droplet Detector

    NASA Astrophysics Data System (ADS)

    Liu, Yi; Sullivan, Clair Julia; d'Errico, Francesco

    2017-07-01

    Direct readability is one advantage of superheated droplet detectors in neutron dosimetry. Utilizing such a distinct characteristic, an imaging readout system analyzes image of the detector for neutron dose readout. To improve the accuracy and precision of algorithms in the imaging readout system, machine learning algorithms were developed. Deep learning neural network and support vector machine algorithms are applied and compared with generally used Hough transform and curvature analysis methods. The machine learning methods showed a much higher accuracy and better precision in recognizing circular gas bubbles.

  7. Measurement uncertainty analysis techniques applied to PV performance measurements

    SciTech Connect

    Wells, C.

    1992-10-01

    The purpose of this presentation is to provide a brief introduction to measurement uncertainty analysis, outline how it is done, and illustrate uncertainty analysis with examples drawn from the PV field, with particular emphasis toward its use in PV performance measurements. The uncertainty information we know and state concerning a PV performance measurement or a module test result determines, to a significant extent, the value and quality of that result. What is measurement uncertainty analysis It is an outgrowth of what has commonly been called error analysis. But uncertainty analysis, a more recent development, gives greater insight into measurement processes and tests, experiments, or calibration results. Uncertainty analysis gives us an estimate of the I interval about a measured value or an experiment's final result within which we believe the true value of that quantity will lie. Why should we take the time to perform an uncertainty analysis A rigorous measurement uncertainty analysis: Increases the credibility and value of research results; allows comparisons of results from different labs; helps improve experiment design and identifies where changes are needed to achieve stated objectives (through use of the pre-test analysis); plays a significant role in validating measurements and experimental results, and in demonstrating (through the post-test analysis) that valid data have been acquired; reduces the risk of making erroneous decisions; demonstrates quality assurance and quality control measures have been accomplished; define Valid Data as data having known and documented paths of: Origin, including theory; measurements; traceability to measurement standards; computations; uncertainty analysis of results.

  8. Classical linear-control analysis applied to business-cycle dynamics and stability

    NASA Technical Reports Server (NTRS)

    Wingrove, R. C.

    1983-01-01

    Linear control analysis is applied as an aid in understanding the fluctuations of business cycles in the past, and to examine monetary policies that might improve stabilization. The analysis shows how different policies change the frequency and damping of the economic system dynamics, and how they modify the amplitude of the fluctuations that are caused by random disturbances. Examples are used to show how policy feedbacks and policy lags can be incorporated, and how different monetary strategies for stabilization can be analytically compared. Representative numerical results are used to illustrate the main points.

  9. Applying TQM/CQI principles to information system selection.

    PubMed

    McConnell, P; Ciotti, V G

    1995-06-01

    Faced with the challenge of selecting a new information system, senior management at North Valley Hospital in Whitefish, Montana, decided to employ principles of total quality management and continuous quality improvement to help ensure an efficient process that involved and empowered those who would use the new system. Senior managers set a vision for the process, then assigned responsibility for system selection to a task force representing departments and disciplines that would be most affected by the system. The task force established a systematic process for identifying information system needs, evaluating existing products, and selecting among those products.

  10. Analysis of Waterway Systems.

    DTIC Science & Technology

    1972-09-01

    9RANDON ROAD 5. Ottawa LOD 6. Peru 3. 7 Henne pin S 8. Peoria DRESDEN ISLAND LSD 9. Pekin /0. Havana 4 I I. Grafton- MARSEILLES ŜD Florence 12. Wood...N. R. and Smith , H. Applied Regression Analysis. Wiley, New York, N. Y., 1966. 29. Verity, W. and Kautz, B. Pearson Product Moment Correlation...Ox- ford Univ. Press, New York, N. Y., 1951. 44. Isard, W. Methods of Regional Analysis. The M.I.T. Press, Cambridge, Mass., 1960. 45. Adams , R. B. An

  11. Applied Missing Data Analysis. Methodology in the Social Sciences Series

    ERIC Educational Resources Information Center

    Enders, Craig K.

    2010-01-01

    Walking readers step by step through complex concepts, this book translates missing data techniques into something that applied researchers and graduate students can understand and utilize in their own research. Enders explains the rationale and procedural details for maximum likelihood estimation, Bayesian estimation, multiple imputation, and…

  12. Geometric Error Analysis in Applied Calculus Problem Solving

    ERIC Educational Resources Information Center

    Usman, Ahmed Ibrahim

    2017-01-01

    The paper investigates geometric errors students made as they tried to use their basic geometric knowledge in the solution of the Applied Calculus Optimization Problem (ACOP). Inaccuracies related to the drawing of geometric diagrams (visualization skills) and those associated with the application of basic differentiation concepts into ACOP…

  13. Applied Missing Data Analysis. Methodology in the Social Sciences Series

    ERIC Educational Resources Information Center

    Enders, Craig K.

    2010-01-01

    Walking readers step by step through complex concepts, this book translates missing data techniques into something that applied researchers and graduate students can understand and utilize in their own research. Enders explains the rationale and procedural details for maximum likelihood estimation, Bayesian estimation, multiple imputation, and…

  14. Duration Analysis Applied to the Adoption of Knowledge.

    ERIC Educational Resources Information Center

    Vega-Cervera, Juan A.; Gordillo, Isabel Cuadrado

    2001-01-01

    Analyzes knowledge acquisition in a sample of 264 pupils in 9 Spanish elementary schools, using time as a dependent variable. Introduces psycho-pedagogical, pedagogical, and social variables into a hazard model applied to the reading process. Auditory discrimination (not intelligence or visual perception) most significantly influences learning to…

  15. Space lab system analysis

    NASA Technical Reports Server (NTRS)

    Rives, T. B.; Ingels, F. M.

    1988-01-01

    An analysis of the Automated Booster Assembly Checkout System (ABACS) has been conducted. A computer simulation of the ETHERNET LAN has been written. The simulation allows one to investigate different structures of the ABACS system. The simulation code is in PASCAL and is VAX compatible.

  16. An Incentivized Capstone Design Team Applying the System Engineering Process

    DTIC Science & Technology

    2015-01-02

    Original Fast Ascender Design The next solution was to simulate the Atlas APA – 5 Lightweight Powered Ascender (Figure 28) using the SP-CW Winch...Figure 28 - Atlas APA - 5   Figure 29 - SP - CW Winch 3.6 Carrying and Organizing Gear Subassembly Purpose: The gear subassembly is...operates the machine using the jog button while the other manually applies pressure on the axial yarns and checks to make sure the braid is correctly

  17. Object-oriented fault tree models applied to system diagnosis

    NASA Technical Reports Server (NTRS)

    Iverson, David L.; Patterson-Hine, F. A.

    1990-01-01

    When a diagnosis system is used in a dynamic environment, such as the distributed computer system planned for use on Space Station Freedom, it must execute quickly and its knowledge base must be easily updated. Representing system knowledge as object-oriented augmented fault trees provides both features. The diagnosis system described here is based on the failure cause identification process of the diagnostic system described by Narayanan and Viswanadham. Their system has been enhanced in this implementation by replacing the knowledge base of if-then rules with an object-oriented fault tree representation. This allows the system to perform its task much faster and facilitates dynamic updating of the knowledge base in a changing diagnosis environment. Accessing the information contained in the objects is more efficient than performing a lookup operation on an indexed rule base. Additionally, the object-oriented fault trees can be easily updated to represent current system status. This paper describes the fault tree representation, the diagnosis algorithm extensions, and an example application of this system. Comparisons are made between the object-oriented fault tree knowledge structure solution and one implementation of a rule-based solution. Plans for future work on this system are also discussed.

  18. CONVEYOR SYSTEM SAFETY ANALYSIS

    SciTech Connect

    M. Salem

    1995-06-23

    The purpose and objective of this analysis is to systematically identify and evaluate hazards related to the Yucca Mountain Project Exploratory Studies Facility (ESF) surface and subsurface conveyor system (for a list of conveyor subsystems see section 3). This process is an integral part of the systems engineering process; whereby safety is considered during planning, design, testing, and construction. A largely qualitative approach was used since a radiological System Safety Analysis is not required. The risk assessment in this analysis characterizes the accident scenarios associated with the conveyor structures/systems/components in terms of relative risk and includes recommendations for mitigating all identified risks. The priority for recommending and implementing mitigation control features is: (1) Incorporate measures to reduce risks and hazards into the structure/system/component (S/S/C) design, (2) add safety devices and capabilities to the designs that reduce risk, (3) provide devices that detect and warn personnel of hazardous conditions, and (4) develop procedures and conduct training to increase worker awareness of potential hazards, on methods to reduce exposure to hazards, and on the actions required to avoid accidents or correct hazardous conditions. The scope of this analysis is limited to the hazards related to the design of conveyor structures/systems/components (S/S/Cs) that occur during normal operation. Hazards occurring during assembly, test and maintenance or ''off normal'' operations have not been included in this analysis. Construction related work activities are specifically excluded per DOE Order 5481.1B section 4. c.

  19. Applying Systems Engineering Principles in Improving Health Care Delivery

    PubMed Central

    Kopach-Konrad, Renata; Criswell, Mike; Hasan, Imran; Chakraborty, Santanu; Pekny, Joseph; Doebbeling, Bradley N.

    2007-01-01

    Background In a highly publicized joint report, the National Academy of Engineering and the Institute of Medicine recently recommended the systematic application of systems engineering approaches for reforming our health care delivery system. For this to happen, medical professionals and managers need to understand and appreciate the power that systems engineering concepts and tools can bring to redesigning and improving health care environments and practices. Objective To present and discuss fundamental concepts and tools of systems engineering and important parallels between systems engineering, health services, and implementation research as it pertains to the care of complex patients. Design An exploratory, qualitative review of systems engineering concepts and overview of ongoing applications of these concepts in the areas of hemodialysis, radiation therapy, and patient flow modeling. Results In this paper, we describe systems engineering as the process of identifying the system of interest, choosing appropriate performance measures, selecting the best modeling tool, studying model properties and behavior under a variety of scenarios, and making design and operational decisions for implementation. Conclusions We discuss challenges and opportunities for bringing people with systems engineering skills into health care. PMID:18026813

  20. How Has Applied Behavior Analysis and Behavior Therapy Changed?: An Historical Analysis of Journals

    ERIC Educational Resources Information Center

    O'Donohue, William; Fryling, Mitch

    2007-01-01

    Applied behavior analysis and behavior therapy are now nearly a half century old. It is interesting to ask if and how these disciplines have changed over time, particularly regarding some of their key internal controversies (e.g., role of cognitions). We examined the first five years and the 2000-2004 five year period of the "Journal of Applied…

  1. Applying Model Analysis to a Resource-Based Analysis of the Force and Motion Conceptual Evaluation

    ERIC Educational Resources Information Center

    Smith, Trevor I.; Wittmann, Michael C.; Carter, Tom

    2014-01-01

    Previously, we analyzed the Force and Motion Conceptual Evaluation in terms of a resources-based model that allows for clustering of questions so as to provide useful information on how students correctly or incorrectly reason about physics. In this paper, we apply model analysis to show that the associated model plots provide more information…

  2. Applying Model Analysis to a Resource-Based Analysis of the Force and Motion Conceptual Evaluation

    ERIC Educational Resources Information Center

    Smith, Trevor I.; Wittmann, Michael C.; Carter, Tom

    2014-01-01

    Previously, we analyzed the Force and Motion Conceptual Evaluation in terms of a resources-based model that allows for clustering of questions so as to provide useful information on how students correctly or incorrectly reason about physics. In this paper, we apply model analysis to show that the associated model plots provide more information…

  3. Coal systems analysis

    SciTech Connect

    Warwick, P.D.

    2005-07-01

    This collection of papers provides an introduction to the concept of coal systems analysis and contains examples of how coal systems analysis can be used to understand, characterize, and evaluate coal and coal gas resources. Chapter are: Coal systems analysis: A new approach to the understanding of coal formation, coal quality and environmental considerations, and coal as a source rock for hydrocarbons by Peter D. Warwick. Appalachian coal assessment: Defining the coal systems of the Appalachian Basin by Robert C. Milici. Subtle structural influences on coal thickness and distribution: Examples from the Lower Broas-Stockton coal (Middle Pennsylvanian), Eastern Kentucky Coal Field, USA by Stephen F. Greb, Cortland F. Eble, and J.C. Hower. Palynology in coal systems analysis The key to floras, climate, and stratigraphy of coal-forming environments by Douglas J. Nichols. A comparison of late Paleocene and late Eocene lignite depositional systems using palynology, upper Wilcox and upper Jackson Groups, east-central Texas by Jennifer M.K. O'Keefe, Recep H. Sancay, Anne L. Raymond, and Thomas E. Yancey. New insights on the hydrocarbon system of the Fruitland Formation coal beds, northern San Juan Basin, Colorado and New Mexico, USA by W.C. Riese, William L. Pelzmann, and Glen T. Snyder.

  4. Measurement uncertainty analysis techniques applied to PV performance measurements

    SciTech Connect

    Wells, C

    1992-10-01

    The purpose of this presentation is to provide a brief introduction to measurement uncertainty analysis, outline how it is done, and illustrate uncertainty analysis with examples drawn from the PV field, with particular emphasis toward its use in PV performance measurements. The uncertainty information we know and state concerning a PV performance measurement or a module test result determines, to a significant extent, the value and quality of that result. What is measurement uncertainty analysis? It is an outgrowth of what has commonly been called error analysis. But uncertainty analysis, a more recent development, gives greater insight into measurement processes and tests, experiments, or calibration results. Uncertainty analysis gives us an estimate of the I interval about a measured value or an experiment`s final result within which we believe the true value of that quantity will lie. Why should we take the time to perform an uncertainty analysis? A rigorous measurement uncertainty analysis: Increases the credibility and value of research results; allows comparisons of results from different labs; helps improve experiment design and identifies where changes are needed to achieve stated objectives (through use of the pre-test analysis); plays a significant role in validating measurements and experimental results, and in demonstrating (through the post-test analysis) that valid data have been acquired; reduces the risk of making erroneous decisions; demonstrates quality assurance and quality control measures have been accomplished; define Valid Data as data having known and documented paths of: Origin, including theory; measurements; traceability to measurement standards; computations; uncertainty analysis of results.

  5. Improving the flash flood frequency analysis applying dendrogeomorphological evidences

    NASA Astrophysics Data System (ADS)

    Ruiz-Villanueva, V.; Ballesteros, J. A.; Bodoque, J. M.; Stoffel, M.; Bollschweiler, M.; Díez-Herrero, A.

    2009-09-01

    Flash floods are one of the natural hazards that cause major damages worldwide. Especially in Mediterranean areas they provoke high economic losses every year. In mountain areas with high stream gradients, floods events are characterized by extremely high flow and debris transport rates. Flash flood analysis in mountain areas presents specific scientific challenges. On one hand, there is a lack of information on precipitation and discharge due to a lack of spatially well distributed gauge stations with long records. On the other hand, gauge stations may not record correctly during extreme events when they are damaged or the discharge exceeds the recordable level. In this case, no systematic data allows improvement of the understanding of the spatial and temporal occurrence of the process. Since historic documentation is normally scarce or even completely missing in mountain areas, tree-ring analysis can provide an alternative approach. Flash floods may influence trees in different ways: (1) tilting of the stem through the unilateral pressure of the flowing mass or individual boulders; (2) root exposure through erosion of the banks; (3) injuries and scars caused by boulders and wood transported in the flow; (4) decapitation of the stem and resulting candelabra growth through the severe impact of boulders; (5) stem burial through deposition of material. The trees react to these disturbances with specific growth changes such as abrupt change of the yearly increment and anatomical changes like reaction wood or callus tissue. In this study, we sampled 90 cross sections and 265 increment cores of trees heavily affected by past flash floods in order to date past events and to reconstruct recurrence intervals in two torrent channels located in the Spanish Central System. The first study site is located along the Pelayo River, a torrent in natural conditions. Based on the external disturbances of trees and their geomorphological position, 114 Pinus pinaster (Ait

  6. Matrix Factorization Methods Applied in Microarray Data Analysis

    PubMed Central

    Kossenkov, Andrew V.

    2010-01-01

    Numerous methods have been applied to microarray data to group genes into clusters that show similar expression patterns. These methods assign each gene to a single group, which does not reflect the widely held view among biologists that most, if not all, genes in eukaryotes are involved in multiple biological processes and therefore will be multiply regulated. Here, we review several methods that have been developed that are capable of identifying patterns of behavior in transcriptional response and assigning genes to multiple patterns. Broadly speaking, these methods define a series of mathematical approaches to matrix factorization with different approaches to the fitting of the model to the data. We focus on these methods in contrast to traditional clustering methods applied to microarray data, which assign one gene to one cluster. PMID:20376923

  7. Natural language interface and database issues in applying expert systems to power systems

    SciTech Connect

    Rumpel, D.; Krost, G. )

    1992-05-01

    Applying expert systems (ES's) to power system operation means imbedding them in an existing environment which consists of computational facilities (SCADA, EMS) on one side, and the operating staff on the other side. To that aim, an interface between process data and ES data is required. Further, for an unambiguous and clear information exchange with, and good acceptance by , the operators and approximate natural language dialog of the ES is desirable. In this paper, several approaches to these problems are described.

  8. Intelligent monitoring system applied to super long distance telerobotic tasks

    NASA Technical Reports Server (NTRS)

    Wakita, Yujin; Hirai, Shigeoki; Machida, Kazuo

    1994-01-01

    Time delay and small capacity of communication are the primary constraint in super long distance telerobotic systems such as astronautical robotic tasks. Intelligent telerobotics is thought to break this constraint. We aim to realize this super long distance telerobotic system with object handling knowledge base and intelligent monitoring. We will discuss physical and technical factors for this purpose.

  9. Applying Language Systems to the Teaching of Writing.

    ERIC Educational Resources Information Center

    Jordan, Michael P.

    1986-01-01

    Uses a single language system--technical description--to illustrate how language systems can now be used as the basis for instruction in technical writing. Provides sample exercises in progressive teaching from simple description to complex continuity devices, showing how they can be used at all stages of the writing process. (HTH)

  10. The Preschool: Setting for Applied Behavior Analysis Research.

    ERIC Educational Resources Information Center

    Essa, Eva L.

    1978-01-01

    This review covers four general areas: modification of undesirable behavior; increasing competence; generalization and maintenance; and behavior analysis methods in assessing preschool procedures. These studies support the usefulness of behavioral analysis; however, a major problem is the lack of systematic follow-up or procedures encouraging…

  11. Applying Financial Portfolio Analysis to Government Program Portfolios

    DTIC Science & Technology

    2007-06-01

    instance, in stock valuation, the Capital Asset Pricing Model ( CAPM ) often provides the returns used in portfolio theory analysis. Further...PORTFOLIO OPTIMIZATION MODELS ................................... 63 1. EVM Optimization Model ...63 2. Constrained EVM Optimization Model ................................. 64 IV. DATA ANALYSIS AND RESULTS

  12. Modal analysis applied to circular, rectangular, and coaxial waveguides

    NASA Technical Reports Server (NTRS)

    Hoppe, D. J.

    1988-01-01

    Recent developments in the analysis of various waveguide components and feedhorns using Modal Analysis (Mode Matching Method) are summarized. A brief description of the theory is presented, and the important features of the method are pointed out. Specific examples in circular, rectangular, and coaxial waveguides are included, with comparisons between the theory and experimental measurements. Extensions to the methods are described.

  13. Quantitative Schlieren analysis applied to holograms of crystals grown on Spacelab 3

    NASA Technical Reports Server (NTRS)

    Brooks, Howard L.

    1986-01-01

    In order to extract additional information about crystals grown in the microgravity environment of Spacelab, a quantitative schlieren analysis technique was developed for use in a Holography Ground System of the Fluid Experiment System. Utilizing the Unidex position controller, it was possible to measure deviation angles produced by refractive index gradients of 0.5 milliradians. Additionally, refractive index gradient maps for any recorded time during the crystal growth were drawn and used to create solute concentration maps for the environment around the crystal. The technique was applied to flight holograms of Cell 204 of the Fluid Experiment System that were recorded during the Spacelab 3 mission on STS 51B. A triglycine sulfate crystal was grown under isothermal conditions in the cell and the data gathered with the quantitative schlieren analysis technique is consistent with a diffusion limited growth process.

  14. Systems Biology Applied to Heart Failure With Normal Ejection Fraction

    PubMed Central

    Mesquita, Evandro Tinoco; Jorge, Antonio Jose Lagoeiro; de Souza, Celso Vale; Cassino, João Paulo Pedroza

    2014-01-01

    Heart failure with normal ejection fraction (HFNEF) is currently the most prevalent clinical phenotype of heart failure. However, the treatments available have shown no reduction in mortality so far. Advances in the omics sciences and techniques of high data processing used in molecular biology have enabled the development of an integrating approach to HFNEF based on systems biology. This study aimed at presenting a systems-biology-based HFNEF model using the bottom-up and top-down approaches. A literature search was conducted for studies published between 1991 and 2013 regarding HFNEF pathophysiology, its biomarkers and systems biology. A conceptual model was developed using bottom-up and top-down approaches of systems biology. The use of systems-biology approaches for HFNEF, a complex clinical syndrome, can be useful to better understand its pathophysiology and to discover new therapeutic targets. PMID:24918915

  15. The Wavefront Clock Technique Applied to Current VLBI Systems

    NASA Astrophysics Data System (ADS)

    Kiuchi, Hitoshi; Kondo, Tetsuro

    1996-02-01

    A prototype wavefront clock system has been developed for application to current VLBI systems. The reference clock for both the front-end and back-end of a VLBI system is controlled at the observing site according to a calculated a priori delay rate. By using this system, fringe stopping is simultaneously performed on all received frequencies and both the upper side band (USB) and the lower side band (LSB) of a frequency-converter signal. Also, it can be used to reduce the large fringe rate (Doppler frequency) inherent in space VLBI, especially when used with an FX-type correlation proces sor. This wavefront clock method will be beneficial to users of conventional VLBI systems and will not require any modification. The reference point for the Doppler compensation can be taken at an ideal point , like the center of the Earth. Moreover, it is possible to use it as a pseudo-fringe simulator for checking the correlation processor.

  16. Distribution patterns of entomopathogenic nematodes applied through drip irrigation systems.

    PubMed

    Wennemann, L; Cone, W W; Wright, L C; Perez, J; Conant, M M

    2003-04-01

    The distribution of entomopathogenic nematodes applied by drip irrigation was evaluated by injecting small volumes of Steinernema carpocapsae (Weiser) All strain, Steinernema feltiae (Filipjev) SN strain, Steinernema glaseri Steiner, and Heterorhabditis bacteriophora HP 88 strain Poinar suspensions into drip irrigation lines. Additionally, Steinernema riobrave Cabanillas, Poinar, & Raulston, and S. carpocapsae were injected in a 10-liter volume of water with an injection pump. Overall, the nematodes were evenly distributed along the drip lines. The total number of nematodes recovered from drip emitters was variable ranging from 42 to 92%. However, drip irrigation lines have potential to deliver entomopathogenic nematodes efficiently into pest habitats.

  17. Aircraft Electric Propulsion Systems Applied Research at NASA

    NASA Technical Reports Server (NTRS)

    Clarke, Sean

    2015-01-01

    Researchers at NASA are investigating the potential for electric propulsion systems to revolutionize the design of aircraft from the small-scale general aviation sector to commuter and transport-class vehicles. Electric propulsion provides new degrees of design freedom that may enable opportunities for tightly coupled design and optimization of the propulsion system with the aircraft structure and control systems. This could lead to extraordinary reductions in ownership and operating costs, greenhouse gas emissions, and noise annoyance levels. We are building testbeds, high-fidelity aircraft simulations, and the first highly distributed electric inhabited flight test vehicle to begin to explore these opportunities.

  18. Applying New Network Security Technologies to SCADA Systems.

    SciTech Connect

    Hurd, Steven A; Stamp, Jason Edwin; Duggan, David P; Chavez, Adrian R.

    2006-11-01

    Supervisory Control and Data Acquisition (SCADA) systems for automation are very important for critical infrastructure and manufacturing operations. They have been implemented to work in a number of physical environments using a variety of hardware, software, networking protocols, and communications technologies, often before security issues became of paramount concern. To offer solutions to security shortcomings in the short/medium term, this project was to identify technologies used to secure "traditional" IT networks and systems, and then assess their efficacy with respect to SCADA systems. These proposed solutions must be relatively simple to implement, reliable, and acceptable to SCADA owners and operators. 4This page intentionally left blank.

  19. [Dichotomizing method applied to calculating equilibrium constant of dimerization system].

    PubMed

    Cheng, Guo-zhong; Ye, Zhi-xiang

    2002-06-01

    The arbitrary trivariate algebraic equations are formed based on the combination principle. The univariata algebraic equation of equilibrium constant kappa for dimerization system is obtained through a series of algebraic transformation, and it depends on the properties of monotonic functions whether the equation is solvable or not. If the equation is solvable, equilibrium constant of dimerization system is obtained by dichotomy and its final equilibrium constant of dimerization system is determined according to the principle of error of fitting. The equilibrium constants of trisulfophthalocyanine and biosulfophthalocyanine obtained with this method are 47,973.4 and 30,271.8 respectively. The results are much better than those reported previously.

  20. EG&G Mound Applied Technologies payroll system

    SciTech Connect

    Not Available

    1992-02-07

    EG&G Mound Applied Technologies, Inc., manages and operates the Mound Facility, Miamisburg, Ohio, under a cost-plus-award-fee contract administered by the Department of Energy`s Albuquerque Field Office. The contractor`s Payroll Department is responsible for prompt payment in the proper amount to all persons entitled to be paid, in compliance with applicable laws, regulations, and legal decisions. The objective was to determine whether controls were in place to avoid erroneous payroll payments. EG&G Mound Applied Technologies, Inc., did not have all the internal controls required by General Accounting Office Title 6, ``Pay, Leave, and Allowances.`` Specifically, they did not have computerized edits, separation of duties and responsibilities, and restricted access to payroll data files. This condition occurred because its managers were not aware of Title 6 requirements. As a result, the contractor could not assure the Department of Energy that payroll costs were processes accurately; and fraud, waste, or abuse of Department of Energy funds could go undetected. Our sample of 212 payroll transactions from a population of 66,000 in FY 1991 disclosed only two minor processing errors and no instances of fraud, waste or abuse.

  1. Spectral Analysis and Filter Theory in Applied Geophysics

    NASA Astrophysics Data System (ADS)

    Buttkus, Burkhard; Newcomb, R. C.

    This book gives a comprehensive picture of the present stage of development of spectral analysis and filter theory in geophysics. The principles and theories behind classical and modern methods are described and the effectiveness of these methods is assessed; selected examples of their practical application in geophysics are discussed. The modern methods include, for example, spectral analysis by fitting random models to the data, the maximum-entropy and maximum-likelihood spectral analysis procedures, the Wiener and Kalman filters, homomorphic deconvolution, and adaptive procedures for non-stationary processes. This book represents a valuable aid in education and research and for solving practical problems in geophysics and related disciplines.

  2. Calibrating and Evaluating Boomless Spray Systems for Applying Forest Herbicides

    Treesearch

    Michael A. Wehr; Russell W. Johnson; Robert L. Sajdak

    1985-01-01

    Describes a testing procedure used to calibrate and evaluate agricultureal boomless spray systems. Tests allow the user to obtain dependable and satisfactory results when used in actual forest situations.

  3. Applying GIS technology to the Regional Information Sharing Systems database

    NASA Astrophysics Data System (ADS)

    Aumond, Karen L.

    1997-02-01

    The Regional Information Sharing Systems (RISS) program was formed as a partnership for information exchange between the federal government and state and local law enforcement. The six regional projects provide member law enforcement agencies in all 50 states with a broad range of intelligence and investigative support services. Recently, the existing RISS databases were redesigned to allow for connectivity among projects and the capability of a nationwide search of over 450,000 suspects. This relational database of intelligence information, along with a photographic imaging system, an operational `critical event' database, and GIS mapping are integrated components of RISSNET. The Geographical-Regional Information Sharing System (G-RISS) application is being prototypes by Graphic Data Systems Corporation at one RISS site, the Western States Information Network in Sacramento, California. G-RISS is a tool that will combine information from various law enforcement resources, map criminal activities to detect trends and assist agencies by being proactive to combat these activities.

  4. Applied estimation for hybrid dynamical systems using perceptional information

    NASA Astrophysics Data System (ADS)

    Plotnik, Aaron M.

    This dissertation uses the motivating example of robotic tracking of mobile deep ocean animals to present innovations in robotic perception and estimation for hybrid dynamical systems. An approach to estimation for hybrid systems is presented that utilizes uncertain perceptional information about the system's mode to improve tracking of its mode and continuous states. This results in significant improvements in situations where previously reported methods of estimation for hybrid systems perform poorly due to poor distinguishability of the modes. The specific application that motivates this research is an automatic underwater robotic observation system that follows and films individual deep ocean animals. A first version of such a system has been developed jointly by the Stanford Aerospace Robotics Laboratory and Monterey Bay Aquarium Research Institute (MBARI). This robotic observation system is successfully fielded on MBARI's ROVs, but agile specimens often evade the system. When a human ROV pilot performs this task, one advantage that he has over the robotic observation system in these situations is the ability to use visual perceptional information about the target, immediately recognizing any changes in the specimen's behavior mode. With the approach of the human pilot in mind, a new version of the robotic observation system is proposed which is extended to (a) derive perceptional information (visual cues) about the behavior mode of the tracked specimen, and (b) merge this dissimilar, discrete and uncertain information with more traditional continuous noisy sensor data by extending existing algorithms for hybrid estimation. These performance enhancements are enabled by integrating techniques in hybrid estimation, computer vision and machine learning. First, real-time computer vision and classification algorithms extract a visual observation of the target's behavior mode. Existing hybrid estimation algorithms are extended to admit this uncertain but discrete

  5. System Configured For Applying Multiple Modifying Agents To A Substrate.

    DOEpatents

    Propp, W. Alan; Argyle, Mark D.; Janikowski, Stuart K.; Fox, Robert V.; Toth, William J.; Ginosar, Daniel M.; Allen, Charles A.; Miller, David L.

    2005-11-08

    The present invention is related to the modifying of substrates with multiple modifying agents in a single continuous system. At least two processing chambers are configured for modifying the substrate in a continuous feed system. The processing chambers can be substantially isolated from one another by interstitial seals. Additionally, the two processing chambers can be substantially isolated from the surrounding atmosphere by end seals. Optionally, expansion chambers can be used to separate the seals from the processing chambers.

  6. APPLIED INFORMATION MANAGEMENT SYSTEM USER’S MANUAL

    DTIC Science & Technology

    Information Management System), an information storage and retrieval system used by the Engineering Department at SDC, is described. AIMS inputs are descriptors manually selected from the body and title of documents by a technique not requiring professional subject matter or abstracting skill. Outputs are special computer generated indexes to locate information stored. The most important is the Permuted Descriptor Index. The most common information handled by AIMS consists of SDC documents, technical articles, manuals and catalogs, correspondence, and the standard office

  7. System configured for applying multiple modifying agents to a substrate

    DOEpatents

    Propp, W. Alan; Argyle, Mark D.; Janikowski, Stuart K.; Fox, Robert V.; Toth, William J.; Ginosar, Daniel M.; Allen, Charles A.; Miller, David L.

    2003-11-25

    The present invention is related to the modifying of substrates with multiple modifying agents in a single continuous system. At least two processing chambers are configured for modifying the substrate in a continuous feed system. The processing chambers can be substantially isolated from one another by interstitial seals. Additionally, the two processing chambers can be substantially isolated from the surrounding atmosphere by end seals. Optionally, expansion chambers can be used to separate the seals from the processing chambers.

  8. Neutron-activation analysis applied to copper ores and artifacts

    NASA Technical Reports Server (NTRS)

    Linder, N. F.

    1970-01-01

    Neutron activation analysis is used for quantitative identification of trace metals in copper. Establishing a unique fingerprint of impurities in Michigan copper would enable identification of artifacts made from this copper.

  9. Calibration and investigation of infrared camera systems applying blackbody radiation

    NASA Astrophysics Data System (ADS)

    Hartmann, Juergen; Fischer, Joachim

    2001-03-01

    An experimental facility is presented, which allows calibration and detailed investigation of infrared camera systems. Various blackbodies operating in the temperature range from -60 degree(s)C up to 3000 degree(s)C serve as standard radiation sources, enabling calibration of camera systems in a wide temperature and spectral range with highest accuracy. Quantitative results and precise long-term investigations, especially in detecting climatic trends, require accurate traceability to the International Temperature Scale of 1990 (ITS-90). For the used blackbodies the traceability to ITS- 90 is either realized by standard platinum resistance thermometers (in the temperature range below 962 degree(s)C) or by absolute and relative radiometry (in the temperature range above 962 degree(s)C). This traceability is fundamental for implementation of quality assurance systems and realization of different standardizations, for example according ISO 9000. For investigation of the angular and the temperature resolution our set-up enables minimum resolvable (MRTD) and minimum detectable temperature difference (MDTD) measurements in the various temperature ranges. A collimator system may be used to image the MRTD and MDTD targets to infinity. As internal calibration of infrared camera systems critically depends on the temperature of the surrounding, the calibration and investigation of the cameras is performed in a climate box, which allows a detailed controlling of the environmental parameters like humidity and temperature. Experimental results obtained for different camera systems are presented and discussed.

  10. Dynamic state and parameter estimation applied to neuromorphic systems.

    PubMed

    Neftci, Emre Ozgur; Toth, Bryan; Indiveri, Giacomo; Abarbanel, Henry D I

    2012-07-01

    Neuroscientists often propose detailed computational models to probe the properties of the neural systems they study. With the advent of neuromorphic engineering, there is an increasing number of hardware electronic analogs of biological neural systems being proposed as well. However, for both biological and hardware systems, it is often difficult to estimate the parameters of the model so that they are meaningful to the experimental system under study, especially when these models involve a large number of states and parameters that cannot be simultaneously measured. We have developed a procedure to solve this problem in the context of interacting neural populations using a recently developed dynamic state and parameter estimation (DSPE) technique. This technique uses synchronization as a tool for dynamically coupling experimentally measured data to its corresponding model to determine its parameters and internal state variables. Typically experimental data are obtained from the biological neural system and the model is simulated in software; here we show that this technique is also efficient in validating proposed network models for neuromorphic spike-based very large-scale integration (VLSI) chips and that it is able to systematically extract network parameters such as synaptic weights, time constants, and other variables that are not accessible by direct observation. Our results suggest that this method can become a very useful tool for model-based identification and configuration of neuromorphic multichip VLSI systems.

  11. Applying Qualitative Hazard Analysis to Support Quantitative Safety Analysis for Proposed Reduced Wake Separation Conops

    NASA Technical Reports Server (NTRS)

    Shortle, John F.; Allocco, Michael

    2005-01-01

    This paper describes a scenario-driven hazard analysis process to identify, eliminate, and control safety-related risks. Within this process, we develop selective criteria to determine the applicability of applying engineering modeling to hypothesized hazard scenarios. This provides a basis for evaluating and prioritizing the scenarios as candidates for further quantitative analysis. We have applied this methodology to proposed concepts of operations for reduced wake separation for closely spaced parallel runways. For arrivals, the process identified 43 core hazard scenarios. Of these, we classified 12 as appropriate for further quantitative modeling, 24 that should be mitigated through controls, recommendations, and / or procedures (that is, scenarios not appropriate for quantitative modeling), and 7 that have the lowest priority for further analysis.

  12. Applying Qualitative Hazard Analysis to Support Quantitative Safety Analysis for Proposed Reduced Wake Separation Conops

    NASA Technical Reports Server (NTRS)

    Shortle, John F.; Allocco, Michael

    2005-01-01

    This paper describes a scenario-driven hazard analysis process to identify, eliminate, and control safety-related risks. Within this process, we develop selective criteria to determine the applicability of applying engineering modeling to hypothesized hazard scenarios. This provides a basis for evaluating and prioritizing the scenarios as candidates for further quantitative analysis. We have applied this methodology to proposed concepts of operations for reduced wake separation for closely spaced parallel runways. For arrivals, the process identified 43 core hazard scenarios. Of these, we classified 12 as appropriate for further quantitative modeling, 24 that should be mitigated through controls, recommendations, and / or procedures (that is, scenarios not appropriate for quantitative modeling), and 7 that have the lowest priority for further analysis.

  13. PIXE measurement applied to trace elemental analysis of human tissues

    NASA Astrophysics Data System (ADS)

    Tanaka, M.; Matsugi, E.; Miyasaki, K.; Yamagata, T.; Inoue, M.; Ogata, H.; Shimoura, S.

    1987-03-01

    PIXE measurement was applied for trace elemental analyses of 40 autoptic human kidneys. To investigate the reproducibility of the PIXE data, 9 targets obtained from one human liver were examined. The targets were prepared by wet-digestion using nitric and sulfuric acid. Yttrium was used as an internal standard. The extracted elemental concentrations for K, Fe, Cu, Zn, and Cd were in reasonable agreement with those obtained by atomic absorption spectrometry (AAS) and flame photometry (FP). Various correlations among the elements K, Ca, Cr, Mn, Fe, Ni, Cu, Zn, Rb, and Cd were examined individually for the renal cortex and renal medulla.

  14. Applied patent RFID systems for building reacting HEPA air ventilation system in hospital operation rooms.

    PubMed

    Lin, Jesun; Pai, Jar-Yuan; Chen, Chih-Cheng

    2012-12-01

    RFID technology, an automatic identification and data capture technology to provide identification, tracing, security and so on, was widely applied to healthcare industry in these years. Employing HEPA ventilation system in hospital is a way to ensure healthful indoor air quality to protect patients and healthcare workers against hospital-acquired infections. However, the system consumes lots of electricity which cost a lot. This study aims to apply the RFID technology to offer a unique medical staff and patient identification, and reacting HEPA air ventilation system in order to reduce the cost, save energy and prevent the prevalence of hospital-acquired infection. The system, reacting HEPA air ventilation system, contains RFID tags (for medical staffs and patients), sensor, and reacting system which receives the information regarding the number of medical staff and the status of the surgery, and controls the air volume of the HEPA air ventilation system accordingly. A pilot program was carried out in a unit of operation rooms of a medical center with 1,500 beds located in central Taiwan from Jan to Aug 2010. The results found the air ventilation system was able to function much more efficiently with less energy consumed. Furthermore, the indoor air quality could still keep qualified and hospital-acquired infection or other occupational diseases could be prevented.

  15. Applied Space Systems Engineering. Chapter 17; Manage Technical Data

    NASA Technical Reports Server (NTRS)

    Kent, Peter

    2008-01-01

    Effective space systems engineering (SSE) is conducted in a fully electronic manner. Competitive hardware, software, and system designs are created in a totally digital environment that enables rapid product design and manufacturing cycles, as well as a multitude of techniques such as modeling, simulation, and lean manufacturing that significantly reduce the lifecycle cost of systems. Because the SSE lifecycle depends on the digital environment, managing the enormous volumes of technical data needed to describe, build, deploy, and operate systems is a critical factor in the success of a project. This chapter presents the key aspects of Technical Data Management (TDM) within the SSE process. It is written from the perspective of the System Engineer tasked with establishing the TDM process and infrastructure for a major project. Additional perspectives are reflected from the point of view of the engineers on the project who work within the digital engineering environment established by the TDM toolset and infrastructure, and from the point of view of the contactors who interface via the TDM infrastructure. Table 17.1 lists the TDM process as it relates to SSE.

  16. Integrated hydrogen/oxygen technology applied to auxiliary propulsion systems

    NASA Technical Reports Server (NTRS)

    Gerhardt, David L.

    1990-01-01

    The purpose of the Integrated Hydrogen/Oxygen Technology (IHOT) study was to determine if the vehicle/mission needs and technology of the 1990's support development of an all cryogenic H2/O2 system. In order to accomplish this, IHOT adopted the approach of designing Integrated Auxiliary Propulsion Systems (IAPS) for a representative manned vehicle; the advanced manned launch system. The primary objectives were to develop IAPS concepts which appeared to offer viable alternatives to state-of-the-art (i.e., hypergolic, or earth-storable) APS approaches. The IHOT study resulted in the definition of three APS concepts; two cryogenic IAPS, and a third concept utilizing hypergolic propellants.

  17. Economical wind powered bioventing systems successfully applied at remote locations

    SciTech Connect

    Graves, D.; Klein, J.; Dillon, T. Jr.; Wilson, B.; Walker, K.

    1996-12-31

    Wind-powered bioventing systems were designed to operate at remote locations in the absence of electrical power. Laboratory measurements of soil respiration under bioventing conditions indicated the biodegradation of up to 25 mg of weathered diesel per kg of site soil per day. Further testing demonstrated the potential for harnessing wind-power to stimulate air movement through vadose zone soil. Several wind-powered bioventing systems were installed near Nome, Alaska. In situ respiration tests, soil gas composition measurements and measurable pressure changes in the soil indicated that the systems were capable of aerating the soil. Diesel range oil measurements indicated contaminant reductions up to 90% after only two treatments seasons. The results demonstrate the effectiveness of wind-powered biovents. The low cost, low maintenance, and simplicity of the biovents make them a very attractive treatment option for windy, remote sites with unsaturated soil impacted by biodegradable contaminants.

  18. Multi-agent cooperative systems applied to precision applications

    SciTech Connect

    McKay, M.D.; Anderson, M.O.; Gunderson, R.W.; Flann, N.; Abbott, B.

    1998-03-01

    Regulatory agencies are imposing limits and constraints to protect the operator and/or the environment. While generally necessary, these controls also tend to increase cost and decrease efficiency and productivity. Intelligent computer systems can be made to perform these hazardous tasks with greater efficiency and precision without danger to the operators. The Idaho national Engineering and Environmental Laboratory and the Center for Self-Organizing and Intelligent Systems at Utah State University have developed a series of autonomous all-terrain multi-agent systems capable of performing automated tasks within hazardous environments. This paper discusses the development and application of cooperative small-scale and large-scale robots for use in various activities associated with radiologically contaminated areas, prescription farming, and unexploded ordinances.

  19. Tributary Analysis Monitoring System

    NASA Technical Reports Server (NTRS)

    Woodard, Stanley, E. (Inventor); Coffey, Neil C. (Inventor); Taylor, Bryant D. (Inventor); Woodman, Keith L. (Inventor)

    2005-01-01

    A monitoring system for a fleet of vehicles includes at least one data acquisition and analysis module (DAAM) mounted on each vehicle in the fleet, a control module on each vehicle in communication with each DAAM, and terminal module located remotely with respect to the vehicles in the fleet. Each DAAM collects/analyzes sensor data to generate analysis results that identify the state of a plurality of systems of the vehicle. Each vehicle's control module collects/analyzes the analysis results from each onboard DAAM to generate vehicle status results that identify potential sources of vehicle anomalies. The terminal module collects/analyzes the analysis results and vehicle status results transmitted from each control module from the fleet of vehicles to identify multiple occurrences of vehicle anomalies and multiple occurrences of those vehicle systems operating at a performance level that is unacceptable. Results of the terminal module's analysis are provided to organizations responsible for the operation, maintenance and manufacturing of the vehicles in the fleet as well as the plurality of systems used in the fleet.

  20. Applying systems biology in drug discovery and development.

    PubMed

    Galizzi, Jean-Pierre; Lockhart, Brian Paul; Bril, Antoine

    2013-01-01

    Translational research is a continuum between clinical and basic research where the patient is the center of the research process. It brings clinical research to a starting point for the drug discovery process, permitting the generation of a more robust pathophysiological hypothesis essential for a better selection of drug targets and candidate optimization. It also establishes the basis of early proof for clinical concept studies, preferably in phase I, for which biomarkers and surrogate endpoints can often be used. Systems biology is a prerequisite approach to translational research where technologies and expertise are integrated and articulated to support efficient and productive realization of this concept. The first component of systems biology relies on omics-based technologies and integrates the changes in variables, such as genes, proteins and metabolites, into networks that are responsible for an organism's normal and diseased state. The second component of systems biology is in the domain of computational methods, where simulation and modeling create hypotheses of signaling pathways, transcription networks, physiological processes or even cell- or organism-based models. The simulations aim to show the origin of perturbations of the system that lead to pathological states and what treatment could be achieved to ameliorate or normalize the system. This review discusses how translational research and systems biology together could improve global understanding of drug targets, suggest new targets and approaches for therapeutics, and provide a deeper understanding of drug effects. Taken together, these types of analyses can lead to new therapeutic options while improving the safety and efficacy of new and existing medications.

  1. Biomedical systems analysis program

    NASA Technical Reports Server (NTRS)

    1979-01-01

    Biomedical monitoring programs which were developed to provide a system analysis context for a unified hypothesis for adaptation to space flight are presented and discussed. A real-time system of data analysis and decision making to assure the greatest possible crew safety and mission success is described. Information about man's abilities, limitations, and characteristic reactions to weightless space flight was analyzed and simulation models were developed. The predictive capabilities of simulation models for fluid-electrolyte regulation, erythropoiesis regulation, and calcium regulation are discussed.

  2. Robust sliding mode control applied to double Inverted pendulum system

    SciTech Connect

    Mahjoub, Sonia; Derbel, Nabil; Mnif, Faical

    2009-03-05

    A three hierarchical sliding mode control is presented for a class of an underactuated system which can overcome the mismatched perturbations. The considered underactuated system is a double inverted pendulum (DIP), can be modeled by three subsystems. Such structure allows the construction of several designs of hierarchies for the controller. For all hierarchical designs, the asymptotic stability of every layer sliding mode surface and the sliding mode surface of subsystems are proved theoretically by Barbalat's lemma. Simulation results show the validity of these methods.

  3. Error behavior of multistep methods applied to unstable differential systems

    NASA Technical Reports Server (NTRS)

    Brown, R. L.

    1977-01-01

    The problem of modeling a dynamic system described by a system of ordinary differential equations which has unstable components for limited periods of time is discussed. It is shown that the global error in a multistep numerical method is the solution to a difference equation initial value problem, and the approximate solution is given for several popular multistep integration formulas. Inspection of the solution leads to the formulation of four criteria for integrators appropriate to unstable problems. A sample problem is solved numerically using three popular formulas and two different stepsizes to illustrate the appropriateness of the criteria.

  4. 40 CFR 63.8030 - What requirements apply to my heat exchange systems?

    Code of Federal Regulations, 2010 CFR

    2010-07-01

    ... 40 Protection of Environment 13 2010-07-01 2010-07-01 false What requirements apply to my heat... apply to my heat exchange systems? (a) You must comply with the requirements specified in Table 6 to this subpart that apply to your heat exchange systems, except as specified in paragraphs (b) through...

  5. 40 CFR 63.8030 - What requirements apply to my heat exchange systems?

    Code of Federal Regulations, 2011 CFR

    2011-07-01

    ... 40 Protection of Environment 13 2011-07-01 2011-07-01 false What requirements apply to my heat... apply to my heat exchange systems? (a) You must comply with the requirements specified in Table 6 to this subpart that apply to your heat exchange systems, except as specified in paragraphs (b) through (e...

  6. 40 CFR 63.8030 - What requirements apply to my heat exchange systems?

    Code of Federal Regulations, 2012 CFR

    2012-07-01

    ... 40 Protection of Environment 14 2012-07-01 2011-07-01 true What requirements apply to my heat... apply to my heat exchange systems? (a) You must comply with the requirements specified in Table 6 to this subpart that apply to your heat exchange systems, except as specified in paragraphs (b) through (e...

  7. 40 CFR 63.8030 - What requirements apply to my heat exchange systems?

    Code of Federal Regulations, 2014 CFR

    2014-07-01

    ... 40 Protection of Environment 14 2014-07-01 2014-07-01 false What requirements apply to my heat... apply to my heat exchange systems? (a) You must comply with the requirements specified in Table 6 to this subpart that apply to your heat exchange systems, except as specified in paragraphs (b) through (e...

  8. Cold-Applied Roofing Systems and Waterproofing and Dampproofing. Roofing Workbook and Tests.

    ERIC Educational Resources Information Center

    Brown, Arthur

    This workbook for students in California roofing apprenticeship programs provides information for classroom work in the area of cold-applied roofing systems and waterproofing and dampproofing. Eight topics are covered: introduction to cold-applied roofing systems and waterproofing and dampproofing, tools and equipment used in cold-applied roofing,…

  9. Integrated analysis environment for high impact systems

    SciTech Connect

    Martinez, M.; Davis, J.; Scott, J.; Sztipanovits, J.; Karsai, G.

    1998-02-01

    Modeling and analysis of high consequence, high assurance systems requires special modeling considerations. System safety and reliability information must be captured in the models. Previously, high consequence systems were modeled using separate, disjoint models for safety, reliability, and security. The MultiGraph Architecture facilitates the implementation of a model integrated system for modeling and analysis of high assurance systems. Model integrated computing allows an integrated modeling technique to be applied to high consequence systems. Among the tools used for analyzing safety and reliability are a behavioral simulator and an automatic fault tree generation and analysis tool. Symbolic model checking techniques are used to efficiently investigate the system models. A method for converting finite state machine models to ordered binary decision diagrams allows the application of symbolic model checking routines to the integrated system models. This integrated approach to modeling and analysis of high consequence systems ensures consistency between the models and the different analysis tools.

  10. GPS ensemble analysis applied to Antarctic vertical velocities

    NASA Astrophysics Data System (ADS)

    Petrie, E. J.; Clarke, P. J.; King, M. A.; Williams, S. D. P.

    2014-12-01

    GPS data is used to provide estimates of vertical land motion caused by e.g. glacial isostatic adjustment (GIA) and hydrologic loading. The vertical velocities estimated from the GPS data are often assimilated into GIA models or used for comparison purposes. GIA models are very important as they provide time-variable gravity corrections needed to estimate ice mass change over Greenland and Antarctica. While state-of-the art global GPS analysis has previously been performed for many Antarctic sites, formal errors in the resulting site velocities are typically obtained from noise analysis of each individual time series without consideration of processing or metadata issues. Here we present analysis of the results from two full global runs including a variety of parameter and reference frame alignment choices, and compare the results to previous work with a view to assessing if the size of the formal errors from the standard method is truly representative.

  11. Orbit Response Matrix Analysis Applied at PEP-II

    SciTech Connect

    Steier, C.; Wolski, A.; Ecklund, S.; Safranek, J.A.; Tenenbaum, P.; Terebilo, A.; Turner, J.L.; Yocky, G.; /SLAC

    2005-05-17

    The analysis of orbit response matrices has been used very successfully to measure and correct the gradient and skew gradient distribution in many accelerators. It allows determination of an accurately calibrated model of the coupled machine lattice, which then can be used to calculate the corrections necessary to improve coupling, dynamic aperture and ultimately luminosity. At PEP-II, the Matlab version of LOCO has been used to analyze coupled response matrices for both the LER and the HER. The large number of elements in PEP-II and the very complicated interaction region present unique challenges to the data analysis. All necessary tools to make the analysis method useable at PEP-II have been implemented and LOCO can now be used as a routine tool for lattice diagnostic.

  12. Joint regression analysis and AMMI model applied to oat improvement

    NASA Astrophysics Data System (ADS)

    Oliveira, A.; Oliveira, T. A.; Mejza, S.

    2012-09-01

    In our work we present an application of some biometrical methods useful in genotype stability evaluation, namely AMMI model, Joint Regression Analysis (JRA) and multiple comparison tests. A genotype stability analysis of oat (Avena Sativa L.) grain yield was carried out using data of the Portuguese Plant Breeding Board, sample of the 22 different genotypes during the years 2002, 2003 and 2004 in six locations. In Ferreira et al. (2006) the authors state the relevance of the regression models and of the Additive Main Effects and Multiplicative Interactions (AMMI) model, to study and to estimate phenotypic stability effects. As computational techniques we use the Zigzag algorithm to estimate the regression coefficients and the agricolae-package available in R software for AMMI model analysis.

  13. Career Counseling: Applying the Systems Theory Framework of Career Development

    ERIC Educational Resources Information Center

    McMahon, Mary

    2005-01-01

    In recent years, career development and career counseling have increasingly been informed by concepts emanating from the constructivist worldview. For example, the Systems Theory Framework (STF; M. McMahon, 2002; M. McMahon & W. Patton, 1995; W. Patton & M. McMahon, 1997, 1999) of career development has been proposed as a metatheoretical account…

  14. System Identification and POD Method Applied to Unsteady Aerodynamics

    NASA Technical Reports Server (NTRS)

    Tang, Deman; Kholodar, Denis; Juang, Jer-Nan; Dowell, Earl H.

    2001-01-01

    The representation of unsteady aerodynamic flow fields in terms of global aerodynamic modes has proven to be a useful method for reducing the size of the aerodynamic model over those representations that use local variables at discrete grid points in the flow field. Eigenmodes and Proper Orthogonal Decomposition (POD) modes have been used for this purpose with good effect. This suggests that system identification models may also be used to represent the aerodynamic flow field. Implicit in the use of a systems identification technique is the notion that a relative small state space model can be useful in describing a dynamical system. The POD model is first used to show that indeed a reduced order model can be obtained from a much larger numerical aerodynamical model (the vortex lattice method is used for illustrative purposes) and the results from the POD and the system identification methods are then compared. For the example considered, the two methods are shown to give comparable results in terms of accuracy and reduced model size. The advantages and limitations of each approach are briefly discussed. Both appear promising and complementary in their characteristics.

  15. AIRS FPA applied to the MIRIADS: powerful infrared systems applications

    NASA Astrophysics Data System (ADS)

    Caulfield, John T.; McCarley, Paul L.; Baxter, Christopher R.; Massie, Mark A.

    2001-10-01

    Raytheon's Infrared Operations (RIO) has invented and developed a new type of focal plane array with 'Image Processing on-the-chip' named the Adaptive IR Sensor (AIRS). The AIRS FPA is based upon the human retina in that it performs signal processing near the photoreceptors. The AIRS FPA has been reduced to practice and adaptively removes detector and optic temperature drift and 1/f induced fixed pattern noise. This 3rd-generation multi-mode IRFPA, also called a Smart FPA, is a 256 X 256-array format capable of operation in four modes: (1) Direct Injection (DI), (2) Adaptive Non-uniformity Correction (NUC), (3) Motion/Edge Detection, and (4) Subframe Averaging. Nova Research has developed a Modular IR Application Development System (MIRIADS) which is a compact single board camera system that is highly integrated with the dewar assembly. The MIRIADS system coupled with the AIRS Smart FPA results in a very high performance wide field of view IR Sensor and processing system with integrated in one of the smallest packages to date.

  16. Quality assurance applied to animal disease surveillance systems.

    PubMed

    Salman; Stärk, K D C; Zepeda, C

    2003-08-01

    Monitoring and surveillance systems (MOSS) are essential activities for official Veterinary Services. In addition, the increased trade in animals and animal products over recent years has increased the importance of international disease reporting. A reliable surveillance system is the key to early warning of a change in the health status of any animal population. Such a system is also essential for providing evidence about the absence of diseases or in determining the extent of a disease which is known to be present. The authors discuss a set of methods and approaches for evaluating the quality of surveillance and survey systems. Certain steps are required when assessing the quality of a service or product. Various approaches for quality assessment are available and the suitability of each method depends on the objective of the evaluation. An essential basic requirement is, however, to use an objective, transparent and systematic approach. The evidence collected and the analyses used to reach conclusions must be of such high quality that the results are acceptable to both the management of the MOSS and the assessor. Repeated discussions and negotiations may be necessary to reach consensus, particularly if the judgement affects activities between trading partners. Well-documented MOSS with specified objectives and integrated quality assurance mechanisms are likely to be easier to evaluate.

  17. Best practices: applying management analysis of excellence to immunization.

    PubMed

    Wishner, Amy; Aronson, Jerold; Kohrt, Alan; Norton, Gary

    2005-01-01

    The authors applied business management tools to analyze and promote excellence and to evaluate differences between average and above-average immunization peformers in private practices. The authors conducted a pilot study of 10 private practices in Pennsylvania using tools common in management to assess practices' organizational climate and managerial style. Authoritative and coaching styles of physician leaders were common to both groups. Managerial styles that emphasized higher levels of clarity and responsibility managerial styles were evident in the large practices; and rewards and flexibility styles were higher in the small above-average practices. The findings of this pilot study match results seen in high performers in other industries. It concludes that the authoritative style appears to have the most impact on performance. It has interesting implications for training/behavior change to improve immunization rates, along with traditional medical interventions.

  18. A value analysis model applied to the management of amblyopia.

    PubMed Central

    Beauchamp, G R; Bane, M C; Stager, D R; Berry, P M; Wright, W W

    1999-01-01

    PURPOSE: To assess the value of amblyopia-related services by utilizing a health value model (HVM). Cost and quality criteria are evaluated in accordance with the interests of patients, physicians, and purchasers. METHODS: We applied an HVM to a hypothetical statistical ("median") child with amblyopia whose visual acuity is 20/80 and to a group of children with amblyopia who are managed by our practice. We applied the model to calculate the value of these services by evaluating the responses of patients and physicians and relating these responses to clinical outcomes. RESULTS: The consensus value of care for the hypothetical median child was calculated to be 0.406 (of 1.000). For those children managed in our practice, the calculated value is 0.682. Clinically, 79% achieved 20/40 or better visual acuity, and the mean final visual acuity was 0.2 logMAR (20/32). Value appraisals revealed significant concerns about the financial aspects of amblyopia-related services, particularly among physicians. Patients rated services more positively than did physicians. CONCLUSIONS: Amblyopia care is difficult, sustained, and important work that requires substantial sensitivity to and support of children and families. Compliance and early detection are essential to success. The value of amblyopia services is rated significantly higher by patients than by physicians. Relative to the measured value, amblyopia care is undercompensated. The HVM is useful to appraise clinical service delivery and its variation. The costs of failure and the benefits of success are high; high-value amblyopia care yields substantial dividends and should be commensurately compensated in the marketplace. PMID:10703133

  19. Applying Score Analysis to a Rehearsal Pedagogy of Expressive Performance

    ERIC Educational Resources Information Center

    Byo, James L.

    2014-01-01

    The discoveries of score analysis (e.g., minor seventh chord, ostinato, phrase elision, melodic fragment, half cadence) are more than just compositional techniques or music vocabulary. They are sounds--fascinating, storytelling, dynamic modes of expression--that when approached as such enrich the rehearsal experience. This article presents a…

  20. Applying Adult Learning Theory through a Character Analysis

    ERIC Educational Resources Information Center

    Baskas, Richard S.

    2011-01-01

    The purpose of this study is to analyze the behavior of a character, Celie, in a movie, 'The Color Purple," through the lens of two adult learning theorists to determine the relationships the character has with each theory. The development and portrayal of characters in movies can be explained and understood by the analysis of adult learning…

  1. Applying MORT maintenance safety analysis in Finnish industry

    NASA Astrophysics Data System (ADS)

    Ruuhilehto, Kaarin; Virolainen, Kimmo

    1992-02-01

    A safety analysis method based on MORT (Management Oversight and Risk Tree) method, especially on the version developed for safety considerations in the evaluation of maintenance programs, is presented. The MORT maintenance safety analysis is intended especially for the use maintenance safety management. The analysis helps managers evaluate the goals of their safety work and measures taken to reach them. The analysis is done by a team or teams. The team ought to have expert knowledge of the organization both vertically and horizontally in order to be able to identify factors that may contribute to accidents or other interruptions in the maintenance work. Identification is made by using the MORT maintenance key question set as a check list. The questions check the way safety matters are connnected with the maintenance planning and managing, as well as the safety management itself. In the second stage, means to eliminate the factors causing problems are developed. New practices are established to improve safety of maintenance planning and managing in the enterprise.

  2. Applying Skinner's Analysis of Verbal Behavior to Persons with Dementia

    ERIC Educational Resources Information Center

    Dixon, Mark; Baker, Jonathan C.; Sadowski, Katherine Ann

    2011-01-01

    Skinner's 1957 analysis of verbal behavior has demonstrated a fair amount of utility to teach language to children with autism and other various disorders. However, the learning of language can be forgotten, as is the case for many elderly suffering from dementia or other degenerative diseases. It appears possible that Skinner's operants may…

  3. Action, Content and Identity in Applied Genre Analysis for ESP

    ERIC Educational Resources Information Center

    Flowerdew, John

    2011-01-01

    Genres are staged, structured, communicative events, motivated by various communicative purposes, and performed by members of specific discourse communities (Swales 1990; Bhatia 1993, 2004; Berkenkotter & Huckin 1995). Since its inception, with the two seminal works on the topic by Swales (1990) and Bhatia (1993), genre analysis has taken pride of…

  4. Applying Score Analysis to a Rehearsal Pedagogy of Expressive Performance

    ERIC Educational Resources Information Center

    Byo, James L.

    2014-01-01

    The discoveries of score analysis (e.g., minor seventh chord, ostinato, phrase elision, melodic fragment, half cadence) are more than just compositional techniques or music vocabulary. They are sounds--fascinating, storytelling, dynamic modes of expression--that when approached as such enrich the rehearsal experience. This article presents a…

  5. Applying Skinner's Analysis of Verbal Behavior to Persons with Dementia

    ERIC Educational Resources Information Center

    Dixon, Mark; Baker, Jonathan C.; Sadowski, Katherine Ann

    2011-01-01

    Skinner's 1957 analysis of verbal behavior has demonstrated a fair amount of utility to teach language to children with autism and other various disorders. However, the learning of language can be forgotten, as is the case for many elderly suffering from dementia or other degenerative diseases. It appears possible that Skinner's operants may…

  6. Applying Contamination Modelling to Spacecraft Propulsion Systems Designs and Operations

    NASA Technical Reports Server (NTRS)

    Chen, Philip T.; Thomson, Shaun; Woronowicz, Michael S.

    2000-01-01

    Molecular and particulate contaminants generated from the operations of a propulsion system may impinge on spacecraft critical surfaces. Plume depositions or clouds may hinder the spacecraft and instruments from performing normal operations. Firing thrusters will generate both molecular and particulate contaminants. How to minimize the contamination impact from the plume becomes very critical for a successful mission. The resulting effect from either molecular or particulate contamination of the thruster firing is very distinct. This paper will discuss the interconnection between the functions of spacecraft contamination modeling and propulsion system implementation. The paper will address an innovative contamination engineering approach implemented from the spacecraft concept design, manufacturing, integration and test (I&T), launch, to on- orbit operations. This paper will also summarize the implementation on several successful missions. Despite other contamination sources, only molecular contamination will be considered here.

  7. Applying twisted boundary conditions for few-body nuclear systems

    NASA Astrophysics Data System (ADS)

    Körber, Christopher; Luu, Thomas

    2016-05-01

    We describe and implement twisted boundary conditions for the deuteron and triton systems within finite volumes using the nuclear lattice EFT formalism. We investigate the finite-volume dependence of these systems with different twist angles. We demonstrate how various finite-volume information can be used to improve calculations of binding energies in such a framework. Our results suggests that with appropriate twisting of boundaries, infinite-volume binding energies can be reliably extracted from calculations using modest volume sizes with cubic length L ≈8 -14 fm. Of particular importance is our derivation and numerical verification of three-body analogs of "i-periodic" twist angles that eliminate the leading-order finite-volume effects to the three-body binding energy.

  8. Parabolic dish systems at work - Applying the concepts

    NASA Technical Reports Server (NTRS)

    Marriott, A. T.

    1981-01-01

    An overview is given of parabolic dish solar concentrator application experiments being conducted by the U.S. Department of Energy. The 'engineering experiments' comprise the testing of (1) a small-community powerplant system, in conjunction with a grid-connected utility; (2) stand-alone applications at remote sites such as military installations, radar stations and villages; and (3) dish modules that can deliver heat for direct use in industrial processes. Applicability projections are based on a dish and receiver that use a Brayton engine with an engine/generator efficiency of 25% and a production level of up to 25,000 units per year. Analyses indicate that parabolic-dish power systems can potentially replace small, oil-fired power plants in all regions of the U.S. between 1985 and 1991.

  9. Simulation program of nonlinearities applied to telecommunication systems

    NASA Technical Reports Server (NTRS)

    Thomas, C.

    1979-01-01

    In any satellite communication system, the problems of distorsion created by nonlinear devices or systems must be considered. The subject of this paper is the use of the Fast Fourier Transform (F.F.T.) in the prediction of the intermodulation performance of amplifiers, mixers, filters. A nonlinear memory-less model is chosen to simulate amplitude and phase nonlinearities of the device in the simulation program written in FORTRAN 4. The experimentally observed nonlinearity parameters of a low noise 3.7-4.2 GHz amplifier are related to the gain and phase coefficients of Fourier Service Series. The measured results are compared with those calculated from the simulation in the cases where the input signal is composed of two, three carriers and noise power density.

  10. Arctic Climate Systems Analysis

    SciTech Connect

    Ivey, Mark D.; Robinson, David G.; Boslough, Mark B.; Backus, George A.; Peterson, Kara J.; van Bloemen Waanders, Bart G.; Swiler, Laura Painton; Desilets, Darin Maurice; Reinert, Rhonda Karen

    2015-03-01

    This study began with a challenge from program area managers at Sandia National Laboratories to technical staff in the energy, climate, and infrastructure security areas: apply a systems-level perspective to existing science and technology program areas in order to determine technology gaps, identify new technical capabilities at Sandia that could be applied to these areas, and identify opportunities for innovation. The Arctic was selected as one of these areas for systems level analyses, and this report documents the results. In this study, an emphasis was placed on the arctic atmosphere since Sandia has been active in atmospheric research in the Arctic since 1997. This study begins with a discussion of the challenges and benefits of analyzing the Arctic as a system. It goes on to discuss current and future needs of the defense, scientific, energy, and intelligence communities for more comprehensive data products related to the Arctic; assess the current state of atmospheric measurement resources available for the Arctic; and explain how the capabilities at Sandia National Laboratories can be used to address the identified technological, data, and modeling needs of the defense, scientific, energy, and intelligence communities for Arctic support.

  11. Interior Ballistics Modeling Applied to Small Arms Systems

    DTIC Science & Technology

    1979-06-01

    equations 1 Xl X2 Xn TA dT1 d2 dn or d = Xldl + X2d2 + .... ÷ Xndn B where X1 is the weight fraction of a narrow fraction of diameter dl. Subscripts A and B...DRDAR-TSS (S) Dover, NJ 07801 Commander Defense Documentation Center (12) Camerson Station Alexandria, VA 22314 Weapon System Concept Team/ CSL ATTN

  12. Applying Systems Engineering Methodologies to the Creative Process

    DTIC Science & Technology

    2014-09-01

    the Goal/Question/Metric Approach (from Vanek, Jackson, and Gryzbowski 2008, 116) ................................................. 70 Figure 14... approach to systems, which can be used to remove this first roadblock (Haskins 2011, 6). Second, the elusiveness of a sound definition of creativity...to be produced and utilized. Third, early creativity study was characterized by “… approaches that have tended to view creativity as an

  13. Nonlinear Waves, Dynamical Systems and Other Applied Mathematics Programs

    DTIC Science & Technology

    1991-10-04

    present a general scheme of perturbation method for perturbed soliton systems, based on the normal form theory and the method of multiple scales. By this...dimension, and discuss possible consequences of the interplay between wavefront- interactions and curvature in two dimensions. Thursday, October 19 All ... normal speed D parametrized by the local mean surface curvature x. Its solution provides a relation D = D(x) which determines the evolution of the front

  14. Applying an integrated neuro-expert system model in a real-time alarm processing system

    NASA Astrophysics Data System (ADS)

    Khosla, Rajiv; Dillon, Tharam S.

    1993-03-01

    In this paper we propose an integrated model which is derived from the combination of a generic neuro-expert system model, an object model, and unix operating system process (UOSP) model. This integrated model reflects the strengths of both artificial neural nets (ANNs) and expert systems (ESs). A formalism of ES object, ANN object, UOSP object, and problem domain object is used for developing a set of generic data structures and methods. These generic data structures and methods help us to build heterogeneous ES-ANN objects with uniform communication interface. The integrated model is applied in a real-time alarm processing system for a non-trivial terminal power station. It is shown how features like hierarchical/distributed ES/ANN objects, inter process communication, and fast concurrent execution help to cope with real-time system constraints like, continuity, data variability, and fast response time.

  15. Current Human Reliability Analysis Methods Applied to Computerized Procedures

    SciTech Connect

    Ronald L. Boring

    2012-06-01

    Computerized procedures (CPs) are an emerging technology within nuclear power plant control rooms. While CPs have been implemented internationally in advanced control rooms, to date no US nuclear power plant has implemented CPs in its main control room (Fink et al., 2009). Yet, CPs are a reality of new plant builds and are an area of considerable interest to existing plants, which see advantages in terms of enhanced ease of use and easier records management by omitting the need for updating hardcopy procedures. The overall intent of this paper is to provide a characterization of human reliability analysis (HRA) issues for computerized procedures. It is beyond the scope of this document to propose a new HRA approach or to recommend specific methods or refinements to those methods. Rather, this paper serves as a review of current HRA as it may be used for the analysis and review of computerized procedures.

  16. Nested sampling applied in Bayesian room-acoustics decay analysis.

    PubMed

    Jasa, Tomislav; Xiang, Ning

    2012-11-01

    Room-acoustic energy decays often exhibit single-rate or multiple-rate characteristics in a wide variety of rooms/halls. Both the energy decay order and decay parameter estimation are of practical significance in architectural acoustics applications, representing two different levels of Bayesian probabilistic inference. This paper discusses a model-based sound energy decay analysis within a Bayesian framework utilizing the nested sampling algorithm. The nested sampling algorithm is specifically developed to evaluate the Bayesian evidence required for determining the energy decay order with decay parameter estimates as a secondary result. Taking the energy decay analysis in architectural acoustics as an example, this paper demonstrates that two different levels of inference, decay model-selection and decay parameter estimation, can be cohesively accomplished by the nested sampling algorithm.

  17. VENTILATION TECHNOLOGY SYSTEMS ANALYSIS

    EPA Science Inventory

    The report gives results of a project to develop a systems analysis of ventilation technology and provide a state-of-the-art assessment of ventilation and indoor air quality (IAQ) research needs. (NOTE: Ventilation technology is defined as the hardware necessary to bring outdoor ...

  18. VENTILATION TECHNOLOGY SYSTEMS ANALYSIS

    EPA Science Inventory

    The report gives results of a project to develop a systems analysis of ventilation technology and provide a state-of-the-art assessment of ventilation and indoor air quality (IAQ) research needs. (NOTE: Ventilation technology is defined as the hardware necessary to bring outdoor ...

  19. URBAN EDUCATION SYSTEMS ANALYSIS.

    ERIC Educational Resources Information Center

    CLARK, STEPHEN C.; AND OTHERS

    ANALYSIS OF URBAN EDUCATIONAL SYSTEMS MAY BE ACHIEVED BY USE OF AN ANALYTICAL MODEL. THE MODEL MAY BE USED IN DECISION-MAKING REGARDING SCHOOL LOCATION, ENROLLMENT, FACILITIES, ORGANIZATION, PROGRAMS AND COSTS. KNOWN DATA SUCH AS MONIES AVAILABLE, STAFF ALLOCATION, AND CURRENT SCHOOL PLANT ARE INTRODUCED INTO THE MODEL. IN THE STRUCTURE OF THE…

  20. Applying Cognitive Work Analysis to Time Critical Targeting Functionality

    DTIC Science & Technology

    2004-10-01

    Target List/Dynamic Target Queue (DTL/ DTQ ) in the same place. Figure 4-27 shows the task steps involved in achieving Goal 7. 4- 30 Figure 4-27...GUI WG to brainstorm the order of columns in the DTL/ DTQ Table, a critical component of the TCTF CUI, with successful results, which were...Cognitive Work Analysis DTD Display Task Description DTL/ DTQ Dynamic Target List/Dynamic Target Queue FDO Fighter Duty Officer FEBA Forward Edge

  1. Applying thiouracil (TU)-tagging for mouse transcriptome analysis

    PubMed Central

    Gay, Leslie; Karfilis, Kate V.; Miller, Michael R.; Doe, Chris Q.; Stankunas, Kryn

    2014-01-01

    Transcriptional profiling is a powerful approach to study mouse development, physiology, and disease models. Here, we describe a protocol for mouse thiouracil-tagging (TU-tagging), a transcriptome analysis technology that includes in vivo covalent labeling, purification, and analysis of cell type-specific RNA. TU-tagging enables 1) the isolation of RNA from a given cell population of a complex tissue, avoiding transcriptional changes induced by cell isolation trauma, and 2) the identification of actively transcribed RNAs and not pre-existing transcripts. Therefore, in contrast to other cell-specific transcriptional profiling methods based on purification of tagged ribosomes or nuclei, TU-tagging provides a direct examination of transcriptional regulation. We describe how to: 1) deliver 4-thiouracil to transgenic mice to thio-label cell lineage-specific transcripts, 2) purify TU-tagged RNA and prepare libraries for Illumina sequencing, and 3) follow a straight-forward bioinformatics workflow to identify cell type-enriched or differentially expressed genes. Tissue containing TU-tagged RNA can be obtained in one day, RNA-Seq libraries generated within two days, and, following sequencing, an initial bioinformatics analysis completed in one additional day. PMID:24457332

  2. Innovative wave energy device applied to coastal observatory systems

    NASA Astrophysics Data System (ADS)

    Marcelli, Marco; Bonamano, Simone; Piermattei, Viviana; Scanu, Sergio; Paladini de Mendoza, Francesco; Martellucci, Riccardo; Maximo, Peviani

    2017-04-01

    Marine environment is one of the most promising sources of renewable energy, whose exploitation could have an effect on several application fields. This work presents the design of an innovative device based on the Oscillating Water Column (OWC), that allows to convert wave energy into electricity, suitable for the typical Mediterranean wave climate. The flexibility of the device permits its installation either in deep or shallow waters, with reduced costs of deployment, maintenance and connection to the grid. Furthermore, the replicability of the design allows the device to be installed in array of several number of similar units. The technical concept is to convey the sea water within a vertical pipe, in which the water movements activate a rotor connected to a generator that transforms the energy of the water motion into electricity. The hydrodynamic design of the pipe is built to minimize the losses due to friction and turbulence and to exploit the maximum possible energy from wave motion. The wave energy is directly absorbed by the rotational movement of the turbine blades located in the water itself allowing a further reduction of the energy losses associated with the transformation of the linear water motion into electrical generation in the air phase (typical configuration of the OWC devices). In this work the device components are described considering two possible configurations that use a Wells turbine or a Bulb type turbine. The system can be realized at a low cost, because of the modularity of the device project, which allows large freedom of sizes and placements, being able to be installed both individually and in arrays. The modularity, associated with the fact that the main elements of the system are available on the market, makes the device particularly attractive from the economic point of view. Finally, it is realized with a high constructive flexibility: the proposed system can be transported floating and moored to existing coastal structures or

  3. An applied study using systems engineering methods to prioritize green systems options

    SciTech Connect

    Lee, Sonya M; Macdonald, John M

    2009-01-01

    For many years, there have been questions about the effectiveness of applying different green solutions. If you're building a home and wish to use green technologies, where do you start? While all technologies sound promising, which will perform the best over time? All this has to be considered within the cost and schedule of the project. The amount of information available on the topic can be overwhelming. We seek to examine if Systems Engineering methods can be used to help people choose and prioritize technologies that fit within their project and budget. Several methods are used to gain perspective into how to select the green technologies, such as the Analytic Hierarchy Process (AHP) and Kepner-Tregoe. In our study, subjects applied these methods to analyze cost, schedule, and trade-offs. Results will document whether the experimental approach is applicable to defining system priorities for green technologies.

  4. Near-infrared radiation curable multilayer coating systems and methods for applying same

    DOEpatents

    Bowman, Mark P; Verdun, Shelley D; Post, Gordon L

    2015-04-28

    Multilayer coating systems, methods of applying and related substrates are disclosed. The coating system may comprise a first coating comprising a near-IR absorber, and a second coating deposited on a least a portion of the first coating. Methods of applying a multilayer coating composition to a substrate may comprise applying a first coating comprising a near-IR absorber, applying a second coating over at least a portion of the first coating and curing the coating with near infrared radiation.

  5. Application of acousto-optic actuator applied in holographic system

    NASA Astrophysics Data System (ADS)

    Ling, FuRi; Wang, Biao

    2002-09-01

    In this paper, we discuss acousto-optical scanning and deflection, and design an acousto-optical actuator for steering the laser beam in the direction of vertical and horizon. In this system a laser whose wavelength is 532 nm is used and is expanded by a cylindrical lens. This horizontal actuator produces the horizontal deflection and the spherical lens following the horizontal actuator rotates the beam to match the aperture of the vertical actuator. The cylindrical lens restores the beam to its original circular cross-section, after which the microscope optics brings it to a focus in the lithium niobate crystal in which we store information.

  6. System And Method Of Applying Energetic Ions For Sterlization

    DOEpatents

    Schmidt, John A.

    2002-06-11

    A method of sterilization of a container is provided whereby a cold plasma is caused to be disposed near a surface to be sterilized, and the cold plasma is then subjected to a pulsed voltage differential for producing energized ions in the plasma. Those energized ions then operate to achieve spore destruction on the surface to be sterilized. Further, a system for sterilization of a container which includes a conductive or non-conductive container, a cold plasma in proximity to the container, and a high voltage source for delivering a pulsed voltage differential between an electrode and the container and across the cold plasma, is provided.

  7. System and method of applying energetic ions for sterilization

    DOEpatents

    Schmidt, John A.

    2003-12-23

    A method of sterilization of a container is provided whereby a cold plasma is caused to be disposed near a surface to be sterilized, and the cold plasma is then subjected to a pulsed voltage differential for producing energized ions in the plasma. Those energized ions then operate to achieve spore destruction on the surface to be sterilized. Further, a system for sterilization of a container which includes a conductive or non-conductive container, a cold plasma in proximity to the container, and a high voltage source for delivering a pulsed voltage differential between an electrode and the container and across the cold plasma, is provided.

  8. Acton mass flow system applied to PFBC feed

    NASA Technical Reports Server (NTRS)

    Homburg, E.

    1977-01-01

    Dense phase pneumatic conveying and the Acton Mass Flow concept are defined with emphasis on the specific advantages to the coal and dolomite feed to the Pressurized Fluidized Bed Combustor. The transport and feed functions are explored with a comparison of designing the process for a combined function or for individual functions. The equipment required to accomplish these functions is described together with a typical example of sizing and air or gas requirements. A general outline of the control system required to obtain a uniform feed rate is provided. The condition of the coal and dolomite and conveying gas as required to obtain reliable transport and feed will be discussed.

  9. Response deprivation and reinforcement in applied settings: A preliminary analysis

    PubMed Central

    Konarski, Edward A.; Johnson, Moses R.; Crowell, Charles R.; Whitman, Thomas L.

    1980-01-01

    First-grade children engaged in seatwork behaviors under reinforcement schedules established according to the Premack Principle and the Response Deprivation Hypothesis. Across two experiments, schedules were presented to the children in a counter-balanced fashion which fulfilled the conditions of one, both, or neither of the hypotheses. Duration of on-task math and coloring in Experiment 1 and on-task math and reading in Experiment 2 were the dependent variables. A modified ABA-type withdrawal design, including a condition to control for the noncontingent effects of a schedule, indicated an increase of on-task instrumental responding only in those schedules where the condition of response deprivation was present but not where it was absent, regardless of the probability differential between the instrumental and contingent responses. These results were consistent with laboratory findings supporting the necessity of response deprivation for producing the reinforcement effect in single response, instrumental schedules. However, the results of the control procedure were equivocal so the contribution of the contingent relationship between the responses to the increases in instrumental behavior could not be determined. Nevertheless, these results provided tentative support for the Response Deprivation Hypothesis as a new approach to establishing reinforcement schedules while indicating the need for further research in this area. The possible advantages of this technique for applied use were identified and discussed. PMID:16795635

  10. Automated synchrogram analysis applied to heartbeat and reconstructed respiration

    NASA Astrophysics Data System (ADS)

    Hamann, Claudia; Bartsch, Ronny P.; Schumann, Aicko Y.; Penzel, Thomas; Havlin, Shlomo; Kantelhardt, Jan W.

    2009-03-01

    Phase synchronization between two weakly coupled oscillators has been studied in chaotic systems for a long time. However, it is difficult to unambiguously detect such synchronization in experimental data from complex physiological systems. In this paper we review our study of phase synchronization between heartbeat and respiration in 150 healthy subjects during sleep using an automated procedure for screening the synchrograms. We found that this synchronization is significantly enhanced during non-rapid-eye-movement (non-REM) sleep (deep sleep and light sleep) and is reduced during REM sleep. In addition, we show that the respiration signal can be reconstructed from the heartbeat recordings in many subjects. Our reconstruction procedure, which works particularly well during non-REM sleep, allows the detection of cardiorespiratory synchronization even if only heartbeat intervals were recorded.

  11. Applying Real Options for Evaluating Investments in ERP Systems

    NASA Astrophysics Data System (ADS)

    Nakagane, Jun; Sekozawa, Teruji

    This paper intends to verify effectiveness of real options approach for evaluating investments in Enterprise Resource Planning systems (ERP) and proves how important it is to disclose shadow options potentially embedded in ERP investment. The net present value (NPV) method is principally adopted to evaluate the value of ERP. However, the NPV method assumes no uncertainties exist in the object. It doesn't satisfy the current business circumstances which are filled with dynamic issues. Since the 1990s the effectiveness of option pricing models for Information System (IS) investment to solve issues in the NPV method has been discussed in the IS literature. This paper presents 3 business cases to review the practical advantages of such techniques for IS investments, especially ERP investments. The first case is EDI development. We evaluate the project by a new approach with lighting one of shadow options, EDI implementation. In the second case we reveal an ERP investment has an “expanding option” in a case of eliminating redundancy. The third case describes an option to contract which is deliberately slotted in ERP development to prepare transferring a manufacturing facility.

  12. Basic and Applied Aspects of Color Tuning of Bioluminescence Systems

    NASA Astrophysics Data System (ADS)

    Ohmiya, Yoshihiro

    2005-09-01

    V. Viviani et al. [Biochemistry 38 (1999) 8271] were the first to succeed in cloning the red-emitting enzyme from the South American railroad worm, which is the only bioluminescent organism known to emit a red-colored light. The application of red bioluminescence has been our goal because the transmittance of longer-wavelength light is superior to that of the other colors for visualization of biological functions in living cells. Now, different color luciferases, which emit with wavelength maxima ranging from 400 to 630 nm, are available and are being used. For example, based on different color luciferases, Nakajima et al. developed a tricolor reporter in vitro assay system based on these different color luciferases in which the expression of three genes can be monitored simultaneously. On the other hand, bioluminescence resonance energy transfer (BRET) is a natural phenomenon caused by the intermolecular interaction between a bioluminescent protein and a fluorophore on a second protein, resulting in the light from the bioluminescence reaction having the spectrum of the fluorophore. Otsuji et al. [Anal. Biochem. 329 (2004) 230] showed that the change in the efficiency of energy transfer in intramolecular BRET can quantify cellular functions in living cells. In this review, I introduce the basic mechanisms of color tuning in bioluminescent systems and new applications based on color tuning in the life sciences.

  13. An Inverse Kinematic Approach Using Groebner Basis Theory Applied to Gait Cycle Analysis

    DTIC Science & Technology

    2013-03-01

    AN INVERSE KINEMATIC APPROACH USING GROEBNER BASIS THEORY APPLIED TO GAIT CYCLE ANALYSIS THESIS Anum Barki AFIT-ENP-13-M-02 DEPARTMENT OF THE AIR...copyright protection in the United States. AFIT-ENP-13-M-02 AN INVERSE KINEMATIC APPROACH USING GROEBNER BASIS THEORY APPLIED TO GAIT CYCLE ANALYSIS THESIS...APPROACH USING GROEBNER BASIS THEORY APPLIED TO GAIT CYCLE ANALYSIS Anum Barki, BS Approved: Dr. Ronald F. Tuttle (Chairman) Date Dr. Kimberly Kendricks

  14. Information Theory Applied to Animal Communication Systems and Its Possible Application to SETI

    NASA Astrophysics Data System (ADS)

    Hanser, Sean F.; Doyle, Laurance R.; McCowan, Brenda; Jenkins, Jon M.

    2004-06-01

    Information theory, as first introduced by Claude Shannon (Shannon &Weaver 1949) quantitatively evaluates the organizational complexity of communication systems. At the same time George Zipf was examining linguistic structure in a way that was mathematically similar to the components of the Shannon first-order entropy (Zipf 1949). Both Shannon's and Zipf's mathematical procedures have been applied to animal communication and recently have been providing insightful results. The Zipf plot is a useful tool for a first estimate of the characterization of a communication system's complexity (which can later be examined for complex structure at deeper levels using Shannon entropic analysis). In this paper we shall discuss some of the applications and pitfalls of using the Zipf distribution as a preliminary evaluator of the communication complexity of a signaling system.

  15. Applying temporal network analysis to the venture capital market

    NASA Astrophysics Data System (ADS)

    Zhang, Xin; Feng, Ling; Zhu, Rongqian; Stanley, H. Eugene

    2015-10-01

    Using complex network theory to study the investment relationships of venture capital firms has produced a number of significant results. However, previous studies have often neglected the temporal properties of those relationships, which in real-world scenarios play a pivotal role. Here we examine the time-evolving dynamics of venture capital investment in China by constructing temporal networks to represent (i) investment relationships between venture capital firms and portfolio companies and (ii) the syndication ties between venture capital investors. The evolution of the networks exhibits rich variations in centrality, connectivity and local topology. We demonstrate that a temporal network approach provides a dynamic and comprehensive analysis of real-world networks.

  16. Applied analysis/computational mathematics. Final report 1993

    SciTech Connect

    Lax, P.; Berger, M.

    1993-12-01

    This is the final report for the Courant Mathematics and Computing Laboratory (CMCL) research program for the years 1991--1993. Our research efforts encompass the formulation of physical problems in terms of mathematical models (both old and new), the mathematical analysis of such models, and their numerical resolution. This last step involves the development and implementation of efficient methods for large scale computation. Our analytic and numerical work often go hand in hand; new theoretical approaches often have numerical counterparts, while numerical experimentation often suggests avenues for analytical investigation.

  17. Inconsistencies in the harmonic analysis applied to pulsating stars

    NASA Astrophysics Data System (ADS)

    Pascual-Granado, J.; Garrido, R.; Suárez, J. C.

    2015-05-01

    Harmonic analysis is the fundamental mathematical method used for the identification of pulsation frequencies in asteroseismology and other fields of physics. Here we introduce a test to evaluate the validity of the hypothesis in which Fourier theorem is based: the convergence of the expansion series. The huge number of difficulties found in the interpretation of the periodograms of pulsating stars observed by CoRoT and Kepler satellites lead us to test whether the function underlying these time series is analytic or not. Surprisingly, the main result is that these are originated from non-analytic functions, therefore, the condition for Parseval's theorem is not guaranteed.

  18. Absorption and adsorption chillers applied to air conditioning systems

    NASA Astrophysics Data System (ADS)

    Kuczyńska, Agnieszka; Szaflik, Władysław

    2010-07-01

    This work presents an application possibility of sorption refrigerators driven by low temperature fluid for air conditioning of buildings. Thermodynamic models were formulated and absorption LiBr-water chiller with 10 kW cooling power as well as adsorption chiller with silica gel bed were investigated. Both of them are using water for desorption process with temperature Tdes = 80 °C. Coefficient of performance (COP) for both cooling cycles was analyzed in the same conditions of the driving heat source, cooling water Tc = 25 °C and temperature in evaporator Tevap = 5 °C. In this study, the computer software EES was used to investigate the performance of absorption heat pump system and its behaviour in configuration with geothermal heat source.

  19. Multivariate calibration applied to the quantitative analysis of infrared spectra

    SciTech Connect

    Haaland, D.M.

    1991-01-01

    Multivariate calibration methods are very useful for improving the precision, accuracy, and reliability of quantitative spectral analyses. Spectroscopists can more effectively use these sophisticated statistical tools if they have a qualitative understanding of the techniques involved. A qualitative picture of the factor analysis multivariate calibration methods of partial least squares (PLS) and principal component regression (PCR) is presented using infrared calibrations based upon spectra of phosphosilicate glass thin films on silicon wafers. Comparisons of the relative prediction abilities of four different multivariate calibration methods are given based on Monte Carlo simulations of spectral calibration and prediction data. The success of multivariate spectral calibrations is demonstrated for several quantitative infrared studies. The infrared absorption and emission spectra of thin-film dielectrics used in the manufacture of microelectronic devices demonstrate rapid, nondestructive at-line and in-situ analyses using PLS calibrations. Finally, the application of multivariate spectral calibrations to reagentless analysis of blood is presented. We have found that the determination of glucose in whole blood taken from diabetics can be precisely monitored from the PLS calibration of either mind- or near-infrared spectra of the blood. Progress toward the non-invasive determination of glucose levels in diabetics is an ultimate goal of this research. 13 refs., 4 figs.

  20. [Failure mode effect analysis applied to preparation of intravenous cytostatics].

    PubMed

    Santos-Rubio, M D; Marín-Gil, R; Muñoz-de la Corte, R; Velázquez-López, M D; Gil-Navarro, M V; Bautista-Paloma, F J

    2016-01-01

    To proactively identify risks in the preparation of intravenous cytostatic drugs, and to prioritise and establish measures to improve safety procedures. Failure Mode Effect Analysis methodology was used. A multidisciplinary team identified potential failure modes of the procedure through a brainstorming session. The impact associated with each failure mode was assessed with the Risk Priority Number (RPN), which involves three variables: occurrence, severity, and detectability. Improvement measures were established for all identified failure modes, with those with RPN>100 considered critical. The final RPN (theoretical) that would result from the proposed measures was also calculated and the process was redesigned. A total of 34 failure modes were identified. The initial accumulated RPN was 3022 (range: 3-252), and after recommended actions the final RPN was 1292 (range: 3-189). RPN scores >100 were obtained in 13 failure modes; only the dispensing sub-process was free of critical points (RPN>100). A final reduction of RPN>50% was achieved in 9 failure modes. This prospective risk analysis methodology allows the weaknesses of the procedure to be prioritised, optimize use of resources, and a substantial improvement in the safety of the preparation of cytostatic drugs through the introduction of double checking and intermediate product labelling. Copyright © 2015 SECA. Published by Elsevier Espana. All rights reserved.

  1. Unsupervised feature relevance analysis applied to improve ECG heartbeat clustering.

    PubMed

    Rodríguez-Sotelo, J L; Peluffo-Ordoñez, D; Cuesta-Frau, D; Castellanos-Domínguez, G

    2012-10-01

    The computer-assisted analysis of biomedical records has become an essential tool in clinical settings. However, current devices provide a growing amount of data that often exceeds the processing capacity of normal computers. As this amount of information rises, new demands for more efficient data extracting methods appear. This paper addresses the task of data mining in physiological records using a feature selection scheme. An unsupervised method based on relevance analysis is described. This scheme uses a least-squares optimization of the input feature matrix in a single iteration. The output of the algorithm is a feature weighting vector. The performance of the method was assessed using a heartbeat clustering test on real ECG records. The quantitative cluster validity measures yielded a correctly classified heartbeat rate of 98.69% (specificity), 85.88% (sensitivity) and 95.04% (general clustering performance), which is even higher than the performance achieved by other similar ECG clustering studies. The number of features was reduced on average from 100 to 18, and the temporal cost was a 43% lower than in previous ECG clustering schemes.

  2. Neptune Aerocapture Systems Analysis

    NASA Technical Reports Server (NTRS)

    Lockwood, Mary Kae

    2004-01-01

    A Neptune Aerocapture Systems Analysis is completed to determine the feasibility, benefit and risk of an aeroshell aerocapture system for Neptune and to identify technology gaps and technology performance goals. The high fidelity systems analysis is completed by a five center NASA team and includes the following disciplines and analyses: science; mission design; aeroshell configuration screening and definition; interplanetary navigation analyses; atmosphere modeling; computational fluid dynamics for aerodynamic performance and database definition; initial stability analyses; guidance development; atmospheric flight simulation; computational fluid dynamics and radiation analyses for aeroheating environment definition; thermal protection system design, concepts and sizing; mass properties; structures; spacecraft design and packaging; and mass sensitivities. Results show that aerocapture can deliver 1.4 times more mass to Neptune orbit than an all-propulsive system for the same launch vehicle. In addition aerocapture results in a 3-4 year reduction in trip time compared to all-propulsive systems. Aerocapture is feasible and performance is adequate for the Neptune aerocapture mission. Monte Carlo simulation results show 100% successful capture for all cases including conservative assumptions on atmosphere and navigation. Enabling technologies for this mission include TPS manufacturing; and aerothermodynamic methods and validation for determining coupled 3-D convection, radiation and ablation aeroheating rates and loads, and the effects on surface recession.

  3. On applying continuous wavelet transform in wheeze analysis.

    PubMed

    Taplidou, Styliani A; Hadjileontiadis, Leontios J; Kitsas, Ilias K; Panoulas, Konstantinos I; Penzel, Thomas; Gross, Volker; Panas, Stavros M

    2004-01-01

    The identification of continuous abnormal lung sounds, like wheezes, in the total breathing cycle is of great importance in the diagnosis of obstructive airways pathologies. To this vein, the current work introduces an efficient method for the detection of wheezes, based on the time-scale representation of breath sound recordings. The employed Continuous Wavelet Transform is proven to be a valuable tool at this direction, when combined with scale-dependent thresholding. Analysis of lung sound recordings from 'wheezing' patients shows promising performance in the detection and extraction of wheezes from the background noise and reveals its potentiality for data-volume reduction in long-term wheezing screening, such as in sleep-laboratories.

  4. Downside Risk analysis applied to the Hedge Funds universe

    NASA Astrophysics Data System (ADS)

    Perelló, Josep

    2007-09-01

    Hedge Funds are considered as one of the portfolio management sectors which shows a fastest growing for the past decade. An optimal Hedge Fund management requires an appropriate risk metrics. The classic CAPM theory and its Ratio Sharpe fail to capture some crucial aspects due to the strong non-Gaussian character of Hedge Funds statistics. A possible way out to this problem while keeping the CAPM simplicity is the so-called Downside Risk analysis. One important benefit lies in distinguishing between good and bad returns, that is: returns greater or lower than investor's goal. We revisit most popular Downside Risk indicators and provide new analytical results on them. We compute these measures by taking the Credit Suisse/Tremont Investable Hedge Fund Index Data and with the Gaussian case as a benchmark. In this way, an unusual transversal lecture of the existing Downside Risk measures is provided.

  5. Applying Machine Learning to GlueX Data Analysis

    NASA Astrophysics Data System (ADS)

    Boettcher, Thomas

    2014-03-01

    GlueX is a high energy physics experiment with the goal of collecting data necessary for understanding confinement in quantum chromodynamics. Beginning in 2015, GlueX will collect huge amounts of data describing billions of particle collisions. In preparation for data collection, efforts are underway to develop a methodology for analyzing these large data sets. One of the primary challenges in GlueX data analysis is isolating events of interest from a proportionally large background. GlueX has recently begun approaching this selection problem using machine learning algorithms, specifically boosted decision trees. Preliminary studies indicate that these algorithms have the potential to offer vast improvements in both signal selection efficiency and purity over more traditional techniques.

  6. Applying Skinner's analysis of verbal behavior to persons with dementia.

    PubMed

    Dixon, Mark; Baker, Jonathan C; Sadowski, Katherine Ann

    2011-03-01

    Skinner's 1957 analysis of verbal behavior has demonstrated a fair amount of utility to teach language to children with autism and other various disorders. However, the learning of language can be forgotten, as is the case for many elderly suffering from dementia or other degenerative diseases. It appears possible that Skinner's operants may facilitate not only acquisition of language but also the ability to recall items or objects that may have appeared to be "forgotten." The present study examined the utility of having a series of adults in long-term care emit tacts, echoics, or intraverbals upon presentation of various visual stimuli. Compared to a no-verbal response condition, it appears that the incorporation of Skinner's verbal operants can in fact improve recall for this population. Implications for the retraining of lost language are presented.

  7. Geostatistical analysis as applied to two environmental radiometric time series.

    PubMed

    Dowdall, Mark; Lind, Bjørn; Gerland, Sebastian; Rudjord, Anne Liv

    2003-03-01

    This article details the results of an investigation into the application of geostatistical data analysis to two environmental radiometric time series. The data series employed consist of 99Tc values for seaweed (Fucus vesiculosus) and seawater samples taken as part of a marine monitoring program conducted on the coast of northern Norway by the Norwegian Radiation Protection Authority. Geostatistical methods were selected in order to provide information on values of the variables at unsampled times and to investigate the temporal correlation exhibited by the data sets. This information is of use in the optimisation of future sampling schemes and for providing information on the temporal behaviour of the variables in question that may not be obtained during a cursory analysis. The results indicate a high degree of temporal correlation within the data sets, the correlation for the seawater and seaweed data being modelled with an exponential and linear function, respectively. The semi-variogram for the seawater data indicates a temporal range of correlation of approximately 395 days with no apparent random component to the overall variance structure and was described best by an exponential function. The temporal structure of the seaweed data was best modelled by a linear function with a small nugget component. Evidence of drift was present in both semi-variograms. Interpolation of the data sets using the fitted models and a simple kriging procedure were compared, using a cross-validation procedure, with simple linear interpolation. Results of this exercise indicate that, for the seawater data, the kriging procedure outperformed the simple interpolation with respect to error distribution and correlation of estimates with actual values. Using the unbounded linear model with the seaweed data produced estimates that were only marginally better than those produced by the simple interpolation.

  8. Can Link Analysis Be Applied to Identify Behavioral Patterns in Train Recorder Data?

    PubMed

    Strathie, Ailsa; Walker, Guy H

    2016-03-01

    A proof-of-concept analysis was conducted to establish whether link analysis could be applied to data from on-train recorders to detect patterns of behavior that could act as leading indicators of potential safety issues. On-train data recorders capture data about driving behavior on thousands of routine journeys every day and offer a source of untapped data that could be used to offer insights into human behavior. Data from 17 journeys undertaken by six drivers on the same route over a 16-hr period were analyzed using link analysis, and four key metrics were examined: number of links, network density, diameter, and sociometric status. The results established that link analysis can be usefully applied to data captured from on-vehicle recorders. The four metrics revealed key differences in normal driver behavior. These differences have promising construct validity as leading indicators. Link analysis is one method that could be usefully applied to exploit data routinely gathered by on-vehicle data recorders. It facilitates a proactive approach to safety based on leading indicators, offers a clearer understanding of what constitutes normal driving behavior, and identifies trends at the interface of people and systems, which is currently a key area of strategic risk. These research findings have direct applications in the field of transport data monitoring. They offer a means of automatically detecting patterns in driver behavior that could act as leading indicators of problems during operation and that could be used in the proactive monitoring of driver competence, risk management, and even infrastructure design. © 2015, Human Factors and Ergonomics Society.

  9. Speckle interferometry applied to asteroids and other solar system objects

    NASA Technical Reports Server (NTRS)

    Drummond, J. D.; Hege, E. K.

    1986-01-01

    Speckle interferometry is a high angular resolution technique that allows study of resolved asteroids. By following the changing size, shape, and orientation of minor planets, and with a few general assumptions (e.g., geometric scattering, triaxial ellipsoid figures, no albedo features), it is possible to directly measure an asteroid's true dimensions and the direction of its spin axis in one or two nights. A particular subset of triaxial ellipsoid figures are equilibrium shapes, and would imply that some asteroids are thoroughly fractured. Such shapes if they exist among the asteroids would allow a determination of bulk density since there is a unique relation among spin period, size, shape, and density. The discovery of even a single rubble pile, (just as the finding of even one binary asteroid by speckle interferometric techniques) would drastically alter the notion of asteroids as small solid planets. The Pluto/Charon system was studied to aid in improving the orbital elements necessary to predict the eclipse/occultation season currently in progress. Four asteroids were reduced to their size, shape, and pole direction: 433 Eros, 532 Herculina, 511 Davida, and 2 Pallas.

  10. Automated document analysis system

    NASA Astrophysics Data System (ADS)

    Black, Jeffrey D.; Dietzel, Robert; Hartnett, David

    2002-08-01

    A software application has been developed to aid law enforcement and government intelligence gathering organizations in the translation and analysis of foreign language documents with potential intelligence content. The Automated Document Analysis System (ADAS) provides the capability to search (data or text mine) documents in English and the most commonly encountered foreign languages, including Arabic. Hardcopy documents are scanned by a high-speed scanner and are optical character recognized (OCR). Documents obtained in an electronic format bypass the OCR and are copied directly to a working directory. For translation and analysis, the script and the language of the documents are first determined. If the document is not in English, the document is machine translated to English. The documents are searched for keywords and key features in either the native language or translated English. The user can quickly review the document to determine if it has any intelligence content and whether detailed, verbatim human translation is required. The documents and document content are cataloged for potential future analysis. The system allows non-linguists to evaluate foreign language documents and allows for the quick analysis of a large quantity of documents. All document processing can be performed manually or automatically on a single document or a batch of documents.

  11. The impact of applied behavior analysis on diverse areas of research.

    PubMed Central

    Kazdin, A E

    1975-01-01

    The impact of applied behavior analysis on various disciplines and areas of research was assessed through two major analyses. First, the relationship of applied behavior analysis to the general area of "behavior modification" was evaluated by examining the citation characteristics of journal articles in JABA and three other behavior-modification journals. Second, the penetration of applied behavior analysis into diverse areas and disciplines, including behavior modification, psychiatry, clinical psychology, education, special education, retardation, speech and hearing, counselling, and law enforcement and correction was assessed. Twenty-five journals representing diverse research areas were evaluated from 1968 to 1974 to assess the extent to which operant techniques were applied for therapeutic, rehabilitative, and educative purposes and the degree to which methodological desiderata of applied behavior analysis were met. The analyses revealed diverse publication outlets for applied behavior analysis in various disciplines. PMID:1184488

  12. A System Analysis Tool

    SciTech Connect

    CAMPBELL,PHILIP L.; ESPINOZA,JUAN

    2000-06-01

    In this paper we describe a tool for analyzing systems. The analysis is based on program slicing. It answers the following question for the software: if the value of a particular variable changes, what other variable values also change, and what is the path in between? program slicing was developed based on intra-procedure control and data flow. It has been expanded commercially to inter-procedure flow. However, we extend slicing to collections of programs and non-program entities, which we term multi-domain systems. The value of our tool is that an analyst can model the entirety of a system, not just the software, and we believe that this makes for a significant increase in power. We are building a prototype system.

  13. Network systems security analysis

    NASA Astrophysics Data System (ADS)

    Yilmaz, Ä.°smail

    2015-05-01

    Network Systems Security Analysis has utmost importance in today's world. Many companies, like banks which give priority to data management, test their own data security systems with "Penetration Tests" by time to time. In this context, companies must also test their own network/server systems and take precautions, as the data security draws attention. Based on this idea, the study cyber-attacks are researched throughoutly and Penetration Test technics are examined. With these information on, classification is made for the cyber-attacks and later network systems' security is tested systematically. After the testing period, all data is reported and filed for future reference. Consequently, it is found out that human beings are the weakest circle of the chain and simple mistakes may unintentionally cause huge problems. Thus, it is clear that some precautions must be taken to avoid such threats like updating the security software.

  14. Aerodynamic Reconstruction Applied to Parachute Test Vehicle Flight Data Analysis

    NASA Technical Reports Server (NTRS)

    Cassady, Leonard D.; Ray, Eric S.; Truong, Tuan H.

    2013-01-01

    The aerodynamics, both static and dynamic, of a test vehicle are critical to determining the performance of the parachute cluster in a drop test and for conducting a successful test. The Capsule Parachute Assembly System (CPAS) project is conducting tests of NASA's Orion Multi-Purpose Crew Vehicle (MPCV) parachutes at the Army Yuma Proving Ground utilizing the Parachute Test Vehicle (PTV). The PTV shape is based on the MPCV, but the height has been reduced in order to fit within the C-17 aircraft for extraction. Therefore, the aerodynamics of the PTV are similar, but not the same as, the MPCV. A small series of wind tunnel tests and computational fluid dynamics cases were run to modify the MPCV aerodynamic database for the PTV, but aerodynamic reconstruction of the flights has proven an effective source for further improvements to the database. The acceleration and rotational rates measured during free flight, before parachute inflation but during deployment, were used to con rm vehicle static aerodynamics. A multibody simulation is utilized to reconstruct the parachute portions of the flight. Aerodynamic or parachute parameters are adjusted in the simulation until the prediction reasonably matches the flight trajectory. Knowledge of the static aerodynamics is critical in the CPAS project because the parachute riser load measurements are scaled based on forebody drag. PTV dynamic damping is critical because the vehicle has no reaction control system to maintain attitude - the vehicle dynamics must be understood and modeled correctly before flight. It will be shown here that aerodynamic reconstruction has successfully contributed to the CPAS project.

  15. Applying axiomatic design to a medication distribution system

    NASA Astrophysics Data System (ADS)

    Raguini, Pepito B.

    As the need to minimize medication errors drives many medical facilities to come up with robust solutions to the most common error that affects patient's safety, these hospitals would be wise to put a concerted effort into finding methodologies that can facilitate an optimized medical distribution system. If the hospitals' upper management is looking for an optimization method that is an ideal fit, it is just as important that the right tool be selected for the application at hand. In the present work, we propose the application of Axiomatic Design (AD), which is a process that focuses on the generation and selection of functional requirements to meet the customer needs for product and/or process design. The appeal of the axiomatic approach is to provide both a formal design process and a set of technical coefficients for meeting the customer's needs. Thus, AD offers a strategy for the effective integration of people, design methods, design tools and design data. Therefore, we propose the AD methodology to medical applications with the main objective of allowing nurses the opportunity to provide cost effective delivery of medications to inpatients, thereby improving quality patient care. The AD methodology will be implemented through the use of focused stores, where medications can be readily stored and can be conveniently located near patients, as well as a mobile apparatus that can also store medications and is commonly used by hospitals, the medication cart. Moreover, a robust methodology called the focused store methodology will be introduced and developed for both the uncapacitated and capacitated case studies, which will set up an appropriate AD framework and design problem for a medication distribution case study.

  16. Applying data mining for the analysis of breast cancer data.

    PubMed

    Liou, Der-Ming; Chang, Wei-Pin

    2015-01-01

    Data mining, also known as Knowledge-Discovery in Databases (KDD), is the process of automatically searching large volumes of data for patterns. For instance, a clinical pattern might indicate a female who have diabetes or hypertension are easier suffered from stroke for 5 years in a future. Then, a physician can learn valuable knowledge from the data mining processes. Here, we present a study focused on the investigation of the application of artificial intelligence and data mining techniques to the prediction models of breast cancer. The artificial neural network, decision tree, logistic regression, and genetic algorithm were used for the comparative studies and the accuracy and positive predictive value of each algorithm were used as the evaluation indicators. 699 records acquired from the breast cancer patients at the University of Wisconsin, nine predictor variables, and one outcome variable were incorporated for the data analysis followed by the tenfold cross-validation. The results revealed that the accuracies of logistic regression model were 0.9434 (sensitivity 0.9716 and specificity 0.9482), the decision tree model 0.9434 (sensitivity 0.9615, specificity 0.9105), the neural network model 0.9502 (sensitivity 0.9628, specificity 0.9273), and the genetic algorithm model 0.9878 (sensitivity 1, specificity 0.9802). The accuracy of the genetic algorithm was significantly higher than the average predicted accuracy of 0.9612. The predicted outcome of the logistic regression model was higher than that of the neural network model but no significant difference was observed. The average predicted accuracy of the decision tree model was 0.9435 which was the lowest of all four predictive models. The standard deviation of the tenfold cross-validation was rather unreliable. This study indicated that the genetic algorithm model yielded better results than other data mining models for the analysis of the data of breast cancer patients in terms of the overall accuracy of

  17. Applied Climate-Change Analysis: The Climate Wizard Tool

    PubMed Central

    Girvetz, Evan H.; Zganjar, Chris; Raber, George T.; Maurer, Edwin P.; Kareiva, Peter; Lawler, Joshua J.

    2009-01-01

    Background Although the message of “global climate change” is catalyzing international action, it is local and regional changes that directly affect people and ecosystems and are of immediate concern to scientists, managers, and policy makers. A major barrier preventing informed climate-change adaptation planning is the difficulty accessing, analyzing, and interpreting climate-change information. To address this problem, we developed a powerful, yet easy to use, web-based tool called Climate Wizard (http://ClimateWizard.org) that provides non-climate specialists with simple analyses and innovative graphical depictions for conveying how climate has and is projected to change within specific geographic areas throughout the world. Methodology/Principal Findings To demonstrate the Climate Wizard, we explored historic trends and future departures (anomalies) in temperature and precipitation globally, and within specific latitudinal zones and countries. We found the greatest temperature increases during 1951–2002 occurred in northern hemisphere countries (especially during January–April), but the latitude of greatest temperature change varied throughout the year, sinusoidally ranging from approximately 50°N during February-March to 10°N during August-September. Precipitation decreases occurred most commonly in countries between 0–20°N, and increases mostly occurred outside of this latitudinal region. Similarly, a quantile ensemble analysis based on projections from 16 General Circulation Models (GCMs) for 2070–2099 identified the median projected change within countries, which showed both latitudinal and regional patterns in projected temperature and precipitation change. Conclusions/Significance The results of these analyses are consistent with those reported by the Intergovernmental Panel on Climate Change, but at the same time, they provide examples of how Climate Wizard can be used to explore regionally- and temporally-specific analyses of climate

  18. Applied and computational harmonic analysis on graphs and networks

    NASA Astrophysics Data System (ADS)

    Irion, Jeff; Saito, Naoki

    2015-09-01

    In recent years, the advent of new sensor technologies and social network infrastructure has provided huge opportunities and challenges for analyzing data recorded on such networks. In the case of data on regular lattices, computational harmonic analysis tools such as the Fourier and wavelet transforms have well-developed theories and proven track records of success. It is therefore quite important to extend such tools from the classical setting of regular lattices to the more general setting of graphs and networks. In this article, we first review basics of graph Laplacian matrices, whose eigenpairs are often interpreted as the frequencies and the Fourier basis vectors on a given graph. We point out, however, that such an interpretation is misleading unless the underlying graph is either an unweighted path or cycle. We then discuss our recent effort of constructing multiscale basis dictionaries on a graph, including the Hierarchical Graph Laplacian Eigenbasis Dictionary and the Generalized Haar-Walsh Wavelet Packet Dictionary, which are viewed as generalizations of the classical hierarchical block DCTs and the Haar-Walsh wavelet packets, respectively, to the graph setting. Finally, we demonstrate the usefulness of our dictionaries by using them to simultaneously segment and denoise 1-D noisy signals sampled on regular lattices, a problem where classical tools have difficulty.

  19. Ion Beam Analysis applied to laser-generated plasmas

    NASA Astrophysics Data System (ADS)

    Cutroneo, M.; Macková, A.; Havranek, V.; Malinsky, P.; Torrisi, L.; Kormunda, M.; Barchuk, M.; Ullschmied, J.; Dudzak, R.

    2016-04-01

    This paper presents the research activity on Ion Beam Analysis methods performed at Tandetron Laboratory (LT) of the Institute of Nuclear Physics AS CR, Rez, Czech Republic. Recently, many groups are paying attention to implantation by laser generated plasma. This process allows to insert a controllable amount of energetic ions into the surface layers of different materials modifying the physical and chemical properties of the surface material. Different substrates are implanted by accelerated ions from plasma through terawatt iodine laser, at nominal intensity of 1015 W/cm2, at the PALS Research Infrastructure AS CR, in the Czech Republic. This regime of the laser matter interaction generates, multi-MeV proton beams, and multi-charged ions that are tightly confined in time (hundreds ps) and space (source radius of a few microns). These ion beams have a much lower transverse temperature, a much shorter duration and a much higher current than those obtainable from conventional accelerators. The implementation of protons and ions acceleration driven by ultra-short high intensity lasers is exhibited by adopting suitable irradiation conditions as well as tailored targets. An overview of implanted targets and their morphological and structural characterizations is presented and discussed.

  20. Applying revised gap analysis model in measuring hotel service quality.

    PubMed

    Lee, Yu-Cheng; Wang, Yu-Che; Chien, Chih-Hung; Wu, Chia-Huei; Lu, Shu-Chiung; Tsai, Sang-Bing; Dong, Weiwei

    2016-01-01

    With the number of tourists coming to Taiwan growing by 10-20 % since 2010, the number has increased due to an increasing number of foreign tourists, particularly after deregulation allowed admitting tourist groups, followed later on by foreign individual tourists, from mainland China. The purpose of this study is to propose a revised gap model to evaluate and improve service quality in Taiwanese hotel industry. Thus, service quality could be clearly measured through gap analysis, which was more effective for offering direction in developing and improving service quality. The HOLSERV instrument was used to identify and analyze service gaps from the perceptions of internal and external customers. The sample for this study included three main categories of respondents: tourists, employees, and managers. The results show that five gaps influenced tourists' evaluations of service quality. In particular, the study revealed that Gap 1 (management perceptions vs. customer expectations) and Gap 9 (service provider perceptions of management perceptions vs. service delivery) were more critical than the others in affecting perceived service quality, making service delivery the main area of improvement. This study contributes toward an evaluation of the service quality of the Taiwanese hotel industry from the perspectives of customers, service providers, and managers, which is considerably valuable for hotel managers. It was the aim of this study to explore all of these together in order to better understand the possible gaps in the hotel industry in Taiwan.

  1. Applied Drama and the Higher Education Learning Spaces: A Reflective Analysis

    ERIC Educational Resources Information Center

    Moyo, Cletus

    2015-01-01

    This paper explores Applied Drama as a teaching approach in Higher Education learning spaces. The exploration takes a reflective analysis approach by first examining the impact that Applied Drama has had on my career as a Lecturer/Educator/Teacher working in Higher Education environments. My engagement with Applied Drama practice and theory is…

  2. Exploratory Factor Analysis as a Construct Validation Tool: (Mis)applications in Applied Linguistics Research

    ERIC Educational Resources Information Center

    Karami, Hossein

    2015-01-01

    Factor analysis has been frequently exploited in applied research to provide evidence about the underlying factors in various measurement instruments. A close inspection of a large number of studies published in leading applied linguistic journals shows that there is a misconception among applied linguists as to the relative merits of exploratory…

  3. Systems analysis of bone.

    PubMed

    Jepsen, Karl J

    2009-01-01

    The genetic variants contributing to variability in skeletal traits has been well studied, and several hundred QTLs have been mapped and several genes contributing to trait variation have been identified. However, many questions remain unanswered. In particular, it is unclear whether variation in a single gene leads to alterations in function. Bone is a highly adaptive system and genetic variants affecting one trait are often accompanied by compensatory changes in other traits. The functional interactions among traits, which is known as phenotypic integration, has been observed in many biological systems, including bone. Phenotypic integration is a property of bone that is critically important for establishing a mechanically functional structure that is capable of supporting the forces imparted during daily activities. In this paper, bone is reviewed as a system and primarily in the context of functionality. A better understanding of the system properties of bone will lead to novel targets for future genetic analyses and the identification of genes that are directly responsible for regulating bone strength. This systems analysis has the added benefit of leaving a trail of valuable information about how the skeletal system works. This information will provide novel approaches to assessing skeletal health during growth and aging and for developing novel treatment strategies to reduce the morbidity and mortality associated with fragility fractures.

  4. Systems thinking applied to safety during manual handling tasks in the transport and storage industry.

    PubMed

    Goode, Natassia; Salmon, Paul M; Lenné, Michael G; Hillard, Peter

    2014-07-01

    Injuries resulting from manual handling tasks represent an on-going problem for the transport and storage industry. This article describes an application of a systems theory-based approach, Rasmussen's (1997. Safety Science 27, 183), risk management framework, to the analysis of the factors influencing safety during manual handling activities in a freight handling organisation. Observations of manual handling activities, cognitive decision method interviews with workers (n=27) and interviews with managers (n=35) were used to gather information about three manual handling activities. Hierarchical task analysis and thematic analysis were used to identify potential risk factors and performance shaping factors across the levels of Rasmussen's framework. These different data sources were then integrated using Rasmussen's Accimap technique to provide an overall analysis of the factors influencing safety during manual handling activities in this context. The findings demonstrate how a systems theory-based approach can be applied to this domain, and suggest that policy-orientated, rather than worker-orientated, changes are required to prevent future manual handling injuries.

  5. A novel bi-level meta-analysis approach: applied to biological pathway analysis.

    PubMed

    Nguyen, Tin; Tagett, Rebecca; Donato, Michele; Mitrea, Cristina; Draghici, Sorin

    2016-02-01

    The accumulation of high-throughput data in public repositories creates a pressing need for integrative analysis of multiple datasets from independent experiments. However, study heterogeneity, study bias, outliers and the lack of power of available methods present real challenge in integrating genomic data. One practical drawback of many P-value-based meta-analysis methods, including Fisher's, Stouffer's, minP and maxP, is that they are sensitive to outliers. Another drawback is that, because they perform just one statistical test for each individual experiment, they may not fully exploit the potentially large number of samples within each study. We propose a novel bi-level meta-analysis approach that employs the additive method and the Central Limit Theorem within each individual experiment and also across multiple experiments. We prove that the bi-level framework is robust against bias, less sensitive to outliers than other methods, and more sensitive to small changes in signal. For comparative analysis, we demonstrate that the intra-experiment analysis has more power than the equivalent statistical test performed on a single large experiment. For pathway analysis, we compare the proposed framework versus classical meta-analysis approaches (Fisher's, Stouffer's and the additive method) as well as against a dedicated pathway meta-analysis package (MetaPath), using 1252 samples from 21 datasets related to three human diseases, acute myeloid leukemia (9 datasets), type II diabetes (5 datasets) and Alzheimer's disease (7 datasets). Our framework outperforms its competitors to correctly identify pathways relevant to the phenotypes. The framework is sufficiently general to be applied to any type of statistical meta-analysis. The R scripts are available on demand from the authors. sorin@wayne.edu Supplementary data are available at Bioinformatics online. © The Author 2015. Published by Oxford University Press. All rights reserved. For Permissions, please e

  6. A novel bi-level meta-analysis approach: applied to biological pathway analysis

    PubMed Central

    Nguyen, Tin; Tagett, Rebecca; Donato, Michele; Mitrea, Cristina; Draghici, Sorin

    2016-01-01

    Motivation: The accumulation of high-throughput data in public repositories creates a pressing need for integrative analysis of multiple datasets from independent experiments. However, study heterogeneity, study bias, outliers and the lack of power of available methods present real challenge in integrating genomic data. One practical drawback of many P-value-based meta-analysis methods, including Fisher’s, Stouffer’s, minP and maxP, is that they are sensitive to outliers. Another drawback is that, because they perform just one statistical test for each individual experiment, they may not fully exploit the potentially large number of samples within each study. Results: We propose a novel bi-level meta-analysis approach that employs the additive method and the Central Limit Theorem within each individual experiment and also across multiple experiments. We prove that the bi-level framework is robust against bias, less sensitive to outliers than other methods, and more sensitive to small changes in signal. For comparative analysis, we demonstrate that the intra-experiment analysis has more power than the equivalent statistical test performed on a single large experiment. For pathway analysis, we compare the proposed framework versus classical meta-analysis approaches (Fisher’s, Stouffer’s and the additive method) as well as against a dedicated pathway meta-analysis package (MetaPath), using 1252 samples from 21 datasets related to three human diseases, acute myeloid leukemia (9 datasets), type II diabetes (5 datasets) and Alzheimer’s disease (7 datasets). Our framework outperforms its competitors to correctly identify pathways relevant to the phenotypes. The framework is sufficiently general to be applied to any type of statistical meta-analysis. Availability and implementation: The R scripts are available on demand from the authors. Contact: sorin@wayne.edu Supplementary Information: Supplementary data are available at Bioinformatics online. PMID:26471455

  7. Versatile microanalytical system with porous polypropylene capillary membrane for calibration gas generation and trace gaseous pollutants sampling applied to the analysis of formaldehyde, formic acid, acetic acid and ammonia in outdoor air.

    PubMed

    Coelho, Lúcia H G; Melchert, Wanessa R; Rocha, Flavio R; Rocha, Fábio R P; Gutz, Ivano G R

    2010-11-15

    The analytical determination of atmospheric pollutants still presents challenges due to the low-level concentrations (frequently in the μg m(-3) range) and their variations with sampling site and time. In this work, a capillary membrane diffusion scrubber (CMDS) was scaled down to match with capillary electrophoresis (CE), a quick separation technique that requires nothing more than some nanoliters of sample and, when combined with capacitively coupled contactless conductometric detection (C(4)D), is particularly favorable for ionic species that do not absorb in the UV-vis region, like the target analytes formaldehyde, formic acid, acetic acid and ammonium. The CMDS was coaxially assembled inside a PTFE tube and fed with acceptor phase (deionized water for species with a high Henry's constant such as formaldehyde and carboxylic acids, or acidic solution for ammonia sampling with equilibrium displacement to the non-volatile ammonium ion) at a low flow rate (8.3 nL s(-1)), while the sample was aspirated through the annular gap of the concentric tubes at 2.5 mL s(-1). A second unit, in all similar to the CMDS, was operated as a capillary membrane diffusion emitter (CMDE), generating a gas flow with know concentrations of ammonia for the evaluation of the CMDS. The fluids of the system were driven with inexpensive aquarium air pumps, and the collected samples were stored in vials cooled by a Peltier element. Complete protocols were developed for the analysis, in air, of NH(3), CH(3)COOH, HCOOH and, with a derivatization setup, CH(2)O, by associating the CMDS collection with the determination by CE-C(4)D. The ammonia concentrations obtained by electrophoresis were checked against the reference spectrophotometric method based on Berthelot's reaction. Sensitivity enhancements of this reference method were achieved by using a modified Berthelot reaction, solenoid micro-pumps for liquid propulsion and a long optical path cell based on a liquid core waveguide (LCW). All

  8. Beyond Time Out and Table Time: Today's Applied Behavior Analysis for Students with Autism

    ERIC Educational Resources Information Center

    Boutot, E. Amanda; Hume, Kara

    2010-01-01

    Recent mandates related to the implementation of evidence-based practices for individuals with autism spectrum disorder (ASD) require that autism professionals both understand and are able to implement practices based on the science of applied behavior analysis (ABA). The use of the term "applied behavior analysis" and its related concepts…

  9. Beyond Time out and Table Time: Today's Applied Behavior Analysis for Students with Autism

    ERIC Educational Resources Information Center

    Boutot, E. Amanda; Hume, Kara

    2012-01-01

    Recent mandates related to the implementation of evidence-based practices for individuals with autism spectrum disorder (ASD) require that autism professionals both understand and are able to implement practices based on the science of applied behavior analysis (ABA). The use of the term "applied behavior analysis" and its related concepts…

  10. Distinguishing Pattern Formation Phenotypes: Applying Minkowski Functionals to Cell Biology Systems

    NASA Astrophysics Data System (ADS)

    Rericha, Erin; Guven, Can; Parent, Carole; Losert, Wolfgang

    2011-03-01

    Spatial Clustering of proteins within cells or cells themselves frequently occur in cell biology systems. However quantifying the underlying order and determining the regulators of these cluster patterns have proved difficult due to the inherent high noise levels in the systems. For instance the patterns formed by wild type and cyclic-AMP regulatory mutant Dictyostelium cells are visually distinctive, yet the large error bars in measurements of the fractal number, area, Euler number, eccentricity, and wavelength making it difficult to quantitatively distinguish between the patterns. We apply a spatial analysis technique based on Minkowski functionals and develop metrics which clearly separate wild type and mutant cell lines into distinct categories. Having such a metric facilitated the development of a computational model for cellular aggregation and its regulators. Supported by NIH-NGHS Nanotechnology (R01GM085574) and the Burroughs Wellcome Fund.

  11. Dental arch analysis system.

    PubMed

    Mutinelli, Sabrina; Cozzani, Mauro; Manfredi, Mario; Siciliani, Giuseppe

    2004-01-01

    The purpose of our research is to describe a computerized method, which enables orthodontists and researchers to analyze variation of dental arch form and dimension. The analysis system is composed of two independent parts: the database, where the images of scanned dental casts are stored, and the software. The operator uses the software to identify some landmarks on dental cast images. Corresponding distances are calculated (interincisive, intercanine and intermolar widths and the curve axis). Then the software algorithms calculate and draw the curves passing trough the selected landmarks (conics, catenary, cubic spline and polynomial curves). In the chosen curves the arch length is measured. The dental cast data, recorded at different times of the subject's life or in different steps of orthodontic treatment, are compared in order to evaluate the change in dimension (arch length and width) and form (shape). The statistical analysis of the data evaluates the variation in form and in dimension separately. The shape change is defined by Euclidean Distance Matrix Analysis (EDMA) and Bootstrap analysis. The computerized method allows orthodontists and researches to evaluate the variation in dimension and form of the dental arch in un-treated subjects and-or orthodontic patients over time. After the analysis of a large sample of patients or subjects, information concerning changes in dental arch dimension and form can be added to previous studies of other authors, distinguishing variation during and after orthodontic treatment, during growth or aging.

  12. Applied statistical training to strengthen analysis and health research capacity in Rwanda.

    PubMed

    Thomson, Dana R; Semakula, Muhammed; Hirschhorn, Lisa R; Murray, Megan; Ndahindwa, Vedaste; Manzi, Anatole; Mukabutera, Assumpta; Karema, Corine; Condo, Jeanine; Hedt-Gauthier, Bethany

    2016-09-29

    To guide efficient investment of limited health resources in sub-Saharan Africa, local researchers need to be involved in, and guide, health system and policy research. While extensive survey and census data are available to health researchers and program officers in resource-limited countries, local involvement and leadership in research is limited due to inadequate experience, lack of dedicated research time and weak interagency connections, among other challenges. Many research-strengthening initiatives host prolonged fellowships out-of-country, yet their approaches have not been evaluated for effectiveness in involvement and development of local leadership in research. We developed, implemented and evaluated a multi-month, deliverable-driven, survey analysis training based in Rwanda to strengthen skills of five local research leaders, 15 statisticians, and a PhD candidate. Research leaders applied with a specific research question relevant to country challenges and committed to leading an analysis to publication. Statisticians with prerequisite statistical training and experience with a statistical software applied to participate in class-based trainings and complete an assigned analysis. Both statisticians and research leaders were provided ongoing in-country mentoring for analysis and manuscript writing. Participants reported a high level of skill, knowledge and collaborator development from class-based trainings and out-of-class mentorship that were sustained 1 year later. Five of six manuscripts were authored by multi-institution teams and submitted to international peer-reviewed scientific journals, and three-quarters of the participants mentored others in survey data analysis or conducted an additional survey analysis in the year following the training. Our model was effective in utilizing existing survey data and strengthening skills among full-time working professionals without disrupting ongoing work commitments and using few resources. Critical to our

  13. Blind deconvolution applied to acoustical systems identification with supporting experimental results

    NASA Astrophysics Data System (ADS)

    Roan, Michael J.; Gramann, Mark R.; Erling, Josh G.; Sibul, Leon H.

    2003-10-01

    Many acoustical applications require the analysis of a signal that is corrupted by an unknown filtering function. Examples arise in the areas of noise or vibration control, room acoustics, structural vibration analysis, and speech processing. Here, the observed signal can be modeled as the convolution of the desired signal with an unknown system impulse response. Blind deconvolution refers to the process of learning the inverse of this unknown impulse response and applying it to the observed signal to remove the filtering effects. Unlike classical deconvolution, which requires prior knowledge of the impulse response, blind deconvolution requires only reasonable prior estimates of the input signal's statistics. The significant contribution of this work lies in experimental verification of a blind deconvolution algorithm in the context of acoustical system identification. Previous experimental work concerning blind deconvolution in acoustics has been minimal, as previous literature concerning blind deconvolution uses computer simulated data. This paper examines experiments involving three classical acoustic systems: driven pipe, driven pipe with open side branch, and driven pipe with Helmholtz resonator side branch. Experimental results confirm that the deconvolution algorithm learns these systems' inverse impulse responses, and that application of these learned inverses removes the effects of the filters.

  14. Computing and Systems Applied in Support of Coordinated Energy, Environmental, and Climate Planning

    EPA Science Inventory

    This talk focuses on how Dr. Loughlin is applying Computing and Systems models, tools and methods to more fully understand the linkages among energy systems, environmental quality, and climate change. Dr. Loughlin will highlight recent and ongoing research activities, including: ...

  15. Computing and Systems Applied in Support of Coordinated Energy, Environmental, and Climate Planning

    EPA Science Inventory

    This talk focuses on how Dr. Loughlin is applying Computing and Systems models, tools and methods to more fully understand the linkages among energy systems, environmental quality, and climate change. Dr. Loughlin will highlight recent and ongoing research activities, including: ...

  16. Applying machine learning techniques to DNA sequence analysis. Progress report, February 14, 1991--February 13, 1992

    SciTech Connect

    Shavlik, J.W.

    1992-04-01

    We are developing a machine learning system that modifies existing knowledge about specific types of biological sequences. It does this by considering sample members and nonmembers of the sequence motif being learned. Using this information (which we call a ``domain theory``), our learning algorithm produces a more accurate representation of the knowledge needed to categorize future sequences. Specifically, the KBANN algorithm maps inference rules, such as consensus sequences, into a neural (connectionist) network. Neural network training techniques then use the training examples of refine these inference rules. We have been applying this approach to several problems in DNA sequence analysis and have also been extending the capabilities of our learning system along several dimensions.

  17. Geostatistics - a tool applied to the distribution of Legionella pneumophila in a hospital water system.

    PubMed

    Laganà, Pasqualina; Moscato, Umberto; Poscia, Andrea; La Milia, Daniele Ignazio; Boccia, Stefania; Avventuroso, Emanuela; Delia, Santi

    2015-01-01

    Legionnaires' disease is normally acquired by inhalation of legionellae from a contaminated environmental source. Water systems of large buildings, such as hospitals, are often contaminated with legionellae and therefore represent a potential risk for the hospital population. The aim of this study was to evaluate the potential contamination of Legionella pneumophila (LP) in a large hospital in Italy through georeferential statistical analysis to assess the possible sources of dispersion and, consequently, the risk of exposure for both health care staff and patients. LP serogroups 1 and 2-14 distribution was considered in the wards housed on two consecutive floors of the hospital building. On the basis of information provided by 53 bacteriological analysis, a 'random' grid of points was chosen and spatial geostatistics or FAIk Kriging was applied and compared with the results of classical statistical analysis. Over 50% of the examined samples were positive for Legionella pneumophila. LP 1 was isolated in 69% of samples from the ground floor and in 60% of sample from the first floor; LP 2-14 in 36% of sample from the ground floor and 24% from the first. The iso-estimation maps show clearly the most contaminated pipe and the difference in the diffusion of the different L. pneumophila serogroups. Experimental work has demonstrated that geostatistical methods applied to the microbiological analysis of water matrices allows a better modeling of the phenomenon under study, a greater potential for risk management and a greater choice of methods of prevention and environmental recovery to be put in place with respect to the classical statistical analysis.

  18. Analysis of bilinear stochastic systems

    NASA Technical Reports Server (NTRS)

    Willsky, A. S.; Martin, D. N.; Marcus, S. I.

    1975-01-01

    Analysis of stochastic dynamical systems that involve multiplicative (bilinear) noise processes. After defining the systems of interest, consideration is given to the evolution of the moments of such systems, the question of stochastic stability, and estimation for bilinear stochastic systems. Both exact and approximate methods of analysis are introduced, and, in particular, the uses of Lie-theoretic concepts and harmonic analysis are discussed.

  19. Launch Vehicle Systems Analysis

    NASA Technical Reports Server (NTRS)

    Olds, John R.

    1999-01-01

    This report summaries the key accomplishments of Georgia Tech's Space Systems Design Laboratory (SSDL) under NASA Grant NAG8-1302 from NASA - Marshall Space Flight Center. The report consists of this summary white paper, copies of technical papers written under this grant, and several viewgraph-style presentations. During the course of this grant four main tasks were completed: (1)Simulated Combined-Cycle Rocket Engine Analysis Module (SCCREAM), a computer analysis tool for predicting the performance of various RBCC engine configurations; (2) Hyperion, a single stage to orbit vehicle capable of delivering 25,000 pound payloads to the International Space Station Orbit; (3) Bantam-X Support - a small payload mission; (4) International Trajectory Support for interplanetary human Mars missions.

  20. Applied Time Domain Stability Margin Assessment for Nonlinear Time-Varying Systems

    NASA Technical Reports Server (NTRS)

    Kiefer, J. M.; Johnson, M. D.; Wall, J. H.; Dominguez, A.

    2016-01-01

    margins. At each time point, the system was linearized about the current operating point using Simulink's built-in solver. Each linearized system in time was evaluated for its rigid-body gain margin (high frequency gain margin), rigid-body phase margin, and aero gain margin (low frequency gain margin) for each control axis. Using the stability margins derived from the baseline linearization approach, the time domain derived stability margins were determined by executing time domain simulations in which axis-specific incremental gain and phase adjustments were made to the nominal system about the expected neutral stability point at specific flight times. The baseline stability margin time histories were used to shift the system gain to various values around the zero margin point such that a precise amount of expected gain margin was maintained throughout flight. When assessing the gain margins, the gain was applied starting at the time point under consideration, thereafter following the variation in the margin found in the linear analysis. When assessing the rigid-body phase margin, a constant time delay was applied to the system starting at the time point under consideration. If the baseline stability margins were correctly determined via the linear analysis, the time domain simulation results should contain unstable behavior at certain gain and phase values. Examples will be shown from repeated simulations with variable added gain and phase lag. Faithfulness of margins calculated from the linear analysis to the nonlinear system will be demonstrated.

  1. Constraints to applying systems thinking concepts in health systems: A regional perspective from surveying stakeholders in Eastern Mediterranean countries

    PubMed Central

    El-Jardali, Fadi; Adam, Taghreed; Ataya, Nour; Jamal, Diana; Jaafar, Maha

    2014-01-01

    Background: Systems Thinking (ST) has recently been promoted as an important approach to health systems strengthening. However, ST is not common practice, particularly in Low- and Middle-Income Countries (LMICs). This paper seeks to explore the barriers that may hinder its application in the Eastern Mediterranean Region (EMR) and possible strategies to mitigate them. Methods: A survey consisting of open-ended questions was conducted with a purposive sample of health policy-makers such as senior officials from the Ministry of Health (MoH), researchers, and other stakeholders such as civil society groups and professional associations from ten countries in the region. A total of 62 respondents participated in the study. Thematic analysis was conducted. Results: There was strong recognition of the relevance and usefulness of ST to health systems policy-making and research, although misconceptions about what ST means were also identified. Experience with applying ST was very limited. Approaches to designing health policies in the EMR were perceived as reactive and fragmented (66%). Commonly perceived constraints to application of ST were: a perceived notion of its costliness combined with lack of the necessary funding to operationalize it (53%), competing political interests and lack of government accountability (50%), lack of awareness about relevance and value (47%), limited capacity to apply it (45%), and difficulty in coordinating and managing stakeholders (39%). Conclusion: While several strategies have been proposed to mitigate most of these constraints, they emphasized the importance of political endorsement and adoption of ST at the leadership level, together with building the necessary capacity to apply it and apply the learning in research and practice. PMID:25489598

  2. Constraints to applying systems thinking concepts in health systems: A regional perspective from surveying stakeholders in Eastern Mediterranean countries.

    PubMed

    El-Jardali, Fadi; Adam, Taghreed; Ataya, Nour; Jamal, Diana; Jaafar, Maha

    2014-12-01

    Systems Thinking (ST) has recently been promoted as an important approach to health systems strengthening. However, ST is not common practice, particularly in Low- and Middle-Income Countries (LMICs). This paper seeks to explore the barriers that may hinder its application in the Eastern Mediterranean Region (EMR) and possible strategies to mitigate them. A survey consisting of open-ended questions was conducted with a purposive sample of health policy-makers such as senior officials from the Ministry of Health (MoH), researchers, and other stakeholders such as civil society groups and professional associations from ten countries in the region. A total of 62 respondents participated in the study. Thematic analysis was conducted. There was strong recognition of the relevance and usefulness of ST to health systems policy-making and research, although misconceptions about what ST means were also identified. Experience with applying ST was very limited. Approaches to designing health policies in the EMR were perceived as reactive and fragmented (66%). Commonly perceived constraints to application of ST were: a perceived notion of its costliness combined with lack of the necessary funding to operationalize it (53%), competing political interests and lack of government accountability (50%), lack of awareness about relevance and value (47%), limited capacity to apply it (45%), and difficulty in coordinating and managing stakeholders (39%). While several strategies have been proposed to mitigate most of these constraints, they emphasized the importance of political endorsement and adoption of ST at the leadership level, together with building the necessary capacity to apply it and apply the learning in research and practice.

  3. Neutron activation analysis system

    DOEpatents

    Taylor, M.C.; Rhodes, J.R.

    1973-12-25

    A neutron activation analysis system for monitoring a generally fluid media, such as slurries, solutions, and fluidized powders, including two separate conduit loops for circulating fluid samples within the range of radiation sources and detectors is described. Associated with the first loop is a neutron source that emits s high flux of slow and thermal neutrons. The second loop employs a fast neutron source, the flux from which is substantially free of thermal neutrons. Adjacent to both loops are gamma counters for spectrographic determination of the fluid constituents. Other gsmma sources and detectors are arranged across a portion of each loop for deterMining the fluid density. (Official Gazette)

  4. Beta systems error analysis

    NASA Technical Reports Server (NTRS)

    1984-01-01

    The atmospheric backscatter coefficient, beta, measured with an airborne CO Laser Doppler Velocimeter (LDV) system operating in a continuous wave, focussed model is discussed. The Single Particle Mode (SPM) algorithm, was developed from concept through analysis of an extensive amount of data obtained with the system on board a NASA aircraft. The SPM algorithm is intended to be employed in situations where one particle at a time appears in the sensitive volume of the LDV. In addition to giving the backscatter coefficient, the SPM algorithm also produces as intermediate results the aerosol density and the aerosol backscatter cross section distribution. A second method, which measures only the atmospheric backscatter coefficient, is called the Volume Mode (VM) and was simultaneously employed. The results of these two methods differed by slightly less than an order of magnitude. The measurement uncertainties or other errors in the results of the two methods are examined.

  5. The Research Mission of Universities of Applied Sciences and the Future Configuration of Higher Education Systems in Europe

    ERIC Educational Resources Information Center

    Lepori, Benedetto; Kyvik, Svein

    2010-01-01

    This article presents a comparative analysis of the development of research in universities of applied sciences (UAS) in eight European countries and its implications for the configuration of the higher education system. The enhancement of research has mostly been seen as a case of academic drift where UAS attempt to become more similar to…

  6. KRON's Method Applied to the Study of Electromagnetic Interference Occurring in Aerospace Systems

    NASA Astrophysics Data System (ADS)

    Leman, S.; Reineix, A.; Hoeppe, F.; Poiré, Y.; Mahoudi, M.; Démoulin, B.; Üstüner, F.; Rodriquez, V. P.

    2012-05-01

    In this paper, spacecraft and aircraft mock-ups are used to simulate the performance of KRON based tools applied to the simulation of large EMC systems. These tools aim to assist engineers in the design phase of complex systems. This is done by effectively evaluating the EM disturbances between antennas, electronic equipment, and Portable Electronic Devices found in large systems. We use a topological analysis of the system to model independent sub-volumes such as antennas, cables, equipments, PED and cavity walls. Each of these sub- volumes is modelled by an appropriate method which can be based on, for example, analytical expressions, transmission line theory or other numerical tools such as the full wave FDFD method. This representation associated with the electrical tensorial method of G.KRON leads to reasonable simulation times (typically a few minutes) and accurate results. Because equivalent sub-models are built separately, the main originality of this method is that each sub- volume can be easily replaced by another one without rebuilding the entire system. Comparisons between measurements and simulations will be also presented.

  7. Air pollution simulation and geographical information systems (GIS) applied to Athens International Airport.

    PubMed

    Theophanides, Mike; Anastassopoulou, Jane

    2009-07-01

    This study presents an improved methodology for analysing atmospheric pollution around airports using Gaussian-plume numerical simulation integrated with Geographical Information Systems (GIS). The new methodology focuses on streamlining the lengthy analysis process for Airport Environmental Impact Assessments by integrating the definition of emission sources, simulating and displaying the results in a GIS environment. One of the objectives of the research is to validate the methodology applied to the Athens International Airport, "Eleftherios Venizelos", to produce a realistic estimate of emission inventories, dispersion simulations and comparison to measured data. The methodology used a combination of the Emission Dispersion and Modelling System (EDMS) and the Atmospheric Dispersion and Modelling system (ADMS) to improve the analysis process. The second objective is to conduct numerical simulations under various adverse conditions (e.g. scenarios) and assess the dispersion in the surrounding areas. The study concludes that the use of GIS in environmental assessments provides a valuable advantage for organizing data and entering accurate geographical/topological information for the simulation engine. Emissions simulation produced estimates within 10% of published values. Dispersion simulations indicate that airport pollution will affect neighbouring cities such as Rafina and Loutsa. Presently, there are no measured controls in these areas. In some cases, airport pollution can contribute to as much as 40% of permissible EU levels in VOCs.

  8. Inclusive Elementary Classroom Teacher Knowledge of and Attitudes toward Applied Behavior Analysis and Autism Spectrum Disorder and Their Use of Applied Behavior Analysis

    ERIC Educational Resources Information Center

    McCormick, Jennifer A.

    2011-01-01

    The purpose of this study was to examine inclusive elementary teacher knowledge and attitude toward Autism Spectrum Disorder (ASD) and applied behavior analysis (ABA) and their use of ABA. Furthermore, this study examined if knowledge and attitude predicted use of ABA. A survey was developed and administered through a web-based program. Of the…

  9. System of systems modeling and analysis.

    SciTech Connect

    Campbell, James E.; Anderson, Dennis James; Longsine, Dennis E.; Shirah, Donald N.

    2005-01-01

    This report documents the results of an LDRD program entitled 'System of Systems Modeling and Analysis' that was conducted during FY 2003 and FY 2004. Systems that themselves consist of multiple systems (referred to here as System of Systems or SoS) introduce a level of complexity to systems performance analysis and optimization that is not readily addressable by existing capabilities. The objective of the 'System of Systems Modeling and Analysis' project was to develop an integrated modeling and simulation environment that addresses the complex SoS modeling and analysis needs. The approach to meeting this objective involved two key efforts. First, a static analysis approach, called state modeling, has been developed that is useful for analyzing the average performance of systems over defined use conditions. The state modeling capability supports analysis and optimization of multiple systems and multiple performance measures or measures of effectiveness. The second effort involves time simulation which represents every system in the simulation using an encapsulated state model (State Model Object or SMO). The time simulation can analyze any number of systems including cross-platform dependencies and a detailed treatment of the logistics required to support the systems in a defined mission.

  10. Applying Dynamical Systems Theory to Optimize Libration Point Orbit Stationkeeping Maneuvers for WIND

    NASA Technical Reports Server (NTRS)

    Brown, Jonathan M.; Petersen, Jeremy D.

    2014-01-01

    NASA's WIND mission has been operating in a large amplitude Lissajous orbit in the vicinity of the interior libration point of the Sun-Earth/Moon system since 2004. Regular stationkeeping maneuvers are required to maintain the orbit due to the instability around the collinear libration points. Historically these stationkeeping maneuvers have been performed by applying an incremental change in velocity, or (delta)v along the spacecraft-Sun vector as projected into the ecliptic plane. Previous studies have shown that the magnitude of libration point stationkeeping maneuvers can be minimized by applying the (delta)v in the direction of the local stable manifold found using dynamical systems theory. This paper presents the analysis of this new maneuver strategy which shows that the magnitude of stationkeeping maneuvers can be decreased by 5 to 25 percent, depending on the location in the orbit where the maneuver is performed. The implementation of the optimized maneuver method into operations is discussed and results are presented for the first two optimized stationkeeping maneuvers executed by WIND.

  11. Basic Concepts of Systems Analysis

    ERIC Educational Resources Information Center

    Packard, Ted; Fuhriman, Addie

    1973-01-01

    Discusses application of systems analysis to social and behavioral science as well as machine and non-machine systems. Presents rationale, principles, techniques, and advantages and disadvantages. (CJ)

  12. Reflections of the social environment in chimpanzee memory: applying rational analysis beyond humans

    PubMed Central

    Marewski, Julian N.; Schooler, Lael J.; Gilby, Ian C.

    2016-01-01

    In cognitive science, the rational analysis framework allows modelling of how physical and social environments impose information-processing demands onto cognitive systems. In humans, for example, past social contact among individuals predicts their future contact with linear and power functions. These features of the human environment constrain the optimal way to remember information and probably shape how memory records are retained and retrieved. We offer a primer on how biologists can apply rational analysis to study animal behaviour. Using chimpanzees (Pan troglodytes) as a case study, we modelled 19 years of observational data on their social contact patterns. Much like humans, the frequency of past encounters in chimpanzees linearly predicted future encounters, and the recency of past encounters predicted future encounters with a power function. Consistent with the rational analyses carried out for human memory, these findings suggest that chimpanzee memory performance should reflect those environmental regularities. In re-analysing existing chimpanzee memory data, we found that chimpanzee memory patterns mirrored their social contact patterns. Our findings hint that human and chimpanzee memory systems may have evolved to solve similar information-processing problems. Overall, rational analysis offers novel theoretical and methodological avenues for the comparative study of cognition. PMID:27853606

  13. Reflections of the social environment in chimpanzee memory: applying rational analysis beyond humans.

    PubMed

    Stevens, Jeffrey R; Marewski, Julian N; Schooler, Lael J; Gilby, Ian C

    2016-08-01

    In cognitive science, the rational analysis framework allows modelling of how physical and social environments impose information-processing demands onto cognitive systems. In humans, for example, past social contact among individuals predicts their future contact with linear and power functions. These features of the human environment constrain the optimal way to remember information and probably shape how memory records are retained and retrieved. We offer a primer on how biologists can apply rational analysis to study animal behaviour. Using chimpanzees (Pan troglodytes) as a case study, we modelled 19 years of observational data on their social contact patterns. Much like humans, the frequency of past encounters in chimpanzees linearly predicted future encounters, and the recency of past encounters predicted future encounters with a power function. Consistent with the rational analyses carried out for human memory, these findings suggest that chimpanzee memory performance should reflect those environmental regularities. In re-analysing existing chimpanzee memory data, we found that chimpanzee memory patterns mirrored their social contact patterns. Our findings hint that human and chimpanzee memory systems may have evolved to solve similar information-processing problems. Overall, rational analysis offers novel theoretical and methodological avenues for the comparative study of cognition.

  14. The Failure Analysis in Traction Power System

    NASA Astrophysics Data System (ADS)

    Kim, Hyungchul; Heo, Guk-bum; Lee, Hyungwoo; Kim, Dong Jin; Kim, Jin O.

    2008-10-01

    This paper presents a failure analysis of traction power systems. The electric railway consists of traction power systems, various vehicles, operating equipment, track, overhead line and electric equipment. It is a fundamental function of traction power systems that they supply customers with acceptable reliability and high quality power. Perhaps the most commonly used reliability assessment for railway systems has been the failure analysis of the traction signal system. The reliability assessment of traction power systems has also been an indispensable issue for reliability assessment. This paper deals with the classification of railway accidents caused by electrification problems, the estimation of failure rate in power equipments and failure analysis using fault trees. In study cases, the fault tree method for failure analysis is applied to railway substations in South Korea.

  15. Quantitative Systems Pharmacology Approaches Applied to Microphysiological Systems (MPS): Data Interpretation and Multi-MPS Integration

    PubMed Central

    Yu, J; Cilfone, NA; Large, EM; Sarkar, U; Wishnok, JS; Tannenbaum, SR; Hughes, DJ; Lauffenburger, DA; Griffith, LG; Stokes, CL; Cirit, M

    2015-01-01

    Our goal in developing Microphysiological Systems (MPS) technology is to provide an improved approach for more predictive preclinical drug discovery via a highly integrated experimental/computational paradigm. Success will require quantitative characterization of MPSs and mechanistic analysis of experimental findings sufficient to translate resulting insights from in vitro to in vivo. We describe herein a systems pharmacology approach to MPS development and utilization that incorporates more mechanistic detail than traditional pharmacokinetic/pharmacodynamic (PK/PD) models. A series of studies illustrates diverse facets of our approach. First, we demonstrate two case studies: a PK data analysis and an inflammation response––focused on a single MPS, the liver/immune MPS. Building on the single MPS modeling, a theoretical investigation of a four-MPS interactome then provides a quantitative way to consider several pharmacological concepts such as absorption, distribution, metabolism, and excretion in the design of multi-MPS interactome operation and experiments. PMID:26535159

  16. Quantitative Systems Pharmacology Approaches Applied to Microphysiological Systems (MPS): Data Interpretation and Multi-MPS Integration.

    PubMed

    Yu, J; Cilfone, N A; Large, E M; Sarkar, U; Wishnok, J S; Tannenbaum, S R; Hughes, D J; Lauffenburger, D A; Griffith, L G; Stokes, C L; Cirit, M

    2015-10-01

    Our goal in developing Microphysiological Systems (MPS) technology is to provide an improved approach for more predictive preclinical drug discovery via a highly integrated experimental/computational paradigm. Success will require quantitative characterization of MPSs and mechanistic analysis of experimental findings sufficient to translate resulting insights from in vitro to in vivo. We describe herein a systems pharmacology approach to MPS development and utilization that incorporates more mechanistic detail than traditional pharmacokinetic/pharmacodynamic (PK/PD) models. A series of studies illustrates diverse facets of our approach. First, we demonstrate two case studies: a PK data analysis and an inflammation response--focused on a single MPS, the liver/immune MPS. Building on the single MPS modeling, a theoretical investigation of a four-MPS interactome then provides a quantitative way to consider several pharmacological concepts such as absorption, distribution, metabolism, and excretion in the design of multi-MPS interactome operation and experiments.

  17. A transversal substructuring mode matching method applied to the acoustic analysis of dissipative mufflers

    NASA Astrophysics Data System (ADS)

    Albelda, J.; Denia, F. D.; Torres, M. I.; Fuenmayor, F. J.

    2007-06-01

    To carry out the acoustic analysis of dissipative silencers with uniform cross-section, the application of the mode matching method at the geometrical discontinuities is an attractive option from a computational point of view. The consideration of this methodology assumes, in general, that the modes associated with the transversal geometry of each element with uniform cross-section are known for the excitation frequencies considered in the analysis. The calculation of the transversal modes is not, however, a simple task when the acoustic system involves perforated elements and absorbent materials. The current work presents a modal approach to calculate the transversal modes and the corresponding axial wavenumbers for dissipative mufflers of uniform (but arbitrary) cross-section. The proposed technique is based on the division of the transversal section into subdomains and the subsequent use of a substructuring procedure with two sets of modes to improve the convergence. The former set of modes fulfils the condition of zero pressure at the common boundary between transversal subdomains while the latter satisfies the condition of zero derivative in the direction normal to the boundary. The approach leads to a versatile methodology with a moderate computational effort that can be applied to mufflers commonly found in real applications. To validate the procedure presented in this work, comparisons are provided with finite element predictions and results available in the literature, showing a good agreement. In addition, the procedure is applied to an example of practical interest.

  18. Applying a social network analysis (SNA) approach to understanding radiologists' performance in reading mammograms

    NASA Astrophysics Data System (ADS)

    Tavakoli Taba, Seyedamir; Hossain, Liaquat; Heard, Robert; Brennan, Patrick; Lee, Warwick; Lewis, Sarah

    2017-03-01

    Rationale and objectives: Observer performance has been widely studied through examining the characteristics of individuals. Applying a systems perspective, while understanding of the system's output, requires a study of the interactions between observers. This research explains a mixed methods approach to applying a social network analysis (SNA), together with a more traditional approach of examining personal/ individual characteristics in understanding observer performance in mammography. Materials and Methods: Using social networks theories and measures in order to understand observer performance, we designed a social networks survey instrument for collecting personal and network data about observers involved in mammography performance studies. We present the results of a study by our group where 31 Australian breast radiologists originally reviewed 60 mammographic cases (comprising of 20 abnormal and 40 normal cases) and then completed an online questionnaire about their social networks and personal characteristics. A jackknife free response operating characteristic (JAFROC) method was used to measure performance of radiologists. JAFROC was tested against various personal and network measures to verify the theoretical model. Results: The results from this study suggest a strong association between social networks and observer performance for Australian radiologists. Network factors accounted for 48% of variance in observer performance, in comparison to 15.5% for the personal characteristics for this study group. Conclusion: This study suggest a strong new direction for research into improving observer performance. Future studies in observer performance should consider social networks' influence as part of their research paradigm, with equal or greater vigour than traditional constructs of personal characteristics.

  19. Space lab system analysis

    NASA Technical Reports Server (NTRS)

    Ingels, F. M.; Rives, T. B.

    1987-01-01

    An analytical analysis of the HOSC Generic Peripheral processing system was conducted. The results are summarized and they indicate that the maximum delay in performing screen change requests should be less than 2.5 sec., occurring for a slow VAX host to video screen I/O rate of 50 KBps. This delay is due to the average I/O rate from the video terminals to their host computer. Software structure of the main computers and the host computers will have greater impact on screen change or refresh response times. The HOSC data system model was updated by a newly coded PASCAL based simulation program which was installed on the HOSC VAX system. This model is described and documented. Suggestions are offered to fine tune the performance of the ETERNET interconnection network. Suggestions for using the Nutcracker by Excelan to trace itinerate packets which appear on the network from time to time were offered in discussions with the HOSC personnel. Several visits to the HOSC facility were to install and demonstrate the simulation model.

  20. A method for work modeling at complex systems: towards applying information systems in family health care units.

    PubMed

    Jatobá, Alessandro; de Carvalho, Paulo Victor R; da Cunha, Amauri Marques

    2012-01-01

    Work in organizations requires a minimum level of consensus on the understanding of the practices performed. To adopt technological devices to support the activities in environments where work is complex, characterized by the interdependence among a large number of variables, understanding about how work is done not only takes an even greater importance, but also becomes a more difficult task. Therefore, this study aims to present a method for modeling of work in complex systems, which allows improving the knowledge about the way activities are performed where these activities do not simply happen by performing procedures. Uniting techniques of Cognitive Task Analysis with the concept of Work Process, this work seeks to provide a method capable of providing a detailed and accurate vision of how people perform their tasks, in order to apply information systems for supporting work in organizations.