High pressure common rail injection system modeling and control.
Wang, H P; Zheng, D; Tian, Y
2016-07-01
In this paper modeling and common-rail pressure control of high pressure common rail injection system (HPCRIS) is presented. The proposed mathematical model of high pressure common rail injection system which contains three sub-systems: high pressure pump sub-model, common rail sub-model and injector sub-model is a relative complicated nonlinear system. The mathematical model is validated by the software Matlab and a virtual detailed simulation environment. For the considered HPCRIS, an effective model free controller which is called Extended State Observer - based intelligent Proportional Integral (ESO-based iPI) controller is designed. And this proposed method is composed mainly of the referred ESO observer, and a time delay estimation based iPI controller. Finally, to demonstrate the performances of the proposed controller, the proposed ESO-based iPI controller is compared with a conventional PID controller and ADRC. Copyright © 2016 ISA. Published by Elsevier Ltd. All rights reserved.
Common Cause Failure Modeling: Aerospace Versus Nuclear
NASA Technical Reports Server (NTRS)
Stott, James E.; Britton, Paul; Ring, Robert W.; Hark, Frank; Hatfield, G. Spencer
2010-01-01
Aggregate nuclear plant failure data is used to produce generic common-cause factors that are specifically for use in the common-cause failure models of NUREG/CR-5485. Furthermore, the models presented in NUREG/CR-5485 are specifically designed to incorporate two significantly distinct assumptions about the methods of surveillance testing from whence this aggregate failure data came. What are the implications of using these NUREG generic factors to model the common-cause failures of aerospace systems? Herein, the implications of using the NUREG generic factors in the modeling of aerospace systems are investigated in detail and strong recommendations for modeling the common-cause failures of aerospace systems are given.
1985-11-01
As a o11066v. nlstle VSuSY £6I5PSAY I’ Iu PAS 11. Title Integrated Information Support System (1SS) Vol V - Common Data Model Subsystem Part 2 - CIMP ...AD-Mel1 236 INTEGRATED INFORMATION SUPPORT SYSTEM (IISS) VOLUME 5 1/2 COMMON DATA MODEL S.. (U) GENERAL ELECTRIC CO SCHENECTADY NY PRODUCTION...Volume V - Common Data Model Subsystem Part 2 - CDMP Test Case Report General Electric Company Production Resources Consulting One River Road
The Use of Object-Oriented Analysis Methods in Surety Analysis
DOE Office of Scientific and Technical Information (OSTI.GOV)
Craft, Richard L.; Funkhouser, Donald R.; Wyss, Gregory D.
1999-05-01
Object-oriented analysis methods have been used in the computer science arena for a number of years to model the behavior of computer-based systems. This report documents how such methods can be applied to surety analysis. By embodying the causality and behavior of a system in a common object-oriented analysis model, surety analysts can make the assumptions that underlie their models explicit and thus better communicate with system designers. Furthermore, given minor extensions to traditional object-oriented analysis methods, it is possible to automatically derive a wide variety of traditional risk and reliability analysis methods from a single common object model. Automaticmore » model extraction helps ensure consistency among analyses and enables the surety analyst to examine a system from a wider variety of viewpoints in a shorter period of time. Thus it provides a deeper understanding of a system's behaviors and surety requirements. This report documents the underlying philosophy behind the common object model representation, the methods by which such common object models can be constructed, and the rules required to interrogate the common object model for derivation of traditional risk and reliability analysis models. The methodology is demonstrated in an extensive example problem.« less
Emergence of a Common Modeling Architecture for Earth System Science (Invited)
NASA Astrophysics Data System (ADS)
Deluca, C.
2010-12-01
Common modeling architecture can be viewed as a natural outcome of common modeling infrastructure. The development of model utility and coupling packages (ESMF, MCT, OpenMI, etc.) over the last decade represents the realization of a community vision for common model infrastructure. The adoption of these packages has led to increased technical communication among modeling centers and newly coupled modeling systems. However, adoption has also exposed aspects of interoperability that must be addressed before easy exchange of model components among different groups can be achieved. These aspects include common physical architecture (how a model is divided into components) and model metadata and usage conventions. The National Unified Operational Prediction Capability (NUOPC), an operational weather prediction consortium, is collaborating with weather and climate researchers to define a common model architecture that encompasses these advanced aspects of interoperability and looks to future needs. The nature and structure of the emergent common modeling architecture will be discussed along with its implications for future model development.
ESPC Common Model Architecture
2014-09-30
1 DISTRIBUTION STATEMENT A. Approved for public release; distribution is unlimited. ESPC Common Model Architecture Earth System Modeling...Operational Prediction Capability (NUOPC) was established between NOAA and Navy to develop common software architecture for easy and efficient...development under a common model architecture and other software-related standards in this project. OBJECTIVES NUOPC proposes to accelerate
Model Data Interoperability for the United States Integrated Ocean Observing System (IOOS)
NASA Astrophysics Data System (ADS)
Signell, Richard P.
2010-05-01
Model data interoperability for the United States Integrated Ocean Observing System (IOOS) was initiated with a focused one year project. The problem was that there were many regional and national providers of oceanographic model data; each had unique file conventions, distribution techniques and analysis tools that made it difficult to compare model results and observational data. To solve this problem, a distributed system was built utilizing a customized middleware layer and a common data model. This allowed each model data provider to keep their existing model and data files unchanged, yet deliver model data via web services in a common form. With standards-based applications that used these web services, end users then had a common way to access data from any of the models. These applications included: (1) a 2D mapping and animation using a web browser application, (2) an advanced 3D visualization and animation using a desktop application, and (3) a toolkit for a common scientific analysis environment. Due to the flexibility and low impact of the approach on providers, rapid progress was made. The system was implemented in all eleven US IOOS regions and at the NOAA National Coastal Data Development Center, allowing common delivery of regional and national oceanographic model forecast and archived results that cover all US waters. The system, based heavily on software technology from the NSF-sponsored Unidata Program Center, is applicable to any structured gridded data, not just oceanographic model data. There is a clear pathway to expand the system to include unstructured grid (e.g. triangular grid) data.
NASA Technical Reports Server (NTRS)
Briggs, Hugh C.
2008-01-01
An error budget is a commonly used tool in design of complex aerospace systems. It represents system performance requirements in terms of allowable errors and flows these down through a hierarchical structure to lower assemblies and components. The requirements may simply be 'allocated' based upon heuristics or experience, or they may be designed through use of physics-based models. This paper presents a basis for developing an error budget for models of the system, as opposed to the system itself. The need for model error budgets arises when system models are a principle design agent as is increasingly more common for poorly testable high performance space systems.
Information Interaction Study for DER and DMS Interoperability
NASA Astrophysics Data System (ADS)
Liu, Haitao; Lu, Yiming; Lv, Guangxian; Liu, Peng; Chen, Yu; Zhang, Xinhui
The Common Information Model (CIM) is an abstract data model that can be used to represent the major objects in Distribution Management System (DMS) applications. Because the Common Information Model (CIM) doesn't modeling the Distributed Energy Resources (DERs), it can't meet the requirements of DER operation and management for Distribution Management System (DMS) advanced applications. Modeling of DER were studied based on a system point of view, the article initially proposed a CIM extended information model. By analysis the basic structure of the message interaction between DMS and DER, a bidirectional messaging mapping method based on data exchange was proposed.
Support System Effects on the NASA Common Research Model
NASA Technical Reports Server (NTRS)
Rivers, S. Melissa B.; Hunter, Craig A.
2012-01-01
An experimental investigation of the NASA Common Research Model was conducted in the NASA Langley National Transonic Facility and NASA Ames 11-Foot Transonic Wind Tunnel Facility for use in the Drag Prediction Workshop. As data from the experimental investigations was collected, a large difference in moment values was seen between the experimental and the computational data from the 4th Drag Prediction Workshop. This difference led to the present work. In this study, a computational assessment has been undertaken to investigate model support system interference effects on the Common Research Model. The configurations computed during this investigation were the wing/body/tail=0deg without the support system and the wing/body/tail=0deg with the support system. The results from this investigation confirm that the addition of the support system to the computational cases does shift the pitching moment in the direction of the experimental results.
Motion Planning in a Society of Intelligent Mobile Agents
NASA Technical Reports Server (NTRS)
Esterline, Albert C.; Shafto, Michael (Technical Monitor)
2002-01-01
The majority of the work on this grant involved formal modeling of human-computer integration. We conceptualize computer resources as a multiagent system so that these resources and human collaborators may be modeled uniformly. In previous work we had used modal for this uniform modeling, and we had developed a process-algebraic agent abstraction. In this work, we applied this abstraction (using CSP) in uniformly modeling agents and users, which allowed us to use tools for investigating CSP models. This work revealed the power of, process-algebraic handshakes in modeling face-to-face conversation. We also investigated specifications of human-computer systems in the style of algebraic specification. This involved specifying the common knowledge required for coordination and process-algebraic patterns of communication actions intended to establish the common knowledge. We investigated the conditions for agents endowed with perception to gain common knowledge and implemented a prototype neural-network system that allows agents to detect when such conditions hold. The literature on multiagent systems conceptualizes communication actions as speech acts. We implemented a prototype system that infers the deontic effects (obligations, permissions, prohibitions) of speech acts and detects violations of these effects. A prototype distributed system was developed that allows users to collaborate in moving proxy agents; it was designed to exploit handshakes and common knowledge Finally. in work carried over from a previous NASA ARC grant, about fifteen undergraduates developed and presented projects on multiagent motion planning.
Modeling and Analysis of Mixed Synchronous/Asynchronous Systems
NASA Technical Reports Server (NTRS)
Driscoll, Kevin R.; Madl. Gabor; Hall, Brendan
2012-01-01
Practical safety-critical distributed systems must integrate safety critical and non-critical data in a common platform. Safety critical systems almost always consist of isochronous components that have synchronous or asynchronous interface with other components. Many of these systems also support a mix of synchronous and asynchronous interfaces. This report presents a study on the modeling and analysis of asynchronous, synchronous, and mixed synchronous/asynchronous systems. We build on the SAE Architecture Analysis and Design Language (AADL) to capture architectures for analysis. We present preliminary work targeted to capture mixed low- and high-criticality data, as well as real-time properties in a common Model of Computation (MoC). An abstract, but representative, test specimen system was created as the system to be modeled.
Energy Efficient Operation of Ammonia Refrigeration Systems
DOE Office of Scientific and Technical Information (OSTI.GOV)
Mohammed, Abdul Qayyum; Wenning, Thomas J; Sever, Franc
Ammonia refrigeration systems typically offer many energy efficiency opportunities because of their size and complexity. This paper develops a model for simulating single-stage ammonia refrigeration systems, describes common energy saving opportunities, and uses the model to quantify those opportunities. The simulation model uses data that are typically available during site visits to ammonia refrigeration plants and can be calibrated to actual consumption and performance data if available. Annual electricity consumption for a base-case ammonia refrigeration system is simulated. The model is then used to quantify energy savings for six specific energy efficiency opportunities; reduce refrigeration load, increase suction pressure, employmore » dual suction, decrease minimum head pressure set-point, increase evaporative condenser capacity, and reclaim heat. Methods and considerations for achieving each saving opportunity are discussed. The model captures synergistic effects that result when more than one component or parameter is changed. This methodology represents an effective method to model and quantify common energy saving opportunities in ammonia refrigeration systems. The results indicate the range of savings that might be expected from common energy efficiency opportunities.« less
2015-09-30
originate from NASA , NOAA , and community modeling efforts, and support for creation of the suite was shared by sponsors from other agencies. ESPS...Framework (ESMF) Software and Application Development Cecelia Deluca NESII/CIRES/ NOAA Earth System Research Laboratory 325 Broadway Boulder, CO...Capability (NUOPC) was established between NOAA and Navy to develop a common software architecture for easy and efficient interoperability. The
NASA Technical Reports Server (NTRS)
Hark, Frank; Britton, Paul; Ring, Robert; Novack, Steven
2015-01-01
Space Launch System (SLS) Agenda: Objective; Key Definitions; Calculating Common Cause; Examples; Defense against Common Cause; Impact of varied Common Cause Failure (CCF) and abortability; Response Surface for various CCF Beta; Takeaways.
The structure of common-envelope remnants
NASA Astrophysics Data System (ADS)
Hall, Philip D.
2015-05-01
We investigate the structure and evolution of the remnants of common-envelope evolution in binary star systems. In a common-envelope phase, two stars become engulfed in a gaseous envelope and, under the influence of drag forces, spiral to smaller separations. They may merge to form a single star or the envelope may be ejected to leave the stars in a shorter period orbit. This process explains the short orbital periods of many observed binary systems, such as cataclysmic variables and low-mass X-ray binary systems. Despite the importance of these systems, and of common-envelope evolution to their formation, it remains poorly understood. Specifically, we are unable to confidently predict the outcome of a common-envelope phase from the properties at its onset. After presenting a review of work on stellar evolution, binary systems, common-envelope evolution and the computer programs used, we describe the results of three computational projects on common-envelope evolution. Our work specifically relates to the methods and prescriptions which are used for predicting the outcome. We use the Cambridge stellar-evolution code STARS to produce detailed models of the structure and evolution of remnants of common-envelope evolution. We compare different assumptions about the uncertain end-of-common envelope structure and envelope mass of remnants which successfully eject their common envelopes. In the first project, we use detailed remnant models to investigate whether planetary nebulae are predicted after common-envelope phases initiated by low-mass red giants. We focus on the requirement that a remnant evolves rapidly enough to photoionize the nebula and compare the predictions for different ideas about the structure at the end of a common-envelope phase. We find that planetary nebulae are possible for some prescriptions for the end-of-common envelope structure. In our second contribution, we compute a large set of single-star models and fit new formulae to the core radii of evolved stars. These formulae can be used to better compute the outcome of common-envelope evolution with rapid evolution codes. We find that the new formulae are necessary for accurate predictions of the properties of post-common envelope systems. Finally, we use detailed remnant models of massive stars to investigate whether hydrogen may be retained after a common-envelope phase to the point of core-collapse and so be observable in supernovae. We find that this is possible and thus common-envelope evolution may contribute to the formation of Type IIb supernovae.
Multitask TSK fuzzy system modeling by mining intertask common hidden structure.
Jiang, Yizhang; Chung, Fu-Lai; Ishibuchi, Hisao; Deng, Zhaohong; Wang, Shitong
2015-03-01
The classical fuzzy system modeling methods implicitly assume data generated from a single task, which is essentially not in accordance with many practical scenarios where data can be acquired from the perspective of multiple tasks. Although one can build an individual fuzzy system model for each task, the result indeed tells us that the individual modeling approach will get poor generalization ability due to ignoring the intertask hidden correlation. In order to circumvent this shortcoming, we consider a general framework for preserving the independent information among different tasks and mining hidden correlation information among all tasks in multitask fuzzy modeling. In this framework, a low-dimensional subspace (structure) is assumed to be shared among all tasks and hence be the hidden correlation information among all tasks. Under this framework, a multitask Takagi-Sugeno-Kang (TSK) fuzzy system model called MTCS-TSK-FS (TSK-FS for multiple tasks with common hidden structure), based on the classical L2-norm TSK fuzzy system, is proposed in this paper. The proposed model can not only take advantage of independent sample information from the original space for each task, but also effectively use the intertask common hidden structure among multiple tasks to enhance the generalization performance of the built fuzzy systems. Experiments on synthetic and real-world datasets demonstrate the applicability and distinctive performance of the proposed multitask fuzzy system model in multitask regression learning scenarios.
NASA Astrophysics Data System (ADS)
Wang, Qinpeng; Yang, Jianguo; Xin, Dong; He, Yuhai; Yu, Yonghua
2018-05-01
In this paper, based on the characteristic analyzing of the mechanical fuel injection system for the marine medium-speed diesel engine, a sectional high-pressure common rail fuel injection system is designed, rated condition rail pressure of which is 160MPa. The system simulation model is built and the performance of the high pressure common rail fuel injection system is analyzed, research results provide the technical foundation for the system engineering development.
NASA Technical Reports Server (NTRS)
1977-01-01
The development of a framework and structure for shuttle era unmanned spacecraft projects and the development of a commonality evaluation model is documented. The methodology developed for model utilization in performing cost trades and comparative evaluations for commonality studies is discussed. The model framework consists of categories of activities associated with the spacecraft system's development process. The model structure describes the physical elements to be treated as separate identifiable entities. Cost estimating relationships for subsystem and program-level components were calculated.
Graphical Modeling Meets Systems Pharmacology.
Lombardo, Rosario; Priami, Corrado
2017-01-01
A main source of failures in systems projects (including systems pharmacology) is poor communication level and different expectations among the stakeholders. A common and not ambiguous language that is naturally comprehensible by all the involved players is a boost to success. We present bStyle, a modeling tool that adopts a graphical language close enough to cartoons to be a common media to exchange ideas and data and that it is at the same time formal enough to enable modeling, analysis, and dynamic simulations of a system. Data analysis and simulation integrated in the same application are fundamental to understand the mechanisms of actions of drugs: a core aspect of systems pharmacology.
Graphical Modeling Meets Systems Pharmacology
Lombardo, Rosario; Priami, Corrado
2017-01-01
A main source of failures in systems projects (including systems pharmacology) is poor communication level and different expectations among the stakeholders. A common and not ambiguous language that is naturally comprehensible by all the involved players is a boost to success. We present bStyle, a modeling tool that adopts a graphical language close enough to cartoons to be a common media to exchange ideas and data and that it is at the same time formal enough to enable modeling, analysis, and dynamic simulations of a system. Data analysis and simulation integrated in the same application are fundamental to understand the mechanisms of actions of drugs: a core aspect of systems pharmacology. PMID:28469411
A conceptual model for megaprogramming
NASA Technical Reports Server (NTRS)
Tracz, Will
1990-01-01
Megaprogramming is component-based software engineering and life-cycle management. Magaprogramming and its relationship to other research initiatives (common prototyping system/common prototyping language, domain specific software architectures, and software understanding) are analyzed. The desirable attributes of megaprogramming software components are identified and a software development model and resulting prototype megaprogramming system (library interconnection language extended by annotated Ada) are described.
TRANSFORM - TRANsient Simulation Framework of Reconfigurable Models
DOE Office of Scientific and Technical Information (OSTI.GOV)
Greenwood, Michael S; Cetiner, Mustafa S; Fugate, David L
Existing development tools for early stage design and scoping of energy systems are often time consuming to use, proprietary, and do not contain the necessary function to model complete systems (i.e., controls, primary, and secondary systems) in a common platform. The Modelica programming language based TRANSFORM tool (1) provides a standardized, common simulation environment for early design of energy systems (i.e., power plants), (2) provides a library of baseline component modules to be assembled into full plant models using available geometry, design, and thermal-hydraulic data, (3) defines modeling conventions for interconnecting component models, and (4) establishes user interfaces and supportmore » tools to facilitate simulation development (i.e., configuration and parameterization), execution, and results display and capture.« less
Colvin, Michael E.; Pierce, Clay; Stewart, Timothy W.
2015-01-01
Food web modeling is recognized as fundamental to understanding the complexities of aquatic systems. Ecopath is the most common mass-balance model used to represent food webs and quantify trophic interactions among groups. We constructed annual Ecopath models for four consecutive years during the first half-decade of a zebra mussel invasion in shallow, eutrophic Clear Lake, Iowa, USA, to evaluate changes in relative biomass and total system consumption among food web groups, evaluate food web impacts of non-native common carp and zebra mussels on food web groups, and to interpret food web impacts in light of on-going lake restoration. Total living biomass increased each year of the study; the majority of the increase due to a doubling in planktonic blue green algae, but several other taxa also increased including a more than two-order of magnitude increase in zebra mussels. Common carp accounted for the largest percentage of total fish biomass throughout the study even with on-going harvest. Chironomids, common carp, and zebra mussels were the top-three ranking consumer groups. Non-native common carp and zebra mussels accounted for an average of 42% of the total system consumption. Despite the relatively high biomass densities of common carp and zebra mussel, food web impacts was minimal due to excessive benthic and primary production in this eutrophic system. Consumption occurring via benthic pathways dominated system consumption in Clear Lake throughout our study, supporting the argument that benthic food webs are significant in shallow, eutrophic lake ecosystems and must be considered if ecosystem-level understanding is to be obtained.
Common modeling system for digital simulation
NASA Technical Reports Server (NTRS)
Painter, Rick
1994-01-01
The Joint Modeling and Simulation System is a tri-service investigation into a common modeling framework for the development digital models. The basis for the success of this framework is an X-window-based, open systems architecture, object-based/oriented methodology, standard interface approach to digital model construction, configuration, execution, and post processing. For years Department of Defense (DOD) agencies have produced various weapon systems/technologies and typically digital representations of the systems/technologies. These digital representations (models) have also been developed for other reasons such as studies and analysis, Cost Effectiveness Analysis (COEA) tradeoffs, etc. Unfortunately, there have been no Modeling and Simulation (M&S) standards, guidelines, or efforts towards commonality in DOD M&S. The typical scenario is an organization hires a contractor to build hardware and in doing so an digital model may be constructed. Until recently, this model was not even obtained by the organization. Even if it was procured, it was on a unique platform, in a unique language, with unique interfaces, and, with the result being UNIQUE maintenance required. Additionally, the constructors of the model expended more effort in writing the 'infrastructure' of the model/simulation (e.g. user interface, database/database management system, data journalizing/archiving, graphical presentations, environment characteristics, other components in the simulation, etc.) than in producing the model of the desired system. Other side effects include: duplication of efforts; varying assumptions; lack of credibility/validation; and decentralization in policy and execution. J-MASS provides the infrastructure, standards, toolset, and architecture to permit M&S developers and analysts to concentrate on the their area of interest.
Chaotic Dynamics and Application of LCR Oscillators Sharing Common Nonlinearity
NASA Astrophysics Data System (ADS)
Jeevarekha, A.; Paul Asir, M.; Philominathan, P.
2016-06-01
This paper addresses the problem of sharing common nonlinearity among nonautonomous and autonomous oscillators. By choosing a suitable common nonlinear element with the driving point characteristics capable of bringing out chaotic motion in a combined system, we obtain identical chaotic states. The dynamics of the coupled system is explored through numerical and experimental studies. Employing the concept of common nonlinearity, a simple chaotic communication system is modeled and its performance is verified through Multisim simulation.
Further Investigation of the Support System Effects and Wing Twist on the NASA Common Research Model
NASA Technical Reports Server (NTRS)
Rivers, Melissa B.; Hunter, Craig A.; Campbell, Richard L.
2012-01-01
An experimental investigation of the NASA Common Research Model was conducted in the NASA Langley National Transonic Facility and NASA Ames 11-foot Transonic Wind Tunnel Facility for use in the Drag Prediction Workshop. As data from the experimental investigations was collected, a large difference in moment values was seen between the experiment and computational data from the 4th Drag Prediction Workshop. This difference led to a computational assessment to investigate model support system interference effects on the Common Research Model. The results from this investigation showed that the addition of the support system to the computational cases did increase the pitching moment so that it more closely matched the experimental results, but there was still a large discrepancy in pitching moment. This large discrepancy led to an investigation into the shape of the as-built model, which in turn led to a change in the computational grids and re-running of all the previous support system cases. The results of these cases are the focus of this paper.
Modeling Common Cause Failures of Thrusters on ISS Visiting Vehicles
NASA Technical Reports Server (NTRS)
Haught, Megan
2014-01-01
This paper discusses the methodology used to model common cause failures of thrusters on the International Space Station (ISS) Visiting Vehicles. The ISS Visiting Vehicles each have as many as 32 thrusters, whose redundancy makes them susceptible to common cause failures. The Global Alpha Model (as described in NUREG/CR-5485) can be used to represent the system common cause contribution, but NUREG/CR-5496 supplies global alpha parameters for groups only up to size six. Because of the large number of redundant thrusters on each vehicle, regression is used to determine parameter values for groups of size larger than six. An additional challenge is that Visiting Vehicle thruster failures must occur in specific combinations in order to fail the propulsion system; not all failure groups of a certain size are critical.
NASA Technical Reports Server (NTRS)
Joshi, Anjali; Heimdahl, Mats P. E.; Miller, Steven P.; Whalen, Mike W.
2006-01-01
System safety analysis techniques are well established and are used extensively during the design of safety-critical systems. Despite this, most of the techniques are highly subjective and dependent on the skill of the practitioner. Since these analyses are usually based on an informal system model, it is unlikely that they will be complete, consistent, and error free. In fact, the lack of precise models of the system architecture and its failure modes often forces the safety analysts to devote much of their effort to gathering architectural details about the system behavior from several sources and embedding this information in the safety artifacts such as the fault trees. This report describes Model-Based Safety Analysis, an approach in which the system and safety engineers share a common system model created using a model-based development process. By extending the system model with a fault model as well as relevant portions of the physical system to be controlled, automated support can be provided for much of the safety analysis. We believe that by using a common model for both system and safety engineering and automating parts of the safety analysis, we can both reduce the cost and improve the quality of the safety analysis. Here we present our vision of model-based safety analysis and discuss the advantages and challenges in making this approach practical.
NASA Technical Reports Server (NTRS)
Platt, M. E.; Lewis, E. E.; Boehm, F.
1991-01-01
A Monte Carlo Fortran computer program was developed that uses two variance reduction techniques for computing system reliability applicable to solving very large highly reliable fault-tolerant systems. The program is consistent with the hybrid automated reliability predictor (HARP) code which employs behavioral decomposition and complex fault-error handling models. This new capability is called MC-HARP which efficiently solves reliability models with non-constant failures rates (Weibull). Common mode failure modeling is also a specialty.
Label Structured Cell Proliferation Models
2010-06-16
and (, + ) are the cell proliferation and death rates , respectively, relative to the moving label coordinate system + . Daughter...proliferation and death rates relative to this new coordinate system. While not common in the biological sciences, it is altogether common in the physical
Schmickl, Thomas; Karsai, Istvan
2014-01-01
We develop a model to produce plausible patterns of task partitioning in the ponerine ant Ectatomma ruidum based on the availability of living prey and prey corpses. The model is based on the organizational capabilities of a “common stomach” through which the colony utilizes the availability of a natural (food) substance as a major communication channel to regulate the income and expenditure of the very same substance. This communication channel has also a central role in regulating task partitioning of collective hunting behavior in a supply&demand-driven manner. Our model shows that task partitioning of the collective hunting behavior in E. ruidum can be explained by regulation due to a common stomach system. The saturation of the common stomach provides accessible information to individual ants so that they can adjust their hunting behavior accordingly by engaging in or by abandoning from stinging or transporting tasks. The common stomach is able to establish and to keep stabilized an effective mix of workforce to exploit the prey population and to transport food into the nest. This system is also able to react to external perturbations in a de-centralized homeostatic way, such as to changes in the prey density or to accumulation of food in the nest. In case of stable conditions the system develops towards an equilibrium concerning colony size and prey density. Our model shows that organization of work through a common stomach system can allow Ectatomma ruidum to collectively forage for food in a robust, reactive and reliable way. The model is compared to previously published models that followed a different modeling approach. Based on our model analysis we also suggest a series of experiments for which our model gives plausible predictions. These predictions are used to formulate a set of testable hypotheses that should be investigated empirically in future experimentation. PMID:25493558
Modeling Common Cause Failures of Thrusters on ISS Visiting Vehicles
NASA Technical Reports Server (NTRS)
Haught, Megan; Duncan, Gary
2014-01-01
This paper discusses the methodology used to model common cause failures of thrusters on the International Space Station (ISS) Visiting Vehicles. The ISS Visiting Vehicles each have as many as 32 thrusters, whose redundancy and similar design make them susceptible to common cause failures. The Global Alpha Model (as described in NUREG/CR-5485) can be used to represent the system common cause contribution, but NUREG/CR-5496 supplies global alpha parameters for groups only up to size six. Because of the large number of redundant thrusters on each vehicle, regression is used to determine parameter values for groups of size larger than six. An additional challenge is that Visiting Vehicle thruster failures must occur in specific combinations in order to fail the propulsion system; not all failure groups of a certain size are critical.
Privacy and security in teleradiology.
Ruotsalainen, Pekka
2010-01-01
Teleradiology is probably the most successful eHealth service available today. Its business model is based on the remote transmission of radiological images (e.g. X-ray and CT-images) over electronic networks, and on the interpretation of the transmitted images for diagnostic purpose. Two basic service models are commonly used teleradiology today. The most common approach is based on the message paradigm (off-line model), but more developed teleradiology systems are based on the interactive use of PACS/RIS systems. Modern teleradiology is also more and more cross-organisational or even cross-border service between service providers having different jurisdictions and security policies. This paper defines the requirements needed to make different teleradiology models trusted. Those requirements include a common security policy that covers all partners and entities, common security and privacy protection principles and requirements, controlled contracts between partners, and the use of security controls and tools that supporting the common security policy. The security and privacy protection of any teleradiology system must be planned in advance, and the necessary security and privacy enhancing tools should be selected (e.g. strong authentication, data encryption, non-repudiation services and audit-logs) based on the risk analysis and requirements set by the legislation. In any case the teleradiology system should fulfil ethical and regulatory requirements. Certification of the whole teleradiology service system including security and privacy is also proposed. In the future, teleradiology services will be an integrated part of pervasive eHealth. Security requirements for this environment including dynamic and context aware security services are also discussed in this paper. Copyright (c) 2009 Elsevier Ireland Ltd. All rights reserved.
THE EPA MULTIMEDIA INTEGRATED MODELING SYSTEM SOFTWARE SUITE
The U.S. EPA is developing a Multimedia Integrated Modeling System (MIMS) framework that will provide a software infrastructure or environment to support constructing, composing, executing, and evaluating complex modeling studies. The framework will include (1) common software ...
A Transparent Translation from Legacy System Model into Common Information Model: Preprint
DOE Office of Scientific and Technical Information (OSTI.GOV)
Ding, Fei; Simpson, Jeffrey; Zhang, Yingchen
Advance in smart grid is forcing utilities towards better monitoring, control and analysis of distribution systems, and requires extensive cyber-based intelligent systems and applications to realize various functionalities. The ability of systems, or components within systems, to interact and exchange services or information with each other is the key to the success of smart grid technologies, and it requires efficient information exchanging and data sharing infrastructure. The Common Information Model (CIM) is a standard that allows different applications to exchange information about an electrical system, and it has become a widely accepted solution for information exchange among different platforms andmore » applications. However, most existing legacy systems are not developed using CIM, but using their own languages. Integrating such legacy systems is a challenge for utilities, and the appropriate utilization of the integrated legacy systems is even more intricate. Thus, this paper has developed an approach and open-source tool in order to translate legacy system models into CIM format. The developed tool is tested for a commercial distribution management system and simulation results have proved its effectiveness.« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Malczynski, Leonard A.
This guide addresses software quality in the construction of Powersim{reg_sign} Studio 8 system dynamics simulation models. It is the result of almost ten years of experience with the Powersim suite of system dynamics modeling tools (Constructor and earlier Studio versions). It is a guide that proposes a common look and feel for the construction of Powersim Studio system dynamics models.
Adaptable state based control system
NASA Technical Reports Server (NTRS)
Rasmussen, Robert D. (Inventor); Dvorak, Daniel L. (Inventor); Gostelow, Kim P. (Inventor); Starbird, Thomas W. (Inventor); Gat, Erann (Inventor); Chien, Steve Ankuo (Inventor); Keller, Robert M. (Inventor)
2004-01-01
An autonomous controller, comprised of a state knowledge manager, a control executor, hardware proxies and a statistical estimator collaborates with a goal elaborator, with which it shares common models of the behavior of the system and the controller. The elaborator uses the common models to generate from temporally indeterminate sets of goals, executable goals to be executed by the controller. The controller may be updated to operate in a different system or environment than that for which it was originally designed by the replacement of shared statistical models and by the instantiation of a new set of state variable objects derived from a state variable class. The adaptation of the controller does not require substantial modification of the goal elaborator for its application to the new system or environment.
Software model of a machine vision system based on the common house fly.
Madsen, Robert; Barrett, Steven; Wilcox, Michael
2005-01-01
The vision system of the common house fly has many properties, such as hyperacuity and parallel structure, which would be advantageous in a machine vision system. A software model has been developed which is ultimately intended to be a tool to guide the design of an analog real time vision system. The model starts by laying out cartridges over an image. The cartridges are analogous to the ommatidium of the fly's eye and contain seven photoreceptors each with a Gaussian profile. The spacing between photoreceptors is variable providing for more or less detail as needed. The cartridges provide information on what type of features they see and neighboring cartridges share information to construct a feature map.
Load Modeling and Calibration Techniques for Power System Studies
DOE Office of Scientific and Technical Information (OSTI.GOV)
Chassin, Forrest S.; Mayhorn, Ebony T.; Elizondo, Marcelo A.
2011-09-23
Load modeling is the most uncertain area in power system simulations. Having an accurate load model is important for power system planning and operation. Here, a review of load modeling and calibration techniques is given. This paper is not comprehensive, but covers some of the techniques most commonly found in the literature. The advantages and disadvantages of each technique are outlined.
Design and modelling of a link monitoring mechanism for the Common Data Link (CDL)
NASA Astrophysics Data System (ADS)
Eichelberger, John W., III
1994-09-01
The Common Data Link (CDL) is a full duplex, point-to-point microwave communications system used in imagery and signals intelligence collection systems. It provides a link between two remote Local Area Networks (LAN's) aboard collection and surface platforms. In a hostile environment, there is an overwhelming need to dynamically monitor the link and thus, limit the impact of jamming. This work describes steps taken to design, model, and evaluate a link monitoring system suitable for the CDL. The monitoring system is based on features and monitoring constructs of the Link Control Protocol (LCP) in the Point-to-Point Protocol (PPP) suite. The CDL model is based on a system of two remote Fiber Distributed Data Interface (FDDI) LAN's. In particular, the policies and mechanisms associated with monitoring are described in detail. An implementation of the required mechanisms using the OPNET network engineering tool is described. Performance data related to monitoring parameters is reported. Finally, integration of the FDDI-CDL model with the OPNET Internet model is described.
Joint Common Architecture Demonstration (JCA Demo) Final Report
2016-07-28
approach for implementing open systems [16], formerly known as the Modular Open Systems Approach (MOSA). OSA is a business and technical strategy to... TECHNICAL REPORT RDMR-AD-16-01 JOINT COMMON ARCHITECTURE DEMONSTRATION (JCA DEMO) FINAL REPORT Scott A. Wigginton... Modular Avionics .......................................................................... 5 E. Model-Based Engineering
Frost, Ram
2012-10-01
I have argued that orthographic processing cannot be understood and modeled without considering the manner in which orthographic structure represents phonological, semantic, and morphological information in a given writing system. A reading theory, therefore, must be a theory of the interaction of the reader with his/her linguistic environment. This outlines a novel approach to studying and modeling visual word recognition, an approach that focuses on the common cognitive principles involved in processing printed words across different writing systems. These claims were challenged by several commentaries that contested the merits of my general theoretical agenda, the relevance of the evolution of writing systems, and the plausibility of finding commonalities in reading across orthographies. Other commentaries extended the scope of the debate by bringing into the discussion additional perspectives. My response addresses all these issues. By considering the constraints of neurobiology on modeling reading, developmental data, and a large scope of cross-linguistic evidence, I argue that front-end implementations of orthographic processing that do not stem from a comprehensive theory of the complex information conveyed by writing systems do not present a viable approach for understanding reading. The common principles by which writing systems have evolved to represent orthographic, phonological, and semantic information in a language reveal the critical distributional characteristics of orthographic structure that govern reading behavior. Models of reading should thus be learning models, primarily constrained by cross-linguistic developmental evidence that describes how the statistical properties of writing systems shape the characteristics of orthographic processing. When this approach is adopted, a universal model of reading is possible.
Robust Synchronization Models for Presentation System Using SMIL-Driven Approach
ERIC Educational Resources Information Center
Asnawi, Rustam; Ahmad, Wan Fatimah Wan; Rambli, Dayang Rohaya Awang
2013-01-01
Current common Presentation System (PS) models are slide based oriented and lack synchronization analysis either with temporal or spatial constraints. Such models, in fact, tend to lead to synchronization problems, particularly on parallel synchronization with spatial constraints between multimedia element presentations. However, parallel…
A practical model for pressure probe system response estimation (with review of existing models)
NASA Astrophysics Data System (ADS)
Hall, B. F.; Povey, T.
2018-04-01
The accurate estimation of the unsteady response (bandwidth) of pneumatic pressure probe systems (probe, line and transducer volume) is a common practical problem encountered in the design of aerodynamic experiments. Understanding the bandwidth of the probe system is necessary to capture unsteady flow features accurately. Where traversing probes are used, the desired traverse speed and spatial gradients in the flow dictate the minimum probe system bandwidth required to resolve the flow. Existing approaches for bandwidth estimation are either complex or inaccurate in implementation, so probes are often designed based on experience. Where probe system bandwidth is characterized, it is often done experimentally, requiring careful experimental set-up and analysis. There is a need for a relatively simple but accurate model for estimation of probe system bandwidth. A new model is presented for the accurate estimation of pressure probe bandwidth for simple probes commonly used in wind tunnel environments; experimental validation is provided. An additional, simple graphical method for air is included for convenience.
Bridging Hydroinformatics Services Between HydroShare and SWATShare
NASA Astrophysics Data System (ADS)
Merwade, V.; Zhao, L.; Song, C. X.; Tarboton, D. G.; Goodall, J. L.; Stealey, M.; Rajib, A.; Morsy, M. M.; Dash, P. K.; Miles, B.; Kim, I. L.
2016-12-01
Many cyberinfrastructure systems in the hydrologic and related domains emerged in the past decade with more being developed to address various data management and modeling needs. Although clearly beneficial to the broad user community, it is a challenging task to build interoperability across these systems due to various obstacles including technological, organizational, semantic, and social issues. This work presents our experience in developing interoperability between two hydrologic cyberinfrastructure systems - SWATShare and HydroShare. HydroShare is a large-scale online system aiming at enabling the hydrologic user community to share their data, models, and analysis online for solving complex hydrologic research questions. On the other side, SWATShare is a focused effort to allow SWAT (Soil and Water Assessment Tool) modelers share, execute and analyze SWAT models using high performance computing resources. Making these two systems interoperable required common sign-in through OAuth, sharing of models through common metadata standards and use of standard web-services for implementing key import/export functionalities. As a result, users from either community can leverage the resources and services across these systems without having to manually importing, exporting, or processing their models. Overall, this use case is an example that can serve as a model for the interoperability among other systems as no one system can provide all the functionality needed to address large interdisciplinary problems.
Franke, O. Lehn; Reilly, Thomas E.; Bennett, Gordon D.
1987-01-01
Accurate definition of boundary and initial conditions is an essential part of conceptualizing and modeling ground-water flow systems. This report describes the properties of the seven most common boundary conditions encountered in ground-water systems and discusses major aspects of their application. It also discusses the significance and specification of initial conditions and evaluates some common errors in applying this concept to ground-water-system models. An appendix is included that discusses what the solution of a differential equation represents and how the solution relates to the boundary conditions defining the specific problem. This report considers only boundary conditions that apply to saturated ground-water systems.
Modeling Common-Sense Decisions in Artificial Intelligence
NASA Technical Reports Server (NTRS)
Zak, Michail
2010-01-01
A methodology has been conceived for efficient synthesis of dynamical models that simulate common-sense decision- making processes. This methodology is intended to contribute to the design of artificial-intelligence systems that could imitate human common-sense decision making or assist humans in making correct decisions in unanticipated circumstances. This methodology is a product of continuing research on mathematical models of the behaviors of single- and multi-agent systems known in biology, economics, and sociology, ranging from a single-cell organism at one extreme to the whole of human society at the other extreme. Earlier results of this research were reported in several prior NASA Tech Briefs articles, the three most recent and relevant being Characteristics of Dynamics of Intelligent Systems (NPO -21037), NASA Tech Briefs, Vol. 26, No. 12 (December 2002), page 48; Self-Supervised Dynamical Systems (NPO-30634), NASA Tech Briefs, Vol. 27, No. 3 (March 2003), page 72; and Complexity for Survival of Living Systems (NPO- 43302), NASA Tech Briefs, Vol. 33, No. 7 (July 2009), page 62. The methodology involves the concepts reported previously, albeit viewed from a different perspective. One of the main underlying ideas is to extend the application of physical first principles to the behaviors of living systems. Models of motor dynamics are used to simulate the observable behaviors of systems or objects of interest, and models of mental dynamics are used to represent the evolution of the corresponding knowledge bases. For a given system, the knowledge base is modeled in the form of probability distributions and the mental dynamics is represented by models of the evolution of the probability densities or, equivalently, models of flows of information. Autonomy is imparted to the decisionmaking process by feedback from mental to motor dynamics. This feedback replaces unavailable external information by information stored in the internal knowledge base. Representation of the dynamical models in a parameterized form reduces the task of common-sense-based decision making to a solution of the following hetero-associated-memory problem: store a set of m predetermined stochastic processes given by their probability distributions in such a way that when presented with an unexpected change in the form of an input out of the set of M inputs, the coupled motormental dynamics converges to the corresponding one of the m pre-assigned stochastic process, and a sample of this process represents the decision.
Polya's bees: A model of decentralized decision-making.
Golman, Russell; Hagmann, David; Miller, John H
2015-09-01
How do social systems make decisions with no single individual in control? We observe that a variety of natural systems, including colonies of ants and bees and perhaps even neurons in the human brain, make decentralized decisions using common processes involving information search with positive feedback and consensus choice through quorum sensing. We model this process with an urn scheme that runs until hitting a threshold, and we characterize an inherent tradeoff between the speed and the accuracy of a decision. The proposed common mechanism provides a robust and effective means by which a decentralized system can navigate the speed-accuracy tradeoff and make reasonably good, quick decisions in a variety of environments. Additionally, consensus choice exhibits systemic risk aversion even while individuals are idiosyncratically risk-neutral. This too is adaptive. The model illustrates how natural systems make decentralized decisions, illuminating a mechanism that engineers of social and artificial systems could imitate.
Polya’s bees: A model of decentralized decision-making
Golman, Russell; Hagmann, David; Miller, John H.
2015-01-01
How do social systems make decisions with no single individual in control? We observe that a variety of natural systems, including colonies of ants and bees and perhaps even neurons in the human brain, make decentralized decisions using common processes involving information search with positive feedback and consensus choice through quorum sensing. We model this process with an urn scheme that runs until hitting a threshold, and we characterize an inherent tradeoff between the speed and the accuracy of a decision. The proposed common mechanism provides a robust and effective means by which a decentralized system can navigate the speed-accuracy tradeoff and make reasonably good, quick decisions in a variety of environments. Additionally, consensus choice exhibits systemic risk aversion even while individuals are idiosyncratically risk-neutral. This too is adaptive. The model illustrates how natural systems make decentralized decisions, illuminating a mechanism that engineers of social and artificial systems could imitate. PMID:26601255
Object-Oriented Technology-Based Software Library for Operations of Water Reclamation Centers
NASA Astrophysics Data System (ADS)
Otani, Tetsuo; Shimada, Takehiro; Yoshida, Norio; Abe, Wataru
SCADA systems in water reclamation centers have been constructed based on hardware and software that each manufacturer produced according to their design. Even though this approach used to be effective to realize real-time and reliable execution, it is an obstacle to cost reduction about system construction and maintenance. A promising solution to address the problem is to set specifications that can be used commonly. In terms of software, information model approach has been adopted in SCADA systems in other field, such as telecommunications and power systems. An information model is a piece of software specification that describes a physical or logical object to be monitored. In this paper, we propose information models for operations of water reclamation centers, which have not ever existed. In addition, we show the feasibility of the information model in terms of common use and processing performance.
NASA Astrophysics Data System (ADS)
Malard, J. J.; Adamowski, J. F.; Wang, L. Y.; Rojas, M.; Carrera, J.; Gálvez, J.; Tuy, H. A.; Melgar-Quiñonez, H.
2015-12-01
The modelling of the impacts of climate change on agriculture requires the inclusion of socio-economic factors. However, while cropping models and economic models of agricultural systems are common, dynamically coupled socio-economic-biophysical models have not received as much success. A promising methodology for modelling the socioeconomic aspects of coupled natural-human systems is participatory system dynamics modelling, in which stakeholders develop mental maps of the socio-economic system that are then turned into quantified simulation models. This methodology has been successful in the water resources management field. However, while the stocks and flows of water resources have also been represented within the system dynamics modelling framework and thus coupled to the socioeconomic portion of the model, cropping models are ill-suited for such reformulation. In addition, most of these system dynamics models were developed without stakeholder input, limiting the scope for the adoption and implementation of their results. We therefore propose a new methodology for the analysis of climate change variability on agroecosystems which uses dynamically coupled system dynamics (socio-economic) and biophysical (cropping) models to represent both physical and socioeconomic aspects of the agricultural system, using two case studies (intensive market-based agricultural development versus subsistence crop-based development) from rural Guatemala. The system dynamics model component is developed with relevant governmental and NGO stakeholders from rural and agricultural development in the case study regions and includes such processes as education, poverty and food security. Common variables with the cropping models (yield and agricultural management choices) are then used to dynamically couple the two models together, allowing for the analysis of the agroeconomic system's response to and resilience against various climatic and socioeconomic shocks.
Expert system development for commonality analysis in space programs
NASA Technical Reports Server (NTRS)
Yeager, Dorian P.
1987-01-01
This report is a combination of foundational mathematics and software design. A mathematical model of the Commonality Analysis problem was developed and some important properties discovered. The complexity of the problem is described herein and techniques, both deterministic and heuristic, for reducing that complexity are presented. Weaknesses are pointed out in the existing software (System Commonality Analysis Tool) and several improvements are recommended. It is recommended that: (1) an expert system for guiding the design of new databases be developed; (2) a distributed knowledge base be created and maintained for the purpose of encoding the commonality relationships between design items in commonality databases; (3) a software module be produced which automatically generates commonality alternative sets from commonality databases using the knowledge associated with those databases; and (4) a more complete commonality analysis module be written which is capable of generating any type of feasible solution.
Conflicting Epistemologies and Inference in Coupled Human and Natural Systems
NASA Astrophysics Data System (ADS)
Garcia, M. E.
2017-12-01
Last year, I presented a model that projects per capita water consumption based on changes in price, population, building codes, and water stress salience. This model applied methods from hydrological science and engineering to relationships both within and beyond their traditional scope. Epistemologically, the development of mathematical models of natural or engineered systems is objectivist while research examining relationships between observations, perceptions and action is commonly constructivist or subjectivist. Drawing on multiple epistemologies is common in, and perhaps central to, the growing fields of coupled human and natural systems, and socio-hydrology. Critically, these philosophical perspectives vary in their view of the nature of the system as mechanistic, adaptive or constructed, and the split between aleatory and epistemic uncertainty. Interdisciplinary research is commonly cited as a way to address the critical and domain crossing challenge of sustainability as synthesis across perspectives can offer a more comprehensive view of system dynamics. However, combining methods and concepts from multiple ontologies and epistemologies can introduce contradictions into the logic of inference. These contractions challenge the evaluation of research products and the implications for practical application of research findings are not fully understood. Reflections on the evaluation, application, and generalization of the water consumption model described above are used to ground these broader questions and offer thoughts on the way forward.
Ganji-Arjenaki, Mahboube; Nasri, Hamid; Rafieian-Kopaei, Mahmoud
2017-07-01
The extra-intestinal manifestations of inflammatory bowel disease (IBD) are common and involve other organs or systems for example; urinary system. For this review, we used a variety of sources by searching through Web of Science, PubMed, EMBASE, Scopus and directory of open access journals (DOAJ). Urinary complications may occur in up to 22% of patients and nephrolithiasis or renal/kidney stones have been suggested to be a common manifestation of disease in forms of uric acid, calcium phosphate or calcium oxalate. We performed a meta-analysis on five clinical trials and reported that correlation between IBD and formation of stone in renal system is positive and significant (Fix-effect model; CI: 95%, P <0.001, and randomeffect model; CI: 95%, P = 0.03). Based on the reports of the clinical trials, calcium oxalate is more prevalent in Crohn's disease (CD) than in ulcerative colitis (UC).
Student Modeling Based on Problem Solving Times
ERIC Educational Resources Information Center
Pelánek, Radek; Jarušek, Petr
2015-01-01
Student modeling in intelligent tutoring systems is mostly concerned with modeling correctness of students' answers. As interactive problem solving activities become increasingly common in educational systems, it is useful to focus also on timing information associated with problem solving. We argue that the focus on timing is natural for certain…
System and method of designing models in a feedback loop
Gosink, Luke C.; Pulsipher, Trenton C.; Sego, Landon H.
2017-02-14
A method and system for designing models is disclosed. The method includes selecting a plurality of models for modeling a common event of interest. The method further includes aggregating the results of the models and analyzing each model compared to the aggregate result to obtain comparative information. The method also includes providing the information back to the plurality of models to design more accurate models through a feedback loop.
Statechart Analysis with Symbolic PathFinder
NASA Technical Reports Server (NTRS)
Pasareanu, Corina S.
2012-01-01
We report here on our on-going work that addresses the automated analysis and test case generation for software systems modeled using multiple Statechart formalisms. The work is motivated by large programs such as NASA Exploration, that involve multiple systems that interact via safety-critical protocols and are designed with different Statechart variants. To verify these safety-critical systems, we have developed Polyglot, a framework for modeling and analysis of model-based software written using different Statechart formalisms. Polyglot uses a common intermediate representation with customizable Statechart semantics and leverages the analysis and test generation capabilities of the Symbolic PathFinder tool. Polyglot is used as follows: First, the structure of the Statechart model (expressed in Matlab Stateflow or Rational Rhapsody) is translated into a common intermediate representation (IR). The IR is then translated into Java code that represents the structure of the model. The semantics are provided as "pluggable" modules.
Common evolutionary trends underlie the four-bar linkage systems of sunfish and mantis shrimp.
Hu, Yinan; Nelson-Maney, Nathan; Anderson, Philip S L
2017-05-01
Comparative biomechanics offers an opportunity to explore the evolution of disparate biological systems that share common underlying mechanics. Four-bar linkage modeling has been applied to various biological systems such as fish jaws and crustacean appendages to explore the relationship between biomechanics and evolutionary diversification. Mechanical sensitivity states that the functional output of a mechanical system will show differential sensitivity to changes in specific morphological components. We document similar patterns of mechanical sensitivity in two disparate four-bar systems from different phyla: the opercular four-bar system in centrarchid fishes and the raptorial appendage of stomatopods. We built dynamic linkage models of 19 centrarchid and 36 stomatopod species and used phylogenetic generalized least squares regression (PGLS) to compare evolutionary shifts in linkage morphology and mechanical outputs derived from the models. In both systems, the kinematics of the four-bar mechanism show significant evolutionary correlation with the output link, while travel distance of the output arm is correlated with the coupler link. This common evolutionary pattern seen in both fish and crustacean taxa is a potential consequence of the mechanical principles underlying four-bar systems. Our results illustrate the potential influence of physical principles on morphological evolution across biological systems with different structures, behaviors, and ecologies. © 2017 The Author(s). Evolution © 2017 The Society for the Study of Evolution.
Sloppy-model universality class and the Vandermonde matrix.
Waterfall, Joshua J; Casey, Fergal P; Gutenkunst, Ryan N; Brown, Kevin S; Myers, Christopher R; Brouwer, Piet W; Elser, Veit; Sethna, James P
2006-10-13
In a variety of contexts, physicists study complex, nonlinear models with many unknown or tunable parameters to explain experimental data. We explain why such systems so often are sloppy: the system behavior depends only on a few "stiff" combinations of the parameters and is unchanged as other "sloppy" parameter combinations vary by orders of magnitude. We observe that the eigenvalue spectra for the sensitivity of sloppy models have a striking, characteristic form with a density of logarithms of eigenvalues which is roughly constant over a large range. We suggest that the common features of sloppy models indicate that they may belong to a common universality class. In particular, we motivate focusing on a Vandermonde ensemble of multiparameter nonlinear models and show in one limit that they exhibit the universal features of sloppy models.
Extending the granularity of representation and control for the MIL-STD CAIS 1.0 node model
NASA Technical Reports Server (NTRS)
Rogers, Kathy L.
1986-01-01
The Common APSE (Ada 1 Program Support Environment) Interface Set (CAIS) (DoD85) node model provides an excellent baseline for interfaces in a single-host development environment. To encompass the entire spectrum of computing, however, the CAIS model should be extended in four areas. It should provide the interface between the engineering workstation and the host system throughout the entire lifecycle of the system. It should provide a basis for communication and integration functions needed by distributed host environments. It should provide common interfaces for communications mechanisms to and among target processors. It should provide facilities for integration, validation, and verification of test beds extending to distributed systems on geographically separate processors with heterogeneous instruction set architectures (ISAS). Additions to the PROCESS NODE model to extend the CAIS into these four areas are proposed.
A Model Plant for a Biology Curriculum: Spider Flower ("Cleome Hasslerana L.")
ERIC Educational Resources Information Center
Marquard, Robert D.; Steinback, Rebecca
2009-01-01
Major advances in fundamental science are developed using model systems. Classic examples of model systems include Mendel's work with the common garden pea ("Pisium sativa"), classic inheritance work by Morgan with the fruit fly ("Drosophila"), developmental studies with the nematode ("C. elegans"), and transposable elements in maize ("Zea…
Waran, V; Pancharatnam, Devaraj; Thambinayagam, Hari Chandran; Raman, Rajagopal; Rathinam, Alwin Kumar; Balakrishnan, Yuwaraj Kumar; Tung, Tan Su; Rahman, Z A
2014-01-01
Navigation in neurosurgery has expanded rapidly; however, suitable models to train end users to use the myriad software and hardware that come with these systems are lacking. Utilizing three-dimensional (3D) industrial rapid prototyping processes, we have been able to create models using actual computed tomography (CT) data from patients with pathology and use these models to simulate a variety of commonly performed neurosurgical procedures with navigation systems. To assess the possibility of utilizing models created from CT scan dataset obtained from patients with cranial pathology to simulate common neurosurgical procedures using navigation systems. Three patients with pathology were selected (hydrocephalus, right frontal cortical lesion, and midline clival meningioma). CT scan data following an image-guidance surgery protocol in DIACOM format and a Rapid Prototyping Machine were taken to create the necessary printed model with the corresponding pathology embedded. The ability in registration, planning, and navigation of two navigation systems using a variety of software and hardware provided by these platforms was assessed. We were able to register all models accurately using both navigation systems and perform the necessary simulations as planned. Models with pathology utilizing 3D rapid prototyping techniques accurately reflect data of actual patients and can be used in the simulation of neurosurgical operations using navigation systems. Georg Thieme Verlag KG Stuttgart · New York.
Learning in a Simple Motor System
ERIC Educational Resources Information Center
Broussard, Dianne M.; Kassardjian, Charles D.
2004-01-01
Motor learning is a very basic, essential form of learning that appears to share common mechanisms across different motor systems. We evaluate and compare a few conceptual models for learning in a relatively simple neural system, the vestibulo-ocular reflex (VOR) of vertebrates. We also compare the different animal models that have been used to…
ERIC Educational Resources Information Center
Guevara, Porfirio
2014-01-01
This article identifies elements and connections that seem to be relevant to explain persistent aggregate behavioral patterns in educational systems when using complex dynamical systems modeling and simulation approaches. Several studies have shown what factors are at play in educational fields, but confusion still remains about the underlying…
The Metadata Cloud: The Last Piece of a Distributed Data System Model
NASA Astrophysics Data System (ADS)
King, T. A.; Cecconi, B.; Hughes, J. S.; Walker, R. J.; Roberts, D.; Thieman, J. R.; Joy, S. P.; Mafi, J. N.; Gangloff, M.
2012-12-01
Distributed data systems have existed ever since systems were networked together. Over the years the model for distributed data systems have evolved from basic file transfer to client-server to multi-tiered to grid and finally to cloud based systems. Initially metadata was tightly coupled to the data either by embedding the metadata in the same file containing the data or by co-locating the metadata in commonly named files. As the sources of data multiplied, data volumes have increased and services have specialized to improve efficiency; a cloud system model has emerged. In a cloud system computing and storage are provided as services with accessibility emphasized over physical location. Computation and data clouds are common implementations. Effectively using the data and computation capabilities requires metadata. When metadata is stored separately from the data; a metadata cloud is formed. With a metadata cloud information and knowledge about data resources can migrate efficiently from system to system, enabling services and allowing the data to remain efficiently stored until used. This is especially important with "Big Data" where movement of the data is limited by bandwidth. We examine how the metadata cloud completes a general distributed data system model, how standards play a role and relate this to the existing types of cloud computing. We also look at the major science data systems in existence and compare each to the generalized cloud system model.
Towards a Framework for Modeling Space Systems Architectures
NASA Technical Reports Server (NTRS)
Shames, Peter; Skipper, Joseph
2006-01-01
Topics covered include: 1) Statement of the problem: a) Space system architecture is complex; b) Existing terrestrial approaches must be adapted for space; c) Need a common architecture methodology and information model; d) Need appropriate set of viewpoints. 2) Requirements on a space systems model. 3) Model Based Engineering and Design (MBED) project: a) Evaluated different methods; b) Adapted and utilized RASDS & RM-ODP; c) Identified useful set of viewpoints; d) Did actual model exchanges among selected subset of tools. 4) Lessons learned & future vision.
Model systems for the study of Enterococcal colonization and infection
Goh, H. M. Sharon; Yong, M. H. Adeline; Chong, Kelvin Kian Long
2017-01-01
ABSTRACT Enterococcus faecalis and Enterococcus faecium are common inhabitants of the human gastrointestinal tract, as well as frequent opportunistic pathogens. Enterococci cause a range of infections including, most frequently, infections of the urinary tract, catheterized urinary tract, bloodstream, wounds and surgical sites, and heart valves in endocarditis. Enterococcal infections are often biofilm-associated, polymicrobial in nature, and resistant to antibiotics of last resort. Understanding Enterococcal mechanisms of colonization and pathogenesis are important for identifying new ways to manage and intervene with these infections. We review vertebrate and invertebrate model systems applied to study the most common E. faecalis and E. faecium infections, with emphasis on recent findings examining Enterococcal-host interactions using these models. We discuss strengths and shortcomings of each model, propose future animal models not yet applied to study mono- and polymicrobial infections involving E. faecalis and E. faecium, and comment on the significance of anti-virulence strategies derived from a fundamental understanding of host-pathogen interactions in model systems. PMID:28102784
Noisy Spins and the Richardson-Gaudin Model
NASA Astrophysics Data System (ADS)
Rowlands, Daniel A.; Lamacraft, Austen
2018-03-01
We study a system of spins (qubits) coupled to a common noisy environment, each precessing at its own frequency. The correlated noise experienced by the spins implies long-lived correlations that relax only due to the differing frequencies. We use a mapping to a non-Hermitian integrable Richardson-Gaudin model to find the exact spectrum of the quantum master equation in the high-temperature limit and, hence, determine the decay rate. Our solution can be used to evaluate the effect of inhomogeneous splittings on a system of qubits coupled to a common bath.
Orbital Noise in the Earth System is a Common Cause of Climate and Greenhouse-Gas Fluctuation
NASA Technical Reports Server (NTRS)
Liu, H. S.; Kolenkiewicz, R.; Wade, C., Jr.; Smith, David E. (Technical Monitor)
2002-01-01
The mismatch between fossil isotopic data and climate models known as the cool-tropic paradox implies that either the data are flawed or we understand very little about the climate models of greenhouse warming. Here we question the validity of the climate models on the scientific background of orbital noise in the Earth system. Our study shows that the insolation pulsation induced by orbital noise is the common cause of climate change and atmospheric concentrations of carbon dioxide and methane. In addition, we find that the intensity of the insolation pulses is dependent on the latitude of the Earth. Thus, orbital noise is the key to understanding the troubling paradox in climate models.
Karam, Eli A; Blow, Adrian J; Sprenkle, Douglas H; Davis, Sean D
2015-04-01
Specific models guide the training of marriage and family therapists (MFTs) as they offer both structure and organization for both therapists and clients. Learning models may also benefit therapists-in-training by instilling confidence and preventing atheoretical eclecticism. The moderate common factors perspective argues that models are essential, but should not be taught as "the absolute truth," given there is no evidence for relative efficacy of one empirically validated model versus another, and no single model works in all instances. The following article provides a blueprint for infusing a common factors perspective into MFT programmes by reviewing innovations in course design, outlining specific teaching strategies, and highlighting potential implementation challenges. © 2014 American Association for Marriage and Family Therapy.
Synthetic Biology Outside the Cell: Linking Computational Tools to Cell-Free Systems
DOE Office of Scientific and Technical Information (OSTI.GOV)
Lewis, Daniel D.; Department of Biomedical Engineering, University of California Davis, Davis, CA; Villarreal, Fernando D.
As mathematical models become more commonly integrated into the study of biology, a common language for describing biological processes is manifesting. Many tools have emerged for the simulation of in vivo synthetic biological systems, with only a few examples of prominent work done on predicting the dynamics of cell-free synthetic systems. At the same time, experimental biologists have begun to study dynamics of in vitro systems encapsulated by amphiphilic molecules, opening the door for the development of a new generation of biomimetic systems. In this review, we explore both in vivo and in vitro models of biochemical networks with amore » special focus on tools that could be applied to the construction of cell-free expression systems. We believe that quantitative studies of complex cellular mechanisms and pathways in synthetic systems can yield important insights into what makes cells different from conventional chemical systems.« less
Synthetic Biology Outside the Cell: Linking Computational Tools to Cell-Free Systems
Lewis, Daniel D.; Villarreal, Fernando D.; Wu, Fan; Tan, Cheemeng
2014-01-01
As mathematical models become more commonly integrated into the study of biology, a common language for describing biological processes is manifesting. Many tools have emerged for the simulation of in vivo synthetic biological systems, with only a few examples of prominent work done on predicting the dynamics of cell-free synthetic systems. At the same time, experimental biologists have begun to study dynamics of in vitro systems encapsulated by amphiphilic molecules, opening the door for the development of a new generation of biomimetic systems. In this review, we explore both in vivo and in vitro models of biochemical networks with a special focus on tools that could be applied to the construction of cell-free expression systems. We believe that quantitative studies of complex cellular mechanisms and pathways in synthetic systems can yield important insights into what makes cells different from conventional chemical systems. PMID:25538941
Synthetic biology outside the cell: linking computational tools to cell-free systems.
Lewis, Daniel D; Villarreal, Fernando D; Wu, Fan; Tan, Cheemeng
2014-01-01
As mathematical models become more commonly integrated into the study of biology, a common language for describing biological processes is manifesting. Many tools have emerged for the simulation of in vivo synthetic biological systems, with only a few examples of prominent work done on predicting the dynamics of cell-free synthetic systems. At the same time, experimental biologists have begun to study dynamics of in vitro systems encapsulated by amphiphilic molecules, opening the door for the development of a new generation of biomimetic systems. In this review, we explore both in vivo and in vitro models of biochemical networks with a special focus on tools that could be applied to the construction of cell-free expression systems. We believe that quantitative studies of complex cellular mechanisms and pathways in synthetic systems can yield important insights into what makes cells different from conventional chemical systems.
A new necessary condition for Turing instabilities.
Elragig, Aiman; Townley, Stuart
2012-09-01
Reactivity (a.k.a initial growth) is necessary for diffusion driven instability (Turing instability). Using a notion of common Lyapunov function we show that this necessary condition is a special case of a more powerful (i.e. tighter) necessary condition. Specifically, we show that if the linearised reaction matrix and the diffusion matrix share a common Lyapunov function, then Turing instability is not possible. The existence of common Lyapunov functions is readily checked using semi-definite programming. We apply this result to the Gierer-Meinhardt system modelling regenerative properties of Hydra, the Oregonator, to a host-parasite-hyperparasite system with diffusion and to a reaction-diffusion-chemotaxis model for a multi-species host-parasitoid community. Copyright © 2012 Elsevier Inc. All rights reserved.
NASA Astrophysics Data System (ADS)
Zakharova, Alexandra A.; Kolegova, Olga A.; Nekrasova, Maria E.
2016-04-01
The paper deals with the issues in program management used for engineering innovative products. The existing project management tools were analyzed. The aim is to develop a decision support system that takes into account the features of program management used for high-tech products: research intensity, a high level of technical risks, unpredictable results due to the impact of various external factors, availability of several implementing agencies. The need for involving experts and using intelligent techniques for information processing is demonstrated. A conceptual model of common information space to support communication between members of the collaboration on high-tech programs has been developed. The structure and objectives of the information analysis system “Geokhod” were formulated with the purpose to implement the conceptual model of common information space in the program “Development and production of new class mining equipment - “Geokhod”.
ERIC Educational Resources Information Center
Chard, David J.
2013-01-01
The majority of school districts implementing response to intervention use a systemwide, multitier model of delivery. This article describes the common features of multitier models and discusses the emerging evidence of their effectiveness. In addition, specific factors that schools should consider to enhance effective implementation of systemic,…
2013-01-01
Background Histopathology has initially been and is still used to diagnose infectious, degenerative or neoplastic diseases in humans or animals. In addition to qualitative diagnoses semiquantitative scoring of a lesion`s magnitude on an ordinal scale is a commonly demanded task for histopathologists. Multiparametric, semiquantitative scoring systems for mouse models histopathology are a common approach to handle these questions and to include histopathologic information in biomedical research. Results Inclusion criteria for scoring systems were a first description of a multiparametric, semiquantiative scoring systems which comprehensibly describe an approach to evaluate morphologic lesion. A comprehensive literature search using these criteria identified 153 originally designed semiquantitative scoring systems for the analysis of morphologic changes in mouse models covering almost all organs systems and a wide variety of disease models. Of these, colitis, experimental autoimmune encephalitis, lupus nephritis and collagen induced osteoarthritis colitis were the disease models with the largest number of different scoring systems. Closer analysis of the identified scoring systems revealed a lack of a rationale for the selection of the scoring parameters or a correlation between scoring parameter value and the magnitude of the clinical symptoms in most studies. Conclusion Although a decision for a particular scoring system is clearly dependent on the respective scientific question this review gives an overview on currently available systems and may therefore allow for a better choice for the respective project. PMID:23800279
An Update on the Conceptual-Production Systems Model of Apraxia: Evidence from Stroke
ERIC Educational Resources Information Center
Stamenova, Vessela; Black, Sandra E.; Roy, Eric A.
2012-01-01
Limb apraxia is a neurological disorder characterized by an inability to pantomime and/or imitate gestures. It is more commonly observed after left hemisphere damage (LHD), but has also been reported after right hemisphere damage (RHD). The Conceptual-Production Systems model (Roy, 1996) suggests that three systems are involved in the control of…
A remark on the GNSS single difference model with common clock scheme for attitude determination
NASA Astrophysics Data System (ADS)
Chen, Wantong
2016-09-01
GNSS-based attitude determination technique is an important field of study, in which two schemes can be used to construct the actual system: the common clock scheme and the non-common clock scheme. Compared with the non-common clock scheme, the common clock scheme can strongly improve both the reliability and the accuracy. However, in order to gain these advantages, specific care must be taken in the implementation. The cares are thus discussed, based on the generating technique of carrier phase measurement in GNSS receivers. A qualitative assessment of potential phase bias contributes is also carried out. Possible technical difficulties are pointed out for the development of single-board multi-antenna GNSS attitude systems with a common clock.
Context-Aware Mobile Collaborative Systems: Conceptual Modeling and Case Study
Benítez-Guerrero, Edgard; Mezura-Godoy, Carmen; Montané-Jiménez, Luis G.
2012-01-01
A Mobile Collaborative System (MCOS) enable the cooperation of the members of a team to achieve a common goal by using a combination of mobile and fixed technologies. MCOS can be enhanced if the context of the group of users is considered in the execution of activities. This paper proposes a novel model for Context-Aware Mobile COllaborative Systems (CAMCOS) and a functional architecture based on that model. In order to validate both the model and the architecture, a prototype system in the tourism domain was implemented and evaluated. PMID:23202007
What can formal methods offer to digital flight control systems design
NASA Technical Reports Server (NTRS)
Good, Donald I.
1990-01-01
Formal methods research begins to produce methods which will enable mathematic modeling of the physical behavior of digital hardware and software systems. The development of these methods directly supports the NASA mission of increasing the scope and effectiveness of flight system modeling capabilities. The conventional, continuous mathematics that is used extensively in modeling flight systems is not adequate for accurate modeling of digital systems. Therefore, the current practice of digital flight control system design has not had the benefits of extensive mathematical modeling which are common in other parts of flight system engineering. Formal methods research shows that by using discrete mathematics, very accurate modeling of digital systems is possible. These discrete modeling methods will bring the traditional benefits of modeling to digital hardware and hardware design. Sound reasoning about accurate mathematical models of flight control systems can be an important part of reducing risk of unsafe flight control.
A Common Core for Active Conceptual Modeling for Learning from Surprises
NASA Astrophysics Data System (ADS)
Liddle, Stephen W.; Embley, David W.
The new field of active conceptual modeling for learning from surprises (ACM-L) may be helpful in preserving life, protecting property, and improving quality of life. The conceptual modeling community has developed sound theory and practices for conceptual modeling that, if properly applied, could help analysts model and predict more accurately. In particular, we need to associate more semantics with links, and we need fully reified high-level objects and relationships that have a clear, formal underlying semantics that follows a natural, ontological approach. We also need to capture more dynamic aspects in our conceptual models to more accurately model complex, dynamic systems. These concepts already exist, and the theory is well developed; what remains is to link them with the ideas needed to predict system evolution, thus enabling risk assessment and response planning. No single researcher or research group will be able to achieve this ambitious vision alone. As a starting point, we recommend that the nascent ACM-L community agree on a common core model that supports all aspects—static and dynamic—needed for active conceptual modeling in support of learning from surprises. A common core will more likely gain the traction needed to sustain the extended ACM-L research effort that will yield the advertised benefits of learning from surprises.
NASA Astrophysics Data System (ADS)
Kutsch, W. L.; Zhao, Z.; Hardisty, A.; Hellström, M.; Chin, Y.; Magagna, B.; Asmi, A.; Papale, D.; Pfeil, B.; Atkinson, M.
2017-12-01
Environmental Research Infrastructures (ENVRIs) are expected to become important pillars not only for supporting their own scientific communities, but also a) for inter-disciplinary research and b) for the European Earth Observation Program Copernicus as a contribution to the Global Earth Observation System of Systems (GEOSS) or global thematic data networks. As such, it is very important that data-related activities of the ENVRIs will be well integrated. This requires common policies, models and e-infrastructure to optimise technological implementation, define workflows, and ensure coordination, harmonisation, integration and interoperability of data, applications and other services. The key is interoperating common metadata systems (utilising a richer metadata model as the `switchboard' for interoperation with formal syntax and declared semantics). The metadata characterises data, services, users and ICT resources (including sensors and detectors). The European Cluster Project ENVRIplus has developed a reference model (ENVRI RM) for common data infrastructure architecture to promote interoperability among ENVRIs. The presentation will provide an overview of recent progress and give examples for the integration of ENVRI data in global integration networks.
NASA Astrophysics Data System (ADS)
Cenek, Martin; Dahl, Spencer K.
2016-11-01
Systems with non-linear dynamics frequently exhibit emergent system behavior, which is important to find and specify rigorously to understand the nature of the modeled phenomena. Through this analysis, it is possible to characterize phenomena such as how systems assemble or dissipate and what behaviors lead to specific final system configurations. Agent Based Modeling (ABM) is one of the modeling techniques used to study the interaction dynamics between a system's agents and its environment. Although the methodology of ABM construction is well understood and practiced, there are no computational, statistically rigorous, comprehensive tools to evaluate an ABM's execution. Often, a human has to observe an ABM's execution in order to analyze how the ABM functions, identify the emergent processes in the agent's behavior, or study a parameter's effect on the system-wide behavior. This paper introduces a new statistically based framework to automatically analyze agents' behavior, identify common system-wide patterns, and record the probability of agents changing their behavior from one pattern of behavior to another. We use network based techniques to analyze the landscape of common behaviors in an ABM's execution. Finally, we test the proposed framework with a series of experiments featuring increasingly emergent behavior. The proposed framework will allow computational comparison of ABM executions, exploration of a model's parameter configuration space, and identification of the behavioral building blocks in a model's dynamics.
Cenek, Martin; Dahl, Spencer K
2016-11-01
Systems with non-linear dynamics frequently exhibit emergent system behavior, which is important to find and specify rigorously to understand the nature of the modeled phenomena. Through this analysis, it is possible to characterize phenomena such as how systems assemble or dissipate and what behaviors lead to specific final system configurations. Agent Based Modeling (ABM) is one of the modeling techniques used to study the interaction dynamics between a system's agents and its environment. Although the methodology of ABM construction is well understood and practiced, there are no computational, statistically rigorous, comprehensive tools to evaluate an ABM's execution. Often, a human has to observe an ABM's execution in order to analyze how the ABM functions, identify the emergent processes in the agent's behavior, or study a parameter's effect on the system-wide behavior. This paper introduces a new statistically based framework to automatically analyze agents' behavior, identify common system-wide patterns, and record the probability of agents changing their behavior from one pattern of behavior to another. We use network based techniques to analyze the landscape of common behaviors in an ABM's execution. Finally, we test the proposed framework with a series of experiments featuring increasingly emergent behavior. The proposed framework will allow computational comparison of ABM executions, exploration of a model's parameter configuration space, and identification of the behavioral building blocks in a model's dynamics.
NASA Astrophysics Data System (ADS)
Cai, Le; Mao, Xiaobing; Ma, Zhexuan
2018-02-01
This study first constructed the nonlinear mathematical model of the high-pressure common rail (HPCR) system in the diesel engine. Then, the nonlinear state transformation was performed using the flow’s calculation and the standard state space equation was acquired. Based on sliding-mode variable structure control (SMVSC) theory, a sliding-mode controller for nonlinear systems was designed for achieving the control of common rail pressure and the diesel engine’s rotational speed. Finally, on the simulation platform of MATLAB, the designed nonlinear HPCR system was simulated. The simulation results demonstrate that sliding-mode variable structure control algorithm shows favorable control performances and overcome the shortcomings of traditional PID control in overshoot, parameter adjustment, system precision, adjustment time and ascending time.
Age effects on explicit and implicit memory
Ward, Emma V.; Berry, Christopher J.; Shanks, David R.
2013-01-01
It is well-documented that explicit memory (e.g., recognition) declines with age. In contrast, many argue that implicit memory (e.g., priming) is preserved in healthy aging. For example, priming on tasks such as perceptual identification is often not statistically different in groups of young and older adults. Such observations are commonly taken as evidence for distinct explicit and implicit learning/memory systems. In this article we discuss several lines of evidence that challenge this view. We describe how patterns of differential age-related decline may arise from differences in the ways in which the two forms of memory are commonly measured, and review recent research suggesting that under improved measurement methods, implicit memory is not age-invariant. Formal computational models are of considerable utility in revealing the nature of underlying systems. We report the results of applying single and multiple-systems models to data on age effects in implicit and explicit memory. Model comparison clearly favors the single-system view. Implications for the memory systems debate are discussed. PMID:24065942
Phillips, Joshua; Chilukuri, Ram; Fragoso, Gilberto; Warzel, Denise; Covitz, Peter A
2006-01-06
Robust, programmatically accessible biomedical information services that syntactically and semantically interoperate with other resources are challenging to construct. Such systems require the adoption of common information models, data representations and terminology standards as well as documented application programming interfaces (APIs). The National Cancer Institute (NCI) developed the cancer common ontologic representation environment (caCORE) to provide the infrastructure necessary to achieve interoperability across the systems it develops or sponsors. The caCORE Software Development Kit (SDK) was designed to provide developers both within and outside the NCI with the tools needed to construct such interoperable software systems. The caCORE SDK requires a Unified Modeling Language (UML) tool to begin the development workflow with the construction of a domain information model in the form of a UML Class Diagram. Models are annotated with concepts and definitions from a description logic terminology source using the Semantic Connector component. The annotated model is registered in the Cancer Data Standards Repository (caDSR) using the UML Loader component. System software is automatically generated using the Codegen component, which produces middleware that runs on an application server. The caCORE SDK was initially tested and validated using a seven-class UML model, and has been used to generate the caCORE production system, which includes models with dozens of classes. The deployed system supports access through object-oriented APIs with consistent syntax for retrieval of any type of data object across all classes in the original UML model. The caCORE SDK is currently being used by several development teams, including by participants in the cancer biomedical informatics grid (caBIG) program, to create compatible data services. caBIG compatibility standards are based upon caCORE resources, and thus the caCORE SDK has emerged as a key enabling technology for caBIG. The caCORE SDK substantially lowers the barrier to implementing systems that are syntactically and semantically interoperable by providing workflow and automation tools that standardize and expedite modeling, development, and deployment. It has gained acceptance among developers in the caBIG program, and is expected to provide a common mechanism for creating data service nodes on the data grid that is under development.
A service-oriented data access control model
NASA Astrophysics Data System (ADS)
Meng, Wei; Li, Fengmin; Pan, Juchen; Song, Song; Bian, Jiali
2017-01-01
The development of mobile computing, cloud computing and distributed computing meets the growing individual service needs. Facing with complex application system, it's an urgent problem to ensure real-time, dynamic, and fine-grained data access control. By analyzing common data access control models, on the basis of mandatory access control model, the paper proposes a service-oriented access control model. By regarding system services as subject and data of databases as object, the model defines access levels and access identification of subject and object, and ensures system services securely to access databases.
Building Dynamic Conceptual Physics Understanding
ERIC Educational Resources Information Center
Trout, Charlotte; Sinex, Scott A.; Ragan, Susan
2011-01-01
Models are essential to the learning and doing of science, and systems thinking is key to appreciating many environmental issues. The National Science Education Standards include models and systems in their unifying concepts and processes standard, while the AAAS Benchmarks include them in their common themes chapter. Hyerle and Marzano argue for…
Cervero-Aragó, Sílvia; Rodríguez-Martínez, Sarah; Puertas-Bennasar, Antoni; Araujo, Rosa M
2015-01-01
Chlorine and thermal treatments are the most commonly used procedures to control and prevent Legionella proliferation in drinking water systems of large buildings. However, cases of legionellosis still occur in facilities with treated water. The purpose of this work was to model the effect of temperature and free chlorine applied in similar exposure conditions as in drinking water systems on five Legionella spp. strains and two amoebal strains of the genera Acanthamoeba. Inactivation models obtained were used to determine the effectiveness of the treatments applied which resulted more effective against Legionella than Acanthamoeba, especially those in cystic stages. Furthermore, to determine the influence of the relationship between L. pneumophila and Acanthamoeba spp. on the treatment effectiveness, inactivation models of the bacteria-associated amoeba were also constructed and compared to the models obtained for the free living bacteria state. The Legionella-amoeba association did not change the inactivation models, but it reduced the effectiveness of the treatments applied. Remarkably, at the lowest free chlorine concentration, 0.5 mg L-1, as well as at the lowest temperatures, 50°C and 55°C, the influence of the Legionella-amoeba associate state was the strongest in reducing the effectiveness of the treatments compared to the free Legionella state. Therefore, the association established between L. pneumophila and amoebae in the water systems indicate an increased health risk in proximal areas of the system (close to the tap) where lower free chlorine concentrations and lower temperatures are commonly observed.
Common world model for unmanned systems
NASA Astrophysics Data System (ADS)
Dean, Robert Michael S.
2013-05-01
The Robotic Collaborative Technology Alliance (RCTA) seeks to provide adaptive robot capabilities which move beyond traditional metric algorithms to include cognitive capabilities. Key to this effort is the Common World Model, which moves beyond the state-of-the-art by representing the world using metric, semantic, and symbolic information. It joins these layers of information to define objects in the world. These objects may be reasoned upon jointly using traditional geometric, symbolic cognitive algorithms and new computational nodes formed by the combination of these disciplines. The Common World Model must understand how these objects relate to each other. Our world model includes the concept of Self-Information about the robot. By encoding current capability, component status, task execution state, and histories we track information which enables the robot to reason and adapt its performance using Meta-Cognition and Machine Learning principles. The world model includes models of how aspects of the environment behave, which enable prediction of future world states. To manage complexity, we adopted a phased implementation approach to the world model. We discuss the design of "Phase 1" of this world model, and interfaces by tracing perception data through the system from the source to the meta-cognitive layers provided by ACT-R and SS-RICS. We close with lessons learned from implementation and how the design relates to Open Architecture.
Plants as models for the study of human pathogenesis.
Guttman, David S
2004-05-01
There are many common disease mechanisms used by bacterial pathogens of plants and humans. They use common means of attachment, secretion and genetic regulation. They share many virulence factors, such as extracellular polysaccharides and some type III secreted effectors. Plant and human innate immune systems also share many similarities. Many of these shared bacterial virulence mechanisms are homologous, but even more appear to have independently converged on a common function. This combination of homologous and analogous systems reveals conserved and critical steps in the disease process. Given these similarities, and the many experimental advantages of plant biology, including ease of replication, stringent genetic and reproductive control, and high throughput with low cost, it is proposed that plants would make excellent models for the study of human pathogenesis.
The Secure Distributed Operating System Design Project
1988-06-01
a di- verse group of people . Its organization isolates different aspects of the project, such as expected results, preliminary results, and technical...modeled after these procedures. " Automation: computers are commonly used to automate tasks previously performed by people ; many of these tasks are... people commonly con- sidered the threats anticipated to the system and mechanisms that are used to prevent those threats. Both hardware and software
Using A Model-Based Systems Engineering Approach For Exploration Medical System Development
NASA Technical Reports Server (NTRS)
Hanson, A.; Mindock, J.; McGuire, K.; Reilly, J.; Cerro, J.; Othon, W.; Rubin, D.; Urbina, M.; Canga, M.
2017-01-01
NASA's Human Research Program's Exploration Medical Capabilities (ExMC) element is defining the medical system needs for exploration class missions. ExMC's Systems Engineering (SE) team will play a critical role in successful design and implementation of the medical system into exploration vehicles. The team's mission is to "Define, develop, validate, and manage the technical system design needed to implement exploration medical capabilities for Mars and test the design in a progression of proving grounds." Development of the medical system is being conducted in parallel with exploration mission architecture and vehicle design development. Successful implementation of the medical system in this environment will require a robust systems engineering approach to enable technical communication across communities to create a common mental model of the emergent engineering and medical systems. Model-Based Systems Engineering (MBSE) improves shared understanding of system needs and constraints between stakeholders and offers a common language for analysis. The ExMC SE team is using MBSE techniques to define operational needs, decompose requirements and architecture, and identify medical capabilities needed to support human exploration. Systems Modeling Language (SysML) is the specific language the SE team is utilizing, within an MBSE approach, to model the medical system functional needs, requirements, and architecture. Modeling methods are being developed through the practice of MBSE within the team, and tools are being selected to support meta-data exchange as integration points to other system models are identified. Use of MBSE is supporting the development of relationships across disciplines and NASA Centers to build trust and enable teamwork, enhance visibility of team goals, foster a culture of unbiased learning and serving, and be responsive to customer needs. The MBSE approach to medical system design offers a paradigm shift toward greater integration between vehicle and the medical system and directly supports the transition of Earth-reliant ISS operations to the Earth-independent operations envisioned for Mars. Here, we describe the methods and approach to building this integrated model.
Content Analysis in Systems Engineering Acquisition Activities
2016-04-30
Acquisition Activities Karen Holness, Assistant Professor, NPS Update on the Department of the Navy Systems Engineering Career Competency Model Clifford...systems engineering toolkit . Having a common analysis tool that is easy to use would support the feedback of observed system performance trends from the
Optimizing Automatic Deployment Using Non-functional Requirement Annotations
NASA Astrophysics Data System (ADS)
Kugele, Stefan; Haberl, Wolfgang; Tautschnig, Michael; Wechs, Martin
Model-driven development has become common practice in design of safety-critical real-time systems. High-level modeling constructs help to reduce the overall system complexity apparent to developers. This abstraction caters for fewer implementation errors in the resulting systems. In order to retain correctness of the model down to the software executed on a concrete platform, human faults during implementation must be avoided. This calls for an automatic, unattended deployment process including allocation, scheduling, and platform configuration.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Mendon, Vrushali V.; Taylor, Zachary T.
ABSTRACT: Recent advances in residential building energy efficiency and codes have resulted in increased interest in detailed residential building energy models using the latest energy simulation software. One of the challenges of developing residential building models to characterize new residential building stock is to allow for flexibility to address variability in house features like geometry, configuration, HVAC systems etc. Researchers solved this problem in a novel way by creating a simulation structure capable of creating fully-functional EnergyPlus batch runs using a completely scalable residential EnergyPlus template system. This system was used to create a set of thirty-two residential prototype buildingmore » models covering single- and multifamily buildings, four common foundation types and four common heating system types found in the United States (US). A weighting scheme with detailed state-wise and national weighting factors was designed to supplement the residential prototype models. The complete set is designed to represent a majority of new residential construction stock. The entire structure consists of a system of utility programs developed around the core EnergyPlus simulation engine to automate the creation and management of large-scale simulation studies with minimal human effort. The simulation structure and the residential prototype building models have been used for numerous large-scale studies, one of which is briefly discussed in this paper.« less
NASA Astrophysics Data System (ADS)
Matussek, Robert; Dzienis, Cezary; Blumschein, Jörg; Schulte, Horst
2014-12-01
In this paper, a generic enhanced protection current transformer (CT) model with saturation effects and transient behavior is presented. The model is used for the purpose of analysis and design of power system protection algorithms. Three major classes of protection CT have been modeled which all take into account the nonlinear inductance with remanence effects. The transient short-circuit currents in power systems are simulated under CT saturation condition. The response of a common power system protection algorithm with respect to robustness to nominal parameter variations and sensitivity against maloperation is demonstrated by simulation studies.
Bialas, Andrzej
2010-01-01
The paper is focused on the security issues of sensors provided with processors and software and used for high-risk applications. Common IT related threats may cause serious consequences for sensor system users. To improve their robustness, sensor systems should be developed in a restricted way that would provide them with assurance. One assurance creation methodology is Common Criteria (ISO/IEC 15408) used for IT products and systems. The paper begins with a primer on the Common Criteria, and then a general security model of the intelligent sensor as an IT product is discussed. The paper presents how the security problem of the intelligent sensor is defined and solved. The contribution of the paper is to provide Common Criteria (CC) related security design patterns and to improve the effectiveness of the sensor development process. PMID:22315571
NASA Technical Reports Server (NTRS)
Johnson, Paul K.; Mason, Lee S.
2006-01-01
This paper provides an analytical evaluation on the operation and performance of a dual Brayton common gas system. The NASA Glenn Research Center in-house computer program Closed Cycle System Simulation (CCSS) was used to construct a model of two identical 50 kWe-class recuperated closed-Brayton-cycle (CBC) power conversion units that share a common gas inventory and single heat source. As operating conditions for each CBC change, the total gas inventory is redistributed between the two units and overall system performance is affected. Several steady-state off-design operating points were analyzed by varying turbine inlet temperature and turbo-alternator shaft rotational speed to investigate the interaction of the two units.
In risk assessment there is a need to accelerate toxicological evaluation of vast numbers of chemicals. New programs focus on identifying common modes of action and on model systems for rapid screening. In this study we address both these issues. Oxidative stress is a good can...
Evaluating a common semi-mechanistic mathematical model of gene-regulatory networks
2015-01-01
Modeling and simulation of gene-regulatory networks (GRNs) has become an important aspect of modern systems biology investigations into mechanisms underlying gene regulation. A key challenge in this area is the automated inference (reverse-engineering) of dynamic, mechanistic GRN models from gene expression time-course data. Common mathematical formalisms for representing such models capture two aspects simultaneously within a single parameter: (1) Whether or not a gene is regulated, and if so, the type of regulator (activator or repressor), and (2) the strength of influence of the regulator (if any) on the target or effector gene. To accommodate both roles, "generous" boundaries or limits for possible values of this parameter are commonly allowed in the reverse-engineering process. This approach has several important drawbacks. First, in the absence of good guidelines, there is no consensus on what limits are reasonable. Second, because the limits may vary greatly among different reverse-engineering experiments, the concrete values obtained for the models may differ considerably, and thus it is difficult to compare models. Third, if high values are chosen as limits, the search space of the model inference process becomes very large, adding unnecessary computational load to the already complex reverse-engineering process. In this study, we demonstrate that restricting the limits to the [−1, +1] interval is sufficient to represent the essential features of GRN systems and offers a reduction of the search space without loss of quality in the resulting models. To show this, we have carried out reverse-engineering studies on data generated from artificial and experimentally determined from real GRN systems. PMID:26356485
Students' Use of the Energy Model to Account for Changes in Physical Systems
ERIC Educational Resources Information Center
Papadouris, Nico; Constantinou, Constantinos P.; Kyratsi, Theodora
2008-01-01
The aim of this study is to explore the ways in which students, aged 11-14 years, account for certain changes in physical systems and the extent to which they draw on an energy model as a common framework for explaining changes observed in diverse systems. Data were combined from two sources: interviews with 20 individuals and an open-ended…
IEA Wind Task 37: Systems Modeling Framework and Ontology for Wind Turbines and Plants
DOE Office of Scientific and Technical Information (OSTI.GOV)
Dykes, Katherine L; Zahle, Frederik; Merz, Karl
This presentation will provide an overview of progress to date in the development of a system modeling framework and ontology for wind turbines and plants as part of the larger IEA Wind Task 37 on wind energy systems engineering. The goals of the effort are to create a set of guidelines for a common conceptual architecture for wind turbines and plants so that practitioners can more easily share descriptions of wind turbines and plants across multiple parties and reduce the effort for translating descriptions between models; integrate different models together and collaborate on model development; and translate models among differentmore » levels of fidelity in the system.« less
The quest for solvable multistate Landau-Zener models
Sinitsyn, Nikolai A.; Chernyak, Vladimir Y.
2017-05-24
Recently, integrability conditions (ICs) in mutistate Landau-Zener (MLZ) theory were proposed. They describe common properties of all known solved systems with linearly time-dependent Hamiltonians. Here we show that ICs enable efficient computer assisted search for new solvable MLZ models that span complexity range from several interacting states to mesoscopic systems with many-body dynamics and combinatorially large phase space. This diversity suggests that nontrivial solvable MLZ models are numerous. Additionally, we refine the formulation of ICs and extend the class of solvable systems to models with points of multiple diabatic level crossing.
General Systems Theory and Instructional Systems Design.
ERIC Educational Resources Information Center
Salisbury, David F.
1990-01-01
Describes basic concepts in the field of general systems theory (GST) and identifies commonalities that exist between GST and instructional systems design (ISD). Models and diagrams that depict system elements in ISD are presented, and two matrices that show how GST has been used in ISD literature are included. (11 references) (LRW)
Intermolecular orbital interaction in π systems
NASA Astrophysics Data System (ADS)
Zhao, Rundong; Zhang, Rui-Qin
2018-04-01
Intermolecular interactions, in regard to which people tend to emphasise the noncovalent van der Waals (vdW) forces when conducting investigations throughout chemistry, can influence the structure, stability and function of molecules and materials. Despite the ubiquitous nature of vdW interactions, a simplified electrostatic model has been popularly adopted to explain common intermolecular interactions, especially those existing in π-involved systems. However, this classical model has come under fire in revealing specific issues such as substituent effects, due to its roughness; and it has been followed in past decades by sundry explanations which sometimes bring in nebulous descriptions. In this account, we try to summarise and present a unified model for describing and analysing the binding mechanism of such systems from the viewpoint of energy decomposition. We also emphasise a commonly ignored factor - orbital interaction, pointing out that the noncovalent intermolecular orbital interactions actually exhibit similar bonding and antibonding phenomena as those in covalent bonds.
A dual system model of preferences under risk.
Mukherjee, Kanchan
2010-01-01
This article presents a dual system model (DSM) of decision making under risk and uncertainty according to which the value of a gamble is a combination of the values assigned to it independently by the affective and deliberative systems. On the basis of research on dual process theories and empirical research in Hsee and Rottenstreich (2004) and Rottenstreich and Hsee (2001) among others, the DSM incorporates (a) individual differences in disposition to rational versus emotional decision making, (b) the affective nature of outcomes, and (c) different task construals within its framework. The model has good descriptive validity and accounts for (a) violation of nontransparent stochastic dominance, (b) fourfold pattern of risk attitudes, (c) ambiguity aversion, (d) common consequence effect, (e) common ratio effect, (f) isolation effect, and (g) coalescing and event-splitting effects. The DSM is also used to make several novel predictions of conditions under which specific behavior patterns may or may not occur.
AI and simulation: What can they learn from each other
NASA Technical Reports Server (NTRS)
Colombano, Silvano P.
1988-01-01
Simulation and Artificial Intelligence share a fertile common ground both from a practical and from a conceptual point of view. Strengths and weaknesses of both Knowledge Based System and Modeling and Simulation are examined and three types of systems that combine the strengths of both technologies are discussed. These types of systems are a practical starting point, however, the real strengths of both technologies will be exploited only when they are combined in a common knowledge representation paradigm. From an even deeper conceptual point of view, one might even argue that the ability to reason from a set of facts (i.e., Expert System) is less representative of human reasoning than the ability to make a model of the world, change it as required, and derive conclusions about the expected behavior of world entities. This is a fundamental problem in AI, and Modeling Theory can contribute to its solution. The application of Knowledge Engineering technology to a Distributed Processing Network Simulator (DPNS) is discussed.
Mathematical properties and parameter estimation for transit compartment pharmacodynamic models.
Yates, James W T
2008-07-03
One feature of recent research in pharmacodynamic modelling has been the move towards more mechanistically based model structures. However, in all of these models there are common sub-systems, such as feedback loops and time-delays, whose properties and contribution to the model behaviour merit some mathematical analysis. In this paper a common pharmacodynamic model sub-structure is considered: the linear transit compartment. These models have a number of interesting properties as the length of the cascade chain is increased. In the limiting case a pure time-delay is achieved [Milsum, J.H., 1966. Biological Control Systems Analysis. McGraw-Hill Book Company, New York] and the initial behaviour becoming increasingly sensitive to parameter value perturbation. It is also shown that the modelled drug effect is attenuated, though the duration of action is longer. Through this analysis the range of behaviours that such models are capable of reproducing are characterised. The properties of these models and the experimental requirements are discussed in order to highlight how mathematical analysis prior to experimentation can enhance the utility of mathematical modelling.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Dana L. Kelly
Typical engineering systems in applications with high failure consequences such as nuclear reactor plants often employ redundancy and diversity of equipment in an effort to lower the probability of failure and therefore risk. However, it has long been recognized that dependencies exist in these redundant and diverse systems. Some dependencies, such as common sources of electrical power, are typically captured in the logic structure of the risk model. Others, usually referred to as intercomponent dependencies, are treated implicitly by introducing one or more statistical parameters into the model. Such common-cause failure models have limitations in a simulation environment. In addition,more » substantial subjectivity is associated with parameter estimation for these models. This paper describes an approach in which system performance is simulated by drawing samples from the joint distributions of dependent variables. The approach relies on the notion of a copula distribution, a notion which has been employed by the actuarial community for ten years or more, but which has seen only limited application in technological risk assessment. The paper also illustrates how equipment failure data can be used in a Bayesian framework to estimate the parameter values in the copula model. This approach avoids much of the subjectivity required to estimate parameters in traditional common-cause failure models. Simulation examples are presented for failures in time. The open-source software package R is used to perform the simulations. The open-source software package WinBUGS is used to perform the Bayesian inference via Markov chain Monte Carlo sampling.« less
[Systemic therapies--a contribution to psychotherapy integration].
Schiepek, Günter
2012-06-01
Some converging lines from neuroscience, neurobiological psychotherapy research, process-outcome-research, internet-based change monitoring and the systems and complexity sciences actually allow for an open and generic definition of systemic therapies. The "family" of systemic therapies as designed here is not restricted to the field of psychotherapy. It is a scientifically founded and engaged, bio-psycho-social multi-level approach to a common or integrative psychotherapy, not restricted to a psychotherapeutic confession or exclusively to family or couples therapy. A core element of systemic therapy is the support of self-organizing processes and the use of data-driven feedback tools. The conclusion goes to a modified concept of evidence-based practice and, vice versa, practice-based evidence, to an integration of the medical model and the common factors model into a self-organization theory of human change processes, and to a list of criteria for scientifically based practice in psychotherapy. © Georg Thieme Verlag KG Stuttgart · New York.
Evaluating North American Electric Grid Reliability Using the Barabasi-Albert Network Model
DOE Office of Scientific and Technical Information (OSTI.GOV)
Chassin, David P.; Posse, Christian
2005-09-15
The reliability of electric transmission systems is examined using a scale-free model of network topology and failure propagation. The topologies of the North American eastern and western electric grids are analyzed to estimate their reliability based on the Barabási-Albert network model. A commonly used power system reliability index is computed using a simple failure propagation model. The results are compared to the values of power system reliability indices previously obtained using other methods and they suggest that scale-free network models are usable to estimate aggregate electric grid reliability.
Evaluating North American Electric Grid Reliability Using the Barabasi-Albert Network Model
DOE Office of Scientific and Technical Information (OSTI.GOV)
Chassin, David P.; Posse, Christian
2005-09-15
The reliability of electric transmission systems is examined using a scale-free model of network topology and failure propagation. The topologies of the North American eastern and western electric grids are analyzed to estimate their reliability based on the Barabasi-Albert network model. A commonly used power system reliability index is computed using a simple failure propagation model. The results are compared to the values of power system reliability indices previously obtained using standard power engineering methods, and they suggest that scale-free network models are usable to estimate aggregate electric grid reliability.
A Model-Based Expert System for Space Power Distribution Diagnostics
NASA Technical Reports Server (NTRS)
Quinn, Todd M.; Schlegelmilch, Richard F.
1994-01-01
When engineers diagnose system failures, they often use models to confirm system operation. This concept has produced a class of advanced expert systems that perform model-based diagnosis. A model-based diagnostic expert system for the Space Station Freedom electrical power distribution test bed is currently being developed at the NASA Lewis Research Center. The objective of this expert system is to autonomously detect and isolate electrical fault conditions. Marple, a software package developed at TRW, provides a model-based environment utilizing constraint suspension. Originally, constraint suspension techniques were developed for digital systems. However, Marple provides the mechanisms for applying this approach to analog systems such as the test bed, as well. The expert system was developed using Marple and Lucid Common Lisp running on a Sun Sparc-2 workstation. The Marple modeling environment has proved to be a useful tool for investigating the various aspects of model-based diagnostics. This report describes work completed to date and lessons learned while employing model-based diagnostics using constraint suspension within an analog system.
Development of a Solid-Oxide Fuel Cell/Gas Turbine Hybrid System Model for Aerospace Applications
NASA Technical Reports Server (NTRS)
Freeh, Joshua E.; Pratt, Joseph W.; Brouwer, Jacob
2004-01-01
Recent interest in fuel cell-gas turbine hybrid applications for the aerospace industry has led to the need for accurate computer simulation models to aid in system design and performance evaluation. To meet this requirement, solid oxide fuel cell (SOFC) and fuel processor models have been developed and incorporated into the Numerical Propulsion Systems Simulation (NPSS) software package. The SOFC and reformer models solve systems of equations governing steady-state performance using common theoretical and semi-empirical terms. An example hybrid configuration is presented that demonstrates the new capability as well as the interaction with pre-existing gas turbine and heat exchanger models. Finally, a comparison of calculated SOFC performance with experimental data is presented to demonstrate model validity. Keywords: Solid Oxide Fuel Cell, Reformer, System Model, Aerospace, Hybrid System, NPSS
Eye-hand coordination during a double-step task: evidence for a common stochastic accumulator
Gopal, Atul
2015-01-01
Many studies of reaching and pointing have shown significant spatial and temporal correlations between eye and hand movements. Nevertheless, it remains unclear whether these correlations are incidental, arising from common inputs (independent model); whether these correlations represent an interaction between otherwise independent eye and hand systems (interactive model); or whether these correlations arise from a single dedicated eye-hand system (common command model). Subjects were instructed to redirect gaze and pointing movements in a double-step task in an attempt to decouple eye-hand movements and causally distinguish between the three architectures. We used a drift-diffusion framework in the context of a race model, which has been previously used to explain redirect behavior for eye and hand movements separately, to predict the pattern of eye-hand decoupling. We found that the common command architecture could best explain the observed frequency of different eye and hand response patterns to the target step. A common stochastic accumulator for eye-hand coordination also predicts comparable variances, despite significant difference in the means of the eye and hand reaction time (RT) distributions, which we tested. Consistent with this prediction, we observed that the variances of the eye and hand RTs were similar, despite much larger hand RTs (∼90 ms). Moreover, changes in mean eye RTs, which also increased eye RT variance, produced a similar increase in mean and variance of the associated hand RT. Taken together, these data suggest that a dedicated circuit underlies coordinated eye-hand planning. PMID:26084906
The Eclipsing Central Stars of the Planetary Nebulae Lo 16 and PHR J1040-5417
NASA Astrophysics Data System (ADS)
Hillwig, Todd C.; Frew, David; Jones, David; Crispo, Danielle
2017-01-01
Binary central stars of planetary nebula are a valuable tool in understanding common envelope evolution. In these cases both the resulting close binary system and the expanding envelope (the planetary nebula) can be studied directly. In order to compare observed systems with common envelope evolution models we need to determine precise physical parameters of the binaries and the nebulae. Eclipsing central stars provide us with the best opportunity to determine high precision values for mass, radius, and temperature of the component stars in these close binaries. We present photometry and spectroscopy for two of these eclipsing systems; the central stars of Lo 16 and PHR 1040-5417. Using light curves and radial velocity curves along with binary modeling we provide physical parameters for the stars in both of these systems.
Magnetic suspension and balance system study
NASA Technical Reports Server (NTRS)
Boom, R. W.; Eyssa, Y. M.; Mcintosh, G. E.; Abdelsalam, M. K.
1984-01-01
A compact design for a superconducting magnetic suspension and balance system is developed for a 8 ft. x 8 ft. transonic wind tunnel. The main features of the design are: a compact superconducting solenoid in the suspended airplane model; permanent magnet wings; one common liquid helium dewar for all superconducting coils; efficient new race track coils for roll torques; use of established 11 kA cryostable AC conductor; acceptable AC losses during 10 Hz control even with all steel structure; and a 560 liter/hour helium liquefier. Considerable design simplicity, reduced magnet weights, and reduced heat leak results from using one common dewar which eliminates most heavy steel structure between coils and the suspended model. Operational availability is thought to approach 100% for such magnet systems. The weight and cost of the magnet system is approximately one-third that of previous less compact designs.
Modeling human-environmental systems
Morgan Grove; Charlie Schweik; Tom Evans; Glen Green
2002-01-01
This chapter focuses on the integration and development of environmental models that include human decision making. While many methodological and technical issues are common to all types of environmental models, our goal is to highlight the unique characteristics that need to be considered when modeling human-environmental dynamics and to identify future directions for...
Hucka, Michael; Bergmann, Frank T.; Dräger, Andreas; Hoops, Stefan; Keating, Sarah M.; Le Novére, Nicolas; Myers, Chris J.; Olivier, Brett G.; Sahle, Sven; Schaff, James C.; Smith, Lucian P.; Waltemath, Dagmar; Wilkinson, Darren J.
2017-01-01
Summary Computational models can help researchers to interpret data, understand biological function, and make quantitative predictions. The Systems Biology Markup Language (SBML) is a file format for representing computational models in a declarative form that can be exchanged between different software systems. SBML is oriented towards describing biological processes of the sort common in research on a number of topics, including metabolic pathways, cell signaling pathways, and many others. By supporting SBML as an input/output format, different tools can all operate on an identical representation of a model, removing opportunities for translation errors and assuring a common starting point for analyses and simulations. This document provides the specification for Version 5 of SBML Level 2. The specification defines the data structures prescribed by SBML as well as their encoding in XML, the eXtensible Markup Language. This specification also defines validation rules that determine the validity of an SBML document, and provides many examples of models in SBML form. Other materials and software are available from the SBML project web site, http://sbml.org/. PMID:26528569
Hucka, Michael; Bergmann, Frank T; Dräger, Andreas; Hoops, Stefan; Keating, Sarah M; Le Novère, Nicolas; Myers, Chris J; Olivier, Brett G; Sahle, Sven; Schaff, James C; Smith, Lucian P; Waltemath, Dagmar; Wilkinson, Darren J
2015-09-04
Computational models can help researchers to interpret data, understand biological function, and make quantitative predictions. The Systems Biology Markup Language (SBML) is a file format for representing computational models in a declarative form that can be exchanged between different software systems. SBML is oriented towards describing biological processes of the sort common in research on a number of topics, including metabolic pathways, cell signaling pathways, and many others. By supporting SBML as an input/output format, different tools can all operate on an identical representation of a model, removing opportunities for translation errors and assuring a common starting point for analyses and simulations. This document provides the specification for Version 5 of SBML Level 2. The specification defines the data structures prescribed by SBML as well as their encoding in XML, the eXtensible Markup Language. This specification also defines validation rules that determine the validity of an SBML document, and provides many examples of models in SBML form. Other materials and software are available from the SBML project web site, http://sbml.org.
Hucka, Michael; Bergmann, Frank T; Dräger, Andreas; Hoops, Stefan; Keating, Sarah M; Le Novère, Nicolas; Myers, Chris J; Olivier, Brett G; Sahle, Sven; Schaff, James C; Smith, Lucian P; Waltemath, Dagmar; Wilkinson, Darren J
2015-06-01
Computational models can help researchers to interpret data, understand biological function, and make quantitative predictions. The Systems Biology Markup Language (SBML) is a file format for representing computational models in a declarative form that can be exchanged between different software systems. SBML is oriented towards describing biological processes of the sort common in research on a number of topics, including metabolic pathways, cell signaling pathways, and many others. By supporting SBML as an input/output format, different tools can all operate on an identical representation of a model, removing opportunities for translation errors and assuring a common starting point for analyses and simulations. This document provides the specification for Version 5 of SBML Level 2. The specification defines the data structures prescribed by SBML as well as their encoding in XML, the eXtensible Markup Language. This specification also defines validation rules that determine the validity of an SBML document, and provides many examples of models in SBML form. Other materials and software are available from the SBML project web site, http://sbml.org/.
Cervero-Aragó, Sílvia; Rodríguez-Martínez, Sarah; Puertas-Bennasar, Antoni; Araujo, Rosa M.
2015-01-01
Chlorine and thermal treatments are the most commonly used procedures to control and prevent Legionella proliferation in drinking water systems of large buildings. However, cases of legionellosis still occur in facilities with treated water. The purpose of this work was to model the effect of temperature and free chlorine applied in similar exposure conditions as in drinking water systems on five Legionella spp. strains and two amoebal strains of the genera Acanthamoeba. Inactivation models obtained were used to determine the effectiveness of the treatments applied which resulted more effective against Legionella than Acanthamoeba, especially those in cystic stages. Furthermore, to determine the influence of the relationship between L. pneumophila and Acanthamoeba spp. on the treatment effectiveness, inactivation models of the bacteria-associated amoeba were also constructed and compared to the models obtained for the free living bacteria state. The Legionella-amoeba association did not change the inactivation models, but it reduced the effectiveness of the treatments applied. Remarkably, at the lowest free chlorine concentration, 0.5 mg L-1, as well as at the lowest temperatures, 50°C and 55°C, the influence of the Legionella-amoeba associate state was the strongest in reducing the effectiveness of the treatments compared to the free Legionella state. Therefore, the association established between L. pneumophila and amoebae in the water systems indicate an increased health risk in proximal areas of the system (close to the tap) where lower free chlorine concentrations and lower temperatures are commonly observed. PMID:26241039
Madi, Mahmoud K; Karameh, Fadi N
2018-05-11
Many physical models of biological processes including neural systems are characterized by parametric nonlinear dynamical relations between driving inputs, internal states, and measured outputs of the process. Fitting such models using experimental data (data assimilation) is a challenging task since the physical process often operates in a noisy, possibly non-stationary environment; moreover, conducting multiple experiments under controlled and repeatable conditions can be impractical, time consuming or costly. The accuracy of model identification, therefore, is dictated principally by the quality and dynamic richness of collected data over single or few experimental sessions. Accordingly, it is highly desirable to design efficient experiments that, by exciting the physical process with smart inputs, yields fast convergence and increased accuracy of the model. We herein introduce an adaptive framework in which optimal input design is integrated with Square root Cubature Kalman Filters (OID-SCKF) to develop an online estimation procedure that first, converges significantly quicker, thereby permitting model fitting over shorter time windows, and second, enhances model accuracy when only few process outputs are accessible. The methodology is demonstrated on common nonlinear models and on a four-area neural mass model with noisy and limited measurements. Estimation quality (speed and accuracy) is benchmarked against high-performance SCKF-based methods that commonly employ dynamically rich informed inputs for accurate model identification. For all the tested models, simulated single-trial and ensemble averages showed that OID-SCKF exhibited (i) faster convergence of parameter estimates and (ii) lower dependence on inter-trial noise variability with gains up to around 1000 msec in speed and 81% increase in variability for the neural mass models. In terms of accuracy, OID-SCKF estimation was superior, and exhibited considerably less variability across experiments, in identifying model parameters of (a) systems with challenging model inversion dynamics and (b) systems with fewer measurable outputs that directly relate to the underlying processes. Fast and accurate identification therefore carries particular promise for modeling of transient (short-lived) neuronal network dynamics using a spatially under-sampled set of noisy measurements, as is commonly encountered in neural engineering applications. © 2018 IOP Publishing Ltd.
Common mechanisms of synaptic plasticity in vertebrates and invertebrates
Glanzman, David L.
2016-01-01
Until recently, the literature on learning-related synaptic plasticity in invertebrates has been dominated by models assuming plasticity is mediated by presynaptic changes, whereas the vertebrate literature has been dominated by models assuming it is mediated by postsynaptic changes. Here I will argue that this situation does not reflect a biological reality and that, in fact, invertebrate and vertebrate nervous systems share a common set of mechanisms of synaptic plasticity. PMID:20152143
Blainey, Joan B.; Faunt, Claudia C.; Hill, Mary C.
2006-01-01
This report is a guide for executing numerical simulations with the transient ground-water flow model of the Death Valley regional ground-water flow system, Nevada and California using the U.S. Geological Survey modular finite-difference ground-water flow model, MODFLOW-2000. Model inputs, including observations of hydraulic head, discharge, and boundary flows, are summarized. Modification of the DVRFS transient ground-water model is discussed for two common uses of the Death Valley regional ground-water flow system model: predictive pumping scenarios that extend beyond the end of the model simulation period (1998), and model simulations with only steady-state conditions.
A Methodology for Quantifying Certain Design Requirements During the Design Phase
NASA Technical Reports Server (NTRS)
Adams, Timothy; Rhodes, Russel
2005-01-01
A methodology for developing and balancing quantitative design requirements for safety, reliability, and maintainability has been proposed. Conceived as the basis of a more rational approach to the design of spacecraft, the methodology would also be applicable to the design of automobiles, washing machines, television receivers, or almost any other commercial product. Heretofore, it has been common practice to start by determining the requirements for reliability of elements of a spacecraft or other system to ensure a given design life for the system. Next, safety requirements are determined by assessing the total reliability of the system and adding redundant components and subsystems necessary to attain safety goals. As thus described, common practice leaves the maintainability burden to fall to chance; therefore, there is no control of recurring costs or of the responsiveness of the system. The means that have been used in assessing maintainability have been oriented toward determining the logistical sparing of components so that the components are available when needed. The process established for developing and balancing quantitative requirements for safety (S), reliability (R), and maintainability (M) derives and integrates NASA s top-level safety requirements and the controls needed to obtain program key objectives for safety and recurring cost (see figure). Being quantitative, the process conveniently uses common mathematical models. Even though the process is shown as being worked from the top down, it can also be worked from the bottom up. This process uses three math models: (1) the binomial distribution (greaterthan- or-equal-to case), (2) reliability for a series system, and (3) the Poisson distribution (less-than-or-equal-to case). The zero-fail case for the binomial distribution approximates the commonly known exponential distribution or "constant failure rate" distribution. Either model can be used. The binomial distribution was selected for modeling flexibility because it conveniently addresses both the zero-fail and failure cases. The failure case is typically used for unmanned spacecraft as with missiles.
Common IED exploitation target set ontology
NASA Astrophysics Data System (ADS)
Russomanno, David J.; Qualls, Joseph; Wowczuk, Zenovy; Franken, Paul; Robinson, William
2010-04-01
The Common IED Exploitation Target Set (CIEDETS) ontology provides a comprehensive semantic data model for capturing knowledge about sensors, platforms, missions, environments, and other aspects of systems under test. The ontology also includes representative IEDs; modeled as explosives, camouflage, concealment objects, and other background objects, which comprise an overall threat scene. The ontology is represented using the Web Ontology Language and the SPARQL Protocol and RDF Query Language, which ensures portability of the acquired knowledge base across applications. The resulting knowledge base is a component of the CIEDETS application, which is intended to support the end user sensor test and evaluation community. CIEDETS associates a system under test to a subset of cataloged threats based on the probability that the system will detect the threat. The associations between systems under test, threats, and the detection probabilities are established based on a hybrid reasoning strategy, which applies a combination of heuristics and simplified modeling techniques. Besides supporting the CIEDETS application, which is focused on efficient and consistent system testing, the ontology can be leveraged in a myriad of other applications, including serving as a knowledge source for mission planning tools.
Traffic flow theory and chaotic behavior
DOT National Transportation Integrated Search
1989-03-01
Many commonly occurring natural systems are modeled with mathematical experessions and exhibit a certain stability. The inherent stability of these equations allows them to serve as the basis for engineering predictions. More complex models, such as ...
CEDS Addresses: Rubric Elements
ERIC Educational Resources Information Center
US Department of Education, 2015
2015-01-01
Common Education Data Standards (CEDS) Version 4 introduced a common data vocabulary for defining rubrics in a data system. The CEDS elements support digital representations of both holistic and analytic rubrics. This document shares examples of holistic and analytic project rubrics, available CEDS Connections, and a logical model showing the…
Model Test of Proposed Loading Rates for Onsite Wastewater Treatment Systems
State regulatory agencies set standards for onsite wastewater treatment system (OWTS), commonly known as septic systems, based on expected hydraulic performance and nitrogen (N) treatment in soils of differing texture. In a previous study, hydraulic loading rates were proposed fo...
Stochastic simulation of multiscale complex systems with PISKaS: A rule-based approach.
Perez-Acle, Tomas; Fuenzalida, Ignacio; Martin, Alberto J M; Santibañez, Rodrigo; Avaria, Rodrigo; Bernardin, Alejandro; Bustos, Alvaro M; Garrido, Daniel; Dushoff, Jonathan; Liu, James H
2018-03-29
Computational simulation is a widely employed methodology to study the dynamic behavior of complex systems. Although common approaches are based either on ordinary differential equations or stochastic differential equations, these techniques make several assumptions which, when it comes to biological processes, could often lead to unrealistic models. Among others, model approaches based on differential equations entangle kinetics and causality, failing when complexity increases, separating knowledge from models, and assuming that the average behavior of the population encompasses any individual deviation. To overcome these limitations, simulations based on the Stochastic Simulation Algorithm (SSA) appear as a suitable approach to model complex biological systems. In this work, we review three different models executed in PISKaS: a rule-based framework to produce multiscale stochastic simulations of complex systems. These models span multiple time and spatial scales ranging from gene regulation up to Game Theory. In the first example, we describe a model of the core regulatory network of gene expression in Escherichia coli highlighting the continuous model improvement capacities of PISKaS. The second example describes a hypothetical outbreak of the Ebola virus occurring in a compartmentalized environment resembling cities and highways. Finally, in the last example, we illustrate a stochastic model for the prisoner's dilemma; a common approach from social sciences describing complex interactions involving trust within human populations. As whole, these models demonstrate the capabilities of PISKaS providing fertile scenarios where to explore the dynamics of complex systems. Copyright © 2017 The Authors. Published by Elsevier Inc. All rights reserved.
ERIC Educational Resources Information Center
Spears, Janine L.; Parrish, James L., Jr.
2013-01-01
This teaching case introduces students to a relatively simple approach to identifying and documenting security requirements within conceptual models that are commonly taught in systems analysis and design courses. An introduction to information security is provided, followed by a classroom example of a fictitious company, "Fun &…
Method and system for dynamic probabilistic risk assessment
NASA Technical Reports Server (NTRS)
Dugan, Joanne Bechta (Inventor); Xu, Hong (Inventor)
2013-01-01
The DEFT methodology, system and computer readable medium extends the applicability of the PRA (Probabilistic Risk Assessment) methodology to computer-based systems, by allowing DFT (Dynamic Fault Tree) nodes as pivot nodes in the Event Tree (ET) model. DEFT includes a mathematical model and solution algorithm, supports all common PRA analysis functions and cutsets. Additional capabilities enabled by the DFT include modularization, phased mission analysis, sequence dependencies, and imperfect coverage.
Threat driven modeling framework using petri nets for e-learning system.
Khamparia, Aditya; Pandey, Babita
2016-01-01
Vulnerabilities at various levels are main cause of security risks in e-learning system. This paper presents a modified threat driven modeling framework, to identify the threats after risk assessment which requires mitigation and how to mitigate those threats. To model those threat mitigations aspects oriented stochastic petri nets are used. This paper included security metrics based on vulnerabilities present in e-learning system. The Common Vulnerability Scoring System designed to provide a normalized method for rating vulnerabilities which will be used as basis in metric definitions and calculations. A case study has been also proposed which shows the need and feasibility of using aspect oriented stochastic petri net models for threat modeling which improves reliability, consistency and robustness of the e-learning system.
Generic Educational Knowledge Representation for Adaptive and Cognitive Systems
ERIC Educational Resources Information Center
Caravantes, Arturo; Galan, Ramon
2011-01-01
The interoperability of educational systems, encouraged by the development of specifications, standards and tools related to the Semantic Web is limited to the exchange of information in domain and student models. High system interoperability requires that a common framework be defined that represents the functional essence of educational systems.…
Advanced Technology Lifecycle Analysis System (ATLAS) Technology Tool Box (TTB)
NASA Technical Reports Server (NTRS)
Doyle, Monica; ONeil, Daniel A.; Christensen, Carissa B.
2005-01-01
The Advanced Technology Lifecycle Analysis System (ATLAS) is a decision support tool designed to aid program managers and strategic planners in determining how to invest technology research and development dollars. It is an Excel-based modeling package that allows a user to build complex space architectures and evaluate the impact of various technology choices. ATLAS contains system models, cost and operations models, a campaign timeline and a centralized technology database. Technology data for all system models is drawn from a common database, the ATLAS Technology Tool Box (TTB). The TTB provides a comprehensive, architecture-independent technology database that is keyed to current and future timeframes.
Phillips, Joshua; Chilukuri, Ram; Fragoso, Gilberto; Warzel, Denise; Covitz, Peter A
2006-01-01
Background Robust, programmatically accessible biomedical information services that syntactically and semantically interoperate with other resources are challenging to construct. Such systems require the adoption of common information models, data representations and terminology standards as well as documented application programming interfaces (APIs). The National Cancer Institute (NCI) developed the cancer common ontologic representation environment (caCORE) to provide the infrastructure necessary to achieve interoperability across the systems it develops or sponsors. The caCORE Software Development Kit (SDK) was designed to provide developers both within and outside the NCI with the tools needed to construct such interoperable software systems. Results The caCORE SDK requires a Unified Modeling Language (UML) tool to begin the development workflow with the construction of a domain information model in the form of a UML Class Diagram. Models are annotated with concepts and definitions from a description logic terminology source using the Semantic Connector component. The annotated model is registered in the Cancer Data Standards Repository (caDSR) using the UML Loader component. System software is automatically generated using the Codegen component, which produces middleware that runs on an application server. The caCORE SDK was initially tested and validated using a seven-class UML model, and has been used to generate the caCORE production system, which includes models with dozens of classes. The deployed system supports access through object-oriented APIs with consistent syntax for retrieval of any type of data object across all classes in the original UML model. The caCORE SDK is currently being used by several development teams, including by participants in the cancer biomedical informatics grid (caBIG) program, to create compatible data services. caBIG compatibility standards are based upon caCORE resources, and thus the caCORE SDK has emerged as a key enabling technology for caBIG. Conclusion The caCORE SDK substantially lowers the barrier to implementing systems that are syntactically and semantically interoperable by providing workflow and automation tools that standardize and expedite modeling, development, and deployment. It has gained acceptance among developers in the caBIG program, and is expected to provide a common mechanism for creating data service nodes on the data grid that is under development. PMID:16398930
The Parallel System for Integrating Impact Models and Sectors (pSIMS)
NASA Technical Reports Server (NTRS)
Elliott, Joshua; Kelly, David; Chryssanthacopoulos, James; Glotter, Michael; Jhunjhnuwala, Kanika; Best, Neil; Wilde, Michael; Foster, Ian
2014-01-01
We present a framework for massively parallel climate impact simulations: the parallel System for Integrating Impact Models and Sectors (pSIMS). This framework comprises a) tools for ingesting and converting large amounts of data to a versatile datatype based on a common geospatial grid; b) tools for translating this datatype into custom formats for site-based models; c) a scalable parallel framework for performing large ensemble simulations, using any one of a number of different impacts models, on clusters, supercomputers, distributed grids, or clouds; d) tools and data standards for reformatting outputs to common datatypes for analysis and visualization; and e) methodologies for aggregating these datatypes to arbitrary spatial scales such as administrative and environmental demarcations. By automating many time-consuming and error-prone aspects of large-scale climate impacts studies, pSIMS accelerates computational research, encourages model intercomparison, and enhances reproducibility of simulation results. We present the pSIMS design and use example assessments to demonstrate its multi-model, multi-scale, and multi-sector versatility.
Reusing models of actors and services in smart homecare to improve sustainability.
Walderhaug, Ståle; Stav, Erlend; Mikalsen, Marius
2008-01-01
Industrial countries are faced with a growing elderly population. Homecare systems with assistive smart house technology enable elderly to live independently at home. Development of such smart home care systems is complex and expensive and there is no common reference model that can facilitate service reuse. This paper proposes reusable actor and service models based on a model-driven development process where end user organizations and domain healthcare experts from four European countries have been involved. The models, specified using UML can be reused actively as assets in the system design and development process and can reduce development costs, and improve interoperability and sustainability of systems. The models are being evaluated in the European IST project MPOWER.
Competent Systems: Effective, Efficient, Deliverable.
ERIC Educational Resources Information Center
Abramson, Bruce
Recent developments in artificial intelligence and decision analysis suggest reassessing the approaches commonly taken to the design of knowledge-based systems. Competent systems are based on models known as influence diagrams, which graphically capture a domain's basic objects and their interrelationships. Among the benefits offered by influence…
Disorders of compulsivity: a common bias towards learning habits
Voon, V; Derbyshire, K; Rück, C; Irvine, M A; Worbe, Y; Enander, J; Schreiber, L R N; Gillan, C; Fineberg, N A; Sahakian, B J; Robbins, T W; Harrison, N A; Wood, J; Daw, N D; Dayan, P; Grant, J E; Bullmore, E T
2015-01-01
Why do we repeat choices that we know are bad for us? Decision making is characterized by the parallel engagement of two distinct systems, goal-directed and habitual, thought to arise from two computational learning mechanisms, model-based and model-free. The habitual system is a candidate source of pathological fixedness. Using a decision task that measures the contribution to learning of either mechanism, we show a bias towards model-free (habit) acquisition in disorders involving both natural (binge eating) and artificial (methamphetamine) rewards, and obsessive-compulsive disorder. This favoring of model-free learning may underlie the repetitive behaviors that ultimately dominate in these disorders. Further, we show that the habit formation bias is associated with lower gray matter volumes in caudate and medial orbitofrontal cortex. Our findings suggest that the dysfunction in a common neurocomputational mechanism may underlie diverse disorders involving compulsion. PMID:24840709
NASA Common Research Model Test Envelope Extension With Active Sting Damping at NTF
NASA Technical Reports Server (NTRS)
Rivers, Melissa B.; Balakrishna, S.
2014-01-01
The NASA Common Research Model (CRM) high Reynolds number transonic wind tunnel testing program was established to generate an experimental database for applied Computational Fluid Dynamics (CFD) validation studies. During transonic wind tunnel tests, the CRM encounters large sting vibrations when the angle of attack approaches the second pitching moment break, which can sometimes become divergent. CRM transonic test data analysis suggests that sting divergent oscillations are related to negative net sting damping episodes associated with flow separation instability. The National Transonic Facility (NTF) has been addressing remedies to extend polar testing up to and beyond the second pitching moment break point of the test articles using an active piezoceramic damper system for both ambient and cryogenic temperatures. This paper reviews CRM test results to gain understanding of sting dynamics with a simple model describing the mechanics of a sting-model system and presents the performance of the damper under cryogenic conditions.
Disorders of compulsivity: a common bias towards learning habits.
Voon, V; Derbyshire, K; Rück, C; Irvine, M A; Worbe, Y; Enander, J; Schreiber, L R N; Gillan, C; Fineberg, N A; Sahakian, B J; Robbins, T W; Harrison, N A; Wood, J; Daw, N D; Dayan, P; Grant, J E; Bullmore, E T
2015-03-01
Why do we repeat choices that we know are bad for us? Decision making is characterized by the parallel engagement of two distinct systems, goal-directed and habitual, thought to arise from two computational learning mechanisms, model-based and model-free. The habitual system is a candidate source of pathological fixedness. Using a decision task that measures the contribution to learning of either mechanism, we show a bias towards model-free (habit) acquisition in disorders involving both natural (binge eating) and artificial (methamphetamine) rewards, and obsessive-compulsive disorder. This favoring of model-free learning may underlie the repetitive behaviors that ultimately dominate in these disorders. Further, we show that the habit formation bias is associated with lower gray matter volumes in caudate and medial orbitofrontal cortex. Our findings suggest that the dysfunction in a common neurocomputational mechanism may underlie diverse disorders involving compulsion.
NASA Astrophysics Data System (ADS)
Laramie, Sydney M.; Milshtein, Jarrod D.; Breault, Tanya M.; Brushett, Fikile R.; Thompson, Levi T.
2016-09-01
Non-aqueous redox flow batteries (NAqRFBs) have recently received considerable attention as promising high energy density, low cost grid-level energy storage technologies. Despite these attractive features, NAqRFBs are still at an early stage of development and innovative design techniques are necessary to improve performance and decrease costs. In this work, we investigate multi-electron transfer, common ion exchange NAqRFBs. Common ion systems decrease the supporting electrolyte requirement, which subsequently improves active material solubility and decreases electrolyte cost. Voltammetric and electrolytic techniques are used to study the electrochemical performance and chemical compatibility of model redox active materials, iron (II) tris(2,2‧-bipyridine) tetrafluoroborate (Fe(bpy)3(BF4)2) and ferrocenylmethyl dimethyl ethyl ammonium tetrafluoroborate (Fc1N112-BF4). These results help disentangle complex cycling behavior observed in flow cell experiments. Further, a simple techno-economic model demonstrates the cost benefits of employing common ion exchange NAqRFBs, afforded by decreasing the salt and solvent contributions to total chemical cost. This study highlights two new concepts, common ion exchange and multi-electron transfer, for NAqRFBs through a demonstration flow cell employing model active species. In addition, the compatibility analysis developed for asymmetric chemistries can apply to other promising species, including organics, metal coordination complexes (MCCs) and mixed MCC/organic systems, enabling the design of low cost NAqRFBs.
Accurate Sloshing Modes Modeling: A New Analytical Solution and its Consequences on Control
NASA Astrophysics Data System (ADS)
Gonidou, Luc-Olivier; Desmariaux, Jean
2014-06-01
This study addresses the issue of sloshing modes modeling for GNC analyses purposes. On European launchers, equivalent mechanical systems are commonly used for modeling sloshing effects on launcher dynamics. The representativeness of such a methodology is discussed here. First an exact analytical formulation of the launcher dynamics fitted with sloshing modes is proposed and discrepancies with equivalent mechanical system approach are emphasized. Then preliminary comparative GNC analyses are performed using the different models of dynamics in order to evaluate the impact of the aforementioned discrepancies from GNC standpoint. Special attention is paid to system stability.
Preparation of a New Oligolamellar Stratum Corneum Lipid Model.
Mueller, Josefin; Schroeter, Annett; Steitz, Roland; Trapp, Marcus; Neubert, Reinhard H H
2016-05-10
In this study, we present a preparation method for a new stratum corneum (SC) model system, which is closer to natural SC than the commonly used multilayer models. The complex setup of the native SC lipid matrix was mimicked by a ternary lipid mixture of ceramide [AP], cholesterol, and stearic acid. A spin coating procedure was applied to realize oligo-layered samples. The influence of lipid concentration, rotation speed, polyethylenimine, methanol content, cholesterol fraction, and annealing on the molecular arrangement of the new SC model was investigated by X-ray reflectivity measurements. The new oligo-SC model is closer to native SC in the total number of lipid membranes found between corneocytes. The reduction in thickness provides the opportunity to study the effects of drugs and/or hydrophilic penetration enhancers on the structure of SC in full detail by X-ray or neutron reflectivity. In addition, the oligo-lamellar systems allows one to infer not only the lamellar spacing, but also the total thickness of the oligo-SC model and changes thereof can be monitored. This improvement is most helpful for the understanding of transdermal drug administration on the nanoscale. The results are compared to the commonly used multilamellar lipid model systems and advantages and disadvantages of both models are discussed.
The application of systems thinking in health: why use systems thinking?
Peters, David H
2014-08-26
This paper explores the question of what systems thinking adds to the field of global health. Observing that elements of systems thinking are already common in public health research, the article discusses which of the large body of theories, methods, and tools associated with systems thinking are more useful. The paper reviews the origins of systems thinking, describing a range of the theories, methods, and tools. A common thread is the idea that the behavior of systems is governed by common principles that can be discovered and expressed. They each address problems of complexity, which is a frequent challenge in global health. The different methods and tools are suited to different types of inquiry and involve both qualitative and quantitative techniques. The paper concludes by emphasizing that explicit models used in systems thinking provide new opportunities to understand and continuously test and revise our understanding of the nature of things, including how to intervene to improve people's health.
Ontology for Life-Cycle Modeling of Electrical Distribution Systems: Model View Definition
2013-06-01
building information models ( BIM ) at the coordinated design stage of building construction. 1.3 Approach To...standard for exchanging Building Information Modeling ( BIM ) data, which defines hundreds of classes for common use in software, currently supported by...specifications, Construction Operations Building in- formation exchange (COBie), Building Information Modeling ( BIM ) 16. SECURITY CLASSIFICATION OF:
On the dynamics of chain systems. [applications in manipulator and human body models
NASA Technical Reports Server (NTRS)
Huston, R. L.; Passerello, C. E.
1974-01-01
A computer-oriented method for obtaining dynamical equations of motion for chain systems is presented. A chain system is defined as an arbitrarily assembled set of rigid bodies such that adjoining bodies have at least one common point and such that closed loops are not formed. The equations of motion are developed through the use of Lagrange's form of d'Alembert's principle. The method and procedure is illustrated with an elementary study of a tripod space manipulator. The method is designed for application with systems such as human body models, chains and cables, and dynamic finite-segment models.
DOE Office of Scientific and Technical Information (OSTI.GOV)
King, D.L.
1995-11-01
The objective of this work was to develop improved performance model for modules and systems for for all operating conditions for use in module specifications, system and BOS component design, and system rating or monitoring. The approach taken was to identify and quantify the influence of dominant factors of solar irradiance, cell temperature, angle-of-incidence; and solar spectrum; use outdoor test procedures to separate the effects of electrical, thermal, and optical performance; use fundamental cell characteristics to improve analysis; and combine factors in simple model using the common variables.
So, H C; Pearl, D L; von Königslöw, T; Louie, M; Chui, L; Svenson, L W
2013-08-01
Molecular typing methods have become a common part of the surveillance of foodborne pathogens. In particular, pulsed-field gel electrophoresis (PFGE) has been used successfully to identify outbreaks of Escherichia coli O157:H7 in humans from a variety of food and environmental sources. However, some PFGE patterns appear commonly in surveillance systems, making it more difficult to distinguish between outbreak and sporadic cases based on molecular data alone. In addition, it is unknown whether these common patterns might have unique epidemiological characteristics reflected in their spatial and temporal distributions. Using E. coli O157:H7 surveillance data from Alberta, collected from 2000 to 2002, we investigated whether E. coli O157:H7 with provincial PFGE pattern 8 (national designation ECXAI.0001) clustered in space, time and space-time relative to other PFGE patterns using the spatial scan statistic. Based on our purely spatial and temporal scans using a Bernoulli model, there did not appear to be strong evidence that isolates of E. coli O157:H7 with provincial PFGE pattern 8 are distributed differently from other PFGE patterns. However, we did identify space-time clusters of isolates with PFGE pattern 8, using a Bernoulli model and a space-time permutation model, which included known outbreaks and potentially unrecognized outbreaks or additional outbreak cases. There were differences between the two models in the space-time clusters identified, which suggests that the use of both models could increase the sensitivity of a quantitative surveillance system for identifying outbreaks involving isolates sharing a common PFGE pattern. © 2012 Blackwell Verlag GmbH.
Distributed Time Synchronization Algorithms and Opinion Dynamics
NASA Astrophysics Data System (ADS)
Manita, Anatoly; Manita, Larisa
2018-01-01
We propose new deterministic and stochastic models for synchronization of clocks in nodes of distributed networks. An external accurate time server is used to ensure convergence of the node clocks to the exact time. These systems have much in common with mathematical models of opinion formation in multiagent systems. There is a direct analogy between the time server/node clocks pair in asynchronous networks and the leader/follower pair in the context of social network models.
Executable Architecture Research at Old Dominion University
NASA Technical Reports Server (NTRS)
Tolk, Andreas; Shuman, Edwin A.; Garcia, Johnny J.
2011-01-01
Executable Architectures allow the evaluation of system architectures not only regarding their static, but also their dynamic behavior. However, the systems engineering community do not agree on a common formal specification of executable architectures. To close this gap and identify necessary elements of an executable architecture, a modeling language, and a modeling formalism is topic of ongoing PhD research. In addition, systems are generally defined and applied in an operational context to provide capabilities and enable missions. To maximize the benefits of executable architectures, a second PhD effort introduces the idea of creating an executable context in addition to the executable architecture. The results move the validation of architectures from the current information domain into the knowledge domain and improve the reliability of such validation efforts. The paper presents research and results of both doctoral research efforts and puts them into a common context of state-of-the-art of systems engineering methods supporting more agility.
FitzHenry, F; Resnic, F S; Robbins, S L; Denton, J; Nookala, L; Meeker, D; Ohno-Machado, L; Matheny, M E
2015-01-01
Adoption of a common data model across health systems is a key infrastructure requirement to allow large scale distributed comparative effectiveness analyses. There are a growing number of common data models (CDM), such as Mini-Sentinel, and the Observational Medical Outcomes Partnership (OMOP) CDMs. In this case study, we describe the challenges and opportunities of a study specific use of the OMOP CDM by two health systems and describe three comparative effectiveness use cases developed from the CDM. The project transformed two health system databases (using crosswalks provided) into the OMOP CDM. Cohorts were developed from the transformed CDMs for three comparative effectiveness use case examples. Administrative/billing, demographic, order history, medication, and laboratory were included in the CDM transformation and cohort development rules. Record counts per person month are presented for the eligible cohorts, highlighting differences between the civilian and federal datasets, e.g. the federal data set had more outpatient visits per person month (6.44 vs. 2.05 per person month). The count of medications per person month reflected the fact that one system's medications were extracted from orders while the other system had pharmacy fills and medication administration records. The federal system also had a higher prevalence of the conditions in all three use cases. Both systems required manual coding of some types of data to convert to the CDM. The data transformation to the CDM was time consuming and resources required were substantial, beyond requirements for collecting native source data. The need to manually code subsets of data limited the conversion. However, once the native data was converted to the CDM, both systems were then able to use the same queries to identify cohorts. Thus, the CDM minimized the effort to develop cohorts and analyze the results across the sites.
Moayyeri, Alireza; Hart, Deborah J; Snieder, Harold; Hammond, Christopher J; Spector, Timothy D; Steves, Claire J
2016-02-01
Little is known about the extent to which aging trajectories of different body systems share common sources of variance. We here present a large twin study investigating the trajectories of change in five systems: cardiovascular, respiratory, skeletal, morphometric, and metabolic. Longitudinal clinical data were collected on 3,508 female twins in the TwinsUK registry (complete pairs:740 monozygotic (MZ), 986 dizygotic (DZ), mean age at entry 48.9 ± 10.4, range 18-75 years; mean follow-up 10.2 ± 2.8 years, range 4-17.8 years). Panel data on multiple age-related variables were used to estimate biological ages for each individual at each time point, in linear mixed effects models. A weighted average approach was used to combine variables within predefined body system groups. Aging trajectories for each system in each individual were then constructed using linear modeling. Multivariate structural equation modeling of these aging trajectories showed low genetic effects (heritability), ranging from 2% in metabolic aging to 22% in cardiovascular aging. However, we found a significant effect of shared environmental factors on the variations in aging trajectories in cardiovascular (54%), skeletal (34%), morphometric (53%), and metabolic systems (53%). The remainder was due to environmental factors unique to each individual plus error. Multivariate Cholesky decomposition showed that among aging trajectories for various body systems there were significant and substantial correlations between the unique environmental latent factors as well as shared environmental factors. However, there was no evidence for a single common factor for aging. This study, the first of its kind in aging, suggests that diverse organ systems share non-genetic sources of variance for aging trajectories. Confirmatory studies are needed using population-based twin cohorts and alternative methods of handling missing data.
2010-09-01
matrix is used in many methods, like Jacobi or Gauss Seidel , for solving linear systems. Also, no partial pivoting is necessary for a strictly column...problems that arise during the procedure, which in general, converges to the solving of a linear system. The most common issue with the solution is the... iterative procedure to find an appropriate subset of parameters that produce an optimal solution commonly known as forward selection. Then, the
Data Driven Model Development for the Supersonic Semispan Transport (S(sup 4)T)
NASA Technical Reports Server (NTRS)
Kukreja, Sunil L.
2011-01-01
We investigate two common approaches to model development for robust control synthesis in the aerospace community; namely, reduced order aeroservoelastic modelling based on structural finite-element and computational fluid dynamics based aerodynamic models and a data-driven system identification procedure. It is shown via analysis of experimental Super- Sonic SemiSpan Transport (S4T) wind-tunnel data using a system identification approach it is possible to estimate a model at a fixed Mach, which is parsimonious and robust across varying dynamic pressures.
Wideband Channel Modeling in Real Atmospheric Environments with Experimental Evaluation
2013-04-01
5] D. F. Gingras and P. Gerstoft. 1997. “The Effect of Propagation on Wideband DS - CDMA Systems in the Suburban Environment,” The First IEEE...are commonly used in spread spectrum communication systems such as Code Division Multiple Access ( CDMA ) systems. Narrowband interference mitigation
Recombination energy in double white dwarf formation
NASA Astrophysics Data System (ADS)
Nandez, J. L. A.; Ivanova, N.; Lombardi, J. C.
2015-06-01
In this Letter, we investigate the role of recombination energy during a common envelope event. We confirm that taking this energy into account helps to avoid the formation of the circumbinary envelope commonly found in previous studies. For the first time, we can model a complete common envelope event, with a clean compact double white dwarf binary system formed at the end. The resulting binary orbit is almost perfectly circular. In addition to considering recombination energy, we also show that between 1/4 and 1/2 of the released orbital energy is taken away by the ejected material. We apply this new method to the case of the double white dwarf system WD 1101+364, and we find that the progenitor system at the start of the common envelope event consisted of an ˜1.5 M⊙ red giant star in an ˜30 d orbit with a white dwarf companion.
A structural model decomposition framework for systems health management
NASA Astrophysics Data System (ADS)
Roychoudhury, I.; Daigle, M.; Bregon, A.; Pulido, B.
Systems health management (SHM) is an important set of technologies aimed at increasing system safety and reliability by detecting, isolating, and identifying faults; and predicting when the system reaches end of life (EOL), so that appropriate fault mitigation and recovery actions can be taken. Model-based SHM approaches typically make use of global, monolithic system models for online analysis, which results in a loss of scalability and efficiency for large-scale systems. Improvement in scalability and efficiency can be achieved by decomposing the system model into smaller local submodels and operating on these submodels instead. In this paper, the global system model is analyzed offline and structurally decomposed into local submodels. We define a common model decomposition framework for extracting submodels from the global model. This framework is then used to develop algorithms for solving model decomposition problems for the design of three separate SHM technologies, namely, estimation (which is useful for fault detection and identification), fault isolation, and EOL prediction. We solve these model decomposition problems using a three-tank system as a case study.
A Structural Model Decomposition Framework for Systems Health Management
NASA Technical Reports Server (NTRS)
Roychoudhury, Indranil; Daigle, Matthew J.; Bregon, Anibal; Pulido, Belamino
2013-01-01
Systems health management (SHM) is an important set of technologies aimed at increasing system safety and reliability by detecting, isolating, and identifying faults; and predicting when the system reaches end of life (EOL), so that appropriate fault mitigation and recovery actions can be taken. Model-based SHM approaches typically make use of global, monolithic system models for online analysis, which results in a loss of scalability and efficiency for large-scale systems. Improvement in scalability and efficiency can be achieved by decomposing the system model into smaller local submodels and operating on these submodels instead. In this paper, the global system model is analyzed offline and structurally decomposed into local submodels. We define a common model decomposition framework for extracting submodels from the global model. This framework is then used to develop algorithms for solving model decomposition problems for the design of three separate SHM technologies, namely, estimation (which is useful for fault detection and identification), fault isolation, and EOL prediction. We solve these model decomposition problems using a three-tank system as a case study.
ERIC Educational Resources Information Center
Tsai, Bor-sheng
1991-01-01
Examines the information communication process and proposes a fuzzy commonality model for improving communication systems. Topics discussed include components of an electronic information programing and processing system and the flow of the formation and transfer of information, including DOS (disk operating system) commands, computer programing…
A Realization of Bias Correction Method in the GMAO Coupled System
NASA Technical Reports Server (NTRS)
Chang, Yehui; Koster, Randal; Wang, Hailan; Schubert, Siegfried; Suarez, Max
2018-01-01
Over the past several decades, a tremendous effort has been made to improve model performance in the simulation of the climate system. The cold or warm sea surface temperature (SST) bias in the tropics is still a problem common to most coupled ocean atmosphere general circulation models (CGCMs). The precipitation biases in CGCMs are also accompanied by SST and surface wind biases. The deficiencies and biases over the equatorial oceans through their influence on the Walker circulation likely contribute the precipitation biases over land surfaces. In this study, we introduce an approach in the CGCM modeling to correct model biases. This approach utilizes the history of the model's short-term forecasting errors and their seasonal dependence to modify model's tendency term and to minimize its climate drift. The study shows that such an approach removes most of model climate biases. A number of other aspects of the model simulation (e.g. extratropical transient activities) are also improved considerably due to the imposed pre-processed initial 3-hour model drift corrections. Because many regional biases in the GEOS-5 CGCM are common amongst other current models, our approaches and findings are applicable to these other models as well.
Applying Hierarchical Model Calibration to Automatically Generated Items.
ERIC Educational Resources Information Center
Williamson, David M.; Johnson, Matthew S.; Sinharay, Sandip; Bejar, Isaac I.
This study explored the application of hierarchical model calibration as a means of reducing, if not eliminating, the need for pretesting of automatically generated items from a common item model prior to operational use. Ultimately the successful development of automatic item generation (AIG) systems capable of producing items with highly similar…
The Self-Help Group Model: A Review
ERIC Educational Resources Information Center
Jaques, Marceline E.; Patterson, Kathleen M.
1974-01-01
Self-help mutual aid groups are organized by peers who share a common problem. Through group identification, mutual support, and modeling, behavior is directed toward learning a new coping life style. The self-help group model is considered here as a viable and necessary part of a total rehabilitation service system. (Author)
Simulation of hardwood log sawing
D.B. Richards; W.K. Adkins; H. Hallock; E.H. Bulgrin
1979-01-01
Mathematical modeling computer programs for several hardwood sawing systems have been developed and are described. One has judgment capabilities. Several of the subroutines are common to all of the models. These models are the basis for further research which examines the question of best-grade sawing method in terms of lumber value yield.
Application of the GRC Stirling Convertor System Dynamic Model
NASA Technical Reports Server (NTRS)
Regan, Timothy F.; Lewandowski, Edward J.; Schreiber, Jeffrey G. (Technical Monitor)
2004-01-01
The GRC Stirling Convertor System Dynamic Model (SDM) has been developed to simulate dynamic performance of power systems incorporating free-piston Stirling convertors. This paper discusses its use in evaluating system dynamics and other systems concerns. Detailed examples are provided showing the use of the model in evaluation of off-nominal operating conditions. The many degrees of freedom in both the mechanical and electrical domains inherent in the Stirling convertor and the nonlinear dynamics make simulation an attractive analysis tool in conjunction with classical analysis. Application of SDM in studying the relationship of the size of the resonant circuit quality factor (commonly referred to as Q) in the various resonant mechanical and electrical sub-systems is discussed.
Application of optimization technique for flood damage modeling in river system
NASA Astrophysics Data System (ADS)
Barman, Sangita Deb; Choudhury, Parthasarathi
2018-04-01
A river system is defined as a network of channels that drains different parts of a basin uniting downstream to form a common outflow. An application of various models found in literatures, to a river system having multiple upstream flows is not always straight forward, involves a lengthy procedure; and with non-availability of data sets model calibration and applications may become difficult. In the case of a river system the flow modeling can be simplified to a large extent if the channel network is replaced by an equivalent single channel. In the present work optimization model formulations based on equivalent flow and applications of the mixed integer programming based pre-emptive goal programming model in evaluating flood control alternatives for a real life river system in India are proposed to be covered in the study.
The use of network theory to model disparate ship design information
NASA Astrophysics Data System (ADS)
Rigterink, Douglas; Piks, Rebecca; Singer, David J.
2014-06-01
This paper introduces the use of network theory to model and analyze disparate ship design information. This work will focus on a ship's distributed systems and their intra- and intersystem structures and interactions. The three system to be analyzed are: a passageway system, an electrical system, and a fire fighting system. These systems will be analyzed individually using common network metrics to glean information regarding their structures and attributes. The systems will also be subjected to community detection algorithms both separately and as a multiplex network to compare their similarities, differences, and interactions. Network theory will be shown to be useful in the early design stage due to its simplicity and ability to model any shipboard system.
Manning, H Charles; Buck, Jason R; Cook, Rebecca S
2016-02-01
Representing an enormous health care and socioeconomic challenge, breast cancer is the second most common cancer in the world and the second most common cause of cancer-related death. Although many of the challenges associated with preventing, treating, and ultimately curing breast cancer are addressable in the laboratory, successful translation of groundbreaking research to clinical populations remains an important barrier. Particularly when compared with research on other types of solid tumors, breast cancer research is hampered by a lack of tractable in vivo model systems that accurately recapitulate the relevant clinical features of the disease. A primary objective of this article was to provide a generalizable overview of the types of in vivo model systems, with an emphasis primarily on murine models, that are widely deployed in preclinical breast cancer research. Major opportunities to advance precision cancer medicine facilitated by molecular imaging of preclinical breast cancer models are discussed. © 2016 by the Society of Nuclear Medicine and Molecular Imaging, Inc.
The Systems Biology Markup Language (SBML): Language Specification for Level 3 Version 1 Core
Hucka, Michael; Bergmann, Frank T.; Hoops, Stefan; Keating, Sarah M.; Sahle, Sven; Schaff, James C.; Smith, Lucian P.; Wilkinson, Darren J.
2017-01-01
Summary Computational models can help researchers to interpret data, understand biological function, and make quantitative predictions. The Systems Biology Markup Language (SBML) is a file format for representing computational models in a declarative form that can be exchanged between different software systems. SBML is oriented towards describing biological processes of the sort common in research on a number of topics, including metabolic pathways, cell signaling pathways, and many others. By supporting SBML as an input/output format, different tools can all operate on an identical representation of a model, removing opportunities for translation errors and assuring a common starting point for analyses and simulations. This document provides the specification for Version 1 of SBML Level 3 Core. The specification defines the data structures prescribed by SBML as well as their encoding in XML, the eXtensible Markup Language. This specification also defines validation rules that determine the validity of an SBML document, and provides many examples of models in SBML form. Other materials and software are available from the SBML project web site, http://sbml.org/. PMID:26528564
The Systems Biology Markup Language (SBML): Language Specification for Level 3 Version 1 Core.
Hucka, Michael; Bergmann, Frank T; Hoops, Stefan; Keating, Sarah M; Sahle, Sven; Schaff, James C; Smith, Lucian P; Wilkinson, Darren J
2015-09-04
Computational models can help researchers to interpret data, understand biological function, and make quantitative predictions. The Systems Biology Markup Language (SBML) is a file format for representing computational models in a declarative form that can be exchanged between different software systems. SBML is oriented towards describing biological processes of the sort common in research on a number of topics, including metabolic pathways, cell signaling pathways, and many others. By supporting SBML as an input/output format, different tools can all operate on an identical representation of a model, removing opportunities for translation errors and assuring a common starting point for analyses and simulations. This document provides the specification for Version 1 of SBML Level 3 Core. The specification defines the data structures prescribed by SBML as well as their encoding in XML, the eXtensible Markup Language. This specification also defines validation rules that determine the validity of an SBML document, and provides many examples of models in SBML form. Other materials and software are available from the SBML project web site, http://sbml.org/.
The Systems Biology Markup Language (SBML): Language Specification for Level 3 Version 1 Core.
Hucka, Michael; Bergmann, Frank T; Hoops, Stefan; Keating, Sarah M; Sahle, Sven; Schaff, James C; Smith, Lucian P; Wilkinson, Darren J
2015-06-01
Computational models can help researchers to interpret data, understand biological function, and make quantitative predictions. The Systems Biology Markup Language (SBML) is a file format for representing computational models in a declarative form that can be exchanged between different software systems. SBML is oriented towards describing biological processes of the sort common in research on a number of topics, including metabolic pathways, cell signaling pathways, and many others. By supporting SBML as an input/output format, different tools can all operate on an identical representation of a model, removing opportunities for translation errors and assuring a common starting point for analyses and simulations. This document provides the specification for Version 1 of SBML Level 3 Core. The specification defines the data structures prescribed by SBML as well as their encoding in XML, the eXtensible Markup Language. This specification also defines validation rules that determine the validity of an SBML document, and provides many examples of models in SBML form. Other materials and software are available from the SBML project web site, http://sbml.org/.
Update on ORNL TRANSFORM Tool: Simulating Multi-Module Advanced Reactor with End-to-End I&C
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hale, Richard Edward; Fugate, David L.; Cetiner, Sacit M.
2015-05-01
The Small Modular Reactor (SMR) Dynamic System Modeling Tool project is in the fourth year of development. The project is designed to support collaborative modeling and study of various advanced SMR (non-light water cooled reactor) concepts, including the use of multiple coupled reactors at a single site. The focus of this report is the development of a steam generator and drum system model that includes the complex dynamics of typical steam drum systems, the development of instrumentation and controls for the steam generator with drum system model, and the development of multi-reactor module models that reflect the full power reactormore » innovative small module design concept. The objective of the project is to provide a common simulation environment and baseline modeling resources to facilitate rapid development of dynamic advanced reactor models; ensure consistency among research products within the Instrumentation, Controls, and Human-Machine Interface technical area; and leverage cross-cutting capabilities while minimizing duplication of effort. The combined simulation environment and suite of models are identified as the TRANSFORM tool. The critical elements of this effort include (1) defining a standardized, common simulation environment that can be applied throughout the Advanced Reactors Technology program; (2) developing a library of baseline component modules that can be assembled into full plant models using available geometry, design, and thermal-hydraulic data; (3) defining modeling conventions for interconnecting component models; and (4) establishing user interfaces and support tools to facilitate simulation development (i.e., configuration and parameterization), execution, and results display and capture.« less
An interactive display system for large-scale 3D models
NASA Astrophysics Data System (ADS)
Liu, Zijian; Sun, Kun; Tao, Wenbing; Liu, Liman
2018-04-01
With the improvement of 3D reconstruction theory and the rapid development of computer hardware technology, the reconstructed 3D models are enlarging in scale and increasing in complexity. Models with tens of thousands of 3D points or triangular meshes are common in practical applications. Due to storage and computing power limitation, it is difficult to achieve real-time display and interaction with large scale 3D models for some common 3D display software, such as MeshLab. In this paper, we propose a display system for large-scale 3D scene models. We construct the LOD (Levels of Detail) model of the reconstructed 3D scene in advance, and then use an out-of-core view-dependent multi-resolution rendering scheme to realize the real-time display of the large-scale 3D model. With the proposed method, our display system is able to render in real time while roaming in the reconstructed scene and 3D camera poses can also be displayed. Furthermore, the memory consumption can be significantly decreased via internal and external memory exchange mechanism, so that it is possible to display a large scale reconstructed scene with over millions of 3D points or triangular meshes in a regular PC with only 4GB RAM.
Cuellar De la Cruz, Yuri
2017-01-01
This article uses studies and organizational trends to understand available solutions to the lack of quality health care access, especially for the poor and needy of local U.S. communities. The U.S. healthcare system seems to be moving toward the World Health Organization's recommendation for universal health coverage for healthcare sustainability. Healthcare trends and offered solutions are varied. Christian healthcare traditionally implements works of mercy guided by a Christian ethos embracing the teachings of human dignity, solidarity, the common good, and subsidiarity. Culture of Life Ministries is one of many new sustainable U.S. healthcare models which implements Christ-centered health care to meet the need of quality and accessible health care for the local community. Culture of Life Ministries employs a model of charity care through volunteerism. Volunteer workers not only improve but also transform the local healthcare system into a personal healing ministry of the highest quality for every person. Summary: The lack of access to quality health care is a common problem in the U.S. despite various solutions offered through legislative and socioeconomic works: universal healthcare models, insurance models, and other business models. U.S. health care would be best transformed by returning to the implementation of a traditional system founded on the Christian principles of human dignity, solidarity, subsidiarity, and the common good. Culture of Life Ministries is an example of such a local ministry in Texas, which has found success in practically applying these Christ-centered, healthcare principles into an emerging not-for-profit, economically sustainable, healthcare model. PMID:28392598
Gravitational Waves from Accreting Neutron Stars Undergoing Common-envelope Inspiral
NASA Astrophysics Data System (ADS)
Holgado, A. Miguel; Ricker, Paul M.; Huerta, E. A.
2018-04-01
The common-envelope phase is a likely formation channel for close binary systems containing compact objects. Neutron stars in common envelopes accrete at a fraction of the Bondi–Hoyle–Lyttleton accretion rate, since the stellar envelope is inhomogeneous, but they may still be able to accrete at hypercritical rates (though not enough to become black holes). We show that common-envelope systems consisting of a neutron star with a massive primary may be gravitational-wave (GW) sources detectable in the Advanced LIGO band as far away as the Magellanic Clouds. To characterize their evolution, we perform orbital integrations using 1D models of 12 M ⊙ and 20 M ⊙ primaries, considering the effects of density gradient on the accretion onto the NS and spin evolution. From the range of possible accretion rates relevant to common-envelope evolution, we find that these systems may be louder GW sources than low-mass X-ray binaries like Sco X-1, which are currently the target of directed searches for continuous GWs. We also find that their strain amplitude signal may allow for novel constraints on the orbital separation and inspiral timescale in common envelopes when combined with pre-common-envelope electromagnetic observations.
Schizophrenia and Depression Co-Morbidity: What We have Learned from Animal Models
Samsom, James N.; Wong, Albert H. C.
2015-01-01
Patients with schizophrenia are at an increased risk for the development of depression. Overlap in the symptoms and genetic risk factors between the two disorders suggests a common etiological mechanism may underlie the presentation of comorbid depression in schizophrenia. Understanding these shared mechanisms will be important in informing the development of new treatments. Rodent models are powerful tools for understanding gene function as it relates to behavior. Examining rodent models relevant to both schizophrenia and depression reveals a number of common mechanisms. Current models which demonstrate endophenotypes of both schizophrenia and depression are reviewed here, including models of CUB and SUSHI multiple domains 1, PDZ and LIM domain 5, glutamate Delta 1 receptor, diabetic db/db mice, neuropeptide Y, disrupted in schizophrenia 1, and its interacting partners, reelin, maternal immune activation, and social isolation. Neurotransmission, brain connectivity, the immune system, the environment, and metabolism emerge as potential common mechanisms linking these models and potentially explaining comorbid depression in schizophrenia. PMID:25762938
Development of structural model of adaptive training complex in ergatic systems for professional use
NASA Astrophysics Data System (ADS)
Obukhov, A. D.; Dedov, D. L.; Arkhipov, A. E.
2018-03-01
The article considers the structural model of the adaptive training complex (ATC), which reflects the interrelations between the hardware, software and mathematical model of ATC and describes the processes in this subject area. The description of the main components of software and hardware complex, their interaction and functioning within the common system are given. Also the article scrutinizers a brief description of mathematical models of personnel activity, a technical system and influences, the interactions of which formalize the regularities of ATC functioning. The studies of main objects of training complexes and connections between them will make it possible to realize practical implementation of ATC in ergatic systems for professional use.
The common engine concept for ALS application - A cost reduction approach
NASA Technical Reports Server (NTRS)
Bair, E. K.; Schindler, C. M.
1989-01-01
Future launch systems require the application of propulsion systems which have been designed and developed to meet mission model needs while providing high degrees of reliability and cost effectiveness. Vehicle configurations which utilize different propellant combinations for booster and core stages can benefit from a common engine approach where a single engine design can be configured to operate on either set of propellants and thus serve as either a booster or core engine. Engine design concepts and mission application for a vehicle employing a common engine are discussed. Engine program cost estimates were made and cost savings, over the design and development of two unique engines, estimated.
The (Mathematical) Modeling Process in Biosciences.
Torres, Nestor V; Santos, Guido
2015-01-01
In this communication, we introduce a general framework and discussion on the role of models and the modeling process in the field of biosciences. The objective is to sum up the common procedures during the formalization and analysis of a biological problem from the perspective of Systems Biology, which approaches the study of biological systems as a whole. We begin by presenting the definitions of (biological) system and model. Particular attention is given to the meaning of mathematical model within the context of biology. Then, we present the process of modeling and analysis of biological systems. Three stages are described in detail: conceptualization of the biological system into a model, mathematical formalization of the previous conceptual model and optimization and system management derived from the analysis of the mathematical model. All along this work the main features and shortcomings of the process are analyzed and a set of rules that could help in the task of modeling any biological system are presented. Special regard is given to the formative requirements and the interdisciplinary nature of this approach. We conclude with some general considerations on the challenges that modeling is posing to current biology.
Traffic Flow of Interacting Self-Driven Particles: Rails and Trails, Vehicles and Vesicles
NASA Astrophysics Data System (ADS)
Chowdhury, Debashish
One common feature of a vehicle, an ant and a kinesin motor is that they all convert chemical energy, derived from fuel or food, into mechanical energy required for their forward movement; such objects have been modelled in recent years as self-driven particles. Cytoskeletal filaments, e.g., microtubules, form a rail network for intra-cellular transport of vesicular cargo by molecular motors like, for example, kinesins. Similarly, ants move along trails while vehicles move along lanes. Therefore, the traffic of vehicles and organisms as well as that of molecular motors can be modelled as systems of interacting self-driven particles; these are of current interest in non-equilibrium statistical mechanics. In this paper we point out the common features of these model systems and emphasize the crucial differences in their physical properties.
NASA Astrophysics Data System (ADS)
Gruszczynska, Marta; Rosat, Severine; Klos, Anna; Gruszczynski, Maciej; Bogusz, Janusz
2018-03-01
We described a spatio-temporal analysis of environmental loading models: atmospheric, continental hydrology, and non-tidal ocean changes, based on multichannel singular spectrum analysis (MSSA). We extracted the common annual signal for 16 different sections related to climate zones: equatorial, arid, warm, snow, polar and continents. We used the loading models estimated for a set of 229 ITRF2014 (International Terrestrial Reference Frame) International GNSS Service (IGS) stations and discussed the amount of variance explained by individual modes, proving that the common annual signal accounts for 16, 24 and 68% of the total variance of non-tidal ocean, atmospheric and hydrological loading models, respectively. Having removed the common environmental MSSA seasonal curve from the corresponding GPS position time series, we found that the residual station-specific annual curve modelled with the least-squares estimation has the amplitude of maximum 2 mm. This means that the environmental loading models underestimate the seasonalities observed by the GPS system. The remaining signal present in the seasonal frequency band arises from the systematic errors which are not of common environmental or geophysical origin. Using common mode error (CME) estimates, we showed that the direct removal of environmental loading models from the GPS series causes an artificial loss in the CME power spectra between 10 and 80 cycles per year. When environmental effect is removed from GPS series with MSSA curves, no influence on the character of spectra of CME estimates was noticed.
NASA Astrophysics Data System (ADS)
Gruszczynska, Marta; Rosat, Severine; Klos, Anna; Gruszczynski, Maciej; Bogusz, Janusz
2018-05-01
We described a spatio-temporal analysis of environmental loading models: atmospheric, continental hydrology, and non-tidal ocean changes, based on multichannel singular spectrum analysis (MSSA). We extracted the common annual signal for 16 different sections related to climate zones: equatorial, arid, warm, snow, polar and continents. We used the loading models estimated for a set of 229 ITRF2014 (International Terrestrial Reference Frame) International GNSS Service (IGS) stations and discussed the amount of variance explained by individual modes, proving that the common annual signal accounts for 16, 24 and 68% of the total variance of non-tidal ocean, atmospheric and hydrological loading models, respectively. Having removed the common environmental MSSA seasonal curve from the corresponding GPS position time series, we found that the residual station-specific annual curve modelled with the least-squares estimation has the amplitude of maximum 2 mm. This means that the environmental loading models underestimate the seasonalities observed by the GPS system. The remaining signal present in the seasonal frequency band arises from the systematic errors which are not of common environmental or geophysical origin. Using common mode error (CME) estimates, we showed that the direct removal of environmental loading models from the GPS series causes an artificial loss in the CME power spectra between 10 and 80 cycles per year. When environmental effect is removed from GPS series with MSSA curves, no influence on the character of spectra of CME estimates was noticed.
Design considerations, architecture, and use of the Mini-Sentinel distributed data system.
Curtis, Lesley H; Weiner, Mark G; Boudreau, Denise M; Cooper, William O; Daniel, Gregory W; Nair, Vinit P; Raebel, Marsha A; Beaulieu, Nicolas U; Rosofsky, Robert; Woodworth, Tiffany S; Brown, Jeffrey S
2012-01-01
We describe the design, implementation, and use of a large, multiorganizational distributed database developed to support the Mini-Sentinel Pilot Program of the US Food and Drug Administration (FDA). As envisioned by the US FDA, this implementation will inform and facilitate the development of an active surveillance system for monitoring the safety of medical products (drugs, biologics, and devices) in the USA. A common data model was designed to address the priorities of the Mini-Sentinel Pilot and to leverage the experience and data of participating organizations and data partners. A review of existing common data models informed the process. Each participating organization designed a process to extract, transform, and load its source data, applying the common data model to create the Mini-Sentinel Distributed Database. Transformed data were characterized and evaluated using a series of programs developed centrally and executed locally by participating organizations. A secure communications portal was designed to facilitate queries of the Mini-Sentinel Distributed Database and transfer of confidential data, analytic tools were developed to facilitate rapid response to common questions, and distributed querying software was implemented to facilitate rapid querying of summary data. As of July 2011, information on 99,260,976 health plan members was included in the Mini-Sentinel Distributed Database. The database includes 316,009,067 person-years of observation time, with members contributing, on average, 27.0 months of observation time. All data partners have successfully executed distributed code and returned findings to the Mini-Sentinel Operations Center. This work demonstrates the feasibility of building a large, multiorganizational distributed data system in which organizations retain possession of their data that are used in an active surveillance system. Copyright © 2012 John Wiley & Sons, Ltd.
Research on the Diesel Engine with Sliding Mode Variable Structure Theory
NASA Astrophysics Data System (ADS)
Ma, Zhexuan; Mao, Xiaobing; Cai, Le
2018-05-01
This study constructed the nonlinear mathematical model of the diesel engine high-pressure common rail (HPCR) system through two polynomial fitting which was treated as a kind of affine nonlinear system. Based on sliding-mode variable structure control (SMVSC) theory, a sliding-mode controller for affine nonlinear systems was designed for achieving the control of common rail pressure and the diesel engine’s rotational speed. Finally, on the simulation platform of MATLAB, the designed nonlinear HPCR system was simulated. The simulation results demonstrated that sliding-mode variable structure control algorithm shows favourable control performances which are overcoming the shortcomings of traditional PID control in overshoot, parameter adjustment, system precision, adjustment time and ascending time.
ERIC Educational Resources Information Center
VanLehn, Kurt; Chung, Greg; Grover, Sachin; Madni, Ayesha; Wetzel, Jon
2016-01-01
A common hypothesis is that students will more deeply understand dynamic systems and other complex phenomena if they construct computational models of them. Attempts to demonstrate the advantages of model construction have been stymied by the long time required for students to acquire skill in model construction. In order to make model…
ERIC Educational Resources Information Center
Butner, Jonathan; Amazeen, Polemnia G.; Mulvey, Genna M.
2005-01-01
The authors present a dynamical multilevel model that captures changes over time in the bidirectional, potentially asymmetric influence of 2 cyclical processes. S. M. Boker and J. Graham's (1998) differential structural equation modeling approach was expanded to the case of a nonlinear coupled oscillator that is common in bimanual coordination…
Process-based modelling of the nutritive value of forages: a review
USDA-ARS?s Scientific Manuscript database
Modelling sward nutritional value (NV) is of particular importance to understand the interactions between grasslands, livestock production, environment and climate-related impacts. Variables describing nutritive value vary significantly between ruminant production systems, but two types are commonly...
External Device to Incrementally Skid the Habitat (E-DISH)
NASA Technical Reports Server (NTRS)
Brazell, J. W.; Introne, Steve; Bedell, Lisa; Credle, Ben; Holp, Graham; Ly, Siao; Tait, Terry
1994-01-01
A Mars habitat transport system was designed as part of the NASA Mars exploration program. The transport system, the External Device to Incrementally Skid the Habitat (E - DISH), will be used to transport Mars habitats from their landing sites to the colony base and will be detached after unloading. The system requirements for Mars were calculated and scaled for model purposes. Specific model materials are commonly found and recommendations for materials for the Mars design are included.
Smoldyn on graphics processing units: massively parallel Brownian dynamics simulations.
Dematté, Lorenzo
2012-01-01
Space is a very important aspect in the simulation of biochemical systems; recently, the need for simulation algorithms able to cope with space is becoming more and more compelling. Complex and detailed models of biochemical systems need to deal with the movement of single molecules and particles, taking into consideration localized fluctuations, transportation phenomena, and diffusion. A common drawback of spatial models lies in their complexity: models can become very large, and their simulation could be time consuming, especially if we want to capture the systems behavior in a reliable way using stochastic methods in conjunction with a high spatial resolution. In order to deliver the promise done by systems biology to be able to understand a system as whole, we need to scale up the size of models we are able to simulate, moving from sequential to parallel simulation algorithms. In this paper, we analyze Smoldyn, a widely diffused algorithm for stochastic simulation of chemical reactions with spatial resolution and single molecule detail, and we propose an alternative, innovative implementation that exploits the parallelism of Graphics Processing Units (GPUs). The implementation executes the most computational demanding steps (computation of diffusion, unimolecular, and bimolecular reaction, as well as the most common cases of molecule-surface interaction) on the GPU, computing them in parallel on each molecule of the system. The implementation offers good speed-ups and real time, high quality graphics output
Tiedeman, Claire; Hill, Mary C.
2007-01-01
When simulating natural and engineered groundwater flow and transport systems, one objective is to produce a model that accurately represents important aspects of the true system. However, using direct measurements of system characteristics, such as hydraulic conductivity, to construct a model often produces simulated values that poorly match observations of the system state, such as hydraulic heads, flows and concentrations (for example, Barth et al., 2001). This occurs because of inaccuracies in the direct measurements and because the measurements commonly characterize system properties at different scales from that of the model aspect to which they are applied. In these circumstances, the conservation of mass equations represented by flow and transport models can be used to test the applicability of the direct measurements, such as by comparing model simulated values to the system state observations. This comparison leads to calibrating the model, by adjusting the model construction and the system properties as represented by model parameter values, so that the model produces simulated values that reasonably match the observations.
Impact of a hospitalist system on length of stay and cost for children with common conditions.
Srivastava, Rajendu; Landrigan, Christopher P; Ross-Degnan, Dennis; Soumerai, Stephen B; Homer, Charles J; Goldmann, Donald A; Muret-Wagstaff, Sharon
2007-08-01
This study examined mechanisms of efficiency in a managed care hospitalist system on length of stay and total costs for common pediatric conditions. We conducted a retrospective cohort study (October 1993 to July 1998) of patients in a not-for-profit staff model (HMO 1) and a non-staff-model (HMO 2) managed care organization at a freestanding children's hospital. HMO 1 introduced a hospitalist system for patients in October 1996. Patients were included if they had 1 of 3 common diagnoses: asthma, dehydration, or viral illness. Linear regression models examining length-of-stay-specific costs for prehospitalist and posthospitalist systems were built. Distribution of length of stay for each diagnosis before and after the system change in both study groups was calculated. Interrupted time series analysis tested whether changes in the trends of length of stay and total costs occurred after implementation of the hospitalist system by HMO1 (HMO 2 as comparison group) for all 3 diagnoses combined. A total of 1970 patients with 1 of the 3 study conditions were cared for in HMO 1, and 1001 in HMO 2. After the hospitalist system was introduced in HMO 1, length of stay was reduced by 0.23 days (13%) for asthma and 0.19 days (11%) for dehydration; there was no difference for patients with viral illness. The largest relative reduction in length of stay occurred in patients with a shorter length of stay whose hospitalizations were reduced from 2 days to 1 day. This shift resulted in an average cost-per-case reduction of $105.51 (9.3%) for patients with asthma and $86.22 (7.8%) for patients with dehydration. During the same period, length of stay and total cost rose in HMO 2. Introduction of a hospitalist system in one health maintenance organization resulted in earlier discharges and reduced costs for children with asthma and dehydration compared with another one, with the largest reductions occurring in reducing some 2-day hospitalizations to 1 day. These findings suggest that hospitalists can increase efficiency and reduce costs for children with common pediatric conditions.
[A web-based biomedical image mosaicing system].
Zhang, Meng; Yan, Zhuang-zhi; Pan, Zhi-jun; Shao, Shi-jie
2006-11-01
This paper describes a web service for biomedical image mosaicing. A web site based on CGI (Common Gateway Interface) is implemented. The system is based on Browser/Server model and is tested in www. Finally implementation examples and experiment results are provided.
Zachary A. Holden; W. Matt Jolly
2011-01-01
Fire danger rating systems commonly ignore fine scale, topographically-induced weather variations. These variations will likely create heterogeneous, landscape-scale fire danger conditions that have never been examined in detail. We modeled the evolution of fuel moistures and the Energy Release Component (ERC) from the US National Fire Danger Rating System across the...
Wind Turbine Blade CAD Models Used as Scaffolding Technique to Teach Design Engineers
ERIC Educational Resources Information Center
Irwin, John
2013-01-01
The Siemens PLM CAD software NX is commonly used for designing mechanical systems, and in complex systems such as the emerging area of wind power, the ability to have a model controlled by design parameters is a certain advantage. Formula driven expressions based on the amount of available wind in an area can drive the amount of effective surface…
Engine Structures Modeling Software System (ESMOSS)
NASA Technical Reports Server (NTRS)
1991-01-01
Engine Structures Modeling Software System (ESMOSS) is the development of a specialized software system for the construction of geometric descriptive and discrete analytical models of engine parts, components, and substructures which can be transferred to finite element analysis programs such as NASTRAN. The NASA Lewis Engine Structures Program is concerned with the development of technology for the rational structural design and analysis of advanced gas turbine engines with emphasis on advanced structural analysis, structural dynamics, structural aspects of aeroelasticity, and life prediction. Fundamental and common to all of these developments is the need for geometric and analytical model descriptions at various engine assembly levels which are generated using ESMOSS.
An intelligent training system for space shuttle flight controllers
NASA Technical Reports Server (NTRS)
Loftin, R. Bowen; Wang, Lui; Baffes, Paul; Hua, Grace
1988-01-01
An autonomous intelligent training system which integrates expert system technology with training/teaching methodologies is described. The system was designed to train Mission Control Center (MCC) Flight Dynamics Officers (FDOs) to deploy a certain type of satellite from the Space Shuttle. The Payload-assist module Deploys/Intelligent Computer-Aided Training (PD/ICAT) system consists of five components: a user interface, a domain expert, a training session manager, a trainee model, and a training scenario generator. The interface provides the trainee with information of the characteristics of the current training session and with on-line help. The domain expert (DeplEx for Deploy Expert) contains the rules and procedural knowledge needed by the FDO to carry out the satellite deploy. The DeplEx also contains mal-rules which permit the identification and diagnosis of common errors made by the trainee. The training session manager (TSM) examines the actions of the trainee and compares them with the actions of DeplEx in order to determine appropriate responses. A trainee model is developed for each individual using the system. The model includes a history of the trainee's interactions with the training system and provides evaluative data on the trainee's current skill level. A training scenario generator (TSG) designs appropriate training exercises for each trainee based on the trainee model and the training goals. All of the expert system components of PD/ICAT communicate via a common blackboard. The PD/ICAT is currently being tested. Ultimately, this project will serve as a vehicle for developing a general architecture for intelligent training systems together with a software environment for creating such systems.
An intelligent training system for space shuttle flight controllers
NASA Technical Reports Server (NTRS)
Loftin, R. Bowen; Wang, Lui; Baffles, Paul; Hua, Grace
1988-01-01
An autonomous intelligent training system which integrates expert system technology with training/teaching methodologies is described. The system was designed to train Mission Control Center (MCC) Flight Dynamics Officers (FDOs) to deploy a certain type of satellite from the Space Shuttle. The Payload-assist module Deploys/Intelligent Computer-Aided Training (PD/ICAT) system consists of five components: a user interface, a domain expert, a training session manager, a trainee model, and a training scenario generator. The interface provides the trainee with information of the characteristics of the current training session and with on-line help. The domain expert (Dep1Ex for Deploy Expert) contains the rules and procedural knowledge needed by the FDO to carry out the satellite deploy. The Dep1Ex also contains mal-rules which permit the identification and diagnosis of common errors made by the trainee. The training session manager (TSM) examines the actions of the trainee and compares them with the actions of Dep1Ex in order to determine appropriate responses. A trainee model is developed for each individual using the system. The model includes a history of the trainee's interactions with the training system and provides evaluative data on the trainee's current skill level. A training scenario generator (TSG) designs appropriate training exercises for each trainee based on the trainee model and the training goals. All of the expert system components of PD/ICAT communicate via a common blackboard. The PD/ICAT is currently being tested. Ultimately, this project will serve as a vehicle for developing a general architecture for intelligent training systems together with a software environment for creating such systems.
Neural models on temperature regulation for cold-stressed animals
NASA Technical Reports Server (NTRS)
Horowitz, J. M.
1975-01-01
The present review evaluates several assumptions common to a variety of current models for thermoregulation in cold-stressed animals. Three areas covered by the models are discussed: signals to and from the central nervous system (CNS), portions of the CNS involved, and the arrangement of neurons within networks. Assumptions in each of these categories are considered. The evaluation of the models is based on the experimental foundations of the assumptions. Regions of the nervous system concerned here include the hypothalamus, the skin, the spinal cord, the hippocampus, and the septal area of the brain.
Integrated Modeling Environment
NASA Technical Reports Server (NTRS)
Mosier, Gary; Stone, Paul; Holtery, Christopher
2006-01-01
The Integrated Modeling Environment (IME) is a software system that establishes a centralized Web-based interface for integrating people (who may be geographically dispersed), processes, and data involved in a common engineering project. The IME includes software tools for life-cycle management, configuration management, visualization, and collaboration.
NASA Astrophysics Data System (ADS)
Madani, K.; Dinar, A.
2013-12-01
Tragedy of the commons is generally recognized as one of the possible destinies for common pool resources (CPRs). To avoid the tragedy of the commons and prolonging the life of CPRs, users may show different behavioral characteristics and use different rationales for CPR planning and management. Furthermore, regulators may adopt different strategies for sustainable management of CPRs. The effectiveness of different regulatory exogenous management institutions cannot be evaluated through conventional CPR models since they assume that either users base their behavior on individual rationality and adopt a selfish behavior (Nash behavior), or that the users seek the system's optimal solution without giving priority to their own interests. Therefore, conventional models fail to reliably predict the outcome of CPR problems in which parties may have a range of behavioral characteristics, putting them somewhere in between the two types of behaviors traditionally considered. This work examines the effectiveness of different regulatory exogenous CPR management institutions through a user-based model (as opposed to a system-based model). The new modeling framework allows for consideration of sensitivity of the results to different behavioral characteristics of interacting CPR users. The suggested modeling approach is applied to a benchmark groundwater management problem. Results indicate that some well-known exogenous management institutions (e.g. taxing) are ineffective in sustainable management of CPRs in most cases. Bankruptcy-based management can be helpful, but determination of the fair level of cutbacks remains challenging under this type of institution. Furthermore, some bankruptcy rules such as the Constrained Equal Award (CEA) method are more beneficial to wealthier users, failing to establish social justice. Quota-based and CPR status-based management perform as the most promising and robust regulatory exogenous institutions in prolonging the CPR's life and increasing the long-term benefits to its users.
Identification of Low Order Equivalent System Models From Flight Test Data
NASA Technical Reports Server (NTRS)
Morelli, Eugene A.
2000-01-01
Identification of low order equivalent system dynamic models from flight test data was studied. Inputs were pilot control deflections, and outputs were aircraft responses, so the models characterized the total aircraft response including bare airframe and flight control system. Theoretical investigations were conducted and related to results found in the literature. Low order equivalent system modeling techniques using output error and equation error parameter estimation in the frequency domain were developed and validated on simulation data. It was found that some common difficulties encountered in identifying closed loop low order equivalent system models from flight test data could be overcome using the developed techniques. Implications for data requirements and experiment design were discussed. The developed methods were demonstrated using realistic simulation cases, then applied to closed loop flight test data from the NASA F-18 High Alpha Research Vehicle.
The famous five factors in teamwork: a case study of fratricide.
Rafferty, Laura A; Stanton, Neville A; Walker, Guy H
2010-10-01
The purpose of this paper is to propose foundations for a theory of errors in teamwork based upon analysis of a case study of fratricide alongside a review of the existing literature. This approach may help to promote a better understanding of interactions within complex systems and help in the formulation of hypotheses and predictions concerning errors in teamwork, particularly incidents of fratricide. It is proposed that a fusion of concepts drawn from error models, with common causal categories taken from teamwork models, could allow for an in-depth exploration of incidents of fratricide. It is argued that such a model has the potential to explore the core causal categories identified as present in an incident of fratricide. This view marks fratricide as a process of errors occurring throughout the military system as a whole, particularly due to problems in teamwork within this complex system. Implications of this viewpoint for the development of a new theory of fratricide are offered. STATEMENT OF RELEVANCE: This article provides an insight into the fusion of existing error and teamwork models for the analysis of an incident of fratricide. Within this paper, a number of commonalities among models of teamwork have been identified allowing for the development of a model.
NASA Astrophysics Data System (ADS)
Skersys, Tomas; Butleris, Rimantas; Kapocius, Kestutis
2013-10-01
Approaches for the analysis and specification of business vocabularies and rules are very relevant topics in both Business Process Management and Information Systems Development disciplines. However, in common practice of Information Systems Development, the Business modeling activities still are of mostly empiric nature. In this paper, basic aspects of the approach for business vocabularies' semi-automated extraction from business process models are presented. The approach is based on novel business modeling-level OMG standards "Business Process Model and Notation" (BPMN) and "Semantics for Business Vocabularies and Business Rules" (SBVR), thus contributing to OMG's vision about Model-Driven Architecture (MDA) and to model-driven development in general.
Metadata-Driven SOA-Based Application for Facilitation of Real-Time Data Warehousing
NASA Astrophysics Data System (ADS)
Pintar, Damir; Vranić, Mihaela; Skočir, Zoran
Service-oriented architecture (SOA) has already been widely recognized as an effective paradigm for achieving integration of diverse information systems. SOA-based applications can cross boundaries of platforms, operation systems and proprietary data standards, commonly through the usage of Web Services technology. On the other side, metadata is also commonly referred to as a potential integration tool given the fact that standardized metadata objects can provide useful information about specifics of unknown information systems with which one has interest in communicating with, using an approach commonly called "model-based integration". This paper presents the result of research regarding possible synergy between those two integration facilitators. This is accomplished with a vertical example of a metadata-driven SOA-based business process that provides ETL (Extraction, Transformation and Loading) and metadata services to a data warehousing system in need of a real-time ETL support.
Why involve families in acute mental healthcare? A collaborative conceptual review
Sandhu, Sima; Giacco, Domenico; Barrett, Katherine; Bennison, Gerry; Collinson, Sue; Priebe, Stefan
2017-01-01
Objectives Family involvement is strongly recommended in clinical guidelines but suffers from poor implementation. To explore this topic at a conceptual level, a multidisciplinary review team including academics, clinicians and individuals with lived experience undertook a review to explore the theoretical background of family involvement models in acute mental health treatment and how this relates to their delivery. Design A conceptual review was undertaken, including a systematic search and narrative synthesis. Included family models were mapped onto the most commonly referenced underlying theories: the diathesis–stress model, systems theories and postmodern theories of mental health. Common components of the models were summarised and compared. Lastly, a thematic analysis was undertaken to explore the role of patients and families in the delivery of the approaches. Setting General adult acute mental health treatment. Results Six distinct family involvement models were identified: Calgary Family Assessment and Intervention Models, ERIC (Equipe Rapide d’Intervention de Crise), Family Psychoeducation Models, Family Systems Approach, Open Dialogue and the Somerset Model. Findings indicated that despite wide variation in the theoretical models underlying family involvement models, there were many commonalities in their components, such as a focus on communication, language use and joint decision-making. Thematic analysis of the role of patients and families identified several issues for implementation. This included potential harms that could emerge during delivery of the models, such as imposing linear ‘patient–carer’ relationships and the risk of perceived coercion. Conclusions We conclude that future staff training may benefit from discussing the chosen family involvement model within the context of other theories of mental health. This may help to clarify the underlying purpose of family involvement and address the diverse needs and world views of patients, families and professionals in acute settings. PMID:28963308
Arabidopsis non-host resistance to powdery mildews.
Lipka, Ulrike; Fuchs, Rene; Lipka, Volker
2008-08-01
Immunity of an entire plant species against all genetic variants of a particular parasite is referred to as non-host resistance. Although non-host resistance represents the most common and durable form of plant resistance in nature, it has thus far been poorly understood at the molecular level. Recently, novel model systems have established the first mechanistic insights. The genetic dissection of Arabidopsis non-host resistance to non-adapted biotrophic powdery mildew fungi provided evidence for functionally redundant but operationally distinct pre- and post-invasion immune responses. Conceptually, these complex and successive defence mechanisms explain the durable and robust nature of non-host resistance. Pathogen lifestyle and infection biology, ecological parameters and the evolutionary relationship of the interaction partners determine differences and commonalities in other model systems.
Collaboration or negotiation: two ways of interacting suggest how shared thinking develops.
Mejía-Arauz, Rebeca; Rogoff, Barbara; Dayton, Andrew; Henne-Ochoa, Richard
2018-03-09
This paper contrasts two ways that shared thinking can be conceptualized: as negotiation, where individuals join their separate ideas, or collaboration, as people mutually engage together in a unified process, as an ensemble. We argue that these paradigms are culturally based, with the negotiation model fitting within an assumption system of separate entities-an assumption system we believe to be common in psychology and in middle-class European American society-and the collaboration model fitting within a holistic worldview that appears to be common in Indigenous-heritage communities of the Americas. We discuss cultural differences in children's interactions-as negotiation or collaboration-that suggest how these distinct paradigms develop. Copyright © 2018 Elsevier Ltd. All rights reserved.
DOT National Transportation Integrated Search
2017-06-01
The structural deterioration of aging infrastructure systems and the costs of repairing these systems is an increasingly important issue worldwide. Structural health monitoring (SHM), most commonly visual inspection and condition rating, has proven t...
A parsimonious land data assimilation system for the SMAP/GPM satellite era
USDA-ARS?s Scientific Manuscript database
Land data assimilation systems typically require complex parameterizations in order to: define required observation operators, quantify observing/forecasting errors and calibrate a land surface assimilation model. These parameters are commonly defined in an arbitrary manner and, if poorly specified,...
Application of fuzzy adaptive control to a MIMO nonlinear time-delay pump-valve system.
Lai, Zhounian; Wu, Peng; Wu, Dazhuan
2015-07-01
In this paper, a control strategy to balance the reliability against efficiency is introduced to overcome the common off-design operation problem in pump-valve systems. The pump-valve system is a nonlinear multi-input-multi-output (MIMO) system with time delays which cannot be accurately measured but can be approximately modeled using Bernoulli Principle. A fuzzy adaptive controller is applied to approximate system parameters and achieve the control of delay-free model since the system model is inaccurate and the direct feedback linearization method cannot be applied. An extended Smith predictor is introduced to compensate time delays of the system using the inaccurate system model. The experiment is carried out to verify the effectiveness of the control strategy whose results show that the control performance is well achieved. Copyright © 2015 ISA. Published by Elsevier Ltd. All rights reserved.
Calculating system reliability with SRFYDO
DOE Office of Scientific and Technical Information (OSTI.GOV)
Morzinski, Jerome; Anderson - Cook, Christine M; Klamann, Richard M
2010-01-01
SRFYDO is a process for estimating reliability of complex systems. Using information from all applicable sources, including full-system (flight) data, component test data, and expert (engineering) judgment, SRFYDO produces reliability estimates and predictions. It is appropriate for series systems with possibly several versions of the system which share some common components. It models reliability as a function of age and up to 2 other lifecycle (usage) covariates. Initial output from its Exploratory Data Analysis mode consists of plots and numerical summaries so that the user can check data entry and model assumptions, and help determine a final form for themore » system model. The System Reliability mode runs a complete reliability calculation using Bayesian methodology. This mode produces results that estimate reliability at the component, sub-system, and system level. The results include estimates of uncertainty, and can predict reliability at some not-too-distant time in the future. This paper presents an overview of the underlying statistical model for the analysis, discusses model assumptions, and demonstrates usage of SRFYDO.« less
Data Driven Model Development for the SuperSonic SemiSpan Transport (S(sup 4)T)
NASA Technical Reports Server (NTRS)
Kukreja, Sunil L.
2011-01-01
In this report, we will investigate two common approaches to model development for robust control synthesis in the aerospace community; namely, reduced order aeroservoelastic modelling based on structural finite-element and computational fluid dynamics based aerodynamic models, and a data-driven system identification procedure. It is shown via analysis of experimental SuperSonic SemiSpan Transport (S4T) wind-tunnel data that by using a system identification approach it is possible to estimate a model at a fixed Mach, which is parsimonious and robust across varying dynamic pressures.
Logic Model Checking of Time-Periodic Real-Time Systems
NASA Technical Reports Server (NTRS)
Florian, Mihai; Gamble, Ed; Holzmann, Gerard
2012-01-01
In this paper we report on the work we performed to extend the logic model checker SPIN with built-in support for the verification of periodic, real-time embedded software systems, as commonly used in aircraft, automobiles, and spacecraft. We first extended the SPIN verification algorithms to model priority based scheduling policies. Next, we added a library to support the modeling of periodic tasks. This library was used in a recent application of the SPIN model checker to verify the engine control software of an automobile, to study the feasibility of software triggers for unintended acceleration events.
Framework for a clinical information system.
Van De Velde, R; Lansiers, R; Antonissen, G
2002-01-01
The design and implementation of Clinical Information System architecture is presented. This architecture has been developed and implemented based on components following a strong underlying conceptual and technological model. Common Object Request Broker and n-tier technology featuring centralised and departmental clinical information systems as the back-end store for all clinical data are used. Servers located in the "middle" tier apply the clinical (business) model and application rules. The main characteristics are the focus on modelling and reuse of both data and business logic. Scalability as well as adaptability to constantly changing requirements via component driven computing are the main reasons for that approach.
ERIC Educational Resources Information Center
Stabback, Philip
2007-01-01
The article addresses the issue of possible curriculum models in post-conflict countries, taking as an example the case of Bosnia and Herzegovina between 1996 and 2004. Following the Dayton agreement, the education system in Bosnia and Herzegovina was divided between 13 ministries administering different Bosnian, Serb and Croat cantons. Despite…
Indonesia’s Electricity Demand Dynamic Modelling
NASA Astrophysics Data System (ADS)
Sulistio, J.; Wirabhuana, A.; Wiratama, M. G.
2017-06-01
Electricity Systems modelling is one of the emerging area in the Global Energy policy studies recently. System Dynamics approach and Computer Simulation has become one the common methods used in energy systems planning and evaluation in many conditions. On the other hand, Indonesia experiencing several major issues in Electricity system such as fossil fuel domination, demand - supply imbalances, distribution inefficiency, and bio-devastation. This paper aims to explain the development of System Dynamics modelling approaches and computer simulation techniques in representing and predicting electricity demand in Indonesia. In addition, this paper also described the typical characteristics and relationship of commercial business sector, industrial sector, and family / domestic sector as electricity subsystems in Indonesia. Moreover, it will be also present direct structure, behavioural, and statistical test as model validation approach and ended by conclusions.
State Analysis: A Control Architecture View of Systems Engineering
NASA Technical Reports Server (NTRS)
Rasmussen, Robert D.
2005-01-01
A viewgraph presentation on the state analysis process is shown. The topics include: 1) Issues with growing complexity; 2) Limits of common practice; 3) Exploiting a control point of view; 4) A glimpse at the State Analysis process; 5) Synergy with model-based systems engineering; and 6) Bridging the systems to software gap.
Intelligent Tutoring Systems: Formalization as Automata and Interface Design Using Neural Networks
ERIC Educational Resources Information Center
Curilem, S. Gloria; Barbosa, Andrea R.; de Azevedo, Fernando M.
2007-01-01
This article proposes a mathematical model of Intelligent Tutoring Systems (ITS), based on observations of the behaviour of these systems. One of the most important problems of pedagogical software is to establish a common language between the knowledge areas involved in their development, basically pedagogical, computing and domain areas. A…
Measuring transferring similarity via local information
NASA Astrophysics Data System (ADS)
Yin, Likang; Deng, Yong
2018-05-01
Recommender systems have developed along with the web science, and how to measure the similarity between users is crucial for processing collaborative filtering recommendation. Many efficient models have been proposed (i.g., the Pearson coefficient) to measure the direct correlation. However, the direct correlation measures are greatly affected by the sparsity of dataset. In other words, the direct correlation measures would present an inauthentic similarity if two users have a very few commonly selected objects. Transferring similarity overcomes this drawback by considering their common neighbors (i.e., the intermediates). Yet, the transferring similarity also has its drawback since it can only provide the interval of similarity. To break the limitations, we propose the Belief Transferring Similarity (BTS) model. The contributions of BTS model are: (1) BTS model addresses the issue of the sparsity of dataset by considering the high-order similarity. (2) BTS model transforms uncertain interval to a certain state based on fuzzy systems theory. (3) BTS model is able to combine the transferring similarity of different intermediates using information fusion method. Finally, we compare BTS models with nine different link prediction methods in nine different networks, and we also illustrate the convergence property and efficiency of the BTS model.
A network-base analysis of CMIP5 "historical" experiments
NASA Astrophysics Data System (ADS)
Bracco, A.; Foudalis, I.; Dovrolis, C.
2012-12-01
In computer science, "complex network analysis" refers to a set of metrics, modeling tools and algorithms commonly used in the study of complex nonlinear dynamical systems. Its main premise is that the underlying topology or network structure of a system has a strong impact on its dynamics and evolution. By allowing to investigate local and non-local statistical interaction, network analysis provides a powerful, but only marginally explored, framework to validate climate models and investigate teleconnections, assessing their strength, range, and impacts on the climate system. In this work we propose a new, fast, robust and scalable methodology to examine, quantify, and visualize climate sensitivity, while constraining general circulation models (GCMs) outputs with observations. The goal of our novel approach is to uncover relations in the climate system that are not (or not fully) captured by more traditional methodologies used in climate science and often adopted from nonlinear dynamical systems analysis, and to explain known climate phenomena in terms of the network structure or its metrics. Our methodology is based on a solid theoretical framework and employs mathematical and statistical tools, exploited only tentatively in climate research so far. Suitably adapted to the climate problem, these tools can assist in visualizing the trade-offs in representing global links and teleconnections among different data sets. Here we present the methodology, and compare network properties for different reanalysis data sets and a suite of CMIP5 coupled GCM outputs. With an extensive model intercomparison in terms of the climate network that each model leads to, we quantify how each model reproduces major teleconnections, rank model performances, and identify common or specific errors in comparing model outputs and observations.
Integrating WEPP into the WEPS infrastructure
USDA-ARS?s Scientific Manuscript database
The Wind Erosion Prediction System (WEPS) and the Water Erosion Prediction Project (WEPP) share a common modeling philosophy, that of moving away from primarily empirically based models based on indices or "average conditions", and toward a more process based approach which can be evaluated using ac...
A Model of Workflow Composition for Emergency Management
NASA Astrophysics Data System (ADS)
Xin, Chen; Bin-ge, Cui; Feng, Zhang; Xue-hui, Xu; Shan-shan, Fu
The common-used workflow technology is not flexible enough in dealing with concurrent emergency situations. The paper proposes a novel model for defining emergency plans, in which workflow segments appear as a constituent part. A formal abstraction, which contains four operations, is defined to compose workflow segments under constraint rule. The software system of the business process resources construction and composition is implemented and integrated into Emergency Plan Management Application System.
An Open Software Platform for Sharing Water Resource Models, Code and Data
NASA Astrophysics Data System (ADS)
Knox, Stephen; Meier, Philipp; Mohamed, Khaled; Korteling, Brett; Matrosov, Evgenii; Huskova, Ivana; Harou, Julien; Rosenberg, David; Tilmant, Amaury; Medellin-Azuara, Josue; Wicks, Jon
2016-04-01
The modelling of managed water resource systems requires new approaches in the face of increasing future uncertainty. Water resources management models, even if applied to diverse problem areas, use common approaches such as representing the problem as a network of nodes and links. We propose a data management software platform, called Hydra, that uses this commonality to allow multiple models using a node-link structure to be managed and run using a single software system. Hydra's user interface allows users to manage network topology and associated data. Hydra feeds this data directly into a model, importing from and exporting to different file formats using Apps. An App connects Hydra to a custom model, a modelling system such as GAMS or MATLAB or to different file formats such as MS Excel, CSV and ESRI Shapefiles. Hydra allows users to manage their data in a single, consistent place. Apps can be used to run domain-specific models and allow users to work with their own required file formats. The Hydra App Store offers a collaborative space where model developers can publish, review and comment on Apps, models and data. Example Apps and open-source libraries are available in a variety of languages (Python, Java and .NET). The App Store can act as a hub for water resource modellers to view and share Apps, models and data easily. This encourages an ecosystem of development using a shared platform, resulting in more model integration and potentially greater unity within resource modelling communities. www.hydraplatform.org www.hydraappstore.com
A common stochastic accumulator with effector-dependent noise can explain eye-hand coordination
Gopal, Atul; Viswanathan, Pooja
2015-01-01
The computational architecture that enables the flexible coupling between otherwise independent eye and hand effector systems is not understood. By using a drift diffusion framework, in which variability of the reaction time (RT) distribution scales with mean RT, we tested the ability of a common stochastic accumulator to explain eye-hand coordination. Using a combination of behavior, computational modeling and electromyography, we show how a single stochastic accumulator to threshold, followed by noisy effector-dependent delays, explains eye-hand RT distributions and their correlation, while an alternate independent, interactive eye and hand accumulator model does not. Interestingly, the common accumulator model did not explain the RT distributions of the same subjects when they made eye and hand movements in isolation. Taken together, these data suggest that a dedicated circuit underlies coordinated eye-hand planning. PMID:25568161
MI-Sim: A MATLAB package for the numerical analysis of microbial ecological interactions.
Wade, Matthew J; Oakley, Jordan; Harbisher, Sophie; Parker, Nicholas G; Dolfing, Jan
2017-01-01
Food-webs and other classes of ecological network motifs, are a means of describing feeding relationships between consumers and producers in an ecosystem. They have application across scales where they differ only in the underlying characteristics of the organisms and substrates describing the system. Mathematical modelling, using mechanistic approaches to describe the dynamic behaviour and properties of the system through sets of ordinary differential equations, has been used extensively in ecology. Models allow simulation of the dynamics of the various motifs and their numerical analysis provides a greater understanding of the interplay between the system components and their intrinsic properties. We have developed the MI-Sim software for use with MATLAB to allow a rigorous and rapid numerical analysis of several common ecological motifs. MI-Sim contains a series of the most commonly used motifs such as cooperation, competition and predation. It does not require detailed knowledge of mathematical analytical techniques and is offered as a single graphical user interface containing all input and output options. The tools available in the current version of MI-Sim include model simulation, steady-state existence and stability analysis, and basin of attraction analysis. The software includes seven ecological interaction motifs and seven growth function models. Unlike other system analysis tools, MI-Sim is designed as a simple and user-friendly tool specific to ecological population type models, allowing for rapid assessment of their dynamical and behavioural properties.
SPASE, Metadata, and the Heliophysics Virtual Observatories
NASA Technical Reports Server (NTRS)
Thieman, James; King, Todd; Roberts, Aaron
2010-01-01
To provide data search and access capability in the field of Heliophysics (the study of the Sun and its effects on the Solar System, especially the Earth) a number of Virtual Observatories (VO) have been established both via direct funding from the U.S. National Aeronautics and Space Administration (NASA) and through other funding agencies in the U.S. and worldwide. At least 15 systems can be labeled as Virtual Observatories in the Heliophysics community, 9 of them funded by NASA. The problem is that different metadata and data search approaches are used by these VO's and a search for data relevant to a particular research question can involve consulting with multiple VO's - needing to learn a different approach for finding and acquiring data for each. The Space Physics Archive Search and Extract (SPASE) project is intended to provide a common data model for Heliophysics data and therefore a common set of metadata for searches of the VO's. The SPASE Data Model has been developed through the common efforts of the Heliophysics Data and Model Consortium (HDMC) representatives over a number of years. We currently have released Version 2.1 of the Data Model. The advantages and disadvantages of the Data Model will be discussed along with the plans for the future. Recent changes requested by new members of the SPASE community indicate some of the directions for further development.
The Large Synoptic Survey Telescope OCS and TCS models
NASA Astrophysics Data System (ADS)
Schumacher, German; Delgado, Francisco
2010-07-01
The Large Synoptic Survey Telescope (LSST) is a project envisioned as a system of systems with demanding science, technical, and operational requirements, that must perform as a fully integrated unit. The design and implementation of such a system poses big engineering challenges when performing requirements analysis, detailed interface definitions, operational modes and control strategy studies. The OMG System Modeling Language (SysML) has been selected as the framework for the systems engineering analysis and documentation for the LSST. Models for the overall system architecture and different observatory subsystems have been built describing requirements, structure, interfaces and behavior. In this paper we show the models for the Observatory Control System (OCS) and the Telescope Control System (TCS), and how this methodology has helped in the clarification of the design and requirements. In one common language, the relationships of the OCS, TCS, Camera and Data management subsystems are captured with models of the structure, behavior, requirements and the traceability between them.
Modeling Common-Sense Decisions
NASA Astrophysics Data System (ADS)
Zak, Michail
This paper presents a methodology for efficient synthesis of dynamical model simulating a common-sense decision making process. The approach is based upon the extension of the physics' First Principles that includes behavior of living systems. The new architecture consists of motor dynamics simulating actual behavior of the object, and mental dynamics representing evolution of the corresponding knowledge-base and incorporating it in the form of information flows into the motor dynamics. The autonomy of the decision making process is achieved by a feedback from mental to motor dynamics. This feedback replaces unavailable external information by an internal knowledgebase stored in the mental model in the form of probability distributions.
Maximizing commonality between military and general aviation fly-by-light helicopter system designs
NASA Astrophysics Data System (ADS)
Enns, Russell; Mossman, David C.
1995-05-01
In the face of shrinking defense budgets, survival of the United States rotorcraft industry is becoming increasingly dependent on increased sales in a highly competitive civil helicopter market. As a result, only the most competitive rotorcraft manufacturers are likely to survive. A key ingredient in improving our competitive position is the ability to produce more versatile, high performance, high quality, and low cost of ownership helicopters. Fiber optic technology offers a path of achieving these objectives. Also, adopting common components and architectures for different helicopter models (while maintaining each models' uniqueness) will further decrease design and production costs. Funds saved (or generated) by exploiting this commonality can be applied to R&D used to further improve the product. In this paper, we define a fiber optics based avionics architecture which provides the pilot a fly-by-light / digital flight control system which can be implemented in both civilian and military helicopters. We then discuss the advantages of such an architecture.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Lorenz, J.C.; Warpinski, N.R.; Sattler, A.R.
A model is presented that suggests that regional fracture systems commonly control permeability in flat-lying reservoirs. Such fractures are distributed in a continuum of sizes and occur in subparallel, en echelon patterns. Few high-angle, orthogonal fractures exist because this system is created by high pore pressures and relatively low differential horizontal (tectonic) stresses rather than by significant structural deformation. Interfracture communication occurs primarily at infrequent, low-angle intersections of fractures. Vertical continuity of such fractures through a reservoir commonly is limited to the numerous lithologic discontinuities inherent in nonmarine sandstones. This type of fracture system has been documented in Mesaverede rocksmore » in the Rulison field of the Piceance Creek basin, northwestern Colorado, by studies of 4,300 ft (1310 m) of core from the U.S. DOE's three Multiwell Experiment (MWX) wells and by studies of the excellent nearby outcrops. Well test results and geologic data from core and outcrop support the model. The described natural fracture system has a significant effect on production and stimulation.« less
The (Mathematical) Modeling Process in Biosciences
Torres, Nestor V.; Santos, Guido
2015-01-01
In this communication, we introduce a general framework and discussion on the role of models and the modeling process in the field of biosciences. The objective is to sum up the common procedures during the formalization and analysis of a biological problem from the perspective of Systems Biology, which approaches the study of biological systems as a whole. We begin by presenting the definitions of (biological) system and model. Particular attention is given to the meaning of mathematical model within the context of biology. Then, we present the process of modeling and analysis of biological systems. Three stages are described in detail: conceptualization of the biological system into a model, mathematical formalization of the previous conceptual model and optimization and system management derived from the analysis of the mathematical model. All along this work the main features and shortcomings of the process are analyzed and a set of rules that could help in the task of modeling any biological system are presented. Special regard is given to the formative requirements and the interdisciplinary nature of this approach. We conclude with some general considerations on the challenges that modeling is posing to current biology. PMID:26734063
Can Moral Hazard Be Resolved by Common-Knowledge in S4n-Knowledge?
NASA Astrophysics Data System (ADS)
Matsuhisa, Takashi
This article investigates the relationship between common-knowledge and agreement in multi-agent system, and to apply the agreement result by common-knowledge to the principal-agent model under non-partition information. We treat the two problems: (1) how we capture the fact that the agents agree on an event or they get consensus on it from epistemic point of view, and (2) how the agreement theorem will be able to make progress to settle a moral hazard problem in the principal-agents model under non-partition information. We shall propose a solution program for the moral hazard in the principal-agents model under non-partition information by common-knowledge. Let us start that the agents have the knowledge structure induced from a reflexive and transitive relation associated with the multi-modal logic S4n. Each agent obtains the membership value of an event under his/her private information, so he/she considers the event as fuzzy set. Specifically consider the situation that the agents commonly know all membership values of the other agents. In this circumstance we shall show the agreement theorem that consensus on the membership values among all agents can still be guaranteed. Furthermore, under certain assumptions we shall show that the moral hazard can be resolved in the principal-agent model when all the expected marginal costs are common-knowledge among the principal and agents.
NASA Astrophysics Data System (ADS)
Abdeljabbar Kharrat, Nourhene; Plateaux, Régis; Miladi Chaabane, Mariem; Choley, Jean-Yves; Karra, Chafik; Haddar, Mohamed
2018-05-01
The present work tackles the modeling of multi-physics systems applying a topological approach while proceeding with a new methodology using a topological modification to the structure of systems. Then the comparison with the Magos' methodology is made. Their common ground is the use of connectivity within systems. The comparison and analysis of the different types of modeling show the importance of the topological methodology through the integration of the topological modification to the topological structure of a multi-physics system. In order to validate this methodology, the case of Pogo-stick is studied. The first step consists in generating a topological graph of the system. Then the connectivity step takes into account the contact with the ground. During the last step of this research; the MGS language (Modeling of General System) is used to model the system through equations. Finally, the results are compared to those obtained by MODELICA. Therefore, this proposed methodology may be generalized to model multi-physics systems that can be considered as a set of local elements.
A hybrid agent-based approach for modeling microbiological systems.
Guo, Zaiyi; Sloot, Peter M A; Tay, Joc Cing
2008-11-21
Models for systems biology commonly adopt Differential Equations or Agent-Based modeling approaches for simulating the processes as a whole. Models based on differential equations presuppose phenomenological intracellular behavioral mechanisms, while models based on Multi-Agent approach often use directly translated, and quantitatively less precise if-then logical rule constructs. We propose an extendible systems model based on a hybrid agent-based approach where biological cells are modeled as individuals (agents) while molecules are represented by quantities. This hybridization in entity representation entails a combined modeling strategy with agent-based behavioral rules and differential equations, thereby balancing the requirements of extendible model granularity with computational tractability. We demonstrate the efficacy of this approach with models of chemotaxis involving an assay of 10(3) cells and 1.2x10(6) molecules. The model produces cell migration patterns that are comparable to laboratory observations.
Comparison of digital elevation models for aquatic data development.
Sharon Clarke; Kelly Burnett
2003-01-01
Thirty-meter digital elevation models (DEMs) produced by the U.S. Geological Survey (USGS) are widely available and commonly used in analyzing aquatic systems. However, these DEMs are of relatively coarse resolution, were inconsistently produced (i.e., Level 1 versus Level 2 DEMs), and lack drainage enforcement. Such issues may hamper efforts to accurately model...
NASA Astrophysics Data System (ADS)
Kadow, Christopher; Illing, Sebastian; Kunst, Oliver; Schartner, Thomas; Kirchner, Ingo; Rust, Henning W.; Cubasch, Ulrich; Ulbrich, Uwe
2016-04-01
The Freie Univ Evaluation System Framework (Freva - freva.met.fu-berlin.de) is a software infrastructure for standardized data and tool solutions in Earth system science. Freva runs on high performance computers to handle customizable evaluation systems of research projects, institutes or universities. It combines different software technologies into one common hybrid infrastructure, including all features present in the shell and web environment. The database interface satisfies the international standards provided by the Earth System Grid Federation (ESGF). Freva indexes different data projects into one common search environment by storing the meta data information of the self-describing model, reanalysis and observational data sets in a database. This implemented meta data system with its advanced but easy-to-handle search tool supports users, developers and their plugins to retrieve the required information. A generic application programming interface (API) allows scientific developers to connect their analysis tools with the evaluation system independently of the programming language used. Users of the evaluation techniques benefit from the common interface of the evaluation system without any need to understand the different scripting languages. Facilitation of the provision and usage of tools and climate data automatically increases the number of scientists working with the data sets and identifying discrepancies. The integrated web-shell (shellinabox) adds a degree of freedom in the choice of the working environment and can be used as a gate to the research projects HPC. Plugins are able to integrate their e.g. post-processed results into the database of the user. This allows e.g. post-processing plugins to feed statistical analysis plugins, which fosters an active exchange between plugin developers of a research project. Additionally, the history and configuration sub-system stores every analysis performed with the evaluation system in a database. Configurations and results of the tools can be shared among scientists via shell or web system. Therefore, plugged-in tools benefit from transparency and reproducibility. Furthermore, if configurations match while starting an evaluation plugin, the system suggests to use results already produced by other users - saving CPU/h, I/O, disk space and time. The efficient interaction between different technologies improves the Earth system modeling science framed by Freva.
NASA Astrophysics Data System (ADS)
Kadow, C.; Illing, S.; Schartner, T.; Grieger, J.; Kirchner, I.; Rust, H.; Cubasch, U.; Ulbrich, U.
2017-12-01
The Freie Univ Evaluation System Framework (Freva - freva.met.fu-berlin.de) is a software infrastructure for standardized data and tool solutions in Earth system science (e.g. www-miklip.dkrz.de, cmip-eval.dkrz.de). Freva runs on high performance computers to handle customizable evaluation systems of research projects, institutes or universities. It combines different software technologies into one common hybrid infrastructure, including all features present in the shell and web environment. The database interface satisfies the international standards provided by the Earth System Grid Federation (ESGF). Freva indexes different data projects into one common search environment by storing the meta data information of the self-describing model, reanalysis and observational data sets in a database. This implemented meta data system with its advanced but easy-to-handle search tool supports users, developers and their plugins to retrieve the required information. A generic application programming interface (API) allows scientific developers to connect their analysis tools with the evaluation system independently of the programming language used. Users of the evaluation techniques benefit from the common interface of the evaluation system without any need to understand the different scripting languages. The integrated web-shell (shellinabox) adds a degree of freedom in the choice of the working environment and can be used as a gate to the research projects HPC. Plugins are able to integrate their e.g. post-processed results into the database of the user. This allows e.g. post-processing plugins to feed statistical analysis plugins, which fosters an active exchange between plugin developers of a research project. Additionally, the history and configuration sub-system stores every analysis performed with the evaluation system in a database. Configurations and results of the tools can be shared among scientists via shell or web system. Furthermore, if configurations match while starting an evaluation plugin, the system suggests to use results already produced by other users - saving CPU/h, I/O, disk space and time. The efficient interaction between different technologies improves the Earth system modeling science framed by Freva.
pyBSM: A Python package for modeling imaging systems
NASA Astrophysics Data System (ADS)
LeMaster, Daniel A.; Eismann, Michael T.
2017-05-01
There are components that are common to all electro-optical and infrared imaging system performance models. The purpose of the Python Based Sensor Model (pyBSM) is to provide open source access to these functions for other researchers to build upon. Specifically, pyBSM implements much of the capability found in the ERIM Image Based Sensor Model (IBSM) V2.0 along with some improvements. The paper also includes two use-case examples. First, performance of an airborne imaging system is modeled using the General Image Quality Equation (GIQE). The results are then decomposed into factors affecting noise and resolution. Second, pyBSM is paired with openCV to evaluate performance of an algorithm used to detect objects in an image.
A hierarchical distributed control model for coordinating intelligent systems
NASA Technical Reports Server (NTRS)
Adler, Richard M.
1991-01-01
A hierarchical distributed control (HDC) model for coordinating cooperative problem-solving among intelligent systems is described. The model was implemented using SOCIAL, an innovative object-oriented tool for integrating heterogeneous, distributed software systems. SOCIAL embeds applications in 'wrapper' objects called Agents, which supply predefined capabilities for distributed communication, control, data specification, and translation. The HDC model is realized in SOCIAL as a 'Manager'Agent that coordinates interactions among application Agents. The HDC Manager: indexes the capabilities of application Agents; routes request messages to suitable server Agents; and stores results in a commonly accessible 'Bulletin-Board'. This centralized control model is illustrated in a fault diagnosis application for launch operations support of the Space Shuttle fleet at NASA, Kennedy Space Center.
Mouse Models for Studying Oral Cancer: Impact in the Era of Cancer Immunotherapy.
Luo, J J; Young, C D; Zhou, H M; Wang, X J
2018-04-01
Model systems for oral cancer research have progressed from tumor epithelial cell cultures to in vivo systems that mimic oral cancer genetics, pathological characteristics, and tumor-stroma interactions of oral cancer patients. In the era of cancer immunotherapy, it is imperative to use model systems to test oral cancer prevention and therapeutic interventions in the presence of an immune system and to discover mechanisms of stromal contributions to oral cancer carcinogenesis. Here, we review in vivo mouse model systems commonly used for studying oral cancer and discuss the impact these models are having in advancing basic mechanisms, chemoprevention, and therapeutic intervention of oral cancer while highlighting recent discoveries concerning the role of immune cells in oral cancer. Improvements to in vivo model systems that highly recapitulate human oral cancer hold the key to identifying features of oral cancer initiation, progression, and invasion as well as molecular and cellular targets for prevention, therapeutic response, and immunotherapy development.
The Systems Biology Markup Language (SBML): Language Specification for Level 3 Version 2 Core.
Hucka, Michael; Bergmann, Frank T; Dräger, Andreas; Hoops, Stefan; Keating, Sarah M; Le Novère, Nicolas; Myers, Chris J; Olivier, Brett G; Sahle, Sven; Schaff, James C; Smith, Lucian P; Waltemath, Dagmar; Wilkinson, Darren J
2018-03-09
Computational models can help researchers to interpret data, understand biological functions, and make quantitative predictions. The Systems Biology Markup Language (SBML) is a file format for representing computational models in a declarative form that different software systems can exchange. SBML is oriented towards describing biological processes of the sort common in research on a number of topics, including metabolic pathways, cell signaling pathways, and many others. By supporting SBML as an input/output format, different tools can all operate on an identical representation of a model, removing opportunities for translation errors and assuring a common starting point for analyses and simulations. This document provides the specification for Version 2 of SBML Level 3 Core. The specification defines the data structures prescribed by SBML, their encoding in XML (the eXtensible Markup Language), validation rules that determine the validity of an SBML document, and examples of models in SBML form. The design of Version 2 differs from Version 1 principally in allowing new MathML constructs, making more child elements optional, and adding identifiers to all SBML elements instead of only selected elements. Other materials and software are available from the SBML project website at http://sbml.org/.
MyOcean Internal Information System (Dial-P)
NASA Astrophysics Data System (ADS)
Blanc, Frederique; Jolibois, Tony; Loubrieu, Thomas; Manzella, Giuseppe; Mazzetti, Paolo; Nativi, Stefano
2010-05-01
MyOcean is a three-year project (2008-2011) which goal is the development and pre-operational validation of the GMES Marine Core Service for ocean monitoring and forecasting. It's a transition project that will conduct the European "operational oceanography" community towards the operational phase of a GMES European service, which demands more European integration, more operationality, and more service. Observations, model-based data, and added-value products will be generated - and enhanced thanks to dedicated expertise - by the following production units: • Five Thematic Assembly Centers, each of them dealing with a specific set of observation data: Sea Level, Ocean colour, Sea Surface Temperature, Sea Ice & Wind, and In Situ data, • Seven Monitoring and Forecasting Centers to serve the Global Ocean, the Arctic area, the Baltic Sea, the Atlantic North-West shelves area, the Atlantic Iberian-Biscay-Ireland area, the Mediterranean Sea and the Black sea. Intermediate and final users will discover, view and get the products by means of a central web desk, a central re-active manned service desk and thematic experts distributed across Europe. The MyOcean Information System (MIS) is considering the various aspects of an interoperable - federated information system. Data models support data and computer systems by providing the definition and format of data. The possibility of including the information in the data file is depending on data model adopted. In general there is little effort in the actual project to develop a ‘generic' data model. A strong push to develop a common model is provided by the EU Directive INSPIRE. At present, there is no single de-facto data format for storing observational data. Data formats are still evolving, with their underlying data models moving towards the concept of Feature Types based on ISO/TC211 standards. For example, Unidata are developing the Common Data Model that can represent scientific data types such as point, trajectory, station, grid, etc., which will be implemented in netCDF format. SeaDataNet is recommending ODV and NetCDF formats. Another problem related to data curation and interoperability is the possibility to use common vocabularies. Common vocabularies are developed in many international initiatives, such as GEMET (promoted by INSPIRE as a multilingual thesaurus), UNIDATA, SeaDataNet, Marine Metadata Initiative (MMI). MIS is considering the SeaDataNet vocabulary as a base for interoperability. Four layers of different abstraction levels of interoperability an be defined: - Technical/basic: this layer is implemented at each TAC or MFC through internet connection and basic services for data transfer and browsing (e.g FTP, HTTP, etc). - Syntactic: allowing the interchange of metadata and protocol elements. This layer corresponds to a definition Core Metadata Set, the format of exchange/delivery for the data and associated metadata and possible software. This layer is implemented by the DIAL-P logical interface (e.g. adoption of INSPIRE compliant metadata set and common data formats). - Functional/pragmatic: based on a common set of functional primitives or on a common set of service definitions. This layer refers to the definition of services based on Web services standards. This layer is implemented by the DIAL-P logical interface (e.g. adoption of INSPIRE compliant network services). - Semantic: allowing to access similar classes of objects and services across multiple sites, with multilinguality of content as one specific aspect. This layer corresponds to MIS interface, terminology and thesaurus. Given the above requirements, the proposed solution is a federation of systems, where the individual participants are self-contained autonomous systems, but together form a consistent wider picture. A mid-tier integration layer mediates between existing systems, adapting their data and service model schema to the MIS. The developed MIS is a read-only system, i.e. does not allow updating (or inserting) data into the participant resource systems. The main advantages of the proposed approach are: • to enable information sources to join the MIS and publish their data and metadata in a secure way, without any modification to their existing resources and procedures and without any restriction to their autonomy; • to enable users to browse and query the MIS, receiving an aggregated result incorporating relevant data and metadata from across different sources; • to accommodate the growth of such a MIS, either in terms of its clients or of its information resources, as well as the evolution of the underlying data model.
A UML model for the description of different brain-computer interface systems.
Quitadamo, Lucia Rita; Abbafati, Manuel; Saggio, Giovanni; Marciani, Maria Grazia; Cardarilli, Gian Carlo; Bianchi, Luigi
2008-01-01
BCI research lacks a universal descriptive language among labs and a unique standard model for the description of BCI systems. This results in a serious problem in comparing performances of different BCI processes and in unifying tools and resources. In such a view we implemented a Unified Modeling Language (UML) model for the description virtually of any BCI protocol and we demonstrated that it can be successfully applied to the most common ones such as P300, mu-rhythms, SCP, SSVEP, fMRI. Finally we illustrated the advantages in utilizing a standard terminology for BCIs and how the same basic structure can be successfully adopted for the implementation of new systems.
Threat Assessment & Remediation Analysis (TARA): Methodology Description Version 1.0
2011-10-01
collectively support this practice. v Table of Contents 1 Introduction...4 1.3.2.3 Common Vulnerability Scoring System (CVSS) ........................................ 4 1.3.2.4 Microsoft Threat Modeling ...6 2.1.1.3 Eliminate Implausible TTPs ........................................................................ 6 2.1.1.4 Apply Scoring Model
Modeling crop residue burning experiments to evaluate smoke emissions and plume transport
Crop residue burning is a common land management practice that results in emissions of a variety of pollutants with negative health impacts. Modeling systems are used to estimate air quality impacts of crop residue burning to support retrospective regulatory assessments and also ...
Simulation and measurement of nonlinear behavior in a high-power test cell.
Harvey, Gerald; Gachagan, Anthony
2011-04-01
High-power ultrasound has many diverse uses in process applications in industries ranging from food to pharmaceutical. Because cavitation is frequently a desirable effect within many high-power, low-frequency systems, these systems are commonly expected to feature highly nonlinear acoustic propagation because of the high input levels employed. This generation of harmonics significantly alters the field profile compared with that of a linear system, making accurate field modeling difficult. However, when the short propagation distances involved are considered, it is not unreasonable to assume that these systems may remain largely linear until the onset of cavitation, in terms of classical acoustic propagation. The purpose of this paper is to investigate the possible nonlinear effects within such systems before the onset of cavitation. A theoretical description of nonlinear propagation will be presented and the merits of common analytical models will be discussed. Following this, a numerical model of nonlinearity will be outlined and the advantages it presents for representing nonlinear effects in bounded fields will be discussed. Next, the driving equipment and transducers will be evaluated for linearity to disengage any effects from those formed in the transmission load. Finally, the linearity of the system will be measured using an acoustic hydrophone and compared with finite element analysis to confirm that nonlinear effects are not prevalent in such systems at the onset of cavitation. © 2011 IEEE
Hay, L.; Knapp, L.
1996-01-01
Investigating natural, potential, and man-induced impacts on hydrological systems commonly requires complex modelling with overlapping data requirements, and massive amounts of one- to four-dimensional data at multiple scales and formats. Given the complexity of most hydrological studies, the requisite software infrastructure must incorporate many components including simulation modelling, spatial analysis and flexible, intuitive displays. There is a general requirement for a set of capabilities to support scientific analysis which, at this time, can only come from an integration of several software components. Integration of geographic information systems (GISs) and scientific visualization systems (SVSs) is a powerful technique for developing and analysing complex models. This paper describes the integration of an orographic precipitation model, a GIS and a SVS. The combination of these individual components provides a robust infrastructure which allows the scientist to work with the full dimensionality of the data and to examine the data in a more intuitive manner.
Virus-mediated EpoR76E Therapy Slows Optic Nerve Axonopathy in Experimental Glaucoma.
Bond, Wesley S; Hines-Beard, Jessica; GoldenMerry, YPaul L; Davis, Mara; Farooque, Alma; Sappington, Rebecca M; Calkins, David J; Rex, Tonia S
2016-02-01
Glaucoma, a common cause of blindness, is currently treated by intraocular pressure (IOP)-lowering interventions. However, this approach is insufficient to completely prevent vision loss. Here, we evaluate an IOP-independent gene therapy strategy using a modified erythropoietin, EPO-R76E, which has reduced erythropoietic function. We used two models of glaucoma, the murine microbead occlusion model and the DBA/2J mouse. Systemic recombinant adeno-associated virus-mediated gene delivery of EpoR76E (rAAV.EpoR76E) was performed concurrent with elevation of IOP. Axon structure and active anterograde transport were preserved in both models. Vision, as determined by the flash visual evoked potential, was preserved in the DBA/2J. These results show that systemic EpoR76E gene therapy protects retinal ganglion cells from glaucomatous degeneration in two different models. This suggests that EPO targets a component of the neurodegenerative pathway that is common to both models. The efficacy of rAAV.EpoR76E delivered at onset of IOP elevation supports clinical relevance of this treatment.
Multiscale Methods for Accurate, Efficient, and Scale-Aware Models of the Earth System
DOE Office of Scientific and Technical Information (OSTI.GOV)
Goldhaber, Steve; Holland, Marika
The major goal of this project was to contribute improvements to the infrastructure of an Earth System Model in order to support research in the Multiscale Methods for Accurate, Efficient, and Scale-Aware models of the Earth System project. In support of this, the NCAR team accomplished two main tasks: improving input/output performance of the model and improving atmospheric model simulation quality. Improvement of the performance and scalability of data input and diagnostic output within the model required a new infrastructure which can efficiently handle the unstructured grids common in multiscale simulations. This allows for a more computationally efficient model, enablingmore » more years of Earth System simulation. The quality of the model simulations was improved by reducing grid-point noise in the spectral element version of the Community Atmosphere Model (CAM-SE). This was achieved by running the physics of the model using grid-cell data on a finite-volume grid.« less
THE EARTH SYSTEM PREDICTION SUITE: Toward a Coordinated U.S. Modeling Capability
Theurich, Gerhard; DeLuca, C.; Campbell, T.; Liu, F.; Saint, K.; Vertenstein, M.; Chen, J.; Oehmke, R.; Doyle, J.; Whitcomb, T.; Wallcraft, A.; Iredell, M.; Black, T.; da Silva, AM; Clune, T.; Ferraro, R.; Li, P.; Kelley, M.; Aleinov, I.; Balaji, V.; Zadeh, N.; Jacob, R.; Kirtman, B.; Giraldo, F.; McCarren, D.; Sandgathe, S.; Peckham, S.; Dunlap, R.
2017-01-01
The Earth System Prediction Suite (ESPS) is a collection of flagship U.S. weather and climate models and model components that are being instrumented to conform to interoperability conventions, documented to follow metadata standards, and made available either under open source terms or to credentialed users. The ESPS represents a culmination of efforts to create a common Earth system model architecture, and the advent of increasingly coordinated model development activities in the U.S. ESPS component interfaces are based on the Earth System Modeling Framework (ESMF), community-developed software for building and coupling models, and the National Unified Operational Prediction Capability (NUOPC) Layer, a set of ESMF-based component templates and interoperability conventions. This shared infrastructure simplifies the process of model coupling by guaranteeing that components conform to a set of technical and semantic behaviors. The ESPS encourages distributed, multi-agency development of coupled modeling systems, controlled experimentation and testing, and exploration of novel model configurations, such as those motivated by research involving managed and interactive ensembles. ESPS codes include the Navy Global Environmental Model (NavGEM), HYbrid Coordinate Ocean Model (HYCOM), and Coupled Ocean Atmosphere Mesoscale Prediction System (COAMPS®); the NOAA Environmental Modeling System (NEMS) and the Modular Ocean Model (MOM); the Community Earth System Model (CESM); and the NASA ModelE climate model and GEOS-5 atmospheric general circulation model. PMID:29568125
THE EARTH SYSTEM PREDICTION SUITE: Toward a Coordinated U.S. Modeling Capability.
Theurich, Gerhard; DeLuca, C; Campbell, T; Liu, F; Saint, K; Vertenstein, M; Chen, J; Oehmke, R; Doyle, J; Whitcomb, T; Wallcraft, A; Iredell, M; Black, T; da Silva, A M; Clune, T; Ferraro, R; Li, P; Kelley, M; Aleinov, I; Balaji, V; Zadeh, N; Jacob, R; Kirtman, B; Giraldo, F; McCarren, D; Sandgathe, S; Peckham, S; Dunlap, R
2016-07-01
The Earth System Prediction Suite (ESPS) is a collection of flagship U.S. weather and climate models and model components that are being instrumented to conform to interoperability conventions, documented to follow metadata standards, and made available either under open source terms or to credentialed users. The ESPS represents a culmination of efforts to create a common Earth system model architecture, and the advent of increasingly coordinated model development activities in the U.S. ESPS component interfaces are based on the Earth System Modeling Framework (ESMF), community-developed software for building and coupling models, and the National Unified Operational Prediction Capability (NUOPC) Layer, a set of ESMF-based component templates and interoperability conventions. This shared infrastructure simplifies the process of model coupling by guaranteeing that components conform to a set of technical and semantic behaviors. The ESPS encourages distributed, multi-agency development of coupled modeling systems, controlled experimentation and testing, and exploration of novel model configurations, such as those motivated by research involving managed and interactive ensembles. ESPS codes include the Navy Global Environmental Model (NavGEM), HYbrid Coordinate Ocean Model (HYCOM), and Coupled Ocean Atmosphere Mesoscale Prediction System (COAMPS ® ); the NOAA Environmental Modeling System (NEMS) and the Modular Ocean Model (MOM); the Community Earth System Model (CESM); and the NASA ModelE climate model and GEOS-5 atmospheric general circulation model.
The Earth System Prediction Suite: Toward a Coordinated U.S. Modeling Capability
NASA Technical Reports Server (NTRS)
Theurich, Gerhard; DeLuca, C.; Campbell, T.; Liu, F.; Saint, K.; Vertenstein, M.; Chen, J.; Oehmke, R.; Doyle, J.; Whitcomb, T.;
2016-01-01
The Earth System Prediction Suite (ESPS) is a collection of flagship U.S. weather and climate models and model components that are being instrumented to conform to interoperability conventions, documented to follow metadata standards, and made available either under open source terms or to credentialed users.The ESPS represents a culmination of efforts to create a common Earth system model architecture, and the advent of increasingly coordinated model development activities in the U.S. ESPS component interfaces are based on the Earth System Modeling Framework (ESMF), community-developed software for building and coupling models, and the National Unified Operational Prediction Capability (NUOPC) Layer, a set of ESMF-based component templates and interoperability conventions. This shared infrastructure simplifies the process of model coupling by guaranteeing that components conform to a set of technical and semantic behaviors. The ESPS encourages distributed, multi-agency development of coupled modeling systems, controlled experimentation and testing, and exploration of novel model configurations, such as those motivated by research involving managed and interactive ensembles. ESPS codes include the Navy Global Environmental Model (NavGEM), HYbrid Coordinate Ocean Model (HYCOM), and Coupled Ocean Atmosphere Mesoscale Prediction System (COAMPS); the NOAA Environmental Modeling System (NEMS) and the Modular Ocean Model (MOM); the Community Earth System Model (CESM); and the NASA ModelE climate model and GEOS-5 atmospheric general circulation model.
Life Support Baseline Values and Assumptions Document
NASA Technical Reports Server (NTRS)
Anderson, Molly S.; Ewert, Michael K.; Keener, John F.
2018-01-01
The Baseline Values and Assumptions Document (BVAD) provides analysts, modelers, and other life support researchers with a common set of values and assumptions which can be used as a baseline in their studies. This baseline, in turn, provides a common point of origin from which many studies in the community may depart, making research results easier to compare and providing researchers with reasonable values to assume for areas outside their experience. This document identifies many specific physical quantities that define life support systems, serving as a general reference for spacecraft life support system technology developers.
Common Lognormal Behavior in Legal Systems
NASA Astrophysics Data System (ADS)
Yamamoto, Ken
2017-07-01
This study characterizes a statistical property of legal systems: the distribution of the number of articles in a law follows a lognormal distribution. This property is common to the Japanese, German, and Singaporean laws. To explain this lognormal behavior, tree structure of the law is analyzed. If the depth of a tree follows a normal distribution, the lognormal distribution of the number of articles can be theoretically derived. We analyze the structure of the Japanese laws using chapters, sections, and other levels of organization, and this analysis demonstrates that the proposed model is quantitatively reasonable.
A Pursuit Theory Account for the Perception of Common Motion in Motion Parallax.
Ratzlaff, Michael; Nawrot, Mark
2016-09-01
The visual system uses an extraretinal pursuit eye movement signal to disambiguate the perception of depth from motion parallax. Visual motion in the same direction as the pursuit is perceived nearer in depth while visual motion in the opposite direction as pursuit is perceived farther in depth. This explanation of depth sign applies to either an allocentric frame of reference centered on the fixation point or an egocentric frame of reference centered on the observer. A related problem is that of depth order when two stimuli have a common direction of motion. The first psychophysical study determined whether perception of egocentric depth order is adequately explained by a model employing an allocentric framework, especially when the motion parallax stimuli have common rather than divergent motion. A second study determined whether a reversal in perceived depth order, produced by a reduction in pursuit velocity, is also explained by this model employing this allocentric framework. The results show than an allocentric model can explain both the egocentric perception of depth order with common motion and the perceptual depth order reversal created by a reduction in pursuit velocity. We conclude that an egocentric model is not the only explanation for perceived depth order in these common motion conditions. © The Author(s) 2016.
NASA Astrophysics Data System (ADS)
Sivapalan, M.; Elshafei, Y.; Srinivasan, V.
2014-12-01
A challenging research puzzle in the research on sustainable water management in the Anthropocene is why some societies successfully recover from "ecological destruction" to transition to "successful adaptation" over decadal timescales, while others fail. We present a conceptual modeling framework to understand and characterize these transitions. In this way, we aim to capture the potential drivers of the desired shift towards achieving sustainability of socio-hydrological systems. This is done through a synthesis of detailed socio-hydrological analyses of four river basins in three continents, carried out using different quantitative socio-hydrologic models: Murrumbidgee River Basin in eastern Australia, Lake Toolibin Catchment in Western Australia, Tarim River Basin in Western China and Kissimmee River Basin, in south-east United States. The case studies are analysed using either place-based models designed specifically to mimic observed long-term socio-hydrologic trends, or generic conceptual models with foundations in diverse strands of literature including sustainability science and resilience theory. A comparative analysis of the four case studies reveals a commonality in the building blocks employed to model these socio-hydrologic systems; including water balance, economic, environmental and human-feedback components. Each model reveals varying interpretations of a common organising principle that could explain the shift between productive (socio-economic) and restorative (environmental) forces that was evident in each of these systems observed over a long time frame. The emergent principle is related to the essential drivers of the human feedback component and rests with a general formulation of human well-being, as reflected by both their economic and environmental well-being. It is envisaged that the understanding of the system drivers gained from such a comparative study would enable more targeted water management strategies that can be administered in developing basins to achieve overall sustainability.
Agent-based modeling in ecological economics.
Heckbert, Scott; Baynes, Tim; Reeson, Andrew
2010-01-01
Interconnected social and environmental systems are the domain of ecological economics, and models can be used to explore feedbacks and adaptations inherent in these systems. Agent-based modeling (ABM) represents autonomous entities, each with dynamic behavior and heterogeneous characteristics. Agents interact with each other and their environment, resulting in emergent outcomes at the macroscale that can be used to quantitatively analyze complex systems. ABM is contributing to research questions in ecological economics in the areas of natural resource management and land-use change, urban systems modeling, market dynamics, changes in consumer attitudes, innovation, and diffusion of technology and management practices, commons dilemmas and self-governance, and psychological aspects to human decision making and behavior change. Frontiers for ABM research in ecological economics involve advancing the empirical calibration and validation of models through mixed methods, including surveys, interviews, participatory modeling, and, notably, experimental economics to test specific decision-making hypotheses. Linking ABM with other modeling techniques at the level of emergent properties will further advance efforts to understand dynamics of social-environmental systems.
CLIMATE CHANGE IN THAILAND AND ITS POTENTIAL IMPACT ON RICE YIELD
Because of the uncertainties surrounding prediction of climate change, it is common to employ climate scenarios to estimate its impacts on a system. Climate scenarios are sets of climatic perturbations used with models to test system sensitivity to projected changes. In this stud...
Non-Darcy flow of water through woodchip media
USDA-ARS?s Scientific Manuscript database
A denitrifying bioreactor is a system where a carbon substrate (commonly woodchips) is used to reduce nitrate concentration in water flow. Knowledge of intrinsic permeability of woodchip media in different types of this system is of great importance for design and modeling. For many years, water flo...
An Evaporative Cooling Model for Teaching Applied Psychrometrics
ERIC Educational Resources Information Center
Johnson, Donald M.
2004-01-01
Evaporative cooling systems are commonly used in controlled environment plant and animal production. These cooling systems operate based on well defined psychrometric principles. However, students often experience considerable difficulty in learning these principles when they are taught in an abstract, verbal manner. This article describes an…
A Reference Architecture for Space Information Management
NASA Technical Reports Server (NTRS)
Mattmann, Chris A.; Crichton, Daniel J.; Hughes, J. Steven; Ramirez, Paul M.; Berrios, Daniel C.
2006-01-01
We describe a reference architecture for space information management systems that elegantly overcomes the rigid design of common information systems in many domains. The reference architecture consists of a set of flexible, reusable, independent models and software components that function in unison, but remain separately managed entities. The main guiding principle of the reference architecture is to separate the various models of information (e.g., data, metadata, etc.) from implemented system code, allowing each to evolve independently. System modularity, systems interoperability, and dynamic evolution of information system components are the primary benefits of the design of the architecture. The architecture requires the use of information models that are substantially more advanced than those used by the vast majority of information systems. These models are more expressive and can be more easily modularized, distributed and maintained than simpler models e.g., configuration files and data dictionaries. Our current work focuses on formalizing the architecture within a CCSDS Green Book and evaluating the architecture within the context of the C3I initiative.
1991-01-01
Foundation FYDP ......... Five Year Defense Plan FSI ............ Fog Stability Index 17 G G ................ gravity, giga- GISM ......... Gridded ...Global Circulation Model GOES-TAP GOES imagery processing & dissemination system GCS .......... grid course GOFS ........ Global Ocean Flux Study GD...Analysis Support System Complex Systems GRID .......... Global Resource Information Data -Base GEMAG ..... geomagnetic GRIST..... grazing-incidence solar
2012-03-01
EMPIRICAL ANALYSIS OF OPTICAL ATTENUATOR PERFORMANCE IN QUANTUM KEY DISTRIBUTION SYSTEMS USING A...DISTRIBUTION IS UNLIMITED AFIT/GCS/ENG/12-01 EMPIRICAL ANALYSIS OF OPTICAL ATTENUATOR PERFORMANCE IN QUANTUM KEY DISTRIBUTION SYSTEMS USING ...challenging as the complexity of actual implementation specifics are considered. Two components common to most quantum key distribution
European environmental research infrastructures are going for common 30 years strategy
NASA Astrophysics Data System (ADS)
Asmi, Ari; Konjin, Jacco; Pursula, Antti
2014-05-01
Environmental Research infrastructures are facilities, resources, systems and related services that are used by research communities to conduct top-level research. Environmental research is addressing processes at very different time scales, and supporting research infrastructures must be designed as long-term facilities in order to meet the requirements of continuous environmental observation, measurement and analysis. This longevity makes the environmental research infrastructures ideal structures to support the long-term development in environmental sciences. ENVRI project is a collaborative action of the major European (ESFRI) Environmental Research Infrastructures working towards increased co-operation and interoperability between the infrastructures. One of the key products of the ENVRI project is to combine the long-term plans of the individual infrastructures towards a common strategy, describing the vision and planned actions. The envisaged vision for environmental research infrastructures toward 2030 is to support the holistic understanding of our planet and it's behavior. The development of a 'Standard Model of the Planet' is a common ambition, a challenge to define an environmental standard model; a framework of all interactions within the Earth System, from solid earth to near space. Indeed scientists feel challenged to contribute to a 'Standard Model of the Planet' with data, models, algorithms and discoveries. Understanding the Earth System as an interlinked system requires a systems approach. The Environmental Sciences are rapidly moving to become a one system-level science. Mainly since modern science, engineering and society are increasingly facing complex problems that can only be understood in the context of the full overall system. The strategy of the supporting collaborating research infrastructures is based on developing three key factors for the Environmental Sciences: the technological, the cultural and the human capital. The technological capital development concentrates on improving the capacities to measure, observe, preserve and compute. This requires staff, technologies, sensors, satellites, floats, software to integrate and to do analysis and modeling, including data storage, computing platforms and networks. The cultural capital development addresses issues such as open access to data, rules, licenses, citation agreements, IPR agreements, technologies for machine-machine interaction, workflows, metadata, and RI community on the policy level. Human capital actions are based on anticipated need of specialists, including data scientists and 'generalists' that oversee more than just their own discipline. Developing these, as interrelated services, should help the scientific community to enter innovative and large projects contributing to a 'Standard Model of the Planet'. To achieve the overall goal, ENVRI will publish a set of action items that contains intermediate aims, bigger and smaller steps to work towards the development of the 'Standard Model of the Planet' approach. This timeline of actions can used as reference and 'common denominator' in defining new projects and research programs. Either within the various environmental scientific disciplines or when cooperating among these disciplines or even when outreaching towards other disciplines like social sciences, physics/chemistry, medical/life sciences etc.
NASA Astrophysics Data System (ADS)
Kadow, C.; Illing, S.; Kunst, O.; Cubasch, U.
2014-12-01
The project 'Integrated Data and Evaluation System for Decadal Scale Prediction' (INTEGRATION) as part of the German decadal prediction project MiKlip develops a central evaluation system. The fully operational hybrid features a HPC shell access and an user friendly web-interface. It employs one common system with a variety of verification tools and validation data from different projects in- and outside of MiKlip. The evaluation system is located at the German Climate Computing Centre (DKRZ) and has direct access to the bulk of its ESGF node including millions of climate model data sets, e.g. from CMIP5 and CORDEX. The database is organized by the international CMOR standard using the meta information of the self-describing model, reanalysis and observational data sets. Apache Solr is used for indexing the different data projects into one common search environment. This implemented meta data system with its advanced but easy to handle search tool supports users, developers and their tools to retrieve the required information. A generic application programming interface (API) allows scientific developers to connect their analysis tools with the evaluation system independently of the programming language used. Users of the evaluation techniques benefit from the common interface of the evaluation system without any need to understand the different scripting languages. Facilitating the provision and usage of tools and climate data increases automatically the number of scientists working with the data sets and identify discrepancies. Additionally, the history and configuration sub-system stores every analysis performed with the evaluation system in a MySQL database. Configurations and results of the tools can be shared among scientists via shell or web-system. Therefore, plugged-in tools gain automatically from transparency and reproducibility. Furthermore, when configurations match while starting a evaluation tool, the system suggests to use results already produced by other users-saving CPU time, I/O and disk space. This study presents the different techniques and advantages of such a hybrid evaluation system making use of a Big Data HPC in climate science. website: www-miklip.dkrz.de visitor-login: guest password: miklip
NASA Astrophysics Data System (ADS)
Kadow, Christopher; Illing, Sebastian; Kunst, Oliver; Ulbrich, Uwe; Cubasch, Ulrich
2015-04-01
The project 'Integrated Data and Evaluation System for Decadal Scale Prediction' (INTEGRATION) as part of the German decadal prediction project MiKlip develops a central evaluation system. The fully operational hybrid features a HPC shell access and an user friendly web-interface. It employs one common system with a variety of verification tools and validation data from different projects in- and outside of MiKlip. The evaluation system is located at the German Climate Computing Centre (DKRZ) and has direct access to the bulk of its ESGF node including millions of climate model data sets, e.g. from CMIP5 and CORDEX. The database is organized by the international CMOR standard using the meta information of the self-describing model, reanalysis and observational data sets. Apache Solr is used for indexing the different data projects into one common search environment. This implemented meta data system with its advanced but easy to handle search tool supports users, developers and their tools to retrieve the required information. A generic application programming interface (API) allows scientific developers to connect their analysis tools with the evaluation system independently of the programming language used. Users of the evaluation techniques benefit from the common interface of the evaluation system without any need to understand the different scripting languages. Facilitating the provision and usage of tools and climate data increases automatically the number of scientists working with the data sets and identify discrepancies. Additionally, the history and configuration sub-system stores every analysis performed with the evaluation system in a MySQL database. Configurations and results of the tools can be shared among scientists via shell or web-system. Therefore, plugged-in tools gain automatically from transparency and reproducibility. Furthermore, when configurations match while starting a evaluation tool, the system suggests to use results already produced by other users-saving CPU time, I/O and disk space. This study presents the different techniques and advantages of such a hybrid evaluation system making use of a Big Data HPC in climate science. website: www-miklip.dkrz.de visitor-login: click on "Guest"
Modelling the urban water cycle as an integrated part of the city: a review.
Urich, Christian; Rauch, Wolfgang
2014-01-01
In contrast to common perceptions, the urban water infrastructure system is a complex and dynamic system that is constantly evolving and adapting to changes in the urban environment, to sustain existing services and provide additional ones. Instead of simplifying urban water infrastructure to a static system that is decoupled from its urban context, new management strategies use the complexity of the system to their advantage by integrating centralised with decentralised solutions and explicitly embedding water systems into their urban form. However, to understand and test possible adaptation strategies, urban water modelling tools are required to support exploration of their effectiveness as the human-technology-environment system coevolves under different future scenarios. The urban water modelling community has taken first steps to developing these new modelling tools. This paper critically reviews the historical development of urban water modelling tools and provides a summary of the current state of integrated modelling approaches. It reflects on the challenges that arise through the current practice of coupling urban water management tools with urban development models and discusses a potential pathway towards a new generation of modelling tools.
Freyre-González, Julio A; Treviño-Quintanilla, Luis G; Valtierra-Gutiérrez, Ilse A; Gutiérrez-Ríos, Rosa María; Alonso-Pavón, José A
2012-10-31
Escherichia coli and Bacillus subtilis are two of the best-studied prokaryotic model organisms. Previous analyses of their transcriptional regulatory networks have shown that they exhibit high plasticity during evolution and suggested that both converge to scale-free-like structures. Nevertheless, beyond this suggestion, no analyses have been carried out to identify the common systems-level components and principles governing these organisms. Here we show that these two phylogenetically distant organisms follow a set of common novel biologically consistent systems principles revealed by the mathematically and biologically founded natural decomposition approach. The discovered common functional architecture is a diamond-shaped, matryoshka-like, three-layer (coordination, processing, and integration) hierarchy exhibiting feedback, which is shaped by four systems-level components: global transcription factors (global TFs), locally autonomous modules, basal machinery and intermodular genes. The first mathematical criterion to identify global TFs, the κ-value, was reassessed on B. subtilis and confirmed its high predictive power by identifying all the previously reported, plus three potential, master regulators and eight sigma factors. The functionally conserved cores of modules, basal cell machinery, and a set of non-orthologous common physiological global responses were identified via both orthologous genes and non-orthologous conserved functions. This study reveals novel common systems principles maintained between two phylogenetically distant organisms and provides a comparison of their lifestyle adaptations. Our results shed new light on the systems-level principles and the fundamental functions required by bacteria to sustain life. Copyright © 2012 Elsevier B.V. All rights reserved.
Condition-based diagnosis of mechatronic systems using a fractional calculus approach
NASA Astrophysics Data System (ADS)
Gutiérrez-Carvajal, Ricardo Enrique; Flávio de Melo, Leonimer; Maurício Rosário, João; Tenreiro Machado, J. A.
2016-07-01
While fractional calculus (FC) is as old as integer calculus, its application has been mainly restricted to mathematics. However, many real systems are better described using FC equations than with integer models. FC is a suitable tool for describing systems characterised by their fractal nature, long-term memory and chaotic behaviour. It is a promising methodology for failure analysis and modelling, since the behaviour of a failing system depends on factors that increase the model's complexity. This paper explores the proficiency of FC in modelling complex behaviour by tuning only a few parameters. This work proposes a novel two-step strategy for diagnosis, first modelling common failure conditions and, second, by comparing these models with real machine signals and using the difference to feed a computational classifier. Our proposal is validated using an electrical motor coupled with a mechanical gear reducer.
An Approach to Model Based Testing of Multiagent Systems
Nadeem, Aamer
2015-01-01
Autonomous agents perform on behalf of the user to achieve defined goals or objectives. They are situated in dynamic environment and are able to operate autonomously to achieve their goals. In a multiagent system, agents cooperate with each other to achieve a common goal. Testing of multiagent systems is a challenging task due to the autonomous and proactive behavior of agents. However, testing is required to build confidence into the working of a multiagent system. Prometheus methodology is a commonly used approach to design multiagents systems. Systematic and thorough testing of each interaction is necessary. This paper proposes a novel approach to testing of multiagent systems based on Prometheus design artifacts. In the proposed approach, different interactions between the agent and actors are considered to test the multiagent system. These interactions include percepts and actions along with messages between the agents which can be modeled in a protocol diagram. The protocol diagram is converted into a protocol graph, on which different coverage criteria are applied to generate test paths that cover interactions between the agents. A prototype tool has been developed to generate test paths from protocol graph according to the specified coverage criterion. PMID:25874263
Host pathogen relations: exploring animal models for fungal pathogens.
Harwood, Catherine G; Rao, Reeta P
2014-06-30
Pathogenic fungi cause superficial infections but pose a significant public health risk when infections spread to deeper tissues, such as the lung. Within the last three decades, fungi have been identified as the leading cause of nosocomial infections making them the focus of research. This review outlines the model systems such as the mouse, zebrafish larvae, flies, and nematodes, as well as ex vivo and in vitro systems available to study common fungal pathogens.
BRIDG: a domain information model for translational and clinical protocol-driven research.
Becnel, Lauren B; Hastak, Smita; Ver Hoef, Wendy; Milius, Robert P; Slack, MaryAnn; Wold, Diane; Glickman, Michael L; Brodsky, Boris; Jaffe, Charles; Kush, Rebecca; Helton, Edward
2017-09-01
It is critical to integrate and analyze data from biological, translational, and clinical studies with data from health systems; however, electronic artifacts are stored in thousands of disparate systems that are often unable to readily exchange data. To facilitate meaningful data exchange, a model that presents a common understanding of biomedical research concepts and their relationships with health care semantics is required. The Biomedical Research Integrated Domain Group (BRIDG) domain information model fulfills this need. Software systems created from BRIDG have shared meaning "baked in," enabling interoperability among disparate systems. For nearly 10 years, the Clinical Data Standards Interchange Consortium, the National Cancer Institute, the US Food and Drug Administration, and Health Level 7 International have been key stakeholders in developing BRIDG. BRIDG is an open-source Unified Modeling Language-class model developed through use cases and harmonization with other models. With its 4+ releases, BRIDG includes clinical and now translational research concepts in its Common, Protocol Representation, Study Conduct, Adverse Events, Regulatory, Statistical Analysis, Experiment, Biospecimen, and Molecular Biology subdomains. The model is a Clinical Data Standards Interchange Consortium, Health Level 7 International, and International Standards Organization standard that has been utilized in national and international standards-based software development projects. It will continue to mature and evolve in the areas of clinical imaging, pathology, ontology, and vocabulary support. BRIDG 4.1.1 and prior releases are freely available at https://bridgmodel.nci.nih.gov . © The Author 2017. Published by Oxford University Press on behalf of the American Medical Informatics Association. All rights reserved. For Permissions, please email: journals.permissions@oup.com
Chen, Bo; Guo, Wei-hua; Li, Peng-yun; Xie, Wen-ping
2014-01-01
This paper presented an overview on the dynamic analysis and control of the transmission tower-line system in the past forty years. The challenges and future developing trends in the dynamic analysis and mitigation of the transmission tower-line system under dynamic excitations are also put forward. It also reviews the analytical models and approaches of the transmission tower, transmission lines, and transmission tower-line systems, respectively, which contain the theoretical model, finite element (FE) model and the equivalent model; shows the advances in wind responses of the transmission tower-line system, which contains the dynamic effects under common wind loading, tornado, downburst, and typhoon; and discusses the dynamic responses under earthquake and ice loads, respectively. The vibration control of the transmission tower-line system is also reviewed, which includes the magnetorheological dampers, friction dampers, tuned mass dampers, and pounding tuned mass dampers. PMID:25105161
Documenting Models for Interoperability and Reusability ...
Many modeling frameworks compartmentalize science via individual models that link sets of small components to create larger modeling workflows. Developing integrated watershed models increasingly requires coupling multidisciplinary, independent models, as well as collaboration between scientific communities, since component-based modeling can integrate models from different disciplines. Integrated Environmental Modeling (IEM) systems focus on transferring information between components by capturing a conceptual site model; establishing local metadata standards for input/output of models and databases; managing data flow between models and throughout the system; facilitating quality control of data exchanges (e.g., checking units, unit conversions, transfers between software languages); warning and error handling; and coordinating sensitivity/uncertainty analyses. Although many computational software systems facilitate communication between, and execution of, components, there are no common approaches, protocols, or standards for turn-key linkages between software systems and models, especially if modifying components is not the intent. Using a standard ontology, this paper reviews how models can be described for discovery, understanding, evaluation, access, and implementation to facilitate interoperability and reusability. In the proceedings of the International Environmental Modelling and Software Society (iEMSs), 8th International Congress on Environmental Mod
An Instructive Model of Entropy
ERIC Educational Resources Information Center
Zimmerman, Seth
2010-01-01
This article first notes the misinterpretation of a common thought experiment, and the misleading comment that "systems tend to flow from less probable to more probable macrostates". It analyses the experiment, generalizes it and introduces a new tool of investigation, the simplectic structure. A time-symmetric model is built upon this structure,…
2009-03-01
model locations, time of day, and video size. The models in the scene consisted of three-dimensional representations of common civilian automobiles in...oats, wheat). Identify automobiles as sedans or station wagons. Identify individual telephone/electric poles in residential neighborhoods. Detect
An instrumental electrode model for solving EIT forward problems.
Zhang, Weida; Li, David
2014-10-01
An instrumental electrode model (IEM) capable of describing the performance of electrical impedance tomography (EIT) systems in the MHz frequency range has been proposed. Compared with the commonly used Complete Electrode Model (CEM), which assumes ideal front-end interfaces, the proposed model considers the effects of non-ideal components in the front-end circuits. This introduces an extra boundary condition in the forward model and offers a more accurate modelling for EIT systems. We have demonstrated its performance using simple geometry structures and compared the results with the CEM and full Maxwell methods. The IEM can provide a significantly more accurate approximation than the CEM in the MHz frequency range, where the full Maxwell methods are favoured over the quasi-static approximation. The improved electrode model will facilitate the future characterization and front-end design of real-world EIT systems.
A geographic data model for representing ground water systems.
Strassberg, Gil; Maidment, David R; Jones, Norm L
2007-01-01
The Arc Hydro ground water data model is a geographic data model for representing spatial and temporal ground water information within a geographic information system (GIS). The data model is a standardized representation of ground water systems within a spatial database that provides a public domain template for GIS users to store, document, and analyze commonly used spatial and temporal ground water data sets. This paper describes the data model framework, a simplified version of the complete ground water data model that includes two-dimensional and three-dimensional (3D) object classes for representing aquifers, wells, and borehole data, and the 3D geospatial context in which these data exist. The framework data model also includes tabular objects for representing temporal information such as water levels and water quality samples that are related with spatial features.
Simultaneous Sensor and Process Fault Diagnostics for Propellant Feed System
NASA Technical Reports Server (NTRS)
Cao, J.; Kwan, C.; Figueroa, F.; Xu, R.
2006-01-01
The main objective of this research is to extract fault features from sensor faults and process faults by using advanced fault detection and isolation (FDI) algorithms. A tank system that has some common characteristics to a NASA testbed at Stennis Space Center was used to verify our proposed algorithms. First, a generic tank system was modeled. Second, a mathematical model suitable for FDI has been derived for the tank system. Third, a new and general FDI procedure has been designed to distinguish process faults and sensor faults. Extensive simulations clearly demonstrated the advantages of the new design.
The Earth System Prediction Suite: Toward a Coordinated U.S. Modeling Capability
Theurich, Gerhard; DeLuca, C.; Campbell, T.; ...
2016-08-22
The Earth System Prediction Suite (ESPS) is a collection of flagship U.S. weather and climate models and model components that are being instrumented to conform to interoperability conventions, documented to follow metadata standards, and made available either under open-source terms or to credentialed users. Furthermore, the ESPS represents a culmination of efforts to create a common Earth system model architecture, and the advent of increasingly coordinated model development activities in the United States. ESPS component interfaces are based on the Earth System Modeling Framework (ESMF), community-developed software for building and coupling models, and the National Unified Operational Prediction Capability (NUOPC)more » Layer, a set of ESMF-based component templates and interoperability conventions. Our shared infrastructure simplifies the process of model coupling by guaranteeing that components conform to a set of technical and semantic behaviors. The ESPS encourages distributed, multiagency development of coupled modeling systems; controlled experimentation and testing; and exploration of novel model configurations, such as those motivated by research involving managed and interactive ensembles. ESPS codes include the Navy Global Environmental Model (NAVGEM), the Hybrid Coordinate Ocean Model (HYCOM), and the Coupled Ocean–Atmosphere Mesoscale Prediction System (COAMPS); the NOAA Environmental Modeling System (NEMS) and the Modular Ocean Model (MOM); the Community Earth System Model (CESM); and the NASA ModelE climate model and the Goddard Earth Observing System Model, version 5 (GEOS-5), atmospheric general circulation model.« less
The Earth System Prediction Suite: Toward a Coordinated U.S. Modeling Capability
DOE Office of Scientific and Technical Information (OSTI.GOV)
Theurich, Gerhard; DeLuca, C.; Campbell, T.
The Earth System Prediction Suite (ESPS) is a collection of flagship U.S. weather and climate models and model components that are being instrumented to conform to interoperability conventions, documented to follow metadata standards, and made available either under open-source terms or to credentialed users. Furthermore, the ESPS represents a culmination of efforts to create a common Earth system model architecture, and the advent of increasingly coordinated model development activities in the United States. ESPS component interfaces are based on the Earth System Modeling Framework (ESMF), community-developed software for building and coupling models, and the National Unified Operational Prediction Capability (NUOPC)more » Layer, a set of ESMF-based component templates and interoperability conventions. Our shared infrastructure simplifies the process of model coupling by guaranteeing that components conform to a set of technical and semantic behaviors. The ESPS encourages distributed, multiagency development of coupled modeling systems; controlled experimentation and testing; and exploration of novel model configurations, such as those motivated by research involving managed and interactive ensembles. ESPS codes include the Navy Global Environmental Model (NAVGEM), the Hybrid Coordinate Ocean Model (HYCOM), and the Coupled Ocean–Atmosphere Mesoscale Prediction System (COAMPS); the NOAA Environmental Modeling System (NEMS) and the Modular Ocean Model (MOM); the Community Earth System Model (CESM); and the NASA ModelE climate model and the Goddard Earth Observing System Model, version 5 (GEOS-5), atmospheric general circulation model.« less
NASA Technical Reports Server (NTRS)
Canfield, Richard C.; De La Beaujardiere, J.-F.; Fan, Yuhong; Leka, K. D.; Mcclymont, A. N.; Metcalf, Thomas R.; Mickey, Donald L.; Wuelser, Jean-Pierre; Lites, Bruce W.
1993-01-01
Electric current systems in solar active regions and their spatial relationship to sites of electron precipitation and high-pressure in flares were studied with the purpose of providing observational evidence for or against the flare models commonly discussed in the literature. The paper describes the instrumentation, the data used, and the data analysis methods, as well as improvements made upon earlier studies. Several flare models are overviewed, and the predictions yielded by each model for the relationships of flares to the vertical current systems are discussed.
Margaliot, Michael; Sontag, Eduardo D; Tuller, Tamir
2014-01-01
Periodic oscillations play an important role in many biomedical systems. Proper functioning of biological systems that respond to periodic signals requires the ability to synchronize with the periodic excitation. For example, the sleep/wake cycle is a manifestation of an internal timing system that synchronizes to the solar day. In the terminology of systems theory, the biological system must entrain or phase-lock to the periodic excitation. Entrainment is also important in synthetic biology. For example, connecting several artificial biological systems that entrain to a common clock may lead to a well-functioning modular system. The cell-cycle is a periodic program that regulates DNA synthesis and cell division. Recent biological studies suggest that cell-cycle related genes entrain to this periodic program at the gene translation level, leading to periodically-varying protein levels of these genes. The ribosome flow model (RFM) is a deterministic model obtained via a mean-field approximation of a stochastic model from statistical physics that has been used to model numerous processes including ribosome flow along the mRNA. Here we analyze the RFM under the assumption that the initiation and/or transition rates vary periodically with a common period T. We show that the ribosome distribution profile in the RFM entrains to this periodic excitation. In particular, the protein synthesis pattern converges to a unique periodic solution with period T. To the best of our knowledge, this is the first proof of entrainment in a mathematical model for translation that encapsulates aspects such as initiation and termination rates, ribosomal movement and interactions, and non-homogeneous elongation speeds along the mRNA. Our results support the conjecture that periodic oscillations in tRNA levels and other factors related to the translation process can induce periodic oscillations in protein levels, and may suggest a new approach for re-engineering genetic systems to obtain a desired, periodic, protein synthesis rate.
The cosmic merger rate of stellar black hole binaries from the Illustris simulation
NASA Astrophysics Data System (ADS)
Mapelli, Michela; Giacobbo, Nicola; Ripamonti, Emanuele; Spera, Mario
2017-12-01
The cosmic merger rate density of black hole binaries (BHBs) can give us an essential clue to constraining the formation channels of BHBs, in light of current and forthcoming gravitational wave detections. Following a Monte Carlo approach, we couple new population-synthesis models of BHBs with the Illustris cosmological simulation, to study the cosmic history of BHB mergers. We explore six population-synthesis models, varying the prescriptions for supernovae, common envelope and natal kicks. In most considered models, the cosmic BHB merger rate follows the same trend as the cosmic star formation rate. The normalization of the cosmic BHB merger rate strongly depends on the treatment of common envelope and on the distribution of natal kicks. We find that most BHBs merging within LIGO's instrumental horizon come from relatively metal-poor progenitors (<0.2 Z⊙). The total masses of merging BHBs span a large range of values, from ∼6 to ∼82 M⊙. In our fiducial model, merging BHBs consistent with GW150914, GW151226 and GW170104 represent ∼6, 3 and 12 per cent of all BHBs merging within the LIGO horizon, respectively. The heavy systems, like GW150914, come from metal-poor progenitors (<0.15 Z⊙). Most GW150914-like systems merging in the local Universe appear to have formed at high redshift, with a long delay time. In contrast, GW151226-like systems form and merge all the way through the cosmic history, from progenitors with a broad range of metallicities. Future detections will be crucial to put constraints on common envelope, on natal kicks, and on the BHB mass function.
An Alternative Proposal for the Graphical Representation of Anticolor Charge
ERIC Educational Resources Information Center
Wiener, Gergried J.; Schmeling, Sascha M.; Hopf, Martin
2017-01-01
We have developed a learning unit based on the Standard Model of particle physics, featuring novel typographic illustrations of elementary particles and particle systems. Since the unit includes antiparticles and systems of antiparticles, a visualization of anticolor charge was required. We propose an alternative to the commonly used…
Automatic classification of animal vocalizations
NASA Astrophysics Data System (ADS)
Clemins, Patrick J.
2005-11-01
Bioacoustics, the study of animal vocalizations, has begun to use increasingly sophisticated analysis techniques in recent years. Some common tasks in bioacoustics are repertoire determination, call detection, individual identification, stress detection, and behavior correlation. Each research study, however, uses a wide variety of different measured variables, called features, and classification systems to accomplish these tasks. The well-established field of human speech processing has developed a number of different techniques to perform many of the aforementioned bioacoustics tasks. Melfrequency cepstral coefficients (MFCCs) and perceptual linear prediction (PLP) coefficients are two popular feature sets. The hidden Markov model (HMM), a statistical model similar to a finite autonoma machine, is the most commonly used supervised classification model and is capable of modeling both temporal and spectral variations. This research designs a framework that applies models from human speech processing for bioacoustic analysis tasks. The development of the generalized perceptual linear prediction (gPLP) feature extraction model is one of the more important novel contributions of the framework. Perceptual information from the species under study can be incorporated into the gPLP feature extraction model to represent the vocalizations as the animals might perceive them. By including this perceptual information and modifying parameters of the HMM classification system, this framework can be applied to a wide range of species. The effectiveness of the framework is shown by analyzing African elephant and beluga whale vocalizations. The features extracted from the African elephant data are used as input to a supervised classification system and compared to results from traditional statistical tests. The gPLP features extracted from the beluga whale data are used in an unsupervised classification system and the results are compared to labels assigned by experts. The development of a framework from which to build animal vocalization classifiers will provide bioacoustics researchers with a consistent platform to analyze and classify vocalizations. A common framework will also allow studies to compare results across species and institutions. In addition, the use of automated classification techniques can speed analysis and uncover behavioral correlations not readily apparent using traditional techniques.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Strons, Philip; Bailey, James L.; Davis, John
2016-03-01
In this work, we apply the CFD in modeling airflow and particulate transport. This modeling is then compared to field validation studies to both inform and validate the modeling assumptions. Based on the results of field tests, modeling assumptions and boundary conditions are refined and the process is repeated until the results are found to be reliable with a high level of confidence.
Conditioning of FRF measurements for use with frequency based substructuring
NASA Astrophysics Data System (ADS)
Nicgorski, Dana; Avitabile, Peter
2010-02-01
Frequency based substructuring approaches have been used for the generation of system models from component data. While numerical models show successful results, there have been many difficulties with actual measurements in many instances. Previous work has identified some of these typical problems using simulated data to incorporate specific measurement difficulties commonly observed along with approaches to overcome some of these difficulties. This paper presents the results using actual measured data for a laboratory structure subjected to both analytical and experimental studies. Various commonly used approaches are shown to illustrate some of the difficulties with measured data. A new approach to better condition the measured functions and purge commonly found data measurement contaminants is utilized to provide dramatically improved results. Several cases are explored to show the difficulties commonly observed as well as the improved conditioning of the measured data to obtain acceptable results.
He, Xiaoming; Bhowmick, Sankha; Bischof, John C
2009-07-01
The Arrhenius and thermal isoeffective dose (TID) models are the two most commonly used models for predicting hyperthermic injury. The TID model is essentially derived from the Arrhenius model, but due to a variety of assumptions and simplifications now leads to different predictions, particularly at temperatures higher than 50 degrees C. In the present study, the two models are compared and their appropriateness tested for predicting hyperthermic injury in both the traditional hyperthermia (usually, 43-50 degrees C) and thermal surgery (or thermal therapy/thermal ablation, usually, >50 degrees C) regime. The kinetic parameters of thermal injury in both models were obtained from the literature (or literature data), tabulated, and analyzed for various prostate and kidney systems. It was found that the kinetic parameters vary widely, and were particularly dependent on the cell or tissue type, injury assay used, and the time when the injury assessment was performed. In order to compare the capability of the two models for thermal injury prediction, thermal thresholds for complete killing (i.e., 99% cell or tissue injury) were predicted using the models in two important urologic systems, viz., the benign prostatic hyperplasia tissue and the normal porcine kidney tissue. The predictions of the two models matched well at temperatures below 50 degrees C. At higher temperatures, however, the thermal thresholds predicted using the TID model with a constant R value of 0.5, the value commonly used in the traditional hyperthermia literature, are much lower than those predicted using the Arrhenius model. This suggests that traditional use of the TID model (i.e., R=0.5) is inappropriate for predicting hyperthermic injury in the thermal surgery regime (>50 degrees C). Finally, the time-temperature relationships for complete killing (i.e., 99% injury) were calculated and analyzed using the Arrhenius model for the various prostate and kidney systems.
Kumarapeli, Pushpa; de Lusignan, Simon; Koczan, Phil; Jones, Beryl; Sheeler, Ian
2007-01-01
UK general practice is universally computerised, with computers used in the consulting room at the point of care. Practices use a range of different brands of computer system, which have developed organically to meet the needs of general practitioners and health service managers. Unified Modelling Language (UML) is a standard modelling and specification notation widely used in software engineering. To examine the feasibility of UML notation to compare the impact of different brands of general practice computer system on the clinical consultation. Multi-channel video recordings of simulated consultation sessions were recorded on three different clinical computer systems in common use (EMIS, iSOFT Synergy and IPS Vision). User action recorder software recorded time logs of keyboard and mouse use, and pattern recognition software captured non-verbal communication. The outputs of these were used to create UML class and sequence diagrams for each consultation. We compared 'definition of the presenting problem' and 'prescribing', as these tasks were present in all the consultations analysed. Class diagrams identified the entities involved in the clinical consultation. Sequence diagrams identified common elements of the consultation (such as prescribing) and enabled comparisons to be made between the different brands of computer system. The clinician and computer system interaction varied greatly between the different brands. UML sequence diagrams are useful in identifying common tasks in the clinical consultation, and for contrasting the impact of the different brands of computer system on the clinical consultation. Further research is needed to see if patterns demonstrated in this pilot study are consistently displayed.
Application of zonal model on indoor air sensor network design
NASA Astrophysics Data System (ADS)
Chen, Y. Lisa; Wen, Jin
2007-04-01
Growing concerns over the safety of the indoor environment have made the use of sensors ubiquitous. Sensors that detect chemical and biological warfare agents can offer early warning of dangerous contaminants. However, current sensor system design is more informed by intuition and experience rather by systematic design. To develop a sensor system design methodology, a proper indoor airflow modeling approach is needed. Various indoor airflow modeling techniques, from complicated computational fluid dynamics approaches to simplified multi-zone approaches, exist in the literature. In this study, the effects of two airflow modeling techniques, multi-zone modeling technique and zonal modeling technique, on indoor air protection sensor system design are discussed. Common building attack scenarios, using a typical CBW agent, are simulated. Both multi-zone and zonal models are used to predict airflows and contaminant dispersion. Genetic Algorithm is then applied to optimize the sensor location and quantity. Differences in the sensor system design resulting from the two airflow models are discussed for a typical office environment and a large hall environment.
Snowden, Thomas J; van der Graaf, Piet H; Tindall, Marcus J
2017-07-01
Complex models of biochemical reaction systems have become increasingly common in the systems biology literature. The complexity of such models can present a number of obstacles for their practical use, often making problems difficult to intuit or computationally intractable. Methods of model reduction can be employed to alleviate the issue of complexity by seeking to eliminate those portions of a reaction network that have little or no effect upon the outcomes of interest, hence yielding simplified systems that retain an accurate predictive capacity. This review paper seeks to provide a brief overview of a range of such methods and their application in the context of biochemical reaction network models. To achieve this, we provide a brief mathematical account of the main methods including timescale exploitation approaches, reduction via sensitivity analysis, optimisation methods, lumping, and singular value decomposition-based approaches. Methods are reviewed in the context of large-scale systems biology type models, and future areas of research are briefly discussed.
Statistics of Shared Components in Complex Component Systems
NASA Astrophysics Data System (ADS)
Mazzolini, Andrea; Gherardi, Marco; Caselle, Michele; Cosentino Lagomarsino, Marco; Osella, Matteo
2018-04-01
Many complex systems are modular. Such systems can be represented as "component systems," i.e., sets of elementary components, such as LEGO bricks in LEGO sets. The bricks found in a LEGO set reflect a target architecture, which can be built following a set-specific list of instructions. In other component systems, instead, the underlying functional design and constraints are not obvious a priori, and their detection is often a challenge of both scientific and practical importance, requiring a clear understanding of component statistics. Importantly, some quantitative invariants appear to be common to many component systems, most notably a common broad distribution of component abundances, which often resembles the well-known Zipf's law. Such "laws" affect in a general and nontrivial way the component statistics, potentially hindering the identification of system-specific functional constraints or generative processes. Here, we specifically focus on the statistics of shared components, i.e., the distribution of the number of components shared by different system realizations, such as the common bricks found in different LEGO sets. To account for the effects of component heterogeneity, we consider a simple null model, which builds system realizations by random draws from a universe of possible components. Under general assumptions on abundance heterogeneity, we provide analytical estimates of component occurrence, which quantify exhaustively the statistics of shared components. Surprisingly, this simple null model can positively explain important features of empirical component-occurrence distributions obtained from large-scale data on bacterial genomes, LEGO sets, and book chapters. Specific architectural features and functional constraints can be detected from occurrence patterns as deviations from these null predictions, as we show for the illustrative case of the "core" genome in bacteria.
Schawo, Saskia J; van Eeren, Hester; Soeteman, Djira I; van der Veldt, Marie-Christine; Noom, Marc J; Brouwer, Werner; Busschbach, Jan J V; Hakkaart, Leona
2012-12-01
Many interventions initiated within and financed from the health care sector are not necessarily primarily aimed at improving health. This poses important questions regarding the operationalisation of economic evaluations in such contexts. We investigated whether assessing cost-effectiveness using state-of-the-art methods commonly applied in health care evaluations is feasible and meaningful when evaluating interventions aimed at reducing youth delinquency. A probabilistic Markov model was constructed to create a framework for the assessment of the cost-effectiveness of systemic interventions in delinquent youth. For illustrative purposes, Functional Family Therapy (FFT), a systemic intervention aimed at improving family functioning and, primarily, reducing delinquent activity in youths, was compared to Treatment as Usual (TAU). "Criminal activity free years" (CAFYs) were introduced as central outcome measure. Criminal activity may e.g. be based on police contacts or committed crimes. In absence of extensive data and for illustrative purposes the current study based criminal activity on available literature on recidivism. Furthermore, a literature search was performed to deduce the model's structure and parameters. Common cost-effectiveness methodology could be applied to interventions for youth delinquency. Model characteristics and parameters were derived from literature and ongoing trial data. The model resulted in an estimate of incremental costs/CAFY and included long-term effects. Illustrative model results point towards dominance of FFT compared to TAU. Using a probabilistic model and the CAFY outcome measure to assess cost-effectiveness of systemic interventions aimed to reduce delinquency is feasible. However, the model structure is limited to three states and the CAFY measure was defined rather crude. Moreover, as the model parameters are retrieved from literature the model results are illustrative in the absence of empirical data. The current model provides a framework to assess the cost-effectiveness of systemic interventions, while taking into account parameter uncertainty and long-term effectiveness. The framework of the model could be used to assess the cost-effectiveness of systemic interventions alongside (clinical) trial data. Consequently, it is suitable to inform reimbursement decisions, since the value for money of systemic interventions can be demonstrated using a decision analytic model. Future research could be focussed on testing the current model based on extensive empirical data, improving the outcome measure and finding appropriate values for that outcome.
Moore, Jason H; Boczko, Erik M; Summar, Marshall L
2005-02-01
Understanding how DNA sequence variations impact human health through a hierarchy of biochemical and physiological systems is expected to improve the diagnosis, prevention, and treatment of common, complex human diseases. We have previously developed a hierarchical dynamic systems approach based on Petri nets for generating biochemical network models that are consistent with genetic models of disease susceptibility. This modeling approach uses an evolutionary computation approach called grammatical evolution as a search strategy for optimal Petri net models. We have previously demonstrated that this approach routinely identifies biochemical network models that are consistent with a variety of genetic models in which disease susceptibility is determined by nonlinear interactions between two or more DNA sequence variations. We review here this approach and then discuss how it can be used to model biochemical and metabolic data in the context of genetic studies of human disease susceptibility.
Alternative Method to Simulate a Sub-idle Engine Operation in Order to Synthesize Its Control System
NASA Astrophysics Data System (ADS)
Sukhovii, Sergii I.; Sirenko, Feliks F.; Yepifanov, Sergiy V.; Loboda, Igor
2016-09-01
The steady-state and transient engine performances in control systems are usually evaluated by applying thermodynamic engine models. Most models operate between the idle and maximum power points, only recently, they sometimes address a sub-idle operating range. The lack of information about the component maps at the sub-idle modes presents a challenging problem. A common method to cope with the problem is to extrapolate the component performances to the sub-idle range. Precise extrapolation is also a challenge. As a rule, many scientists concern only particular aspects of the problem such as the lighting combustion chamber or the turbine operation under the turned-off conditions of the combustion chamber. However, there are no reports about a model that considers all of these aspects and simulates the engine starting. The proposed paper addresses a new method to simulate the starting. The method substitutes the non-linear thermodynamic model with a linear dynamic model, which is supplemented with a simplified static model. The latter model is the set of direct relations between parameters that are used in the control algorithms instead of commonly used component performances. Specifically, this model consists of simplified relations between the gas path parameters and the corrected rotational speed.
Discrete Address Beacon System (DABS) Baseline Test and Evaluation.
1980-04-01
Organization ReportNo 7. ~/ - 9. PorTorming Organisation Name and Address 10. Work Unit No. (TRALS) Federal Aviation Administration National Aviation...version of the Common International Civil Aviation Organization (ICAO) Data Interchange Network (CIDIN) protocol used in the DABS engineering model. 8. All...grouped into two subsets, one for surveillance data communications and one for Common International Civil Aviation Organization (ICAO) Data Interchange
Control algorithms and applications of the wavefront sensorless adaptive optics
NASA Astrophysics Data System (ADS)
Ma, Liang; Wang, Bin; Zhou, Yuanshen; Yang, Huizhen
2017-10-01
Compared with the conventional adaptive optics (AO) system, the wavefront sensorless (WFSless) AO system need not to measure the wavefront and reconstruct it. It is simpler than the conventional AO in system architecture and can be applied to the complex conditions. Based on the analysis of principle and system model of the WFSless AO system, wavefront correction methods of the WFSless AO system were divided into two categories: model-free-based and model-based control algorithms. The WFSless AO system based on model-free-based control algorithms commonly considers the performance metric as a function of the control parameters and then uses certain control algorithm to improve the performance metric. The model-based control algorithms include modal control algorithms, nonlinear control algorithms and control algorithms based on geometrical optics. Based on the brief description of above typical control algorithms, hybrid methods combining the model-free-based control algorithm with the model-based control algorithm were generalized. Additionally, characteristics of various control algorithms were compared and analyzed. We also discussed the extensive applications of WFSless AO system in free space optical communication (FSO), retinal imaging in the human eye, confocal microscope, coherent beam combination (CBC) techniques and extended objects.
NASA Astrophysics Data System (ADS)
Gelikonov, V. M.; Romashov, V. N.; Shabanov, D. V.; Ksenofontov, S. Yu.; Terpelov, D. A.; Shilyagin, P. A.; Gelikonov, G. V.; Vitkin, I. A.
2018-05-01
We consider a cross-polarization optical coherence tomography system with a common path for the sounding and reference waves and active maintenance of the circular polarization of a sounding wave. The system is based on the formation of birefringent characteristics of the total optical path, which are equivalent to a quarter-wave plate with a 45° orientation of its optical axes with respect to the linearly polarized reference wave. Conditions under which any light-polarization state can be obtained using a two-element phase controller are obtained. The dependence of the local cross-scattering coefficient of light in a model medium and biological tissue on the sounding-wave polarization state is demonstrated. The necessity of active maintenance of the circular polarization of a sounding wave in this common path system (including a flexible probe) is shown to realize uniform optimal conditions for cross-polarization studies of biological tissue.
A characterization of workflow management systems for extreme-scale applications
DOE Office of Scientific and Technical Information (OSTI.GOV)
Ferreira da Silva, Rafael; Filgueira, Rosa; Pietri, Ilia
We present that the automation of the execution of computational tasks is at the heart of improving scientific productivity. Over the last years, scientific workflows have been established as an important abstraction that captures data processing and computation of large and complex scientific applications. By allowing scientists to model and express entire data processing steps and their dependencies, workflow management systems relieve scientists from the details of an application and manage its execution on a computational infrastructure. As the resource requirements of today’s computational and data science applications that process vast amounts of data keep increasing, there is a compellingmore » case for a new generation of advances in high-performance computing, commonly termed as extreme-scale computing, which will bring forth multiple challenges for the design of workflow applications and management systems. This paper presents a novel characterization of workflow management systems using features commonly associated with extreme-scale computing applications. We classify 15 popular workflow management systems in terms of workflow execution models, heterogeneous computing environments, and data access methods. Finally, the paper also surveys workflow applications and identifies gaps for future research on the road to extreme-scale workflows and management systems.« less
A characterization of workflow management systems for extreme-scale applications
Ferreira da Silva, Rafael; Filgueira, Rosa; Pietri, Ilia; ...
2017-02-16
We present that the automation of the execution of computational tasks is at the heart of improving scientific productivity. Over the last years, scientific workflows have been established as an important abstraction that captures data processing and computation of large and complex scientific applications. By allowing scientists to model and express entire data processing steps and their dependencies, workflow management systems relieve scientists from the details of an application and manage its execution on a computational infrastructure. As the resource requirements of today’s computational and data science applications that process vast amounts of data keep increasing, there is a compellingmore » case for a new generation of advances in high-performance computing, commonly termed as extreme-scale computing, which will bring forth multiple challenges for the design of workflow applications and management systems. This paper presents a novel characterization of workflow management systems using features commonly associated with extreme-scale computing applications. We classify 15 popular workflow management systems in terms of workflow execution models, heterogeneous computing environments, and data access methods. Finally, the paper also surveys workflow applications and identifies gaps for future research on the road to extreme-scale workflows and management systems.« less
Models of an In-Situ Propellant Production Plant for Mars Exploration
NASA Technical Reports Server (NTRS)
Goodrich, Charlie; Kurien, James; Millar, Bill; Sweet, Adam; Waterman, Sue; Clancy, Daniel (Technical Monitor)
2001-01-01
An in-situ propellant production system (ISPP) is designed to make rocket fuel from chemicals in the Martian atmosphere in order to reduce the amount of materials that would need to be brought from Earth to support Mars missions. We have developed a description of a hypothetical ISPP system that we would like to make available to researchers who are interested in the problem of automatically diagnosing failures in complex NASA systems. This problem description will help researchers to investigate problems of interest to NASA. We would like to make the following material publicly available: (1) a 'common sense' model of an ISPP system; (2) low- and medium-fidelity simulations of the ISPP system written in Microsoft Excel and HCC; and (3) previously published data and diagrams concerning ISPP components. We do not believe there are any export considerations on these materials for the following reasons: (1) These models are not useful for guidance and real time control of vehicles, encrpytion, or any other software purpose categorized under the Export Control Classification Numbers; and (2) The models are very high level and would not by themselves enable real-time control of a real hardware system. The models are at the level of common sense. They capture, for example, that if a heater is turned on an increase in temperature should result(see the attached excerpt). We do not believe there is any commercial value to this material, given the low commercial demand for propellant plants on mars. We have spoken to acting Code IC Division Chief Dan Clancy, and he concurs with our desire to make these materials publicly available via a technical report.
NASA Integrated Network Monitor and Control Software Architecture
NASA Technical Reports Server (NTRS)
Shames, Peter; Anderson, Michael; Kowal, Steve; Levesque, Michael; Sindiy, Oleg; Donahue, Kenneth; Barnes, Patrick
2012-01-01
The National Aeronautics and Space Administration (NASA) Space Communications and Navigation office (SCaN) has commissioned a series of trade studies to define a new architecture intended to integrate the three existing networks that it operates, the Deep Space Network (DSN), Space Network (SN), and Near Earth Network (NEN), into one integrated network that offers users a set of common, standardized, services and interfaces. The integrated monitor and control architecture utilizes common software and common operator interfaces that can be deployed at all three network elements. This software uses state-of-the-art concepts such as a pool of re-programmable equipment that acts like a configurable software radio, distributed hierarchical control, and centralized management of the whole SCaN integrated network. For this trade space study a model-based approach using SysML was adopted to describe and analyze several possible options for the integrated network monitor and control architecture. This model was used to refine the design and to drive the costing of the four different software options. This trade study modeled the three existing self standing network elements at point of departure, and then described how to integrate them using variations of new and existing monitor and control system components for the different proposed deployments under consideration. This paper will describe the trade space explored, the selected system architecture, the modeling and trade study methods, and some observations on useful approaches to implementing such model based trade space representation and analysis.
Finite-dimensional modeling of network-induced delays for real-time control systems
NASA Technical Reports Server (NTRS)
Ray, Asok; Halevi, Yoram
1988-01-01
In integrated control systems (ICS), a feedback loop is closed by the common communication channel, which multiplexes digital data from the sensor to the controller and from the controller to the actuator along with the data traffic from other control loops and management functions. Due to asynchronous time-division multiplexing in the network access protocols, time-varying delays are introduced in the control loop, which degrade the system dynamic performance and are a potential source of instability. The delayed control system is represented by a finite-dimensional, time-varying, discrete-time model which is less complex than the existing continuous-time models for time-varying delays; this approach allows for simpler schemes for analysis and simulation of the ICS.
Archetypes for Organisational Safety
NASA Technical Reports Server (NTRS)
Marais, Karen; Leveson, Nancy G.
2003-01-01
We propose a framework using system dynamics to model the dynamic behavior of organizations in accident analysis. Most current accident analysis techniques are event-based and do not adequately capture the dynamic complexity and non-linear interactions that characterize accidents in complex systems. In this paper we propose a set of system safety archetypes that model common safety culture flaws in organizations, i.e., the dynamic behaviour of organizations that often leads to accidents. As accident analysis and investigation tools, the archetypes can be used to develop dynamic models that describe the systemic and organizational factors contributing to the accident. The archetypes help clarify why safety-related decisions do not always result in the desired behavior, and how independent decisions in different parts of the organization can combine to impact safety.
Parameterized reduced-order models using hyper-dual numbers.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Fike, Jeffrey A.; Brake, Matthew Robert
2013-10-01
The goal of most computational simulations is to accurately predict the behavior of a real, physical system. Accurate predictions often require very computationally expensive analyses and so reduced order models (ROMs) are commonly used. ROMs aim to reduce the computational cost of the simulations while still providing accurate results by including all of the salient physics of the real system in the ROM. However, real, physical systems often deviate from the idealized models used in simulations due to variations in manufacturing or other factors. One approach to this issue is to create a parameterized model in order to characterize themore » effect of perturbations from the nominal model on the behavior of the system. This report presents a methodology for developing parameterized ROMs, which is based on Craig-Bampton component mode synthesis and the use of hyper-dual numbers to calculate the derivatives necessary for the parameterization.« less
Tu, Rui; Zhang, Pengfei; Zhang, Rui; Liu, Jinhai; Lu, Xiaochun
2018-03-29
This study proposes two models for precise time transfer using the BeiDou Navigation Satellite System triple-frequency signals: ionosphere-free (IF) combined precise point positioning (PPP) model with two dual-frequency combinations (IF-PPP1) and ionosphere-free combined PPP model with a single triple-frequency combination (IF-PPP2). A dataset with a short baseline (with a common external time frequency) and a long baseline are used for performance assessments. The results show that IF-PPP1 and IF-PPP2 models can both be used for precise time transfer using BeiDou Navigation Satellite System (BDS) triple-frequency signals, and the accuracy and stability of time transfer is the same in both cases, except for a constant system bias caused by the hardware delay of different frequencies, which can be removed by the parameter estimation and prediction with long time datasets or by a priori calibration.
Practical Techniques for Modeling Gas Turbine Engine Performance
NASA Technical Reports Server (NTRS)
Chapman, Jeffryes W.; Lavelle, Thomas M.; Litt, Jonathan S.
2016-01-01
The cost and risk associated with the design and operation of gas turbine engine systems has led to an increasing dependence on mathematical models. In this paper, the fundamentals of engine simulation will be reviewed, an example performance analysis will be performed, and relationships useful for engine control system development will be highlighted. The focus will be on thermodynamic modeling utilizing techniques common in industry, such as: the Brayton cycle, component performance maps, map scaling, and design point criteria generation. In general, these topics will be viewed from the standpoint of an example turbojet engine model; however, demonstrated concepts may be adapted to other gas turbine systems, such as gas generators, marine engines, or high bypass aircraft engines. The purpose of this paper is to provide an example of gas turbine model generation and system performance analysis for educational uses, such as curriculum creation or student reference.
CDEP Consortium on Ocean Data Assimilation for Seasonal-to-Interannual Prediction (ODASI)
NASA Technical Reports Server (NTRS)
Rienecker, Michele; Zebiak, Stephen; Kinter, James; Behringer, David; Rosati, Antonio; Kaplan, Alexey
2005-01-01
The ODASI consortium is focused activity of the NOAA/OGP/Climate Diagnostics and Experimental Prediction Program with the goal of improving ocean data assimilation methods and their implementations in support of seasonal forecasts with coupled general circulation models. The consortium is undertaking coordinated assimilation experiments, with common forcing data sets and common input data streams. With different assimilation systems and different models, we aim to understand what approach works best in improving forecast skill in the equatorial Pacific. The presentation will provide an overview of the consortium goals and plans and recent results focused towards evaluating data impacts.
Modeling and Verification of Dependable Electronic Power System Architecture
NASA Astrophysics Data System (ADS)
Yuan, Ling; Fan, Ping; Zhang, Xiao-fang
The electronic power system can be viewed as a system composed of a set of concurrently interacting subsystems to generate, transmit, and distribute electric power. The complex interaction among sub-systems makes the design of electronic power system complicated. Furthermore, in order to guarantee the safe generation and distribution of electronic power, the fault tolerant mechanisms are incorporated in the system design to satisfy high reliability requirements. As a result, the incorporation makes the design of such system more complicated. We propose a dependable electronic power system architecture, which can provide a generic framework to guide the development of electronic power system to ease the development complexity. In order to provide common idioms and patterns to the system *designers, we formally model the electronic power system architecture by using the PVS formal language. Based on the PVS model of this system architecture, we formally verify the fault tolerant properties of the system architecture by using the PVS theorem prover, which can guarantee that the system architecture can satisfy high reliability requirements.
Dynamics and control simulation of the Spacelab Experiment Pointing Mount
NASA Technical Reports Server (NTRS)
Marsh, E. L.; Ward, R. S.
1977-01-01
Computer simulations were developed to evaluate the performance of four Experiment Pointing Mounts (EPM) being considered for Spacelab experiments in the 1980-1990 time frame. The system modeled compromises a multibody system consisting of the shuttle, a mechanical isolation device, the EPM, celestial and inertial sensors, bearings, gimbal torque motors and associated nonlinearities, the experiment payload, and control and estimator algorithms. Each mount was subjected to a common disturbance (shuttle vernier thruster firing and man push off) and command (stellar pointing or solar raster scan) input. The fundamental limitation common to all mounts was found to be sensor noise. System dynamics and hardware nonlinearities have secondary effects on pointing performance for sufficiently high bandwidth.
NASA Astrophysics Data System (ADS)
Aufdenkampe, A. K.; Tarboton, D. G.; Horsburgh, J. S.; Mayorga, E.; McFarland, M.; Robbins, A.; Haag, S.; Shokoufandeh, A.; Evans, B. M.; Arscott, D. B.
2017-12-01
The Model My Watershed Web app (https://app.wikiwatershed.org/) and the BiG-CZ Data Portal (http://portal.bigcz.org/) and are web applications that share a common codebase and a common goal to deliver high-performance discovery, visualization and analysis of geospatial data in an intuitive user interface in web browser. Model My Watershed (MMW) was designed as a decision support system for watershed conservation implementation. BiG CZ Data Portal was designed to provide context and background data for research sites. Users begin by creating an Area of Interest, via an automated watershed delineation tool, a free draw tool, selection of a predefined area such as a county or USGS Hydrological Unit (HUC), or uploading a custom polygon. Both Web apps visualize and provide summary statistics of land use, soil groups, streams, climate and other geospatial information. MMW then allows users to run a watershed model to simulate different scenarios of human impacts on stormwater runoff and water-quality. BiG CZ Data Portal allows users to search for scientific and monitoring data within the Area of Interest, which also serves as a prototype for the upcoming Monitor My Watershed web app. Both systems integrate with CUAHSI cyberinfrastructure, including visualizing observational data from CUAHSI Water Data Center and storing user data via CUAHSI HydroShare. Both systems also integrate with the new EnviroDIY Water Quality Data Portal (http://data.envirodiy.org/), a system for crowd-sourcing environmental monitoring data using open-source sensor stations (http://envirodiy.org/mayfly/) and based on the Observations Data Model v2.
Rattanatamrong, Prapaporn; Matsunaga, Andrea; Raiturkar, Pooja; Mesa, Diego; Zhao, Ming; Mahmoudi, Babak; Digiovanna, Jack; Principe, Jose; Figueiredo, Renato; Sanchez, Justin; Fortes, Jose
2010-01-01
The CyberWorkstation (CW) is an advanced cyber-infrastructure for Brain-Machine Interface (BMI) research. It allows the development, configuration and execution of BMI computational models using high-performance computing resources. The CW's concept is implemented using a software structure in which an "experiment engine" is used to coordinate all software modules needed to capture, communicate and process brain signals and motor-control commands. A generic BMI-model template, which specifies a common interface to the CW's experiment engine, and a common communication protocol enable easy addition, removal or replacement of models without disrupting system operation. This paper reviews the essential components of the CW and shows how templates can facilitate the processes of BMI model development, testing and incorporation into the CW. It also discusses the ongoing work towards making this process infrastructure independent.
Single stage queueing/manufacturing system model that involves emission variable
NASA Astrophysics Data System (ADS)
Murdapa, P. S.; Pujawan, I. N.; Karningsih, P. D.; Nasution, A. H.
2018-04-01
Queueing is commonly occured at every industry. The basic model of queueing theory gives a foundation for modeling a manufacturing system. Nowadays, carbon emission is an important and inevitable issue due to its huge impact to our environment. However, existing model of queuing applied for analysis of single stage manufacturing system has not taken Carbon emissions into consideration. If it is applied to manufacturing context, it may lead to improper decisisions. By taking into account of emission variables into queuing models, not only the model become more comprehensive but also it creates awareness on the issue to many parties that involves in the system. This paper discusses the single stage M/M/1 queueing model that involves emission variable. Hopefully it could be a starting point for the next more complex models. It has a main objective for determining how carbon emissions could fit into the basic queueing theory. It turned out that the involvement of emission variables into the model has modified the traditional model of a single stage queue to a calculation model of production lot quantity allowed per period.
A design pathfinder with material correlation points for inflatable systems
NASA Astrophysics Data System (ADS)
Fulcher, Jared Terrell
The incorporation of inflatable structures into aerospace systems can produce significant advantages in stowed volume to mechanical effectiveness and overall weight. Many applications of these ultra-lightweight systems are designed to precisely control internal or external surfaces, or both, to achieve desired performance. The modeling of these structures becomes complex due to the material nonlinearities inherent to the majority of construction materials used in inflatable structures. Furthermore, accurately modeling the response and behavior of the interfacing boundaries that are common to many inflatable systems will lead to better understanding of the entire class of structures. The research presented involved using nonlinear finite element simulations correlated with photogrammetry testing to develop a procedure for defining material properties for commercially available polyurethane-coated woven nylon fabric, which is representative of coated materials that have been proven materials for use in many inflatable systems. Further, the new material model was used to design and develop an inflatable pathfinder system which employs only internal pressure to control an assembly of internal membranes. This canonical inflatable system will be used for exploration and development of general understanding of efficient design methodology and analysis of future systems. Canonical structures are incorporated into the design of the phased pathfinder system to allow for more universal insight. Nonlinear finite element simulations were performed to evaluate the effect of various boundary conditions, loading configurations, and material orientations on the geometric precision of geometries representing typical internal/external surfaces commonly incorporated into inflatable pathfinder system. The response of the inflatable system to possible damage was also studied using nonlinear finite element simulations. Development of a correlated material model for analysis of the inflatable pathfinder system has improved the efficiency of design and analysis techniques of future inflatable structures. KEYWORDS: Nonlinear Finite Element, Inflatable Structures, Gossamer Space Systems, Photogrammetry Measurements, Coated Woven Fabric.
A Model for Enhancing Online Course Development
ERIC Educational Resources Information Center
Knowles, Evelyn; Kalata, Kathleen
2008-01-01
In order to meet the growing demand for quality online education, Park University has adopted a model that provides a common framework for all of its online courses. Evelyn Knowles and Kathleen Kalata discuss the circumstances leading to the current system and describe the university's implementation of a course development process that ensures…
The ISO Edi Conceptual Model Activity and Its Relationship to OSI.
ERIC Educational Resources Information Center
Fincher, Judith A.
1990-01-01
The edi conceptual model is being developed to define common structures, services, and processes that syntax-specific standards like X12 and EDIFACT could adopt. Open Systems Interconnection (OSI) is of interest to edi because of its potential to help enable global interoperability across Electronic Data Interchange (EDI) functional groups. A…
An Analytic Hierarchy Process for School Quality and Inspection: Model Development and Application
ERIC Educational Resources Information Center
Al Qubaisi, Amal; Badri, Masood; Mohaidat, Jihad; Al Dhaheri, Hamad; Yang, Guang; Al Rashedi, Asma; Greer, Kenneth
2016-01-01
Purpose: The purpose of this paper is to develop an analytic hierarchy planning-based framework to establish criteria weights and to develop a school performance system commonly called school inspections. Design/methodology/approach: The analytic hierarchy process (AHP) model uses pairwise comparisons and a measurement scale to generate the…
Model Based Usability Heuristics for Constructivist E-Learning
ERIC Educational Resources Information Center
Katre, Dinesh S.
2007-01-01
Many e-learning applications and games have been studied to identify the common interaction models of constructivist learning, namely: 1. Move the object to appropriate location; 2. Place objects in appropriate order and location(s); 3. Click to identify; 4. Change the variable factors to observe the effects; and 5. System personification and…
Improvements in continuum modeling for biomolecular systems
NASA Astrophysics Data System (ADS)
Yu, Qiao; Ben-Zhuo, Lu
2016-01-01
Modeling of biomolecular systems plays an essential role in understanding biological processes, such as ionic flow across channels, protein modification or interaction, and cell signaling. The continuum model described by the Poisson- Boltzmann (PB)/Poisson-Nernst-Planck (PNP) equations has made great contributions towards simulation of these processes. However, the model has shortcomings in its commonly used form and cannot capture (or cannot accurately capture) some important physical properties of the biological systems. Considerable efforts have been made to improve the continuum model to account for discrete particle interactions and to make progress in numerical methods to provide accurate and efficient simulations. This review will summarize recent main improvements in continuum modeling for biomolecular systems, with focus on the size-modified models, the coupling of the classical density functional theory and the PNP equations, the coupling of polar and nonpolar interactions, and numerical progress. Project supported by the National Natural Science Foundation of China (Grant No. 91230106) and the Chinese Academy of Sciences Program for Cross & Cooperative Team of the Science & Technology Innovation.
Rumination time as a potential predictor of common diseases in high-productive Holstein dairy cows.
Moretti, Riccardo; Biffani, Stefano; Tiezzi, Francesco; Maltecca, Christian; Chessa, Stefania; Bozzi, Riccardo
2017-11-01
We examined the hypothesis that rumination time (RT) could serve as a useful predictor of various common diseases of high producing dairy cows and hence improve herd management and animal wellbeing. We measured the changes in rumination time (RT) in the days before the recording of diseases (specifically: mastitis, reproductive system diseases, locomotor system issues, and gastroenteric diseases). We built predictive models to assess the association between RT and these diseases, using the former as the outcome variable, and to study the effects of the latter on the former. The average Pseudo-R 2 of the fitted models was moderate to low, and this could be due to the fact that RT is influenced by other additional factors which have a greater effect than the predictors used here. Although remaining in a moderate-to-low range, the average Pseudo-R 2 of the models regarding locomotion issues and gastroenteric diseases was higher than the others, suggesting the greater effect of these diseases on RT. The results are encouraging, but further work is needed if these models are to become useful predictors.
Adaptive Neural Network Based Control of Noncanonical Nonlinear Systems.
Zhang, Yanjun; Tao, Gang; Chen, Mou
2016-09-01
This paper presents a new study on the adaptive neural network-based control of a class of noncanonical nonlinear systems with large parametric uncertainties. Unlike commonly studied canonical form nonlinear systems whose neural network approximation system models have explicit relative degree structures, which can directly be used to derive parameterized controllers for adaptation, noncanonical form nonlinear systems usually do not have explicit relative degrees, and thus their approximation system models are also in noncanonical forms. It is well-known that the adaptive control of noncanonical form nonlinear systems involves the parameterization of system dynamics. As demonstrated in this paper, it is also the case for noncanonical neural network approximation system models. Effective control of such systems is an open research problem, especially in the presence of uncertain parameters. This paper shows that it is necessary to reparameterize such neural network system models for adaptive control design, and that such reparameterization can be realized using a relative degree formulation, a concept yet to be studied for general neural network system models. This paper then derives the parameterized controllers that guarantee closed-loop stability and asymptotic output tracking for noncanonical form neural network system models. An illustrative example is presented with the simulation results to demonstrate the control design procedure, and to verify the effectiveness of such a new design method.
NASA Astrophysics Data System (ADS)
Voldoire, Aurore; Decharme, Bertrand; Pianezze, Joris; Lebeaupin Brossier, Cindy; Sevault, Florence; Seyfried, Léo; Garnier, Valérie; Bielli, Soline; Valcke, Sophie; Alias, Antoinette; Accensi, Mickael; Ardhuin, Fabrice; Bouin, Marie-Noëlle; Ducrocq, Véronique; Faroux, Stéphanie; Giordani, Hervé; Léger, Fabien; Marsaleix, Patrick; Rainaud, Romain; Redelsperger, Jean-Luc; Richard, Evelyne; Riette, Sébastien
2017-11-01
This study presents the principles of the new coupling interface based on the SURFEX multi-surface model and the OASIS3-MCT coupler. As SURFEX can be plugged into several atmospheric models, it can be used in a wide range of applications, from global and regional coupled climate systems to high-resolution numerical weather prediction systems or very fine-scale models dedicated to process studies. The objective of this development is to build and share a common structure for the atmosphere-surface coupling of all these applications, involving on the one hand atmospheric models and on the other hand ocean, ice, hydrology, and wave models. The numerical and physical principles of SURFEX interface between the different component models are described, and the different coupled systems in which the SURFEX OASIS3-MCT-based coupling interface is already implemented are presented.
Retrosynthetic Reaction Prediction Using Neural Sequence-to-Sequence Models
2017-01-01
We describe a fully data driven model that learns to perform a retrosynthetic reaction prediction task, which is treated as a sequence-to-sequence mapping problem. The end-to-end trained model has an encoder–decoder architecture that consists of two recurrent neural networks, which has previously shown great success in solving other sequence-to-sequence prediction tasks such as machine translation. The model is trained on 50,000 experimental reaction examples from the United States patent literature, which span 10 broad reaction types that are commonly used by medicinal chemists. We find that our model performs comparably with a rule-based expert system baseline model, and also overcomes certain limitations associated with rule-based expert systems and with any machine learning approach that contains a rule-based expert system component. Our model provides an important first step toward solving the challenging problem of computational retrosynthetic analysis. PMID:29104927
Using Petri Net Tools to Study Properties and Dynamics of Biological Systems
Peleg, Mor; Rubin, Daniel; Altman, Russ B.
2005-01-01
Petri Nets (PNs) and their extensions are promising methods for modeling and simulating biological systems. We surveyed PN formalisms and tools and compared them based on their mathematical capabilities as well as by their appropriateness to represent typical biological processes. We measured the ability of these tools to model specific features of biological systems and answer a set of biological questions that we defined. We found that different tools are required to provide all capabilities that we assessed. We created software to translate a generic PN model into most of the formalisms and tools discussed. We have also made available three models and suggest that a library of such models would catalyze progress in qualitative modeling via PNs. Development and wide adoption of common formats would enable researchers to share models and use different tools to analyze them without the need to convert to proprietary formats. PMID:15561791
Standardized reporting of functioning information on ICF-based common metrics.
Prodinger, Birgit; Tennant, Alan; Stucki, Gerold
2018-02-01
In clinical practice and research a variety of clinical data collection tools are used to collect information on people's functioning for clinical practice and research and national health information systems. Reporting on ICF-based common metrics enables standardized documentation of functioning information in national health information systems. The objective of this methodological note on applying the ICF in rehabilitation is to demonstrate how to report functioning information collected with a data collection tool on ICF-based common metrics. We first specify the requirements for the standardized reporting of functioning information. Secondly, we introduce the methods needed for transforming functioning data to ICF-based common metrics. Finally, we provide an example. The requirements for standardized reporting are as follows: 1) having a common conceptual framework to enable content comparability between any health information; and 2) a measurement framework so that scores between two or more clinical data collection tools can be directly compared. The methods needed to achieve these requirements are the ICF Linking Rules and the Rasch measurement model. Using data collected incorporating the 36-item Short Form Health Survey (SF-36), the World Health Organization Disability Assessment Schedule 2.0 (WHODAS 2.0), and the Stroke Impact Scale 3.0 (SIS 3.0), the application of the standardized reporting based on common metrics is demonstrated. A subset of items from the three tools linked to common chapters of the ICF (d4 Mobility, d5 Self-care and d6 Domestic life), were entered as "super items" into the Rasch model. Good fit was achieved with no residual local dependency and a unidimensional metric. A transformation table allows for comparison between scales, and between a scale and the reporting common metric. Being able to report functioning information collected with commonly used clinical data collection tools with ICF-based common metrics enables clinicians and researchers to continue using their tools while still being able to compare and aggregate the information within and across tools.
Klein, E S; Barbier, M R; Watson, J R
2017-08-01
Understanding how and when cooperative human behaviour forms in common-pool resource systems is critical to illuminating social-ecological systems and designing governance institutions that promote sustainable resource use. Before assessing the full complexity of social dynamics, it is essential to understand, concretely and mechanistically, how resource dynamics and human actions interact to create incentives and pay-offs for social behaviours. Here, we investigated how such incentives for information sharing are affected by spatial dynamics and management in a common-pool resource system. Using interviews with fishermen to inform an agent-based model, we reveal generic mechanisms through which, for a given ecological setting characterized by the spatial dynamics of the resource, the two 'human factors' of information sharing and management may heterogeneously impact various members of a group for whom theory would otherwise predict the same strategy. When users can deplete the resource, these interactions are further affected by the management approach. Finally, we discuss the implications of alternative motivations, such as equity among fishermen and consistency of the fleet's output. Our results indicate that resource spatial dynamics, form of management and level of depletion can interact to alter the sociality of people in common-pool resource systems, providing necessary insight for future study of strategic decision processes.
Computational models for the nonlinear analysis of reinforced concrete plates
NASA Technical Reports Server (NTRS)
Hinton, E.; Rahman, H. H. A.; Huq, M. M.
1980-01-01
A finite element computational model for the nonlinear analysis of reinforced concrete solid, stiffened and cellular plates is briefly outlined. Typically, Mindlin elements are used to model the plates whereas eccentric Timoshenko elements are adopted to represent the beams. The layering technique, common in the analysis of reinforced concrete flexural systems, is incorporated in the model. The proposed model provides an inexpensive and reasonably accurate approach which can be extended for use with voided plates.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Honrubia-Escribano, A.; Gomez Lazaro, E.; Jimenez-Buendia, F.
The International Electrotechnical Commission Standard 61400-27-1 was published in February 2015. This standard deals with the development of generic terms and parameters to specify the electrical characteristics of wind turbines. Generic models of very complex technological systems, such as wind turbines, are thus defined based on the four common configurations available in the market. Due to its recent publication, the comparison of the response of generic models with specific vendor models plays a key role in ensuring the widespread use of this standard. This paper compares the response of a specific Gamesa dynamic wind turbine model to the corresponding genericmore » IEC Type III wind turbine model response when the wind turbine is subjected to a three-phase voltage dip. This Type III model represents the doubly-fed induction generator wind turbine, which is not only one of the most commonly sold and installed technologies in the current market but also a complex variable-speed operation implementation. In fact, active and reactive power transients are observed due to the voltage reduction. Special attention is given to the reactive power injection provided by the wind turbine models because it is a requirement of current grid codes. Further, the boundaries of the generic models associated with transient events that cannot be represented exactly are included in the paper.« less
Chun, Seokjoon; Harris, Alexa; Carrion, Margely; Rojas, Elizabeth; Stark, Stephen; Lejuez, Carl; Lechner, William V.; Bornovalova, Marina A.
2016-01-01
The comorbidity between Borderline Personality Disorder (BPD) and Antisocial Personality Disorder (ASPD) is well-established, and the two disorders share many similarities. However, there are also differences across disorders: most notably, BPD is diagnosed more frequently in females and ASPD in males. We investigated if a) comorbidity between BPD and ASPD is attributable to two discrete disorders or the expression of common underlying processes, and b) if the model of comorbidity is true across sex. Using a clinical sample of 1400 drug users in residential substance abuse treatment, we tested three competing models to explore whether the comorbidity of ASPD and BPD should be represented by a single common factor, two correlated factors, or a bifactor structure involving a general and disorder-specific factors. Next, we tested whether our resulting model was meaningful by examining its relationship with criterion variables previously reported to be associated with BPD and ASPD. The bifactor model provided the best fit and was invariant across sex. Overall, the general factor of the bifactor model significantly accounted for a large percentage of the variance in criterion variables, whereas the BPD and AAB specific factors added little to the models. The association of the general and specific factor with all criterion variables was equal for males and females. Our results suggest common underlying vulnerability accounts for both the comorbidity between BPD and AAB (across sex), and this common vulnerability drives the association with other psychopathology and maladaptive behavior. This in turn has implications for diagnostic classification systems and treatment. General scientific summary This study found that, for both males and females, borderline and antisocial personality disorders show a large degree of overlap, and little uniqueness. The commonality between BPD and ASPD mainly accounted for associations with criterion variables. This suggests that BPD and ASPD show a large common core that accounts for their comorbidity. PMID:27808543
Performance evaluation of the croissant production line with reparable machines
NASA Astrophysics Data System (ADS)
Tsarouhas, Panagiotis H.
2015-03-01
In this study, the analytical probability models for an automated serial production system, bufferless that consists of n-machines in series with common transfer mechanism and control system was developed. Both time to failure and time to repair a failure are assumed to follow exponential distribution. Applying those models, the effect of system parameters on system performance in actual croissant production line was studied. The production line consists of six workstations with different numbers of reparable machines in series. Mathematical models of the croissant production line have been developed using Markov process. The strength of this study is in the classification of the whole system in states, representing failures of different machines. Failure and repair data from the actual production environment have been used to estimate reliability and maintainability for each machine, workstation, and the entire line is based on analytical models. The analysis provides a useful insight into the system's behaviour, helps to find design inherent faults and suggests optimal modifications to upgrade the system and improve its performance.
Why involve families in acute mental healthcare? A collaborative conceptual review.
Dirik, Aysegul; Sandhu, Sima; Giacco, Domenico; Barrett, Katherine; Bennison, Gerry; Collinson, Sue; Priebe, Stefan
2017-09-27
Family involvement is strongly recommended in clinical guidelines but suffers from poor implementation. To explore this topic at a conceptual level, a multidisciplinary review team including academics, clinicians and individuals with lived experience undertook a review to explore the theoretical background of family involvement models in acute mental health treatment and how this relates to their delivery. A conceptual review was undertaken, including a systematic search and narrative synthesis. Included family models were mapped onto the most commonly referenced underlying theories: the diathesis-stress model, systems theories and postmodern theories of mental health. Common components of the models were summarised and compared. Lastly, a thematic analysis was undertaken to explore the role of patients and families in the delivery of the approaches. General adult acute mental health treatment. Six distinct family involvement models were identified: Calgary Family Assessment and Intervention Models, ERIC (Equipe Rapide d'Intervention de Crise), Family Psychoeducation Models, Family Systems Approach, Open Dialogue and the Somerset Model. Findings indicated that despite wide variation in the theoretical models underlying family involvement models, there were many commonalities in their components, such as a focus on communication, language use and joint decision-making. Thematic analysis of the role of patients and families identified several issues for implementation. This included potential harms that could emerge during delivery of the models, such as imposing linear 'patient-carer' relationships and the risk of perceived coercion. We conclude that future staff training may benefit from discussing the chosen family involvement model within the context of other theories of mental health. This may help to clarify the underlying purpose of family involvement and address the diverse needs and world views of patients, families and professionals in acute settings. © Article author(s) (or their employer(s) unless otherwise stated in the text of the article) 2017. All rights reserved. No commercial use is permitted unless otherwise expressly granted.
A Generic Modeling Process to Support Functional Fault Model Development
NASA Technical Reports Server (NTRS)
Maul, William A.; Hemminger, Joseph A.; Oostdyk, Rebecca; Bis, Rachael A.
2016-01-01
Functional fault models (FFMs) are qualitative representations of a system's failure space that are used to provide a diagnostic of the modeled system. An FFM simulates the failure effect propagation paths within a system between failure modes and observation points. These models contain a significant amount of information about the system including the design, operation and off nominal behavior. The development and verification of the models can be costly in both time and resources. In addition, models depicting similar components can be distinct, both in appearance and function, when created individually, because there are numerous ways of representing the failure space within each component. Generic application of FFMs has the advantages of software code reuse: reduction of time and resources in both development and verification, and a standard set of component models from which future system models can be generated with common appearance and diagnostic performance. This paper outlines the motivation to develop a generic modeling process for FFMs at the component level and the effort to implement that process through modeling conventions and a software tool. The implementation of this generic modeling process within a fault isolation demonstration for NASA's Advanced Ground System Maintenance (AGSM) Integrated Health Management (IHM) project is presented and the impact discussed.
Oleribe, Obinna Ositadimma; Oladipo, Olabisi Abiodun; Ezieme, Iheaka Paul; Crossey, Mary Margaret Elizabeth; Taylor-Robinson, Simon David
2016-01-01
Access to quality care is essential for improved health outcomes. Decentralization improves access to healthcare services at lower levels of care, but it does not dismantle structural, funding and programming restrictions to access, resulting in inequity and inequality in population health. Unlike decentralization, Commonization Model of care reduces health inequalities and inequity, dismantles structural, funding and other program related obstacles to population health. Excellence and Friends Management Care Center (EFMC) using Commonization Model (CM), fully integrated HIV services into core health services in 121 supported facilities. This initiative improved access to care, treatment, support services, reduced stigmatization/discrimination, and improved uptake of HTC. We call on governments to adequately finance CM for health systems restructuring towards better health outcomes.
Equivalent reduced model technique development for nonlinear system dynamic response
NASA Astrophysics Data System (ADS)
Thibault, Louis; Avitabile, Peter; Foley, Jason; Wolfson, Janet
2013-04-01
The dynamic response of structural systems commonly involves nonlinear effects. Often times, structural systems are made up of several components, whose individual behavior is essentially linear compared to the total assembled system. However, the assembly of linear components using highly nonlinear connection elements or contact regions causes the entire system to become nonlinear. Conventional transient nonlinear integration of the equations of motion can be extremely computationally intensive, especially when the finite element models describing the components are very large and detailed. In this work, the equivalent reduced model technique (ERMT) is developed to address complicated nonlinear contact problems. ERMT utilizes a highly accurate model reduction scheme, the System equivalent reduction expansion process (SEREP). Extremely reduced order models that provide dynamic characteristics of linear components, which are interconnected with highly nonlinear connection elements, are formulated with SEREP for the dynamic response evaluation using direct integration techniques. The full-space solution will be compared to the response obtained using drastically reduced models to make evident the usefulness of the technique for a variety of analytical cases.
Cellular signaling identifiability analysis: a case study.
Roper, Ryan T; Pia Saccomani, Maria; Vicini, Paolo
2010-05-21
Two primary purposes for mathematical modeling in cell biology are (1) simulation for making predictions of experimental outcomes and (2) parameter estimation for drawing inferences from experimental data about unobserved aspects of biological systems. While the former purpose has become common in the biological sciences, the latter is less common, particularly when studying cellular and subcellular phenomena such as signaling-the focus of the current study. Data are difficult to obtain at this level. Therefore, even models of only modest complexity can contain parameters for which the available data are insufficient for estimation. In the present study, we use a set of published cellular signaling models to address issues related to global parameter identifiability. That is, we address the following question: assuming known time courses for some model variables, which parameters is it theoretically impossible to estimate, even with continuous, noise-free data? Following an introduction to this problem and its relevance, we perform a full identifiability analysis on a set of cellular signaling models using DAISY (Differential Algebra for the Identifiability of SYstems). We use our analysis to bring to light important issues related to parameter identifiability in ordinary differential equation (ODE) models. We contend that this is, as of yet, an under-appreciated issue in biological modeling and, more particularly, cell biology. Copyright (c) 2010 Elsevier Ltd. All rights reserved.
Interoperability Gap Challenges for Learning Object Repositories & Learning Management Systems
ERIC Educational Resources Information Center
Mason, Robert T.
2011-01-01
An interoperability gap exists between Learning Management Systems (LMSs) and Learning Object Repositories (LORs). Learning Objects (LOs) and the associated Learning Object Metadata (LOM) that is stored within LORs adhere to a variety of LOM standards. A common LOM standard found in LORs is the Sharable Content Object Reference Model (SCORM)…
A Process-Based Knowledge Management System for Schools: A Case Study in Taiwan
ERIC Educational Resources Information Center
Lee, Chi-Lung; Lu, Hsi-Peng; Yang, Chyan; Hou, Huei-Tse
2010-01-01
Knowledge management systems, or KMSs, have been widely adopted in business organizations, yet little research exists on the actual integration of the knowledge management model and the application of KMSs in secondary schools. In the present study, the common difficulties and limitations regarding the implementation of knowledge management into…
ERIC Educational Resources Information Center
Mavor, A. S.; And Others
Part of a sustained program that has involved the design of personally tailored information systems responsive to the needs of scientists performing common research and teaching tasks, this project focuses on the procedural and content requirements for accomplishing need diagnosis and presents these requirements as specifications for an…
System-of-Systems Technology-Portfolio-Analysis Tool
NASA Technical Reports Server (NTRS)
O'Neil, Daniel; Mankins, John; Feingold, Harvey; Johnson, Wayne
2012-01-01
Advanced Technology Life-cycle Analysis System (ATLAS) is a system-of-systems technology-portfolio-analysis software tool. ATLAS affords capabilities to (1) compare estimates of the mass and cost of an engineering system based on competing technological concepts; (2) estimate life-cycle costs of an outer-space-exploration architecture for a specified technology portfolio; (3) collect data on state-of-the-art and forecasted technology performance, and on operations and programs; and (4) calculate an index of the relative programmatic value of a technology portfolio. ATLAS facilitates analysis by providing a library of analytical spreadsheet models for a variety of systems. A single analyst can assemble a representation of a system of systems from the models and build a technology portfolio. Each system model estimates mass, and life-cycle costs are estimated by a common set of cost models. Other components of ATLAS include graphical-user-interface (GUI) software, algorithms for calculating the aforementioned index, a technology database, a report generator, and a form generator for creating the GUI for the system models. At the time of this reporting, ATLAS is a prototype, embodied in Microsoft Excel and several thousand lines of Visual Basic for Applications that run on both Windows and Macintosh computers.
Acoustic surface perception from naturally occurring step sounds of a dexterous hexapod robot
NASA Astrophysics Data System (ADS)
Cuneyitoglu Ozkul, Mine; Saranli, Afsar; Yazicioglu, Yigit
2013-10-01
Legged robots that exhibit dynamic dexterity naturally interact with the surface to generate complex acoustic signals carrying rich information on the surface as well as the robot platform itself. However, the nature of a legged robot, which is a complex, hybrid dynamic system, renders the more common approach of model-based system identification impractical. The present paper focuses on acoustic surface identification and proposes a non-model-based analysis and classification approach adopted from the speech processing literature. A novel feature set composed of spectral band energies augmented by their vector time derivatives and time-domain averaged zero crossing rate is proposed. Using a multi-dimensional vector classifier, these features carry enough information to accurately classify a range of commonly occurring indoor and outdoor surfaces without using of any mechanical system model. A comparative experimental study is carried out and classification performance and computational complexity are characterized. Different feature combinations, classifiers and changes in critical design parameters are investigated. A realistic and representative acoustic data set is collected with the robot moving at different speeds on a number of surfaces. The study demonstrates promising performance of this non-model-based approach, even in an acoustically uncontrolled environment. The approach also has good chance of performing in real-time.
Automated space vehicle control for rendezvous proximity operations
NASA Technical Reports Server (NTRS)
Lea, Robert N.
1988-01-01
Rendezvous during the unmanned space exploration missions, such as a Mars Rover/Sample Return will require a completely automatic system from liftoff to docking. A conceptual design of an automated rendezvous, proximity operations, and docking system is being implemented and validated at the Johnson Space Center (JSC). The emphasis is on the progress of the development and testing of a prototype system for control of the rendezvous vehicle during proximity operations that is currently being developed at JSC. Fuzzy sets are used to model the human capability of common sense reasoning in decision making tasks and such models are integrated with the expert systems and engineering control system technology to create a system that performs comparably to a manned system.
Automated space vehicle control for rendezvous proximity operations
NASA Technical Reports Server (NTRS)
Lea, Robert N.
1988-01-01
Rendezvous during the unmanned space exploration missions, such as a Mars Rover/Sample Return will require a completely automatic system from liftoff to docking. A conceptual design of an automated rendezvous, proximity operations, and docking system is being implemented and validated at the Johnson Space Center (JSC). The emphasis is on the progress of the development and testing of a prototype system for control of the rendezvous vehicle during proximity operations that is currently being developed at JSC. Fuzzy sets are used to model the human capability of common sense reasoning in decision-making tasks and such models are integrated with the expert systems and engineering control system technology to create a system that performs comparably to a manned system.
Model based design introduction: modeling game controllers to microprocessor architectures
NASA Astrophysics Data System (ADS)
Jungwirth, Patrick; Badawy, Abdel-Hameed
2017-04-01
We present an introduction to model based design. Model based design is a visual representation, generally a block diagram, to model and incrementally develop a complex system. Model based design is a commonly used design methodology for digital signal processing, control systems, and embedded systems. Model based design's philosophy is: to solve a problem - a step at a time. The approach can be compared to a series of steps to converge to a solution. A block diagram simulation tool allows a design to be simulated with real world measurement data. For example, if an analog control system is being upgraded to a digital control system, the analog sensor input signals can be recorded. The digital control algorithm can be simulated with the real world sensor data. The output from the simulated digital control system can then be compared to the old analog based control system. Model based design can compared to Agile software develop. The Agile software development goal is to develop working software in incremental steps. Progress is measured in completed and tested code units. Progress is measured in model based design by completed and tested blocks. We present a concept for a video game controller and then use model based design to iterate the design towards a working system. We will also describe a model based design effort to develop an OS Friendly Microprocessor Architecture based on the RISC-V.
Stochastic phase segregation on surfaces
Gera, Prerna
2017-01-01
Phase separation and coarsening is a phenomenon commonly seen in binary physical and chemical systems that occur in nature. Often, thermal fluctuations, modelled as stochastic noise, are present in the system and the phase segregation process occurs on a surface. In this work, the segregation process is modelled via the Cahn–Hilliard–Cook model, which is a fourth-order parabolic stochastic system. Coarsening is analysed on two sample surfaces: a unit sphere and a dumbbell. On both surfaces, a statistical analysis of the growth rate is performed, and the influence of noise level and mobility is also investigated. For the spherical interface, it is also shown that a lognormal distribution fits the growth rate well. PMID:28878994
Multiscale information modelling for heart morphogenesis
NASA Astrophysics Data System (ADS)
Abdulla, T.; Imms, R.; Schleich, J. M.; Summers, R.
2010-07-01
Science is made feasible by the adoption of common systems of units. As research has become more data intensive, especially in the biomedical domain, it requires the adoption of a common system of information models, to make explicit the relationship between one set of data and another, regardless of format. This is being realised through the OBO Foundry to develop a suite of reference ontologies, and NCBO Bioportal to provide services to integrate biomedical resources and functionality to visualise and create mappings between ontology terms. Biomedical experts tend to be focused at one level of spatial scale, be it biochemistry, cell biology, or anatomy. Likewise, the ontologies they use tend to be focused at a particular level of scale. There is increasing interest in a multiscale systems approach, which attempts to integrate between different levels of scale to gain understanding of emergent effects. This is a return to physiological medicine with a computational emphasis, exemplified by the worldwide Physiome initiative, and the European Union funded Network of Excellence in the Virtual Physiological Human. However, little work has been done on how information modelling itself may be tailored to a multiscale systems approach. We demonstrate how this can be done for the complex process of heart morphogenesis, which requires multiscale understanding in both time and spatial domains. Such an effort enables the integration of multiscale metrology.
Mechanism Design for Incentivizing Social Media Contributions
NASA Astrophysics Data System (ADS)
Singh, Vivek K.; Jain, Ramesh; Kankanhalli, Mohan
Despite recent advancements in user-driven social media platforms, tools for studying user behavior patterns and motivations remain primitive. We highlight the voluntary nature of user contributions and that users can choose when (and when not) to contribute to the common media pool. A Game theoretic framework is proposed to study the dynamics of social media networks where contribution costs are individual but gains are common. We model users as rational selfish agents, and consider domain attributes like voluntary participation, virtual reward structure, network effect, and public-sharing to model the dynamics of this interaction. The created model describes the most appropriate contribution strategy from each user's perspective and also highlights issues like 'free-rider' problem and individual rationality leading to irrational (i.e. sub-optimal) group behavior. We also consider the perspective of the system designer who is interested in finding the best incentive mechanisms to influence the selfish end-users so that the overall system utility is maximized. We propose and compare multiple mechanisms (based on optimal bonus payment, social incentive leveraging, and second price auction) to study how a system designer can exploit the selfishness of its users, to design incentive mechanisms which improve the overall task-completion probability and system performance, while possibly still benefiting the individual users.
Simulating the Interactions Among Land Use, Transportation ...
In most transportation studies, computer models that forecast travel behavior statistics for a future year use static projections of the spatial distribution of future population and employment growth as inputs. As a result, they are unable to account for the temporally dynamic and non-linear interactions among transportation, land use, and socioeconomic systems. System dynamics (SD) provides a common framework for modeling the complex interactions among transportation and other related systems. This study uses a SD model to simulate the cascading impacts of a proposed light rail transit (LRT) system in central North Carolina, USA. The Durham-Orange Light Rail Project (D-O LRP) SD model incorporates relationships among the land use, transportation, and economy sectors to simulate the complex feedbacks that give rise to the travel behavior changes forecasted by the region’s transportation model. This paper demonstrates the sensitivity of changes in travel behavior to the proposed LRT system and the assumptions that went into the transportation modeling, and compares those results to the impacts of an alternative fare-free transit system. SD models such as the D-O LRP SD model can complement transportation studies by providing valuable insight into the interdependent community systems that collectively contribute to travel behavior changes. Presented at the 35th International Conference of the System Dynamics Society in Cambridge, MA, July 18th, 2017
Development of fish-based model systems with various microstructures.
Verheyen, Davy; Baka, Maria; Glorieux, Seline; Duquenne, Barbara; Fraeye, Ilse; Skåra, Torstein; Van Impe, Jan F
2018-04-01
The effectiveness of predictive microbiology is limited by the lack of knowledge concerning the influence of food microstructure on microbial dynamics. Therefore, future modelling attempts should be based on experiments in structured food model systems as well as liquid systems. In this study, fish-based model systems with various microstructures were developed, i.e., two liquid systems (with and without xanthan gum), an emulsion, an aqueous gel, and a gelled emulsion. The microstructural effect was isolated by minimising compositional and physico-chemical changes among the different model systems. The systems were suitable for common growth and mild thermal inactivation experiments involving both homogeneous and surface inoculation. Average pH of the model systems was 6.36±0.03 and average a w was 0.988±0.002. The liquid system without xanthan gum behaved like a Newtonian fluid, while the emulsion and the liquid containing xanthan gum exhibited (non-Newtonian) pseudo-plastic behaviour. Both the aqueous gel and gelled emulsion were classified as strong gels, with a hardness of 1.35±0.07N and 1.25±0.05N, respectively. Fat droplet size of the emulsion and gelled emulsion model systems was evenly distributed around 1μm. In general, the set of model systems was proven to be suitable to study the influence of important aspects of food microstructure on microbial dynamics. Copyright © 2017. Published by Elsevier Ltd.
Common world model for unmanned systems: Phase 2
NASA Astrophysics Data System (ADS)
Dean, Robert M. S.; Oh, Jean; Vinokurov, Jerry
2014-06-01
The Robotics Collaborative Technology Alliance (RCTA) seeks to provide adaptive robot capabilities which move beyond traditional metric algorithms to include cognitive capabilities. Key to this effort is the Common World Model, which moves beyond the state-of-the-art by representing the world using semantic and symbolic as well as metric information. It joins these layers of information to define objects in the world. These objects may be reasoned upon jointly using traditional geometric, symbolic cognitive algorithms and new computational nodes formed by the combination of these disciplines to address Symbol Grounding and Uncertainty. The Common World Model must understand how these objects relate to each other. It includes the concept of Self-Information about the robot. By encoding current capability, component status, task execution state, and their histories we track information which enables the robot to reason and adapt its performance using Meta-Cognition and Machine Learning principles. The world model also includes models of how entities in the environment behave which enable prediction of future world states. To manage complexity, we have adopted a phased implementation approach. Phase 1, published in these proceedings in 2013 [1], presented the approach for linking metric with symbolic information and interfaces for traditional planners and cognitive reasoning. Here we discuss the design of "Phase 2" of this world model, which extends the Phase 1 design API, data structures, and reviews the use of the Common World Model as part of a semantic navigation use case.
Kenow, Kevin P.; Ge, Zhongfu; Fara, Luke J.; Houdek, Steven C.; Lubinski, Brian R.
2016-01-01
Avian botulism type E is responsible for extensive waterbird mortality on the Great Lakes, yet the actual site of toxin exposure remains unclear. Beached carcasses are often used to describe the spatial aspects of botulism mortality outbreaks, but lack specificity of offshore toxin source locations. We detail methodology for developing a neural network model used for predicting waterbird carcass motions in response to wind, wave, and current forcing, in lieu of a complex analytical relationship. This empirically trained model uses current velocity, wind velocity, significant wave height, and wave peak period in Lake Michigan simulated by the Great Lakes Coastal Forecasting System. A detailed procedure is further developed to use the model for back-tracing waterbird carcasses found on beaches in various parts of Lake Michigan, which was validated using drift data for radiomarked common loon (Gavia immer) carcasses deployed at a variety of locations in northern Lake Michigan during September and October of 2013. The back-tracing model was further used on 22 non-radiomarked common loon carcasses found along the shoreline of northern Lake Michigan in October and November of 2012. The model-estimated origins of those cases pointed to some common source locations offshore that coincide with concentrations of common loons observed during aerial surveys. The neural network source tracking model provides a promising approach for identifying locations of botulinum neurotoxin type E intoxication and, in turn, contributes to developing an understanding of the dynamics of toxin production and possible trophic transfer pathways.
Dynamic analysis of clamp band joint system subjected to axial vibration
NASA Astrophysics Data System (ADS)
Qin, Z. Y.; Yan, S. Z.; Chu, F. L.
2010-10-01
Clamp band joints are commonly used for connecting circular components together in industry. Some of the systems jointed by clamp band are subjected to dynamic load. However, very little research on the dynamic characteristics for this kind of joint can be found in the literature. In this paper, a dynamic model for clamp band joint system is developed. Contact and frictional slip between the components are accommodated in this model. Nonlinear finite element analysis is conducted to identify the model parameters. Then static experiments are carried out on a scaled model of the clamp band joint to validate the joint model. Finally, the model is adopted to study the dynamic characteristics of the clamp band joint system subjected to axial harmonic excitation and the effects of the wedge angle of the clamp band joint and the preload on the response. The model proposed in this paper can represent the nonlinearity of the clamp band joint and be used conveniently to investigate the effects of the structural and loading parameters on the dynamic characteristics of this type of joint system.
NASA Astrophysics Data System (ADS)
Ikeda, Fujio; Toyama, Shigehiro; Ishiduki, Souta; Seta, Hiroaki
2016-09-01
Maritime accidents of small ships continue to increase in number. One of the major factors is poor manoeuvrability of the Manual Hydraulic Steering Mechanism (MHSM) in common use. The manoeuvrability can be improved by using the Electronic Control Steering Mechanism (ECSM). This paper conducts stability analyses of a pleasure boat controlled by human models in view of path following on a target course, in order to establish design guidelines for the ECSM. First, to analyse the stability region, the research derives the linear approximated model in a planar global coordinate system. Then, several human models are assumed to develop closed-loop human-machine controlled systems. These human models include basic proportional, derivative, integral and time-delay actions. The stability analysis simulations for those human-machine systems are carried out. The results show that the stability region tends to spread as a ship's velocity increases in the case of the basic proportional human model. The derivative action and time-delay action of human models are effective in spreading the stability region in their respective ranges of frontal gazing points.
Coppola, Jennifer J.; Disney, Anita A.
2018-01-01
Acetylcholine (ACh) is believed to act as a neuromodulator in cortical circuits that support cognition, specifically in processes including learning, memory consolidation, vigilance, arousal and attention. The cholinergic modulation of cortical processes is studied in many model systems including rodents, cats and primates. Further, these studies are performed in cortical areas ranging from the primary visual cortex to the prefrontal cortex and using diverse methodologies. The results of these studies have been combined into singular models of function—a practice based on an implicit assumption that the various model systems are equivalent and interchangeable. However, comparative anatomy both within and across species reveals important differences in the structure of the cholinergic system. Here, we will review anatomical data including innervation patterns, receptor expression, synthesis and release compared across species and cortical area with a focus on rodents and primates. We argue that these data suggest no canonical cortical model system exists for the cholinergic system. Further, we will argue that as a result, care must be taken both in combining data from studies across cortical areas and species, and in choosing the best model systems to improve our understanding and support of human health. PMID:29440996
Coppola, Jennifer J; Disney, Anita A
2018-01-01
Acetylcholine (ACh) is believed to act as a neuromodulator in cortical circuits that support cognition, specifically in processes including learning, memory consolidation, vigilance, arousal and attention. The cholinergic modulation of cortical processes is studied in many model systems including rodents, cats and primates. Further, these studies are performed in cortical areas ranging from the primary visual cortex to the prefrontal cortex and using diverse methodologies. The results of these studies have been combined into singular models of function-a practice based on an implicit assumption that the various model systems are equivalent and interchangeable. However, comparative anatomy both within and across species reveals important differences in the structure of the cholinergic system. Here, we will review anatomical data including innervation patterns, receptor expression, synthesis and release compared across species and cortical area with a focus on rodents and primates. We argue that these data suggest no canonical cortical model system exists for the cholinergic system. Further, we will argue that as a result, care must be taken both in combining data from studies across cortical areas and species, and in choosing the best model systems to improve our understanding and support of human health.
Noise and coupling induced synchronization in a network of chaotic neurons
NASA Astrophysics Data System (ADS)
Ciszak, Marzena; Euzzor, Stefano; Geltrude, Andrea; Tito Arecchi, F.; Meucci, Riccardo
2013-04-01
The synchronization in four forced FitzHugh-Nagumo (FHN) systems is studied, both experimentally and by numerical simulations of a model. We show that synchronization may be achieved either by coupling of systems through bidirectional diffusive interactions, by introducing a common noise to all systems or by combining both ingredients, noise and coupling together. Here we consider white and colored noises, showing that the colored noise is more efficient in synchronizing the systems respect to white noise. Moreover, a small addition of common noise allows the synchronization to occur at smaller values of the coupling strength. When the diffusive coupling in the absence of noise is considered, the system undergoes the transition to subthreshold oscillations, giving a spike suppression regime. We show that noise destroys the appearance of this dynamical regime induced by coupling.
Network-driven design principles for neuromorphic systems.
Partzsch, Johannes; Schüffny, Rene
2015-01-01
Synaptic connectivity is typically the most resource-demanding part of neuromorphic systems. Commonly, the architecture of these systems is chosen mainly on technical considerations. As a consequence, the potential for optimization arising from the inherent constraints of connectivity models is left unused. In this article, we develop an alternative, network-driven approach to neuromorphic architecture design. We describe methods to analyse performance of existing neuromorphic architectures in emulating certain connectivity models. Furthermore, we show step-by-step how to derive a neuromorphic architecture from a given connectivity model. For this, we introduce a generalized description for architectures with a synapse matrix, which takes into account shared use of circuit components for reducing total silicon area. Architectures designed with this approach are fitted to a connectivity model, essentially adapting to its connection density. They are guaranteeing faithful reproduction of the model on chip, while requiring less total silicon area. In total, our methods allow designers to implement more area-efficient neuromorphic systems and verify usability of the connectivity resources in these systems.
Network-driven design principles for neuromorphic systems
Partzsch, Johannes; Schüffny, Rene
2015-01-01
Synaptic connectivity is typically the most resource-demanding part of neuromorphic systems. Commonly, the architecture of these systems is chosen mainly on technical considerations. As a consequence, the potential for optimization arising from the inherent constraints of connectivity models is left unused. In this article, we develop an alternative, network-driven approach to neuromorphic architecture design. We describe methods to analyse performance of existing neuromorphic architectures in emulating certain connectivity models. Furthermore, we show step-by-step how to derive a neuromorphic architecture from a given connectivity model. For this, we introduce a generalized description for architectures with a synapse matrix, which takes into account shared use of circuit components for reducing total silicon area. Architectures designed with this approach are fitted to a connectivity model, essentially adapting to its connection density. They are guaranteeing faithful reproduction of the model on chip, while requiring less total silicon area. In total, our methods allow designers to implement more area-efficient neuromorphic systems and verify usability of the connectivity resources in these systems. PMID:26539079
Leavesley, G.H.; Markstrom, S.L.; Restrepo, Pedro J.; Viger, R.J.
2002-01-01
A modular approach to model design and construction provides a flexible framework in which to focus the multidisciplinary research and operational efforts needed to facilitate the development, selection, and application of the most robust distributed modelling methods. A variety of modular approaches have been developed, but with little consideration for compatibility among systems and concepts. Several systems are proprietary, limiting any user interaction. The US Geological Survey modular modelling system (MMS) is a modular modelling framework that uses an open source software approach to enable all members of the scientific community to address collaboratively the many complex issues associated with the design, development, and application of distributed hydrological and environmental models. Implementation of a common modular concept is not a trivial task. However, it brings the resources of a larger community to bear on the problems of distributed modelling, provides a framework in which to compare alternative modelling approaches objectively, and provides a means of sharing the latest modelling advances. The concepts and components of the MMS are described and an example application of the MMS, in a decision-support system context, is presented to demonstrate current system capabilities. Copyright ?? 2002 John Wiley and Sons, Ltd.
Ma, Xiao; Bibby, Kyle
2017-09-01
Fungi are near-ubiquitous in potable water distribution systems, but the disinfection kinetics of commonly identified fungi are poorly studied. In the present study, laboratory scale experiments were conducted to evaluate the inactivation kinetics of Aspergillus fumigatus, Aspergillus versicolor, and Penicillium purpurogenum by free chlorine and monochloramine. The observed inactivation data were then fit to a delayed Chick-Watson model. Based on the model parameter estimation, the Ct values (integrated product of disinfectant concentration C and contact time t over defined time intervals) for 99.9% inactivation of the tested fungal strains ranged from 48.99 mg min/L to 194.7 mg min/L for free chlorine and from 90.33 mg min/L to 531.3 mg min/L for monochloramine. Fungal isolates from a drinking water system (Aspergillus versicolor and Penicillium purpurogenum) were more disinfection resistant than Aspergillus fumigatus type and clinical isolates. The required 99.9% inactivation Ct values for the tested fungal strains are higher than E. coli, a commonly monitored indicator bacteria, and within a similar range for bacteria commonly identified within water distribution systems, such as Mycobacterium spp. and Legionella spp. Copyright © 2017 Elsevier Ltd. All rights reserved.
Animal Models of Peripheral Neuropathy Due to Environmental Toxicants
Rao, Deepa B.; Jortner, Bernard S.; Sills, Robert C.
2014-01-01
Despite the progress in our understanding of pathogeneses and the identification of etiologies of peripheral neuropathy, idiopathic neuropathy remains common. Typically, attention to peripheral neuropathies resulting from exposure to environmental agents is limited relative to more commonly diagnosed causes of peripheral neuropathy (diabetes and chemotherapeutic agents). Given that there are more than 80,000 chemicals in commerce registered with the Environmental Protection Agency and that at least 1000 chemicals are known to have neurotoxic potential, very few chemicals have been established to affect the peripheral nervous system (mainly after occupational exposures). A wide spectrum of exposures, including pesticides, metals, solvents, nutritional sources, and pharmaceutical agents, has been related, both historically and recently, to environmental toxicant-induced peripheral neuropathy. A review of the literature shows that the toxicity and pathogeneses of chemicals adversely affecting the peripheral nervous system have been studied using animal models. This article includes an overview of five prototypical environmental agents known to cause peripheral neuropathy—namely, organophosphates, carbon disulfide, pyridoxine (Vitamin B6), acrylamide, and hexacarbons (mainly n-hexane, 2,5-hexanedione, methyl n-butyl ketone). Also included is a brief introduction to the structural components of the peripheral nervous system and pointers on common methodologies for histopathologic evaluation of the peripheral nerves. PMID:24615445
NASA Technical Reports Server (NTRS)
Dominick, Wayne D. (Editor); Farooq, Mohammad U.
1986-01-01
The definition of proposed research addressing the development and validation of a methodology for the design and evaluation of user interfaces for interactive information systems is given. The major objectives of this research are: the development of a comprehensive, objective, and generalizable methodology for the design and evaluation of user interfaces for information systems; the development of equations and/or analytical models to characterize user behavior and the performance of a designed interface; the design of a prototype system for the development and administration of user interfaces; and the design and use of controlled experiments to support the research and test/validate the proposed methodology. The proposed design methodology views the user interface as a virtual machine composed of three layers: an interactive layer, a dialogue manager layer, and an application interface layer. A command language model of user system interactions is presented because of its inherent simplicity and structured approach based on interaction events. All interaction events have a common structure based on common generic elements necessary for a successful dialogue. It is shown that, using this model, various types of interfaces could be designed and implemented to accommodate various categories of users. The implementation methodology is discussed in terms of how to store and organize the information.
NASA Astrophysics Data System (ADS)
Wattawa, Scott
1995-11-01
Offering interactive services and data in a hybrid fiber/coax cable system requires the coordination of a host of operations and business support systems. New service offerings and network growth and evolution create never-ending changes in the network infrastructure. Agent-based enterprise models provide a flexible mechanism for systems integration of service and support systems. Agent models also provide a mechanism to decouple interactive services from network architecture. By using the Java programming language, agents may be made safe, portable, and intelligent. This paper investigates the application of the Object Management Group's Common Object Request Brokering Architecture to the integration of a multiple services metropolitan area network.
ERIC Educational Resources Information Center
Carver, Charles S.; Johnson, Sheri L.; Joormann, Jutta
2008-01-01
Evidence from diverse literatures supports the viewpoint that two modes of self-regulation exist, a lower-order system that responds quickly to associative cues of the moment and a higher-order system that responds more reflectively and planfully; that low serotonergic function is linked to relative dominance of the lower-order system; that how…
A common type system for clinical natural language processing
2013-01-01
Background One challenge in reusing clinical data stored in electronic medical records is that these data are heterogenous. Clinical Natural Language Processing (NLP) plays an important role in transforming information in clinical text to a standard representation that is comparable and interoperable. Information may be processed and shared when a type system specifies the allowable data structures. Therefore, we aim to define a common type system for clinical NLP that enables interoperability between structured and unstructured data generated in different clinical settings. Results We describe a common type system for clinical NLP that has an end target of deep semantics based on Clinical Element Models (CEMs), thus interoperating with structured data and accommodating diverse NLP approaches. The type system has been implemented in UIMA (Unstructured Information Management Architecture) and is fully functional in a popular open-source clinical NLP system, cTAKES (clinical Text Analysis and Knowledge Extraction System) versions 2.0 and later. Conclusions We have created a type system that targets deep semantics, thereby allowing for NLP systems to encapsulate knowledge from text and share it alongside heterogenous clinical data sources. Rather than surface semantics that are typically the end product of NLP algorithms, CEM-based semantics explicitly build in deep clinical semantics as the point of interoperability with more structured data types. PMID:23286462
A common type system for clinical natural language processing.
Wu, Stephen T; Kaggal, Vinod C; Dligach, Dmitriy; Masanz, James J; Chen, Pei; Becker, Lee; Chapman, Wendy W; Savova, Guergana K; Liu, Hongfang; Chute, Christopher G
2013-01-03
One challenge in reusing clinical data stored in electronic medical records is that these data are heterogenous. Clinical Natural Language Processing (NLP) plays an important role in transforming information in clinical text to a standard representation that is comparable and interoperable. Information may be processed and shared when a type system specifies the allowable data structures. Therefore, we aim to define a common type system for clinical NLP that enables interoperability between structured and unstructured data generated in different clinical settings. We describe a common type system for clinical NLP that has an end target of deep semantics based on Clinical Element Models (CEMs), thus interoperating with structured data and accommodating diverse NLP approaches. The type system has been implemented in UIMA (Unstructured Information Management Architecture) and is fully functional in a popular open-source clinical NLP system, cTAKES (clinical Text Analysis and Knowledge Extraction System) versions 2.0 and later. We have created a type system that targets deep semantics, thereby allowing for NLP systems to encapsulate knowledge from text and share it alongside heterogenous clinical data sources. Rather than surface semantics that are typically the end product of NLP algorithms, CEM-based semantics explicitly build in deep clinical semantics as the point of interoperability with more structured data types.
NASA Technical Reports Server (NTRS)
Nauda, A.
1982-01-01
Performance and reliability models of alternate microcomputer architectures as a methodology for optimizing system design were examined. A methodology for selecting an optimum microcomputer architecture for autonomous operation of planetary spacecraft power systems was developed. Various microcomputer system architectures are analyzed to determine their application to spacecraft power systems. It is suggested that no standardization formula or common set of guidelines exists which provides an optimum configuration for a given set of specifications.
Improving Cyber-Security of Smart Grid Systems via Anomaly Detection and Linguistic Domain Knowledge
DOE Office of Scientific and Technical Information (OSTI.GOV)
Ondrej Linda; Todd Vollmer; Milos Manic
The planned large scale deployment of smart grid network devices will generate a large amount of information exchanged over various types of communication networks. The implementation of these critical systems will require appropriate cyber-security measures. A network anomaly detection solution is considered in this work. In common network architectures multiple communications streams are simultaneously present, making it difficult to build an anomaly detection solution for the entire system. In addition, common anomaly detection algorithms require specification of a sensitivity threshold, which inevitably leads to a tradeoff between false positives and false negatives rates. In order to alleviate these issues, thismore » paper proposes a novel anomaly detection architecture. The designed system applies the previously developed network security cyber-sensor method to individual selected communication streams allowing for learning accurate normal network behavior models. Furthermore, the developed system dynamically adjusts the sensitivity threshold of each anomaly detection algorithm based on domain knowledge about the specific network system. It is proposed to model this domain knowledge using Interval Type-2 Fuzzy Logic rules, which linguistically describe the relationship between various features of the network communication and the possibility of a cyber attack. The proposed method was tested on experimental smart grid system demonstrating enhanced cyber-security.« less
PRESS-based EFOR algorithm for the dynamic parametrical modeling of nonlinear MDOF systems
NASA Astrophysics Data System (ADS)
Liu, Haopeng; Zhu, Yunpeng; Luo, Zhong; Han, Qingkai
2017-09-01
In response to the identification problem concerning multi-degree of freedom (MDOF) nonlinear systems, this study presents the extended forward orthogonal regression (EFOR) based on predicted residual sums of squares (PRESS) to construct a nonlinear dynamic parametrical model. The proposed parametrical model is based on the non-linear autoregressive with exogenous inputs (NARX) model and aims to explicitly reveal the physical design parameters of the system. The PRESS-based EFOR algorithm is proposed to identify such a model for MDOF systems. By using the algorithm, we built a common-structured model based on the fundamental concept of evaluating its generalization capability through cross-validation. The resulting model aims to prevent over-fitting with poor generalization performance caused by the average error reduction ratio (AERR)-based EFOR algorithm. Then, a functional relationship is established between the coefficients of the terms and the design parameters of the unified model. Moreover, a 5-DOF nonlinear system is taken as a case to illustrate the modeling of the proposed algorithm. Finally, a dynamic parametrical model of a cantilever beam is constructed from experimental data. Results indicate that the dynamic parametrical model of nonlinear systems, which depends on the PRESS-based EFOR, can accurately predict the output response, thus providing a theoretical basis for the optimal design of modeling methods for MDOF nonlinear systems.
Framework for a clinical information system.
Van de Velde, R
2000-01-01
The current status of our work towards the design and implementation of a reference architecture for a Clinical Information System is presented. This architecture has been developed and implemented based on components following a strong underlying conceptual and technological model. Common Object Request Broker and n-tier technology featuring centralised and departmental clinical information systems as the back-end store for all clinical data are used. Servers located in the 'middle' tier apply the clinical (business) model and application rules to communicate with so-called 'thin client' workstations. The main characteristics are the focus on modelling and reuse of both data and business logic as there is a shift away from data and functional modelling towards object modelling. Scalability as well as adaptability to constantly changing requirements via component driven computing are the main reasons for that approach.
Jones, James W; Antle, John M; Basso, Bruno; Boote, Kenneth J; Conant, Richard T; Foster, Ian; Godfray, H Charles J; Herrero, Mario; Howitt, Richard E; Janssen, Sander; Keating, Brian A; Munoz-Carpena, Rafael; Porter, Cheryl H; Rosenzweig, Cynthia; Wheeler, Tim R
2017-07-01
We review the current state of agricultural systems science, focusing in particular on the capabilities and limitations of agricultural systems models. We discuss the state of models relative to five different Use Cases spanning field, farm, landscape, regional, and global spatial scales and engaging questions in past, current, and future time periods. Contributions from multiple disciplines have made major advances relevant to a wide range of agricultural system model applications at various spatial and temporal scales. Although current agricultural systems models have features that are needed for the Use Cases, we found that all of them have limitations and need to be improved. We identified common limitations across all Use Cases, namely 1) a scarcity of data for developing, evaluating, and applying agricultural system models and 2) inadequate knowledge systems that effectively communicate model results to society. We argue that these limitations are greater obstacles to progress than gaps in conceptual theory or available methods for using system models. New initiatives on open data show promise for addressing the data problem, but there also needs to be a cultural change among agricultural researchers to ensure that data for addressing the range of Use Cases are available for future model improvements and applications. We conclude that multiple platforms and multiple models are needed for model applications for different purposes. The Use Cases provide a useful framework for considering capabilities and limitations of existing models and data.
NASA Technical Reports Server (NTRS)
Jones, James W.; Antle, John M.; Basso, Bruno; Boote, Kenneth J.; Conant, Richard T.; Foster, Ian; Godfray, H. Charles J.; Herrero, Mario; Howitt, Richard E.; Janssen, Sander;
2016-01-01
We review the current state of agricultural systems science, focusing in particular on the capabilities and limitations of agricultural systems models. We discuss the state of models relative to five different Use Cases spanning field, farm, landscape, regional, and global spatial scales and engaging questions in past, current, and future time periods. Contributions from multiple disciplines have made major advances relevant to a wide range of agricultural system model applications at various spatial and temporal scales. Although current agricultural systems models have features that are needed for the Use Cases, we found that all of them have limitations and need to be improved. We identified common limitations across all Use Cases, namely 1) a scarcity of data for developing, evaluating, and applying agricultural system models and 2) inadequate knowledge systems that effectively communicate model results to society. We argue that these limitations are greater obstacles to progress than gaps in conceptual theory or available methods for using system models. New initiatives on open data show promise for addressing the data problem, but there also needs to be a cultural change among agricultural researchers to ensure that data for addressing the range of Use Cases are available for future model improvements and applications. We conclude that multiple platforms and multiple models are needed for model applications for different purposes. The Use Cases provide a useful framework for considering capabilities and limitations of existing models and data.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Jones, James W.; Antle, John M.; Basso, Bruno
We review the current state of agricultural systems science, focusing in particular on the capabilities and limitations of agricultural systems models. We discuss the state of models relative to five different Use Cases spanning field, farm, landscape, regional, and global spatial scales and engaging questions in past, current, and future time periods. Contributions from multiple disciplines have made major advances relevant to a wide range of agricultural system model applications at various spatial and temporal scales. Although current agricultural systems models have features that are needed for the Use Cases, we found that all of them have limitations and needmore » to be improved. We identified common limitations across all Use Cases, namely 1) a scarcity of data for developing, evaluating, and applying agricultural system models and 2) inadequate knowledge systems that effectively communicate model results to society. We argue that these limitations are greater obstacles to progress than gaps in conceptual theory or available methods for using system models. New initiatives on open data show promise for addressing the data problem, but there also needs to be a cultural change among agricultural researchers to ensure that data for addressing the range of Use Cases are available for future model improvements and applications. We conclude that multiple platforms and multiple models are needed for model applications for different purposes. The Use Cases provide a useful framework for considering capabilities and limitations of existing models and data.« less
Crisis Management Systems: A Case Study for Aspect-Oriented Modeling
NASA Astrophysics Data System (ADS)
Kienzle, Jörg; Guelfi, Nicolas; Mustafiz, Sadaf
The intent of this document is to define a common case study for the aspect-oriented modeling research community. The domain of the case study is crisis management systems, i.e., systems that help in identifying, assessing, and handling a crisis situation by orchestrating the communication between all parties involved in handling the crisis, by allocating and managing resources, and by providing access to relevant crisis-related information to authorized users. This document contains informal requirements of crisis management systems (CMSs) in general, a feature model for a CMS product line, use case models for a car crash CMS (CCCMS), a domain model for the CCCMS, an informal physical architecture description of the CCCMS, as well as some design models of a possible object-oriented implementation of parts of the CCCMS backend. AOM researchers who want to demonstrate the power of their AOM approach or technique can hence apply the approach at the most appropriate level of abstraction.
A two-fluid model for avalanche and debris flows.
Pitman, E Bruce; Le, Long
2005-07-15
Geophysical mass flows--debris flows, avalanches, landslides--can contain O(10(6)-10(10)) m(3) or more of material, often a mixture of soil and rocks with a significant quantity of interstitial fluid. These flows can be tens of meters in depth and hundreds of meters in length. The range of scales and the rheology of this mixture presents significant modelling and computational challenges. This paper describes a depth-averaged 'thin layer' model of geophysical mass flows containing a mixture of solid material and fluid. The model is derived from a 'two-phase' or 'two-fluid' system of equations commonly used in engineering research. Phenomenological modelling and depth averaging combine to yield a tractable set of equations, a hyperbolic system that describes the motion of the two constituent phases. If the fluid inertia is small, a reduced model system that is easier to solve may be derived.
Noyes, Jane; Brenner, Maria; Fox, Patricia; Guerin, Ashleigh
2014-05-01
To report a novel review to develop a health systems model of successful transition of children with complex healthcare needs from hospital to home. Children with complex healthcare needs commonly experience an expensive, ineffectual and prolonged nurse-led discharge process. Children gain no benefit from prolonged hospitalization and are exposed to significant harm. Research to enable intervention development and process evaluation across the entire health system is lacking. Novel mixed-method integrative review informed by health systems theory. DATA CINAHL, PsychInfo, EMBASE, PubMed, citation searching, personal contact. REVIEW Informed by consultation with experts. English language studies, opinion/discussion papers reporting research, best practice and experiences of children, parents and healthcare professionals and purposively selected policies/guidelines from 2002-December 2012 were abstracted using Framework synthesis, followed by iterative theory development. Seven critical factors derived from thirty-four sources across five health system levels explained successful discharge (new programme theory). All seven factors are required in an integrated care pathway, with a dynamic communication loop to facilitate effective discharge (new programme logic). Current health system responses were frequently static and critical success factors were commonly absent, thereby explaining ineffectual discharge. The novel evidence-based model, which reconceptualizes 'discharge' as a highly complex longitudinal health system intervention, makes a significant contribution to global knowledge to drive practice development. Research is required to develop process and outcome measures at different time points in the discharge process and future trials are needed to determine the effectiveness of integrated health system discharge models. © 2013 John Wiley & Sons Ltd.
Surrogate Based Uni/Multi-Objective Optimization and Distribution Estimation Methods
NASA Astrophysics Data System (ADS)
Gong, W.; Duan, Q.; Huo, X.
2017-12-01
Parameter calibration has been demonstrated as an effective way to improve the performance of dynamic models, such as hydrological models, land surface models, weather and climate models etc. Traditional optimization algorithms usually cost a huge number of model evaluations, making dynamic model calibration very difficult, or even computationally prohibitive. With the help of a serious of recently developed adaptive surrogate-modelling based optimization methods: uni-objective optimization method ASMO, multi-objective optimization method MO-ASMO, and probability distribution estimation method ASMO-PODE, the number of model evaluations can be significantly reduced to several hundreds, making it possible to calibrate very expensive dynamic models, such as regional high resolution land surface models, weather forecast models such as WRF, and intermediate complexity earth system models such as LOVECLIM. This presentation provides a brief introduction to the common framework of adaptive surrogate-based optimization algorithms of ASMO, MO-ASMO and ASMO-PODE, a case study of Common Land Model (CoLM) calibration in Heihe river basin in Northwest China, and an outlook of the potential applications of the surrogate-based optimization methods.
Australian health system restructuring - what problem is being solved?
Dwyer, Judith M
2004-11-19
BACKGROUND: In recent years, Australian state and territory governments have reviewed and restructured the health systems they lead and regulate. This paper examines the outcomes of the most recent official published reviews of systems and structures; identifies the common themes; and addresses two questions: what problems are being addressed? And how would we know if the changes were successful? RESULTS: In all the broad, systemic reviews, the main health system problems identified were money, hospital utilisation and a weak primary health care system. The solutions are various, but there is a common trend towards centralisation of governance, often at state health authority level, and stronger accountability measures. Other common themes are hospital substitution (services to avoid the need for admission); calls for cooperation across the Commonwealth:state divide, or for its abolition; and the expected range of current efficiency and effectiveness measures (eg amalgamate pathology and support services) and ideas in good currency (eg call centres). The top-down nature of the public review process is noted, along with the political nature of the immediate catalysts for calling on a review. CONCLUSION: The long-standing tension between the pull to centralisation of authority and the need for innovation in care models is heightened by recent changes, which may be counterproductive in an era dominated by the burden of chronic disease. I argue that the current reforms will not succeed in achieving the stated goals unless they make a difference for people with chronic illness. And if this is correct, the most useful focus for evaluation of the success of the reforms may be their impact on the system's ability to develop and deliver better models of care for this growing group of patients.
A Clinical Decision Support System for Breast Cancer Patients
NASA Astrophysics Data System (ADS)
Fernandes, Ana S.; Alves, Pedro; Jarman, Ian H.; Etchells, Terence A.; Fonseca, José M.; Lisboa, Paulo J. G.
This paper proposes a Web clinical decision support system for clinical oncologists and for breast cancer patients making prognostic assessments, using the particular characteristics of the individual patient. This system comprises three different prognostic modelling methodologies: the clinically widely used Nottingham prognostic index (NPI); the Cox regression modelling and a partial logistic artificial neural network with automatic relevance determination (PLANN-ARD). All three models yield a different prognostic index that can be analysed together in order to obtain a more accurate prognostic assessment of the patient. Missing data is incorporated in the mentioned models, a common issue in medical data that was overcome using multiple imputation techniques. Risk group assignments are also provided through a methodology based on regression trees, where Boolean rules can be obtained expressed with patient characteristics.
NASA Astrophysics Data System (ADS)
McIntosh, Bryan
The LSO scintillator crystal commonly used in PET scanners contains a low level of intrinsic radioactivity due to a small amount of Lu-176. This is not usually a concern in routine scanning but can become an issue in small animal imaging, especially when imaging low tracer activity levels. Previously there had been no systematic validation of simulations of this activity; this thesis discusses the validation of a GATE model of intrinsic Lu-176 against results from a bench-top pair of detectors and a Siemens Inveon preclinical PET system. The simulation results matched those from the bench-top system very well, but did not agree as well with results from the complete Inveon system due to a drop-off in system sensitivity at low energies that was not modelled. With this validation the model can now be used with confidence to predict the effects of Lu-176 activity in future PET systems.
NASA Technical Reports Server (NTRS)
Blumenthal, Brennan T.; Elmiligui, Alaa; Geiselhart, Karl A.; Campbell, Richard L.; Maughmer, Mark D.; Schmitz, Sven
2016-01-01
The present paper examines potential propulsive and aerodynamic benefits of integrating a Boundary-Layer Ingestion (BLI) propulsion system into a typical commercial aircraft using the Common Research Model (CRM) geometry and the NASA Tetrahedral Unstructured Software System (TetrUSS). The Numerical Propulsion System Simulation (NPSS) environment is used to generate engine conditions for CFD analysis. Improvements to the BLI geometry are made using the Constrained Direct Iterative Surface Curvature (CDISC) design method. Previous studies have shown reductions of up to 25% in terms of propulsive power required for cruise for other axisymmetric geometries using the BLI concept. An analysis of engine power requirements, drag, and lift coefficients using the baseline and BLI geometries coupled with the NPSS model are shown. Potential benefits of the BLI system relating to cruise propulsive power are quantified using a power balance method, and a comparison to the baseline case is made. Iterations of the BLI geometric design are shown and any improvements between subsequent BLI designs presented. Simulations are conducted for a cruise flight condition of Mach 0.85 at an altitude of 38,500 feet and an angle of attack of 2 deg for all geometries. A comparison between available wind tunnel data, previous computational results, and the original CRM model is presented for model verification purposes along with full results for BLI power savings. Results indicate a 14.4% reduction in engine power requirements at cruise for the BLI configuration over the baseline geometry. Minor shaping of the aft portion of the fuselage using CDISC has been shown to increase the benefit from Boundary-Layer Ingestion further, resulting in a 15.6% reduction in power requirements for cruise as well as a drag reduction of eighteen counts over the baseline geometry.
NASA Technical Reports Server (NTRS)
Blumenthal, Brennan
2016-01-01
This thesis will examine potential propulsive and aerodynamic benefits of integrating a boundary-layer ingestion (BLI) propulsion system with a typical commercial aircraft using the Common Research Model geometry and the NASA Tetrahedral Unstructured Software System (TetrUSS). The Numerical Propulsion System Simulation (NPSS) environment will be used to generate engine conditions for CFD analysis. Improvements to the BLI geometry will be made using the Constrained Direct Iterative Surface Curvature (CDISC) design method. Previous studies have shown reductions of up to 25% in terms of propulsive power required for cruise for other axisymmetric geometries using the BLI concept. An analysis of engine power requirements, drag, and lift coefficients using the baseline and BLI geometries coupled with the NPSS model are shown. Potential benefits of the BLI system relating to cruise propulsive power are quantified using a power balance method and a comparison to the baseline case is made. Iterations of the BLI geometric design are shown and any improvements between subsequent BLI designs presented. Simulations are conducted for a cruise flight condition of Mach 0.85 at an altitude of 38,500 feet and an angle of attack of 2deg for all geometries. A comparison between available wind tunnel data, previous computational results, and the original CRM model is presented for model verification purposes along with full results for BLI power savings. Results indicate a 14.3% reduction in engine power requirements at cruise for the BLI configuration over the baseline geometry. Minor shaping of the aft portion of the fuselage using CDISC has been shown to increase the benefit from boundary-layer ingestion further, resulting in a 15.6% reduction in power requirements for cruise as well as a drag reduction of eighteen counts over the baseline geometry.
Latent geometry of bipartite networks
NASA Astrophysics Data System (ADS)
Kitsak, Maksim; Papadopoulos, Fragkiskos; Krioukov, Dmitri
2017-03-01
Despite the abundance of bipartite networked systems, their organizing principles are less studied compared to unipartite networks. Bipartite networks are often analyzed after projecting them onto one of the two sets of nodes. As a result of the projection, nodes of the same set are linked together if they have at least one neighbor in common in the bipartite network. Even though these projections allow one to study bipartite networks using tools developed for unipartite networks, one-mode projections lead to significant loss of information and artificial inflation of the projected network with fully connected subgraphs. Here we pursue a different approach for analyzing bipartite systems that is based on the observation that such systems have a latent metric structure: network nodes are points in a latent metric space, while connections are more likely to form between nodes separated by shorter distances. This approach has been developed for unipartite networks, and relatively little is known about its applicability to bipartite systems. Here, we fully analyze a simple latent-geometric model of bipartite networks and show that this model explains the peculiar structural properties of many real bipartite systems, including the distributions of common neighbors and bipartite clustering. We also analyze the geometric information loss in one-mode projections in this model and propose an efficient method to infer the latent pairwise distances between nodes. Uncovering the latent geometry underlying real bipartite networks can find applications in diverse domains, ranging from constructing efficient recommender systems to understanding cell metabolism.
Hettinger, Lawrence J.; Kirlik, Alex; Goh, Yang Miang; Buckle, Peter
2015-01-01
Accurate comprehension and analysis of complex sociotechnical systems is a daunting task. Empirically examining, or simply envisioning the structure and behaviour of such systems challenges traditional analytic and experimental approaches as well as our everyday cognitive capabilities. Computer-based models and simulations afford potentially useful means of accomplishing sociotechnical system design and analysis objectives. From a design perspective, they can provide a basis for a common mental model among stakeholders, thereby facilitating accurate comprehension of factors impacting system performance and potential effects of system modifications. From a research perspective, models and simulations afford the means to study aspects of sociotechnical system design and operation, including the potential impact of modifications to structural and dynamic system properties, in ways not feasible with traditional experimental approaches. This paper describes issues involved in the design and use of such models and simulations and describes a proposed path forward to their development and implementation. Practitioner Summary: The size and complexity of real-world sociotechnical systems can present significant barriers to their design, comprehension and empirical analysis. This article describes the potential advantages of computer-based models and simulations for understanding factors that impact sociotechnical system design and operation, particularly with respect to process and occupational safety. PMID:25761227
ERIC Educational Resources Information Center
Lee, Ming-Chi
2010-01-01
Although e-learning has been prompted to various education levels, the intention to continue using such systems is still very low, and the acceptance-discontinuance anomaly phenomenon (i.e., users discontinue using e-learning after initially accepting it) is a common occurrence. This paper synthesizes the expectation-confirmation model (ECM), the…
Toward a Working Model for the Analysis of Cohesion and Coherence in Writing.
ERIC Educational Resources Information Center
Marzano, Robert J.
Although most models of connected discourse are strikingly similar in the types of relationships they describe, they are strikingly different in two areas: the unit of analysis and the dimensions on which cohesion versus coherence are described. Common systems for analyzing written text use the sentence, T-unit, clause, or proposition as the unit…
Modeling crop residue burning experiments to evaluate smoke emissions and plume transport
Luxi Zhou; Kirk R. Baker; Sergey L. Napelenok; George Pouliot; Robert Elleman; Susan M. O' Neill; Shawn P. Urbanski; David C. Wong
2018-01-01
Crop residue burning is a common land management practice that results in emissions of a variety of pollutants with negative health impacts. Modeling systems are used to estimate air quality impacts of crop residue burning to support retrospective regulatory assessments and also for forecasting purposes. Ground and airborne measurements from a recent field experiment...
The Aegean Sea marine security decision support system
NASA Astrophysics Data System (ADS)
Perivoliotis, L.; Krokos, G.; Nittis, K.; Korres, G.
2011-05-01
As part of the integrated ECOOP (European Coastal Sea Operational observing and Forecasting System) project, HCMR upgraded the already existing standalone Oil Spill Forecasting System for the Aegean Sea, initially developed for the Greek Operational Oceanography System (POSEIDON), into an active element of the European Decision Support System (EuroDeSS). The system is accessible through a user friendly web interface where the case scenarios can be fed into the oil spill drift model component, while the synthetic output contains detailed information about the distribution of oil spill particles and the oil spill budget and it is provided both in text based ECOOP common output format and as a series of sequential graphics. The main development steps that were necessary for this transition were the modification of the forcing input data module in order to allow the import of other system products which are usually provided in standard formats such as NetCDF and the transformation of the model's calculation routines to allow use of current, density and diffusivities data in z instead of sigma coordinates. During the implementation of the Aegean DeSS, the system was used in operational mode in order support the Greek marine authorities in handling a real accident that took place in North Aegean area. Furthermore, the introduction of common input and output files by all the partners of EuroDeSS extended the system's interoperability thus facilitating data exchanges and comparison experiments.
The Aegean sea marine security decision support system
NASA Astrophysics Data System (ADS)
Perivoliotis, L.; Krokos, G.; Nittis, K.; Korres, G.
2011-10-01
As part of the integrated ECOOP (European Coastal Sea Operational observing and Forecasting System) project, HCMR upgraded the already existing standalone Oil Spill Forecasting System for the Aegean Sea, initially developed for the Greek Operational Oceanography System (POSEIDON), into an active element of the European Decision Support System (EuroDeSS). The system is accessible through a user friendly web interface where the case scenarios can be fed into the oil spill drift model component, while the synthetic output contains detailed information about the distribution of oil spill particles and the oil spill budget and it is provided both in text based ECOOP common output format and as a series of sequential graphics. The main development steps that were necessary for this transition were the modification of the forcing input data module in order to allow the import of other system products which are usually provided in standard formats such as NetCDF and the transformation of the model's calculation routines to allow use of current, density and diffusivities data in z instead of sigma coordinates. During the implementation of the Aegean DeSS, the system was used in operational mode in order to support the Greek marine authorities in handling a real accident that took place in North Aegean area. Furthermore, the introduction of common input and output files by all the partners of EuroDeSS extended the system's interoperability thus facilitating data exchanges and comparison experiments.
Carrió, Pau; López, Oriol; Sanz, Ferran; Pastor, Manuel
2015-01-01
Computational models based in Quantitative-Structure Activity Relationship (QSAR) methodologies are widely used tools for predicting the biological properties of new compounds. In many instances, such models are used as a routine in the industry (e.g. food, cosmetic or pharmaceutical industry) for the early assessment of the biological properties of new compounds. However, most of the tools currently available for developing QSAR models are not well suited for supporting the whole QSAR model life cycle in production environments. We have developed eTOXlab; an open source modeling framework designed to be used at the core of a self-contained virtual machine that can be easily deployed in production environments, providing predictions as web services. eTOXlab consists on a collection of object-oriented Python modules with methods mapping common tasks of standard modeling workflows. This framework allows building and validating QSAR models as well as predicting the properties of new compounds using either a command line interface or a graphic user interface (GUI). Simple models can be easily generated by setting a few parameters, while more complex models can be implemented by overriding pieces of the original source code. eTOXlab benefits from the object-oriented capabilities of Python for providing high flexibility: any model implemented using eTOXlab inherits the features implemented in the parent model, like common tools and services or the automatic exposure of the models as prediction web services. The particular eTOXlab architecture as a self-contained, portable prediction engine allows building models with confidential information within corporate facilities, which can be safely exported and used for prediction without disclosing the structures of the training series. The software presented here provides full support to the specific needs of users that want to develop, use and maintain predictive models in corporate environments. The technologies used by eTOXlab (web services, VM, object-oriented programming) provide an elegant solution to common practical issues; the system can be installed easily in heterogeneous environments and integrates well with other software. Moreover, the system provides a simple and safe solution for building models with confidential structures that can be shared without disclosing sensitive information.
NASA Technical Reports Server (NTRS)
Hark, Frank; Britton, Paul; Ring, Rob; Novack, Steven D.
2016-01-01
Common Cause Failures (CCFs) are a known and documented phenomenon that defeats system redundancy. CCFS are a set of dependent type of failures that can be caused by: system environments; manufacturing; transportation; storage; maintenance; and assembly, as examples. Since there are many factors that contribute to CCFs, the effects can be reduced, but they are difficult to eliminate entirely. Furthermore, failure databases sometimes fail to differentiate between independent and CCF (dependent) failure and data is limited, especially for launch vehicles. The Probabilistic Risk Assessment (PRA) of NASA's Safety and Mission Assurance Directorate at Marshal Space Flight Center (MFSC) is using generic data from the Nuclear Regulatory Commission's database of common cause failures at nuclear power plants to estimate CCF due to the lack of a more appropriate data source. There remains uncertainty in the actual magnitude of the common cause risk estimates for different systems at this stage of the design. Given the limited data about launch vehicle CCF and that launch vehicles are a highly redundant system by design, it is important to make design decisions to account for a range of values for independent and CCFs. When investigating the design of the one-out-of-two component redundant system for launch vehicles, a response surface was constructed to represent the impact of the independent failure rate versus a common cause beta factor effect on a system's failure probability. This presentation will define a CCF and review estimation calculations. It gives a summary of reduction methodologies and a review of examples of historical CCFs. Finally, it presents the response surface and discusses the results of the different CCFs on the reliability of a one-out-of-two system.
NASA Technical Reports Server (NTRS)
Hark, Frank; Britton, Paul; Ring, Rob; Novack, Steven D.
2015-01-01
Common Cause Failures (CCFs) are a known and documented phenomenon that defeats system redundancy. CCFS are a set of dependent type of failures that can be caused by: system environments; manufacturing; transportation; storage; maintenance; and assembly, as examples. Since there are many factors that contribute to CCFs, the effects can be reduced, but they are difficult to eliminate entirely. Furthermore, failure databases sometimes fail to differentiate between independent and CCF (dependent) failure and data is limited, especially for launch vehicles. The Probabilistic Risk Assessment (PRA) of NASA's Safety and Mission Assurance Directorate at Marshall Space Flight Center (MFSC) is using generic data from the Nuclear Regulatory Commission's database of common cause failures at nuclear power plants to estimate CCF due to the lack of a more appropriate data source. There remains uncertainty in the actual magnitude of the common cause risk estimates for different systems at this stage of the design. Given the limited data about launch vehicle CCF and that launch vehicles are a highly redundant system by design, it is important to make design decisions to account for a range of values for independent and CCFs. When investigating the design of the one-out-of-two component redundant system for launch vehicles, a response surface was constructed to represent the impact of the independent failure rate versus a common cause beta factor effect on a system's failure probability. This presentation will define a CCF and review estimation calculations. It gives a summary of reduction methodologies and a review of examples of historical CCFs. Finally, it presents the response surface and discusses the results of the different CCFs on the reliability of a one-out-of-two system.
Weeks, Margaret R; Li, Jianghong; Lounsbury, David; Green, Helena Danielle; Abbott, Maryann; Berman, Marcie; Rohena, Lucy; Gonzalez, Rosely; Lang, Shawn; Mosher, Heather
2017-12-01
Achieving community-level goals to eliminate the HIV epidemic requires coordinated efforts through community consortia with a common purpose to examine and critique their own HIV testing and treatment (T&T) care system and build effective tools to guide their efforts to improve it. Participatory system dynamics (SD) modeling offers conceptual, methodological, and analytical tools to engage diverse stakeholders in systems conceptualization and visual mapping of dynamics that undermine community-level health outcomes and identify those that can be leveraged for systems improvement. We recruited and engaged a 25-member multi-stakeholder Task Force, whose members provide or utilize HIV-related services, to participate in SD modeling to examine and address problems of their local HIV T&T service system. Findings from the iterative model building sessions indicated Task Force members' increasingly complex understanding of the local HIV care system and demonstrated their improved capacity to visualize and critique multiple models of the HIV T&T service system and identify areas of potential leverage. Findings also showed members' enhanced communication and consensus in seeking deeper systems understanding and options for solutions. We discuss implications of using these visual SD models for subsequent simulation modeling of the T&T system and for other community applications to improve system effectiveness. © Society for Community Research and Action 2017.
Qualitative models for space system engineering
NASA Technical Reports Server (NTRS)
Forbus, Kenneth D.
1990-01-01
The objectives of this project were: (1) to investigate the implications of qualitative modeling techniques for problems arising in the monitoring, diagnosis, and design of Space Station subsystems and procedures; (2) to identify the issues involved in using qualitative models to enhance and automate engineering functions. These issues include representing operational criteria, fault models, alternate ontologies, and modeling continuous signals at a functional level of description; and (3) to develop a prototype collection of qualitative models for fluid and thermal systems commonly found in Space Station subsystems. Potential applications of qualitative modeling to space-systems engineering, including the notion of intelligent computer-aided engineering are summarized. Emphasis is given to determining which systems of the proposed Space Station provide the most leverage for study, given the current state of the art. Progress on using qualitative models, including development of the molecular collection ontology for reasoning about fluids, the interaction of qualitative and quantitative knowledge in analyzing thermodynamic cycles, and an experiment on building a natural language interface to qualitative reasoning is reported. Finally, some recommendations are made for future research.
NASA Astrophysics Data System (ADS)
Wahid, A.; Taqwallah, H. M. H.
2018-03-01
Compressors and a steam reformer are the important units in biohydrogen from biomass plant. The compressors are useful for achieving high-pressure operating conditions while the steam reformer is the main process to produce H2 gas. To control them, in this research used a model predictive control (MPC) expected to have better controller performance than conventional controllers. Because of the explicit model empowerment in MPC, obtaining a better model is the main objective before employing MPC. The common way to get the empirical model is through the identification system, so that obtained a first-order plus dead-time (FOPDT) model. This study has already improved that way since used the system re-identification (SRI) based on closed loop mode. Based on this method the results of the compressor pressure control and temperature control of steam reformer were that MPC based on system re-identification (MPC-SRI) has better performance than MPC without system re-identification (MPCWSRI) and the proportional-integral (PI) controller, by % improvement of 73% against MPCWSRI and 75% against the PI controller.
ERIC Educational Resources Information Center
Dörnyei, Zoltán
2014-01-01
While approaching second language acquisition from a complex dynamic systems perspective makes a lot of intuitive sense, it is difficult for a number of reasons to operationalise such a dynamic approach in research terms. For example, the most common research paradigms in the social sciences tend to examine variables in relative isolation rather…
Human Factors in Command-and-Control System Procurement,
1985-12-01
more common terms, whether workload is too high. Generally, workload is a concept that is open to many intepretations . From a modeller’s viewpoint...solution, according to Meister and Farr (1966). may be to provide designers with better means of analysing conceptual systems. The goals of this report are thus closely aligned with that philosophy. DATE FILMED
EPA’s Risk-Informed Materials Management (RIMM) tool system is a modeling approach that helps risk assessors evaluate the safety of managing raw, reused, or waste material streams via a variety of common scenarios (e.g., application to farms, use as a component in road cons...
Nonlinear Spectroscopy Study of Vibrational Self-Trapping in Hydrogen Bonded Crystals
NASA Astrophysics Data System (ADS)
Edler, Julian; Hamm, Peter
Femtosecond pump probe spectroscopy proves that self-trapping occurs in the NH and amide I band of crystalline acetanilide (ACN). The phonon modes that mediate the self-trapping are identified. Comparison between ACN and N-methylacetamide, both model systems for proteins, shows that self-trapping is a common feature in hydrogen bonded systems.
Measuring the Resilience of Advanced Life Support Systems
NASA Technical Reports Server (NTRS)
Bell, Ann Maria; Dearden, Richard; Levri, Julie A.
2002-01-01
Despite the central importance of crew safety in designing and operating a life support system, the metric commonly used to evaluate alternative Advanced Life Support (ALS) technologies does not currently provide explicit techniques for measuring safety. The resilience of a system, or the system s ability to meet performance requirements and recover from component-level faults, is fundamentally a dynamic property. This paper motivates the use of computer models as a tool to understand and improve system resilience throughout the design process. Extensive simulation of a hybrid computational model of a water revitalization subsystem (WRS) with probabilistic, component-level faults provides data about off-nominal behavior of the system. The data can then be used to test alternative measures of resilience as predictors of the system s ability to recover from component-level faults. A novel approach to measuring system resilience using a Markov chain model of performance data is also developed. Results emphasize that resilience depends on the complex interaction of faults, controls, and system dynamics, rather than on simple fault probabilities.
Omics analysis of mouse brain models of human diseases.
Paban, Véronique; Loriod, Béatrice; Villard, Claude; Buee, Luc; Blum, David; Pietropaolo, Susanna; Cho, Yoon H; Gory-Faure, Sylvie; Mansour, Elodie; Gharbi, Ali; Alescio-Lautier, Béatrice
2017-02-05
The identification of common gene/protein profiles related to brain alterations, if they exist, may indicate the convergence of the pathogenic mechanisms driving brain disorders. Six genetically engineered mouse lines modelling neurodegenerative diseases and neuropsychiatric disorders were considered. Omics approaches, including transcriptomic and proteomic methods, were used. The gene/protein lists were used for inter-disease comparisons and further functional and network investigations. When the inter-disease comparison was performed using the gene symbol identifiers, the number of genes/proteins involved in multiple diseases decreased rapidly. Thus, no genes/proteins were shared by all 6 mouse models. Only one gene/protein (Gfap) was shared among 4 disorders, providing strong evidence that a common molecular signature does not exist among brain diseases. The inter-disease comparison of functional processes showed the involvement of a few major biological processes indicating that brain diseases of diverse aetiologies might utilize common biological pathways in the nervous system, without necessarily involving similar molecules. Copyright © 2016 Elsevier B.V. All rights reserved.
Designing and encoding models for synthetic biology.
Endler, Lukas; Rodriguez, Nicolas; Juty, Nick; Chelliah, Vijayalakshmi; Laibe, Camille; Li, Chen; Le Novère, Nicolas
2009-08-06
A key component of any synthetic biology effort is the use of quantitative models. These models and their corresponding simulations allow optimization of a system design, as well as guiding their subsequent analysis. Once a domain mostly reserved for experts, dynamical modelling of gene regulatory and reaction networks has been an area of growth over the last decade. There has been a concomitant increase in the number of software tools and standards, thereby facilitating model exchange and reuse. We give here an overview of the model creation and analysis processes as well as some software tools in common use. Using markup language to encode the model and associated annotation, we describe the mining of components, their integration in relational models, formularization and parametrization. Evaluation of simulation results and validation of the model close the systems biology 'loop'.
Brossier, Nicole M; Carroll, Steven L
2012-05-01
Neurofibromatosis type 1 (NF1), the most common genetic disorder affecting the human nervous system, is characterized by the development of multiple benign Schwann cell tumors in skin and large peripheral nerves. These neoplasms, which are termed dermal and plexiform neurofibromas respectively, have distinct clinical courses; of particular note, plexiform, but not dermal, neurofibromas often undergo malignant progression to form malignant peripheral nerve sheath tumors (MPNSTs), the most common malignancy occurring in NF1 patients. In recent years, a number of genetically engineered mouse models have been created to investigate the molecular mechanisms driving the pathogenesis of these tumors. These models have been designed to address key questions including: (1) whether NF1 loss in the Schwann cell lineage is essential for tumorigenesis; (2) what cell type(s) in the Schwann cell lineage gives rise to dermal neurofibromas, plexiform neurofibromas and MPNSTs; (3) how the tumor microenvironment contributes to neoplasia; (4) what additional mutations contribute to neurofibroma-MPNST progression; (5) what role different neurofibromin-regulated Ras proteins play in this process and (6) how dysregulated growth factor signaling facilitates PNS tumorigenesis. In this review, we summarize the major findings from each of these models and their limitations as well as how discrepancies between these models may be reconciled. We also discuss how information gleaned from these models can be synthesized to into a comprehensive model of tumor formation in peripheral nervous system and consider several of the major questions that remain unanswered about this process. Copyright © 2011 Elsevier Inc. All rights reserved.
Modulation instability in high power laser amplifiers.
Rubenchik, Alexander M; Turitsyn, Sergey K; Fedoruk, Michail P
2010-01-18
The modulation instability (MI) is one of the main factors responsible for the degradation of beam quality in high-power laser systems. The so-called B-integral restriction is commonly used as the criteria for MI control in passive optics devices. For amplifiers the adiabatic model, assuming locally the Bespalov-Talanov expression for MI growth, is commonly used to estimate the destructive impact of the instability. We present here the exact solution of MI development in amplifiers. We determine the parameters which control the effect of MI in amplifiers and calculate the MI growth rate as a function of those parameters. The safety range of operational parameters is presented. The results of the exact calculations are compared with the adiabatic model, and the range of validity of the latest is determined. We demonstrate that for practical situations the adiabatic approximation noticeably overestimates MI. The additional margin of laser system design is quantified.
Treatment of systematic errors in land data assimilation systems
NASA Astrophysics Data System (ADS)
Crow, W. T.; Yilmaz, M.
2012-12-01
Data assimilation systems are generally designed to minimize the influence of random error on the estimation of system states. Yet, experience with land data assimilation systems has also revealed the presence of large systematic differences between model-derived and remotely-sensed estimates of land surface states. Such differences are commonly resolved prior to data assimilation through implementation of a pre-processing rescaling step whereby observations are scaled (or non-linearly transformed) to somehow "match" comparable predictions made by an assimilation model. While the rationale for removing systematic differences in means (i.e., bias) between models and observations is well-established, relatively little theoretical guidance is currently available to determine the appropriate treatment of higher-order moments during rescaling. This talk presents a simple analytical argument to define an optimal linear-rescaling strategy for observations prior to their assimilation into a land surface model. While a technique based on triple collocation theory is shown to replicate this optimal strategy, commonly-applied rescaling techniques (e.g., so called "least-squares regression" and "variance matching" approaches) are shown to represent only sub-optimal approximations to it. Since the triple collocation approach is likely infeasible in many real-world circumstances, general advice for deciding between various feasible (yet sub-optimal) rescaling approaches will be presented with an emphasis of the implications of this work for the case of directly assimilating satellite radiances. While the bulk of the analysis will deal with linear rescaling techniques, its extension to nonlinear cases will also be discussed.
NASA Technical Reports Server (NTRS)
Kostov, Veselin B.; Moore, Keavin; Tamayo, Daniel; Jayawardhana, Ray; Rinehart, Stephen A.
2016-01-01
Inspired by the recent Kepler discoveries of circumbinary planets orbiting nine close binary stars, we explore the fate of the former as the latter evolve off the main sequence. We combine binary star evolution models with dynamical simulations to study the orbital evolution of these planets as their hosts undergo common-envelope stages, losing in the process a tremendous amount of mass on dynamical timescales. Five of the systems experience at least one Roche-lobe overflow and common-envelope stages (Kepler-1647 experiences three), and the binary stars either shrink to very short orbits or coalesce; two systems trigger a double-degenerate supernova explosion. Kepler's circumbinary planets predominantly remain gravitationally bound at the end of the common-envelope phase, migrate to larger orbits, and may gain significant eccentricity; their orbital expansion can be more than an order of magnitude and can occur over the course of a single planetary orbit. The orbits these planets can reach are qualitatively consistent with those of the currently known post-common-envelope, eclipse-time variations circumbinary candidates. Our results also show that circumbinary planets can experience both modes of orbital expansion (adiabatic and non-adiabatic) if their host binaries undergo more than one common-envelope stage; multiplanet circumbinary systems like Kepler-47 can experience both modes during the same common-envelope stage. Additionally, unlike Mercury orbiting the Sun, a circumbinary planet with the same semi-major axis can survive the common envelope evolution of a close binary star with a total mass of 1 Solar Mass.
Isolation transformers for utility-interactive photovoltaic systems
NASA Astrophysics Data System (ADS)
Kern, E. C., Jr.
1982-12-01
Isolation transformers are used in some photovoltaic systems to isolate the photovoltaic system common mode voltage from the utility distribution system. In early system experiments with grid connected photovoltaics, such transformers were the source of significant power losses. A project at the Lincoln Laboratory and at Allied Chemical Corporation developed an improved isolation transformer to minimize such power losses. Experimental results and an analytical model of conventional and improved transformers are presented, showing considerable reductions of losses associated with the improved transformer.
Transactions in domain-specific information systems
NASA Astrophysics Data System (ADS)
Zacek, Jaroslav
2017-07-01
Substantial number of the current information system (IS) implementations is based on transaction approach. In addition, most of the implementations are domain-specific (e.g. accounting IS, resource planning IS). Therefore, we have to have a generic transaction model to build and verify domain-specific IS. The paper proposes a new transaction model for domain-specific ontologies. This model is based on value oriented business process modelling technique. The transaction model is formalized by the Petri Net theory. First part of the paper presents common business processes and analyses related to business process modeling. Second part defines the transactional model delimited by REA enterprise ontology paradigm and introduces states of the generic transaction model. The generic model proposal is defined and visualized by the Petri Net modelling tool. Third part shows application of the generic transaction model. Last part of the paper concludes results and discusses a practical usability of the generic transaction model.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Brouillette, Greg A.
These are slides for various presentations on C41SR and urban disasters disasters response and recovery tools. These are all mainly charts and images of disaster response and recovery tools. Slides included have headings such as the following: vignette of a disaster response, situational awareness and common operating picture available to EOC, plume modeling capability, Program ASPECT Chemical Response Products, EPA ASPECT - Hurricane RITA Response 9/25/2005, Angel Fire Imagery, incident commander's view/police chief's view/ EMS' view, common situational awareness and collaborative planning, exercise, training capability, systems diagram, Austere Challenge 06 Sim/C4 Requirements, common situational awareness and collaborative planning, exercise, trainingmore » environment, common situational awareness, real world, crisis response, and consequence management.« less
ERIC Educational Resources Information Center
Painter, Kirstin
2012-01-01
Background: Systems of care is a family centered, strengths-based service delivery model for treating youth experiencing a serious emotional disturbance. Wraparound is the most common method of service delivery adopted by states and communities as a way to adhere to systems of care philosophy. Objective: The purpose of this study was to evaluate…
Managing Variation in Services in a Software Product Line Context
2010-05-01
Oriented Domain Analysis ( FODA ) Feasibility Study (CMU/SEI-90-TR-021, ADA235785). Software Engineering Institute, Carnegie Mellon University, 1990...the systems in the product line, and a plan for building the systems. Product line scope and product line analysis define the boundaries and...systems, as well as expected ways in which they may vary. Product line analysis applies established modeling techniques to engineer the common and
OntoTrader: An Ontological Web Trading Agent Approach for Environmental Information Retrieval
Iribarne, Luis; Padilla, Nicolás; Ayala, Rosa; Asensio, José A.; Criado, Javier
2014-01-01
Modern Web-based Information Systems (WIS) are becoming increasingly necessary to provide support for users who are in different places with different types of information, by facilitating their access to the information, decision making, workgroups, and so forth. Design of these systems requires the use of standardized methods and techniques that enable a common vocabulary to be defined to represent the underlying knowledge. Thus, mediation elements such as traders enrich the interoperability of web components in open distributed systems. These traders must operate with other third-party traders and/or agents in the system, which must also use a common vocabulary for communication between them. This paper presents the OntoTrader architecture, an Ontological Web Trading agent based on the OMG ODP trading standard. It also presents the ontology needed by some system agents to communicate with the trading agent and the behavioral framework for the SOLERES OntoTrader agent, an Environmental Management Information System (EMIS). This framework implements a “Query-Searching/Recovering-Response” information retrieval model using a trading service, SPARQL notation, and the JADE platform. The paper also presents reflection, delegation and, federation mediation models and describes formalization, an experimental testing environment in three scenarios, and a tool which allows our proposal to be evaluated and validated. PMID:24977211
OntoTrader: an ontological Web trading agent approach for environmental information retrieval.
Iribarne, Luis; Padilla, Nicolás; Ayala, Rosa; Asensio, José A; Criado, Javier
2014-01-01
Modern Web-based Information Systems (WIS) are becoming increasingly necessary to provide support for users who are in different places with different types of information, by facilitating their access to the information, decision making, workgroups, and so forth. Design of these systems requires the use of standardized methods and techniques that enable a common vocabulary to be defined to represent the underlying knowledge. Thus, mediation elements such as traders enrich the interoperability of web components in open distributed systems. These traders must operate with other third-party traders and/or agents in the system, which must also use a common vocabulary for communication between them. This paper presents the OntoTrader architecture, an Ontological Web Trading agent based on the OMG ODP trading standard. It also presents the ontology needed by some system agents to communicate with the trading agent and the behavioral framework for the SOLERES OntoTrader agent, an Environmental Management Information System (EMIS). This framework implements a "Query-Searching/Recovering-Response" information retrieval model using a trading service, SPARQL notation, and the JADE platform. The paper also presents reflection, delegation and, federation mediation models and describes formalization, an experimental testing environment in three scenarios, and a tool which allows our proposal to be evaluated and validated.
Design optimization of a prescribed vibration system using conjoint value analysis
NASA Astrophysics Data System (ADS)
Malinga, Bongani; Buckner, Gregory D.
2016-12-01
This article details a novel design optimization strategy for a prescribed vibration system (PVS) used to mechanically filter solids from fluids in oil and gas drilling operations. A dynamic model of the PVS is developed, and the effects of disturbance torques are detailed. This model is used to predict the effects of design parameters on system performance and efficiency, as quantified by system attributes. Conjoint value analysis, a statistical technique commonly used in marketing science, is utilized to incorporate designer preferences. This approach effectively quantifies and optimizes preference-based trade-offs in the design process. The effects of designer preferences on system performance and efficiency are simulated. This novel optimization strategy yields improvements in all system attributes across all simulated vibration profiles, and is applicable to other industrial electromechanical systems.
Tougas-Tellier, Marie-Andrée; Morin, Jean; Hatin, Daniel; Lavoie, Claude
2015-01-01
Climate change will likely affect flooding regimes, which have a large influence on the functioning of freshwater riparian wetlands. Low water levels predicted for several fluvial systems make wetlands especially vulnerable to the spread of invaders, such as the common reed (Phragmites australis), one of the most invasive species in North America. We developed a model to map the distribution of potential germination grounds of the common reed in freshwater wetlands of the St. Lawrence River (Québec, Canada) under current climate conditions and used this model to predict their future distribution under two climate change scenarios simulated for 2050. We gathered historical and recent (remote sensing) data on the distribution of common reed stands for model calibration and validation purposes, then determined the parameters controlling the species establishment by seed. A two-dimensional model and the identified parameters were used to simulate the current (2010) and future (2050) distribution of germination grounds. Common reed stands are not widespread along the St. Lawrence River (212 ha), but our model suggests that current climate conditions are already conducive to considerable further expansion (>16,000 ha). Climate change may also exacerbate the expansion, particularly if river water levels drop, which will expose large bare areas propitious to seed germination. This phenomenon may be particularly important in one sector of the river, where existing common reed stands could increase their areas by a factor of 100, potentially creating the most extensive reedbed complex in North America. After colonizing salt and brackishwater marshes, the common reed could considerably expand into the freshwater marshes of North America which cover several million hectares. The effects of common reed expansion on biodiversity are difficult to predict, but likely to be highly deleterious given the competitiveness of the invader and the biological richness of freshwater wetlands. PMID:26380675
Single-phase power distribution system power flow and fault analysis
NASA Technical Reports Server (NTRS)
Halpin, S. M.; Grigsby, L. L.
1992-01-01
Alternative methods for power flow and fault analysis of single-phase distribution systems are presented. The algorithms for both power flow and fault analysis utilize a generalized approach to network modeling. The generalized admittance matrix, formed using elements of linear graph theory, is an accurate network model for all possible single-phase network configurations. Unlike the standard nodal admittance matrix formulation algorithms, the generalized approach uses generalized component models for the transmission line and transformer. The standard assumption of a common node voltage reference point is not required to construct the generalized admittance matrix. Therefore, truly accurate simulation results can be obtained for networks that cannot be modeled using traditional techniques.
Validation of a model for investigating red cell mass changes during weightlessness
NASA Technical Reports Server (NTRS)
Leonard, J. I.
1976-01-01
The model, both the conceptual model and simulation model, provided a convenient framework on which to demonstrate the commonality between such diverse stresses as descent from altitude, red cell infusions, bed rest, and weightlessness. The results suggest that all of these stresses induce an increased blood hematocrit leading to tissue hyperoxia and eventual inhibition of the erythyocyte producing circuit until the hyperoxic condition is relieved. The erythropoietic system was acting, in these situations, as if it were an hematocrit sensor and regulator. In these terms the decreases in red cell mass during Skylab may be explained in terms of normal feedback regulation of the erythropoietic system in the face of sustained decreases in plasma colume.
Modeling and formal analysis of urban road traffic
NASA Astrophysics Data System (ADS)
Avram, Camelia; Machado, José; Aştilean, Adina
2013-10-01
Modern life in cities leads to complex urban traffic road and, sometimes, to go from one point to another, in a city, is a hard and very complex task. The use of assisted systems for helping drivers on their task of reaching the desired destination is being common, mainly systems like GPS location systems or other similar systems. The main gap of those systems is that they are not able to assist drivers when some unexpected changes occur, like accidents, or another unexpected situations. In this context, it would be desirable to have a dynamic system to inform the drivers, about everything that is happening "online". This work is inserted in this context and the work presented here is one part of a bigger project that has, as main goal, to be a dynamic system for assisting drivers under hard conditions of urban road traffic. In this paper is modeled, and formally analyzed, the intersection of four street segments, in order to take some considerations about this subject. This paper presents the model of the considered system, using timed automata formalism. The validation and verification of the road traffic model it is realized using UPPAAL model-checker.
Method for automated building of spindle thermal model with use of CAE system
NASA Astrophysics Data System (ADS)
Kamenev, S. V.
2018-03-01
The spindle is one of the most important units of the metal-cutting machine tool. Its performance is critical to minimize the machining error, especially the thermal error. Various methods are applied to improve the thermal behaviour of spindle units. One of the most important methods is mathematical modelling based on the finite element analysis. The most common approach for its realization is the use of CAE systems. This approach, however, is not capable to address the number of important effects that need to be taken into consideration for proper simulation. In the present article, the authors propose the solution to overcome these disadvantages using automated thermal model building for the spindle unit utilizing the CAE system ANSYS.
Investigating the Effect of Damage Progression Model Choice on Prognostics Performance
NASA Technical Reports Server (NTRS)
Daigle, Matthew; Roychoudhury, Indranil; Narasimhan, Sriram; Saha, Sankalita; Saha, Bhaskar; Goebel, Kai
2011-01-01
The success of model-based approaches to systems health management depends largely on the quality of the underlying models. In model-based prognostics, it is especially the quality of the damage progression models, i.e., the models describing how damage evolves as the system operates, that determines the accuracy and precision of remaining useful life predictions. Several common forms of these models are generally assumed in the literature, but are often not supported by physical evidence or physics-based analysis. In this paper, using a centrifugal pump as a case study, we develop different damage progression models. In simulation, we investigate how model changes influence prognostics performance. Results demonstrate that, in some cases, simple damage progression models are sufficient. But, in general, the results show a clear need for damage progression models that are accurate over long time horizons under varied loading conditions.
The Hydrofacies Approach and Why ln K σ 2 <5-10 is Unlikely
NASA Astrophysics Data System (ADS)
Fogg, G. E.
2004-12-01
When heterogeneity of geologic systems is characterized in terms of hydrofacies rather than solely based on K measurements, the resulting flow and transport models typically contain not only aquifer materials but also significant volumes (10-70%) of aquitard materials. This leads to clear, heuristic rationale for the ln K σ 2 commonly exceeding 5 to 10, contradicting published data on ln K σ 2. I will explain the inconsistencies between commonly held assumptions of low (<1-2) ln K σ 2 and abundant geologic and hydrologic field data that indicate substantially larger values. The K data commonly cited in support of the low ln K σ 2 assumption have been misinterpreted because of unintentional, biased sampling. Geologic fundamentals and field data indicate that ln K σ 2 is commonly >10 and can easily exceed 20 in typical sedimentary deposits (not surficial soils) at spatial scales on the order of 101 to 102 m. Presence of large ln K σ 2 can be paramount in transport models and is often requisite for modeling observed transport phenomena such as preferential flow, extreme tailing, difficult remediation including frequent pump-and-treat failure, and significant, unanticipated mixing of groundwater ages.
Dynamics of land use and common-resource pressures in terrestrial-aquatic environments
NASA Astrophysics Data System (ADS)
Lazarus, E.; Bell, K. P.
2010-12-01
Common-pool resource problems can arise in aquatic systems such as lakes, rivers, and open coastlines where individual land-use decisions produce collective, emergent effects at the watershed scale. A body of highly generalized modeling experiments has illustrated ways in which simple, opposing tendencies among individuals—imitative versus self-initiated actions, for example—can result in richly complex behaviors. If the dynamics of those opposing tendencies are translated into different land uses (development, extraction, working land, conservation), each entailing a different environmental consequence, then feedbacks between land-use decisions and resulting changes to the physical environment (which in turn influence subsequent land-use decisions) cause the environmental and social systems of the watershed to become coupled. We present the early results of an exploratory, spatially-extended model that couples a simplified riparian system to a hypothetical group of landowners, each of whom can choose between property development, placing property under a conservation easement, or taking no action and effectively maintaining the property's status quo. As in the generalized experiments, landowner behavior depends on two sets of opposing tendencies, one of which is imitation versus self-initiation. The other is a preference for property consolidation or subdivision; large property holdings can be subdivided into smaller parcels, and small parcels can be bought up by an owner to amass a larger property. Land-use decisions taken by landowners in the model affect the riparian system in process-based ways (sedimentation, eutrophication, water quality) that then inform subsequent social interactions and decisions. Isolating the basic dynamics of this kind of socio-environmental system allows us to pursue a number of questions relevant to resource management. What social and environmental circumstances in this framework result in continuous conservation corridors as opposed to patchworks of easements? How do different kinds of easements, such as those that permit private development and subdivision, affect the integrity of the commons? More broadly, how do changing, heterogeneous patterns of mixed land uses (with different inherent time scales) affect environmental resilience? Dynamical insights from the idealized model dovetail with empirical economic analyses of aquatic-system conservation currently underway in the State of Maine, where opportunities for case studies of specific coupled human/aquatic systems and cross-case comparisons are abundant. This investigation is part of an interdisciplinary project under the auspices of the Sustainability Solutions Initiative at the University of Maine.
NASA Technical Reports Server (NTRS)
Bonanne, Kevin H.
2011-01-01
Model-based Systems Engineering (MBSE) is an emerging methodology that can be leveraged to enhance many system development processes. MBSE allows for the centralization of an architecture description that would otherwise be stored in various locations and formats, thus simplifying communication among the project stakeholders, inducing commonality in representation, and expediting report generation. This paper outlines the MBSE approach taken to capture the processes of two different, but related, architectures by employing the Systems Modeling Language (SysML) as a standard for architecture description and the modeling tool MagicDraw. The overarching goal of this study was to demonstrate the effectiveness of MBSE as a means of capturing and designing a mission systems architecture. The first portion of the project focused on capturing the necessary system engineering activities that occur when designing, developing, and deploying a mission systems architecture for a space mission. The second part applies activities from the first to an application problem - the system engineering of the Orion Flight Test 1 (OFT-1) End-to-End Information System (EEIS). By modeling the activities required to create a space mission architecture and then implementing those activities in an application problem, the utility of MBSE as an approach to systems engineering can be demonstrated.
A Cost and Performance System (CAPS) in a Federal agency
NASA Technical Reports Server (NTRS)
Huseonia, W. F.; Penton, P. G.
1994-01-01
Cost and Performance System (CAPS) is an automated system used from the planning phase through implementation to analysis and documentation. Data is retrievable or available for analysis of cost versus performance anomalies. CAPS provides a uniform system across intra- and international elements. A common system is recommended throughout an entire cost or profit center. Data can be easily accumulated and aggregated into higher levels of tracking and reporting of cost and performance.The level and quality of performance or productivity is indicated in the CAPS model and its process. The CAPS model provides the necessary decision information and insight to the principal investigator/project engineer for a successful project management experience. CAPS provides all levels of management with the appropriate detailed level of data.
Analysis, requirements and development of a collaborative social and medical services data model.
Bobroff, R B; Petermann, C A; Beck, J R; Buffone, G J
1994-01-01
In any medical and social service setting, patient data must be readily shared among multiple providers for delivery of expeditious, quality care. This paper describes the development and implementation of a generalized social and medical services data model for an ambulatory population. The model, part of the Collaborative Social and Medical Services System Project, is based on the data needs of the Baylor College of Medicine Teen Health Clinics and follows the guidelines of the ANSI HISPP/MSDS JWG for a Common Data Model. Design details were determined by informal staff interviews, operational observations, and examination of clinic guidelines and forms. The social and medical services data model is implemented using object-oriented data modeling techniques and will be implemented in C++ using an Object-Oriented Database Management System.
The Early History of Bioenergy
NASA Astrophysics Data System (ADS)
Radu, Popa
Energy is most commonly defined as the potential to do work. The maintenance of the living state requires a constant flow of energy through the system. The concept of energy is not easily implemented in computational models of life and is therefore often ignored in artificial life models. Some models even regard as irrelevant the energetic problematic (dissipation, irreversibility, couplings, energy currencies), in the physical realization of a biological system" (Ruiz-Mirazo et al. 1998). Examples of such models are Rosen's (M,R)-system, Varela's autopoietic models, Kauffman's autocatalytic set, and Fontana's algorithmic chemistry (see Appendix A). However, many origin-of-life theories maintain the primordial importance of energy for early life. Although everyone accepts that energetic constraints are important when describing material-based living systems, a problem arises when we have to consider whether or not they affect the very logic of the organization (Morán et al. 1999). It is argued here that energy considerations are not only primordial, but intimately related to the essence of life as well.
Simple models for rope substructure mechanics: application to electro-mechanical lifts
NASA Astrophysics Data System (ADS)
Herrera, I.; Kaczmarczyk, S.
2016-05-01
Mechanical systems modelled as rigid mass elements connected by tensioned slender structural members such as ropes and cables represent quite common substructures used in lift engineering and hoisting applications. Special interest is devoted by engineers and researchers to the vibratory response of such systems for optimum performance and durability. This paper presents simplified models that can be employed to determine the natural frequencies of systems having substructures of two rigid masses constrained by tensioned rope/cable elements. The exact solution for free un-damped longitudinal displacement response is discussed in the context of simple two-degree-of-freedom models. The results are compared and the influence of characteristics parameters such as the ratio of the average mass of the two rigid masses with respect to the rope mass and the deviation ratio of the two rigid masses with respect to the average mass is analyzed. This analysis gives criteria for the application of such simplified models in complex elevator and hoisting system configurations.
Practical robustness measures in multivariable control system analysis. Ph.D. Thesis
NASA Technical Reports Server (NTRS)
Lehtomaki, N. A.
1981-01-01
The robustness of the stability of multivariable linear time invariant feedback control systems with respect to model uncertainty is considered using frequency domain criteria. Available robustness tests are unified under a common framework based on the nature and structure of model errors. These results are derived using a multivariable version of Nyquist's stability theorem in which the minimum singular value of the return difference transfer matrix is shown to be the multivariable generalization of the distance to the critical point on a single input, single output Nyquist diagram. Using the return difference transfer matrix, a very general robustness theorem is presented from which all of the robustness tests dealing with specific model errors may be derived. The robustness tests that explicitly utilized model error structure are able to guarantee feedback system stability in the face of model errors of larger magnitude than those robustness tests that do not. The robustness of linear quadratic Gaussian control systems are analyzed.
A Petri net synthesis theory for modeling flexible manufacturing systems.
Jeng, M D
1997-01-01
A theory that synthesizes Petri nets for modeling flexible manufacturing systems is presented. The theory adopts a bottom-up or modular-composition approach to construct net models. Each module is modeled as a resource control net (RCN), which represents a subsystem that controls a resource type in a flexible manufacturing system. Interactions among the modules are described as the common transition and transition subnets. The net obtained by merging the modules with two minimal restrictions is shown to be conservative and thus bounded. An algorithm is developed to detect two sufficient conditions for structural liveness of the net. The algorithm examines only the net's structure and the initial marking, and appears to be more efficient than state enumeration techniques such as the reachability tree method. In this paper, the sufficient conditions for liveness are shown to be related to some structural objects called siphons. To demonstrate the applicability of the theory, a flexible manufacturing system of a moderate size is modeled and analyzed using the proposed theory.
Zhang, Pengfei; Zhang, Rui; Liu, Jinhai; Lu, Xiaochun
2018-01-01
This study proposes two models for precise time transfer using the BeiDou Navigation Satellite System triple-frequency signals: ionosphere-free (IF) combined precise point positioning (PPP) model with two dual-frequency combinations (IF-PPP1) and ionosphere-free combined PPP model with a single triple-frequency combination (IF-PPP2). A dataset with a short baseline (with a common external time frequency) and a long baseline are used for performance assessments. The results show that IF-PPP1 and IF-PPP2 models can both be used for precise time transfer using BeiDou Navigation Satellite System (BDS) triple-frequency signals, and the accuracy and stability of time transfer is the same in both cases, except for a constant system bias caused by the hardware delay of different frequencies, which can be removed by the parameter estimation and prediction with long time datasets or by a priori calibration. PMID:29596330
Learning to learn causal models.
Kemp, Charles; Goodman, Noah D; Tenenbaum, Joshua B
2010-09-01
Learning to understand a single causal system can be an achievement, but humans must learn about multiple causal systems over the course of a lifetime. We present a hierarchical Bayesian framework that helps to explain how learning about several causal systems can accelerate learning about systems that are subsequently encountered. Given experience with a set of objects, our framework learns a causal model for each object and a causal schema that captures commonalities among these causal models. The schema organizes the objects into categories and specifies the causal powers and characteristic features of these categories and the characteristic causal interactions between categories. A schema of this kind allows causal models for subsequent objects to be rapidly learned, and we explore this accelerated learning in four experiments. Our results confirm that humans learn rapidly about the causal powers of novel objects, and we show that our framework accounts better for our data than alternative models of causal learning. Copyright © 2010 Cognitive Science Society, Inc.
An Analytic Model for the Success Rate of a Robotic Actuator System in Hitting Random Targets.
Bradley, Stuart
2015-11-20
Autonomous robotic systems are increasingly being used in a wide range of applications such as precision agriculture, medicine, and the military. These systems have common features which often includes an action by an "actuator" interacting with a target. While simulations and measurements exist for the success rate of hitting targets by some systems, there is a dearth of analytic models which can give insight into, and guidance on optimization, of new robotic systems. The present paper develops a simple model for estimation of the success rate for hitting random targets from a moving platform. The model has two main dimensionless parameters: the ratio of actuator spacing to target diameter; and the ratio of platform distance moved (between actuator "firings") to the target diameter. It is found that regions of parameter space having specified high success are described by simple equations, providing guidance on design. The role of a "cost function" is introduced which, when minimized, provides optimization of design, operating, and risk mitigation costs.
Modeling cutinase enzyme regulation in polyethylene terepthalate plastic biodegradation
NASA Astrophysics Data System (ADS)
Apri, M.; Silmi, M.; Heryanto, T. E.; Moeis, M. R.
2016-04-01
PET (Polyethylene terephthalate) is a plastic material that is commonly used in our daily life. The high production of PET and others plastics that can be up to three hundred million tons per year, is not matched by its degradation rate and hence leads to environmental pollution. To overcome this problem, we develop a biodegradation system. This system utilizes LC Cutinase enzyme produced by engineered escherichia coli bacteria to degrade PET. To make the system works efficaciously, it is important to understand the mechanism underlying its enzyme regulation. Therefore, we construct a mathematical model to describe the regulation of LC Cutinase production. The stability of the model is analyzed. We show that the designated biodegradation system can give an oscillatory behavior that is very important to control the amount of inclusion body (the miss-folded proteins that reduce the efficiency of the biodegradation system).
Modeling cutinase enzyme regulation in polyethylene terepthalate plastic biodegradation
DOE Office of Scientific and Technical Information (OSTI.GOV)
Apri, M., E-mail: m.apri@math.itb.ac.id; Silmi, M.; Heryanto, T. E.
PET (Polyethylene terephthalate) is a plastic material that is commonly used in our daily life. The high production of PET and others plastics that can be up to three hundred million tons per year, is not matched by its degradation rate and hence leads to environmental pollution. To overcome this problem, we develop a biodegradation system. This system utilizes LC Cutinase enzyme produced by engineered escherichia coli bacteria to degrade PET. To make the system works efficaciously, it is important to understand the mechanism underlying its enzyme regulation. Therefore, we construct a mathematical model to describe the regulation of LCmore » Cutinase production. The stability of the model is analyzed. We show that the designated biodegradation system can give an oscillatory behavior that is very important to control the amount of inclusion body (the miss-folded proteins that reduce the efficiency of the biodegradation system).« less
DOE Office of Scientific and Technical Information (OSTI.GOV)
Jones, Andew; Di Vittorio, Alan; Collins, William
The integrated Earth system model (iESM) has been developed as a new tool for projecting the joint human/climate system. The iESM is based upon coupling an integrated assessment model (IAM) and an Earth system model (ESM) into a common modeling infrastructure. IAMs are the primary tool for describing the human-Earth system, including the sources of global greenhouse gases (GHGs) and short-lived species (SLS), land use and land cover change (LULCC), and other resource-related drivers of anthropogenic climate change. ESMs are the primary scientific tools for examining the physical, chemical, and biogeochemical impacts of human-induced changes to the climate system. Themore » iESM project integrates the economic and human-dimension modeling of an IAM and a fully coupled ESM within a single simulation system while maintaining the separability of each model if needed. Both IAM and ESM codes are developed and used by large communities and have been extensively applied in recent national and international climate assessments. By introducing heretofore-omitted feedbacks between natural and societal drivers, we can improve scientific understanding of the human-Earth system dynamics. Potential applications include studies of the interactions and feedbacks leading to the timing, scale, and geographic distribution of emissions trajectories and other human influences, corresponding climate effects, and the subsequent impacts of a changing climate on human and natural systems.« less
Lifelong Learning and Employability: Is the European Model of Vocational Training in Crisis?
ERIC Educational Resources Information Center
Heidemann, Winfried
This paper explores the traditional European model of vocational training in light of a new focus on employability and lifelong learning that is becoming more common in Europe. It includes the following four sections: (1) an overview of some examples of vocational training systems in Europe and the proposal that they share enough to be considered…
Assessing Resilience in the Global Undersea Cable Infrastructure
2012-06-01
ABBREVIATIONS ACMA Australian Communications and Media Authority AD Attacker-Defender FSSCC Financial Services Sector Coordinating Council...after a disruption to the value delivery of the system before the disruption. Finally, their article also highlights the critical importance of...Chang et al. (2006), gravity models take their name from Newton’s law of gravitation, and are commonly used by social scientists to model or
FPGA-based firmware model for extended measurement systems with data quality monitoring
NASA Astrophysics Data System (ADS)
Wojenski, A.; Pozniak, K. T.; Mazon, D.; Chernyshova, M.
2017-08-01
Modern physics experiments requires construction of advanced, modular measurement systems for data processing and registration purposes. Components are often designed in one of the common mechanical and electrical standards, e.g. VME or uTCA. The paper is focused on measurement systems using FPGAs as data processing blocks, especially for plasma diagnostics using GEM detectors with data quality monitoring aspects. In the article is proposed standardized model of HDL FPGA firmware implementation, for use in a wide range of different measurement system. The effort was made in term of flexible implementation of data quality monitoring along with source data dynamic selection. In the paper is discussed standard measurement system model followed by detailed model of FPGA firmware for modular measurement systems. Considered are both: functional blocks and data buses. In the summary, necessary blocks and signal lines are described. Implementation of firmware following the presented rules should provide modular design, with ease of change different parts of it. The key benefit is construction of universal, modular HDL design, that can be applied in different measurement system with simple adjustments.
Increased sediment oxygen flux in lakes and reservoirs: The impact of hypolimnetic oxygenation
NASA Astrophysics Data System (ADS)
Bierlein, Kevin A.; Rezvani, Maryam; Socolofsky, Scott A.; Bryant, Lee D.; Wüest, Alfred; Little, John C.
2017-06-01
Hypolimnetic oxygenation is an increasingly common lake management strategy for mitigating hypoxia/anoxia and associated deleterious effects on water quality. A common effect of oxygenation is increased oxygen consumption in the hypolimnion and predicting the magnitude of this increase is the crux of effective oxygenation system design. Simultaneous measurements of sediment oxygen flux (JO2) and turbulence in the bottom boundary layer of two oxygenated lakes were used to investigate the impact of oxygenation on JO2. Oxygenation increased JO2 in both lakes by increasing the bulk oxygen concentration, which in turn steepens the diffusive gradient across the diffusive boundary layer. At high flow rates, the diffusive boundary layer thickness decreased as well. A transect along one of the lakes showed JO2 to be spatially quite variable, with near-field and far-field JO2 differing by a factor of 4. Using these in situ measurements, physical models of interfacial flux were compared to microprofile-derived JO2 to determine which models adequately predict JO2 in oxygenated lakes. Models based on friction velocity, turbulence dissipation rate, and the integral scale of turbulence agreed with microprofile-derived JO2 in both lakes. These models could potentially be used to predict oxygenation-induced oxygen flux and improve oxygenation system design methods for a broad range of reservoir systems.
Portal of medical data models: information infrastructure for medical research and healthcare.
Dugas, Martin; Neuhaus, Philipp; Meidt, Alexandra; Doods, Justin; Storck, Michael; Bruland, Philipp; Varghese, Julian
2016-01-01
Information systems are a key success factor for medical research and healthcare. Currently, most of these systems apply heterogeneous and proprietary data models, which impede data exchange and integrated data analysis for scientific purposes. Due to the complexity of medical terminology, the overall number of medical data models is very high. At present, the vast majority of these models are not available to the scientific community. The objective of the Portal of Medical Data Models (MDM, https://medical-data-models.org) is to foster sharing of medical data models. MDM is a registered European information infrastructure. It provides a multilingual platform for exchange and discussion of data models in medicine, both for medical research and healthcare. The system is developed in collaboration with the University Library of Münster to ensure sustainability. A web front-end enables users to search, view, download and discuss data models. Eleven different export formats are available (ODM, PDF, CDA, CSV, MACRO-XML, REDCap, SQL, SPSS, ADL, R, XLSX). MDM contents were analysed with descriptive statistics. MDM contains 4387 current versions of data models (in total 10,963 versions). 2475 of these models belong to oncology trials. The most common keyword (n = 3826) is 'Clinical Trial'; most frequent diseases are breast cancer, leukemia, lung and colorectal neoplasms. Most common languages of data elements are English (n = 328,557) and German (n = 68,738). Semantic annotations (UMLS codes) are available for 108,412 data items, 2453 item groups and 35,361 code list items. Overall 335,087 UMLS codes are assigned with 21,847 unique codes. Few UMLS codes are used several thousand times, but there is a long tail of rarely used codes in the frequency distribution. Expected benefits of the MDM portal are improved and accelerated design of medical data models by sharing best practice, more standardised data models with semantic annotation and better information exchange between information systems, in particular Electronic Data Capture (EDC) and Electronic Health Records (EHR) systems. Contents of the MDM portal need to be further expanded to reach broad coverage of all relevant medical domains. Database URL: https://medical-data-models.org. © The Author(s) 2016. Published by Oxford University Press.
Smith, Andrew B; López-Villarejo, Juan; Diago-Navarro, Elizabeth; Mitchenall, Lesley A; Barendregt, Arjan; Heck, Albert J; Lemonnier, Marc; Maxwell, Anthony; Díaz-Orejas, Ramón
2012-01-01
Bacterial toxin-antitoxin (TA) systems encode two proteins, a potent inhibitor of cell proliferation (toxin) and its specific antidote (antitoxin). Structural data has revealed striking similarities between the two model TA toxins CcdB, a DNA gyrase inhibitor encoded by the ccd system of plasmid F, and Kid, a site-specific endoribonuclease encoded by the parD system of plasmid R1. While a common structural fold seemed at odds with the two clearly different modes of action of these toxins, the possibility of functional crosstalk between the parD and ccd systems, which would further point to their common evolutionary origin, has not been documented. Here, we show that the cleavage of RNA and the inhibition of protein synthesis by the Kid toxin, two activities that are specifically counteracted by its cognate Kis antitoxin, are altered, but not inhibited, by the CcdA antitoxin. In addition, Kis was able to inhibit the stimulation of DNA gyrase-mediated cleavage of DNA by CcdB, albeit less efficiently than CcdA. We further show that physical interactions between the toxins and antitoxins of the different systems do occur and define the stoichiometry of the complexes formed. We found that CcdB did not degrade RNA nor did Kid have any reproducible effect on the tested DNA gyrase activities, suggesting that these toxins evolved to reach different, rather than common, cellular targets.
Mitchenall, Lesley A.; Barendregt, Arjan; Heck, Albert J.; Lemonnier, Marc; Maxwell, Anthony; Díaz-Orejas, Ramón
2012-01-01
Bacterial toxin-antitoxin (TA) systems encode two proteins, a potent inhibitor of cell proliferation (toxin) and its specific antidote (antitoxin). Structural data has revealed striking similarities between the two model TA toxins CcdB, a DNA gyrase inhibitor encoded by the ccd system of plasmid F, and Kid, a site-specific endoribonuclease encoded by the parD system of plasmid R1. While a common structural fold seemed at odds with the two clearly different modes of action of these toxins, the possibility of functional crosstalk between the parD and ccd systems, which would further point to their common evolutionary origin, has not been documented. Here, we show that the cleavage of RNA and the inhibition of protein synthesis by the Kid toxin, two activities that are specifically counteracted by its cognate Kis antitoxin, are altered, but not inhibited, by the CcdA antitoxin. In addition, Kis was able to inhibit the stimulation of DNA gyrase-mediated cleavage of DNA by CcdB, albeit less efficiently than CcdA. We further show that physical interactions between the toxins and antitoxins of the different systems do occur and define the stoichiometry of the complexes formed. We found that CcdB did not degrade RNA nor did Kid have any reproducible effect on the tested DNA gyrase activities, suggesting that these toxins evolved to reach different, rather than common, cellular targets. PMID:23029540
Common sense reasoning about petroleum flow
DOE Office of Scientific and Technical Information (OSTI.GOV)
Rosenberg, S.
1981-02-01
This paper describes an expert system for understanding and Reasoning in a petroleum resources domain. A basic model is implemented in FRL (Frame Representation Language). Expertise is encoded as rule frames. The model consists of a set of episodic contexts which are sequentially generated over time. Reasoning occurs in separate reasoning contexts consisting of a buffer frame and packets of rules. These function similar to small production systems. reasoning is linked to the model through an interface of Sentinels (instance driven demons) which notice anomalous conditions. Heuristics and metaknowledge are used through the creation of further reasoning contexts which overlaymore » the simpler ones.« less
A risk evaluation model and its application in online retailing trustfulness
NASA Astrophysics Data System (ADS)
Ye, Ruyi; Xu, Yingcheng
2017-08-01
Building a general model for risks evaluation in advance could improve the convenience, normality and comparability of the results of repeating risks evaluation in the case that the repeating risks evaluating are in the same area and for a similar purpose. One of the most convenient and common risks evaluation models is an index system including of several index, according weights and crediting method. One method to build a risk evaluation index system that guarantees the proportional relationship between the resulting credit and the expected risk loss is proposed and an application example is provided in online retailing in this article.
Linear Power-Flow Models in Multiphase Distribution Networks: Preprint
DOE Office of Scientific and Technical Information (OSTI.GOV)
Bernstein, Andrey; Dall'Anese, Emiliano
This paper considers multiphase unbalanced distribution systems and develops approximate power-flow models where bus-voltages, line-currents, and powers at the point of common coupling are linearly related to the nodal net power injections. The linearization approach is grounded on a fixed-point interpretation of the AC power-flow equations, and it is applicable to distribution systems featuring (i) wye connections; (ii) ungrounded delta connections; (iii) a combination of wye-connected and delta-connected sources/loads; and, (iv) a combination of line-to-line and line-to-grounded-neutral devices at the secondary of distribution transformers. The proposed linear models can facilitate the development of computationally-affordable optimization and control applications -- frommore » advanced distribution management systems settings to online and distributed optimization routines. Performance of the proposed models is evaluated on different test feeders.« less
A Comparison and Evaluation of Real-Time Software Systems Modeling Languages
NASA Technical Reports Server (NTRS)
Evensen, Kenneth D.; Weiss, Kathryn Anne
2010-01-01
A model-driven approach to real-time software systems development enables the conceptualization of software, fostering a more thorough understanding of its often complex architecture and behavior while promoting the documentation and analysis of concerns common to real-time embedded systems such as scheduling, resource allocation, and performance. Several modeling languages have been developed to assist in the model-driven software engineering effort for real-time systems, and these languages are beginning to gain traction with practitioners throughout the aerospace industry. This paper presents a survey of several real-time software system modeling languages, namely the Architectural Analysis and Design Language (AADL), the Unified Modeling Language (UML), Systems Modeling Language (SysML), the Modeling and Analysis of Real-Time Embedded Systems (MARTE) UML profile, and the AADL for UML profile. Each language has its advantages and disadvantages, and in order to adequately describe a real-time software system's architecture, a complementary use of multiple languages is almost certainly necessary. This paper aims to explore these languages in the context of understanding the value each brings to the model-driven software engineering effort and to determine if it is feasible and practical to combine aspects of the various modeling languages to achieve more complete coverage in architectural descriptions. To this end, each language is evaluated with respect to a set of criteria such as scope, formalisms, and architectural coverage. An example is used to help illustrate the capabilities of the various languages.
Hospital information system: reusability, designing, modelling, recommendations for implementing.
Huet, B
1998-01-01
The aims of this paper are to precise some essential conditions for building reuse models for hospital information systems (HIS) and to present an application for hospital clinical laboratories. Reusability is a general trend in software, however reuse can involve a more or less part of design, classes, programs; consequently, a project involving reusability must be precisely defined. In the introduction it is seen trends in software, the stakes of reuse models for HIS and the special use case constituted with a HIS. The main three parts of this paper are: 1) Designing a reuse model (which objects are common to several information systems?) 2) A reuse model for hospital clinical laboratories (a genspec object model is presented for all laboratories: biochemistry, bacteriology, parasitology, pharmacology, ...) 3) Recommendations for generating plug-compatible software components (a reuse model can be implemented as a framework, concrete factors that increase reusability are presented). In conclusion reusability is a subtle exercise of which project must be previously and carefully defined.
Designing a model for trauma system management using public health approach: the case of Iran.
Tarighi, Payam; Tabibi, Seyed Jamaledin; Motevalian, Seyed Abbas; Tofighi, Shahram; Maleki, Mohammad Reza; Delgoshaei, Bahram; Panahi, Farzad; Masoomi, Gholam Reza
2012-01-01
Trauma is a leading cause of death and disability around the world. Injuries are responsible for about six million deaths annually, of which ninety percent occur in developing countries. In Iran, injuries are the most common cause of death among age groups below fifty. Trauma system development is a systematic and comprehensive approach to injury prevention and treatment whose effectiveness has been proved. The present study aims at designing a trauma system management model as the first step toward trauma system establishment in Iran. In this qualitative research, a conceptual framework was developed based on the public health approach and three well-known trauma system models. We used Benchmarks, Indicators and Scoring (BIS) to analyze the current situation of Iran trauma care system. Then the trauma system management was designed using the policy development phase of public health approach The trauma system management model, validated by a panel of experts, describes lead agency, trauma system plan, policy-making councils, and data-based control according to the four main functions of management: leading, planning, organizing and controlling. This model may be implemented in two phases: the exclusive phase, focusing on resource integration and the inclusive phase, which concentrates on system development. The model could facilitate the development of trauma system in Iran through pilot studies as the assurance phase of public health approach. Furthermore, the model can provide a practical framework for trauma system management at the international level.
Reuseable Objects Software Environment (ROSE): Introduction to Air Force Software Reuse Workshop
NASA Technical Reports Server (NTRS)
Cottrell, William L.
1994-01-01
The Reusable Objects Software Environment (ROSE) is a common, consistent, consolidated implementation of software functionality using modern object oriented software engineering including designed-in reuse and adaptable requirements. ROSE is designed to minimize abstraction and reduce complexity. A planning model for the reverse engineering of selected objects through object oriented analysis is depicted. Dynamic and functional modeling are used to develop a system design, the object design, the language, and a database management system. The return on investment for a ROSE pilot program and timelines are charted.
ELISA, a demonstrator environment for information systems architecture design
NASA Technical Reports Server (NTRS)
Panem, Chantal
1994-01-01
This paper describes an approach of reusability of software engineering technology in the area of ground space system design. System engineers have lots of needs similar to software developers: sharing of a common data base, capitalization of knowledge, definition of a common design process, communication between different technical domains. Moreover system designers need to simulate dynamically their system as early as possible. Software development environments, methods and tools now become operational and widely used. Their architecture is based on a unique object base, a set of common management services and they host a family of tools for each life cycle activity. In late '92, CNES decided to develop a demonstrative software environment supporting some system activities. The design of ground space data processing systems was chosen as the application domain. ELISA (Integrated Software Environment for Architectures Specification) was specified as a 'demonstrator', i.e. a sufficient basis for demonstrations, evaluation and future operational enhancements. A process with three phases was implemented: system requirements definition, design of system architectures models, and selection of physical architectures. Each phase is composed of several activities that can be performed in parallel, with the provision of Commercial Off the Shelves Tools. ELISA has been delivered to CNES in January 94, currently used for demonstrations and evaluations on real projects (e.g. SPOT4 Satellite Control Center). It is on the way of new evolutions.
High Lift Common Research Model for Wind Tunnel Testing: An Active Flow Control Perspective
NASA Technical Reports Server (NTRS)
Lin, John C.; Melton, Latunia P.; Viken, Sally A.; Andino, Marlyn Y.; Koklu, Mehti; Hannon, Judith A.; Vatsa, Veer N.
2017-01-01
This paper provides an overview of a research and development effort sponsored by the NASA Advanced Air Transport Technology Project to achieve the required high-lift performance using active flow control (AFC) on simple hinged flaps while reducing the cruise drag associated with the external mechanisms on slotted flaps of a generic modern transport aircraft. The removal of the external fairings for the Fowler flap mechanism could help to reduce drag by 3.3 counts. The main challenge is to develop an AFC system that can provide the necessary lift recovery on a simple hinged flap high-lift system while using the limited pneumatic power available on the aircraft. Innovative low-power AFC concepts will be investigated in the flap shoulder region. The AFC concepts being explored include steady blowing and unsteady blowing operating in the spatial and/or temporal domain. Both conventional and AFC-enabled high-lift configurations were designed for the current effort. The high-lift configurations share the cruise geometry that is based on the NASA Common Research Model, and therefore, are also open geometries. A 10%-scale High Lift Common Research Model (HL-CRM) is being designed for testing at the NASA Langley Research Center 14- by 22-Foot Subsonic Tunnel during fiscal year 2018. The overall project plan, status, HL-CRM configurations, and AFC objectives for the wind tunnel test are described.
The analysis of magnesium oxide hydration in three-phase reaction system
DOE Office of Scientific and Technical Information (OSTI.GOV)
Tang, Xiaojia; Guo, Lin; Chen, Chen
In order to investigate the magnesium oxide hydration process in gas–liquid–solid (three-phase) reaction system, magnesium hydroxide was prepared by magnesium oxide hydration in liquid–solid (two-phase) and three-phase reaction systems. A semi-empirical model and the classical shrinking core model were used to fit the experimental data. The fitting result shows that both models describe well the hydration process of three-phase system, while only the semi-empirical model right for the hydration process of two-phase system. The characterization of the hydration product using X-Ray diffraction (XRD) and scanning electron microscope (SEM) was performed. The XRD and SEM show hydration process in the two-phasemore » system follows common dissolution/precipitation mechanism. While in the three-phase system, the hydration process undergo MgO dissolution, Mg(OH){sub 2} precipitation, Mg(OH){sub 2} peeling off from MgO particle and leaving behind fresh MgO surface. - Graphical abstract: There was existence of a peeling-off process in the gas–liquid–solid (three-phase) MgO hydration system. - Highlights: • Magnesium oxide hydration in gas–liquid–solid system was investigated. • The experimental data in three-phase system could be fitted well by two models. • The morphology analysis suggested that there was existence of a peel-off process.« less
Simulation of shoreline development in a groyne system, with a case study Sanur Bali beach
NASA Astrophysics Data System (ADS)
Gunawan, P. H.; Pudjaprasetya, S. R.
2018-03-01
The process of shoreline changes due to transport of sediment by littoral drift is studied in this paper. Pelnard-Considère is the commonly adopted model. This model is based on the principle of sediment conservation, without diffraction. In this research, we adopt the Pelnard-Considère equation with diffraction, and a numerical scheme based on the finite volume method is implemented. Shoreline development in a groyne system is then simulated. For a case study, the Sanur Bali Beach, Indonesia is considered, in which from Google Earth photos, the beach experiences changes of coastline caused by sediment trapped in a groyne system.
A Model Independent S/W Framework for Search-Based Software Testing
Baik, Jongmoon
2014-01-01
In Model-Based Testing (MBT) area, Search-Based Software Testing (SBST) has been employed to generate test cases from the model of a system under test. However, many types of models have been used in MBT. If the type of a model has changed from one to another, all functions of a search technique must be reimplemented because the types of models are different even if the same search technique has been applied. It requires too much time and effort to implement the same algorithm over and over again. We propose a model-independent software framework for SBST, which can reduce redundant works. The framework provides a reusable common software platform to reduce time and effort. The software framework not only presents design patterns to find test cases for a target model but also reduces development time by using common functions provided in the framework. We show the effectiveness and efficiency of the proposed framework with two case studies. The framework improves the productivity by about 50% when changing the type of a model. PMID:25302314
UNH Data Cooperative: A Cyber Infrastructure for Earth System Studies
NASA Astrophysics Data System (ADS)
Braswell, B. H.; Fekete, B. M.; Prusevich, A.; Gliden, S.; Magill, A.; Vorosmarty, C. J.
2007-12-01
Earth system scientists and managers have a continuously growing demand for a wide array of earth observations derived from various data sources including (a) modern satellite retrievals, (b) "in-situ" records, (c) various simulation outputs, and (d) assimilated data products combining model results with observational records. The sheer quantity of data, and formatting inconsistencies make it difficult for users to take full advantage of this important information resource. Thus the system could benefit from a thorough retooling of our current data processing procedures and infrastructure. Emerging technologies, like OPeNDAP and OGC map services, open standard data formats (NetCDF, HDF) data cataloging systems (NASA-Echo, Global Change Master Directory, etc.) are providing the basis for a new approach in data management and processing, where web- services are increasingly designed to serve computer-to-computer communications without human interactions and complex analysis can be carried out over distributed computer resources interconnected via cyber infrastructure. The UNH Earth System Data Collaborative is designed to utilize the aforementioned emerging web technologies to offer new means of access to earth system data. While the UNH Data Collaborative serves a wide array of data ranging from weather station data (Climate Portal) to ocean buoy records and ship tracks (Portsmouth Harbor Initiative) to land cover characteristics, etc. the underlaying data architecture shares common components for data mining and data dissemination via web-services. Perhaps the most unique element of the UNH Data Cooperative's IT infrastructure is its prototype modeling environment for regional ecosystem surveillance over the Northeast corridor, which allows the integration of complex earth system model components with the Cooperative's data services. While the complexity of the IT infrastructure to perform complex computations is continuously increasing, scientists are often forced to spend considerable amount of time to solve basic data management and preprocessing tasks and deal with low level computational design problems like parallelization of model codes. Our modeling infrastructure is designed to take care the bulk of the common tasks found in complex earth system models like I/O handling, computational domain and time management, parallel execution of the modeling tasks, etc. The modeling infrastructure allows scientists to focus on the numerical implementation of the physical processes on a single computational objects(typically grid cells) while the framework takes care of the preprocessing of input data, establishing of the data exchange between computation objects and the execution of the science code. In our presentation, we will discuss the key concepts of our modeling infrastructure. We will demonstrate integration of our modeling framework with data services offered by the UNH Earth System Data Collaborative via web interfaces. We will layout the road map to turn our prototype modeling environment into a truly community framework for wide range of earth system scientists and environmental managers.
The Mechanisms of Water Exchange: The Regulatory Roles of Multiple Interactions in Social Wasps.
Agrawal, Devanshu; Karsai, Istvan
2016-01-01
Evolutionary benefits of task fidelity and improving information acquisition via multiple transfers of materials between individuals in a task partitioned system have been shown before, but in this paper we provide a mechanistic explanation of these phenomena. Using a simple mathematical model describing the individual interactions of the wasps, we explain the functioning of the common stomach, an information center, which governs construction behavior and task change. Our central hypothesis is a symmetry between foragers who deposit water and foragers who withdraw water into and out of the common stomach. We combine this with a trade-off between acceptance and resistance to water transfer. We ultimately derive a mathematical function that relates the number of interactions that foragers complete with common stomach wasps during a foraging cycle. We use field data and additional model assumptions to calculate values of our model parameters, and we use these to explain why the fullness of the common stomach stabilizes just below 50 percent, why the average number of successful interactions between foragers and the wasps forming the common stomach is between 5 and 7, and why there is a variation in this number of interactions over time. Our explanation is that our proposed water exchange mechanism places natural bounds on the number of successful interactions possible, water exchange is set to optimize mediation of water through the common stomach, and the chance that foragers abort their task prematurely is very low.
The Mechanisms of Water Exchange: The Regulatory Roles of Multiple Interactions in Social Wasps
Agrawal, Devanshu; Karsai, Istvan
2016-01-01
Evolutionary benefits of task fidelity and improving information acquisition via multiple transfers of materials between individuals in a task partitioned system have been shown before, but in this paper we provide a mechanistic explanation of these phenomena. Using a simple mathematical model describing the individual interactions of the wasps, we explain the functioning of the common stomach, an information center, which governs construction behavior and task change. Our central hypothesis is a symmetry between foragers who deposit water and foragers who withdraw water into and out of the common stomach. We combine this with a trade-off between acceptance and resistance to water transfer. We ultimately derive a mathematical function that relates the number of interactions that foragers complete with common stomach wasps during a foraging cycle. We use field data and additional model assumptions to calculate values of our model parameters, and we use these to explain why the fullness of the common stomach stabilizes just below 50 percent, why the average number of successful interactions between foragers and the wasps forming the common stomach is between 5 and 7, and why there is a variation in this number of interactions over time. Our explanation is that our proposed water exchange mechanism places natural bounds on the number of successful interactions possible, water exchange is set to optimize mediation of water through the common stomach, and the chance that foragers abort their task prematurely is very low. PMID:26751076
Building a Common Pediatric Research Terminology for Accelerating Child Health Research
Bailey, L. Charles; Forrest, Christopher B.; Padula, Michael A.; Hirschfeld, Steven
2014-01-01
Longitudinal observational clinical data on pediatric patients in electronic format is becoming widely available. A new era of multi-institutional data networks that study pediatric diseases and outcomes across disparate health delivery models and care settings are also enabling an innovative collaborative rapid improvement paradigm called the Learning Health System. However, the potential alignment of routine clinical care, observational clinical research, pragmatic clinical trials, and health systems improvement requires a data infrastructure capable of combining information from systems and workflows that historically have been isolated from each other. Removing barriers to integrating and reusing data collected in different settings will permit new opportunities to develop a more complete picture of a patient’s care and to leverage data from related research studies. One key barrier is the lack of a common terminology that provides uniform definitions and descriptions of clinical observations and data. A well-characterized terminology ensures a common meaning and supports data reuse and integration. A common terminology allows studies to build upon previous findings and to reuse data collection tools and data management processes. We present the current state of terminology harmonization and describe a governance structure and mechanism for coordinating the development of a common pediatric research terminology that links to clinical terminologies and can be used to align existing terminologies. By reducing the barriers between clinical care and clinical research, a Learning Health System can leverage and reuse not only its own data resources but also broader extant data resources. PMID:24534404
Zisner, Aimee; Beauchaine, Theodore P
2016-11-01
Trait impulsivity, which is often defined as a strong preference for immediate over delayed rewards and results in behaviors that are socially inappropriate, maladaptive, and short-sighted, is a predisposing vulnerability to all externalizing spectrum disorders. In contrast, anhedonia is characterized by chronically low motivation and reduced capacity to experience pleasure, and is common to depressive disorders. Although externalizing and depressive disorders have virtually nonoverlapping diagnostic criteria in the fifth edition of the Diagnostic and Statistical Manual of Mental Disorders, heterotypic comorbidity between them is common. Here, we review common neural substrates of trait impulsivity, anhedonia, and irritability, which include both low tonic mesolimbic dopamine activity and low phasic mesolimbic dopamine responding to incentives during reward anticipation and associative learning. We also consider how other neural networks, including bottom-up emotion generation systems and top-down emotion regulation systems, interact with mesolimbic dysfunction to result in alternative manifestations of psychiatric illness. Finally, we present a model that emphasizes a translational, transdiagnostic approach to understanding externalizing/depression comorbidity. This model should refine ways in which internalizing and externalizing disorders are studied, classified, and treated.
Murray, L K; Jordans, M J D
2016-07-12
Global mental health is a growing field intricately connected to broader health, violence and economic issues. Despite the high prevalence and cost of mental health disorders, an estimated 75 % of those with need in lower resource settings do not receive intervention. Most studies to date have examined the effectiveness of single-disorder mental health treatments - an approach that may be a significant challenge to scale-up and sustainability in lower resource settings. This paper presents a brief overview of the scientific progress in global mental health, and suggests consideration of an internal stepped care delivery approach. An internal stepped care model is one idea of a delivery system, utilizing a common elements approach, where the same provider could navigate between different elements based on severity and type of problems of the client. It is distinct from traditional stepped care models in that clients remain with the same provider, rather than relying on referral systems. An internal stepped care delivery system based on a simplified common elements approach could be more efficient, scalable, sustainable, and reduce the loss of clients to referrals in lower resource settings.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Knirsch, Fabian; Engel, Dominik; Neureiter, Christian
In a smart grid, data and information are transported, transmitted, stored, and processed with various stakeholders having to cooperate effectively. Furthermore, personal data is the key to many smart grid applications and therefore privacy impacts have to be taken into account. For an effective smart grid, well integrated solutions are crucial and for achieving a high degree of customer acceptance, privacy should already be considered at design time of the system. To assist system engineers in early design phase, frameworks for the automated privacy evaluation of use cases are important. For evaluation, use cases for services and software architectures needmore » to be formally captured in a standardized and commonly understood manner. In order to ensure this common understanding for all kinds of stakeholders, reference models have recently been developed. In this paper we present a model-driven approach for the automated assessment of such services and software architectures in the smart grid that builds on the standardized reference models. The focus of qualitative and quantitative evaluation is on privacy. For evaluation, the framework draws on use cases from the University of Southern California microgrid.« less
Study on the standard architecture for geoinformation common services
NASA Astrophysics Data System (ADS)
Zha, Z.; Zhang, L.; Wang, C.; Jiang, J.; Huang, W.
2014-04-01
The construction of platform for geoinformation common services was completed or on going in in most provinces and cities in these years in China, and the platforms plays an important role in the economic and social activities. Geoinfromation and geoinfromation based services are the key issues in the platform. The standards on geoinormation common services play as bridges among the users, systems and designers of the platform. The standard architecture for geoinformation common services is the guideline for designing and using the standard system in which the standards integrated to each other to promote the development, sharing and services of geoinformation resources. To establish the standard architecture for geoinformation common services is one of the tasks of "Study on important standards for geonformation common services and management of public facilities in city". The scope of the standard architecture is defined, such as data or information model, interoperability interface or service, information management. Some Research work on the status of international standards of geoinormation common services in organization and countries, like ISO/TC 211, OGC and other countries or unions like USA, EU, Japan have done. Some principles are set up to evaluate the standard, such as availability, suitability and extensible ability. Then the development requirement and practical situation are analyzed, and a framework of the standard architecture for geoinformation common services are proposed. Finally, a summary and prospects of the geoinformation standards are made.
Patch models and their applications to multivehicle command and control.
Rao, Venkatesh G; D'Andrea, Raffaello
2007-06-01
We introduce patch models, a computational modeling formalism for multivehicle combat domains, based on spatiotemporal abstraction methods developed in the computer science community. The framework yields models that are expressive enough to accommodate nontrivial controlled vehicle dynamics while being within the representational capabilities of common artificial intelligence techniques used in the construction of autonomous systems. The framework allows several key design requirements of next-generation network-centric command and control systems, such as maintenance of shared situation awareness, to be achieved. Major features include support for multiple situation models at each decision node and rapid mission plan adaptation. We describe the formal specification of patch models and our prototype implementation, i.e., Patchworks. The capabilities of patch models are validated through a combat mission simulation in Patchworks, which involves two defending teams protecting a camp from an enemy attacking team.
NASA Astrophysics Data System (ADS)
Yu, Y.; Jordanova, V. K.; McGranaghan, R. M.; Solomon, S. C.
2017-12-01
The ionospheric conductance, height-integrated electric conductivity, can regulate both the ionospheric electrodynamics and the magnetospheric dynamics because of its key role in determining the electric field within the coupled magnetosphere-ionosphere system. State-of-the-art global magnetosphere models commonly adopt empirical conductance calculators to obtain the auroral conductance. Such specification can bypass the complexity of the ionosphere-thermosphere chemistry but on the other hand breaks the self-consistent link within the coupled system. In this study, we couple a kinetic ring current model RAM-SCB-E that solves for anisotropic particle distributions with a two-stream electron transport code (GLOW) to more self-consistently compute the height-dependent electric conductivity, provided the auroral electron precipitation from the ring current model. Comparisons with the traditional empirical formula are carried out. It is found that the newly coupled modeling framework reveals smaller Hall and Pedersen conductance, resulting in a larger electric field. As a consequence, the subauroral polarization streams demonstrate a better agreement with observations from DMSP satellites. It is further found that the commonly assumed Maxwellian spectrum of the particle precipitation is not globally appropriate. Instead, a full precipitation spectrum resulted from wave particle interactions in the ring current accounts for a more comprehensive precipitation spectrum.
NASA Technical Reports Server (NTRS)
Cole, Bjorn; Chung, Seung
2012-01-01
One of the challenges of systems engineering is in working multidisciplinary problems in a cohesive manner. When planning analysis of these problems, system engineers must trade between time and cost for analysis quality and quantity. The quality often correlates with greater run time in multidisciplinary models and the quantity is associated with the number of alternatives that can be analyzed. The trade-off is due to the resource intensive process of creating a cohesive multidisciplinary systems model and analysis. Furthermore, reuse or extension of the models used in one stage of a product life cycle for another is a major challenge. Recent developments have enabled a much less resource-intensive and more rigorous approach than hand-written translation scripts between multi-disciplinary models and their analyses. The key is to work from a core systems model defined in a MOF-based language such as SysML and in leveraging the emerging tool ecosystem, such as Query/View/Transformation (QVT), from the OMG community. SysML was designed to model multidisciplinary systems. The QVT standard was designed to transform SysML models into other models, including those leveraged by engineering analyses. The Europa Habitability Mission (EHM) team has begun to exploit these capabilities. In one case, a Matlab/Simulink model is generated on the fly from a system description for power analysis written in SysML. In a more general case, symbolic analysis (supported by Wolfram Mathematica) is coordinated by data objects transformed from the systems model, enabling extremely flexible and powerful design exploration and analytical investigations of expected system performance.
Diagnostic layer integration in FPGA-based pipeline measurement systems for HEP experiments
NASA Astrophysics Data System (ADS)
Pozniak, Krzysztof T.
2007-08-01
Integrated triggering and data acquisition systems for high energy physics experiments may be considered as fast, multichannel, synchronous, distributed, pipeline measurement systems. A considerable extension of functional, technological and monitoring demands, which has recently been imposed on them, forced a common usage of large field-programmable gate array (FPGA), digital signal processing-enhanced matrices and fast optical transmission for their realization. This paper discusses modelling, design, realization and testing of pipeline measurement systems. A distribution of synchronous data stream flows is considered in the network. A general functional structure of a single network node is presented. A suggested, novel block structure of the node model facilitates full implementation in the FPGA chip, circuit standardization and parametrization, as well as integration of functional and diagnostic layers. A general method for pipeline system design was derived. This method is based on a unified model of the synchronous data network node. A few examples of practically realized, FPGA-based, pipeline measurement systems were presented. The described systems were applied in ZEUS and CMS.
Toward Failure Modeling In Complex Dynamic Systems: Impact of Design and Manufacturing Variations
NASA Technical Reports Server (NTRS)
Tumer, Irem Y.; McAdams, Daniel A.; Clancy, Daniel (Technical Monitor)
2001-01-01
When designing vehicle vibration monitoring systems for aerospace devices, it is common to use well-established models of vibration features to determine whether failures or defects exist. Most of the algorithms used for failure detection rely on these models to detect significant changes during a flight environment. In actual practice, however, most vehicle vibration monitoring systems are corrupted by high rates of false alarms and missed detections. Research conducted at the NASA Ames Research Center has determined that a major reason for the high rates of false alarms and missed detections is the numerous sources of statistical variations that are not taken into account in the. modeling assumptions. In this paper, we address one such source of variations, namely, those caused during the design and manufacturing of rotating machinery components that make up aerospace systems. We present a novel way of modeling the vibration response by including design variations via probabilistic methods. The results demonstrate initial feasibility of the method, showing great promise in developing a general methodology for designing more accurate aerospace vehicle vibration monitoring systems.
An improved predictive functional control method with application to PMSM systems
NASA Astrophysics Data System (ADS)
Li, Shihua; Liu, Huixian; Fu, Wenshu
2017-01-01
In common design of prediction model-based control method, usually disturbances are not considered in the prediction model as well as the control design. For the control systems with large amplitude or strong disturbances, it is difficult to precisely predict the future outputs according to the conventional prediction model, and thus the desired optimal closed-loop performance will be degraded to some extent. To this end, an improved predictive functional control (PFC) method is developed in this paper by embedding disturbance information into the system model. Here, a composite prediction model is thus obtained by embedding the estimated value of disturbances, where disturbance observer (DOB) is employed to estimate the lumped disturbances. So the influence of disturbances on system is taken into account in optimisation procedure. Finally, considering the speed control problem for permanent magnet synchronous motor (PMSM) servo system, a control scheme based on the improved PFC method is designed to ensure an optimal closed-loop performance even in the presence of disturbances. Simulation and experimental results based on a hardware platform are provided to confirm the effectiveness of the proposed algorithm.
Computational Systems Biology in Cancer: Modeling Methods and Applications
Materi, Wayne; Wishart, David S.
2007-01-01
In recent years it has become clear that carcinogenesis is a complex process, both at the molecular and cellular levels. Understanding the origins, growth and spread of cancer, therefore requires an integrated or system-wide approach. Computational systems biology is an emerging sub-discipline in systems biology that utilizes the wealth of data from genomic, proteomic and metabolomic studies to build computer simulations of intra and intercellular processes. Several useful descriptive and predictive models of the origin, growth and spread of cancers have been developed in an effort to better understand the disease and potential therapeutic approaches. In this review we describe and assess the practical and theoretical underpinnings of commonly-used modeling approaches, including ordinary and partial differential equations, petri nets, cellular automata, agent based models and hybrid systems. A number of computer-based formalisms have been implemented to improve the accessibility of the various approaches to researchers whose primary interest lies outside of model development. We discuss several of these and describe how they have led to novel insights into tumor genesis, growth, apoptosis, vascularization and therapy. PMID:19936081
MONALISA for stochastic simulations of Petri net models of biochemical systems.
Balazki, Pavel; Lindauer, Klaus; Einloft, Jens; Ackermann, Jörg; Koch, Ina
2015-07-10
The concept of Petri nets (PN) is widely used in systems biology and allows modeling of complex biochemical systems like metabolic systems, signal transduction pathways, and gene expression networks. In particular, PN allows the topological analysis based on structural properties, which is important and useful when quantitative (kinetic) data are incomplete or unknown. Knowing the kinetic parameters, the simulation of time evolution of such models can help to study the dynamic behavior of the underlying system. If the number of involved entities (molecules) is low, a stochastic simulation should be preferred against the classical deterministic approach of solving ordinary differential equations. The Stochastic Simulation Algorithm (SSA) is a common method for such simulations. The combination of the qualitative and semi-quantitative PN modeling and stochastic analysis techniques provides a valuable approach in the field of systems biology. Here, we describe the implementation of stochastic analysis in a PN environment. We extended MONALISA - an open-source software for creation, visualization and analysis of PN - by several stochastic simulation methods. The simulation module offers four simulation modes, among them the stochastic mode with constant firing rates and Gillespie's algorithm as exact and approximate versions. The simulator is operated by a user-friendly graphical interface and accepts input data such as concentrations and reaction rate constants that are common parameters in the biological context. The key features of the simulation module are visualization of simulation, interactive plotting, export of results into a text file, mathematical expressions for describing simulation parameters, and up to 500 parallel simulations of the same parameter sets. To illustrate the method we discuss a model for insulin receptor recycling as case study. We present a software that combines the modeling power of Petri nets with stochastic simulation of dynamic processes in a user-friendly environment supported by an intuitive graphical interface. The program offers a valuable alternative to modeling, using ordinary differential equations, especially when simulating single-cell experiments with low molecule counts. The ability to use mathematical expressions provides an additional flexibility in describing the simulation parameters. The open-source distribution allows further extensions by third-party developers. The software is cross-platform and is licensed under the Artistic License 2.0.
NASA Technical Reports Server (NTRS)
Torres-Pomales, Wilfredo
2014-01-01
A system is safety-critical if its failure can endanger human life or cause significant damage to property or the environment. State-of-the-art computer systems on commercial aircraft are highly complex, software-intensive, functionally integrated, and network-centric systems of systems. Ensuring that such systems are safe and comply with existing safety regulations is costly and time-consuming as the level of rigor in the development process, especially the validation and verification activities, is determined by considerations of system complexity and safety criticality. A significant degree of care and deep insight into the operational principles of these systems is required to ensure adequate coverage of all design implications relevant to system safety. Model-based development methodologies, methods, tools, and techniques facilitate collaboration and enable the use of common design artifacts among groups dealing with different aspects of the development of a system. This paper examines the application of model-based development to complex and safety-critical aircraft computer systems. Benefits and detriments are identified and an overall assessment of the approach is given.
Forum: The challenge of global change
NASA Astrophysics Data System (ADS)
Roederer, Juan G.
1990-09-01
How can we sustain a public sense of the common danger of global change while remaining honest in view of the realities of scientific uncertainty? How can we nurture this sense of common danger without making statements based on half-baked ideas, statistically unreliable results, or oversimplified models? How can we strike a balance between the need to overstate a case to attract the attention of the media and the obligation to adhere strictly to the ethos of science?The task of achieving a scientific understanding of the inner workings of the terrestrial environment is one of the most difficult and ambitious endeavors of humankind. It is full of traps, temptations and deceptions for the participating scientists. We are dealing with a horrendously complex, strongly interactive, highly non-linear system. Lessons learned from disciplines such as plasma physics and solid state physics which have been dealing with complex non-linear systems for decades, are not very encouraging. The first thing one learns is that there are intrinsic, physical limits to the quantitative predictability of a complex system that have nothing to do with the particular techniques employed to model it.
Porter-Thomas distribution in unstable many-body systems
DOE Office of Scientific and Technical Information (OSTI.GOV)
Volya, Alexander
We use the continuum shell model approach to explore the resonance width distribution in unstable many-body systems. The single-particle nature of a decay, the few-body character of the interaction Hamiltonian, and the collectivity that emerges in nonstationary systems due to the coupling to the continuum of reaction states are discussed. Correlations between the structures of the parent and daughter nuclear systems in the common Fock space are found to result in deviations of decay width statistics from the Porter-Thomas distribution.
2015-12-01
the MIS System/Subsystem Specification ( SSS ), and supplementary BAA document. On June 26, 2014, the SEI provided a draft interim report of the...findings and issues. The SEI team also received July 3, 2014, versions of the MIS Stakeholder Requirements, MIS SSS , and build plan and July 17, 2014...versions of the MIS SSS together with the MIS system model. On July 14–15, 2014, the SEI presented a summary of the issues at the two contractors
2015-06-01
very coarse architectural model proposed in Section 2.4 into something that might be implemented . Figure 11 shows the model we have created based ...interoperability through common data models . So many of the pieces are either in place or are being developed currently. However, SEA still needs: • A core...of knowledge derived through the scientific method. In NATO, S&T is addressed using different business models , namely a collaborative business model
Integrating Multiscale Modeling with Drug Effects for Cancer Treatment.
Li, Xiangfang L; Oduola, Wasiu O; Qian, Lijun; Dougherty, Edward R
2015-01-01
In this paper, we review multiscale modeling for cancer treatment with the incorporation of drug effects from an applied system's pharmacology perspective. Both the classical pharmacology and systems biology are inherently quantitative; however, systems biology focuses more on networks and multi factorial controls over biological processes rather than on drugs and targets in isolation, whereas systems pharmacology has a strong focus on studying drugs with regard to the pharmacokinetic (PK) and pharmacodynamic (PD) relations accompanying drug interactions with multiscale physiology as well as the prediction of dosage-exposure responses and economic potentials of drugs. Thus, it requires multiscale methods to address the need for integrating models from the molecular levels to the cellular, tissue, and organism levels. It is a common belief that tumorigenesis and tumor growth can be best understood and tackled by employing and integrating a multifaceted approach that includes in vivo and in vitro experiments, in silico models, multiscale tumor modeling, continuous/discrete modeling, agent-based modeling, and multiscale modeling with PK/PD drug effect inputs. We provide an example application of multiscale modeling employing stochastic hybrid system for a colon cancer cell line HCT-116 with the application of Lapatinib drug. It is observed that the simulation results are similar to those observed from the setup of the wet-lab experiments at the Translational Genomics Research Institute.
Atlas Career Path Guidebook: Patterns and Common Practices in Systems Engineers’ Development
2018-01-16
Overview of Atlas Proficiency Model .............................................................................. 68 5.1.2. Math /Science/General... Math /Science/General Engineering ................................ 72 Figure 42. Distribution for individuals with highest proficiency self...assessment in Math /Science/General Engineering ..................................................................................... 73 Figure 43
Monitoring Bloom Dynamics of a Common Coastal Bioluminescent Ctenophore
2010-09-30
photodiodes. IMPACT/APPLICATIONS More frequent and more rapidly developing jellyfish blooms, especially Mnemiopsis leidyi as well as Harmful Algal...To meet the need for a bioluminescent jellyfish monitoring and forecasting system, predictive models will depend upon dense networks of sensor
A coupled duration-focused architecture for real-time music-to-score alignment.
Cont, Arshia
2010-06-01
The capacity for real-time synchronization and coordination is a common ability among trained musicians performing a music score that presents an interesting challenge for machine intelligence. Compared to speech recognition, which has influenced many music information retrieval systems, music's temporal dynamics and complexity pose challenging problems to common approximations regarding time modeling of data streams. In this paper, we propose a design for a real-time music-to-score alignment system. Given a live recording of a musician playing a music score, the system is capable of following the musician in real time within the score and decoding the tempo (or pace) of its performance. The proposed design features two coupled audio and tempo agents within a unique probabilistic inference framework that adaptively updates its parameters based on the real-time context. Online decoding is achieved through the collaboration of the coupled agents in a Hidden Hybrid Markov/semi-Markov framework, where prediction feedback of one agent affects the behavior of the other. We perform evaluations for both real-time alignment and the proposed temporal model. An implementation of the presented system has been widely used in real concert situations worldwide and the readers are encouraged to access the actual system and experiment the results.
A General Water Resources Regulation Software System in China
NASA Astrophysics Data System (ADS)
LEI, X.
2017-12-01
To avoid iterative development of core modules in water resource normal regulation and emergency regulation and improve the capability of maintenance and optimization upgrading of regulation models and business logics, a general water resources regulation software framework was developed based on the collection and analysis of common demands for water resources regulation and emergency management. It can provide a customizable, secondary developed and extensible software framework for the three-level platform "MWR-Basin-Province". Meanwhile, this general software system can realize business collaboration and information sharing of water resources regulation schemes among the three-level platforms, so as to improve the decision-making ability of national water resources regulation. There are four main modules involved in the general software system: 1) A complete set of general water resources regulation modules allows secondary developer to custom-develop water resources regulation decision-making systems; 2) A complete set of model base and model computing software released in the form of Cloud services; 3) A complete set of tools to build the concept map and model system of basin water resources regulation, as well as a model management system to calibrate and configure model parameters; 4) A database which satisfies business functions and functional requirements of general water resources regulation software can finally provide technical support for building basin or regional water resources regulation models.
Yuksel, Mustafa; Gonul, Suat; Laleci Erturkmen, Gokce Banu; Sinaci, Ali Anil; Invernizzi, Paolo; Facchinetti, Sara; Migliavacca, Andrea; Bergvall, Tomas; Depraetere, Kristof; De Roo, Jos
2016-01-01
Depending mostly on voluntarily sent spontaneous reports, pharmacovigilance studies are hampered by low quantity and quality of patient data. Our objective is to improve postmarket safety studies by enabling safety analysts to seamlessly access a wide range of EHR sources for collecting deidentified medical data sets of selected patient populations and tracing the reported incidents back to original EHRs. We have developed an ontological framework where EHR sources and target clinical research systems can continue using their own local data models, interfaces, and terminology systems, while structural interoperability and Semantic Interoperability are handled through rule-based reasoning on formal representations of different models and terminology systems maintained in the SALUS Semantic Resource Set. SALUS Common Information Model at the core of this set acts as the common mediator. We demonstrate the capabilities of our framework through one of the SALUS safety analysis tools, namely, the Case Series Characterization Tool, which have been deployed on top of regional EHR Data Warehouse of the Lombardy Region containing about 1 billion records from 16 million patients and validated by several pharmacovigilance researchers with real-life cases. The results confirm significant improvements in signal detection and evaluation compared to traditional methods with the missing background information. PMID:27123451
Polarization division multiplexing for optical data communications
NASA Astrophysics Data System (ADS)
Ivanovich, Darko; Powell, Samuel B.; Gruev, Viktor; Chamberlain, Roger D.
2018-02-01
Multiple parallel channels are ubiquitous in optical communications, with spatial division multiplexing (separate physical paths) and wavelength division multiplexing (separate optical wavelengths) being the most common forms. Here, we investigate the viability of polarization division multiplexing, the separation of distinct parallel optical communication channels through the polarization properties of light. Two or more linearly polarized optical signals (at different polarization angles) are transmitted through a common medium, filtered using aluminum nanowire optical filters fabricated on-chip, and received using individual silicon photodetectors (one per channel). The entire receiver (including optics) is compatible with standard CMOS fabrication processes. The filter model is based upon an input optical signal formed as the sum of the Stokes vectors for each individual channel, transformed by the Mueller matrix that models the filter proper, resulting in an output optical signal that impinges on each photodiode. The results show that two- and three-channel systems can operate with a fixed-threshold comparator in the receiver circuit, but four-channel systems (and larger) will require channel coding of some form. For example, in the four-channel system, 10 of 16 distinct bit patterns are separable by the receiver. The model supports investigation of the range of variability tolerable in the fabrication of the on-chip polarization filters.
Tracking interface and common curve dynamics for two-fluid flow in porous media
Mcclure, James E.; Miller, Cass T.; Gray, W. G.; ...
2016-04-29
Pore-scale studies of multiphase flow in porous medium systems can be used to understand transport mechanisms and quantitatively determine closure relations that better incorporate microscale physics into macroscale models. Multiphase flow simulators constructed using the lattice Boltzmann method provide a means to conduct such studies, including both the equilibrium and dynamic aspects. Moving, storing, and analyzing the large state space presents a computational challenge when highly-resolved models are applied. We present an approach to simulate multiphase flow processes in which in-situ analysis is applied to track multiphase flow dynamics at high temporal resolution. We compute a comprehensive set of measuresmore » of the phase distributions and the system dynamics, which can be used to aid fundamental understanding and inform closure relations for macroscale models. The measures computed include microscale point representations and macroscale averages of fluid saturations, the pressure and velocity of the fluid phases, interfacial areas, interfacial curvatures, interface and common curve velocities, interfacial orientation tensors, phase velocities and the contact angle between the fluid-fluid interface and the solid surface. Test cases are studied to validate the approach and illustrate how measures of system state can be obtained and used to inform macroscopic theory.« less
Using emergent order to shape a space society
NASA Technical Reports Server (NTRS)
Graps, Amara L.
1993-01-01
A fast-growing movement in the scientific community is reshaping the way that we view the world around us. The short-hand name for this movement is 'chaos'. Chaos is a science of the global, nonlinear nature of systems. The center of this set of ideas is that simple, deterministic systems can breed complexity. Systems as complex as the human body, ecology, the mind or a human society. While it is true that simple laws can breed complexity, the other side is that complex systems can breed order. It is the latter that I will focus on in this paper. In the past, nonlinear was nearly synonymous with unsolvable because no general analytic solutions exist. Mathematically, an essential difference exists between linear and nonlinear systems. For linear systems, you just break up the complicated system into many simple pieces and patch together the separated solutions for each piece to form a solution to the full problem. In contrast, solutions to a nonlinear system cannot be added to form a new solution. The system must be treated in its full complexity. While it is true that no general analytical approach exists for reducing a complex system such as a society, it can be modeled. The technical involves a mathematical construct called phase space. In this space stable structures can appear which I use as analogies for the stable structures that appear in a complex system such as an ecology, the mind or a society. The common denominator in all of these systems is that they rely on a process called feedback loops. Feedback loops link the microscopic (individual) parts to the macroscopic (global) parts. The key, then, in shaping a space society, is in effectively using feedback loops. This paper will illustrate how one can model a space society by using methods that chaoticists have developed over the last hundred years. And I will show that common threads exist in the modeling of biological, economical, philosophical, and sociological systems.
Mass storage system reference model, Version 4
NASA Technical Reports Server (NTRS)
Coleman, Sam (Editor); Miller, Steve (Editor)
1993-01-01
The high-level abstractions that underlie modern storage systems are identified. The information to generate the model was collected from major practitioners who have built and operated large storage facilities, and represents a distillation of the wisdom they have acquired over the years. The model provides a common terminology and set of concepts to allow existing systems to be examined and new systems to be discussed and built. It is intended that the model and the interfaces identified from it will allow and encourage vendors to develop mutually-compatible storage components that can be combined to form integrated storage systems and services. The reference model presents an abstract view of the concepts and organization of storage systems. From this abstraction will come the identification of the interfaces and modules that will be used in IEEE storage system standards. The model is not yet suitable as a standard; it does not contain implementation decisions, such as how abstract objects should be broken up into software modules or how software modules should be mapped to hosts; it does not give policy specifications, such as when files should be migrated; does not describe how the abstract objects should be used or connected; and does not refer to specific hardware components. In particular, it does not fully specify the interfaces.
NASA Astrophysics Data System (ADS)
Ehlmann, Bryon K.
Current scientific experiments are often characterized by massive amounts of very complex data and the need for complex data analysis software. Object-oriented database (OODB) systems have the potential of improving the description of the structure and semantics of this data and of integrating the analysis software with the data. This dissertation results from research to enhance OODB functionality and methodology to support scientific databases (SDBs) and, more specifically, to support a nuclear physics experiments database for the Continuous Electron Beam Accelerator Facility (CEBAF). This research to date has identified a number of problems related to the practical application of OODB technology to the conceptual design of the CEBAF experiments database and other SDBs: the lack of a generally accepted OODB design methodology, the lack of a standard OODB model, the lack of a clear conceptual level in existing OODB models, and the limited support in existing OODB systems for many common object relationships inherent in SDBs. To address these problems, the dissertation describes an Object-Relationship Diagram (ORD) and an Object-oriented Database Definition Language (ODDL) that provide tools that allow SDB design and development to proceed systematically and independently of existing OODB systems. These tools define multi-level, conceptual data models for SDB design, which incorporate a simple notation for describing common types of relationships that occur in SDBs. ODDL allows these relationships and other desirable SDB capabilities to be supported by an extended OODB system. A conceptual model of the CEBAF experiments database is presented in terms of ORDs and the ODDL to demonstrate their functionality and use and provide a foundation for future development of experimental nuclear physics software using an OODB approach.
Genetics of common forms of heart failure: challenges and potential solutions.
Rau, Christoph D; Lusis, Aldons J; Wang, Yibin
2015-05-01
In contrast to many other human diseases, the use of genome-wide association studies (GWAS) to identify genes for heart failure (HF) has had limited success. We will discuss the underlying challenges as well as potential new approaches to understanding the genetics of common forms of HF. Recent research using intermediate phenotypes, more detailed and quantitative stratification of HF symptoms, founder populations and novel animal models has begun to allow researchers to make headway toward explaining the genetics underlying HF using GWAS techniques. By expanding analyses of HF to improved clinical traits, additional HF classifications and innovative model systems, the intractability of human HF GWAS should be ameliorated significantly.
Neurobiological roots of language in primate audition: common computational properties.
Bornkessel-Schlesewsky, Ina; Schlesewsky, Matthias; Small, Steven L; Rauschecker, Josef P
2015-03-01
Here, we present a new perspective on an old question: how does the neurobiology of human language relate to brain systems in nonhuman primates? We argue that higher-order language combinatorics, including sentence and discourse processing, can be situated in a unified, cross-species dorsal-ventral streams architecture for higher auditory processing, and that the functions of the dorsal and ventral streams in higher-order language processing can be grounded in their respective computational properties in primate audition. This view challenges an assumption, common in the cognitive sciences, that a nonhuman primate model forms an inherently inadequate basis for modeling higher-level language functions. Copyright © 2014 Elsevier Ltd. All rights reserved.
[Cost-effectiveness of a TLC-NOSF polyurethane foam dressing].
Arroyo Ana, Abejón; Alvarez Vázquez, Juan Carlos; Blasco García, Carmen; Bermejo Martínez, Mariano; López Casanova, Pablo; Cuesta Cuesta, Juan José; De Haro Fernández, Francisco; Mateo Marín, Emilia; Segovia Gómez, Teresa; Villar Rojas, Antonio Erasto
2012-11-01
Chronic wounds represent a drain on the Spanish health system, nowdays is necessary an optimization of the resources used and that is for this that is necessary justify the use of the products over others through cost-effective studies for to show the economic benefit to professionals and the life quality of patient. This article compares the use of a new technology for format polyurethane foam, TLC-NOSF, with the most commonly used products for treating wounds. This comparison is made using a cost-effectiveness model (Markov Model). The results demonstrate that treatment with polyurethane foam dressing with TLC-NOSF are cost-effective versus treatments with polyurethane foams most commonly used in Spain.
A study of ionospheric grid modification technique for BDS/GPS receiver
NASA Astrophysics Data System (ADS)
Liu, Xuelin; Li, Meina; Zhang, Lei
2017-07-01
For the single-frequency GPS receiver, ionospheric delay is an important factor affecting the positioning performance. There are many kinds of ionospheric correction methods, common models are Bent model, IRI model, Klobuchar model, Ne Quick model and so on. The US Global Positioning System (GPS) uses the Klobuchar coefficients transmitted in the satellite signal to correct the ionospheric delay error for a single frequency GPS receiver, but this model can only reduce the ionospheric error of about 50% in the mid-latitudes. In the Beidou system, the accuracy of the correction delay is higher. Therefore, this paper proposes a method that using BD grid information to correct GPS ionospheric delay to improve the ionospheric delay for the BDS/GPS compatible positioning receiver. In this paper, the principle of ionospheric grid algorithm is introduced in detail, and the positioning accuracy of GPS system and BDS/GPS compatible positioning system is compared and analyzed by the real measured data. The results show that the method can effectively improve the positioning accuracy of the receiver in a more concise way.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hale, Richard Edward; Cetiner, Sacit M.; Fugate, David L.
The Small Modular Reactor (SMR) Dynamic System Modeling Tool project is in the third year of development. The project is designed to support collaborative modeling and study of various advanced SMR (non-light water cooled) concepts, including the use of multiple coupled reactors at a single site. The objective of the project is to provide a common simulation environment and baseline modeling resources to facilitate rapid development of dynamic advanced reactor SMR models, ensure consistency among research products within the Instrumentation, Controls, and Human-Machine Interface (ICHMI) technical area, and leverage cross-cutting capabilities while minimizing duplication of effort. The combined simulation environmentmore » and suite of models are identified as the Modular Dynamic SIMulation (MoDSIM) tool. The critical elements of this effort include (1) defining a standardized, common simulation environment that can be applied throughout the program, (2) developing a library of baseline component modules that can be assembled into full plant models using existing geometry and thermal-hydraulic data, (3) defining modeling conventions for interconnecting component models, and (4) establishing user interfaces and support tools to facilitate simulation development (i.e., configuration and parameterization), execution, and results display and capture.« less
An overload behavior detection system for engineering transport vehicles based on deep learning
NASA Astrophysics Data System (ADS)
Zhou, Libo; Wu, Gang
2018-04-01
This paper builds an overloaded truck detect system called ITMD to help traffic department automatically identify the engineering transport vehicles (commonly known as `dirt truck') in CCTV and determine whether the truck is overloaded or not. We build the ITMD system based on the Single Shot MultiBox Detector (SSD) model. By constructing the image dataset of the truck and adjusting hyper-parameters of the original SSD neural network, we successfully trained a basic network model which the ITMD system depends on. The basic ITMD system achieves 83.01% mAP on classifying overload/non-overload truck, which is a not bad result. Still, some shortcomings of basic ITMD system have been targeted to enhance: it is easy for the ITMD system to misclassify other similar vehicle as truck. In response to this problem, we optimized the basic ITMD system, which effectively reduced basic model's false recognition rate. The optimized ITMD system achieved 86.18% mAP on the test set, which is better than the 83.01% mAP of the basic ITMD system.
NASA Technical Reports Server (NTRS)
Cole, Bjorn; Chung, Seung H.
2012-01-01
One of the challenges of systems engineering is in working multidisciplinary problems in a cohesive manner. When planning analysis of these problems, system engineers must tradeoff time and cost for analysis quality and quantity. The quality is associated with the fidelity of the multidisciplinary models and the quantity is associated with the design space that can be analyzed. The tradeoff is due to the resource intensive process of creating a cohesive multidisciplinary system model and analysis. Furthermore, reuse or extension of the models used in one stage of a product life cycle for another is a major challenge. Recent developments have enabled a much less resource-intensive and more rigorous approach than handwritten translation scripts or codes of multidisciplinary models and their analyses. The key is to work from a core system model defined in a MOF-based language such as SysML and in leveraging the emerging tool ecosystem, such as Query-View- Transform (QVT), from the OMG community. SysML was designed to model multidisciplinary systems and analyses. The QVT standard was designed to transform SysML models. The Europa Hability Mission (EHM) team has begun to exploit these capabilities. In one case, a Matlab/Simulink model is generated on the fly from a system description for power analysis written in SysML. In a more general case, a symbolic mathematical framework (supported by Wolfram Mathematica) is coordinated by data objects transformed from the system model, enabling extremely flexible and powerful tradespace exploration and analytical investigations of expected system performance.
New approach to analyzing soil-building systems
Safak, E.
1998-01-01
A new method of analyzing seismic response of soil-building systems is introduced. The method is based on the discrete-time formulation of wave propagation in layered media for vertically propagating plane shear waves. Buildings are modeled as an extension of the layered soil media by assuming that each story in the building is another layer. The seismic response is expressed in terms of wave travel times between the layers, and the wave reflection and transmission coefficients at layer interfaces. The calculation of the response is reduced to a pair of simple finite-difference equations for each layer, which are solved recursively starting from the bedrock. Compared with commonly used vibration formulation, the wave propagation formulation provides several advantages, including the ability to incorporate soil layers, simplicity of the calculations, improved accuracy in modeling the mass and damping, and better tools for system identification and damage detection.A new method of analyzing seismic response of soil-building systems is introduced. The method is based on the discrete-time formulation of wave propagation in layered media for vertically propagating plane shear waves. Buildings are modeled as an extension of the layered soil media by assuming that each story in the building is another layer. The seismic response is expressed in terms of wave travel times between the layers, and the wave reflection and transmission coefficients at layer interfaces. The calculation of the response is reduced to a pair of simple finite-difference equations for each layer, which are solved recursively starting from the bedrock. Compared with commonly used vibration formulation, the wave propagation formulation provides several advantages, including the ability to incorporate soil layers, simplicity of the calculations, improved accuracy in modeling the mass and damping, and better tools for system identification and damage detection.