Science.gov

Sample records for agile process models

  1. Planning and scheduling for agile manufacturers: The Pantex Process Model

    SciTech Connect

    Kjeldgaard, E.A.; Jones, D.A.; List, G.F.; Tumquist, M.A.

    1998-02-01

    Effective use of resources that are shared among multiple products or processes is critical for agile manufacturing. This paper describes the development and implementation of a computerized model to support production planning in a complex manufacturing system at the Pantex Plant, a US Department of Energy facility. The model integrates two different production processes (nuclear weapon disposal and stockpile evaluation) that use common facilities and personnel at the plant. The two production processes are characteristic of flow-shop and job shop operations. The model reflects the interactions of scheduling constraints, material flow constraints, and the availability of required technicians and facilities. Operational results show significant productivity increases from use of the model.

  2. Are we unnecessarily constraining the agility of complex process-based models?

    NASA Astrophysics Data System (ADS)

    Mendoza, Pablo A.; Clark, Martyn P.; Barlage, Michael; Rajagopalan, Balaji; Samaniego, Luis; Abramowitz, Gab; Gupta, Hoshin

    2015-01-01

    In this commentary we suggest that hydrologists and land-surface modelers may be unnecessarily constraining the behavioral agility of very complex physics-based models. We argue that the relatively poor performance of such models can occur due to restrictions on their ability to refine their portrayal of physical processes, in part because of strong a priori constraints in: (i) the representation of spatial variability and hydrologic connectivity, (ii) the choice of model parameterizations, and (iii) the choice of model parameter values. We provide a specific example of problems associated with strong a priori constraints on parameters in a land surface model. Moving forward, we assert that improving hydrological models requires integrating the strengths of the "physics-based" modeling philosophy (which relies on prior knowledge of hydrologic processes) with the strengths of the "conceptual" modeling philosophy (which relies on data driven inference). Such integration will accelerate progress on methods to define and discriminate among competing modeling options, which should be ideally incorporated in agile modeling frameworks and tested through a diagnostic evaluation approach.

  3. Opening up the Agile Innovation Process

    NASA Astrophysics Data System (ADS)

    Conboy, Kieran; Donnellan, Brian; Morgan, Lorraine; Wang, Xiaofeng

    The objective of this panel is to discuss how firms can operate both an open and agile innovation process. In an era of unprecedented changes, companies need to be open and agile in order to adapt rapidly and maximize their innovation processes. Proponents of agile methods claim that one of the main distinctions between agile methods and their traditional bureaucratic counterparts is their drive toward creativity and innovation. However, agile methods are rarely adopted in their textbook, "vanilla" format, and are usually adopted in part or are tailored or modified to suit the organization. While we are aware that this happens, there is still limited understanding of what is actually happening in practice. Using innovation adoption theory, this panel will discuss the issues and challenges surrounding the successful adoption of agile practices. In addition, this panel will report on the obstacles and benefits reported by over 20 industrial partners engaged in a pan-European research project into agile practices between 2006 and 2009.

  4. Agile

    NASA Technical Reports Server (NTRS)

    Trimble, Jay Phillip

    2013-01-01

    This is based on a previous talk on agile development. Methods for delivering software on a short cycle are described, including interactions with the customer, the affect on the team, and how to be more effective, streamlined and efficient.

  5. Pilot users in agile development processes: motivational factors.

    PubMed

    Johannessen, Liv Karen; Gammon, Deede

    2010-01-01

    Despite a wealth of research on user participation, few studies offer insights into how to involve multi-organizational users in agile development methods. This paper is a case study of user involvement in developing a system for electronic laboratory requisitions using agile methodologies in a multi-organizational context. Building on an interpretive approach, we illuminate questions such as: How does collaboration between users and developers evolve and how might it be improved? What key motivational aspects are at play when users volunteer and continue contributing in the face of considerable added burdens? The study highlights how agile methods in themselves appear to facilitate mutually motivating collaboration between user groups and developers. Lessons learned for leveraging the advantages of agile development processes include acknowledging the substantial and ongoing contributions of users and their roles as co-designers of the system. PMID:20543366

  6. Pilot users in agile development processes: motivational factors.

    PubMed

    Johannessen, Liv Karen; Gammon, Deede

    2010-01-01

    Despite a wealth of research on user participation, few studies offer insights into how to involve multi-organizational users in agile development methods. This paper is a case study of user involvement in developing a system for electronic laboratory requisitions using agile methodologies in a multi-organizational context. Building on an interpretive approach, we illuminate questions such as: How does collaboration between users and developers evolve and how might it be improved? What key motivational aspects are at play when users volunteer and continue contributing in the face of considerable added burdens? The study highlights how agile methods in themselves appear to facilitate mutually motivating collaboration between user groups and developers. Lessons learned for leveraging the advantages of agile development processes include acknowledging the substantial and ongoing contributions of users and their roles as co-designers of the system.

  7. Thinking Outside the Box: Agile Business Models for CNOs

    NASA Astrophysics Data System (ADS)

    Loss, Leandro; Crave, Servane

    This paper introduces the idea of an agile Business Model for CNOs grounded on a new model of innovation based on the effects of globalization and of Knowledge Economy. The agile Business Model considers the resources that are spread out and available worldwide as well as the need for each customer to receive a unique customer experience. It aims at reinforcing in the context of the Knowledge Economy the different business models approaches developed so far. The paper also identifies the levers and the barriers of Agile Business Models Innovation in CNOs.

  8. Agility and mixed-model furniture production

    NASA Astrophysics Data System (ADS)

    Yao, Andrew C.

    2000-10-01

    The manufacture of upholstered furniture provides an excellent opportunity to analyze the effect of a comprehensive communication system on classical production management functions. The objective of the research is to study the scheduling heuristics that embrace the concepts inherent in MRP, JIT and TQM while recognizing the need for agility in a somewhat complex and demanding environment. An on-line, real-time data capture system provides the status and location of production lots, components, subassemblies for schedule control. Current inventory status of raw material and purchased items are required in order to develop and adhere to schedules. For the large variety of styles and fabrics customers may order, the communication system must provide timely, accurate and comprehensive information for intelligent decisions with respect to the product mix and production resources.

  9. RFID-Based Critical Path Expert System for Agility Manufacture Process Management

    NASA Astrophysics Data System (ADS)

    Cheng, Haifang; Xiang, Yuli

    This paper presents a critical path expert system for the agility manufacture process management based on radio frequency identification (RFID) technology. The paper explores that the agility manufacture processes can be visible and controllable with RFID. The critical paths or activities can be easily found out and tracked by the RFID tracing technology. And the expert system can optimize the bottle neck of the task process of the agility management with the critical path adjusting and reforming method. Finally, the paper gives a simple application example of the system to discuss how to adjust the critical paths and how to make the process more agility and flexibility with the critical path expert system. With an RFID-based critical path expert system, the agility manufacture process management will be more effective and efficient.

  10. The NERV Methodology: Non-Functional Requirements Elicitation, Reasoning and Validation in Agile Processes

    ERIC Educational Resources Information Center

    Domah, Darshan

    2013-01-01

    Agile software development has become very popular around the world in recent years, with methods such as Scrum and Extreme Programming (XP). Literature suggests that functionality is the primary focus in Agile processes while non-functional requirements (NFR) are either ignored or ill-defined. However, for software to be of good quality both…

  11. Bringing Agility to Business Process Management: Rules Deployment in an SOA

    NASA Astrophysics Data System (ADS)

    El Kharbili, Marwane; Keil, Tobias

    Business process management (BPM) has emerged as paradigm for integrating business strategies and enterprise architecture (EA). In this context, BPM implementation on top of web-service-based service oriented architectures is an accepted approach as shown by great amount of literature. One concern in this regard is how-to make BPs reactive to change. Our approach to the problem is the integration of business rule management (BRM) and BPM by allowing modeling of decisions hard-coded in BPs as separate business rules (BRs). These BRs become EA assets and need to be exploited when executing BPs. We motivate why BPM needs agility and discuss what requirements on BPM this poses. This paper presents prototyping work conducted at a BP modeling and analysis vendor which weeks to showcase how using business rule management (BRM) as a mean for modeling decisions can help achieve a much sought-after agility to BPM. This prototype relies on the integrated modeling of business rules (BRs) and BPs, and rule deployment as web services part of an SOA.

  12. Information Models, Data Requirements, and Agile Data Curation

    NASA Astrophysics Data System (ADS)

    Hughes, John S.; Crichton, Dan; Ritschel, Bernd; Hardman, Sean; Joyner, Ron

    2015-04-01

    The Planetary Data System's next generation system, PDS4, is an example of the successful use of an ontology-based Information Model (IM) to drive the development and operations of a data system. In traditional systems engineering, requirements or statements about what is necessary for the system are collected and analyzed for input into the design stage of systems development. With the advent of big data the requirements associated with data have begun to dominate and an ontology-based information model can be used to provide a formalized and rigorous set of data requirements. These requirements address not only the usual issues of data quantity, quality, and disposition but also data representation, integrity, provenance, context, and semantics. In addition the use of these data requirements during system's development has many characteristics of Agile Curation as proposed by Young et al. [Taking Another Look at the Data Management Life Cycle: Deconstruction, Agile, and Community, AGU 2014], namely adaptive planning, evolutionary development, early delivery, continuous improvement, and rapid and flexible response to change. For example customers can be satisfied through early and continuous delivery of system software and services that are configured directly from the information model. This presentation will describe the PDS4 architecture and its three principle parts: the ontology-based Information Model (IM), the federated registries and repositories, and the REST-based service layer for search, retrieval, and distribution. The development of the IM will be highlighted with special emphasis on knowledge acquisition, the impact of the IM on development and operations, and the use of shared ontologies at multiple governance levels to promote system interoperability and data correlation.

  13. Modeling the Agility MLC in the Monaco treatment planning system.

    PubMed

    Snyder, Michael; Halford, Robert; Knill, Cory; Adams, Jeffrey N; Bossenberger, Todd; Nalichowski, Adrian; Hammoud, Ahmad; Burmeister, Jay

    2016-01-01

    We investigate the relationship between the various parameters in the Monaco MLC model and dose calculation accuracy for an Elekta Agility MLC. The vendor-provided MLC modeling procedure - completed first with external vendor participation and then exclusively in-house - was used in combination with our own procedures to investigate several sets of MLC modeling parameters to determine their effect on dose distributions and point-dose measurements. Simple plans provided in the vendor procedure were used to elucidate specific mechanical characteristics of the MLC, while ten complex treatment plans - five IMRT and five VMAT - created using TG-119-based structure sets were used to test clinical dosimetric effects of particular parameter choices. EDR2 film was used for the vendor fields to give high spatial resolution, while a combination of MapCHECK and ion chambers were used for the in-house TG-119-based proced-ures. The vendor-determined parameter set provided a reasonable starting point for the MLC model and largely delivered acceptable gamma pass rates for clinical plans - including a passing external evaluation using the IROC H&N phantom. However, the vendor model did not provide point-dose accuracy consistent with that seen in other treatment systems at our center. Through further internal testing it was found that there existed many sets of MLC parameters, often at opposite ends of their allowable ranges, that provided similar dosimetric characteristics and good agreement with planar and point-dose measurements. In particular, the leaf offset and tip leakage parameters compensated for one another if adjusted in opposite directions, which provided a level curve of acceptable parameter sets across all plans. Interestingly, gamma pass rates of the plans were less dependent upon parameter choices than point-dose measurements, suggesting that MLC modeling using only gamma evaluation may be generally an insufficient approach. It was also found that exploring all

  14. A process for the agile product realization of electro-mechanical devices

    SciTech Connect

    Forsythe, C.; Ashby, M.R.; Benavides, G.L.; Diegert, K.V.; Jones, R.E.; Longcope, D.B.; Parratt, S.W.

    1995-09-01

    This paper describes a product realization process developed and demonstrated at Sandia by the A-PRIMED (Agile Product Realization for Innovative Electro MEchanical Devices) project that integrates many of the key components of ``agile manufacturing`` into a complete, design-to-production process. Evidence indicates that the process has reduced the product realization cycle and assured product quality. Products included discriminators for a robotic quick change adapter and for an electronic defense system. These discriminators, built using A-PRIMED, met random vibration requirements and had life cycles that far surpass the performance obtained from earlier efforts.

  15. Impact of Agile Software Development Model on Software Maintainability

    ERIC Educational Resources Information Center

    Gawali, Ajay R.

    2012-01-01

    Software maintenance and support costs account for up to 60% of the overall software life cycle cost and often burdens tightly budgeted information technology (IT) organizations. Agile software development approach delivers business value early, but implications on software maintainability are still unknown. The purpose of this quantitative study…

  16. Unsteady aerodynamic models for agile flight at low Reynolds numbers

    NASA Astrophysics Data System (ADS)

    Brunton, Steven L.

    This work develops low-order models for the unsteady aerodynamic forces on a wing in response to agile maneuvers at low Reynolds number. Model performance is assessed on the basis of accuracy across a range of parameters and frequencies as well as of computational efficiency and compatibility with existing control techniques and flight dynamic models. The result is a flexible modeling procedure that yields accurate, low-dimensional, state-space models. The modeling procedures are developed and tested on direct numerical simulations of a two-dimensional flat plate airfoil in motion at low Reynolds number, Re=100, and in a wind tunnel experiment at the Illinois Institute of Technology involving a NACA 0006 airfoil pitching and plunging at Reynolds number Re=65,000. In both instances, low-order models are obtained that accurately capture the unsteady aerodynamic forces at all frequencies. These cases demonstrate the utility of the modeling procedure developed in this thesis for obtaining accurate models for different geometries and Reynolds numbers. Linear reduced-order models are constructed from either the indicial response (step response) or realistic input/output maneuvers using a flexible modeling procedure. The method is based on identifying stability derivatives and modeling the remaining dynamics with the eigensystem realization algorithm. A hierarchy of models is developed, based on linearizing the flow at various operating conditions. These models are shown to be accurate and efficient for plunging, pitching about various points, and combined pitch and plunge maneuvers, at various angle of attack and Reynolds number. Models are compared against the classical unsteady aerodynamic models of Wagner and Theodorsen over a large range of Strouhal number and reduced frequency for a baseline comparison. Additionally, state-space representations are developed for Wagner's and Theodorsen's models, making them compatible with modern control-system analysis. A number of

  17. Agile Walker.

    PubMed

    Katz, Reuven

    2015-01-01

    The goal of the Agile Walker is to improve the outdoor mobility of healthy elderly people with some mobility limitations. It is a newly developed, all-terrain walker, equipped with an electric drive system and speed control that can assists elderly people to walk outdoors or to hike. The walker has a unique product design with an attractive look that will appeal to "active-agers" population. This paper describes product design requirements and the development process of the Agile Walker, its features and some preliminary testing results.

  18. Agile based "Semi-"Automated Data ingest process : ORNL DAAC example

    NASA Astrophysics Data System (ADS)

    Santhana Vannan, S. K.; Beaty, T.; Cook, R. B.; Devarakonda, R.; Hook, L.; Wei, Y.; Wright, D.

    2015-12-01

    The ORNL DAAC archives and publishes data and information relevant to biogeochemical, ecological, and environmental processes. The data archived at the ORNL DAAC must be well formatted, self-descriptive, and documented, as well as referenced in a peer-reviewed publication. The ORNL DAAC ingest team curates diverse data sets from multiple data providers simultaneously. To streamline the ingest process, the data set submission process at the ORNL DAAC has been recently updated to use an agile process and a semi-automated workflow system has been developed to provide a consistent data provider experience and to create a uniform data product. The goals of semi-automated agile ingest process are to: 1.Provide the ability to track a data set from acceptance to publication 2. Automate steps that can be automated to improve efficiencies and reduce redundancy 3.Update legacy ingest infrastructure 4.Provide a centralized system to manage the various aspects of ingest. This talk will cover the agile methodology, workflow, and tools developed through this system.

  19. Exploring the possibility of modeling a genetic counseling guideline using agile methodology.

    PubMed

    Choi, Jeeyae

    2013-01-01

    Increased demand of genetic counseling services heightened the necessity of a computerized genetic counseling decision support system. In order to develop an effective and efficient computerized system, modeling of genetic counseling guideline is an essential step. Throughout this pilot study, Agile methodology with United Modeling Language (UML) was utilized to model a guideline. 13 tasks and 14 associated elements were extracted. Successfully constructed conceptual class and activity diagrams revealed that Agile methodology with UML was a suitable tool to modeling a genetic counseling guideline.

  20. Developing a model for agile supply: an empirical study from Iranian pharmaceutical supply chain.

    PubMed

    Rajabzadeh Ghatari, Ali; Mehralian, Gholamhossein; Zarenezhad, Forouzandeh; Rasekh, Hamid Reza

    2013-01-01

    Agility is the fundamental characteristic of a supply chain needed for survival in turbulent markets, where environmental forces create additional uncertainty resulting in higher risk in the supply chain management. In addition, agility helps providing the right product, at the right time to the consumer. The main goal of this research is therefore to promote supplier selection in pharmaceutical industry according to the formative basic factors. Moreover, this paper can configure its supply network to achieve the agile supply chain. The present article analyzes the supply part of supply chain based on SCOR model, used to assess agile supply chains by highlighting their specific characteristics and applicability in providing the active pharmaceutical ingredient (API). This methodology provides an analytical modeling; the model enables potential suppliers to be assessed against the multiple criteria using both quantitative and qualitative measures. In addition, for making priority of critical factors, TOPSIS algorithm has been used as a common technique of MADM model. Finally, several factors such as delivery speed, planning and reorder segmentation, trust development and material quantity adjustment are identified and prioritized as critical factors for being agile in supply of API. PMID:24250689

  1. Developing a Model for Agile Supply: an Empirical Study from Iranian Pharmaceutical Supply Chain

    PubMed Central

    Rajabzadeh Ghatari, Ali; Mehralian, Gholamhossein; Zarenezhad, Forouzandeh; Rasekh, Hamid Reza

    2013-01-01

    Agility is the fundamental characteristic of a supply chain needed for survival in turbulent markets, where environmental forces create additional uncertainty resulting in higher risk in the supply chain management. In addition, agility helps providing the right product, at the right time to the consumer. The main goal of this research is therefore to promote supplier selection in pharmaceutical industry according to the formative basic factors. Moreover, this paper can configure its supply network to achieve the agile supply chain. The present article analyzes the supply part of supply chain based on SCOR model, used to assess agile supply chains by highlighting their specific characteristics and applicability in providing the active pharmaceutical ingredient (API). This methodology provides an analytical modeling; the model enables potential suppliers to be assessed against the multiple criteria using both quantitative and qualitative measures. In addition, for making priority of critical factors, TOPSIS algorithm has been used as a common technique of MADM model. Finally, several factors such as delivery speed, planning and reorder segmentation, trust development and material quantity adjustment are identified and prioritized as critical factors for being agile in supply of API. PMID:24250689

  2. Developing a model for agile supply: an empirical study from Iranian pharmaceutical supply chain.

    PubMed

    Rajabzadeh Ghatari, Ali; Mehralian, Gholamhossein; Zarenezhad, Forouzandeh; Rasekh, Hamid Reza

    2013-01-01

    Agility is the fundamental characteristic of a supply chain needed for survival in turbulent markets, where environmental forces create additional uncertainty resulting in higher risk in the supply chain management. In addition, agility helps providing the right product, at the right time to the consumer. The main goal of this research is therefore to promote supplier selection in pharmaceutical industry according to the formative basic factors. Moreover, this paper can configure its supply network to achieve the agile supply chain. The present article analyzes the supply part of supply chain based on SCOR model, used to assess agile supply chains by highlighting their specific characteristics and applicability in providing the active pharmaceutical ingredient (API). This methodology provides an analytical modeling; the model enables potential suppliers to be assessed against the multiple criteria using both quantitative and qualitative measures. In addition, for making priority of critical factors, TOPSIS algorithm has been used as a common technique of MADM model. Finally, several factors such as delivery speed, planning and reorder segmentation, trust development and material quantity adjustment are identified and prioritized as critical factors for being agile in supply of API.

  3. Agile Development Processes: Delivering a Successful Data Management Platform Now and in the Future

    NASA Astrophysics Data System (ADS)

    Deaubl, E.; Lowry, S.

    2007-10-01

    Developing a flexible, extensible architecture for scientific data archival and management is a monumental task under older, big design, up-front methodologies. We will describe how we are using agile development techniques in our service oriented architecture (SOA)-based platform to integrate astronomer and operator input into the development process, deliver functional software earlier, and ensure that the software is maintainable and extensible in the future.

  4. Agile waveforms for joint SAR-GMTI processing

    NASA Astrophysics Data System (ADS)

    Jaroszewski, Steven; Corbeil, Allan; McMurray, Stephen; Majumder, Uttam; Bell, Mark R.; Corbeil, Jeffrey; Minardi, Michael

    2016-05-01

    Wideband radar waveforms that employ spread-spectrum techniques were investigated and experimentally tested. The waveforms combine bi-phase coding with a traditional LFM chirp and are applicable to joint SAR-GMTI processing. After de-spreading, the received signals can be processed to support simultaneous GMTI and high resolution SAR imaging missions by airborne radars. The spread spectrum coding techniques can provide nearly orthogonal waveforms and offer enhanced operations in some environments by distributing the transmitted energy over a large instantaneous bandwidth. The LFM component offers the desired Doppler tolerance. In this paper, the waveforms are formulated and a shift-register approach for de-spreading the received signals is described. Hardware loop-back testing has shown the feasibility of using these waveforms in experimental radar test bed.

  5. Agile IT: Thinking in User-Centric Models

    NASA Astrophysics Data System (ADS)

    Margaria, Tiziana; Steffen, Bernhard

    We advocate a new teaching direction for modern CS curricula: extreme model-driven development (XMDD), a new development paradigm designed to continuously involve the customer/application expert throughout the whole systems' life cycle. Based on the `One-Thing Approach', which works by successively enriching and refining one single artifact, system development becomes in essence a user-centric orchestration of intuitive service functionality. XMDD differs radically from classical software development, which, in our opinion is no longer adequate for the bulk of application programming - in particular when it comes to heterogeneous, cross organizational systems which must adapt to rapidly changing market requirements. Thus there is a need for new curricula addressing this model-driven, lightweight, and cooperative development paradigm that puts the user process in the center of the development and the application expert in control of the process evolution.

  6. Positioning Agility

    NASA Astrophysics Data System (ADS)

    Oza, Nilay; Abrahamsson, Pekka; Conboy, Kieran

    Agile methods are increasingly adopted by European companies. Academics too are conducting numerous studies on different tenets of agile methods. Companies often feel proud in marketing themselves as ‘agile’. However, the true notion of ‘being agile’ seems to have been overlooked due to lack of positioning of oneself for agility. This raises a call for more research and interactions between academia and the industry. The proposed workshop refers to this call. It will be highly relevant to participants, interested in positioning their company’s agility from organizational, group or project perspectives. The positioning of agility will help companies to better align their agile practices with stakeholder values. Results of the workshop will be shared across participants and they will also have opportunity to continue their work on agile positioning in their companies. At broader level, the work done in this workshop will contribute towards developing Agile Positioning System.

  7. Modeling and Developing the Information System for the SuperAGILE Experiment

    NASA Astrophysics Data System (ADS)

    Lazzarotto, F.; Costa, E.; del Monte, E.; Feroci, M.

    2004-07-01

    We will present some formal description of the SuperAGILE (SA) detection system data, the relationships among them and the operations applied on data, with the aid of instruments such as Entity-Relationship (E-R) and UML diagrams. We just realized functions of reception, pre-processing, archiving and analysis on SA data making use of Object Oriented and SQL open source software instruments.

  8. Agile Methods for Open Source Safety-Critical Software

    PubMed Central

    Enquobahrie, Andinet; Ibanez, Luis; Cheng, Patrick; Yaniv, Ziv; Cleary, Kevin; Kokoori, Shylaja; Muffih, Benjamin; Heidenreich, John

    2011-01-01

    The introduction of software technology in a life-dependent environment requires the development team to execute a process that ensures a high level of software reliability and correctness. Despite their popularity, agile methods are generally assumed to be inappropriate as a process family in these environments due to their lack of emphasis on documentation, traceability, and other formal techniques. Agile methods, notably Scrum, favor empirical process control, or small constant adjustments in a tight feedback loop. This paper challenges the assumption that agile methods are inappropriate for safety-critical software development. Agile methods are flexible enough to encourage the right amount of ceremony; therefore if safety-critical systems require greater emphasis on activities like formal specification and requirements management, then an agile process will include these as necessary activities. Furthermore, agile methods focus more on continuous process management and code-level quality than classic software engineering process models. We present our experiences on the image-guided surgical toolkit (IGSTK) project as a backdrop. IGSTK is an open source software project employing agile practices since 2004. We started with the assumption that a lighter process is better, focused on evolving code, and only adding process elements as the need arose. IGSTK has been adopted by teaching hospitals and research labs, and used for clinical trials. Agile methods have matured since the academic community suggested they are not suitable for safety-critical systems almost a decade ago, we present our experiences as a case study for renewing the discussion. PMID:21799545

  9. Agile Methods for Open Source Safety-Critical Software.

    PubMed

    Gary, Kevin; Enquobahrie, Andinet; Ibanez, Luis; Cheng, Patrick; Yaniv, Ziv; Cleary, Kevin; Kokoori, Shylaja; Muffih, Benjamin; Heidenreich, John

    2011-08-01

    The introduction of software technology in a life-dependent environment requires the development team to execute a process that ensures a high level of software reliability and correctness. Despite their popularity, agile methods are generally assumed to be inappropriate as a process family in these environments due to their lack of emphasis on documentation, traceability, and other formal techniques. Agile methods, notably Scrum, favor empirical process control, or small constant adjustments in a tight feedback loop. This paper challenges the assumption that agile methods are inappropriate for safety-critical software development. Agile methods are flexible enough to encourage the rightamount of ceremony; therefore if safety-critical systems require greater emphasis on activities like formal specification and requirements management, then an agile process will include these as necessary activities. Furthermore, agile methods focus more on continuous process management and code-level quality than classic software engineering process models. We present our experiences on the image-guided surgical toolkit (IGSTK) project as a backdrop. IGSTK is an open source software project employing agile practices since 2004. We started with the assumption that a lighter process is better, focused on evolving code, and only adding process elements as the need arose. IGSTK has been adopted by teaching hospitals and research labs, and used for clinical trials. Agile methods have matured since the academic community suggested they are not suitable for safety-critical systems almost a decade ago, we present our experiences as a case study for renewing the discussion.

  10. Agile Methods for Open Source Safety-Critical Software.

    PubMed

    Gary, Kevin; Enquobahrie, Andinet; Ibanez, Luis; Cheng, Patrick; Yaniv, Ziv; Cleary, Kevin; Kokoori, Shylaja; Muffih, Benjamin; Heidenreich, John

    2011-08-01

    The introduction of software technology in a life-dependent environment requires the development team to execute a process that ensures a high level of software reliability and correctness. Despite their popularity, agile methods are generally assumed to be inappropriate as a process family in these environments due to their lack of emphasis on documentation, traceability, and other formal techniques. Agile methods, notably Scrum, favor empirical process control, or small constant adjustments in a tight feedback loop. This paper challenges the assumption that agile methods are inappropriate for safety-critical software development. Agile methods are flexible enough to encourage the rightamount of ceremony; therefore if safety-critical systems require greater emphasis on activities like formal specification and requirements management, then an agile process will include these as necessary activities. Furthermore, agile methods focus more on continuous process management and code-level quality than classic software engineering process models. We present our experiences on the image-guided surgical toolkit (IGSTK) project as a backdrop. IGSTK is an open source software project employing agile practices since 2004. We started with the assumption that a lighter process is better, focused on evolving code, and only adding process elements as the need arose. IGSTK has been adopted by teaching hospitals and research labs, and used for clinical trials. Agile methods have matured since the academic community suggested they are not suitable for safety-critical systems almost a decade ago, we present our experiences as a case study for renewing the discussion. PMID:21799545

  11. A process for the agile product realization of electromechanical devices (A-primed)

    SciTech Connect

    Forsythe, C.; Ashby, M.R.; Benavides, G.L.; Diegert, K.V.; Jones, R.E.; Longcope, D.B.; Parratt, S.W.

    1996-02-01

    This paper describes a product realization process developed at Sandia National Laboratories by the A-PRIMED project that integrates many of the key components of ``agile manufacturing`` (Nagel & Dove, 1992) into a complete, step-by-step, design-to-production process. For two separate product realization efforts, each geared to a different set of requirements, A-PRIMED demonstrated product realization of a custom device in less than a month. A-PRIMED used a discriminator (a precision electro mechanical device) as the demonstration device, but the process is readily adaptable to other electro mechanical products. The process begins with a qualified design parameter space (Diegert et al, 1995). From that point, the product realization process encompasses all facets of requirements development, analysis and testing, design, manufacturing, robot assembly and quality assurance, as well as product data management and concurrent engineering. In developing the product realization process, A-PRIMED employed an iterative approach whereby after each build, the process was reviewed and refinements were made on the basis of lessons learned. This paper describes the integration of project functions and product realization technologies to develop a product realization process that on repeated iterations, was proven successful.

  12. Allometric multilevel modelling of agility and dribbling speed by skeletal age and playing position in youth soccer players.

    PubMed

    Valente-dos-Santos, J; Coelho-e-Silva, M J; Duarte, J; Pereira, J; Rebelo-Gonçalves, R; Figueiredo, A; Mazzuco, M A; Sherar, L B; Elferink-Gemser, M T; Malina, R M

    2014-08-01

    This study evaluates the contributions of age, skeletal maturation, body size and composition, training and playing position to the development of agility and dribbling speed in young male soccer players (10-18 years) followed longitudinally. 83 players [defenders (n=35), midfielders (n=27), forwards (n=21)] were followed annually over 5 years (average: 4.4 observations per player). Skeletal age (SA), stature, body mass, triceps and subscapular skinfolds, agility and dribbling speed were measured annually. Body composition was estimated from the 2 skinfolds. Annual training volume was estimated from weekly participation forms completed by coaches. The multiplicative allometric models with the best statistical fit showed that statural growth of 1 cm predicts 1.334 s and 1.927 s of improvement in agility and dribbling speed, respectively. Significant independent effects of fat-free mass and annual volume training were found for agility and dribbling speed, respectively (P<0.05). Predicted agility (from 12 to 18 years of SA) and dribbling speed (from 13 to 18 years of SA) differed significantly among players by playing positions (midfielders>forwards>defenders). The present results provide developmental models for the interpretation of intra- and inter-individual variability in agility and dribbling speed among youth soccer players across adolescence, and may provide a framework for trainers and coaches to develop and evaluate individualized training protocols.

  13. A minimal limit-cycle model to profile movement patterns of individuals during agility drill performance: Effects of skill level.

    PubMed

    Mehdizadeh, Sina; Arshi, Ahmed Reza; Davids, Keith

    2015-06-01

    Identification of control strategies during agility performance is significant in understanding movement behavior. This study aimed at providing a fundamental mathematical model for describing the motion of participants during an agility drill and to determine whether skill level constrained model components. Motion patterns of two groups of skilled and unskilled participants (n=8 in each) during performance of a forward/backward agility drill modeled as limit-cycles. Participant movements were recorded by motion capture of a reflective marker attached to the sacrum of each individual. Graphical and regression analyses of movement kinematics in Hooke's plane, phase plane and velocity profile were performed to determine components of the models. Results showed that the models of both skilled and unskilled groups had terms from Duffing stiffness as well as Van der Pol damping oscillators. Data also indicated that the proposed models captured on average 97% of the variance for both skilled and unskilled groups. Findings from this study revealed the movement patterning associated with skilled and unskilled performance in a typical forward/backward agility drill which might be helpful for trainers and physiotherapists in enhancing agility. PMID:25828582

  14. Agility Meets Systems Engineering: A Catalogue of Success Factors from Industry Practice

    NASA Astrophysics Data System (ADS)

    Stelzmann, Ernst; Kreiner, Christian; Spork, Gunther; Messnarz, Richard; Koenig, Frank

    Agile software development methods are widely accepted and valued in software-dominated industries. In more complex setups like multidisciplinary system development the adoption of an agile development paradigm is much less straightforward. Bigger teams, longer development cycles, process and product standard compliance and products lacking flexibility make an agile behaviour more difficult to achieve. Focusing on the fundamental underlying problem of dealing with ever ongoing change, this paper presents an agile Systems Engineering approach as a potential solution. Therefore a generic Systems Engineering action model was upgraded respecting agile principles and adapted according to practical needs discovered in an empirical study. This study was conducted among the partners of the S2QI agile workgroup made up from experts of automotive, logistics and electronics industries. Additionally to an agile Systems Engineering action model, a list of 15 practical success factors that should be considered when using an agile Systems Engineering approach is one of the main outcomes of this survey. It was also found that an agile behaviour in Systems Engineering could be supported in many different areas within companies. These areas are listed and it is also shown how the agile action model and the agile success factors are related to them.

  15. Organizational Culture and the Deployment of Agile Methods: The Competing Values Model View

    NASA Astrophysics Data System (ADS)

    Iivari, Juhani; Iivari, Netta

    A number of researchers have identified organizational culture as a factor that potentially affects the deployment of agile systems development methods. Inspired by the study of Iivari and Huisman (2007), which focused on the deployment of traditional systems development methods, the present paper proposes a number of hypotheses about the influence of organizational culture on the deployment of agile methods.

  16. Improving Global Development Using Agile

    NASA Astrophysics Data System (ADS)

    Avritzer, Alberto; Bronsard, Francois; Matos, Gilberto

    Global development promises important productivity and capability advantages over centralized work by optimally allocating tasks according to locality, expertise or cost. All too often, global development also introduces a different set of communication and coordination challenges that can negate all the expected benefits and even cause project failures. Most common problems have to do with building trust or quick feedback loops between distributed teams, or with the integration of globally developed components. Agile processes tend to emphasize the intensity of communication, and would seem to be negatively impacted by team distribution. In our experience, these challenges can be overcome, and agile processes can address some of the pitfalls of global development more effectively than plan-driven development. This chapter discusses how to address the difficulties faced when adapting agile processes to global development and the improvements to global development that adopting agile can produce.

  17. The Telemetry Agile Manufacturing Effort

    SciTech Connect

    Brown, K.D.

    1995-01-01

    The Telemetry Agile Manufacturing Effort (TAME) is an agile enterprising demonstration sponsored by the US Department of Energy (DOE). The project experimented with new approaches to product realization and assessed their impacts on performance, cost, flow time, and agility. The purpose of the project was to design the electrical and mechanical features of an integrated telemetry processor, establish the manufacturing processes, and produce an initial production lot of two to six units. This paper outlines the major methodologies utilized by the TAME, describes the accomplishments that can be attributed to each methodology, and finally, examines the lessons learned and explores the opportunities for improvement associated with the overall effort. The areas for improvement are discussed relative to an ideal vision of the future for agile enterprises. By the end of the experiment, the TAME reduced production flow time by approximately 50% and life cycle cost by more than 30%. Product performance was improved compared with conventional DOE production approaches.

  18. Agile manufacturing prototyping system (AMPS)

    SciTech Connect

    Garcia, P.

    1998-05-09

    The Agile Manufacturing Prototyping System (AMPS) is being integrated at Sandia National Laboratories. AMPS consists of state of the industry flexible manufacturing hardware and software enhanced with Sandia advancements in sensor and model based control; automated programming, assembly and task planning; flexible fixturing; and automated reconfiguration technology. AMPS is focused on the agile production of complex electromechanical parts. It currently includes 7 robots (4 Adept One, 2 Adept 505, 1 Staubli RX90), conveyance equipment, and a collection of process equipment to form a flexible production line capable of assembling a wide range of electromechanical products. This system became operational in September 1995. Additional smart manufacturing processes will be integrated in the future. An automated spray cleaning workcell capable of handling alcohol and similar solvents was added in 1996 as well as parts cleaning and encapsulation equipment, automated deburring, and automated vision inspection stations. Plans for 1997 and out years include adding manufacturing processes for the rapid prototyping of electronic components such as soldering, paste dispensing and pick-and-place hardware.

  19. Utilization of an agility assessment module in analysis and optimization of preliminary fighter configuration

    NASA Technical Reports Server (NTRS)

    Ngan, Angelen; Biezad, Daniel

    1996-01-01

    A study has been conducted to develop and to analyze a FORTRAN computer code for performing agility analysis on fighter aircraft configurations. This program is one of the modules of the NASA Ames ACSYNT (AirCraft SYNThesis) design code. The background of the agility research in the aircraft industry and a survey of a few agility metrics are discussed. The methodology, techniques, and models developed for the code are presented. The validity of the existing code was evaluated by comparing with existing flight test data. A FORTRAN program was developed for a specific metric, PM (Pointing Margin), as part of the agility module. Example trade studies using the agility module along with ACSYNT were conducted using a McDonnell Douglas F/A-18 Hornet aircraft model. Tile sensitivity of thrust loading, wing loading, and thrust vectoring on agility criteria were investigated. The module can compare the agility potential between different configurations and has capability to optimize agility performance in the preliminary design process. This research provides a new and useful design tool for analyzing fighter performance during air combat engagements in the preliminary design.

  20. Human factors in agile manufacturing

    SciTech Connect

    Forsythe, C.

    1995-03-01

    As industries position themselves for the competitive markets of today, and the increasingly competitive global markets of the 21st century, agility, or the ability to rapidly develop and produce new products, represents a common trend. Agility manifests itself in many different forms, with the agile manufacturing paradigm proposed by the Iacocca Institute offering a generally accepted, long-term vision. In its many forms, common elements of agility or agile manufacturing include: changes in business, engineering and production practices, seamless information flow from design through production, integration of computer and information technologies into all facets of the product development and production process, application of communications technologies to enable collaborative work between geographically dispersed product development team members and introduction of flexible automation of production processes. Industry has rarely experienced as dramatic an infusion of new technologies or as extensive a change in culture and work practices. Human factors will not only play a vital role in accomplishing the technical and social objectives of agile manufacturing. but has an opportunity to participate in shaping the evolution of industry paradigms for the 21st century.

  1. TriBITS lifecycle model. Version 1.0, a lean/agile software lifecycle model for research-based computational science and engineering and applied mathematical software.

    SciTech Connect

    Willenbring, James M.; Bartlett, Roscoe Ainsworth; Heroux, Michael Allen

    2012-01-01

    Software lifecycles are becoming an increasingly important issue for computational science and engineering (CSE) software. The process by which a piece of CSE software begins life as a set of research requirements and then matures into a trusted high-quality capability is both commonplace and extremely challenging. Although an implicit lifecycle is obviously being used in any effort, the challenges of this process - respecting the competing needs of research vs. production - cannot be overstated. Here we describe a proposal for a well-defined software lifecycle process based on modern Lean/Agile software engineering principles. What we propose is appropriate for many CSE software projects that are initially heavily focused on research but also are expected to eventually produce usable high-quality capabilities. The model is related to TriBITS, a build, integration and testing system, which serves as a strong foundation for this lifecycle model, and aspects of this lifecycle model are ingrained in the TriBITS system. Here, we advocate three to four phases or maturity levels that address the appropriate handling of many issues associated with the transition from research to production software. The goals of this lifecycle model are to better communicate maturity levels with customers and to help to identify and promote Software Engineering (SE) practices that will help to improve productivity and produce better software. An important collection of software in this domain is Trilinos, which is used as the motivation and the initial target for this lifecycle model. However, many other related and similar CSE (and non-CSE) software projects can also make good use of this lifecycle model, especially those that use the TriBITS system. Indeed this lifecycle process, if followed, will enable large-scale sustainable integration of many complex CSE software efforts across several institutions.

  2. CT-assisted agile manufacturing

    NASA Astrophysics Data System (ADS)

    Stanley, James H.; Yancey, Robert N.

    1996-11-01

    The next century will witness at least two great revolutions in the way goods are produced. First, workers will use the medium of virtual reality in all aspects of marketing, research, development, prototyping, manufacturing, sales and service. Second, market forces will drive manufacturing towards small-lot production and just-in-time delivery. Already, we can discern the merging of these megatrends into what some are calling agile manufacturing. Under this new paradigm, parts and processes will be designed and engineered within the mind of a computer, tooled and manufactured by the offspring of today's rapid prototyping equipment, and evaluated for performance and reliability by advanced nondestructive evaluation (NDE) techniques and sophisticated computational models. Computed tomography (CT) is the premier example of an NDE method suitable for future agile manufacturing activities. It is the only modality that provides convenient access to the full suite of engineering data that users will need to avail themselves of computer- aided design, computer-aided manufacturing, and computer- aided engineering capabilities, as well as newly emerging reverse engineering, rapid prototyping and solid freeform fabrication technologies. As such, CT is assured a central, utilitarian role in future industrial operations. An overview of this exciting future for industrial CT is presented.

  3. ESTIMATION OF THE NEUTRINO FLUX AND RESULTING CONSTRAINTS ON HADRONIC EMISSION MODELS FOR Cyg X-3 USING AGILE DATA

    SciTech Connect

    Baerwald, P.; Guetta, D.

    2013-08-20

    In this work, we give an estimate of the neutrino flux that can be expected from the microquasar Cyg X-3. We calculate the muon neutrino flux expected here on Earth as well as the corresponding number of neutrino events in the IceCube telescope based on the so-called hypersoft X-ray state of Cyg X-3. If the average emission from Cyg X-3 over a period of 5 yr were as high as during the used X-ray state, a total of 0.8 events should be observed by the full IceCube telescope. We also show that this conclusion holds by a factor of a few when we consider the other measured X-ray states. Using the correlation of AGILE data on the flaring episodes in 2009 June and July to the hypersoft X-ray state, we calculate that the upper limits on the neutrino flux given by IceCube are starting to constrain the hadronic models, which have been introduced to interpret the high-energy emission detected by AGILE.

  4. Aircraft agility maneuvers

    NASA Technical Reports Server (NTRS)

    Cliff, Eugene M.; Thompson, Brian G.

    1992-01-01

    A new dynamic model for aircraft motions is presented. This model can be viewed as intermediate between a point-mass model, in which the body attitude angles are control-like, and a rigid-body model, in which the body-attitude angles evolve according to Newton's Laws. Specifically, consideration is given to the case of symmetric flight, and a model is constructed in which the body roll-rate and the body pitch-rate are the controls. In terms of this body-rate model a minimum-time heading change maneuver is formulated. When the bounds on the body-rates are large the results are similar to the point-mass model in that the model can very quickly change the applied forces and produce an acceleration to turn the vehicle. With finite bounds on these rates, the forces change in a smooth way. This leads to a measurable effect of agility.

  5. Some Findings Concerning Requirements in Agile Methodologies

    NASA Astrophysics Data System (ADS)

    Rodríguez, Pilar; Yagüe, Agustín; Alarcón, Pedro P.; Garbajosa, Juan

    Agile methods have appeared as an attractive alternative to conventional methodologies. These methods try to reduce the time to market and, indirectly, the cost of the product through flexible development and deep customer involvement. The processes related to requirements have been extensively studied in literature, in most cases in the frame of conventional methods. However, conclusions of conventional methodologies could not be necessarily valid for Agile; in some issues, conventional and Agile processes are radically different. As recent surveys report, inadequate project requirements is one of the most conflictive issues in agile approaches and better understanding about this is needed. This paper describes some findings concerning requirements activities in a project developed under an agile methodology. The project intended to evolve an existing product and, therefore, some background information was available. The major difficulties encountered were related to non-functional needs and management of requirements dependencies.

  6. Implementing Kanban for agile process management within the ALMA Software Operations Group

    NASA Astrophysics Data System (ADS)

    Reveco, Johnny; Mora, Matias; Shen, Tzu-Chiang; Soto, Ruben; Sepulveda, Jorge; Ibsen, Jorge

    2014-07-01

    After the inauguration of the Atacama Large Millimeter/submillimeter Array (ALMA), the Software Operations Group in Chile has refocused its objectives to: (1) providing software support to tasks related to System Integration, Scientific Commissioning and Verification, as well as Early Science observations; (2) testing the remaining software features, still under development by the Integrated Computing Team across the world; and (3) designing and developing processes to optimize and increase the level of automation of operational tasks. Due to their different stakeholders, each of these tasks presents a wide diversity of importances, lifespans and complexities. Aiming to provide the proper priority and traceability for every task without stressing our engineers, we introduced the Kanban methodology in our processes in order to balance the demand on the team against the throughput of the delivered work. The aim of this paper is to share experiences gained during the implementation of Kanban in our processes, describing the difficulties we have found, solutions and adaptations that led us to our current but still evolving implementation, which has greatly improved our throughput, prioritization and problem traceability.

  7. Agile manufacturing from a statistical perspective

    SciTech Connect

    Easterling, R.G.

    1995-10-01

    The objective of agile manufacturing is to provide the ability to quickly realize high-quality, highly-customized, in-demand products at a cost commensurate with mass production. More broadly, agility in manufacturing, or any other endeavor, is defined as change-proficiency; the ability to thrive in an environment of unpredictable change. This report discusses the general direction of the agile manufacturing initiative, including research programs at the National Institute of Standards and Technology (NIST), the Department of Energy, and other government agencies, but focuses on agile manufacturing from a statistical perspective. The role of statistics can be important because agile manufacturing requires the collection and communication of process characterization and capability information, much of which will be data-based. The statistical community should initiate collaborative work in this important area.

  8. Social Protocols for Agile Virtual Teams

    NASA Astrophysics Data System (ADS)

    Picard, Willy

    Despite many works on collaborative networked organizations (CNOs), CSCW, groupware, workflow systems and social networks, computer support for virtual teams is still insufficient, especially support for agility, i.e. the capability of virtual team members to rapidly and cost efficiently adapt the way they interact to changes. In this paper, requirements for computer support for agile virtual teams are presented. Next, an extension of the concept of social protocol is proposed as a novel model supporting agile interactions within virtual teams. The extended concept of social protocol consists of an extended social network and a workflow model.

  9. Agile Software Development

    ERIC Educational Resources Information Center

    Biju, Soly Mathew

    2008-01-01

    Many software development firms are now adopting the agile software development method. This method involves the customer at every level of software development, thus reducing the impact of change in the requirement at a later stage. In this article, the principles of the agile method for software development are explored and there is a focus on…

  10. Research on modeling of the agile satellite using a single gimbal magnetically suspended CMG and the disturbance feedforward compensation for rotors.

    PubMed

    Cui, Peiling; Yan, Ning

    2012-01-01

    The magnetically suspended Control Moment Gyroscope (CMG) has the advantages of long-life, micro-vibration and being non-lubricating, and is the ideal actuator for agile maneuver satellite attitude control. However, the stability of the rotor in magnetic bearing and the precision of the output torque of a magnetically suspended CMG are affected by the rapid maneuvers of satellites. In this paper, a dynamic model of the agile satellite including a magnetically suspended single gimbal control moment gyroscope is built and the equivalent disturbance torque effected on the rotor is obtained. The feedforward compensation control method is used to depress the disturbance on the rotor. Simulation results are given to show that the rotor displacement is obviously reduced.

  11. Research on Modeling of the Agile Satellite Using a Single Gimbal Magnetically Suspended CMG and the Disturbance Feedforward Compensation for Rotors

    PubMed Central

    Cui, Peiling; Yan, Ning

    2012-01-01

    The magnetically suspended Control Moment Gyroscope (CMG) has the advantages of long-life, micro-vibration and being non-lubricating, and is the ideal actuator for agile maneuver satellite attitude control. However, the stability of the rotor in magnetic bearing and the precision of the output torque of a magnetically suspended CMG are affected by the rapid maneuvers of satellites. In this paper, a dynamic model of the agile satellite including a magnetically suspended single gimbal control moment gyroscope is built and the equivalent disturbance torque effected on the rotor is obtained. The feedforward compensation control method is used to depress the disturbance on the rotor. Simulation results are given to show that the rotor displacement is obviously reduced. PMID:23235442

  12. Integrating a distributed, agile, virtual enterprise in the TEAM program

    NASA Astrophysics Data System (ADS)

    Cobb, C. K.; Gray, W. Harvey; Hewgley, Robert E.; Klages, Edward J.; Neal, Richard E.

    1997-01-01

    The technologies enabling agile manufacturing (TEAM) program enhances industrial capability by advancing and deploying manufacturing technologies that promote agility. TEAM has developed a product realization process that features the integration of product design and manufacturing groups. TEAM uses the tools it collects, develops, and integrates in support of the product realization process to demonstrate and deploy agile manufacturing capabilities for three high- priority processes identified by industry: material removal, forming, and electromechanical assembly. In order to provide a proof-of-principle, the material removal process has been addressed first and has been successfully demonstrate din an 'interconnected' mode. An internet-accessible intersite file manager (IFM) application has been deployed to allow geographically distributed TEAM participants to share and distribute information as the product realization process is executed. An automated inspection planning application has been demonstrated, importing a solid model form the IFM, generating an inspection plan and a part program to be used in the inspection process, and then distributing the part program to the inspection site via the IFM. TEAM seeks to demonstrate the material removal process in an integrated mode in June 1997 complete with an object-oriented framework and infrastructure. The current status and future plans for this project are presented here.

  13. Production planning tools and techniques for agile manufacturing

    SciTech Connect

    Kjeldgaard, E.A.; Jones, D.A.; List, G.F.; Turnquist, M.A.

    1996-10-01

    Effective use of resources shared among multiple products or processes is critical for agile manufacturing. This paper describes development and implementation of a computerized model to support production planning in a complex manufacturing system at Pantex Plant. The model integrates two different production processes (nuclear weapon dismantlement and stockpile evaluation) which use common facilities and personnel, and reflects the interactions of scheduling constraints, material flow constraints, and resource availability. These two processes reflect characteristics of flow-shop and job-shop operations in a single facility. Operational results from using the model are also discussed.

  14. An agile implementation of SCRUM

    NASA Astrophysics Data System (ADS)

    Gannon, Michele

    Is Agile a way to cut corners? To some, the use of an Agile Software Development Methodology has a negative connotation - “ Oh, you're just not producing any documentation” . So can a team with no experience in Agile successfully implement and use SCRUM?

  15. Strategic agility for nursing leadership.

    PubMed

    Shirey, Maria R

    2015-06-01

    This department highlights change management strategies that may be successful in strategically planning and executing organizational change. In this article, the author discusses strategic agility as an important leadership competency and offers approaches for incorporating strategic agility in healthcare systems. A strategic agility checklist and infrastructure-building approach are presented. PMID:26010278

  16. Strategic agility for nursing leadership.

    PubMed

    Shirey, Maria R

    2015-06-01

    This department highlights change management strategies that may be successful in strategically planning and executing organizational change. In this article, the author discusses strategic agility as an important leadership competency and offers approaches for incorporating strategic agility in healthcare systems. A strategic agility checklist and infrastructure-building approach are presented.

  17. A Quantitative Examination of Critical Success Factors Comparing Agile and Waterfall Project Management Methodologies

    ERIC Educational Resources Information Center

    Pedersen, Mitra

    2013-01-01

    This study investigated the rate of success for IT projects using agile and standard project management methodologies. Any successful project requires use of project methodology. Specifically, large projects require formal project management methodologies or models, which establish a blueprint of processes and project planning activities. This…

  18. Development of a Computer Program for Analyzing Preliminary Aircraft Configurations in Relationship to Emerging Agility Metrics

    NASA Technical Reports Server (NTRS)

    Bauer, Brent

    1993-01-01

    This paper discusses the development of a FORTRAN computer code to perform agility analysis on aircraft configurations. This code is to be part of the NASA-Ames ACSYNT (AirCraft SYNThesis) design code. This paper begins with a discussion of contemporary agility research in the aircraft industry and a survey of a few agility metrics. The methodology, techniques and models developed for the code are then presented. Finally, example trade studies using the agility module along with ACSYNT are illustrated. These trade studies were conducted using a Northrop F-20 Tigershark aircraft model. The studies show that the agility module is effective in analyzing the influence of common parameters such as thrust-to-weight ratio and wing loading on agility criteria. The module can compare the agility potential between different configurations. In addition, one study illustrates the module's ability to optimize a configuration's agility performance.

  19. Analysis and optimization of preliminary aircraft configurations in relationship to emerging agility metrics

    NASA Technical Reports Server (NTRS)

    Sandlin, Doral R.; Bauer, Brent Alan

    1993-01-01

    This paper discusses the development of a FORTRAN computer code to perform agility analysis on aircraft configurations. This code is to be part of the NASA-Ames ACSYNT (AirCraft SYNThesis) design code. This paper begins with a discussion of contemporary agility research in the aircraft industry and a survey of a few agility metrics. The methodology, techniques and models developed for the code are then presented. Finally, example trade studies using the agility module along with ACSYNT are illustrated. These trade studies were conducted using a Northrop F-20 Tigershark aircraft model. The studies show that the agility module is effective in analyzing the influence of common parameters such as thrust-to-weight ratio and wing loading on agility criteria. The module can compare the agility potential between different configurations. In addition one study illustrates the module's ability to optimize a configuration's agility performance.

  20. Creating IT agility.

    PubMed

    Glaser, John

    2008-04-01

    Seven steps healthcare organizations can take to improve IT agility are: Pay attention to the capabilities of IT applications. Establish short project phases. Stage the release of capital and new IT positions. Cross-train IT staff. Adopt technology standards. Shorten IT plan time horizons. Align IT with organizational strategies and priorities.

  1. Agile robotic edge finishing system research

    SciTech Connect

    Powell, M.A.

    1995-07-01

    This paper describes a new project undertaken by Sandia National Laboratories to develop an agile, automated, high-precision edge finishing system. The project has a two-year duration and was initiated in October, 1994. This project involves re-designing and adding additional capabilities to an existing finishing workcell at Sandia; and developing intelligent methods for automating process definition and for controlling finishing processes. The resulting system will serve as a prototype for systems that will be deployed into highly flexible automated production lines. The production systems will be used to produce a wide variety of products with limited production quantities and quick turnaround requirements. The prototype system is designed to allow programming, process definition, fixture re-configuration, and process verification to be performed off-line for new products. CAD/CAM (Computer Aided Design/Computer Aided Manufacturing) models of the part will be used to assist with the automated process development and process control tasks. To achieve Sandia`s performance goals, the system will be employ advanced path planning, burr prediction expert systems, automated process definition, statistical process models in a process database, and a two-level control scheme using hybrid position-force control and fuzzy logic control. In this paper, we discuss the progress and the planned system development under this project.

  2. 5th Annual AGILE Science Workshop

    NASA Technical Reports Server (NTRS)

    Hunter, Stanley

    2008-01-01

    The EGRET model of the galactic diffuse gamma-ray emission (GALDIF) has been extended to provide full-sky coverage and improved to address the discrepancies with the EGRET data. This improved model is compared with the AGILE results from the Galactic center. The comparison is discussed.

  3. A Roadmap for Using Agile Development in a Traditional Environment

    NASA Technical Reports Server (NTRS)

    Streiffert, Barbara; Starbird, Thomas; Grenander, Sven

    2006-01-01

    One of the newer classes of software engineering techniques is called 'Agile Development'. In Agile Development software engineers take small implementation steps and, in some cases, they program in pairs. In addition, they develop automatic tests prior to implementing their small functional piece. Agile Development focuses on rapid turnaround, incremental planning, customer involvement and continuous integration. Agile Development is not the traditional waterfall method or even a rapid prototyping method (although this methodology is closer to Agile Development). At the Jet Propulsion Laboratory (JPL) a few groups have begun Agile Development software implementations. The difficulty with this approach becomes apparent when Agile Development is used in an organization that has specific criteria and requirements handed down for how software development is to be performed. The work at the JPL is performed for the National Aeronautics and Space Agency (NASA). Both organizations have specific requirements, rules and processes for developing software. This paper will discuss some of the initial uses of the Agile Development methodology, the spread of this method and the current status of the successful incorporation into the current JPL development policies and processes.

  4. Peridigm summary report : lessons learned in development with agile components.

    SciTech Connect

    Salinger, Andrew Gerhard; Mitchell, John Anthony; Littlewood, David John; Parks, Michael L.

    2011-09-01

    This report details efforts to deploy Agile Components for rapid development of a peridynamics code, Peridigm. The goal of Agile Components is to enable the efficient development of production-quality software by providing a well-defined, unifying interface to a powerful set of component-based software. Specifically, Agile Components facilitate interoperability among packages within the Trilinos Project, including data management, time integration, uncertainty quantification, and optimization. Development of the Peridigm code served as a testbed for Agile Components and resulted in a number of recommendations for future development. Agile Components successfully enabled rapid integration of Trilinos packages into Peridigm. A cost of this approach, however, was a set of restrictions on Peridigm's architecture which impacted the ability to track history-dependent material data, dynamically modify the model discretization, and interject user-defined routines into the time integration algorithm. These restrictions resulted in modifications to the Agile Components approach, as implemented in Peridigm, and in a set of recommendations for future Agile Components development. Specific recommendations include improved handling of material states, a more flexible flow control model, and improved documentation. A demonstration mini-application, SimpleODE, was developed at the onset of this project and is offered as a potential supplement to Agile Components documentation.

  5. Optical flows method for lightweight agile remote sensor design and instrumentation

    NASA Astrophysics Data System (ADS)

    Wang, Chong; Xing, Fei; Wang, Hongjian; You, Zheng

    2013-08-01

    Lightweight agile remote sensors have become one type of the most important payloads and were widely utilized in space reconnaissance and resource survey. These imaging sensors are designed to obtain the high spatial, temporary and spectral resolution imageries. Key techniques in instrumentation include flexible maneuvering, advanced imaging control algorithms and integrative measuring techniques, which are closely correlative or even acting as the bottle-necks for each other. Therefore, mutual restrictive problems must be solved and optimized. Optical flow is the critical model which to be fully represented in the information transferring as well as radiation energy flowing in dynamic imaging. For agile sensors, especially with wide-field-of view, imaging optical flows may distort and deviate seriously when they perform large angle attitude maneuvering imaging. The phenomena are mainly attributed to the geometrical characteristics of the three-dimensional earth surface as well as the coupled effects due to the complicated relative motion between the sensor and scene. Under this circumstance, velocity fields distribute nonlinearly, the imageries may badly be smeared or probably the geometrical structures are changed since the image velocity matching errors are not having been eliminated perfectly. In this paper, precise imaging optical flow model is established for agile remote sensors, for which optical flows evolving is factorized by two forms, which respectively due to translational movement and image shape changing. Moreover, base on that, agile remote sensors instrumentation was investigated. The main techniques which concern optical flow modeling include integrative design with lightweight star sensors along with micro inertial measurement units and corresponding data fusion, the assemblies of focal plane layout and control, imageries post processing for agile remote sensors etc. Some experiments show that the optical analyzing method is effective to

  6. Introduction to Stand-up Meetings in Agile Methods

    NASA Astrophysics Data System (ADS)

    Hasnain, Eisha; Hall, Tracy

    2009-05-01

    In recent years, agile methods have become more popular in the software industry. Agile methods are a new approach compared to plan-driven approaches. One of the most important shifts in adopting an agile approach is the central focus given to people in the process. This is exemplified by the independence afforded to developers in the development work they do. This work investigates the opinions of practitioners about daily stand-up meetings in the agile methods and the role of developer in that. For our investigation we joined a yahoo group called "Extreme Programming". Our investigation suggests that although trust is an important factor in agile methods. But stand-ups are not the place to build trust.

  7. Supporting Agile Development of Authorization Rules for SME Applications

    NASA Astrophysics Data System (ADS)

    Bartsch, Steffen; Sohr, Karsten; Bormann, Carsten

    Custom SME applications for collaboration and workflow have become affordable when implemented as Web applications employing Agile methodologies. Security engineering is still difficult with Agile development, though: heavy-weight processes put the improvements of Agile development at risk. We propose Agile security engineering and increased end-user involvement to improve Agile development with respect to authorization policy development. To support the authorization policy development, we introduce a simple and readable authorization rules language implemented in a Ruby on Rails authorization plugin that is employed in a real-world SME collaboration and workflow application. Also, we report on early findings of the language’s use in authorization policy development with domain experts.

  8. Achieving agility through parameter space qualification

    SciTech Connect

    Diegert, K.V.; Easterling, R.G.; Ashby, M.R.; Benavides, G.L.; Forsythe, C.; Jones, R.E.; Longcope, D.B.; Parratt, S.W.

    1995-02-01

    The A-primed (Agile Product Realization of Innovative electro-Mechanical Devices) project is defining and proving processes for agile product realization for the Department of Energy complex. Like other agile production efforts reported in the literature, A-primed uses concurrent engineering and information automation technologies to enhance information transfer. A unique aspect of our approach to agility is the qualification during development of a family of related product designs and their production processes, rather than a single design and its attendant processes. Applying engineering principles and statistical design of experiments, economies of test and analytic effort are realized for the qualification of the device family as a whole. Thus the need is minimized for test and analysis to qualify future devices from this family, thereby further reducing the design-to-production cycle time. As a measure of the success of the A-primed approach, the first design took 24 days to produce, and operated correctly on the first attempt. A flow diagram for the qualification process is presented. Guidelines are given for implementation, based on the authors experiences as members of the A-primed qualification team.

  9. An investigation of fighter aircraft agility

    NASA Technical Reports Server (NTRS)

    Valasek, John; Downing, David R.

    1993-01-01

    of how to test and measure the metric, including any special data reduction requirements; typical values for the metric obtained using one or more aircraft types; and a sensitivity analysis if applicable. The report is organized as follows. The first chapter in the report presents a historical review of air combat trends which demonstrate the need for agility metrics in assessing the combat performance of fighter aircraft in a modern, all-aspect missile environment. The second chapter presents a framework for classifying each candidate metric according to time scale (transient, functional, instantaneous), further subdivided by axis (pitch, lateral, axial). The report is then broadly divided into two parts, with the transient agility metrics (pitch lateral, axial) covered in chapters three, four, and five, and the functional agility metrics covered in chapter six. Conclusions, recommendations, and an extensive reference list and biography are also included. Five appendices contain a comprehensive list of the definitions of all the candidate metrics; a description of the aircraft models and flight simulation programs used for testing the metrics; several relations and concepts which are fundamental to the study of lateral agility; an in-depth analysis of the axial agility metrics; and a derivation of the relations for the instantaneous agility and their approximations.

  10. Perspectives on Agile Coaching

    NASA Astrophysics Data System (ADS)

    Fraser, Steven; Lundh, Erik; Davies, Rachel; Eckstein, Jutta; Larsen, Diana; Vilkki, Kati

    There are many perspectives to agile coaching including: growing coaching expertise, selecting the appropriate coach for your context; and eva luating value. A coach is often an itinerant who may observe, mentor, negotiate, influence, lead, and/or architect everything from team organization to system architecture. With roots in diverse fields ranging from technology to sociology coaches have differing motivations and experience bases. This panel will bring together coaches to debate and discuss various perspectives on agile coaching. Some of the questions to be addressed will include: What are the skills required for effective coaching? What should be the expectations for teams or individu als being coached? Should coaches be: a corporate resource (internal team of consultants working with multiple internal teams); an integral part of a specific team; or external contractors? How should coaches exercise influence and au thority? How should management assess the value of a coaching engagement? Do you have what it takes to be a coach? - This panel will bring together sea soned agile coaches to offer their experience and advice on how to be the best you can be!

  11. Pinnacle3 modeling and end-to-end dosimetric testing of a Versa HD linear accelerator with the Agility head and flattening filter-free modes.

    PubMed

    Saenz, Daniel L; Narayanasamy, Ganesh; Cruz, Wilbert; Papanikolaou, Nikos; Stathakis, Sotirios

    2016-01-01

    The Elekta Versa HD incorporates a variety of upgrades to the line of Elekta linear accelerators, primarily including the Agility head and flattening filter-free (FFF) photon beam delivery. The completely distinct dosimetric output of the head from its predecessors, combined with the FFF beams, requires a new investigation of modeling in treatment planning systems. A model was created in Pinnacle3 v9.8 with the commissioned beam data. A phantom consisting of several plastic water and Styrofoam slabs was scanned and imported into Pinnacle3, where beams of different field sizes, source-to-surface distances (SSDs), wedges, and gantry angles were devised. Beams included all of the available photon energies (6, 10, 18, 6FFF, and 10 FFF MV), as well as the four electron energies commissioned for clinical use (6, 9, 12, and 15 MeV). The plans were verified at calculation points by measurement with a calibrated ionization chamber. Homogeneous and hetero-geneous point-dose measurements agreed within 2% relative to maximum dose for all photon and electron beams. AP photon open field measurements along the central axis at 100 cm SSD passed within 1%. In addition, IMRT testing was also performed with three standard plans (step and shoot IMRT, as well as a small- and large-field VMAT plan). The IMRT plans were delivered on the Delta4 IMRT QA phantom, for which a gamma passing rate was > 99.5% for all plans with a 3% dose deviation, 3 mm distance-to-agreement, and 10% dose threshold. The IMRT QA results for the first 23 patients yielded gamma passing rates of 97.4% ± 2.3%. Such testing ensures confidence in the ability of Pinnacle3 to model photon and electron beams with the Agility head. PMID:26894352

  12. Response of the Italian agile frog (Rana latastei) to a Ranavirus, frog virus 3: a model for viral emergence in naïve populations.

    PubMed

    Pearman, Peter B; Garner, Trenton W J; Straub, Monika; Greber, Urs F

    2004-10-01

    Ranavirus (family Iridoviridae) is a genus of pathogens of poikilotherms, and some ranaviruses may play a role in widespread mortality of amphibians. Ecology of viral transmission in amphibians is poorly known but can be addressed through experimentation in the laboratory. In this study, we use the Ranavirus frog virus 3 (FV3) as an experimental model for pathogen emergence in naive populations of tadpoles. We simulated emerging disease by exposing tadpoles of the Italian agile frog (Rana latastei), to the North American Ranavirus FV3. We demonstrated that mortality occurred due to viral exposure, exposure of tadpoles to decreasing concentrations of FV3 in the laboratory produced dose-dependent survival rates, and cannibalism of virus-carrying carcasses increased mortality due to FV3. These experiments suggest the potential for ecological mechanisms to affect the level of exposure of tadpoles to Ranavirus and to impact transmission of viral pathogens in aquatic systems.

  13. Modelling fire-fighter responses to exercise and asymmetric infrared radiation using a dynamic multi-mode model of human physiology and results from the sweating agile thermal manikin.

    PubMed

    Richards, M G M; Fiala, D

    2004-09-01

    In this study, predicted dynamic physiological responses are compared with wear trials results for firefighter suits: impermeable (A), semi-permeable (B) and permeable (C), and underwear. Wear trials consisted of three rest phases and two moderate work phases, with a frontal infrared (IR) radiation exposure of 500 W/m2 for the last 15 min of each work phase. Simulations were performed by detailed modelling of the experimental boundary conditions, including the inhomogeneous IR radiation combined with clothing properties for still and walking conditions measured using the Sweating Agile thermal Manikin. Accounting for the effect of sweat gland activity suppression with increased skin wettedness, the predicted total moisture loss was insignificantly different (P<0.05) from the wear trial value for suits B and C but was 37% too high for suit A. Predicted evolution of core, mean skin and local skin temperatures agreed well with the wear trial results for all clothing. Root mean square deviations ranged from 0.11 degrees C to 0.26 degrees C for core temperatures and from 0.28 degrees C to 0.38 degrees C for mean skin temperatures, which where typically lower than the experimental error. Transient thermodynamic processes occurring within suit A may account for the delayed/reduced fall in core temperature following exercise.

  14. Aging contributes to inflammation in upper extremity tendons and declines in forelimb agility in a rat model of upper extremity overuse.

    PubMed

    Kietrys, David M; Barr-Gillespie, Ann E; Amin, Mamta; Wade, Christine K; Popoff, Steve N; Barbe, Mary F

    2012-01-01

    We sought to determine if tendon inflammatory and histopathological responses increase in aged rats compared to young rats performing a voluntary upper extremity repetitive task, and if these changes are associated with motor declines. Ninety-six female Sprague-Dawley rats were used in the rat model of upper extremity overuse: 67 aged and 29 young adult rats. After a training period of 4 weeks, task rats performed a voluntary high repetition low force (HRLF) handle-pulling task for 2 hrs/day, 3 days/wk for up to 12 weeks. Upper extremity motor function was assessed, as were inflammatory and histomorphological changes in flexor digitorum and supraspinatus tendons. The percentage of successful reaches improved in young adult HRLF rats, but not in aged HRLF rats. Forelimb agility decreased transiently in young adult HRLF rats, but persistently in aged HRLF rats. HRLF task performance for 12 weeks lead to increased IL-1beta and IL-6 in flexor digitorum tendons of aged HRLF rats, compared to aged normal control (NC) as well as young adult HRLF rats. In contrast, TNF-alpha increased more in flexor digitorum tendons of young adult 12-week HRLF rats than in aged HRLF rats. Vascularity and collagen fibril organization were not affected by task performance in flexor digitorum tendons of either age group, although cellularity increased in both. By week 12 of HRLF task performance, vascularity and cellularity increased in the supraspinatus tendons of only aged rats. The increased cellularity was due to increased macrophages and connective tissue growth factor (CTGF)-immunoreactive fibroblasts in the peritendon. In conclusion, aged rat tendons were overall more affected by the HRLF task than young adult tendons, particularly supraspinatus tendons. Greater inflammatory changes in aged HRLF rat tendons were observed, increases associated temporally with decreased forelimb agility and lack of improvement in task success.

  15. Supply chain network design problem for a new market opportunity in an agile manufacturing system

    NASA Astrophysics Data System (ADS)

    Babazadeh, Reza; Razmi, Jafar; Ghodsi, Reza

    2012-08-01

    The characteristics of today's competitive environment, such as the speed with which products are designed, manufactured, and distributed, and the need for higher responsiveness and lower operational cost, are forcing companies to search for innovative ways to do business. The concept of agile manufacturing has been proposed in response to these challenges for companies. This paper copes with the strategic and tactical level decisions in agile supply chain network design. An efficient mixed-integer linear programming model that is able to consider the key characteristics of agile supply chain such as direct shipments, outsourcing, different transportation modes, discount, alliance (process and information integration) between opened facilities, and maximum waiting time of customers for deliveries is developed. In addition, in the proposed model, the capacity of facilities is determined as decision variables, which are often assumed to be fixed. Computational results illustrate that the proposed model can be applied as a power tool in agile supply chain network design as well as in the integration of strategic decisions with tactical decisions.

  16. Agile green process design for the intensified Kolbe-Schmitt synthesis by accompanying (simplified) life cycle assessment.

    PubMed

    Kressirer, Sabine; Kralisch, Dana; Stark, Annegret; Krtschil, Ulrich; Hessel, Volker

    2013-05-21

    In order to investigate the potential for process intensification, various reaction conditions were applied to the Kolbe-Schmitt synthesis starting from resorcinol. Different CO₂ precursors such as aqueous potassium hydrogencarbonate, hydrogencarbonate-based ionic liquids, DIMCARB, or sc-CO₂, the application of microwave irradiation for fast volumetric heating of the reaction mixture, and the effect of harsh reaction conditions were investigated. The experiments, carried out in conventional batch-wise as well as in continuously operated microstructured reactors, aimed at the development of an environmentally benign process for the preparation of 2,4-dihydroxybenzoic acid. To provide decision support toward a green process design, a research-accompanying simplified life cycle assessment (SLCA) was performed throughout the whole investigation. Following this approach, it was found that convective heating methods such as oil bath or electrical heating were more beneficial than the application of microwave irradiation. Furthermore, the consideration of workup procedures was crucial for a holistic view on the environmental burdens.

  17. Agile Infrastructure Monitoring

    NASA Astrophysics Data System (ADS)

    Andrade, P.; Ascenso, J.; Fedorko, I.; Fiorini, B.; Paladin, M.; Pigueiras, L.; Santos, M.

    2014-06-01

    At the present time, data centres are facing a massive rise in virtualisation and cloud computing. The Agile Infrastructure (AI) project is working to deliver new solutions to ease the management of CERN data centres. Part of the solution consists in a new "shared monitoring architecture" which collects and manages monitoring data from all data centre resources. In this article, we present the building blocks of this new monitoring architecture, the different open source technologies selected for each architecture layer, and how we are building a community around this common effort.

  18. Agile Walking Robot

    NASA Technical Reports Server (NTRS)

    Larimer, Stanley J.; Lisec, Thomas R.; Spiessbach, Andrew J.; Waldron, Kenneth J.

    1990-01-01

    Proposed agile walking robot operates over rocky, sandy, and sloping terrain. Offers stability and climbing ability superior to other conceptual mobile robots. Equipped with six articulated legs like those of insect, continually feels ground under leg before applying weight to it. If leg sensed unexpected object or failed to make contact with ground at expected point, seeks alternative position within radius of 20 cm. Failing that, robot halts, examines area around foot in detail with laser ranging imager, and replans entire cycle of steps for all legs before proceeding.

  19. Moving target detection for frequency agility radar by sparse reconstruction

    NASA Astrophysics Data System (ADS)

    Quan, Yinghui; Li, YaChao; Wu, Yaojun; Ran, Lei; Xing, Mengdao; Liu, Mengqi

    2016-09-01

    Frequency agility radar, with randomly varied carrier frequency from pulse to pulse, exhibits superior performance compared to the conventional fixed carrier frequency pulse-Doppler radar against the electromagnetic interference. A novel moving target detection (MTD) method is proposed for the estimation of the target's velocity of frequency agility radar based on pulses within a coherent processing interval by using sparse reconstruction. Hardware implementation of orthogonal matching pursuit algorithm is executed on Xilinx Virtex-7 Field Programmable Gata Array (FPGA) to perform sparse optimization. Finally, a series of experiments are performed to evaluate the performance of proposed MTD method for frequency agility radar systems.

  20. An Approach for Prioritizing Agile Practices for Adaptation

    NASA Astrophysics Data System (ADS)

    Mikulenas, Gytenis; Kapocius, Kestutis

    Agile software development approaches offer a strong alternative to the traditional plan-driven methodologies that have not been able to warrant successfulness of the software projects. However, the move toward Agile is often hampered by the wealth of alternative practices that are accompanied by numerous success or failure stories. Clearly, the formal methods for choosing most suitable practices are lacking. In this chapter, we present an overview of this problem and propose an approach for prioritization of available practices in accordance to the particular circumstances. The proposal combines ideas from Analytic Hierarchy Process (AHP) decision-making technique, cost-value analysis, and Rule-Description-Practice (RDP) technique. Assumption that such approach could facilitate the Agile adaptation process was supported by the case study of the approach illustrating the process of choosing most suitable Agile practices within a real-life project.

  1. A Rule-Based Modeling for the Description of Flexible and Self-healing Business Processes

    NASA Astrophysics Data System (ADS)

    Boukhebouze, Mohamed; Amghar, Youssef; Benharkat, Aïcha-Nabila; Maamar, Zakaria

    In this paper we discuss the importance of ensuring that business processes are label robust and agile at the same time robust and agile. To this end, we consider reviewing the way business processes are managed. For instance we consider offering a flexible way to model processes so that changes in regulations are handled through some self-healing mechanisms. These changes may raise exceptions at run-time if not properly reflected on these processes. To this end we propose a new rule based model that adopts the ECA rules and is built upon formal tools. The business logic of a process can be summarized with a set of rules that implement an organization’s policies. Each business rule is formalized using our ECAPE formalism (Event-Condition-Action-Post condition- post Event). This formalism allows translating a process into a graph of rules that is analyzed in terms of reliably and flexibility.

  2. Frequency agile optical parametric oscillator

    DOEpatents

    Velsko, S.P.

    1998-11-24

    The frequency agile OPO device converts a fixed wavelength pump laser beam to arbitrary wavelengths within a specified range with pulse to pulse agility, at a rate limited only by the repetition rate of the pump laser. Uses of this invention include Laser radar, LIDAR, active remote sensing of effluents/pollutants, environmental monitoring, antisensor lasers, and spectroscopy. 14 figs.

  3. Frequency agile optical parametric oscillator

    DOEpatents

    Velsko, Stephan P.

    1998-01-01

    The frequency agile OPO device converts a fixed wavelength pump laser beam to arbitrary wavelengths within a specified range with pulse to pulse agility, at a rate limited only by the repetition rate of the pump laser. Uses of this invention include Laser radar, LIDAR, active remote sensing of effluents/pollutants, environmental monitoring, antisensor lasers, and spectroscopy.

  4. Final Report of the NASA Office of Safety and Mission Assurance Agile Benchmarking Team

    NASA Technical Reports Server (NTRS)

    Wetherholt, Martha

    2016-01-01

    To ensure that the NASA Safety and Mission Assurance (SMA) community remains in a position to perform reliable Software Assurance (SA) on NASAs critical software (SW) systems with the software industry rapidly transitioning from waterfall to Agile processes, Terry Wilcutt, Chief, Safety and Mission Assurance, Office of Safety and Mission Assurance (OSMA) established the Agile Benchmarking Team (ABT). The Team's tasks were: 1. Research background literature on current Agile processes, 2. Perform benchmark activities with other organizations that are involved in software Agile processes to determine best practices, 3. Collect information on Agile-developed systems to enable improvements to the current NASA standards and processes to enhance their ability to perform reliable software assurance on NASA Agile-developed systems, 4. Suggest additional guidance and recommendations for updates to those standards and processes, as needed. The ABT's findings and recommendations for software management, engineering and software assurance are addressed herein.

  5. PDS4 - Some Principles for Agile Data Curation

    NASA Astrophysics Data System (ADS)

    Hughes, J. S.; Crichton, D. J.; Hardman, S. H.; Joyner, R.; Algermissen, S.; Padams, J.

    2015-12-01

    PDS4, a research data management and curation system for NASA's Planetary Science Archive, was developed using principles that promote the characteristics of agile development. The result is an efficient system that produces better research data products while using less resources (time, effort, and money) and maximizes their usefulness for current and future scientists. The key principle is architectural. The PDS4 information architecture is developed and maintained independent of the infrastructure's process, application and technology architectures. The information architecture is based on an ontology-based information model developed to leverage best practices from standard reference models for digital archives, digital object registries, and metadata registries and capture domain knowledge from a panel of planetary science domain experts. The information model provides a sharable, stable, and formal set of information requirements for the system and is the primary source for information to configure most system components, including the product registry, search engine, validation and display tools, and production pipelines. Multi-level governance is also allowed for the effective management of the informational elements at the common, discipline, and project level. This presentation will describe the development principles, components, and uses of the information model and how an information model-driven architecture exhibits characteristics of agile curation including early delivery, evolutionary development, adaptive planning, continuous improvement, and rapid and flexible response to change.

  6. Agile manufacturing: The factory of the future

    NASA Technical Reports Server (NTRS)

    Loibl, Joseph M.; Bossieux, Terry A.

    1994-01-01

    The factory of the future will require an operating methodology which effectively utilizes all of the elements of product design, manufacturing and delivery. The process must respond rapidly to changes in product demand, product mix, design changes or changes in the raw materials. To achieve agility in a manufacturing operation, the design and development of the manufacturing processes must focus on customer satisfaction. Achieving greatest results requires that the manufacturing process be considered from product concept through sales. This provides the best opportunity to build a quality product for the customer at a reasonable rate. The primary elements of a manufacturing system include people, equipment, materials, methods and the environment. The most significant and most agile element in any process is the human resource. Only with a highly trained, knowledgeable work force can the proper methods be applied to efficiently process materials with machinery which is predictable, reliable and flexible. This paper discusses the affect of each element on the development of agile manufacturing systems.

  7. Agile manufacturing concept

    NASA Astrophysics Data System (ADS)

    Goldman, Steven L.

    1994-03-01

    The initial conceptualization of agile manufacturing was the result of a 1991 study -- chaired by Lehigh Professor Roger N. Nagel and California-based entrepreneur Rick Dove, President of Paradigm Shifts, International -- of what it would take for U.S. industry to regain global manufacturing competitiveness by the early twenty-first century. This industry-led study, reviewed by senior management at over 100 companies before its release, concluded that incremental improvement of the current system of manufacturing would not be enough to be competitive in today's global marketplace. Computer-based information and production technologies that were becoming available to industry opened up the possibility of an altogether new system of manufacturing, one that would be characterized by a distinctive integration of people and technologies; of management and labor; of customers, producers, suppliers, and society.

  8. Compact, Automated, Frequency-Agile Microspectrofluorimeter

    NASA Technical Reports Server (NTRS)

    Fernandez, Salvador M.; Guignon, Ernest F.

    1995-01-01

    Compact, reliable, rugged, automated cell-culture and frequency-agile microspectrofluorimetric apparatus developed to perform experiments involving photometric imaging observations of single live cells. In original application, apparatus operates mostly unattended aboard spacecraft; potential terrestrial applications include automated or semiautomated diagnosis of pathological tissues in clinical laboratories, biomedical instrumentation, monitoring of biological process streams, and portable instrumentation for testing biological conditions in various environments. Offers obvious advantages over present laboratory instrumentation.

  9. Development of EarthCube Governance: An Agile Approach

    NASA Astrophysics Data System (ADS)

    Pearthree, G.; Allison, M. L.; Patten, K.

    2013-12-01

    Governance of geosciences cyberinfrastructure is a complex and essential undertaking, critical in enabling distributed knowledge communities to collaborate and communicate across disciplines, distances, and cultures. Advancing science with respect to 'grand challenges," such as global climate change, weather prediction, and core fundamental science, depends not just on technical cyber systems, but also on social systems for strategic planning, decision-making, project management, learning, teaching, and building a community of practice. Simply put, a robust, agile technical system depends on an equally robust and agile social system. Cyberinfrastructure development is wrapped in social, organizational and governance challenges, which may significantly impede progress. An agile development process is underway for governance of transformative investments in geosciences cyberinfrastructure through the NSF EarthCube initiative. Agile development is iterative and incremental, and promotes adaptive planning and rapid and flexible response. Such iterative deployment across a variety of EarthCube stakeholders encourages transparency, consensus, accountability, and inclusiveness. A project Secretariat acts as the coordinating body, carrying out duties for planning, organizing, communicating, and reporting. A broad coalition of stakeholder groups comprises an Assembly (Mainstream Scientists, Cyberinfrastructure Institutions, Information Technology/Computer Sciences, NSF EarthCube Investigators, Science Communities, EarthCube End-User Workshop Organizers, Professional Societies) to serve as a preliminary venue for identifying, evaluating, and testing potential governance models. To offer opportunity for broader end-user input, a crowd-source approach will engage stakeholders not involved otherwise. An Advisory Committee from the Earth, ocean, atmosphere, social, computer and library sciences is guiding the process from a high-level policy point of view. Developmental

  10. Elements of an Art - Agile Coaching

    NASA Astrophysics Data System (ADS)

    Lundh, Erik

    This tutorial gives you a lead on becoming or redefining yourself as an Agile Coach. Introduction to elements and dimensions of state-of-the-art Agile Coaching. How to position the agile coach to be effective in a larger setting. Making the agile transition - from a single team to thousands of people. How to support multiple teams as a coach. How to build a coaches network in your company. Challenges when the agile coach is a consultant and the organization is large.

  11. What Does an Agile Coach Do?

    NASA Astrophysics Data System (ADS)

    Davies, Rachel; Pullicino, James

    The surge in Agile adoption has created a demand for project managers rather than direct their teams. A sign of this trend is the ever-increasing number of people getting certified as scrum masters and agile leaders. Training courses that introduce agile practices are easy to find. But making the transition to coach is not as simple as understanding what agile practices are. Your challenge as an Agile Coach is to support your team in learning how to wield their new Agile tools in creating great software.

  12. Software ``Best'' Practices: Agile Deconstructed

    NASA Astrophysics Data System (ADS)

    Fraser, Steven

    Software “best” practices depend entirely on context - in terms of the problem domain, the system constructed, the software designers, and the “customers” ultimately deriving value from the system. Agile practices no longer have the luxury of “choosing” small non-mission critical projects with co-located teams. Project stakeholders are selecting and adapting practices based on a combina tion of interest, need and staffing. For example, growing product portfolios through a merger or the acquisition of a company exposes legacy systems to new staff, new software integration challenges, and new ideas. Innovation in communications (tools and processes) to span the growth and contraction of both information and organizations, while managing the adoption of changing software practices, is imperative for success. Traditional web-based tools such as web pages, document libraries, and forums are not suf ficient. A blend of tweeting, blogs, wikis, instant messaging, web-based confer encing, and telepresence creates a new dimension of communication “best” practices.

  13. The influence of physical and cognitive factors on reactive agility performance in men basketball players.

    PubMed

    Scanlan, Aaron; Humphries, Brendan; Tucker, Patrick S; Dalbo, Vincent

    2014-01-01

    This study explored the influence of physical and cognitive measures on reactive agility performance in basketball players. Twelve men basketball players performed multiple sprint, Change of Direction Speed Test, and Reactive Agility Test trials. Pearson's correlation analyses were used to determine relationships between the predictor variables (stature, mass, body composition, 5-m, 10-m and 20-m sprint times, peak speed, closed-skill agility time, response time and decision-making time) and reactive agility time (response variable). Simple and stepwise regression analyses determined the individual influence of each predictor variable and the best predictor model for reactive agility time. Morphological (r = -0.45 to 0.19), sprint (r = -0.40 to 0.41) and change-of-direction speed measures (r = 0.43) had small to moderate correlations with reactive agility time. Response time (r = 0.76, P = 0.004) and decision-making time (r = 0.58, P = 0.049) had large to very large relationships with reactive agility time. Response time was identified as the sole predictor variable for reactive agility time in the stepwise model (R(2) = 0.58, P = 0.004). In conclusion, cognitive measures had the greatest influence on reactive agility performance in men basketball players. These findings suggest reaction and decision-making drills should be incorporated in basketball training programmes.

  14. Piloted simulator assessments of agility

    NASA Technical Reports Server (NTRS)

    Schneider, Edward T.

    1990-01-01

    NASA has utilized piloted simulators for nearly two decades to study high-angle-of-attack flying qualities, agility, and air-to-air combat. These studies have included assessments of an F-16XL aircraft equipped with thrust vectoring, an assessment of the F-18 HARV maneuvering requirements to assist in thrust vectoring control system design, and an agility assessment of the F-18. The F-18 agility assessment was compared with in-flight testing. Open-loop maneuvers such as 180-deg rolls to measure roll rate showed favorable simulator/in-flight comparison. Closed-loop maneuvers such as rolls to 90 deg with precision stops or certain maximum longitudinal pitching maneuvers showed poorer performance due to reduced aggressiveness of pilot inputs in flight to remain within flight envelope limits.

  15. Taking Another Look at the Data Management Life Cycle: Deconstruction, Agile, and Community

    NASA Astrophysics Data System (ADS)

    Young, J. W.; Lenhardt, W. C.; Parsons, M. A.; Benedict, K. K.

    2014-12-01

    The data life cycle has figured prominently in describing the context of digital scientific data stewardship and cyberinfractructure in support of science. There are many different versions of the data life cycle, but they all follow a similar basic pattern: plan, collect, ingest, asses, preserve, discover, and reuse. The process is often interpreted in a fairly linear fashion despite it being a cycle conceptually. More recently at GeoData 2014 and elsewhere, questions have been raised about the utility of the data life cycle as it is currently represented. We are proposing to the community a re-examination of the data life cycle using an agile lens. Our goal is not to deploy agile methods, but to use agile principles as a heuristic to think about how to incorporate data stewardship across the scientific process from proposal stage to research and beyond. We will present alternative conceptualizations of the data life cycle with a goal to solicit feedback and to develop a new model for conceiving and describing the overall data stewardship process. We seek to re-examine past assumptions and shed new light on the challenges and necessity of data stewardship. The ultimate goal is to support new science through enhanced data interoperability, usability, and preservation.

  16. The agile alert system for gamma-ray transients

    SciTech Connect

    Bulgarelli, A.; Trifoglio, M.; Gianotti, F.; Fioretti, V.; Chen, A. W.; Pittori, C.; Verrecchia, F.; Lucarelli, F.; Santolamazza, P.; Fanari, G.; Giommi, P.; Pellizzoni, A.; and others

    2014-01-20

    In recent years, a new generation of space missions has offered great opportunities for discovery in high-energy astrophysics. In this article we focus on the scientific operations of the Gamma-Ray Imaging Detector (GRID) on board the AGILE space mission. AGILE-GRID, sensitive in the energy range of 30 MeV-30 GeV, has detected many γ-ray transients of both galactic and extragalactic origin. This work presents the AGILE innovative approach to fast γ-ray transient detection, which is a challenging task and a crucial part of the AGILE scientific program. The goals are to describe (1) the AGILE Gamma-Ray Alert System, (2) a new algorithm for blind search identification of transients within a short processing time, (3) the AGILE procedure for γ-ray transient alert management, and (4) the likelihood of ratio tests that are necessary to evaluate the post-trial statistical significance of the results. Special algorithms and an optimized sequence of tasks are necessary to reach our goal. Data are automatically analyzed at every orbital downlink by an alert pipeline operating on different timescales. As proper flux thresholds are exceeded, alerts are automatically generated and sent as SMS messages to cellular telephones, via e-mail, and via push notifications from an application for smartphones and tablets. These alerts are crosschecked with the results of two pipelines, and a manual analysis is performed. Being a small scientific-class mission, AGILE is characterized by optimization of both scientific analysis and ground-segment resources. The system is capable of generating alerts within two to three hours of a data downlink, an unprecedented reaction time in γ-ray astrophysics.

  17. The AGILE Alert System for Gamma-Ray Transients

    NASA Astrophysics Data System (ADS)

    Bulgarelli, A.; Trifoglio, M.; Gianotti, F.; Tavani, M.; Parmiggiani, N.; Fioretti, V.; Chen, A. W.; Vercellone, S.; Pittori, C.; Verrecchia, F.; Lucarelli, F.; Santolamazza, P.; Fanari, G.; Giommi, P.; Beneventano, D.; Argan, A.; Trois, A.; Scalise, E.; Longo, F.; Pellizzoni, A.; Pucella, G.; Colafrancesco, S.; Conforti, V.; Tempesta, P.; Cerone, M.; Sabatini, P.; Annoni, G.; Valentini, G.; Salotti, L.

    2014-01-01

    In recent years, a new generation of space missions has offered great opportunities for discovery in high-energy astrophysics. In this article we focus on the scientific operations of the Gamma-Ray Imaging Detector (GRID) on board the AGILE space mission. AGILE-GRID, sensitive in the energy range of 30 MeV-30 GeV, has detected many γ-ray transients of both galactic and extragalactic origin. This work presents the AGILE innovative approach to fast γ-ray transient detection, which is a challenging task and a crucial part of the AGILE scientific program. The goals are to describe (1) the AGILE Gamma-Ray Alert System, (2) a new algorithm for blind search identification of transients within a short processing time, (3) the AGILE procedure for γ-ray transient alert management, and (4) the likelihood of ratio tests that are necessary to evaluate the post-trial statistical significance of the results. Special algorithms and an optimized sequence of tasks are necessary to reach our goal. Data are automatically analyzed at every orbital downlink by an alert pipeline operating on different timescales. As proper flux thresholds are exceeded, alerts are automatically generated and sent as SMS messages to cellular telephones, via e-mail, and via push notifications from an application for smartphones and tablets. These alerts are crosschecked with the results of two pipelines, and a manual analysis is performed. Being a small scientific-class mission, AGILE is characterized by optimization of both scientific analysis and ground-segment resources. The system is capable of generating alerts within two to three hours of a data downlink, an unprecedented reaction time in γ-ray astrophysics.

  18. Software ``Best'' Practices: Agile Deconstructed

    NASA Astrophysics Data System (ADS)

    Fraser, Steven

    This workshop will explore the intersection of agility and software development in a world of legacy code-bases and large teams. Organizations with hundreds of developers and code-bases exceeding a million or tens of millions of lines of code are seeking new ways to expedite development while retaining and attracting staff who desire to apply “agile” methods. This is a situation where specific agile practices may be embraced outside of their usual zone of applicability. Here is where practitioners must understand both what “best practices” already exist in the organization - and how they might be improved or modified by applying “agile” approaches.

  19. Biosphere Process Model Report

    SciTech Connect

    J. Schmitt

    2000-05-25

    To evaluate the postclosure performance of a potential monitored geologic repository at Yucca Mountain, a Total System Performance Assessment (TSPA) will be conducted. Nine Process Model Reports (PMRs), including this document, are being developed to summarize the technical basis for each of the process models supporting the TSPA model. These reports cover the following areas: (1) Integrated Site Model; (2) Unsaturated Zone Flow and Transport; (3) Near Field Environment; (4) Engineered Barrier System Degradation, Flow, and Transport; (5) Waste Package Degradation; (6) Waste Form Degradation; (7) Saturated Zone Flow and Transport; (8) Biosphere; and (9) Disruptive Events. Analysis/Model Reports (AMRs) contain the more detailed technical information used to support TSPA and the PMRs. The AMRs consists of data, analyses, models, software, and supporting documentation that will be used to defend the applicability of each process model for evaluating the postclosure performance of the potential Yucca Mountain repository system. This documentation will ensure the traceability of information from its source through its ultimate use in the TSPA-Site Recommendation (SR) and in the National Environmental Policy Act (NEPA) analysis processes. The objective of the Biosphere PMR is to summarize (1) the development of the biosphere model, and (2) the Biosphere Dose Conversion Factors (BDCFs) developed for use in TSPA. The Biosphere PMR does not present or summarize estimates of potential radiation doses to human receptors. Dose calculations are performed as part of TSPA and will be presented in the TSPA documentation. The biosphere model is a component of the process to evaluate postclosure repository performance and regulatory compliance for a potential monitored geologic repository at Yucca Mountain, Nevada. The biosphere model describes those exposure pathways in the biosphere by which radionuclides released from a potential repository could reach a human receptor

  20. Lean and Agile Development of the AITS Ground Software System

    NASA Astrophysics Data System (ADS)

    Richters, Mark; Dutruel, Etienne; Mecredy, Nicolas

    2013-08-01

    We present the ongoing development of a new ground software system used for integrating, testing and operating spacecraft. The Advanced Integration and Test Services (AITS) project aims at providing a solution for electrical ground support equipment and mission control systems in future Astrium Space Transportation missions. Traditionally ESA ground or flight software development projects are conducted according to a waterfall-like process as specified in the ECSS-E-40 standard promoted by ESA in the European industry. In AITS a decision was taken to adopt an agile development process. This work could serve as a reference for future ESA software projects willing to apply agile concepts.

  1. Agile Development Methods for Space Operations

    NASA Technical Reports Server (NTRS)

    Trimble, Jay; Webster, Chris

    2012-01-01

    Main stream industry software development practice has gone from a traditional waterfall process to agile iterative development that allows for fast response to customer inputs and produces higher quality software at lower cost. How can we, the space ops community, adopt state of the art software development practice, achieve greater productivity at lower cost, and maintain safe and effective space flight operations? At NASA Ames, we are developing Mission Control Technologies Software, in collaboration with Johnson Space Center (JSC) and, more recently, the Jet Propulsion Laboratory (JPL).

  2. Foam process models.

    SciTech Connect

    Moffat, Harry K.; Noble, David R.; Baer, Thomas A.; Adolf, Douglas Brian; Rao, Rekha Ranjana; Mondy, Lisa Ann

    2008-09-01

    In this report, we summarize our work on developing a production level foam processing computational model suitable for predicting the self-expansion of foam in complex geometries. The model is based on a finite element representation of the equations of motion, with the movement of the free surface represented using the level set method, and has been implemented in SIERRA/ARIA. An empirically based time- and temperature-dependent density model is used to encapsulate the complex physics of foam nucleation and growth in a numerically tractable model. The change in density with time is at the heart of the foam self-expansion as it creates the motion of the foam. This continuum-level model uses an homogenized description of foam, which does not include the gas explicitly. Results from the model are compared to temperature-instrumented flow visualization experiments giving the location of the foam front as a function of time for our EFAR model system.

  3. Flight dynamics research for highly agile aircraft

    NASA Technical Reports Server (NTRS)

    Nguyen, Luat T.

    1989-01-01

    This paper highlights recent results of research conducted at the NASA Langley Research Center as part of a broad flight dynamics program aimed at developing technology that will enable future combat aircraft to achieve greatly enhanced agility capability at subsonic combat conditions. Studies of advanced control concepts encompassing both propulsive and aerodynamic approaches are reviewed. Dynamic stall phenomena and their potential impact on maneuvering performance and stability are summarized. Finally, issues of mathematical modeling of complex aerodynamics occurring during rapid, large amplitude maneuvers are discussed.

  4. Development of perceived competence, tactical skills, motivation, technical skills, and speed and agility in young soccer players.

    PubMed

    Forsman, Hannele; Gråstén, Arto; Blomqvist, Minna; Davids, Keith; Liukkonen, Jarmo; Konttinen, Niilo

    2016-07-01

    The objective of this 1-year, longitudinal study was to examine the development of perceived competence, tactical skills, motivation, technical skills, and speed and agility characteristics of young Finnish soccer players. We also examined associations between latent growth models of perceived competence and other recorded variables. Participants were 288 competitive male soccer players ranging from 12 to 14 years (12.7 ± 0.6) from 16 soccer clubs. Players completed the self-assessments of perceived competence, tactical skills, and motivation, and participated in technical, and speed and agility tests. Results of this study showed that players' levels of perceived competence, tactical skills, motivation, technical skills, and speed and agility characteristics remained relatively high and stable across the period of 1 year. Positive relationships were found between these levels and changes in perceived competence and motivation, and levels of perceived competence and speed and agility characteristics. Together these results illustrate the multi-dimensional nature of talent development processes in soccer. Moreover, it seems crucial in coaching to support the development of perceived competence and motivation in young soccer players and that it might be even more important in later maturing players. PMID:26708723

  5. Decision Support for Iteration Scheduling in Agile Environments

    NASA Astrophysics Data System (ADS)

    Szőke, Ákos

    Today’s software business development projects often lay claim to low-risk value to the customers in order to be financed. Emerging agile processes offer shorter investment periods, faster time-to-market and better customer satisfaction. To date, however, in agile environments there is no sound methodological schedule support contrary to the traditional plan-based approaches. To address this situation, we present an agile iteration scheduling method whose usefulness is evaluated with post-mortem simulation. It demonstrates that the method can significantly improve load balancing of resources (cca. 5×), produce higher quality and lower-risk feasible schedule, and provide more informed and established decisions by optimized schedule production. Finally, the paper analyzes benefits and issues from the use of this method.

  6. Onshore and Offshore Outsourcing with Agility: Lessons Learned

    NASA Astrophysics Data System (ADS)

    Kussmaul, Clifton

    This chapter reflects on case study based an agile distributed project that ran for approximately three years (from spring 2003 to spring 2006). The project involved (a) a customer organization with key personnel distributed across the US, developing an application with rapidly changing requirements; (b) onshore consultants with expertise in project management, development processes, offshoring, and relevant technologies; and (c) an external offsite development team in a CMM-5 organization in southern India. This chapter is based on surveys and discussions with multiple participants. The several years since the project was completed allow greater perspective on both the strengths and weaknesses, since the participants can reflect on the entire life of the project, and compare it to subsequent experiences. Our findings emphasize the potential for agile project management in distributed software development, and the importance of people and interactions, taking many small steps to find and correct errors, and matching the structures of the project and product to support implementation of agility.

  7. SU-E-T-610: Comparison of Treatment Times Between the MLCi and Agility Multileaf Collimators

    SciTech Connect

    Ramsey, C; Bowling, J

    2014-06-01

    Purpose: The Agility is a new 160-leaf MLC developed by Elekta for use in their Infinity and Versa HD linacs. As compared to the MLCi, the Agility increased the maximum leaf speed from 2 cm/s to 3.5 cm/s, and the maximum primary collimator speed from 1.5 cm/s to 9.0 cm/s. The purpose of this study was to determine if the Agility MLC resulted in improved plan quality and/or shorter treatment times. Methods: An Elekta Infinity that was originally equipped with a 80 leaf MLCi was upgraded to an 160 leaf Agility. Treatment plan quality was evaluated using the Pinnacle planning system with SmartArc. Optimization was performed once for the MLCi and once for the Agility beam models using the same optimization parameters and the same number of iterations. Patient treatment times were measured for all IMRT, VMAT, and SBRT patients treated on the Infinity with the MLCi and Agility MLCs. Treatment times were extracted from the EMR and measured from when the patient first walked into the treatment room until exiting the treatment room. Results: 11,380 delivery times were measured for patients treated with the MLCi, and 1,827 measurements have been made for the Agility MLC. The average treatment times were 19.1 minutes for the MLCi and 20.8 minutes for the Agility. Using a t-test analysis, there was no difference between the two groups (t = 0.22). The dose differences between patients planned with the MLCi and the Agility MLC were minimal. For example, the dose difference for the PTV, GTV, and cord for a head and neck patient planned using Pinnacle were effectively equivalent. However, the dose to the parotid glands was slightly worse with the Agility MLC. Conclusion: There was no statistical difference in treatment time, or any significant dosimetric difference between the Agility MLC and the MLCi.

  8. Evaluation of a novel laparoscopic camera for characterization of renal ischemia in a porcine model using digital light processing (DLP) hyperspectral imaging

    NASA Astrophysics Data System (ADS)

    Olweny, Ephrem O.; Tan, Yung K.; Faddegon, Stephen; Jackson, Neil; Wehner, Eleanor F.; Best, Sara L.; Park, Samuel K.; Thapa, Abhas; Cadeddu, Jeffrey A.; Zuzak, Karel J.

    2012-03-01

    Digital light processing hyperspectral imaging (DLP® HSI) was adapted for use during laparoscopic surgery by coupling a conventional laparoscopic light guide with a DLP-based Agile Light source (OL 490, Optronic Laboratories, Orlando, FL), incorporating a 0° laparoscope, and a customized digital CCD camera (DVC, Austin, TX). The system was used to characterize renal ischemia in a porcine model.

  9. Biorobotics: using robots to emulate and investigate agile locomotion.

    PubMed

    Ijspeert, Auke J

    2014-10-10

    The graceful and agile movements of animals are difficult to analyze and emulate because locomotion is the result of a complex interplay of many components: the central and peripheral nervous systems, the musculoskeletal system, and the environment. The goals of biorobotics are to take inspiration from biological principles to design robots that match the agility of animals, and to use robots as scientific tools to investigate animal adaptive behavior. Used as physical models, biorobots contribute to hypothesis testing in fields such as hydrodynamics, biomechanics, neuroscience, and prosthetics. Their use may contribute to the design of prosthetic devices that more closely take human locomotion principles into account.

  10. A Framework for Decomposition and Analysis of Agile Methodologies During Their Adaptation

    NASA Astrophysics Data System (ADS)

    Mikulenas, Gytenis; Kapocius, Kestutis

    In recent years there has been a steady increase of interest in Agile software development methodologies and techniques, which are often positioned as proven alternatives to the traditional plan-driven approaches. However, although there is no shortage of Agile methodologies to choose from, the formal methods for actually choosing or adapting the right one are lacking. The aim of the presented research was to define the formal way of preparing Agile methodologies for adaptation and creating an adaptation process framework. We argue that Agile methodologies can be successfully broken down into individual parts that can be specified on three different levels and later analyzed with regard to problem/concern areas. Results of such decomposition can form the foundation for the decisions on the adaptation of the specific Agile methodology. A case study is included in this chapter to further clarify the proposed approach.

  11. Agile methods in biomedical software development: a multi-site experience report

    PubMed Central

    Kane, David W; Hohman, Moses M; Cerami, Ethan G; McCormick, Michael W; Kuhlmman, Karl F; Byrd, Jeff A

    2006-01-01

    Background Agile is an iterative approach to software development that relies on strong collaboration and automation to keep pace with dynamic environments. We have successfully used agile development approaches to create and maintain biomedical software, including software for bioinformatics. This paper reports on a qualitative study of our experiences using these methods. Results We have found that agile methods are well suited to the exploratory and iterative nature of scientific inquiry. They provide a robust framework for reproducing scientific results and for developing clinical support systems. The agile development approach also provides a model for collaboration between software engineers and researchers. We present our experience using agile methodologies in projects at six different biomedical software development organizations. The organizations include academic, commercial and government development teams, and included both bioinformatics and clinical support applications. We found that agile practices were a match for the needs of our biomedical projects and contributed to the success of our organizations. Conclusion We found that the agile development approach was a good fit for our organizations, and that these practices should be applicable and valuable to other biomedical software development efforts. Although we found differences in how agile methods were used, we were also able to identify a set of core practices that were common to all of the groups, and that could be a focus for others seeking to adopt these methods. PMID:16734914

  12. Evaluation of agile designs in first-in-human (FIH) trials--a simulation study.

    PubMed

    Perlstein, Itay; Bolognese, James A; Krishna, Rajesh; Wagner, John A

    2009-12-01

    The aim of the investigation was to evaluate alternatives to standard first-in-human (FIH) designs in order to optimize the information gained from such studies by employing novel agile trial designs. Agile designs combine adaptive and flexible elements to enable optimized use of prior information either before and/or during conduct of the study to seamlessly update the study design. A comparison of the traditional 6 + 2 (active + placebo) subjects per cohort design with alternative, reduced sample size, agile designs was performed by using discrete event simulation. Agile designs were evaluated for specific adverse event models and rates as well as dose-proportional, saturated, and steep-accumulation pharmacokinetic profiles. Alternative, reduced sample size (hereafter referred to as agile) designs are proposed for cases where prior knowledge about pharmacokinetics and/or adverse event relationships are available or appropriately assumed. Additionally, preferred alternatives are proposed for a general case when prior knowledge is limited or unavailable. Within the tested conditions and stated assumptions, some agile designs were found to be as efficient as traditional designs. Thus, simulations demonstrated that the agile design is a robust and feasible approach to FIH clinical trials, with no meaningful loss of relevant information, as it relates to PK and AE assumptions. In some circumstances, applying agile designs may decrease the duration and resources required for Phase I studies, increasing the efficiency of early clinical development. We highlight the value and importance of useful prior information when specifying key assumptions related to safety, tolerability, and PK.

  13. AGILE integration into APC for high mix logic fab

    NASA Astrophysics Data System (ADS)

    Gatefait, M.; Lam, A.; Le Gratiet, B.; Mikolajczak, M.; Morin, V.; Chojnowski, N.; Kocsis, Z.; Smith, I.; Decaunes, J.; Ostrovsky, A.; Monget, C.

    2015-09-01

    For C040 technology and below, photolithographic depth of focus control and dispersion improvement is essential to secure product functionality. Critical 193nm immersion layers present initial focus process windows close to machine control capability. For previous technologies, the standard scanner sensor (Level sensor - LS) was used to map wafer topology and expose the wafer at the right Focus. Such optical embedded metrology, based on light reflection, suffers from reading issues that cannot be neglected anymore. Metrology errors are correlated to inspected product area for which material types and densities change, and so optical properties are not constant. Various optical phenomena occur across the product field during wafer inspection and have an effect on the quality and position of the reflected light. This can result in incorrect heights being recorded and exposures possibly being done out of focus. Focus inaccuracy associated to aggressive process windows on critical layers will directly impact product realization and therefore functionality and yield. ASML has introduced an air gauge sensor to complement the optical level sensor and lead to optimal topology metrology. The use of this new sensor is managed by the AGILE (Air Gauge Improved process LEveling) application. This measurement with no optical dependency will correct for optical inaccuracy of level sensor, and so improve best focus dispersion across the product. Due to the fact that stack complexity is more and more important through process steps flow, optical perturbation of standard Level sensor metrology is increasing and is becoming maximum for metallization layers. For these reasons AGILE feature implementation was first considered for contact and all metal layers. Another key point is that standard metrology will be sensitive to layer and reticle/product density. The gain of Agile will be enhanced for multiple product contribution mask and for complex System on Chip. Into ST context (High

  14. An Investigation of Agility Issues in Scrum Teams Using Agility Indicators

    NASA Astrophysics Data System (ADS)

    Pikkarainen, Minna; Wang, Xiaofeng

    Agile software development methods have emerged and become increasingly popular in recent years; yet the issues encountered by software development teams that strive to achieve agility using agile methods are yet to be explored systematically. Built upon a previous study that has established a set of indicators of agility, this study investigates what issues are manifested in software development teams using agile methods. It is focussed on Scrum teams particularly. In other words, the goal of the chapter is to evaluate Scrum teams using agility indicators and therefore to further validate previously presented agility indicators within the additional cases. A multiple case study research method is employed. The findings of the study reveal that the teams using Scrum do not necessarily achieve agility in terms of team autonomy, sharing, stability and embraced uncertainty. The possible reasons include previous organizational plan-driven culture, resistance towards the Scrum roles and changing resources.

  15. The AGILE gamma-ray astronomy mission

    NASA Astrophysics Data System (ADS)

    Mereghetti, S.; Tavani, M.; Argan, A.; Barbiellini, G.; Caraveo, P.; Chen, A.; Cocco, V.; Costa, E.; Di Cocco, G.; Feroci, M.; Labanti, C.; Lapshov, I.; Lipari, P.; Longo, F.; Morselli, A.; Perotti, F.; Picozza, P.; Pittori, C.; Prest, M.; Rubini, A.; Soffitta, P.; Vallazza, E.; Vercellone, S.; Zanello, D.

    2001-09-01

    We describe the AGILE satellite: a unique tool for high-energy astrophysics in the 30 MeV - 50 GeV range before GLAST. The scientific performances of AGILE are comparable to those of EGRET, despite the much smaller weight and dimensions. The AGILE mission will be optimized for the imaging capabilities above 30 MeV and for the study of transient phenomena, complemented by simultaneous monitoring in the hard X-ray band (10 - 40 keV).

  16. Agile manufacturing and constraints management: a strategic perspective

    NASA Astrophysics Data System (ADS)

    Stratton, Roy; Yusuf, Yahaya Y.

    2000-10-01

    The definition of the agile paradigm has proved elusive and is often viewed as a panacea, in contention with more traditional approaches to operations strategy development and Larkin its own methodology and tools. The Theory of Constraints (TOC) is also poorly understood, as it is commonly solely associated with production planning and control systems and bottleneck management. This paper will demonstrate the synergy between these two approaches together with the Theory of Inventive Problem Solving (TRIZ), and establish how the systematic elimination of trade-offs can support the agile paradigm. Whereas agility is often seen as a trade-off free destination, both TOC and TRIZ may be considered to be route finders, as they comprise methodologies that focus on the identification and elimination of the trade-offs that constrain the purposeful improvement of a system, be it organizational or mechanical. This paper will also show how the TOC thinking process may be combined with the TRIZ knowledge based approach and used in breaking contradictions within agile logistics.

  17. Network configuration management : paving the way to network agility.

    SciTech Connect

    Maestas, Joseph H.

    2007-08-01

    Sandia networks consist of nearly nine hundred routers and switches and nearly one million lines of command code, and each line ideally contributes to the capabilities of the network to convey information from one location to another. Sandia's Cyber Infrastructure Development and Deployment organizations recognize that it is therefore essential to standardize network configurations and enforce conformance to industry best business practices and documented internal configuration standards to provide a network that is agile, adaptable, and highly available. This is especially important in times of constrained budgets as members of the workforce are called upon to improve efficiency, effectiveness, and customer focus. Best business practices recommend using the standardized configurations in the enforcement process so that when root cause analysis results in recommended configuration changes, subsequent configuration auditing will improve compliance to the standard. Ultimately, this minimizes mean time to repair, maintains the network security posture, improves network availability, and enables efficient transition to new technologies. Network standardization brings improved network agility, which in turn enables enterprise agility, because the network touches all facets of corporate business. Improved network agility improves the business enterprise as a whole.

  18. Tailoring Agility: Promiscuous Pair Story Authoring and Value Calculation

    NASA Astrophysics Data System (ADS)

    Tendon, Steve

    This chapter describes how a multi-national software organization created a business plan involving business units from eight countries that followed an agile way, after two previously failed attempts with traditional approaches. The case is told by the consultant who initiated implementation of agility into requirements gathering, estimation and planning processes in an international setting. The agile approach was inspired by XP, but then tailored to meet the peculiar requirements. Two innovations were critical. The first innovation was promiscuous pair story authoring, where user stories were written by two people (similarly to pair programming), and the pairing changed very often (as frequently as every 15-20 minutes) to achieve promiscuity and cater for diverse point of views. The second innovation was an economic value evaluation (and not the cost) which was attributed to stories. Continuous recalculation of the financial value of the stories allowed to assess the projects financial return. In this case implementation of agility in the international context allowed the involved team members to reach consensus and unanimity of decisions, vision and purpose.

  19. Current State of Agile User-Centered Design: A Survey

    NASA Astrophysics Data System (ADS)

    Hussain, Zahid; Slany, Wolfgang; Holzinger, Andreas

    Agile software development methods are quite popular nowadays and are being adopted at an increasing rate in the industry every year. However, these methods are still lacking usability awareness in their development lifecycle, and the integration of usability/User-Centered Design (UCD) into agile methods is not adequately addressed. This paper presents the preliminary results of a recently conducted online survey regarding the current state of the integration of agile methods and usability/UCD. A world wide response of 92 practitioners was received. The results show that the majority of practitioners perceive that the integration of agile methods with usability/UCD has added value to their adopted processes and to their teams; has resulted in the improvement of usability and quality of the product developed; and has increased the satisfaction of the end-users of the product developed. The top most used HCI techniques are low-fidelity prototyping, conceptual designs, observational studies of users, usability expert evaluations, field studies, personas, rapid iterative testing, and laboratory usability testing.

  20. Neuromuscular strategies contributing to faster multidirectional agility performance.

    PubMed

    Spiteri, Tania; Newton, Robert U; Nimphius, Sophia

    2015-08-01

    The aim of this study was to first determine differences in neuromuscular strategy between a faster and slower agility performance, and second compare differences in muscle activation strategy employed when performing two closely executed agility movements. Participants recruited from an elite female basketball team completed an ultrasound to determine quadriceps muscle-cross sectional area; reactive isometric mid-thigh pull to determine the rate of muscle activation, rate of force development, pre-motor time and motor time; and multidirectional agility tests completing two directional changes in response to a visual stimulus. Peak and average relative muscle activation of the rectus femoris, vastus medialis, vastus lateralis, biceps femoris, semitendinosus and gastrocnemius were measured 100ms prior to heel strike (pre-heel strike) and across stance phase for both directional changes. Faster agility performance was characterized by greater pre-heel strike muscle activity and greater anterior muscle activation during stance phase resulting in greater hip and knee extension increasing propulsive impulse. Differences between directional changes appear to result from processing speed, where a greater delay in refractory times during the second directional change resulted in greater anterior muscle activation, decelerating the body while movement direction was determined.

  1. On the biomimetic design of agile-robot legs.

    PubMed

    Garcia, Elena; Arevalo, Juan Carlos; Muñoz, Gustavo; Gonzalez-de-Santos, Pablo

    2011-01-01

    The development of functional legged robots has encountered its limits in human-made actuation technology. This paper describes research on the biomimetic design of legs for agile quadrupeds. A biomimetic leg concept that extracts key principles from horse legs which are responsible for the agile and powerful locomotion of these animals is presented. The proposed biomimetic leg model defines the effective leg length, leg kinematics, limb mass distribution, actuator power, and elastic energy recovery as determinants of agile locomotion, and values for these five key elements are given. The transfer of the extracted principles to technological instantiations is analyzed in detail, considering the availability of current materials, structures and actuators. A real leg prototype has been developed following the biomimetic leg concept proposed. The actuation system is based on the hybrid use of series elasticity and magneto-rheological dampers which provides variable compliance for natural motion. From the experimental evaluation of this prototype, conclusions on the current technological barriers to achieve real functional legged robots to walk dynamically in agile locomotion are presented.

  2. On the Biomimetic Design of Agile-Robot Legs

    PubMed Central

    Garcia, Elena; Arevalo, Juan Carlos; Muñoz, Gustavo; Gonzalez-de-Santos, Pablo

    2011-01-01

    The development of functional legged robots has encountered its limits in human-made actuation technology. This paper describes research on the biomimetic design of legs for agile quadrupeds. A biomimetic leg concept that extracts key principles from horse legs which are responsible for the agile and powerful locomotion of these animals is presented. The proposed biomimetic leg model defines the effective leg length, leg kinematics, limb mass distribution, actuator power, and elastic energy recovery as determinants of agile locomotion, and values for these five key elements are given. The transfer of the extracted principles to technological instantiations is analyzed in detail, considering the availability of current materials, structures and actuators. A real leg prototype has been developed following the biomimetic leg concept proposed. The actuation system is based on the hybrid use of series elasticity and magneto-rheological dampers which provides variable compliance for natural motion. From the experimental evaluation of this prototype, conclusions on the current technological barriers to achieve real functional legged robots to walk dynamically in agile locomotion are presented. PMID:22247667

  3. Multiply-agile encryption in high speed communication networks

    SciTech Connect

    Pierson, L.G.; Witzke, E.L.

    1997-05-01

    Different applications have different security requirements for data privacy, data integrity, and authentication. Encryption is one technique that addresses these requirements. Encryption hardware, designed for use in high-speed communications networks, can satisfy a wide variety of security requirements if that hardware is key-agile, robustness-agile and algorithm-agile. Hence, multiply-agile encryption provides enhanced solutions to the secrecy, interoperability and quality of service issues in high-speed networks. This paper defines these three types of agile encryption. Next, implementation issues are discussed. While single-algorithm, key-agile encryptors exist, robustness-agile and algorithm-agile encryptors are still research topics.

  4. How Can Agile Practices Minimize Global Software Development Co-ordination Risks?

    NASA Astrophysics Data System (ADS)

    Hossain, Emam; Babar, Muhammad Ali; Verner, June

    The distribution of project stakeholders in Global Software Development (GSD) projects provides significant risks related to project communication, coordination and control processes. There is growing interest in applying agile practices in GSD projects in order to leverage the advantages of both approaches. In some cases, GSD project managers use agile practices to reduce project distribution challenges. We use an existing coordination framework to identify GSD coordination problems due to temporal, geographical and socio-cultural distances. An industry-based case study is used to describe, explore and explain the use of agile practices to reduce development coordination challenges.

  5. Transitioning from Distributed and Traditional to Distributed and Agile: An Experience Report

    NASA Astrophysics Data System (ADS)

    Wildt, Daniel; Prikladnicki, Rafael

    Global companies that experienced extensive waterfall phased plans are trying to improve their existing processes to expedite team engagement. Agile methodologies have become an acceptable path to follow because it comprises project management as part of its practices. Agile practices have been used with the objective of simplifying project control through simple processes, easy to update documentation and higher team iteration over exhaustive documentation, focusing rather on team continuous improvement and aiming to add value to business processes. The purpose of this chapter is to describe the experience of a global multinational company on transitioning from distributed and traditional to distributed and agile. This company has development centers across North America, South America and Asia. This chapter covers challenges faced by the project teams of two pilot projects, including strengths of using agile practices in a globally distributed environment and practical recommendations for similar endeavors.

  6. The GLAST-AGILE Support Program (GASP)

    NASA Astrophysics Data System (ADS)

    Villata, M.; Raiteri, C. M.; Webt Collaboration

    2008-10-01

    The GLAST-AGILE Support Program (GASP) was organized within the Whole Earth Blazar Telescope to provide optical-to-radio long-term continuous monitoring of a list of selected gamma-ray-loud blazars during the operation of the AGILE and GLAST satellites. We present some results obtained since its birth, in September 2007.

  7. Teaching Agile Software Development: A Case Study

    ERIC Educational Resources Information Center

    Devedzic, V.; Milenkovic, S. R.

    2011-01-01

    This paper describes the authors' experience of teaching agile software development to students of computer science, software engineering, and other related disciplines, and comments on the implications of this and the lessons learned. It is based on the authors' eight years of experience in teaching agile software methodologies to various groups…

  8. An Agile Course-Delivery Approach

    ERIC Educational Resources Information Center

    Capellan, Mirkeya

    2009-01-01

    In the world of software development, agile methodologies have gained popularity thanks to their lightweight methodologies and flexible approach. Many advocates believe that agile methodologies can provide significant benefits if applied in the educational environment as a teaching method. The need for an approach that engages and motivates…

  9. The Introduction of Agility into Albania.

    ERIC Educational Resources Information Center

    Smith-Stevens, Eileen J.; Shkurti, Drita

    1998-01-01

    Describes a plan to introduce and achieve a national awareness of agility (and easy entry into the world market) for Albania through the relatively stable higher-education order. Agility's four strategic principles are enriching the customer, cooperating to enhance competitiveness, organizing to master change and uncertainty, and leveraging the…

  10. GREENSCOPE: Sustainable Process Modeling

    EPA Science Inventory

    EPA researchers are responding to environmental problems by incorporating sustainability into process design and evaluation. EPA researchers are also developing a tool that allows users to assess modifications to existing and new chemical processes to determine whether changes in...

  11. Agile parallel bioinformatics workflow management using Pwrake

    PubMed Central

    2011-01-01

    Background In bioinformatics projects, scientific workflow systems are widely used to manage computational procedures. Full-featured workflow systems have been proposed to fulfil the demand for workflow management. However, such systems tend to be over-weighted for actual bioinformatics practices. We realize that quick deployment of cutting-edge software implementing advanced algorithms and data formats, and continuous adaptation to changes in computational resources and the environment are often prioritized in scientific workflow management. These features have a greater affinity with the agile software development method through iterative development phases after trial and error. Here, we show the application of a scientific workflow system Pwrake to bioinformatics workflows. Pwrake is a parallel workflow extension of Ruby's standard build tool Rake, the flexibility of which has been demonstrated in the astronomy domain. Therefore, we hypothesize that Pwrake also has advantages in actual bioinformatics workflows. Findings We implemented the Pwrake workflows to process next generation sequencing data using the Genomic Analysis Toolkit (GATK) and Dindel. GATK and Dindel workflows are typical examples of sequential and parallel workflows, respectively. We found that in practice, actual scientific workflow development iterates over two phases, the workflow definition phase and the parameter adjustment phase. We introduced separate workflow definitions to help focus on each of the two developmental phases, as well as helper methods to simplify the descriptions. This approach increased iterative development efficiency. Moreover, we implemented combined workflows to demonstrate modularity of the GATK and Dindel workflows. Conclusions Pwrake enables agile management of scientific workflows in the bioinformatics domain. The internal domain specific language design built on Ruby gives the flexibility of rakefiles for writing scientific workflows. Furthermore, readability

  12. Integrating Low-Cost Rapid Usability Testing into Agile System Development of Healthcare IT: A Methodological Perspective.

    PubMed

    Kushniruk, Andre W; Borycki, Elizabeth M

    2015-01-01

    The development of more usable and effective healthcare information systems has become a critical issue. In the software industry methodologies such as agile and iterative development processes have emerged to lead to more effective and usable systems. These approaches highlight focusing on user needs and promoting iterative and flexible development practices. Evaluation and testing of iterative agile development cycles is considered an important part of the agile methodology and iterative processes for system design and re-design. However, the issue of how to effectively integrate usability testing methods into rapid and flexible agile design cycles has remained to be fully explored. In this paper we describe our application of an approach known as low-cost rapid usability testing as it has been applied within agile system development in healthcare. The advantages of the integrative approach are described, along with current methodological considerations. PMID:25991130

  13. Analysis on critical success factors for agile manufacturing evaluation in original equipment manufacturing industry-an AHP approach

    NASA Astrophysics Data System (ADS)

    Ajay Guru Dev, C.; Senthil Kumar, V. S.

    2016-08-01

    Manufacturing industries are facing challenges in the implementation of agile manufacturing in their products and processes. Agility is widely accepted as a new competitive concept in the manufacturing sector in fulfilling varying customer demand. Thus, evaluation of agile manufacturing in industries has become a necessity. The success of an organisation depends on its ability to manage finding the critical success factors and give them special and continued attention in order to bring about high performance. This paper proposes a set of critical success factors (CSFs) for evaluating agile manufacturing considered appropriate for the manufacturing sector. The analytical hierarchy process (AHP) method is applied for prioritizing the success factors, by summarizing the opinions of experts. It is believed that the proposed CSFs enable and assist manufacturing industries to achieve a higher performance in agile manufacturing so as to increase competitiveness.

  14. Integrating Low-Cost Rapid Usability Testing into Agile System Development of Healthcare IT: A Methodological Perspective.

    PubMed

    Kushniruk, Andre W; Borycki, Elizabeth M

    2015-01-01

    The development of more usable and effective healthcare information systems has become a critical issue. In the software industry methodologies such as agile and iterative development processes have emerged to lead to more effective and usable systems. These approaches highlight focusing on user needs and promoting iterative and flexible development practices. Evaluation and testing of iterative agile development cycles is considered an important part of the agile methodology and iterative processes for system design and re-design. However, the issue of how to effectively integrate usability testing methods into rapid and flexible agile design cycles has remained to be fully explored. In this paper we describe our application of an approach known as low-cost rapid usability testing as it has been applied within agile system development in healthcare. The advantages of the integrative approach are described, along with current methodological considerations.

  15. Adopting best practices: "Agility" moves from software development to healthcare project management.

    PubMed

    Kitzmiller, Rebecca; Hunt, Eleanor; Sproat, Sara Breckenridge

    2006-01-01

    It is time for a change in mindset in how nurses operationalize system implementations and manage projects. Computers and systems have evolved over time from unwieldy mysterious machines of the past to ubiquitous computer use in every aspect of daily lives and work sites. Yet, disconcertingly, the process used to implement these systems has not evolved. Technology implementation does not need to be a struggle. It is time to adapt traditional plan-driven implementation methods to incorporate agile techniques. Agility is a concept borrowed from software development and is presented here because it encourages flexibility, adaptation, and continuous learning as part of the implementation process. Agility values communication and harnesses change to an advantage, which facilitates the natural evolution of an adaptable implementation process. Specific examples of agility in an implementation are described, and plan-driven implementation stages are adapted to incorporate relevant agile techniques. This comparison demonstrates how an agile approach enhances traditional implementation techniques to meet the demands of today's complex healthcare environments. PMID:16554690

  16. Adopting best practices: "Agility" moves from software development to healthcare project management.

    PubMed

    Kitzmiller, Rebecca; Hunt, Eleanor; Sproat, Sara Breckenridge

    2006-01-01

    It is time for a change in mindset in how nurses operationalize system implementations and manage projects. Computers and systems have evolved over time from unwieldy mysterious machines of the past to ubiquitous computer use in every aspect of daily lives and work sites. Yet, disconcertingly, the process used to implement these systems has not evolved. Technology implementation does not need to be a struggle. It is time to adapt traditional plan-driven implementation methods to incorporate agile techniques. Agility is a concept borrowed from software development and is presented here because it encourages flexibility, adaptation, and continuous learning as part of the implementation process. Agility values communication and harnesses change to an advantage, which facilitates the natural evolution of an adaptable implementation process. Specific examples of agility in an implementation are described, and plan-driven implementation stages are adapted to incorporate relevant agile techniques. This comparison demonstrates how an agile approach enhances traditional implementation techniques to meet the demands of today's complex healthcare environments.

  17. Towards a Better Understanding of CMMI and Agile Integration - Multiple Case Study of Four Companies

    NASA Astrophysics Data System (ADS)

    Pikkarainen, Minna

    The amount of software is increasing in the different domains in Europe. This provides the industries in smaller countries good opportunities to work in the international markets. Success in the global markets however demands the rapid production of high quality, error free software. Both CMMI and agile methods seem to provide a ready solution for quality and lead time improvements. There is not, however, much empirical evidence available either about 1) how the integration of these two aspects can be done in practice or 2) what it actually demands from assessors and software process improvement groups. The goal of this paper is to increase the understanding of CMMI and agile integration, in particular, focusing on the research question: how to use ‘lightweight’ style of CMMI assessments in agile contexts. This is done via four case studies in which assessments were conducted using the goals of CMMI integrated project management and collaboration and coordination with relevant stakeholder process areas and practices from XP and Scrum. The study shows that the use of agile practices may support the fulfilment of the goals of CMMI process areas but there are still many challenges for the agile teams to be solved within the continuous improvement programs. It also identifies practical advices to the assessors and improvement groups to take into consideration when conducting assessment in the context of agile software development.

  18. A combined exercise model for improving muscle strength, balance, walking distance, and motor agility in multiple sclerosis patients: A randomized clinical trial

    PubMed Central

    Sangelaji, Bahram; Kordi, Mohammadreza; Banihashemi, Farzaneh; Nabavi, Seyed Massood; Khodadadeh, Sara; Dastoorpoor, Maryam

    2016-01-01

    Background: Multiple sclerosis (MS) is a neurological disease with a variety of signs and symptoms. Exercise therapy has been shown to improve physical functions in MS. However, questions about an optimal exercise therapy remain. In this regard, we suggest a combined exercise therapy including aerobic and resistance exercises for MS patients. The study is designed to observe, test and compare the effects of proposed combined exercises on strength, balance, agility, fatigue, speed, and walking distance in people with mild to moderate MS [0 < expanded disability status scale (EDSS) < 5]. Methods: A total of 40 people with relapse-remitting MS (16 male, 0 < EDSS < 5) were randomized into one of the four groups (3 intervention and one control). The intervention consisted of various combinations of aerobic and resistance exercises with different repetition rates. Pre- and post-intervention scores of fatigue severity scale (FSS), timed up and go (TUG) test, 6-minute walk test (6MWT), 10- and 20-MWT, Berg balance scale (BBS), and one repetition maximum (1RM) test were recorded and analyzed. Results: For most tests, post-intervention values of the group 1, with 3-aerobic and 1-resistance exercises, were significantly higher compared to control group (P < 0.050). However, no significant progression was observed in the other two intervention groups. Conclusion: A combination of three aerobic exercises with one resistance exercise may result in improved balance, locomotion, and endurance in MS patients. PMID:27648171

  19. A combined exercise model for improving muscle strength, balance, walking distance, and motor agility in multiple sclerosis patients: A randomized clinical trial

    PubMed Central

    Sangelaji, Bahram; Kordi, Mohammadreza; Banihashemi, Farzaneh; Nabavi, Seyed Massood; Khodadadeh, Sara; Dastoorpoor, Maryam

    2016-01-01

    Background: Multiple sclerosis (MS) is a neurological disease with a variety of signs and symptoms. Exercise therapy has been shown to improve physical functions in MS. However, questions about an optimal exercise therapy remain. In this regard, we suggest a combined exercise therapy including aerobic and resistance exercises for MS patients. The study is designed to observe, test and compare the effects of proposed combined exercises on strength, balance, agility, fatigue, speed, and walking distance in people with mild to moderate MS [0 < expanded disability status scale (EDSS) < 5]. Methods: A total of 40 people with relapse-remitting MS (16 male, 0 < EDSS < 5) were randomized into one of the four groups (3 intervention and one control). The intervention consisted of various combinations of aerobic and resistance exercises with different repetition rates. Pre- and post-intervention scores of fatigue severity scale (FSS), timed up and go (TUG) test, 6-minute walk test (6MWT), 10- and 20-MWT, Berg balance scale (BBS), and one repetition maximum (1RM) test were recorded and analyzed. Results: For most tests, post-intervention values of the group 1, with 3-aerobic and 1-resistance exercises, were significantly higher compared to control group (P < 0.050). However, no significant progression was observed in the other two intervention groups. Conclusion: A combination of three aerobic exercises with one resistance exercise may result in improved balance, locomotion, and endurance in MS patients.

  20. A Big Data-driven Model for the Optimization of Healthcare Processes.

    PubMed

    Koufi, Vassiliki; Malamateniou, Flora; Vassilacopoulos, George

    2015-01-01

    Healthcare organizations increasingly navigate a highly volatile, complex environment in which technological advancements and new healthcare delivery business models are the only constants. In their effort to out-perform in this environment, healthcare organizations need to be agile enough in order to become responsive to these increasingly changing conditions. To act with agility, healthcare organizations need to discover new ways to optimize their operations. To this end, they focus on healthcare processes that guide healthcare delivery and on the technologies that support them. Business process management (BPM) and Service-Oriented Architecture (SOA) can provide a flexible, dynamic, cloud-ready infrastructure where business process analytics can be utilized to extract useful insights from mountains of raw data, and make them work in ways beyond the abilities of human brains, or IT systems from just a year ago. This paper presents a framework which provides healthcare professionals gain better insight within and across your business processes. In particular, it performs real-time analysis on process-related data in order reveal areas of potential process improvement.

  1. Agile science: creating useful products for behavior change in the real world.

    PubMed

    Hekler, Eric B; Klasnja, Predrag; Riley, William T; Buman, Matthew P; Huberty, Jennifer; Rivera, Daniel E; Martin, Cesar A

    2016-06-01

    Evidence-based practice is important for behavioral interventions but there is debate on how best to support real-world behavior change. The purpose of this paper is to define products and a preliminary process for efficiently and adaptively creating and curating a knowledge base for behavior change for real-world implementation. We look to evidence-based practice suggestions and draw parallels to software development. We argue to target three products: (1) the smallest, meaningful, self-contained, and repurposable behavior change modules of an intervention; (2) "computational models" that define the interaction between modules, individuals, and context; and (3) "personalization" algorithms, which are decision rules for intervention adaptation. The "agile science" process includes a generation phase whereby contender operational definitions and constructs of the three products are created and assessed for feasibility and an evaluation phase, whereby effect size estimates/casual inferences are created. The process emphasizes early-and-often sharing. If correct, agile science could enable a more robust knowledge base for behavior change.

  2. Agile science: creating useful products for behavior change in the real world.

    PubMed

    Hekler, Eric B; Klasnja, Predrag; Riley, William T; Buman, Matthew P; Huberty, Jennifer; Rivera, Daniel E; Martin, Cesar A

    2016-06-01

    Evidence-based practice is important for behavioral interventions but there is debate on how best to support real-world behavior change. The purpose of this paper is to define products and a preliminary process for efficiently and adaptively creating and curating a knowledge base for behavior change for real-world implementation. We look to evidence-based practice suggestions and draw parallels to software development. We argue to target three products: (1) the smallest, meaningful, self-contained, and repurposable behavior change modules of an intervention; (2) "computational models" that define the interaction between modules, individuals, and context; and (3) "personalization" algorithms, which are decision rules for intervention adaptation. The "agile science" process includes a generation phase whereby contender operational definitions and constructs of the three products are created and assessed for feasibility and an evaluation phase, whereby effect size estimates/casual inferences are created. The process emphasizes early-and-often sharing. If correct, agile science could enable a more robust knowledge base for behavior change. PMID:27357001

  3. Agile multi-scale decompositions for automatic image registration

    NASA Astrophysics Data System (ADS)

    Murphy, James M.; Leija, Omar Navarro; Le Moigne, Jacqueline

    2016-05-01

    In recent works, the first and third authors developed an automatic image registration algorithm based on a multiscale hybrid image decomposition with anisotropic shearlets and isotropic wavelets. This prototype showed strong performance, improving robustness over registration with wavelets alone. However, this method imposed a strict hierarchy on the order in which shearlet and wavelet features were used in the registration process, and also involved an unintegrated mixture of MATLAB and C code. In this paper, we introduce a more agile model for generating features, in which a flexible and user-guided mix of shearlet and wavelet features are computed. Compared to the previous prototype, this method introduces a flexibility to the order in which shearlet and wavelet features are used in the registration process. Moreover, the present algorithm is now fully coded in C, making it more efficient and portable than the mixed MATLAB and C prototype. We demonstrate the versatility and computational efficiency of this approach by performing registration experiments with the fully-integrated C algorithm. In particular, meaningful timing studies can now be performed, to give a concrete analysis of the computational costs of the flexible feature extraction. Examples of synthetically warped and real multi-modal images are analyzed.

  4. Enhancing Users' Participation in Business Process Modeling through Ontology-Based Training

    NASA Astrophysics Data System (ADS)

    Macris, A.; Malamateniou, F.; Vassilacopoulos, G.

    Successful business process design requires active participation of users who are familiar with organizational activities and business process modelling concepts. Hence, there is a need to provide users with reusable, flexible, agile and adaptable training material in order to enable them instil their knowledge and expertise in business process design and automation activities. Knowledge reusability is of paramount importance in designing training material on process modelling since it enables users participate actively in process design/redesign activities stimulated by the changing business environment. This paper presents a prototype approach for the design and use of training material that provides significant advantages to both the designer (knowledge - content reusability and semantic web enabling) and the user (semantic search, knowledge navigation and knowledge dissemination). The approach is based on externalizing domain knowledge in the form of ontology-based knowledge networks (i.e. training scenarios serving specific training needs) so that it is made reusable.

  5. Preparing your Offshore Organization for Agility: Experiences in India

    NASA Astrophysics Data System (ADS)

    Srinivasan, Jayakanth

    Two strategies that have significantly changed the way we conventionally think about managing software development and sustainment are the family of development approaches collectively referred to as agile methods, and the distribution of development efforts on a global scale. When you combine the two strategies, organizations have to address not only the technical challenges that arise from introducing new ways of working, but more importantly have to manage the 'soft' factors that if ignored lead to hard challenges. Using two case studies of distributed agile software development in India we illustrate the areas that organizations need to be aware of when transitioning work to India. The key issues that we emphasize are the need to recruit and retain personnel; the importance of teaching, mentoring and coaching; the need to manage customer expectations; the criticality of well-articulated senior leadership vision and commitment; and the reality of operating in a heterogeneous process environment.

  6. Enabling Agile Testing through Continuous Integration

    SciTech Connect

    Stolberg, Sean E.

    2009-08-24

    A Continuous Integration system is often considered one of the key elements involved in supporting an agile software development and testing environment. As a traditional software tester transitioning to an agile development environment it became clear to me that I would need to put this essential infrastructure in place and promote improved development practices in order to make the transition to agile testing possible. This experience report discusses a continuous integration implementation I lead last year. The initial motivations for implementing continuous integration are discussed and a pre and post-assessment using Martin Fowler's "Practices of Continuous Integration" is provided along with the technical specifics of the implementation. Finally, I’ll wrap up with a retrospective of my experiences implementing and promoting continuous integration within the context of agile testing.

  7. Participatory Design Activities and Agile Software Development

    NASA Astrophysics Data System (ADS)

    Kautz, Karlheinz

    This paper contributes to the studies of design activities in information systems development. It provides a case study of a large agile development project and focusses on how customers and users participated in agile development and design activities in practice. The investigated project utilized the agile method eXtreme Programming. Planning games, user stories and story cards, working software, and acceptance tests structured the customer and user involvement. We found genuine customer and user involvement in the design activities in the form of both direct and indirect participation in the agile development project. The involved customer representatives played informative, consultative, and participative roles in the project. This led to their functional empowerment— the users were enabled to carry out their work to their own satisfaction and in an effective, efficient, and economical manner.

  8. Agile Data Management with the Global Change Information System

    NASA Astrophysics Data System (ADS)

    Duggan, B.; Aulenbach, S.; Tilmes, C.; Goldstein, J.

    2013-12-01

    We describe experiences applying agile software development techniques to the realm of data management during the development of the Global Change Information System (GCIS), a web service and API for authoritative global change information under development by the US Global Change Research Program. Some of the challenges during system design and implementation have been : (1) balancing the need for a rigorous mechanism for ensuring information quality with the realities of large data sets whose contents are often in flux, (2) utilizing existing data to inform decisions about the scope and nature of new data, and (3) continuously incorporating new knowledge and concepts into a relational data model. The workflow for managing the content of the system has much in common with the development of the system itself. We examine various aspects of agile software development and discuss whether or how we have been able to use them for data curation as well as software development.

  9. Radiolysis Process Model

    SciTech Connect

    Buck, Edgar C.; Wittman, Richard S.; Skomurski, Frances N.; Cantrell, Kirk J.; McNamara, Bruce K.; Soderquist, Chuck Z.

    2012-07-17

    Assessing the performance of spent (used) nuclear fuel in geological repository requires quantification of time-dependent phenomena that may influence its behavior on a time-scale up to millions of years. A high-level waste repository environment will be a dynamic redox system because of the time-dependent generation of radiolytic oxidants and reductants and the corrosion of Fe-bearing canister materials. One major difference between used fuel and natural analogues, including unirradiated UO2, is the intense radiolytic field. The radiation emitted by used fuel can produce radiolysis products in the presence of water vapor or a thin-film of water (including OH• and H• radicals, O2-, eaq, H2O2, H2, and O2) that may increase the waste form degradation rate and change radionuclide behavior. H2O2 is the dominant oxidant for spent nuclear fuel in an O2 depleted water environment, the most sensitive parameters have been identified with respect to predictions of a radiolysis model under typical conditions. As compared with the full model with about 100 reactions it was found that only 30-40 of the reactions are required to determine [H2O2] to one part in 10–5 and to preserve most of the predictions for major species. This allows a systematic approach for model simplification and offers guidance in designing experiments for validation.

  10. Balancing Plan-Driven and Agile Methods in Software Engineering Project Courses

    NASA Astrophysics Data System (ADS)

    Boehm, Barry; Port, Dan; Winsor Brown, A.

    2002-09-01

    For the past 6 years, we have been teaching a two-semester software engineering project course. The students organize into 5-person teams and develop largely web-based electronic services projects for real USC campus clients. We have been using and evolving a method called Model- Based (System) Architecting and Software Engineering (MBASE) for use in both the course and in industrial applications. The MBASE Guidelines include a lot of documents. We teach risk-driven documentation: if it is risky to document something, and not risky to leave it out (e.g., GUI screen placements), leave it out. Even so, students tend to associate more documentation with higher grades, although our grading eventually discourages this. We are always on the lookout for ways to have students learn best practices without having to produce excessive documentation. Thus, we were very interested in analyzing the various emerging agile methods. We found that agile methods and milestone plan-driven methods are part of a “how much planning is enough?” spectrum. Both agile and plan-driven methods have home grounds of project characteristics where they clearly work best, and where the other will have difficulties. Hybrid agile/plan-driven approaches are feasible, and necessary for projects having a mix of agile and plan-driven home ground characteristics. Information technology trends are going more toward the agile methods' home ground characteristics of emergent requirements and rapid change, although there is a concurrent increase in concern with dependability. As a result, we are currently experimenting with risk-driven combinations of MBASE and agile methods, such as integrating requirements, test plans, peer reviews, and pair programming into “agile quality management.”

  11. Chemical Process Modeling and Control.

    ERIC Educational Resources Information Center

    Bartusiak, R. Donald; Price, Randel M.

    1987-01-01

    Describes some of the features of Lehigh University's (Pennsylvania) process modeling and control program. Highlights the creation and operation of the Chemical Process Modeling and Control Center (PMC). Outlines the program's philosophy, faculty, technical program, current research projects, and facilities. (TW)

  12. SAR imagery using chaotic carrier frequency agility pulses

    NASA Astrophysics Data System (ADS)

    Xu, Xiaojian; Feng, Xiangzhi

    2011-06-01

    Synthetic aperture radar (SAR) systems are getting more and more applications in both civilian and military remote sensing missions. With the increasing deployment of electronic countermeasures (ECM) on modern battlefields, SAR encounters more and more interference jamming signals. The ECM jamming signals cause the SAR system to receive and process erroneous information which results in severe degradations in the output SAR images and/or formation of phony images of nonexistent targets. As a consequence, development of the electronic counter-countermeasures (ECCM) capability becomes one of the key problems in SAR system design. This paper develops radar signaling strategies and algorithms that enhance the ability of synthetic aperture radar to image targets under conditions of electronic jamming. The concept of SAR using chaotic carrier frequency agility pulses (CCFAP-SAR) is first proposed. Then the imaging procedure for CCFAP-SAR is discussed in detail. The ECCM performance of CCFAP-SAR for both depressive noise jamming and deceptive repeat jamming is analyzed. The impact of the carrier frequency agility range on the image quality of CCFAP-SAR is also studied. Simulation results demonstrate that, with adequate agility range of the carrier frequency, the proposed CCFAP-SAR performs as well as conventional radar with linear frequency modulation (LFM) waveform in image quality and slightly better in anti-noise depressive jamming; while performs very well in anti-deception jamming which cannot be rejected by LFM-SAR.

  13. An agile mask data preparation and writer dispatching approach

    NASA Astrophysics Data System (ADS)

    Hsu, Chih-tung; Chen, Y. S.; Hsin, S. C.; Tuo, Laurent C.; Schulze, Steffen F.

    2004-08-01

    An agile mask data preparation (MDP) approach is proposed to cut re-fracture cycle time as incurred by mask writer dispatching policy changes. Shorter re-fracture cycle time increases the flexibility of mask writer dispatching, as a result, mask writer's capacity can be utilized to its optimum. Preliminary results demonstrate promising benefits in MDP cycle time reduction and writer dispatching flexibility improvement. The agile MDP can save up to 40% of re-fracture cycle time. OASIS (Open Artwork System Interchange Standard) was proposed to address the GDSII file size explosion problem. However, OASIS has yet to gain wide acceptance in the mask industry. The authors envision OASIS adoption by the mask industry as a three-phase process and identify key issues of each phase from the mask manufacturer's perspective. As a long-term MDP flow reengineering project, an agile MDP and writer dispatching approach based on OASIS is proposed. The paper describes the results of an extensive evaluation on OASIS performance compared to that of GDSII, both original GDSII and post-OPC GDSII files. The file size of eighty percent of the original GDSII files is more than ten times larger compared to that of its OASIS counterpart.

  14. Modeling nuclear processes by Simulink

    SciTech Connect

    Rashid, Nahrul Khair Alang Md

    2015-04-29

    Modelling and simulation are essential parts in the study of dynamic systems behaviours. In nuclear engineering, modelling and simulation are important to assess the expected results of an experiment before the actual experiment is conducted or in the design of nuclear facilities. In education, modelling can give insight into the dynamic of systems and processes. Most nuclear processes can be described by ordinary or partial differential equations. Efforts expended to solve the equations using analytical or numerical solutions consume time and distract attention from the objectives of modelling itself. This paper presents the use of Simulink, a MATLAB toolbox software that is widely used in control engineering, as a modelling platform for the study of nuclear processes including nuclear reactor behaviours. Starting from the describing equations, Simulink models for heat transfer, radionuclide decay process, delayed neutrons effect, reactor point kinetic equations with delayed neutron groups, and the effect of temperature feedback are used as examples.

  15. Modeling nuclear processes by Simulink

    NASA Astrophysics Data System (ADS)

    Rashid, Nahrul Khair Alang Md

    2015-04-01

    Modelling and simulation are essential parts in the study of dynamic systems behaviours. In nuclear engineering, modelling and simulation are important to assess the expected results of an experiment before the actual experiment is conducted or in the design of nuclear facilities. In education, modelling can give insight into the dynamic of systems and processes. Most nuclear processes can be described by ordinary or partial differential equations. Efforts expended to solve the equations using analytical or numerical solutions consume time and distract attention from the objectives of modelling itself. This paper presents the use of Simulink, a MATLAB toolbox software that is widely used in control engineering, as a modelling platform for the study of nuclear processes including nuclear reactor behaviours. Starting from the describing equations, Simulink models for heat transfer, radionuclide decay process, delayed neutrons effect, reactor point kinetic equations with delayed neutron groups, and the effect of temperature feedback are used as examples.

  16. Business process modeling in healthcare.

    PubMed

    Ruiz, Francisco; Garcia, Felix; Calahorra, Luis; Llorente, César; Gonçalves, Luis; Daniel, Christel; Blobel, Bernd

    2012-01-01

    The importance of the process point of view is not restricted to a specific enterprise sector. In the field of health, as a result of the nature of the service offered, health institutions' processes are also the basis for decision making which is focused on achieving their objective of providing quality medical assistance. In this chapter the application of business process modelling - using the Business Process Modelling Notation (BPMN) standard is described. Main challenges of business process modelling in healthcare are the definition of healthcare processes, the multi-disciplinary nature of healthcare, the flexibility and variability of the activities involved in health care processes, the need of interoperability between multiple information systems, and the continuous updating of scientific knowledge in healthcare. PMID:22925789

  17. Conceptual models of information processing

    NASA Technical Reports Server (NTRS)

    Stewart, L. J.

    1983-01-01

    The conceptual information processing issues are examined. Human information processing is defined as an active cognitive process that is analogous to a system. It is the flow and transformation of information within a human. The human is viewed as an active information seeker who is constantly receiving, processing, and acting upon the surrounding environmental stimuli. Human information processing models are conceptual representations of cognitive behaviors. Models of information processing are useful in representing the different theoretical positions and in attempting to define the limits and capabilities of human memory. It is concluded that an understanding of conceptual human information processing models and their applications to systems design leads to a better human factors approach.

  18. Business process modeling in healthcare.

    PubMed

    Ruiz, Francisco; Garcia, Felix; Calahorra, Luis; Llorente, César; Gonçalves, Luis; Daniel, Christel; Blobel, Bernd

    2012-01-01

    The importance of the process point of view is not restricted to a specific enterprise sector. In the field of health, as a result of the nature of the service offered, health institutions' processes are also the basis for decision making which is focused on achieving their objective of providing quality medical assistance. In this chapter the application of business process modelling - using the Business Process Modelling Notation (BPMN) standard is described. Main challenges of business process modelling in healthcare are the definition of healthcare processes, the multi-disciplinary nature of healthcare, the flexibility and variability of the activities involved in health care processes, the need of interoperability between multiple information systems, and the continuous updating of scientific knowledge in healthcare.

  19. Agile Data Curation at a State Geological Survey

    NASA Astrophysics Data System (ADS)

    Hills, D. J.

    2015-12-01

    State agencies, including geological surveys, are often the gatekeepers for myriad data products essential for scientific research and economic development. For example, the Geological Survey of Alabama (GSA) is mandated to explore for, characterize, and report Alabama's mineral, energy, water, and biological resources in support of economic development, conservation, management, and public policy for the betterment of Alabama's citizens, communities, and businesses. As part of that mandate, the GSA has increasingly been called upon to make our data more accessible to stakeholders. Even as demand for greater data accessibility grows, budgets for such efforts are often small, meaning that agencies must do more for less. Agile software development has yielded efficient, effective products, most often at lower cost and in shorter time. Taking guidance from the agile software development model, the GSA is working towards more agile data management and curation. To date, the GSA's work has been focused primarily on data rescue. By using workflows that maximize clear communication while encouraging simplicity (e.g., maximizing the amount of work not done or that can be automated), the GSA is bringing decades of dark data into the light. Regular checks by the data rescuer with the data provider (or their proxy) provides quality control without adding an overt burden on either party. Moving forward, these workflows will also allow for more efficient and effective data management.

  20. Electrophysiological models of neural processing.

    PubMed

    Nelson, Mark E

    2011-01-01

    The brain is an amazing information processing system that allows organisms to adaptively monitor and control complex dynamic interactions with their environment across multiple spatial and temporal scales. Mathematical modeling and computer simulation techniques have become essential tools in understanding diverse aspects of neural processing ranging from sub-millisecond temporal coding in the sound localization circuity of barn owls to long-term memory storage and retrieval in humans that can span decades. The processing capabilities of individual neurons lie at the core of these models, with the emphasis shifting upward and downward across different levels of biological organization depending on the nature of the questions being addressed. This review provides an introduction to the techniques for constructing biophysically based models of individual neurons and local networks. Topics include Hodgkin-Huxley-type models of macroscopic membrane currents, Markov models of individual ion-channel currents, compartmental models of neuronal morphology, and network models involving synaptic interactions among multiple neurons. PMID:21064164

  1. Models of the Reading Process

    PubMed Central

    Rayner, Keith; Reichle, Erik D.

    2010-01-01

    Reading is a complex skill involving the orchestration of a number of components. Researchers often talk about a “model of reading” when talking about only one aspect of the reading process (for example, models of word identification are often referred to as “models of reading”). Here, we review prominent models that are designed to account for (1) word identification, (2) syntactic parsing, (3) discourse representations, and (4) how certain aspects of language processing (e.g., word identification), in conjunction with other constraints (e g., limited visual acuity, saccadic error, etc.), guide readers’ eyes. Unfortunately, it is the case that these various models addressing specific aspects of the reading process seldom make contact with models dealing with other aspects of reading. Thus, for example, the models of word identification seldom make contact with models of eye movement control, and vice versa. While this may be unfortunate in some ways, it is quite understandable in other ways because reading itself is a very complex process. We discuss prototypical models of aspects of the reading process in the order mentioned above. We do not review all possible models, but rather focus on those we view as being representative and most highly recognized. PMID:21170142

  2. Models of the Reading Process.

    PubMed

    Rayner, Keith; Reichle, Erik D

    2010-11-01

    Reading is a complex skill involving the orchestration of a number of components. Researchers often talk about a "model of reading" when talking about only one aspect of the reading process (for example, models of word identification are often referred to as "models of reading"). Here, we review prominent models that are designed to account for (1) word identification, (2) syntactic parsing, (3) discourse representations, and (4) how certain aspects of language processing (e.g., word identification), in conjunction with other constraints (e g., limited visual acuity, saccadic error, etc.), guide readers' eyes. Unfortunately, it is the case that these various models addressing specific aspects of the reading process seldom make contact with models dealing with other aspects of reading. Thus, for example, the models of word identification seldom make contact with models of eye movement control, and vice versa. While this may be unfortunate in some ways, it is quite understandable in other ways because reading itself is a very complex process. We discuss prototypical models of aspects of the reading process in the order mentioned above. We do not review all possible models, but rather focus on those we view as being representative and most highly recognized.

  3. Design and characterization of frequency agile RF and microwave devices using ferroelectrics

    NASA Astrophysics Data System (ADS)

    Nath, Jayesh

    A methodology for the optimized design of tunable distributed resonators is introduced and verified. This technique enables maximum tuning with minimum degradation in quality (Q) factor. The concept of a network transformation factor and a new figure-of-merit for tunable resonators is introduced and applied to experimental data. The figure-of-merit quantifies the trade-off between tunability and Q factor for a given tuning ratio of the variable reactance device. As such, it can be extended to the design of filters, phase shifters, antennas, matching networks and other frequency-agile devices where resonant elements are used. Varactors utilizing Barium Strontium Titanate (BST) thin-film were designed and fabricated in integrated form and also in discrete form as standard 0603 components. High frequency characterization and modeling of BST varactors is described. A novel characterization technique for the intrinsic loss extraction of symmetrical two-port networks was developed and verified experimentally. Both integrated and discrete BST thin-film varactors were used to design, fabricate and characterize frequency-agile circuits. Tunable bandpass and bandstop filters and matching networks are described. A dual-mode, narrowband microstrip patch antenna with independently tunable modes was developed and characterized. Tuning and nonlinear characterization results are presented. Investigation for the use of BST thin-film varactors for voltage-controlled oscillators and phase shifters are also presented. Design parameters, fabrication issues, and processing challenges are discussed.

  4. Distributed agile software development for the SKA

    NASA Astrophysics Data System (ADS)

    Wicenec, Andreas; Parsons, Rebecca; Kitaeff, Slava; Vinsen, Kevin; Wu, Chen; Nelson, Paul; Reed, David

    2012-09-01

    The SKA software will most probably be developed by many groups distributed across the globe and coming from dierent backgrounds, like industries and research institutions. The SKA software subsystems will have to cover a very wide range of dierent areas, but still they have to react and work together like a single system to achieve the scientic goals and satisfy the challenging data ow requirements. Designing and developing such a system in a distributed fashion requires proper tools and the setup of an environment to allow for ecient detection and tracking of interface and integration issues in particular in a timely way. Agile development can provide much faster feedback mechanisms and also much tighter collaboration between the customer (scientist) and the developer. Continuous integration and continuous deployment on the other hand can provide much faster feedback of integration issues from the system level to the subsystem developers. This paper describes the results obtained from trialing a potential SKA development environment based on existing science software development processes like ALMA, the expected distribution of the groups potentially involved in the SKA development and experience gained in the development of large scale commercial software projects.

  5. Modelling of CWS combustion process

    NASA Astrophysics Data System (ADS)

    Rybenko, I. A.; Ermakova, L. A.

    2016-10-01

    The paper considers the combustion process of coal water slurry (CWS) drops. The physico-chemical process scheme consisting of several independent parallel-sequential stages is offered. This scheme of drops combustion process is proved by the particle size distribution test and research stereomicroscopic analysis of combustion products. The results of mathematical modelling and optimization of stationary regimes of CWS combustion are provided. During modeling the problem of defining possible equilibrium composition of products, which can be obtained as a result of CWS combustion processes at different temperatures, is solved.

  6. Gamma-ray Astrophysics with AGILE

    SciTech Connect

    Longo, Francesco |; Tavani, M.; Barbiellini, G.; Argan, A.; Basset, M.; Boffelli, F.; Bulgarelli, A.; Caraveo, P.; Cattaneo, P.; Chen, A.; Costa, E.; Del Monte, E.; Di Cocco, G.; Di Persio, G.; Donnarumma, I.; Feroci, M.; Fiorini, M.; Foggetta, L.; Froysland, T.; Frutti, M.

    2007-07-12

    AGILE will explore the gamma-ray Universe with a very innovative instrument combining for the first time a gamma-ray imager and a hard X-ray imager. AGILE will be operational in spring 2007 and it will provide crucial data for the study of Active Galactic Nuclei, Gamma-Ray Bursts, unidentified gamma-ray sources. Galactic compact objects, supernova remnants, TeV sources, and fundamental physics by microsecond timing. The AGILE instrument is designed to simultaneously detect and image photons in the 30 MeV - 50 GeV and 15 - 45 keV energy bands with excellent imaging and timing capabilities, and a large field of view covering {approx} 1/5 of the entire sky at energies above 30 MeV. A CsI calorimeter is capable of GRB triggering in the energy band 0.3-50 MeV AGILE is now (March 2007) undergoing launcher integration and testing. The PLSV launch is planned in spring 2007. AGILE is then foreseen to be fully operational during the summer of 2007.

  7. Fighter agility metrics. M.S. Thesis

    NASA Technical Reports Server (NTRS)

    Liefer, Randall K.

    1990-01-01

    Fighter flying qualities and combat capabilities are currently measured and compared in terms relating to vehicle energy, angular rates and sustained acceleration. Criteria based on these measurable quantities have evolved over the past several decades and are routinely used to design aircraft structures, aerodynamics, propulsion and control systems. While these criteria, or metrics, have the advantage of being well understood, easily verified and repeatable during test, they tend to measure the steady state capability of the aircraft and not its ability to transition quickly from one state to another. Proposed new metrics to assess fighter aircraft agility are collected and analyzed. A framework for classification of these new agility metrics is developed and applied. A complete set of transient agility metrics is evaluated with a high fidelity, nonlinear F-18 simulation. Test techniques and data reduction methods are proposed. A method of providing cuing information to the pilot during flight test is discussed. The sensitivity of longitudinal and lateral agility metrics to deviations from the pilot cues is studied in detail. The metrics are shown to be largely insensitive to reasonable deviations from the nominal test pilot commands. Instrumentation required to quantify agility via flight test is also considered. With one exception, each of the proposed new metrics may be measured with instrumentation currently available.

  8. Kinetic Modeling of Microbiological Processes

    SciTech Connect

    Liu, Chongxuan; Fang, Yilin

    2012-08-26

    Kinetic description of microbiological processes is vital for the design and control of microbe-based biotechnologies such as waste water treatment, petroleum oil recovery, and contaminant attenuation and remediation. Various models have been proposed to describe microbiological processes. This editorial article discusses the advantages and limiation of these modeling approaches in cluding tranditional, Monod-type models and derivatives, and recently developed constraint-based approaches. The article also offers the future direction of modeling researches that best suit for petroleum and environmental biotechnologies.

  9. Social Models: Blueprints or Processes?

    ERIC Educational Resources Information Center

    Little, Graham R.

    1981-01-01

    Discusses the nature and implications of two different models for societal planning: (1) the problem-solving process approach based on Karl Popper; and (2) the goal-setting "blueprint" approach based on Karl Marx. (DC)

  10. Contribution of Agility to Successful Distributed Software Development

    NASA Astrophysics Data System (ADS)

    Sarker, Saonee; Munson, Charles L.; Sarker, Suprateek; Chakraborty, Suranjan

    In recent times, both researchers and practitioners have touted agility as the latest innovation in distributed software development (DSD). In spite of this acknowledgement, there is little understanding and evidence surrounding the effect of agility on distributed project success. This chapter reports on a study that examines practitioner views surrounding the relative importance of different sub-types of agility to DSD project success. Preliminary results indicate that practitioners view on-time completion of DSD projects, and effective collaboration amongst stakeholders as the top two criteria of DSD project success, with lower emphasis on within-budget considerations. Among the many agility sub-types examined, people-based agility, communication-based agility, methodological agility, and time-based agility emerged as the most important for practitioners in terms of ensuring DSD project success.

  11. SuperAGILE and Gamma Ray Bursts

    SciTech Connect

    Pacciani, Luigi; Costa, Enrico; Del Monte, Ettore; Donnarumma, Immacolata; Evangelista, Yuri; Feroci, Marco; Frutti, Massimo; Lazzarotto, Francesco; Lapshov, Igor; Rubini, Alda; Soffitta, Paolo; Tavani, Marco; Barbiellini, Guido; Mastropietro, Marcello; Morelli, Ennio; Rapisarda, Massimo

    2006-05-19

    The solid-state hard X-ray imager of AGILE gamma-ray mission -- SuperAGILE -- has a six arcmin on-axis angular resolution in the 15-45 keV range, a field of view in excess of 1 steradian. The instrument is very light: 5 kg only. It is equipped with an on-board self triggering logic, image deconvolution, and it is able to transmit the coordinates of a GRB to the ground in real-time through the ORBCOMM constellation of satellites. Photon by photon Scientific Data are sent to the Malindi ground station at every contact. In this paper we review the performance of the SuperAGILE experiment (scheduled for a launch in the middle of 2006), after its first onground calibrations, and show the perspectives for Gamma Ray Bursts.

  12. Command Process Modeling & Risk Analysis

    NASA Technical Reports Server (NTRS)

    Meshkat, Leila

    2011-01-01

    Commanding Errors may be caused by a variety of root causes. It's important to understand the relative significance of each of these causes for making institutional investment decisions. One of these causes is the lack of standardized processes and procedures for command and control. We mitigate this problem by building periodic tables and models corresponding to key functions within it. These models include simulation analysis and probabilistic risk assessment models.

  13. Explaining the Obvious - How Do You Teach Agile?

    NASA Astrophysics Data System (ADS)

    Lundh, Erik

    Agile is now a hot topic and many organizations decide on adopting “agile” without really knowing how and why. This workshop will explore how fresh and seasoned agile coaches teach traditional and novel agile concepts, by example, with discussions. All participants are invited to show and tell about agile with an audience of peers. It might be the fresh first time with an audience, or golden hits that served you well for years.

  14. Agile interferometry: a non-traditional approach

    NASA Astrophysics Data System (ADS)

    Riza, Nabeel A.; Yaqoob, Zahid

    2004-11-01

    A new approach called agile interferometry is introduced to attain interferometric information with high sensitivity and scenario-based intelligence. Compared to traditional interferometric techniques, the proposed method thrives on dynamic control of the reference signal strength and detector integration time for efficient interferometric detection with high signal-to-noise ratio and significantly improved detected signal dynamic range capabilities. Theoretical analysis is presented with the operational methodology of the new approach. A high-speed optical attenuator is required in the interferometer reference arm to implement the proposed agile interferometer.

  15. Control design for future agile fighters

    NASA Technical Reports Server (NTRS)

    Murphy, Patrick C.; Davidson, John B.

    1991-01-01

    The CRAFT control design methodology is presented. CRAFT stands for the design objectives addressed, namely, Control power, Robustness, Agility, and Flying Qualities Tradeoffs. The approach combines eigenspace assignment, which allows for direct specification of eigenvalues and eigenvectors, and a graphical approach for representing control design metrics that captures numerous design goals in one composite illustration. The methodology makes use of control design metrics from four design objective areas, namely, control power, robustness, agility, and flying qualities. An example of the CRAFT methodology as well as associated design issues are presented.

  16. Agile enterprise development framework utilizing services principles for building pervasive security

    NASA Astrophysics Data System (ADS)

    Farroha, Deborah; Farroha, Bassam

    2011-06-01

    We are in an environment of continuously changing mission requirements and therefore our Information Systems must adapt to accomplish new tasks, quicker, in a more proficient manner. Agility is the only way we will be able to keep up with this change. But there are subtleties that must be considered as we adopt various agile methods: secure, protect, control and authenticate are all elements needed to posture our Information Technology systems to counteract the real and perceived threats in today's environment. Many systems have been tasked to ingest process and analyze different data sets than they were originally designed for and they have to interact with multiple new systems that were unaccounted for at design time. Leveraging the tenets of security, we have devised a new framework that takes agility into a new realm where the product will built to work in a service-based environment but is developed using agile processes. Even though these two criteria promise to hone the development effort, they actually contradict each other in philosophy where Services require stable interfaces, while Agile focuses on being flexible and tolerate changes up to much later stages of development. This framework is focused on enabling a successful product development that capitalizes on both philosophies.

  17. Neuroscientific Model of Motivational Process

    PubMed Central

    Kim, Sung-il

    2013-01-01

    Considering the neuroscientific findings on reward, learning, value, decision-making, and cognitive control, motivation can be parsed into three sub processes, a process of generating motivation, a process of maintaining motivation, and a process of regulating motivation. I propose a tentative neuroscientific model of motivational processes which consists of three distinct but continuous sub processes, namely reward-driven approach, value-based decision-making, and goal-directed control. Reward-driven approach is the process in which motivation is generated by reward anticipation and selective approach behaviors toward reward. This process recruits the ventral striatum (reward area) in which basic stimulus-action association is formed, and is classified as an automatic motivation to which relatively less attention is assigned. By contrast, value-based decision-making is the process of evaluating various outcomes of actions, learning through positive prediction error, and calculating the value continuously. The striatum and the orbitofrontal cortex (valuation area) play crucial roles in sustaining motivation. Lastly, the goal-directed control is the process of regulating motivation through cognitive control to achieve goals. This consciously controlled motivation is associated with higher-level cognitive functions such as planning, retaining the goal, monitoring the performance, and regulating action. The anterior cingulate cortex (attention area) and the dorsolateral prefrontal cortex (cognitive control area) are the main neural circuits related to regulation of motivation. These three sub processes interact with each other by sending reward prediction error signals through dopaminergic pathway from the striatum and to the prefrontal cortex. The neuroscientific model of motivational process suggests several educational implications with regard to the generation, maintenance, and regulation of motivation to learn in the learning environment. PMID:23459598

  18. Neuroscientific model of motivational process.

    PubMed

    Kim, Sung-Il

    2013-01-01

    Considering the neuroscientific findings on reward, learning, value, decision-making, and cognitive control, motivation can be parsed into three sub processes, a process of generating motivation, a process of maintaining motivation, and a process of regulating motivation. I propose a tentative neuroscientific model of motivational processes which consists of three distinct but continuous sub processes, namely reward-driven approach, value-based decision-making, and goal-directed control. Reward-driven approach is the process in which motivation is generated by reward anticipation and selective approach behaviors toward reward. This process recruits the ventral striatum (reward area) in which basic stimulus-action association is formed, and is classified as an automatic motivation to which relatively less attention is assigned. By contrast, value-based decision-making is the process of evaluating various outcomes of actions, learning through positive prediction error, and calculating the value continuously. The striatum and the orbitofrontal cortex (valuation area) play crucial roles in sustaining motivation. Lastly, the goal-directed control is the process of regulating motivation through cognitive control to achieve goals. This consciously controlled motivation is associated with higher-level cognitive functions such as planning, retaining the goal, monitoring the performance, and regulating action. The anterior cingulate cortex (attention area) and the dorsolateral prefrontal cortex (cognitive control area) are the main neural circuits related to regulation of motivation. These three sub processes interact with each other by sending reward prediction error signals through dopaminergic pathway from the striatum and to the prefrontal cortex. The neuroscientific model of motivational process suggests several educational implications with regard to the generation, maintenance, and regulation of motivation to learn in the learning environment.

  19. An agile enterprise regulation architecture for health information security management.

    PubMed

    Chen, Ying-Pei; Hsieh, Sung-Huai; Cheng, Po-Hsun; Chien, Tsan-Nan; Chen, Heng-Shuen; Luh, Jer-Junn; Lai, Jin-Shin; Lai, Feipei; Chen, Sao-Jie

    2010-09-01

    Information security management for healthcare enterprises is complex as well as mission critical. Information technology requests from clinical users are of such urgency that the information office should do its best to achieve as many user requests as possible at a high service level using swift security policies. This research proposes the Agile Enterprise Regulation Architecture (AERA) of information security management for healthcare enterprises to implement as part of the electronic health record process. Survey outcomes and evidential experiences from a sample of medical center users proved that AERA encourages the information officials and enterprise administrators to overcome the challenges faced within an electronically equipped hospital.

  20. An Agile Enterprise Regulation Architecture for Health Information Security Management

    PubMed Central

    Chen, Ying-Pei; Hsieh, Sung-Huai; Chien, Tsan-Nan; Chen, Heng-Shuen; Luh, Jer-Junn; Lai, Jin-Shin; Lai, Feipei; Chen, Sao-Jie

    2010-01-01

    Abstract Information security management for healthcare enterprises is complex as well as mission critical. Information technology requests from clinical users are of such urgency that the information office should do its best to achieve as many user requests as possible at a high service level using swift security policies. This research proposes the Agile Enterprise Regulation Architecture (AERA) of information security management for healthcare enterprises to implement as part of the electronic health record process. Survey outcomes and evidential experiences from a sample of medical center users proved that AERA encourages the information officials and enterprise administrators to overcome the challenges faced within an electronically equipped hospital. PMID:20815748

  1. Lean and Agile: An Epistemological Reflection

    ERIC Educational Resources Information Center

    Browaeys, Marie-Joelle; Fisser, Sandra

    2012-01-01

    Purpose: The aim of the paper is to contribute to the discussion of treating the concepts of lean and agile in isolation or combination by presenting an alternative view from complexity thinking on these concepts, considering an epistemological approach to this topic. Design/methodology/approach: The paper adopts an epistemological approach, using…

  2. A Roadmap for using Agile Development in a Traditional System

    NASA Technical Reports Server (NTRS)

    Streiffert, Barbara; Starbird, Thomas

    2006-01-01

    I. Ensemble Development Group: a) Produces activity planning software for in spacecraft; b) Built on Eclipse Rich Client Platform (open source development and runtime software); c) Funded by multiple sources including the Mars Technology Program; d) Incorporated the use of Agile Development. II. Next Generation Uplink Planning System: a) Researches the Activity Planning and Sequencing Subsystem for Mars Science Laboratory (APSS); b) APSS includes Ensemble, Activity Modeling, Constraint Checking, Command Editing and Sequencing tools plus other uplink generation utilities; c) Funded by the Mars Technology Program; d) Integrates all of the tools for APSS.

  3. Comparison of a New Test For Agility and Skill in Soccer With Other Agility Tests

    PubMed Central

    Kutlu, Mehmet; Yapıcı, Hakan; Yoncalık, Oğuzhan; Çelik, Serkan

    2012-01-01

    The purpose of this study was both to develop a novel test to measure run, shuttle run and directional change agility, and soccer shots on goal with decision making and to compare it with other agility tests. Multiple comparisons and assessments were conducted, including test-retest, Illinois, Zig-Zag, 30 m, Bosco, T-drill agility, and Wingate peak power tests. A total of 113 Turkish amateur and professional soccer players and tertiary-level students participated in the study. Test-retest and inter-tester reliability testing measures were conducted with athletes. The correlation coefficient of the new test was .88, with no significant difference (p> 0.01> 0.01) between the test results obtained in the first and second test sessions. The results of an analysis of variance revealed a significant (p < 0.01) difference between the T-drill agility and power test results for soccer players. The new agility and skill test is an acceptable and reliable test when considering test-retest reliability and inter-rater reliability. The findings in this study suggest that the novel soccer-specific agility and shooting test can be utilized in the testing and identification of soccer players’ talents. PMID:23486732

  4. AGILE and Gamma-Ray Bursts

    SciTech Connect

    Longo, Francesco; Tavani, M.; Barbiellini, G.; Argan, A.; Basset, M.; Boffelli, F.; Bulgarelli, A.; Caraveo, P.; Cattaneo, P.; Chen, A.; Costa, E.; Del Monte, E.; Di Cocco, G.; Di Persio, G.; Donnarumma, I.; Feroci, M.; Fiorini, M.; Foggetta, L.; Froysland, T.; Frutti, M.

    2006-05-19

    AGILE is a Scientific Mission dedicated to high-energy astrophysics supported by ASI with scientific participation of INAF and INFN. The AGILE instrument is designed to simultaneously detect and image photons in the 30 MeV - 50 GeV and 15 - 45 keV energy bands with excellent imaging and timing capabilities, and a large field of view covering {approx} 1/5 of the entire sky at energies above 30 MeV. A CsI calorimeter is capable of GRB triggering in the energy band 0.3-50 MeV. The broadband detection of GRBs and the study of implications for particle acceleration and high energy emission are primary goals of th emission. AGILE can image GRBs with 2-3 arcminutes error boxes in the hard X-ray range, and provide broadband photon-by photon detection in the 15-45 keV, 03-50 MeV, and 30 MeV-30 GeV energy ranges. Microsecond on-board photon tagging and a {approx} 100 microsecond gamma-ray detection deadtime will be crucial for fast GRB timing. On-board calculated GRB coordinates and energy fluxes will be quickly transmitted to the ground by an ORBCOMM transceiver. AGILE have recently (December 2005) completed its gamma-ray calibration. It is now (January 2006) undergoing satellite integration and testing. The PLSV launch is planned in early 2006. AGILE is then foreseen to be fully operational during the summer of 2006. It will be the only mission entirely dedicated to high-energy astrophysics above 30 MeV during the period mid-2006/mid-2007.

  5. Autonomous Guidance of Agile Small-scale Rotorcraft

    NASA Technical Reports Server (NTRS)

    Mettler, Bernard; Feron, Eric

    2004-01-01

    This report describes a guidance system for agile vehicles based on a hybrid closed-loop model of the vehicle dynamics. The hybrid model represents the vehicle dynamics through a combination of linear-time-invariant control modes and pre-programmed, finite-duration maneuvers. This particular hybrid structure can be realized through a control system that combines trim controllers and a maneuvering control logic. The former enable precise trajectory tracking, and the latter enables trajectories at the edge of the vehicle capabilities. The closed-loop model is much simpler than the full vehicle equations of motion, yet it can capture a broad range of dynamic behaviors. It also supports a consistent link between the physical layer and the decision-making layer. The trajectory generation was formulated as an optimization problem using mixed-integer-linear-programming. The optimization is solved in a receding horizon fashion. Several techniques to improve the computational tractability were investigate. Simulation experiments using NASA Ames 'R-50 model show that this approach fully exploits the vehicle's agility.

  6. AGILE OBSERVATIONS OF THE 'SOFT' GAMMA-RAY PULSAR PSR B1509 - 58

    SciTech Connect

    Pilia, M.; Pellizzoni, A.; Esposito, P.; Burgay, M.; Possenti, A.; D'Amico, N.; Trois, A.; Monte, E. Del; Tavani, M.; Argan, A.; Costa, E.; Weltevrede, P.; Johnston, S.; Fuschino, F.

    2010-11-01

    We present the results of new AGILE observations of PSR B1509 - 58 performed over a period of {approx}2.5 years following the detection obtained with a subset of the present data. The modulation significance of the light curve above 30 MeV is at a 5{sigma} confidence level and the light curve is similar to those found earlier by COMPTEL up to 30 MeV: a broad asymmetric first peak reaching its maximum 0.39 {+-} 0.02 cycles after the radio peak plus a second peak at 0.94 {+-} 0.03. The gamma-ray spectral energy distribution of the pulsed flux detected by COMPTEL and AGILE is well described by a power law (photon index {alpha} = 1.87 {+-} 0.09) with a remarkable cutoff at E{sub c} = 81 {+-} 20 MeV, representing the softest spectrum observed among gamma-ray pulsars so far. The pulsar luminosity at E > 1 MeV is L{sub {gamma}} = 4.2{sup +0.5}{sub -0.2} x 10{sup 35} erg s{sup -1}, assuming a distance of 5.2 kpc, which implies a spin-down conversion efficiency to gamma rays of {approx}0.03. The unusual soft break in the spectrum of PSR B1509 - 58 has been interpreted in the framework of polar cap models as a signature of the exotic photon-splitting process in the strong magnetic field of this pulsar. In this interpretation, our spectrum constrains the magnetic altitude of the emission point(s) at 3 km above the neutron star surface, implying that the attenuation may not be as strong as formerly suggested because pair production can substitute photon splitting into regions of the magnetosphere where the magnetic field becomes too low to sustain photon splitting. In the case of an outer-gap scenario or the two-pole caustic model, better constraints on the geometry of the emission would be needed from the radio band in order to establish whether the conditions required by the models to reproduce AGILE light curves and spectra match the polarization measurements.

  7. Thermoplastic matrix composite processing model

    NASA Technical Reports Server (NTRS)

    Dara, P. H.; Loos, A. C.

    1985-01-01

    The effects the processing parameters pressure, temperature, and time have on the quality of continuous graphite fiber reinforced thermoplastic matrix composites were quantitatively accessed by defining the extent to which intimate contact and bond formation has occurred at successive ply interfaces. Two models are presented predicting the extents to which the ply interfaces have achieved intimate contact and cohesive strength. The models are based on experimental observation of compression molded laminates and neat resin conditions, respectively. Identified as the mechanism explaining the phenomenon by which the plies bond to themselves is the theory of autohesion (or self diffusion). Theoretical predictions from the Reptation Theory between autohesive strength and contact time are used to explain the effects of the processing parameters on the observed experimental strengths. The application of a time-temperature relationship for autohesive strength predictions is evaluated. A viscoelastic compression molding model of a tow was developed to explain the phenomenon by which the prepreg ply interfaces develop intimate contact.

  8. Towards an Understanding of the Conceptual Underpinnings of Agile Development Methodologies

    NASA Astrophysics Data System (ADS)

    Nerur, Sridhar; Cannon, Alan; Balijepally, Venugopal; Bond, Philip

    While the growing popularity of agile development methodologies is undeniable, there has been little systematic exploration of its intellectual foundation. Such an effort would be an important first step in understanding this paradigm's underlying premises. This understanding, in turn, would be invaluable in our assessment of current practices as well as in our efforts to advance the field of software engineering. Drawing on a variety of sources, both within and outside the discipline, we argue that the concepts underpinning agile development methodologies are by no means novel. In the tradition of General Systems Theory this paper advocates a transdisciplinary examination of agile development methodologies to extend the intellectual boundaries of software development. This is particularly important as the field moves beyond instrumental processes aimed at satisfying mere technical considerations.

  9. Array Databases: Agile Analytics (not just) for the Earth Sciences

    NASA Astrophysics Data System (ADS)

    Baumann, P.; Misev, D.

    2015-12-01

    Gridded data, such as images, image timeseries, and climate datacubes, today are managed separately from the metadata, and with different, restricted retrieval capabilities. While databases are good at metadata modelled in tables, XML hierarchies, or RDF graphs, they traditionally do not support multi-dimensional arrays.This gap is being closed by Array Databases, pioneered by the scalable rasdaman ("raster data manager") array engine. Its declarative query language, rasql, extends SQL with array operators which are optimized and parallelized on server side. Installations can easily be mashed up securely, thereby enabling large-scale location-transparent query processing in federations. Domain experts value the integration with their commonly used tools leading to a quick learning curve.Earth, Space, and Life sciences, but also Social sciences as well as business have massive amounts of data and complex analysis challenges that are answered by rasdaman. As of today, rasdaman is mature and in operational use on hundreds of Terabytes of timeseries datacubes, with transparent query distribution across more than 1,000 nodes. Additionally, its concepts have shaped international Big Data standards in the field, including the forthcoming array extension to ISO SQL, many of which are supported by both open-source and commercial systems meantime. In the geo field, rasdaman is reference implementation for the Open Geospatial Consortium (OGC) Big Data standard, WCS, now also under adoption by ISO. Further, rasdaman is in the final stage of OSGeo incubation.In this contribution we present array queries a la rasdaman, describe the architecture and novel optimization and parallelization techniques introduced in 2015, and put this in context of the intercontinental EarthServer initiative which utilizes rasdaman for enabling agile analytics on Petascale datacubes.

  10. Welding process modelling and control

    NASA Technical Reports Server (NTRS)

    Romine, Peter L.; Adenwala, Jinen A.

    1993-01-01

    The research and analysis performed, and software developed, and hardware/software recommendations made during 1992 in development of the PC-based data acquisition system for support of Welding Process Modeling and Control is reported. A need was identified by the Metals Processing Branch of NASA Marshall Space Flight Center, for a mobile data aquisition and analysis system, customized for welding measurement and calibration. Several hardware configurations were evaluated and a PC-based system was chosen. The Welding Measurement System (WMS) is a dedicated instrument, strictly for the use of data aquisition and analysis. Although the WMS supports many of the functions associated with the process control, it is not the intention for this system to be used for welding process control.

  11. Architecture and performances of the AGILE Telemetry Preprocessing System (TMPPS)

    NASA Astrophysics Data System (ADS)

    Trifoglio, M.; Bulgarelli, A.; Gianotti, F.; Lazzarotto, F.; Di Cocco, G.; Fuschino, F.; Tavani, M.

    2008-07-01

    AGILE is an Italian Space Agency (ASI) satellite dedicated to high energy Astrophysics. It was launched successfully on 23 April 2007, and it has been operated by the AGILE Ground Segment, consisting of the Ground Station located in Malindi (Kenia), the Mission Operations Centre (MOC) and the AGILE Data Centre (ADC) established in Italy, at Telespazio in Fucino and at the ASI Science Data Centre (ASDC) in Frascati respectively. Due to the low equatorial orbit at ~ 530 Km. with inclination angle of ~ 2.5°, the satellite passes over the Ground Station every ~ 100'. During the visibility period of . ~ 12', the Telemetry (TM) is down linked through two separated virtual channels, VC0 and VC1. The former is devoted to the real time TM generated during the pass at the average rate of 50 Kbit/s and is directly relayed to the Control Centre. The latter is used to downlink TM data collected on the satellite on-board mass memory during the non visibility period. This generates at the Ground Station a raw TM file of up to 37 MByte. Within 20' after the end of the contact, both the real time and mass memory TM arrive at ADC through the dedicated VPN ASINet. Here they are automatically detected and ingested by the TMPPS pipeline in less than 5 minutes. The TMPPS archives each TM file and sorts its packets into one stream for each of the different TM layout. Each stream is processed in parallel in order to unpack the various telemetry field and archive them into suitable FITS files. Each operation is tracked into a MySQL data base which interfaces the TMPPS pipeline to the rest of the scientific pipeline running at ADC. In this paper the architecture and the performance of the TMPPS will be described and discussed.

  12. Insights into Global Health Practice from the Agile Software Development Movement.

    PubMed

    Flood, David; Chary, Anita; Austad, Kirsten; Diaz, Anne Kraemer; García, Pablo; Martinez, Boris; Canú, Waleska López; Rohloff, Peter

    2016-01-01

    Global health practitioners may feel frustration that current models of global health research, delivery, and implementation are overly focused on specific interventions, slow to provide health services in the field, and relatively ill-equipped to adapt to local contexts. Adapting design principles from the agile software development movement, we propose an analogous approach to designing global health programs that emphasizes tight integration between research and implementation, early involvement of ground-level health workers and program beneficiaries, and rapid cycles of iterative program improvement. Using examples from our own fieldwork, we illustrate the potential of 'agile global health' and reflect on the limitations, trade-offs, and implications of this approach.

  13. Lesson Learned from AGILE and LARES ASI Projects About MATED Data Collection and Post Analysis

    NASA Astrophysics Data System (ADS)

    Carpentiero, Rita; Mrchetti, Ernesto; Natalucci, Silvia; Portelli, Claudio

    2012-07-01

    ASI has managed and collected data on project development of two scientific all-Italian missions: AGILE and LARES. Collection of the Model And Test Effectiveness Database (MATED) data, concerning Project, AIV (Assembly Integration and Verification) and NCR (Non Conformance Report) aspects has been performed by the Italian Space Agency (ASI), using available technical documentation of both AGILE e LARES projects. In this paper some consideration on the needs of 'real time' data collection is made, together with proposal of front end improvement of this tool. In addition a preliminary analysis of MATED effectiveness related to the above ASI projects will be presented in a bottom-up and post verification approach.

  14. Animal models and conserved processes

    PubMed Central

    2012-01-01

    Background The concept of conserved processes presents unique opportunities for using nonhuman animal models in biomedical research. However, the concept must be examined in the context that humans and nonhuman animals are evolved, complex, adaptive systems. Given that nonhuman animals are examples of living systems that are differently complex from humans, what does the existence of a conserved gene or process imply for inter-species extrapolation? Methods We surveyed the literature including philosophy of science, biological complexity, conserved processes, evolutionary biology, comparative medicine, anti-neoplastic agents, inhalational anesthetics, and drug development journals in order to determine the value of nonhuman animal models when studying conserved processes. Results Evolution through natural selection has employed components and processes both to produce the same outcomes among species but also to generate different functions and traits. Many genes and processes are conserved, but new combinations of these processes or different regulation of the genes involved in these processes have resulted in unique organisms. Further, there is a hierarchy of organization in complex living systems. At some levels, the components are simple systems that can be analyzed by mathematics or the physical sciences, while at other levels the system cannot be fully analyzed by reducing it to a physical system. The study of complex living systems must alternate between focusing on the parts and examining the intact whole organism while taking into account the connections between the two. Systems biology aims for this holism. We examined the actions of inhalational anesthetic agents and anti-neoplastic agents in order to address what the characteristics of complex living systems imply for inter-species extrapolation of traits and responses related to conserved processes. Conclusion We conclude that even the presence of conserved processes is insufficient for inter

  15. Face Processing: Models For Recognition

    NASA Astrophysics Data System (ADS)

    Turk, Matthew A.; Pentland, Alexander P.

    1990-03-01

    The human ability to process faces is remarkable. We can identify perhaps thousands of faces learned throughout our lifetime and read facial expression to understand such subtle qualities as emotion. These skills are quite robust, despite sometimes large changes in the visual stimulus due to expression, aging, and distractions such as glasses or changes in hairstyle or facial hair. Computers which model and recognize faces will be useful in a variety of applications, including criminal identification, human-computer interface, and animation. We discuss models for representing faces and their applicability to the task of recognition, and present techniques for identifying faces and detecting eye blinks.

  16. Theoretical Models of Astrochemical Processes

    NASA Technical Reports Server (NTRS)

    Charnley, Steven

    2009-01-01

    Interstellar chemistry provides a natural laboratory for studying exotic species and processes at densities, temperatures, and reaction rates. that are difficult or impractical to address in the laboratory. Thus, many chemical reactions considered too sloe by the standards of terrestrial chemistry, can be 'observed and modeled. Curious proposals concerning the nature and chemistry of complex interstellar organic molecules will be described. Catalytic reactions on "rain surfaces can, in principle, lead to a lame variety of species and this has motivated many laboratory and theoretical studies. Gas phase processes may also build lame species in molecular clouds. Future laboratory data and computational tools needed to construct accurate chemical models of various astronomical sources to be observed by Herschel and ALMA will be outlined.

  17. The Test Equipment of the AGILE Minicalorimeter Prototype

    SciTech Connect

    Trifoglio, M.; Bulgarelli, A.; Gianotti, F.; Celesti, E.; Di Cocco, G.; Labanti, C.; Mauri, A.; Prest, M.; Vallazza, E.; Froysland, T.

    2004-09-28

    AGILE is an ASI (Italian Space Agency) Small Space Mission for high energy astrophysics in the range 30 MeV - 50 GeV. The AGILE satellite is currently in the C phase and is planned to be launched in 2005. The Payload shall consist of a Tungsten-Silicon Tracker, a CsI Minicalorimeter, an anticoincidence system and a X-Ray detector sensitive in the 10-40 KeV range. The purpose of the Minicalorimeter (MCAL) is twofold. It shall work in conjunction with the Tracker in order to evaluate the energy of the interacting photons, and it shall operate autonomously in the energy range 250KeV-250 MeV for detection of transients and gamma ray burst events and for the measurement of gamma ray background fluctuations. We present the architecture of the Test Equipment we have designed and developed in order to test and verify the MCAL Simplified Electrical Model prototype which has been manufactured in order to validate the design of the MCAL Proto Flight Model.

  18. Cross Sectional Study of Agile Software Development Methods and Project Performance

    ERIC Educational Resources Information Center

    Lambert, Tracy

    2011-01-01

    Agile software development methods, characterized by delivering customer value via incremental and iterative time-boxed development processes, have moved into the mainstream of the Information Technology (IT) industry. However, despite a growing body of research which suggests that a predictive manufacturing approach, with big up-front…

  19. An Agile Constructionist Mentoring Methodology for Software Projects in the High School

    ERIC Educational Resources Information Center

    Meerbaum-Salant, Orni; Hazzan, Orit

    2010-01-01

    This article describes the construction process and evaluation of the Agile Constructionist Mentoring Methodology (ACMM), a mentoring method for guiding software development projects in the high school. The need for such a methodology has arisen due to the complexity of mentoring software project development in the high school. We introduce the…

  20. Sharing environmental models: An Approach using GitHub repositories and Web Processing Services

    NASA Astrophysics Data System (ADS)

    Stasch, Christoph; Nuest, Daniel; Pross, Benjamin

    2016-04-01

    accordingly. The admin tool of the 52°North WPS was extended to support automated retrieval and deployment of computational models from GitHub repositories. Once the R code is available in the GitHub repo, the contained process can be easily deployed and executed by simply defining the GitHub repository URL in the WPS admin tool. We illustrate the usage of the approach by sharing and running a model for land use system archetypes developed by the Helmholtz Centre for Environmental Research (UFZ, see Vaclavik et al.). The original R code was extended and published in the 52°North WPS using both, public and non-public datasets (Nüst et al., see also https://github.com/52North/glues-wps). Hosting the analysis in a Git repository now allows WPS administrators, client developers, and modelers to easily work together on new versions or completely new web processes using the powerful GitHub collaboration platform. References: Hinz, M. et. al. (2013): Spatial Statistics on the Geospatial Web. In: The 16th AGILE International Conference on Geographic Information Science, Short Papers. http://www.agile-online.org/Conference_Paper/CDs/agile_2013/Short_Papers/SP_S3.1_Hinz.pdf Nüst, D. et. al.: (2015): Open and reproducible global land use classification. In: EGU General Assembly Conference Abstracts . Vol. 17. European Geophysical Union, 2015, p. 9125, http://meetingorganizer.copernicus. org/EGU2015/EGU2015- 9125.pdf Vaclavik, T., et. al. (2013): Mapping global land system archetypes. Global Environmental Change 23(6): 1637-1647. Online available: October 9, 2013, DOI: 10.1016/j.gloenvcha.2013.09.004

  1. Prospects for High Energy Detection of Microquasars with the AGILE and GLAST Gamma-Ray Telescopes

    SciTech Connect

    Santolamazza, Patrizia; Pittori, Carlotta; Verrecchia, Francesco

    2007-08-21

    We estimate the sensitivities of the AGILE and GLAST {gamma}-ray experiments taking into account two cases for the galactic {gamma}-ray diffuse background (at high galactic latitude and toward the galactic center). Then we use sensitivities to estimate microquasar observability with the two experiments, assuming the {gamma}-ray emission above 100 MeV of a recent microquasar model.

  2. Architecture-Centric Methods and Agile Approaches

    NASA Astrophysics Data System (ADS)

    Babar, Muhammad Ali; Abrahamsson, Pekka

    Agile software development approaches have had significant impact on industrial software development practices. Despite becoming widely popular, there is an increasing perplexity about the role and importance of a system’s software architecture in agile approaches [1, 2]. Advocates of the vital role of architecture in achieving quality goals of large-scale-software-intensive-systems are skeptics of the scalability of any development approach that does not pay sufficient attention to architectural issues. However, the proponents of agile approaches usually perceive the upfront design and evaluation of architecture as being of less value to the customers of a system. According to them, for example, re-factoring can help fix most of the problems. Many experiences show that large-scale re-factoring often results in significant defects, which are very costly to address later in the development cycle. It is considered that re-factoring is worthwhile as long as the high-level design is good enough to limit the need for large-scale re-factoring [1, 3, 4].

  3. First GRB detections with the AGILE Minicalorimeter

    NASA Astrophysics Data System (ADS)

    Marisaldi, M.; Labanti, C.; Fuschino, F.; Galli, M.; Tavani, M.; Bulgarelli, A.; Gianotti, F.; Trifoglio, M.; Argan, A.

    2008-05-01

    The Minicalorimeter (MCAL) onboard the AGILE satellite is a 1400 cm2 scintillation detector sensitive in the energy range 0.3-200 MeV. MCAL works both as a slave of the AGILE Silicon Tracker and as an autonomous detector for transient events (BURST mode). A dedicated onboard Burst Search logic scans BURST mode data in search of count rate increase. Peculiar characteristics of the detector are the high energy spectral coverage and a timing resolution of about 2 microseconds. Even if a trigger is not issued, BURST mode data are used to build a broad band energy spectrum (scientific ratemeters) organized in 11 bands for each of the two MCAL detection planes, with a time resolution of 1 second. After the first engineering commissioning phase, following the AGILE launch on 23rd April 2007, between 22nd June and 5th November 2007 eighteen GRBs were detected offline in the scientific ratemeters data, with a detection rate of about one per week. In this paper the capabilities of the detector will be described and an overview of the first detected GRBs will be given.

  4. First GRB detections with the AGILE Minicalorimeter

    SciTech Connect

    Marisaldi, M.; Labanti, C.; Fuschino, F.; Bulgarelli, A.; Gianotti, F.; Trifoglio, M.; Galli, M.; Tavani, M.; Argan, A.

    2008-05-22

    The Minicalorimeter (MCAL) onboard the AGILE satellite is a 1400 cm{sup 2} scintillation detector sensitive in the energy range 0.3-200 MeV. MCAL works both as a slave of the AGILE Silicon Tracker and as an autonomous detector for transient events (BURST mode). A dedicated onboard Burst Search logic scans BURST mode data in search of count rate increase. Peculiar characteristics of the detector are the high energy spectral coverage and a timing resolution of about 2 microseconds. Even if a trigger is not issued, BURST mode data are used to build a broad band energy spectrum (scientific ratemeters) organized in 11 bands for each of the two MCAL detection planes, with a time resolution of 1 second. After the first engineering commissioning phase, following the AGILE launch on 23rd April 2007, between 22nd June and 5th November 2007 eighteen GRBs were detected offline in the scientific ratemeters data, with a detection rate of about one per week. In this paper the capabilities of the detector will be described and an overview of the first detected GRBs will be given.

  5. Frequency-agile CO2 DIAL for environmental monitoring

    NASA Astrophysics Data System (ADS)

    Carr, Lewis W.; Fletcher, Leland; Crittenden, Max; Carlisle, Clinton B.; Gotoff, Steve W.; Reyes, Felix; D'Amico, Francis

    1994-06-01

    SRI International has designed and developed a fully automated frequency-agile CO2 DIAL (differential absorption lidar) system. The system sensor head consists of a single, frequency- agile, CO2, TEA laser; a 10-inch receiver telescope, a liquid-nitrogen-cooled HgCdTe detector; and a transmit energy monitor. The sensor head and its auxiliary equipment (including the data acquisition and processing system, laser power supply, and water cooler) are mounted in a Grumman-Olson 11-ft step van. The self-contained, mobile system can be used to detect and quantify many volatile organic compounds (VOCs) at parts per million sensitivities over open-path ranges to 5 km. Characterization and demonstration of the system is ongoing. However, data collected on benzene, toluene, xylene, methanol, ethyl acetate, acetic anhydride, and other VOCs will be described herein. The system could be used by industry and government agencies in stand-off monitoring to map VOC emission sources and transport patterns into surrounding communities. A single mobile system could be used for several locations to verify compliance with environmental regulations such as the 1990 Clean Air Act Amendments.

  6. Gamma-ray blazars: The view from AGILE

    NASA Astrophysics Data System (ADS)

    D'Ammando, F.; Bulgarelli, A.; Chen, A. W.; Donnarumma, I.; Giuliani, A.; Longo, F.; Pacciani, L.; Pucella, G.; Striani, E.; Tavani, M.; Vercellone, S.; Vittorini, V.; Covino, S.; Krimm, H. A.; Raiteri, C. M.; Romano, P.; Villata, M.

    2011-07-01

    During the first 3 years of operation the Gamma-Ray Imaging Detector onboard the AGILE satellite detected several blazars in a high γ-ray activity: 3C 279, 3C 454.3, PKS 1510-089, S5 0716+714, 3C 273, W Comae, Mrk 421, PKS 0537-441 and 4C +21.35. Thanks to the rapid dissemination of our alerts, we were able to obtain multiwavelength data from other observatories such as Spitzer, Swift, RXTE, Suzaku, INTEGRAL, MAGIC, VERITAS, and ARGO as well as radio-to-optical coverage by means of the GASP Project of the WEBT and the REM Telescope. This large multifrequency coverage gave us the opportunity to study the variability correlations between the emission at different frequencies and to obtain simultaneous Spectral Energy Distributions of these sources from radio to γ-ray energy bands, investigating the different mechanisms responsible for their emission and uncovering in some cases a more complex behavior with respect to the standard models. We present a review of the most interesting AGILE results on these γ-ray blazars and their multifrequency data.

  7. Candidate control design metrics for an agile fighter

    NASA Technical Reports Server (NTRS)

    Murphy, Patrick C.; Bailey, Melvin L.; Ostroff, Aaron J.

    1991-01-01

    Success in the fighter combat environment of the future will certainly demand increasing capability from aircraft technology. These advanced capabilities in the form of superagility and supermaneuverability will require special design techniques which translate advanced air combat maneuvering requirements into design criteria. Control design metrics can provide some of these techniques for the control designer. Thus study presents an overview of control design metrics and investigates metrics for advanced fighter agility. The objectives of various metric users, such as airframe designers and pilots, are differentiated from the objectives of the control designer. Using an advanced fighter model, metric values are documented over a portion of the flight envelope through piloted simulation. These metric values provide a baseline against which future control system improvements can be compared and against which a control design methodology can be developed. Agility is measured for axial, pitch, and roll axes. Axial metrics highlight acceleration and deceleration capabilities under different flight loads and include specific excess power measurements to characterize energy meneuverability. Pitch metrics cover both body-axis and wind-axis pitch rates and accelerations. Included in pitch metrics are nose pointing metrics which highlight displacement capability between the nose and the velocity vector. Roll metrics (or torsion metrics) focus on rotational capability about the wind axis.

  8. Future Research in Agile Systems Development: Applying Open Innovation Principles Within the Agile Organisation

    NASA Astrophysics Data System (ADS)

    Conboy, Kieran; Morgan, Lorraine

    A particular strength of agile approaches is that they move away from ‘introverted' development and intimately involve the customer in all areas of development, supposedly leading to the development of a more innovative and hence more valuable information system. However, we argue that a single customer representative is too narrow a focus to adopt and that involvement of stakeholders beyond the software development itself is still often quite weak and in some cases non-existent. In response, we argue that current thinking regarding innovation in agile development needs to be extended to include multiple stakeholders outside the business unit. This paper explores the intra-organisational applicability and implications of open innovation in agile systems development. Additionally, it argues for a different perspective of project management that includes collaboration and knowledge-sharing with other business units, customers, partners, and other relevant stakeholders pertinent to the business success of an organisation, thus embracing open innovation principles.

  9. End users transforming experiences into formal information and process models for personalised health interventions.

    PubMed

    Lindgren, Helena; Lundin-Olsson, Lillemor; Pohl, Petra; Sandlund, Marlene

    2014-01-01

    Five physiotherapists organised a user-centric design process of a knowledge-based support system for promoting exercise and preventing falls. The process integrated focus group studies with 17 older adults and prototyping. The transformation of informal medical and rehabilitation expertise and older adults' experiences into formal information and process models during the development was studied. As tool they used ACKTUS, a development platform for knowledge-based applications. The process became agile and incremental, partly due to the diversity of expectations and preferences among both older adults and physiotherapists, and the participatory approach to design and development. In addition, there was a need to develop the knowledge content alongside with the formal models and their presentations, which allowed the participants to test hands-on and evaluate the ideas, content and design. The resulting application is modular, extendable, flexible and adaptable to the individual end user. Moreover, the physiotherapists are able to modify the information and process models, and in this way further develop the application. The main constraint was found to be the lack of support for the initial phase of concept modelling, which lead to a redesigned user interface and functionality of ACKTUS.

  10. Applying Agile MethodstoWeapon/Weapon-Related Software

    SciTech Connect

    Adams, D; Armendariz, M; Blackledge, M; Campbell, F; Cloninger, M; Cox, L; Davis, J; Elliott, M; Granger, K; Hans, S; Kuhn, C; Lackner, M; Loo, P; Matthews, S; Morrell, K; Owens, C; Peercy, D; Pope, G; Quirk, R; Schilling, D; Stewart, A; Tran, A; Ward, R; Williamson, M

    2007-05-02

    This white paper provides information and guidance to the Department of Energy (DOE) sites on Agile software development methods and the impact of their application on weapon/weapon-related software development. The purpose of this white paper is to provide an overview of Agile methods, examine the accepted interpretations/uses/practices of these methodologies, and discuss the applicability of Agile methods with respect to Nuclear Weapons Complex (NWC) Technical Business Practices (TBPs). It also provides recommendations on the application of Agile methods to the development of weapon/weapon-related software.

  11. NUMERICAL MODELING OF FINE SEDIMENT PHYSICAL PROCESSES.

    USGS Publications Warehouse

    Schoellhamer, David H.

    1985-01-01

    Fine sediment in channels, rivers, estuaries, and coastal waters undergo several physical processes including flocculation, floc disruption, deposition, bed consolidation, and resuspension. This paper presents a conceptual model and reviews mathematical models of these physical processes. Several general fine sediment models that simulate some of these processes are reviewed. These general models do not directly simulate flocculation and floc disruption, but the conceptual model and existing functions are shown to adequately model these two processes for one set of laboratory data.

  12. Wideband Agile Digital Microwave Radiometer

    NASA Technical Reports Server (NTRS)

    Gaier, Todd C.; Brown, Shannon T.; Ruf, Christopher; Gross, Steven

    2012-01-01

    The objectives of this work were to take the initial steps needed to develop a field programmable gate array (FPGA)- based wideband digital radiometer backend (>500 MHz bandwidth) that will enable passive microwave observations with minimal performance degradation in a radiofrequency-interference (RFI)-rich environment. As manmade RF emissions increase over time and fill more of the microwave spectrum, microwave radiometer science applications will be increasingly impacted in a negative way, and the current generation of spaceborne microwave radiometers that use broadband analog back ends will become severely compromised or unusable over an increasing fraction of time on orbit. There is a need to develop a digital radiometer back end that, for each observation period, uses digital signal processing (DSP) algorithms to identify the maximum amount of RFI-free spectrum across the radiometer band to preserve bandwidth to minimize radiometer noise (which is inversely related to the bandwidth). Ultimately, the objective is to incorporate all processing necessary in the back end to take contaminated input spectra and produce a single output value free of manmade signals to minimize data rates for spaceborne radiometer missions. But, to meet these objectives, several intermediate processing algorithms had to be developed, and their performance characterized relative to typical brightness temperature accuracy re quirements for current and future microwave radiometer missions, including those for measuring salinity, soil moisture, and snow pack.

  13. Between Oais and Agile a Dynamic Data Management Approach

    NASA Astrophysics Data System (ADS)

    Bennett, V. L.; Conway, E. A.; Waterfall, A. M.; Pepler, S.

    2015-12-01

    In this paper we decribe an approach to the integration of existing archival activities which lies between compliance with the more rigid OAIS/TRAC standards and a more flexible "Agile" approach to the curation and preservation of Earth Observation data. We provide a high level overview of existing practice and discuss how these procedures can be extended and supported through the description of preservation state. The aim of which is to facilitate the dynamic controlled management of scientific data through its lifecycle. While processes are considered they are not statically defined but rather driven by human interactions in the form of risk management/review procedure that produce actionable plans, which are responsive to change. We then proceed by describing the feasibility testing of extended risk management and planning procedures which integrate current practices. This was done through the CEDA Archival Format Audit which inspected British Atmospheric Data Centre and NERC Earth Observation Data Centre Archival holdings. These holdings are extensive, comprising of around 2 Petabytes of data and 137 million individual files, which were analysed and characterised in terms of format, based risk. We are then able to present an overview of the format based risk burden faced by a large scale archive attempting to maintain the usability of heterogeneous environmental data sets We continue by presenting a dynamic data management information model and provide discussion of the following core model entities and their relationships: Aspirational entities, which include Data Entity definitions and their associated Preservation Objectives. Risk entities, which act as drivers for change within the data lifecycle. These include Acquisitional Risks, Technical Risks, Strategic Risks and External Risks Plan entities, which detail the actions to bring about change within an archive. These include Acquisition Plans, Preservation Plans and Monitoring plans which support

  14. Chaste: using agile programming techniques to develop computational biology software.

    PubMed

    Pitt-Francis, Joe; Bernabeu, Miguel O; Cooper, Jonathan; Garny, Alan; Momtahan, Lee; Osborne, James; Pathmanathan, Pras; Rodriguez, Blanca; Whiteley, Jonathan P; Gavaghan, David J

    2008-09-13

    Cardiac modelling is the area of physiome modelling where the available simulation software is perhaps most mature, and it therefore provides an excellent starting point for considering the software requirements for the wider physiome community. In this paper, we will begin by introducing some of the most advanced existing software packages for simulating cardiac electrical activity. We consider the software development methods used in producing codes of this type, and discuss their use of numerical algorithms, relative computational efficiency, usability, robustness and extensibility. We then go on to describe a class of software development methodologies known as test-driven agile methods and argue that such methods are more suitable for scientific software development than the traditional academic approaches. A case study is a project of our own, Cancer, Heart and Soft Tissue Environment, which is a library of computational biology software that began as an experiment in the use of agile programming methods. We present our experiences with a review of our progress thus far, focusing on the advantages and disadvantages of this new approach compared with the development methods used in some existing packages. We conclude by considering whether the likely wider needs of the cardiac modelling community are currently being met and suggest that, in order to respond effectively to changing requirements, it is essential that these codes should be more malleable. Such codes will allow for reliable extensions to include both detailed mathematical models--of the heart and other organs--and more efficient numerical techniques that are currently being developed by many research groups worldwide. PMID:18565813

  15. Chaste: using agile programming techniques to develop computational biology software.

    PubMed

    Pitt-Francis, Joe; Bernabeu, Miguel O; Cooper, Jonathan; Garny, Alan; Momtahan, Lee; Osborne, James; Pathmanathan, Pras; Rodriguez, Blanca; Whiteley, Jonathan P; Gavaghan, David J

    2008-09-13

    Cardiac modelling is the area of physiome modelling where the available simulation software is perhaps most mature, and it therefore provides an excellent starting point for considering the software requirements for the wider physiome community. In this paper, we will begin by introducing some of the most advanced existing software packages for simulating cardiac electrical activity. We consider the software development methods used in producing codes of this type, and discuss their use of numerical algorithms, relative computational efficiency, usability, robustness and extensibility. We then go on to describe a class of software development methodologies known as test-driven agile methods and argue that such methods are more suitable for scientific software development than the traditional academic approaches. A case study is a project of our own, Cancer, Heart and Soft Tissue Environment, which is a library of computational biology software that began as an experiment in the use of agile programming methods. We present our experiences with a review of our progress thus far, focusing on the advantages and disadvantages of this new approach compared with the development methods used in some existing packages. We conclude by considering whether the likely wider needs of the cardiac modelling community are currently being met and suggest that, in order to respond effectively to changing requirements, it is essential that these codes should be more malleable. Such codes will allow for reliable extensions to include both detailed mathematical models--of the heart and other organs--and more efficient numerical techniques that are currently being developed by many research groups worldwide.

  16. Agile informatics: application of agile project management to the development of a personal health application.

    PubMed

    Chung, Jeanhee; Pankey, Evan; Norris, Ryan J

    2007-01-01

    We describe the application of the Agile method-- a short iteration cycle, user responsive, measurable software development approach-- to the project management of a modular personal health record, iHealthSpace, to be deployed to the patients and providers of a large academic primary care practice. PMID:18694014

  17. Agile informatics: application of agile project management to the development of a personal health application.

    PubMed

    Chung, Jeanhee; Pankey, Evan; Norris, Ryan J

    2007-10-11

    We describe the application of the Agile method-- a short iteration cycle, user responsive, measurable software development approach-- to the project management of a modular personal health record, iHealthSpace, to be deployed to the patients and providers of a large academic primary care practice.

  18. Process modeling and control in foundry operations

    NASA Astrophysics Data System (ADS)

    Piwonka, T. S.

    1989-02-01

    Initial uses of process modeling were limited to phenomenological descriptions of the physical processes in foundry operations, with the aim of decreasing scrap and rework. It is now clear that process modeling can be used to select, design and optimize foundry processes so that on-line process control can be achieved. Computational, analogue and empirical process models have been developed for sand casting operations, and they are being applied in the foundry with beneficial effects.

  19. Managing Analysis Models in the Design Process

    NASA Technical Reports Server (NTRS)

    Briggs, Clark

    2006-01-01

    Design of large, complex space systems depends on significant model-based support for exploration of the design space. Integrated models predict system performance in mission-relevant terms given design descriptions and multiple physics-based numerical models. Both the design activities and the modeling activities warrant explicit process definitions and active process management to protect the project from excessive risk. Software and systems engineering processes have been formalized and similar formal process activities are under development for design engineering and integrated modeling. JPL is establishing a modeling process to define development and application of such system-level models.

  20. Scaling Agile Infrastructure to People

    NASA Astrophysics Data System (ADS)

    Jones, B.; McCance, G.; Traylen, S.; Barrientos Arias, N.

    2015-12-01

    When CERN migrated its infrastructure away from homegrown fabric management tools to emerging industry-standard open-source solutions, the immediate technical challenges and motivation were clear. The move to a multi-site Cloud Computing model meant that the tool chains that were growing around this ecosystem would be a good choice, the challenge was to leverage them. The use of open-source tools brings challenges other than merely how to deploy them. Homegrown software, for all the deficiencies identified at the outset of the project, has the benefit of growing with the organization. This paper will examine what challenges there were in adapting open-source tools to the needs of the organization, particularly in the areas of multi-group development and security. Additionally, the increase in scale of the plant required changes to how Change Management was organized and managed. Continuous Integration techniques are used in order to manage the rate of change across multiple groups, and the tools and workflow for this will be examined.

  1. Frequency/phase agile microwave circuits on ferroelectric films

    NASA Astrophysics Data System (ADS)

    Romanofsky, Robert Raymond

    This work describes novel microwave circuits that can be tuned in either frequency or phase through the use of nonlinear dielectrics, specifically thin ferroelectric films. These frequency and phase agile circuits in many cases provide a new capability or offer the potential for lower cost alternatives in satellite and terrestrial communications and sensor applications. A brief introduction to nonlinear dielectrics and a summary of some of the special challenges confronting the practical insertion of ferroelectric technology into commercial systems is provided. A theoretical solution for the propagation characteristics of the multi-layer structures, with emphasis on a new type of phase shifter based on coupled microstrip, lines, is developed. The quasi-TEM analysis is based on a variational solution for line capacitance and an extension of coupled transmission line theory. It is shown that the theoretical model is applicable to a broad class of multi-layer transmission lines. The critical role that ferroelectric film thickness plays in loss and phase-shift is closely examined. Experimental data for both thin film BaxSr1-xTiO 3 phase shifters near room temperature and SMO3 phase shifters at cryogenic temperatures on MgO and LaAlO3 substrates is included. Some of these devices demonstrated an insertion loss of less than 5 dB at Ku-band with continuously variable phase shift in excess of 360 degrees. The performance of these devices is superior to the state-of-the-art semiconductor counterparts. Frequency and phase agile antenna prototypes including a microstrip patch that can operate at multiple microwave frequency bands and a new type of phased array antenna concept called the ferroelectric reflectarray are introduced. Modeled data for tunable microstrip patch antennas is presented for various ferroelectric film thickness. A prototype linear phased array, with a conventional beam-forming manifold, and an electronic controller is described. This is the first

  2. Cupola Furnace Computer Process Model

    SciTech Connect

    Seymour Katz

    2004-12-31

    The cupola furnace generates more than 50% of the liquid iron used to produce the 9+ million tons of castings annually. The cupola converts iron and steel into cast iron. The main advantages of the cupola furnace are lower energy costs than those of competing furnaces (electric) and the ability to melt less expensive metallic scrap than the competing furnaces. However the chemical and physical processes that take place in the cupola furnace are highly complex making it difficult to operate the furnace in optimal fashion. The results are low energy efficiency and poor recovery of important and expensive alloy elements due to oxidation. Between 1990 and 2004 under the auspices of the Department of Energy, the American Foundry Society and General Motors Corp. a computer simulation of the cupola furnace was developed that accurately describes the complex behavior of the furnace. When provided with the furnace input conditions the model provides accurate values of the output conditions in a matter of seconds. It also provides key diagnostics. Using clues from the diagnostics a trained specialist can infer changes in the operation that will move the system toward higher efficiency. Repeating the process in an iterative fashion leads to near optimum operating conditions with just a few iterations. More advanced uses of the program have been examined. The program is currently being combined with an ''Expert System'' to permit optimization in real time. The program has been combined with ''neural network'' programs to affect very easy scanning of a wide range of furnace operation. Rudimentary efforts were successfully made to operate the furnace using a computer. References to these more advanced systems will be found in the ''Cupola Handbook''. Chapter 27, American Foundry Society, Des Plaines, IL (1999).

  3. Process-Response Modeling and the Scientific Process.

    ERIC Educational Resources Information Center

    Fichter, Lynn S.

    1988-01-01

    Discusses the process-response model (PRM) in its theoretical and practical forms. Describes how geologists attempt to reconstruct the process from the response (the geologic phenomenon) being studied. (TW)

  4. A Roadmap for Using Agile Development in a Traditional Environment

    NASA Technical Reports Server (NTRS)

    Streiffert, Barbara A.; Starbird, Thomas; Grenander, Sven

    2006-01-01

    One of the newer classes of software engineering techniques is called 'Agile Development'. In Agile Development software engineers take small implementation steps and, in some cases they program in pairs. In addition, they develop automatic tests prior to implementing their small functional piece. Agile Development focuses on rapid turnaround, incremental planning, customer involvement and continuous integration. Agile Development is not the traditional waterfall method or even a rapid prototyping method (although this methodology is closer to Agile Development). At Jet Propulsion Laboratory (JPL) a few groups have begun Agile Development software implementations. The difficulty with this approach becomes apparent when Agile Development is used in an organization that has specific criteria and requirements handed down for how software development is to be performed. The work at the JPL is performed for the National Aeronautics and Space Agency (NASA). Both organizations have specific requirements, rules and procedure for developing software. This paper will discuss the some of the initial uses of the Agile Development methodology, the spread of this method and the current status of the successful incorporation into the current JPL development policies.

  5. Agile manufacturing in Intelligence, Surveillance and Reconnaissance (ISR)

    NASA Astrophysics Data System (ADS)

    DiPadua, Mark; Dalton, George

    2016-05-01

    The objective of the Agile Manufacturing for Intelligence, Surveillance, and Reconnaissance (AMISR) effort is to research, develop, design and build a prototype multi-intelligence (multi-INT), reconfigurable pod demonstrating benefits of agile manufacturing and a modular open systems approach (MOSA) to make podded intelligence, surveillance, and reconnaissance (ISR) capability more affordable and operationally flexible.

  6. Agile Bodies: A New Imperative in Neoliberal Governance

    ERIC Educational Resources Information Center

    Gillies, Donald

    2011-01-01

    Modern business discourse suggests that a key bulwark against market fluctuation and the threat of failure is for organizations to become "agile'", a more dynamic and proactive position than that previously afforded by mere "flexibility". The same idea is also directed at the personal level, it being argued that the "agile" individual is better…

  7. Process modeling and industrial energy use

    SciTech Connect

    Howe, S O; Pilati, D A; Sparrow, F T

    1980-11-01

    How the process models developed at BNL are used to analyze industrial energy use is described and illustrated. Following a brief overview of the industry modeling program, the general methodology of process modeling is discussed. The discussion highlights the important concepts, contents, inputs, and outputs of a typical process model. A model of the US pulp and paper industry is then discussed as a specific application of process modeling methodology. Case study results from the pulp and paper model illustrate how process models can be used to analyze a variety of issues. Applications addressed with the case study results include projections of energy demand, conservation technology assessment, energy-related tax policies, and sensitivity analysis. A subsequent discussion of these results supports the conclusion that industry process models are versatile and powerful tools for energy end-use modeling and conservation analysis. Information on the current status of industry models at BNL is tabulated.

  8. Software Product Line Engineering Approach for Enhancing Agile Methodologies

    NASA Astrophysics Data System (ADS)

    Martinez, Jabier; Diaz, Jessica; Perez, Jennifer; Garbajosa, Juan

    One of the main principles of Agile methodologies consists in the early and continuous delivery of valuable software by short time-framed iterations. After each iteration, a working product is delivered according to the requirements defined at the beginning of the iteration. Testing tools facilitate the task of checking if the system provides the expected behavior according to the specified requirements. However, since testing tools need to be adapted in order to test new working products in each iteration, a significant effort has to be invested. This work presents a Software Product Line Engineering (SPLE) approach that allows flexibility in the adaption of testing tools with the working products in an iterative way. A case study is also presented using PLUM (Product Line Unified Modeller) as the tool suite for SPL implementation and management.

  9. Agile Machining and Inspection Non-Nuclear Report (NNR) Project

    SciTech Connect

    Lazarus, Lloyd

    2009-02-19

    This report is a high level summary of the eight major projects funded by the Agile Machining and Inspection Non-Nuclear Readiness (NNR) project (FY06.0422.3.04.R1). The largest project of the group is the Rapid Response project in which the six major sub categories are summarized. This project focused on the operations of the machining departments that will comprise Special Applications Machining (SAM) in the Kansas City Responsive Infrastructure Manufacturing & Sourcing (KCRIMS) project. This project was aimed at upgrading older machine tools, developing new inspection tools, eliminating Classified Removable Electronic Media (CREM) in the handling of classified Numerical Control (NC) programs by installing the CRONOS network, and developing methods to automatically load Coordinated-Measuring Machine (CMM) inspection data into bomb books and product score cards. Finally, the project personnel leaned perations of some of the machine tool cells, and now have the model to continue this activity.

  10. Analog modelling of obduction processes

    NASA Astrophysics Data System (ADS)

    Agard, P.; Zuo, X.; Funiciello, F.; Bellahsen, N.; Faccenna, C.; Savva, D.

    2012-04-01

    Obduction corresponds to one of plate tectonics oddities, whereby dense, oceanic rocks (ophiolites) are presumably 'thrust' on top of light, continental ones, as for the short-lived, almost synchronous Peri-Arabic obduction (which took place along thousands of km from Turkey to Oman in c. 5-10 Ma). Analog modelling experiments were performed to study the mechanisms of obduction initiation and test various triggering hypotheses (i.e., plate acceleration, slab hitting the 660 km discontinuity, ridge subduction; Agard et al., 2007). The experimental setup comprises (1) an upper mantle, modelled as a low-viscosity transparent Newtonian glucose syrup filling a rigid Plexiglas tank and (2) high-viscosity silicone plates (Rhodrosil Gomme with PDMS iron fillers to reproduce densities of continental or oceanic plates), located at the centre of the tank above the syrup to simulate the subducting and the overriding plates - and avoid friction on the sides of the tank. Convergence is simulated by pushing on a piston at one end of the model with velocities comparable to those of plate tectonics (i.e., in the range 1-10 cm/yr). The reference set-up includes, from one end to the other (~60 cm): (i) the piston, (ii) a continental margin containing a transition zone to the adjacent oceanic plate, (iii) a weakness zone with variable resistance and dip (W), (iv) an oceanic plate - with or without a spreading ridge, (v) a subduction zone (S) dipping away from the piston and (vi) an upper, active continental margin, below which the oceanic plate is being subducted at the start of the experiment (as is known to have been the case in Oman). Several configurations were tested and over thirty different parametric tests were performed. Special emphasis was placed on comparing different types of weakness zone (W) and the extent of mechanical coupling across them, particularly when plates were accelerated. Displacements, together with along-strike and across-strike internal deformation in all

  11. Combining Agile and Traditional: Customer Communication in Distributed Environment

    NASA Astrophysics Data System (ADS)

    Korkala, Mikko; Pikkarainen, Minna; Conboy, Kieran

    Distributed development is a radically increasing phenomenon in modern software development environments. At the same time, traditional and agile methodologies and combinations of those are being used in the industry. Agile approaches place a large emphasis on customer communication. However, existing knowledge on customer communication in distributed agile development seems to be lacking. In order to shed light on this topic and provide practical guidelines for companies in distributed agile environments, a qualitative case study was conducted in a large globally distributed software company. The key finding was that it might be difficult for an agile organization to get relevant information from a traditional type of customer organization, even though the customer communication was indicated to be active and utilized via multiple different communication media. Several challenges discussed in this paper referred to "information blackout" indicating the importance of an environment fostering meaningful communication. In order to evaluate if this environment can be created a set of guidelines is proposed.

  12. A Case Study of Coordination in Distributed Agile Software Development

    NASA Astrophysics Data System (ADS)

    Hole, Steinar; Moe, Nils Brede

    Global Software Development (GSD) has gained significant popularity as an emerging paradigm. Companies also show interest in applying agile approaches in distributed development to combine the advantages of both approaches. However, in their most radical forms, agile and GSD can be placed in each end of a plan-based/agile spectrum because of how work is coordinated. We describe how three GSD projects applying agile methods coordinate their work. We found that trust is needed to reduce the need of standardization and direct supervision when coordinating work in a GSD project, and that electronic chatting supports mutual adjustment. Further, co-location and modularization mitigates communication problems, enables agility in at least part of a GSD project, and renders the implementation of Scrum of Scrums possible.

  13. Analysis of VLF signals associated to AGILE Terrestrial Gamma-ray Flashes detected over Central America

    NASA Astrophysics Data System (ADS)

    Marisaldi, Martino; Lyu, Fanchao; Cummer, Steven; Ursi, Alessandro

    2016-04-01

    Analysis of radio signals detected on ground and associated to Terrestrial Gamma-ray Flashes (TGFs) have proven to be a successful tool to extract information on the TGF itself and the possible associated lightning process. Triangulation of Very Low Frequency (VLF) signals by means of the Time Of Arrival technique provides TGF location with few km accuracy. The AGILE satellite is routinely observing TGFs on a narrow band across the Equator, limited by the small satellite orbital inclination (2.5°). However, until recently it was not possible to provide firm associations between AGILE TGFs and radio signals, because of two main limiting factors. First, dead-time effects led to a bias towards long duration events in AGILE TGF sample, which are less likely associated to strong radio pulses. In addition, most VLF detection networks are less sensitive along the equatorial region. Since the end of March 2015 a major change in the AGILE MiniCalorimeter instrument configuration resulted in a ten fold increase in TGF detection rate, and in the detection of events as short as 20 microseconds. 14% of the events in the new sample resulted simultaneous (within 200 microseconds) to sferics detected by the World Wide Lightning Location Network (WWLLN), therefore a source localisation is available for these events. We present here the first analysis of VLF waveforms associated to AGILE TGFs observed above Central America, detected by magnetic field sensors deployed in Puerto Rico. Among the seven TGFs with a WWLLN location at a distance lower than 10000 km from the sensors, four of them have detectable signals. These events are the closest to the sensors, with distance less than 7500 km. We present here the properties of these TGFs and the characteristics of the associated radio waveforms.

  14. Process correlation analysis model for process improvement identification.

    PubMed

    Choi, Su-jin; Kim, Dae-Kyoo; Park, Sooyong

    2014-01-01

    Software process improvement aims at improving the development process of software systems. It is initiated by process assessment identifying strengths and weaknesses and based on the findings, improvement plans are developed. In general, a process reference model (e.g., CMMI) is used throughout the process of software process improvement as the base. CMMI defines a set of process areas involved in software development and what to be carried out in process areas in terms of goals and practices. Process areas and their elements (goals and practices) are often correlated due to the iterative nature of software development process. However, in the current practice, correlations of process elements are often overlooked in the development of an improvement plan, which diminishes the efficiency of the plan. This is mainly attributed to significant efforts and the lack of required expertise. In this paper, we present a process correlation analysis model that helps identify correlations of process elements from the results of process assessment. This model is defined based on CMMI and empirical data of improvement practices. We evaluate the model using industrial data.

  15. Process Correlation Analysis Model for Process Improvement Identification

    PubMed Central

    Park, Sooyong

    2014-01-01

    Software process improvement aims at improving the development process of software systems. It is initiated by process assessment identifying strengths and weaknesses and based on the findings, improvement plans are developed. In general, a process reference model (e.g., CMMI) is used throughout the process of software process improvement as the base. CMMI defines a set of process areas involved in software development and what to be carried out in process areas in terms of goals and practices. Process areas and their elements (goals and practices) are often correlated due to the iterative nature of software development process. However, in the current practice, correlations of process elements are often overlooked in the development of an improvement plan, which diminishes the efficiency of the plan. This is mainly attributed to significant efforts and the lack of required expertise. In this paper, we present a process correlation analysis model that helps identify correlations of process elements from the results of process assessment. This model is defined based on CMMI and empirical data of improvement practices. We evaluate the model using industrial data. PMID:24977170

  16. Model-based software process improvement

    NASA Technical Reports Server (NTRS)

    Zettervall, Brenda T.

    1994-01-01

    The activities of a field test site for the Software Engineering Institute's software process definition project are discussed. Products tested included the improvement model itself, descriptive modeling techniques, the CMM level 2 framework document, and the use of process definition guidelines and templates. The software process improvement model represents a five stage cyclic approach for organizational process improvement. The cycles consist of the initiating, diagnosing, establishing, acting, and leveraging phases.

  17. Sequential detection and robust estimation of vapor concentration using frequency-agile lidar time series data

    NASA Astrophysics Data System (ADS)

    Warren, Russell E.; Vanderbeek, Richard G.; D'Amico, Francis M.; Ben-David, Avishai

    1999-01-01

    This paper extends an earlier optimal approach for frequency-agile lidar using fixed-size samples of data to include the time series aspect of data collection. The likelihood ratio test methodology for deterministic but unknown vapor concentration is replaced by a Bayesian formalism in which the path integral of vapor concentration CL evolves in time through a random walk model. The fixed- sample maximum likelihood estimates of CL derived earlier are replaced by Kalman filter estimates, and the log- likelihood ratio is generalized to a sequential test statistic written in terms of the Kalman estimates. In addition to the time series aspect, the earlier approach is generalized by (1) including the transmitted energy on a short-by-shot basis in a statistically optimum manner, (2) adding a linear slope component to the transmitter and received data models, and (3) replacing the nominal multivariate normal statistical assumption by a robust model in the Huber sensor for mitigating the effects of occasional data spikes caused by laser misfiring or EMI. The estimation and detection algorithms are compared with fixed-sample processing by the DIAL method on FAL data collected by ERDEC during vapor chamber testing at Dugway, Utah.

  18. Transforming Collaborative Process Models into Interface Process Models by Applying an MDA Approach

    NASA Astrophysics Data System (ADS)

    Lazarte, Ivanna M.; Chiotti, Omar; Villarreal, Pablo D.

    Collaborative business models among enterprises require defining collaborative business processes. Enterprises implement B2B collaborations to execute these processes. In B2B collaborations the integration and interoperability of processes and systems of the enterprises are required to support the execution of collaborative processes. From a collaborative process model, which describes the global view of the enterprise interactions, each enterprise must define the interface process that represents the role it performs in the collaborative process in order to implement the process in a Business Process Management System. Hence, in this work we propose a method for the automatic generation of the interface process model of each enterprise from a collaborative process model. This method is based on a Model-Driven Architecture to transform collaborative process models into interface process models. By applying this method, interface processes are guaranteed to be interoperable and defined according to a collaborative process.

  19. Birth/death process model

    NASA Technical Reports Server (NTRS)

    Solloway, C. B.; Wakeland, W.

    1976-01-01

    First-order Markov model developed on digital computer for population with specific characteristics. System is user interactive, self-documenting, and does not require user to have complete understanding of underlying model details. Contains thorough error-checking algorithms on input and default capabilities.

  20. Mathematical Modeling: A Structured Process

    ERIC Educational Resources Information Center

    Anhalt, Cynthia Oropesa; Cortez, Ricardo

    2015-01-01

    Mathematical modeling, in which students use mathematics to explain or interpret physical, social, or scientific phenomena, is an essential component of the high school curriculum. The Common Core State Standards for Mathematics (CCSSM) classify modeling as a K-12 standard for mathematical practice and as a conceptual category for high school…

  1. SuperAGILE: The Hard X-ray Imager of AGILE

    SciTech Connect

    Feroci, M.; Costa, E.; Barbanera, L.; Del Monte, E.; Di Persio, G.; Frutti, M.; Lapshov, I.; Lazzarotto, F.; Pacciani, L.; Porrovecchio, G.; Preger, B.; Rapisarda, M.; Rubini, A.; Soffitta, P.; Tavani, M.; Mastropietro, M.; Morelli, E.; Argan, A.; Ghirlanda, G.; Mereghetti, S.

    2004-09-28

    SuperAGILE is the hard X-ray (10-40 keV) imager for the gamma-ray mission AGILE, currently scheduled for launch in mid-2005. It is based on 4 Si-microstrip detectors, with a total geometric area of 1444 cm{sup 2} (max effective about 300 cm{sup 2}), equipped with one-dimensional coded masks. The 4 detectors are perpendicularly oriented, in order to provide pairs of orthogonal one-dimensional images of the X-ray sky. The field of view of each 1-D detector is 107 deg. x 68 deg., at zero response, with an overlap in the central 68 deg. x 68 deg. area. The angular resolution on axis is 6 arcmin (pixel size). We present here the current status of the hardware development and the scientific potential for GRBs, for which an onboard trigger and imaging system will allow distributing locations through a fast communication telemetry link from AGILE to the ground.

  2. A Comparative of business process modelling techniques

    NASA Astrophysics Data System (ADS)

    Tangkawarow, I. R. H. T.; Waworuntu, J.

    2016-04-01

    In this era, there is a lot of business process modeling techniques. This article is the research about differences of business process modeling techniques. For each technique will explain about the definition and the structure. This paper presents a comparative analysis of some popular business process modelling techniques. The comparative framework is based on 2 criteria: notation and how it works when implemented in Somerleyton Animal Park. Each technique will end with the advantages and disadvantages. The final conclusion will give recommend of business process modeling techniques that easy to use and serve the basis for evaluating further modelling techniques.

  3. Agile: From Software to Mission System

    NASA Technical Reports Server (NTRS)

    Trimble, Jay; Shirley, Mark H.; Hobart, Sarah Groves

    2016-01-01

    The Resource Prospector (RP) is an in-situ resource utilization (ISRU) technology demonstration mission, designed to search for volatiles at the Lunar South Pole. This is NASA's first near real time tele-operated rover on the Moon. The primary objective is to search for volatiles at one of the Lunar Poles. The combination of short mission duration, a solar powered rover, and the requirement to explore shadowed regions makes for an operationally challenging mission. To maximize efficiency and flexibility in Mission System design and thus to improve the performance and reliability of the resulting Mission System, we are tailoring Agile principles that we have used effectively in ground data system software development and applying those principles to the design of elements of the mission operations system.

  4. Wavelength agile holmium-doped fiber laser

    NASA Astrophysics Data System (ADS)

    Simakov, N.; Daniel, J. M. O.; Ward, J.; Clarkson, W. A.; Hemming, A.; Haub, J.

    2016-03-01

    For the first time, an electronically-controlled, wavelength-agile tuneable holmium-doped fibre laser is presented. A narrow-band acousto-optic tuneable filter was characterized and used as the wavelength selective element to avoid any inertial effects associated with opto-mechanical tuning mechanisms. We demonstrate operation over a 90 nm wavelength range spanning 2040 - 2130 nm. The laser produced >150 mW over this entire range with a signal-to-noise ratio of >45 dB and line-width of ~0.16 nm. Switching times of ~35 μs and sweep rates of up to 9 nm/ms were also demonstrated.

  5. Compact, flexible, frequency agile parametric wavelength converter

    DOEpatents

    Velsko, Stephan P.; Yang, Steven T.

    2002-01-01

    This improved Frequency Agile Optical Parametric Oscillator provides near on-axis pumping of a single QPMC with a tilted periodically poled grating to overcome the necessity to find a particular crystal that will permit collinear birefringence in order to obtain a desired tuning range. A tilted grating design and the elongation of the transverse profile of the pump beam in the angle tuning plane of the FA-OPO reduces the rate of change of the overlap between the pumped volume in the crystal and the resonated and non-resonated wave mode volumes as the pump beam angle is changed. A folded mirror set relays the pivot point for beam steering from a beam deflector to the center of the FA-OPO crystal. This reduces the footprint of the device by as much as a factor of two over that obtained when using the refractive telescope design.

  6. Negative Binomial Process Count and Mixture Modeling.

    PubMed

    Zhou, Mingyuan; Carin, Lawrence

    2013-10-17

    The seemingly disjoint problems of count and mixture modeling are united under the negative binomial (NB) process. A gamma process is employed to model the rate measure of a Poisson process, whose normalization provides a random probability measure for mixture modeling and whose marginalization leads to a NB process for count modeling. A draw from the NB process consists of a Poisson distributed finite number of distinct atoms, each of which is associated with a logarithmic distributed number of data samples. We reveal relationships between various count- and mixture-modeling distributions distributions, and construct a Poisson-logarithmic bivariate distribution that connects the NB and Chinese restaurant table distributions. Fundamental properties of the models are developed, and we derive efficient Bayesian inference. It is shown that with augmentation and normalization, the NB process and gamma-NB process can be reduced to the Dirichlet process and hierarchical Dirichlet process, respectively. These relationships highlight theoretical, structural and computational advantages of the NB process. A variety of NB processes, including the beta-geometric, beta-NB, marked-beta-NB, marked-gamma-NB and zero-inflated-NB processes, with distinct sharing mechanisms, are also constructed. These models are applied to topic modeling, with connections made to existing algorithms under Poisson factor analysis. Example results show the importance of inferring both the NB dispersion and probability parameters. PMID:24144977

  7. Negative Binomial Process Count and Mixture Modeling.

    PubMed

    Zhou, Mingyuan; Carin, Lawrence

    2015-02-01

    The seemingly disjoint problems of count and mixture modeling are united under the negative binomial (NB) process. A gamma process is employed to model the rate measure of a Poisson process, whose normalization provides a random probability measure for mixture modeling and whose marginalization leads to an NB process for count modeling. A draw from the NB process consists of a Poisson distributed finite number of distinct atoms, each of which is associated with a logarithmic distributed number of data samples. We reveal relationships between various count- and mixture-modeling distributions and construct a Poisson-logarithmic bivariate distribution that connects the NB and Chinese restaurant table distributions. Fundamental properties of the models are developed, and we derive efficient Bayesian inference. It is shown that with augmentation and normalization, the NB process and gamma-NB process can be reduced to the Dirichlet process and hierarchical Dirichlet process, respectively. These relationships highlight theoretical, structural, and computational advantages of the NB process. A variety of NB processes, including the beta-geometric, beta-NB, marked-beta-NB, marked-gamma-NB and zero-inflated-NB processes, with distinct sharing mechanisms, are also constructed. These models are applied to topic modeling, with connections made to existing algorithms under Poisson factor analysis. Example results show the importance of inferring both the NB dispersion and probability parameters. PMID:26353243

  8. Information-Processing Models and Curriculum Design

    ERIC Educational Resources Information Center

    Calfee, Robert C.

    1970-01-01

    "This paper consists of three sections--(a) the relation of theoretical analyses of learning to curriculum design, (b) the role of information-processing models in analyses of learning processes, and (c) selected examples of the application of information-processing models to curriculum design problems." (Author)

  9. Ramping up for agility: Development of a concurrent engineering communications infrastructure

    SciTech Connect

    Forsythe, C.; Ashby, M.R.

    1995-09-01

    A-PRIMED (Agile Product Realization for Innovative Electro MEchanical Devices) demonstrated new product development in24 days accompanied by improved product quality, through ability enabling technologies. A concurrent engineering communications infrastructure was developed that provided electronic data communications, information access, enterprise integration of computers and applications, and collaborative work tools. This paper describes how A-PRIMED did it through attention to technologies, processes, and people.

  10. A Hierarchical Process-Dissociation Model

    ERIC Educational Resources Information Center

    Rouder, Jeffrey N.; Lu, Jun; Morey, Richard D.; Sun, Dongchu; Speckman, Paul L.

    2008-01-01

    In fitting the process-dissociation model (L. L. Jacoby, 1991) to observed data, researchers aggregate outcomes across participant, items, or both. T. Curran and D. L. Hintzman (1995) demonstrated how biases from aggregation may lead to artifactual support for the model. The authors develop a hierarchical process-dissociation model that does not…

  11. Agile rediscovering values: Similarities to continuous improvement strategies

    NASA Astrophysics Data System (ADS)

    Díaz de Mera, P.; Arenas, J. M.; González, C.

    2012-04-01

    Research in the late 80's on technological companies that develop products of high value innovation, with sufficient speed and flexibility to adapt quickly to changing market conditions, gave rise to the new set of methodologies known as Agile Management Approach. In the current changing economic scenario, we considered very interesting to study the similarities of these Agile Methodologies with other practices whose effectiveness has been amply demonstrated in both the West and Japan. Strategies such as Kaizen, Lean, World Class Manufacturing, Concurrent Engineering, etc, would be analyzed to check the values they have in common with the Agile Approach.

  12. Computational Process Modeling for Additive Manufacturing

    NASA Technical Reports Server (NTRS)

    Bagg, Stacey; Zhang, Wei

    2014-01-01

    Computational Process and Material Modeling of Powder Bed additive manufacturing of IN 718. Optimize material build parameters with reduced time and cost through modeling. Increase understanding of build properties. Increase reliability of builds. Decrease time to adoption of process for critical hardware. Potential to decrease post-build heat treatments. Conduct single-track and coupon builds at various build parameters. Record build parameter information and QM Meltpool data. Refine Applied Optimization powder bed AM process model using data. Report thermal modeling results. Conduct metallography of build samples. Calibrate STK models using metallography findings. Run STK models using AO thermal profiles and report STK modeling results. Validate modeling with additional build. Photodiode Intensity measurements highly linear with power input. Melt Pool Intensity highly correlated to Melt Pool Size. Melt Pool size and intensity increase with power. Applied Optimization will use data to develop powder bed additive manufacturing process model.

  13. Simulation Modeling of Software Development Processes

    NASA Technical Reports Server (NTRS)

    Calavaro, G. F.; Basili, V. R.; Iazeolla, G.

    1996-01-01

    A simulation modeling approach is proposed for the prediction of software process productivity indices, such as cost and time-to-market, and the sensitivity analysis of such indices to changes in the organization parameters and user requirements. The approach uses a timed Petri Net and Object Oriented top-down model specification. Results demonstrate the model representativeness, and its usefulness in verifying process conformance to expectations, and in performing continuous process improvement and optimization.

  14. GRB 070724B: the first Gamma Ray Burst localized by SuperAGILE

    SciTech Connect

    Del Monte, E.; Costa, E.; Donnarumma, I.; Feroci, M.; Lapshov, I.; Lazzarotto, F.; Soffitta, P.; Argan, A.; Pucella, G.; Trois, A.; Vittorini, V.; Evangelista, Y.; Rapisarda, M.; Barbiellini, G.; Longo, F.; Basset, M.; Foggetta, L.; Vallazza, E.; Bulgarelli, A.; Di Cocco, G.

    2008-05-22

    GRB070724B is the first Gamma Ray Burst localized by the SuperAGILE instrument aboard the AGILE space mission. The SuperAGILE localization has been confirmed after the after-glow observation by the XRT aboard the Swift satellite. No significant gamma ray emission above 50 MeV has been detected for this GRB. In this paper we describe the SuperAGILE capabilities in detecting Gamma Ray Burst and the AGILE observation of GRB 070724B.

  15. The influence of ankle dorsiflexion on jumping capacity and the modified agility t-test performance.

    PubMed

    Salinero, Juan J; Abian-Vicen, Javier; Del Coso, Juan; González-Millán, Cristina

    2014-01-01

    Dorsiflexion sport shoes aim to increase jumping capacity and speed by means of a lower position of the heel in comparison with the forefoot, favouring additional stretching of the ankle plantar flexors. In previous studies, contradictory results have been found on the benefits of using this type of shoe. With the aim of comparing a dorsiflexion sport shoe model (DF) with a conventional sport shoe (CS), 41 participants performed a countermovement jump (CMJ) test and an agility test (MAT) with both models of shoe. There were no significant differences in the jump test [CS=35.3 cm (6.4) and DF=35.6 cm (6.4), P>0.05]. In the agility test, the conventional shoe obtained better results than the model with dorsiflexion with regard to time taken to complete the circuit [CS=6236 ms (540) and DF=6377 ms (507), P<0.05)]. In spite of producing pre-stretching of the plantar muscles, the DF sport shoes were not effective for improving either jump power or agility in a specific test.

  16. Modeling the Reading Process: Promise and Problems.

    ERIC Educational Resources Information Center

    Geyer, John J.

    The problems of modeling a process as complex as reading are discussed, including such factors as the lack of agreement surrounding definitions of modeling, varying levels of rigor within and between models, the disjunctive categories within which models fall, and the difficulty of synthesis across fields which employ very different technical…

  17. Agile text mining for the 2014 i2b2/UTHealth Cardiac risk factors challenge.

    PubMed

    Cormack, James; Nath, Chinmoy; Milward, David; Raja, Kalpana; Jonnalagadda, Siddhartha R

    2015-12-01

    This paper describes the use of an agile text mining platform (Linguamatics' Interactive Information Extraction Platform, I2E) to extract document-level cardiac risk factors in patient records as defined in the i2b2/UTHealth 2014 challenge. The approach uses a data-driven rule-based methodology with the addition of a simple supervised classifier. We demonstrate that agile text mining allows for rapid optimization of extraction strategies, while post-processing can leverage annotation guidelines, corpus statistics and logic inferred from the gold standard data. We also show how data imbalance in a training set affects performance. Evaluation of this approach on the test data gave an F-Score of 91.7%, one percent behind the top performing system.

  18. Frequency agile OPO-based transmitters for multiwavelength DIAL

    SciTech Connect

    Velsko, S.P.; Ruggiero, A.; Herman, M.

    1996-09-01

    We describe a first generation mid-infrared transmitter with pulse to pulse frequency agility and both wide and narrow band capability. This transmitter was used to make multicomponent Differential Absorption LIDAR (DIAL) measurements in the field.

  19. Value Creation by Agile Projects: Methodology or Mystery?

    NASA Astrophysics Data System (ADS)

    Racheva, Zornitza; Daneva, Maya; Sikkel, Klaas

    Business value is a key concept in agile software development approaches. This paper presents results of a systematic review of literature on how business value is created by agile projects. We found that with very few exceptions, most published studies take the concept of business value for granted and do not state what it means in general as well as in the specific study context. We could find no study which clearly indicates how exactly individual agile practices or groups of those create value and keep accumulating it over time. The key implication for research is that we have an incentive to pursue the study of value creation in agile project by deploying empirical research methods.

  20. Frequency agile OPO-based transmitters for multiwavelength DIAL

    SciTech Connect

    Velsko, S.P.; Ruggiero, A.; Herman, M.

    1996-09-01

    We describe a first generation mid-infrared transmitter with pulse-to- pulse frequency agility and both wide and narrow band capability. This transmitter was used to make multicomponent DIAL measurements in the field.

  1. Investigation into the impact of agility on conceptual fighter design

    NASA Technical Reports Server (NTRS)

    Engelbeck, R. M.

    1995-01-01

    The Agility Design Study was performed by the Boeing Defense and Space Group for the NASA Langley Research Center. The objective of the study was to assess the impact of agility requirements on new fighter configurations. Global trade issues investigated were the level of agility, the mission role of the aircraft (air-to-ground, multi-role, or air-to-air), and whether the customer is Air force, Navy, or joint service. Mission profiles and design objectives were supplied by NASA. An extensive technology assessment was conducted to establish the available technologies to industry for the aircraft. Conceptual level methodology is presented to assess the five NASA-supplied agility metrics. Twelve configurations were developed to address the global trade issues. Three-view drawings, inboard profiles, and performance estimates were made and are included in the report. A critical assessment and lessons learned from the study are also presented.

  2. Laterality and performance of agility-trained dogs.

    PubMed

    Siniscalchi, Marcello; Bertino, Daniele; Quaranta, Angelo

    2014-01-01

    Correlations between lateralised behaviour and performance were investigated in 19 agility-trained dogs (Canis familiaris) by scoring paw preference to hold a food object and relating it to performance during typical agility obstacles (jump/A-frame and weave poles). In addition, because recent behavioural studies reported that visual stimuli of emotional valence presented to one visual hemifield at a time affect visually guided motor responses in dogs, the possibility that the position of the owner respectively in the left and in the right canine visual hemifield might be associated with quality of performance during agility was considered. Dogs' temperament was also measured by an owner-rated questionnaire. The most relevant finding was that agility-trained dogs displayed longer latencies to complete the obstacles with the owner located in their left visual hemifield compared to the right. Interestingly, the results showed that this phenomenon was significantly linked to both dogs' trainability and the strength of paw preference.

  3. Modern Enterprise Systems as Enablers of Agile Development

    NASA Astrophysics Data System (ADS)

    Fredriksson, Odd; Ljung, Lennart

    Traditional ES technology and traditional project management methods are supporting and matching each other. But they are not supporting the critical success conditions for ES development in an effective way. Although the findings from one case study of a successful modern ES change project is not strong empirical evidence, we carefully propose that the new modern ES technology is supporting and matching agile project management methods. In other words, it provides the required flexibility which makes it possible to put into practice the agile way of running projects, both for the system supplier and for the customer. In addition, we propose that the combination of modern ES technology and agile project management methods are more appropriate for supporting the realization of critical success conditions for ES development. The main purpose of this chapter is to compare critical success conditions for modern enterprise systems development projects with critical success conditions for agile information systems development projects.

  4. Insights into Global Health Practice from the Agile Software Development Movement

    PubMed Central

    Flood, David; Chary, Anita; Austad, Kirsten; Diaz, Anne Kraemer; García, Pablo; Martinez, Boris; Canú, Waleska López; Rohloff, Peter

    2016-01-01

    Global health practitioners may feel frustration that current models of global health research, delivery, and implementation are overly focused on specific interventions, slow to provide health services in the field, and relatively ill-equipped to adapt to local contexts. Adapting design principles from the agile software development movement, we propose an analogous approach to designing global health programs that emphasizes tight integration between research and implementation, early involvement of ground-level health workers and program beneficiaries, and rapid cycles of iterative program improvement. Using examples from our own fieldwork, we illustrate the potential of ‘agile global health’ and reflect on the limitations, trade-offs, and implications of this approach. PMID:27134081

  5. Drought processes, modeling, and mitigation

    NASA Astrophysics Data System (ADS)

    Mishra, Ashok K.; Sivakumar, Bellie; Singh, Vijay P.

    2015-07-01

    Accurate assessment of droughts is crucial for proper planning and management of our water resources, environment, and ecosystems. The combined influence of increasing water demands and the anticipated impacts of global climate change has already raised serious concerns about worsening drought conditions in the future and their social, economic, and environmental impacts. As a result, studies on droughts are currently a major focal point for a broad range of research communities, including civil engineers, hydrologists, environmentalists, ecologists, meteorologists, geologists, agricultural scientists, economists, policy makers, and water managers. There is, therefore, an urgent need for enhancing our understanding of droughts (e.g. occurrence, modeling), making more reliable assessments of their impacts on various sectors of our society (e.g. domestic, agricultural, industrial), and undertaking appropriate adaptation and mitigation measures, especially in the face of global climate change.

  6. The impact of flying qualities on helicopter operational agility

    NASA Technical Reports Server (NTRS)

    Padfield, Gareth D.; Lappos, Nick; Hodgkinson, John

    1993-01-01

    Flying qualities standards are formally set to ensure safe flight and therefore reflect minimum, rather than optimum, requirements. Agility is a flying quality but relates to operations at high, if not maximum, performance. While the quality metrics and test procedures for flying, as covered for example in ADS33C, may provide an adequate structure to encompass agility, they do not currently address flight at high performance. This is also true in the fixed-wing world and a current concern in both communities is the absence of substantiated agility criteria and possible conflicts between flying qualities and high performance. AGARD is sponsoring a working group (WG19) title 'Operational Agility' that deals with these and a range of related issues. This paper is condensed from contributions by the three authors to WG19, relating to flying qualities. Novel perspectives on the subject are presented including the agility factor, that quantifies performance margins in flying qualities terms; a new parameter, based on maneuver acceleration is introduced as a potential candidate for defining upper limits to flying qualities. Finally, a probabilistic analysis of pilot handling qualities ratings is presented that suggests a powerful relationship between inherent airframe flying qualities and operational agility.

  7. Declarative business process modelling: principles and modelling languages

    NASA Astrophysics Data System (ADS)

    Goedertier, Stijn; Vanthienen, Jan; Caron, Filip

    2015-02-01

    The business process literature has proposed a multitude of business process modelling approaches or paradigms, each in response to a different business process type with a unique set of requirements. Two polar paradigms, i.e. the imperative and the declarative paradigm, appear to define the extreme positions on the paradigm spectrum. While imperative approaches focus on explicitly defining how an organisational goal should be reached, the declarative approaches focus on the directives, policies and regulations restricting the potential ways to achieve the organisational goal. In between, a variety of hybrid-paradigms can be distinguished, e.g. the advanced and adaptive case management. This article focuses on the less-exposed declarative approach on process modelling. An outline of the declarative process modelling and the modelling approaches is presented, followed by an overview of the observed declarative process modelling principles and an evaluation of the declarative process modelling approaches.

  8. A Review of Agile and Lean Manufacturing as Issues in Selected International and National Research and Development Programs and Roadmaps

    ERIC Educational Resources Information Center

    Castro, Helio; Putnik, Goran D.; Shah, Vaibhav

    2012-01-01

    Purpose: The aim of this paper is to analyze international and national research and development (R&D) programs and roadmaps for the manufacturing sector, presenting how agile and lean manufacturing models are addressed in these programs. Design/methodology/approach: In this review, several manufacturing research and development programs and…

  9. Lithography process window analysis with calibrated model

    NASA Astrophysics Data System (ADS)

    Zhou, Wenzhan; Yu, Jin; Lo, James; Liu, Johnson

    2004-05-01

    As critical-dimension shrink below 0.13 μm, the SPC (Statistical Process Control) based on CD (Critical Dimension) control in lithography process becomes more difficult. Increasing requirements of a shrinking process window have called on the need for more accurate decision of process window center. However in practical fabrication, we found that systematic error introduced by metrology and/or resist process can significantly impact the process window analysis result. Especially, when the simple polynomial functions are used to fit the lithographic data from focus exposure matrix (FEM), the model will fit these systematic errors rather than filter them out. This will definitely impact the process window analysis and determination of the best process condition. In this paper, we proposed to use a calibrated first principle model to do process window analysis. With this method, the systematic metrology error can be filtered out efficiently and give a more reasonable window analysis result.

  10. Fusion Process Model Implementation Case Studies

    NASA Astrophysics Data System (ADS)

    Kaur, Rupinder; Sengupta, Jyotsna

    2012-07-01

    In this paper we have discussed, three case studies. The first one is applied at Web Shrub Solutions, a software development organization, second is applied at web based job portal (stepintojob.com) for leading Indian firm and the third is web design and development for SCL limited, to observe the results of Fusion Process Model. Fusion Process Model follows component driven approach; it applies 3C Model to generalize the process of solving the problem in each phase, which provides firm control over the software development process.

  11. Distillation modeling for a uranium refining process

    SciTech Connect

    Westphal, B.R.

    1996-03-01

    As part of the spent fuel treatment program at Argonne National Laboratory, a vacuum distillation process is being employed for the recovery of uranium following an electrorefining process. Distillation of a salt electrolyte, containing a eutectic mixture of lithium and potassium chlorides, from uranium is achieved by a simple batch operation and is termed {open_quotes}cathode processing{close_quotes}. The incremental distillation of electrolyte salt will be modeled by an equilibrium expression and on a molecular basis since the operation is conducted under moderate vacuum conditions. As processing continues, the two models will be compared and analyzed for correlation with actual operating results. Possible factors that may contribute to aberrations from the models include impurities at the vapor-liquid boundary, distillate reflux, anomalous pressure gradients, and mass transport phenomena at the evaporating surface. Ultimately, the purpose of either process model is to enable the parametric optimization of the process.

  12. VARTM Process Modeling of Aerospace Composite Structures

    NASA Technical Reports Server (NTRS)

    Song, Xiao-Lan; Grimsley, Brian W.; Hubert, Pascal; Cano, Roberto J.; Loos, Alfred C.

    2003-01-01

    A three-dimensional model was developed to simulate the VARTM composite manufacturing process. The model considers the two important mechanisms that occur during the process: resin flow, and compaction and relaxation of the preform. The model was used to simulate infiltration of a carbon preform with an epoxy resin by the VARTM process. The model predicted flow patterns and preform thickness changes agreed qualitatively with the measured values. However, the predicted total infiltration times were much longer than measured most likely due to the inaccurate preform permeability values used in the simulation.

  13. INTEGRATED FISCHER TROPSCH MODULAR PROCESS MODEL

    SciTech Connect

    Donna Post Guillen; Richard Boardman; Anastasia M. Gribik; Rick A. Wood; Robert A. Carrington

    2007-12-01

    With declining petroleum reserves, increased world demand, and unstable politics in some of the world’s richest oil producing regions, the capability for the U.S. to produce synthetic liquid fuels from domestic resources is critical to national security and economic stability. Coal, biomass and other carbonaceous materials can be converted to liquid fuels using several conversion processes. The leading candidate for large-scale conversion of coal to liquid fuels is the Fischer Tropsch (FT) process. Process configuration, component selection, and performance are interrelated and dependent on feed characteristics. This paper outlines a flexible modular approach to model an integrated FT process that utilizes a library of key component models, supporting kinetic data and materials and transport properties allowing rapid development of custom integrated plant models. The modular construction will permit rapid assessment of alternative designs and feed stocks. The modeling approach consists of three thrust areas, or “strands” – model/module development, integration of the model elements into an end to end integrated system model, and utilization of the model for plant design. Strand 1, model/module development, entails identifying, developing, and assembling a library of codes, user blocks, and data for FT process unit operations for a custom feedstock and plant description. Strand 2, integration development, provides the framework for linking these component and subsystem models to form an integrated FT plant simulation. Strand 3, plant design, includes testing and validation of the comprehensive model and performing design evaluation analyses.

  14. Communicative processes: a model of communication

    SciTech Connect

    Kimura, T.D.; Gillett, W.D.

    1982-01-01

    The authors introduce a conceptual model of communicative organization as a part of the formal semantic study of distributed computation. The model includes, as communication primitives, three independent modes of communication: mailing, posting and broadcasting. Mailing models thin-wire communication, and posting models shared memory communication. While broadcasting is not prominent in today's parallel programming languages, it has an important role to play in distributed computation. Other fundamental notions in the model are process, symbol, site, process class, symbol class and site class. 8 references.

  15. Models of Problem Solving Processes and Abilities.

    ERIC Educational Resources Information Center

    Feldhusen, John F.; Guthrie, Virginia A.

    1979-01-01

    This paper reviews current models of problem solving to identify results relevant to teachers or instructional developers. Four areas are covered: information processing models, approaches stressing human abilities and factors, creative problem solving models, and other aspects of problem solving. Part of a theme issue on intelligence. (Author/SJL)

  16. "Agile" Battery Technology Transfer-Lessons Learnt

    NASA Astrophysics Data System (ADS)

    Sabatini, P.; Annoni, G.; Grossi, R.; Alia, Sergio; Reulier, David

    2008-09-01

    AGILE, the high energy astrophysics mission of the Italian Space Agency launched on April 23rd 2007, is the first LEO satellite to be powered by Saft's commercially available space qualified MPS176065 rechargeable lithium ion batteries.Saft and Carlo Gavazzi Space (CGS) have achieved a successful technology transfer replacing Ni-H2 batteries with high energy lithium ion batteries in a full speed program (4 months) and with a cost effective approach. The battery system comprises 2 x 24 Saft MPS176065 space qualified Li-ion cells in an 8s3p configuration (3 parallel arrays each composed by 8 series cell) with a nominal capacity of 2 x 480 Wh and an integral autonomous cell balancing system that ensures the maximum possible battery life.The MPS176065 space qualified cell is based on Saft's well proven MP series of prismatic rechargeable Li-ion batteries. It offers an extremely high capacity made possible by the stainless steel prismatic container that makes use of the volume which is otherwise lost when conventional cylindrical cells are packed together. A single prismatic cell has about 20% more volumetric energy density than an equivalent pack of cylindrical cells.

  17. Embedding Agile Practices within a Plan-Driven Hierarchical Project Life Cycle

    SciTech Connect

    Millard, W. David; Johnson, Daniel M.; Henderson, John M.; Lombardo, Nicholas J.; Bass, Robert B.; Smith, Jason E.

    2014-07-28

    Organizations use structured, plan-driven approaches to provide continuity, direction, and control to large, multi-year programs. Projects within these programs vary greatly in size, complexity, level of maturity, technical risk, and clarity of the development objectives. Organizations that perform exploratory research, evolutionary development, and other R&D activities can obtain the benefits of Agile practices without losing the benefits of their program’s overarching plan-driven structure. This paper describes application of Agile development methods on a large plan-driven sensor integration program. While the client employed plan-driven, requirements flow-down methodologies, tight project schedules and complex interfaces called for frequent end-to-end demonstrations to provide feedback during system development. The development process maintained the many benefits of plan-driven project execution with the rapid prototyping, integration, demonstration, and client feedback possible through Agile development methods. This paper also describes some of the tools and implementing mechanisms used to transition between and take advantage of each methodology, and presents lessons learned from the project management, system engineering, and developer’s perspectives.

  18. The mechanism and realization of a band-agile coaxial relativistic backward-wave oscillator

    SciTech Connect

    Ge, Xingjun; Zhang, Jun; Zhong, Huihuang; Qian, Baoliang; Wang, Haitao

    2014-11-03

    The mechanism and realization of a band-agile coaxial relativistic backward-wave oscillator (RBWO) are presented. The operation frequency tuning can be easily achieved by merely altering the inner-conductor length. The key effects of the inner-conductor length contributing to the mechanical frequency tunability are investigated theoretically and experimentally. There is a specific inner-conductor length where the operation frequency can jump from one mode to another mode, which belongs to a different operation band. In addition, the operation frequency is tunable within each operation band. During simulation, the L-band microwave with a frequency of 1.61 GHz is radiated when the inner-conductor length is 39 cm. Meanwhile, the S-band microwave with a frequency of 2.32 GHz is radiated when the inner-conductor length is 5 cm. The frequency adjustment bandwidths of L-band and S-band are about 8.5% and 2%, respectively. Moreover, the online mechanical tunability process is described in detail. In the initial experiment, the generated microwave frequencies remain approximately 1.59 GHz and 2.35 GHz when the inner-conductor lengths are 39 cm and 5 cm. In brief, this technical route of the band-agile coaxial RBWO is feasible and provides a guide to design other types of band-agile high power microwaves sources.

  19. Preform Characterization in VARTM Process Model Development

    NASA Technical Reports Server (NTRS)

    Grimsley, Brian W.; Cano, Roberto J.; Hubert, Pascal; Loos, Alfred C.; Kellen, Charles B.; Jensen, Brian J.

    2004-01-01

    Vacuum-Assisted Resin Transfer Molding (VARTM) is a Liquid Composite Molding (LCM) process where both resin injection and fiber compaction are achieved under pressures of 101.3 kPa or less. Originally developed over a decade ago for marine composite fabrication, VARTM is now considered a viable process for the fabrication of aerospace composites (1,2). In order to optimize and further improve the process, a finite element analysis (FEA) process model is being developed to include the coupled phenomenon of resin flow, preform compaction and resin cure. The model input parameters are obtained from resin and fiber-preform characterization tests. In this study, the compaction behavior and the Darcy permeability of a commercially available carbon fabric are characterized. The resulting empirical model equations are input to the 3- Dimensional Infiltration, version 5 (3DINFILv.5) process model to simulate infiltration of a composite panel.

  20. Modeling Cellular Processes in 3-D

    PubMed Central

    Mogilner, Alex; Odde, David

    2011-01-01

    Summary Recent advances in photonic imaging and fluorescent protein technology offer unprecedented views of molecular space-time dynamics in living cells. At the same time, advances in computing hardware and software enable modeling of ever more complex systems, from global climate to cell division. As modeling and experiment become more closely integrated, we must address the issue of modeling cellular processes in 3-D. Here, we highlight recent advances related to 3-D modeling in cell biology. While some processes require full 3-D analysis, we suggest that others are more naturally described in 2-D or 1-D. Keeping the dimensionality as low as possible reduces computational time and makes models more intuitively comprehensible; however, the ability to test full 3-D models will build greater confidence in models generally and remains an important emerging area of cell biological modeling. PMID:22036197

  1. Software-Engineering Process Simulation (SEPS) model

    NASA Technical Reports Server (NTRS)

    Lin, C. Y.; Abdel-Hamid, T.; Sherif, J. S.

    1992-01-01

    The Software Engineering Process Simulation (SEPS) model is described which was developed at JPL. SEPS is a dynamic simulation model of the software project development process. It uses the feedback principles of system dynamics to simulate the dynamic interactions among various software life cycle development activities and management decision making processes. The model is designed to be a planning tool to examine tradeoffs of cost, schedule, and functionality, and to test the implications of different managerial policies on a project's outcome. Furthermore, SEPS will enable software managers to gain a better understanding of the dynamics of software project development and perform postmodern assessments.

  2. Agile data management for curation of genomes to watershed datasets

    NASA Astrophysics Data System (ADS)

    Varadharajan, C.; Agarwal, D.; Faybishenko, B.; Versteeg, R.

    2015-12-01

    A software platform is being developed for data management and assimilation [DMA] as part of the U.S. Department of Energy's Genomes to Watershed Sustainable Systems Science Focus Area 2.0. The DMA components and capabilities are driven by the project science priorities and the development is based on agile development techniques. The goal of the DMA software platform is to enable users to integrate and synthesize diverse and disparate field, laboratory, and simulation datasets, including geological, geochemical, geophysical, microbiological, hydrological, and meteorological data across a range of spatial and temporal scales. The DMA objectives are (a) developing an integrated interface to the datasets, (b) storing field monitoring data, laboratory analytical results of water and sediments samples collected into a database, (c) providing automated QA/QC analysis of data and (d) working with data providers to modify high-priority field and laboratory data collection and reporting procedures as needed. The first three objectives are driven by user needs, while the last objective is driven by data management needs. The project needs and priorities are reassessed regularly with the users. After each user session we identify development priorities to match the identified user priorities. For instance, data QA/QC and collection activities have focused on the data and products needed for on-going scientific analyses (e.g. water level and geochemistry). We have also developed, tested and released a broker and portal that integrates diverse datasets from two different databases used for curation of project data. The development of the user interface was based on a user-centered design process involving several user interviews and constant interaction with data providers. The initial version focuses on the most requested feature - i.e. finding the data needed for analyses through an intuitive interface. Once the data is found, the user can immediately plot and download data

  3. Probabilistic models of language processing and acquisition.

    PubMed

    Chater, Nick; Manning, Christopher D

    2006-07-01

    Probabilistic methods are providing new explanatory approaches to fundamental cognitive science questions of how humans structure, process and acquire language. This review examines probabilistic models defined over traditional symbolic structures. Language comprehension and production involve probabilistic inference in such models; and acquisition involves choosing the best model, given innate constraints and linguistic and other input. Probabilistic models can account for the learning and processing of language, while maintaining the sophistication of symbolic models. A recent burgeoning of theoretical developments and online corpus creation has enabled large models to be tested, revealing probabilistic constraints in processing, undermining acquisition arguments based on a perceived poverty of the stimulus, and suggesting fruitful links with probabilistic theories of categorization and ambiguity resolution in perception.

  4. Fuel Conditioning Facility Electrorefiner Process Model

    SciTech Connect

    DeeEarl Vaden

    2005-10-01

    The Fuel Conditioning Facility at the Idaho National Laboratory processes spent nuclear fuel from the Experimental Breeder Reactor II using electro-metallurgical treatment. To process fuel without waiting for periodic sample analyses to assess process conditions, an electrorefiner process model predicts the composition of the electrorefiner inventory and effluent streams. For the chemical equilibrium portion of the model, the two common methods for solving chemical equilibrium problems, stoichiometric and non stoichiometric, were investigated. In conclusion, the stoichiometric method produced equilibrium compositions close to the measured results whereas the non stoichiometric method did not.

  5. Cost Models for MMC Manufacturing Processes

    NASA Technical Reports Server (NTRS)

    Elzey, Dana M.; Wadley, Haydn N. G.

    1996-01-01

    Processes for the manufacture of advanced metal matrix composites are rapidly approaching maturity in the research laboratory and there is growing interest in their transition to industrial production. However, research conducted to date has almost exclusively focused on overcoming the technical barriers to producing high-quality material and little attention has been given to the economical feasibility of these laboratory approaches and process cost issues. A quantitative cost modeling (QCM) approach was developed to address these issues. QCM are cost analysis tools based on predictive process models relating process conditions to the attributes of the final product. An important attribute, of the QCM approach is the ability to predict the sensitivity of material production costs to product quality and to quantitatively explore trade-offs between cost and quality. Applications of the cost models allow more efficient direction of future MMC process technology development and a more accurate assessment of MMC market potential. Cost models were developed for two state-of-the art metal matrix composite (MMC) manufacturing processes: tape casting and plasma spray deposition. Quality and Cost models are presented for both processes and the resulting predicted quality-cost curves are presented and discussed.

  6. Modeling approach for business process reengineering

    NASA Astrophysics Data System (ADS)

    Tseng, Mitchell M.; Chen, Yuliu

    1995-08-01

    The purpose of this paper is to introduce a modeling approach to define, simulate, animate, and control business processes. The intent is to introduce the undergoing methodology to build tools for designing and managing business processes. Similar to computer aided design (CAD) for mechanical parts, CAD tools are needed for designing business processes. It emphasizes the dynamic behavior of business process. The proposed modeling technique consists of a definition of each individual activity, the network of activities, a control mechanism that describes coordination of these activities, and events that will flow through these activities. Based on the formalism introduced in this modeling technique, users will be able to define business process with minimum ambiguity, take snap shots of particular events in the process, describe the accountability of participants, and view a replay of event streams in the process flow. This modeling approach, mapped on top of a commercial software, has been tested by using examples from real life business process. The examples and testing helped us to identify some of the strengths and weaknesses of this proposed approach.

  7. The Southern Argentine Agile Meteor Radar (SAAMER)

    NASA Astrophysics Data System (ADS)

    Janches, Diego

    2014-11-01

    The Southern Argentina Agile Meteor Radar (SAAMER) is a new generation system deployed in Rio Grande, Tierra del Fuego, Argentina (53 S) in May 2008. SAAMER transmits 10 times more power than regular meteor radars, and uses a newly developed transmitting array, which focuses power upward instead of the traditional single-antenna-all-sky configuration. The system is configured such that the transmitter array can also be utilized as a receiver. The new design greatly increases the sensitivity of the radar enabling the detection of large number of particles at low zenith angles. The more concentrated transmitted power enables additional meteor studies besides those typical of these systems based on the detection of specular reflections, such as routine detections of head echoes and non-specular trails, previously only possible with High Power and Large Aperture radars. In August 2010, SAAMER was upgraded to a system capable to determine meteoroid orbital parameters. This was achieved by adding two remote receiving stations approximately 10 km away from the main site in near perpendicular directions. The upgrade significantly expands the science that is achieved with this new radar enabling us to study the orbital properties of the interplanetary dust environment. Because of the unique geographical location, SAAMER allows for additional inter-hemispheric comparison with measurements from Canadian Meteor Orbit Radar, which is geographically conjugate. Initial surveys show, for example, that SAAMER observes a very strong contribution of the South Toroidal Sporadic meteor source, of which limited observational data is available. In addition, SAAMER offers similar unique capabilities for meteor showers and streams studies given the range of ecliptic latitudes that the system enables detailed study of showers at high southern latitudes (e.g July Phoenicids or Puppids complex). Finally, SAAMER is ideal for the deployment of complementary instrumentation in both, permanent

  8. The (Mathematical) Modeling Process in Biosciences

    PubMed Central

    Torres, Nestor V.; Santos, Guido

    2015-01-01

    In this communication, we introduce a general framework and discussion on the role of models and the modeling process in the field of biosciences. The objective is to sum up the common procedures during the formalization and analysis of a biological problem from the perspective of Systems Biology, which approaches the study of biological systems as a whole. We begin by presenting the definitions of (biological) system and model. Particular attention is given to the meaning of mathematical model within the context of biology. Then, we present the process of modeling and analysis of biological systems. Three stages are described in detail: conceptualization of the biological system into a model, mathematical formalization of the previous conceptual model and optimization and system management derived from the analysis of the mathematical model. All along this work the main features and shortcomings of the process are analyzed and a set of rules that could help in the task of modeling any biological system are presented. Special regard is given to the formative requirements and the interdisciplinary nature of this approach. We conclude with some general considerations on the challenges that modeling is posing to current biology. PMID:26734063

  9. Modeling aerosol processes at the local scale

    SciTech Connect

    Lazaridis, M.; Isukapalli, S.S.; Georgopoulos, P.G.

    1998-12-31

    This work presents an approach for modeling photochemical gaseous and aerosol phase processes in subgrid plumes from major localized (e.g. point) sources (plume-in-grid modeling), thus improving the ability to quantify the relationship between emission source activity and ambient air quality. This approach employs the Reactive Plume Model (RPM-AERO) which extends the regulatory model RPM-IV by incorporating aerosol processes and heterogeneous chemistry. The physics and chemistry of elemental carbon, organic carbon, sulfate, sodium, chloride and crustal material of aerosols are treated and attributed to the PM size distribution. A modified version of the Carbon Bond IV chemical mechanism is included to model the formation of organic aerosol, and the inorganic multicomponent atmospheric aerosol equilibrium model, SEQUILIB is used for calculating the amounts of inorganic species in particulate matter. Aerosol dynamics modeled include mechanisms of nucleation, condensation and gas/particle partitioning of organic matter. An integrated trajectory-in-grid modeling system, UAM/RPM-AERO, is under continuing development for extracting boundary and initial conditions from the mesoscale photochemical/aerosol model UAM-AERO. The RPM-AERO is applied here to case studies involving emissions from point sources to study sulfate particle formation in plumes. Model calculations show that homogeneous nucleation is an efficient process for new particle formation in plumes, in agreement with previous field studies and theoretical predictions.

  10. The Epidemic Process and The Contagion Model

    ERIC Educational Resources Information Center

    Worthen, Dennis B.

    1973-01-01

    Goffman's epidemic theory is presented and compared to the contagion theory developed by Menzel. An attempt is made to compare the two models presented and examine their similarities and differences. The conclusion drawn is that the two models are very similar in their approach to understanding communication processes. (14 references) (Author/SJ)

  11. Dynamic process modeling with recurrent neural networks

    SciTech Connect

    You, Yong; Nikolaou, M. . Dept. of Chemical Engineering)

    1993-10-01

    Mathematical models play an important role in control system synthesis. However, due to the inherent nonlinearity, complexity and uncertainty of chemical processes, it is usually difficult to obtain an accurate model for a chemical engineering system. A method of nonlinear static and dynamic process modeling via recurrent neural networks (RNNs) is studied. An RNN model is a set of coupled nonlinear ordinary differential equations in continuous time domain with nonlinear dynamic node characteristics as well as both feed forward and feedback connections. For such networks, each physical input to a system corresponds to exactly one input to the network. The system's dynamics are captured by the internal structure of the network. The structure of RNN models may be more natural and attractive than that of feed forward neural network models, but computation time for training is longer. Simulation results show that RNNs can learn both steady-state relationships and process dynamics of continuous and batch, single-input/single-output and multi-input/multi-output systems in a simple and direct manner. Training of RNNs shows only small degradation in the presence of noise in the training data. Thus, RNNs constitute a feasible alternative to layered feed forward back propagation neural networks in steady-state and dynamic process modeling and model-based control.

  12. Modeling of fluidized bed silicon deposition process

    NASA Technical Reports Server (NTRS)

    Kim, K.; Hsu, G.; Lutwack, R.; PRATURI A. K.

    1977-01-01

    The model is intended for use as a means of improving fluidized bed reactor design and for the formulation of the research program in support of the contracts of Silicon Material Task for the development of the fluidized bed silicon deposition process. A computer program derived from the simple modeling is also described. Results of some sample calculations using the computer program are shown.

  13. Theoretical Models and Processes of Reading.

    ERIC Educational Resources Information Center

    Singer, Harry, Ed.; Ruddell, Robert B., Ed.

    The first section of this two-part collection of articles contains six papers and their discussions read at a symposium on Theoretical Models and Processes of Reading. The papers cover the linguistic, perceptual, and cognitive components involved in reading. The models attempt to integrate the variables that influence the perception, recognition,…

  14. Modeling biochemical transformation processes and information processing with Narrator

    PubMed Central

    Mandel, Johannes J; Fuß, Hendrik; Palfreyman, Niall M; Dubitzky, Werner

    2007-01-01

    Background Software tools that model and simulate the dynamics of biological processes and systems are becoming increasingly important. Some of these tools offer sophisticated graphical user interfaces (GUIs), which greatly enhance their acceptance by users. Such GUIs are based on symbolic or graphical notations used to describe, interact and communicate the developed models. Typically, these graphical notations are geared towards conventional biochemical pathway diagrams. They permit the user to represent the transport and transformation of chemical species and to define inhibitory and stimulatory dependencies. A critical weakness of existing tools is their lack of supporting an integrative representation of transport, transformation as well as biological information processing. Results Narrator is a software tool facilitating the development and simulation of biological systems as Co-dependence models. The Co-dependence Methodology complements the representation of species transport and transformation together with an explicit mechanism to express biological information processing. Thus, Co-dependence models explicitly capture, for instance, signal processing structures and the influence of exogenous factors or events affecting certain parts of a biological system or process. This combined set of features provides the system biologist with a powerful tool to describe and explore the dynamics of life phenomena. Narrator's GUI is based on an expressive graphical notation which forms an integral part of the Co-dependence Methodology. Behind the user-friendly GUI, Narrator hides a flexible feature which makes it relatively easy to map models defined via the graphical notation to mathematical formalisms and languages such as ordinary differential equations, the Systems Biology Markup Language or Gillespie's direct method. This powerful feature facilitates reuse, interoperability and conceptual model development. Conclusion Narrator is a flexible and intuitive systems

  15. Spiked Dirichlet Process Priors for Gaussian Process Models

    PubMed Central

    Savitsky, Terrance; Vannucci, Marina

    2013-01-01

    We expand a framework for Bayesian variable selection for Gaussian process (GP) models by employing spiked Dirichlet process (DP) prior constructions over set partitions containing covariates. Our approach results in a nonparametric treatment of the distribution of the covariance parameters of the GP covariance matrix that in turn induces a clustering of the covariates. We evaluate two prior constructions: the first one employs a mixture of a point-mass and a continuous distribution as the centering distribution for the DP prior, therefore, clustering all covariates. The second one employs a mixture of a spike and a DP prior with a continuous distribution as the centering distribution, which induces clustering of the selected covariates only. DP models borrow information across covariates through model-based clustering. Our simulation results, in particular, show a reduction in posterior sampling variability and, in turn, enhanced prediction performances. In our model formulations, we accomplish posterior inference by employing novel combinations and extensions of existing algorithms for inference with DP prior models and compare performances under the two prior constructions. PMID:23950763

  16. Filament winding cylinders. I - Process model

    NASA Technical Reports Server (NTRS)

    Lee, Soo-Yong; Springer, George S.

    1990-01-01

    A model was developed which describes the filament winding process of composite cylinders. The model relates the significant process variables such as winding speed, fiber tension, and applied temperature to the thermal, chemical and mechanical behavior of the composite cylinder and the mandrel. Based on the model, a user friendly code was written which can be used to calculate (1) the temperature in the cylinder and the mandrel, (2) the degree of cure and viscosity in the cylinder, (3) the fiber tensions and fiber positions, (4) the stresses and strains in the cylinder and in the mandrel, and (5) the void diameters in the cylinder.

  17. Using Perspective to Model Complex Processes

    SciTech Connect

    Kelsey, R.L.; Bisset, K.R.

    1999-04-04

    The notion of perspective, when supported in an object-based knowledge representation, can facilitate better abstractions of reality for modeling and simulation. The object modeling of complex physical and chemical processes is made more difficult in part due to the poor abstractions of state and phase changes available in these models. The notion of perspective can be used to create different views to represent the different states of matter in a process. These techniques can lead to a more understandable model. Additionally, the ability to record the progress of a process from start to finish is problematic. It is desirable to have a historic record of the entire process, not just the end result of the process. A historic record should facilitate backtracking and re-start of a process at different points in time. The same representation structures and techniques can be used to create a sequence of process markers to represent a historic record. By using perspective, the sequence of markers can have multiple and varying views tailored for a particular user's context of interest.

  18. Modeling the VARTM Composite Manufacturing Process

    NASA Technical Reports Server (NTRS)

    Song, Xiao-Lan; Loos, Alfred C.; Grimsley, Brian W.; Cano, Roberto J.; Hubert, Pascal

    2004-01-01

    A comprehensive simulation model of the Vacuum Assisted Resin Transfer Modeling (VARTM) composite manufacturing process has been developed. For isothermal resin infiltration, the model incorporates submodels which describe cure of the resin and changes in resin viscosity due to cure, resin flow through the reinforcement preform and distribution medium and compaction of the preform during the infiltration. The accuracy of the model was validated by measuring the flow patterns during resin infiltration of flat preforms. The modeling software was used to evaluate the effects of the distribution medium on resin infiltration of a flat preform. Different distribution medium configurations were examined using the model and the results were compared with data collected during resin infiltration of a carbon fabric preform. The results of the simulations show that the approach used to model the distribution medium can significantly effect the predicted resin infiltration times. Resin infiltration into the preform can be accurately predicted only when the distribution medium is modeled correctly.

  19. From Business Value Model to Coordination Process Model

    NASA Astrophysics Data System (ADS)

    Fatemi, Hassan; van Sinderen, Marten; Wieringa, Roel

    The increased complexity of business webs calls for modeling the collaboration of enterprises from different perspectives, in particular the business and process perspectives, and for mutually aligning these perspectives. Business value modeling and coordination process modeling both are necessary for a good e-business design, but these activities have different goals and use different concepts. Nevertheless, the resulting models should be consistent with each other because they refer to the same system from different perspectives. Hence, checking the consistency between these models or producing one based on the other would be of high value. In this paper we discuss the issue of achieving consistency in multi-level e-business design and give guidelines to produce consistent coordination process models from business value models in a stepwise manner.

  20. Mathematical modeling of the coating process.

    PubMed

    Toschkoff, Gregor; Khinast, Johannes G

    2013-12-01

    Coating of tablets is a common unit operation in the pharmaceutical industry. In most cases, the final product must meet strict quality requirements; to meet them, a detailed understanding of the coating process is required. To this end, numerous experiment studies have been performed. However, to acquire a mechanistic understanding, experimental data must be interpreted in the light of mathematical models. In recent years, a combination of analytical modeling and computational simulations enabled deeper insights into the nature of the coating process. This paper presents an overview of modeling and simulation approaches of the coating process, covering various relevant aspects from scale-up considerations to coating mass uniformity investigations and models for drop atomization. The most important analytical and computational concepts are presented and the findings are compared.

  1. Incorporating process variability into stormwater quality modelling.

    PubMed

    Wijesiri, Buddhi; Egodawatta, Prasanna; McGree, James; Goonetilleke, Ashantha

    2015-11-15

    Process variability in pollutant build-up and wash-off generates inherent uncertainty that affects the outcomes of stormwater quality models. Poor characterisation of process variability constrains the accurate accounting of the uncertainty associated with pollutant processes. This acts as a significant limitation to effective decision making in relation to stormwater pollution mitigation. The study undertaken developed three theoretical scenarios based on research findings that variations in particle size fractions <150 μm and >150 μm during pollutant build-up and wash-off primarily determine the variability associated with these processes. These scenarios, which combine pollutant build-up and wash-off processes that takes place on a continuous timeline, are able to explain process variability under different field conditions. Given the variability characteristics of a specific build-up or wash-off event, the theoretical scenarios help to infer the variability characteristics of the associated pollutant process that follows. Mathematical formulation of the theoretical scenarios enables the incorporation of variability characteristics of pollutant build-up and wash-off processes in stormwater quality models. The research study outcomes will contribute to the quantitative assessment of uncertainty as an integral part of the interpretation of stormwater quality modelling outcomes.

  2. Mathematical modeling of biomass fuels formation process.

    PubMed

    Gaska, Krzysztof; Wandrasz, Andrzej J

    2008-01-01

    The increasing demand for thermal and electric energy in many branches of industry and municipal management accounts for a drastic diminishing of natural resources (fossil fuels). Meanwhile, in numerous technical processes, a huge mass of wastes is produced. A segregated and converted combustible fraction of the wastes, with relatively high calorific value, may be used as a component of formed fuels. The utilization of the formed fuel components from segregated groups of waste in associated processes of co-combustion with conventional fuels causes significant savings resulting from partial replacement of fossil fuels, and reduction of environmental pollution resulting directly from the limitation of waste migration to the environment (soil, atmospheric air, surface and underground water). The realization of technological processes with the utilization of formed fuel in associated thermal systems should be qualified by technical criteria, which means that elementary processes as well as factors of sustainable development, from a global viewpoint, must not be disturbed. The utilization of post-process waste should be preceded by detailed technical, ecological and economic analyses. In order to optimize the mixing process of fuel components, a mathematical model of the forming process was created. The model is defined as a group of data structures which uniquely identify a real process and conversion of this data in algorithms based on a problem of linear programming. The paper also presents the optimization of parameters in the process of forming fuels using a modified simplex algorithm with a polynomial worktime. This model is a datum-point in the numerical modeling of real processes, allowing a precise determination of the optimal elementary composition of formed fuels components, with assumed constraints and decision variables of the task.

  3. Architected Agile Solutions for Software-Reliant Systems

    NASA Astrophysics Data System (ADS)

    Boehm, Barry; Lane, Jo Ann; Koolmanojwong, Supannika; Turner, Richard

    Systems are becoming increasingly reliant on software due to needs for rapid fielding of “70% capabilities,” interoperability, net-centricity, and rapid adaptation to change. The latter need has led to increased interest in agile methods of software development, in which teams rely on shared tacit interpersonal knowledge rather than explicit documented knowledge. However, such systems often need to be scaled up to higher level of performance and assurance, requiring stronger architectural support. Several organizations have recently transformed themselves by developing successful combinations of agility and architecture that can scale to projects of up to 100 personnel. This chapter identifies a set of key principles for such architected agile solutions for software-reliant systems, provides guidance for how much architecting is enough, and illustrates the key principles with several case studies.

  4. Quantum mechanical Hamiltonian models of discrete processes

    SciTech Connect

    Benioff, P.

    1981-03-01

    Here the results of other work on quantum mechanical Hamiltonian models of Turing machines are extended to include any discrete process T on a countably infinite set A. The models are constructed here by use of scattering phase shifts from successive scatterers to turn on successive step interactions. Also a locality requirement is imposed. The construction is done by first associating with each process T a model quantum system M with associated Hilbert space H/sub M/ and step operator U/sub T/. Since U/sub T/ is not unitary in general, M, H/sub M/, and U/sub T/ are extended into a (continuous time) Hamiltonian model on a larger space which satisfies the locality requirement. The construction is compared with the minimal unitary dilation of U/sub T/. It is seen that the model constructed here is larger than the minimal one. However, the minimal one does not satisfy the locality requirement.

  5. The AGILE Mission and Gamma-Ray Bursts

    SciTech Connect

    Longo, Francesco; Tavani, M.; Barbiellini, G.; Argan, A.; Basset, M.; Boffelli, F.; Bulgarelli, A.; Caraveo, P.; Cattaneo, P.; Chen, A.; Costa, E.; Del Monte, E.; Di Cocco, G.; Di Persio, G.; Donnarumma, I.; Feroci, M.; Fiorini, M.; Foggetta, L.; Froysland, T.; Frutti, M.

    2007-05-01

    The AGILE Mission will explore the gamma-ray Universe with a very innovative instrument combining for the first time a gamma-ray imager and a hard X-ray imager. AGILE will be operational at the beginning of 2007 and it will provide crucial data for the study of Active Galactic Nuclei, Gamma-Ray Bursts, unidentified gamma-ray sources, Galactic compact objects, supernova remnants, TeV sources, and fundamental physics by microsecond timing. The AGILE instrument is designed to simultaneously detect and image photons in the 30 MeV - 50 GeV and 15 - 45 keV energy bands with excellent imaging and timing capabilities, and a large field of view covering {approx} 1/5 of the entire sky at energies above 30 MeV. A CsI calorimeter is capable of GRB triggering in the energy band 0.3-50 MeV. The broadband detection of GRBs and the study of implications for particle acceleration and high energy emission are primary goals of the mission. AGILE can image GRBs with 2-3 arcminute error boxes in the hard X-ray range, and provide broadband photon-by photon detection in the 15-45 keV, 03-50 MeV, and 30 MeV-30 GeV energy ranges. Microsecond on-board photon tagging and a {approx} 100 microsecond gamma-ray detection deadtime will be crucial for fast GRB timing. On-board calculated GRB coordinates and energy fluxes will be quickly transmitted to the ground by an ORBCOMM transceiver. AGILE is now (January 2007) undergoing final satellite integration and testing. The PLS V launch is planned in spring 2007. AGILE is then foreseen to be fully operational during the summer of 2007.

  6. Software Engineering Laboratory (SEL) cleanroom process model

    NASA Technical Reports Server (NTRS)

    Green, Scott; Basili, Victor; Godfrey, Sally; Mcgarry, Frank; Pajerski, Rose; Waligora, Sharon

    1991-01-01

    The Software Engineering Laboratory (SEL) cleanroom process model is described. The term 'cleanroom' originates in the integrated circuit (IC) production process, where IC's are assembled in dust free 'clean rooms' to prevent the destructive effects of dust. When applying the clean room methodology to the development of software systems, the primary focus is on software defect prevention rather than defect removal. The model is based on data and analysis from previous cleanroom efforts within the SEL and is tailored to serve as a guideline in applying the methodology to future production software efforts. The phases that are part of the process model life cycle from the delivery of requirements to the start of acceptance testing are described. For each defined phase, a set of specific activities is discussed, and the appropriate data flow is described. Pertinent managerial issues, key similarities and differences between the SEL's cleanroom process model and the standard development approach used on SEL projects, and significant lessons learned from prior cleanroom projects are presented. It is intended that the process model described here will be further tailored as additional SEL cleanroom projects are analyzed.

  7. Computational Modeling in Structural Materials Processing

    NASA Technical Reports Server (NTRS)

    Meyyappan, Meyya; Arnold, James O. (Technical Monitor)

    1997-01-01

    High temperature materials such as silicon carbide, a variety of nitrides, and ceramic matrix composites find use in aerospace, automotive, machine tool industries and in high speed civil transport applications. Chemical vapor deposition (CVD) is widely used in processing such structural materials. Variations of CVD include deposition on substrates, coating of fibers, inside cavities and on complex objects, and infiltration within preforms called chemical vapor infiltration (CVI). Our current knowledge of the process mechanisms, ability to optimize processes, and scale-up for large scale manufacturing is limited. In this regard, computational modeling of the processes is valuable since a validated model can be used as a design tool. The effort is similar to traditional chemically reacting flow modeling with emphasis on multicomponent diffusion, thermal diffusion, large sets of homogeneous reactions, and surface chemistry. In the case of CVI, models for pore infiltration are needed. In the present talk, examples of SiC nitride, and Boron deposition from the author's past work will be used to illustrate the utility of computational process modeling.

  8. Dynamic occupancy models for explicit colonization processes.

    PubMed

    Broms, Kristin M; Hooten, Mevin B; Johnson, Devin S; Altwegg, Res; Conquest, Loveday L

    2016-01-01

    The dynamic, multi-season occupancy model framework has become a popular tool for modeling open populations with occupancies that change over time through local colonizations and extinctions. However, few versions of the model relate these probabilities to the occupancies of neighboring sites or patches. We present a modeling framework that incorporates this information and is capable of describing a wide variety of spatiotemporal colonization and extinction processes. A key feature of the model is that it is based on a simple set of small-scale rules describing how the process evolves. The result is a dynamic process that can account for complicated large-scale features. In our model, a site is more likely to be colonized if more of its neighbors were previously occupied and if it provides more appealing environmental characteristics than its neighboring sites. Additionally, a site without occupied neighbors may also become colonized through the inclusion of a long-distance dispersal process. Although similar model specifications have been developed for epidemiological applications, ours formally accounts for detectability using the well-known occupancy modeling framework. After demonstrating the viability and potential of this new form of dynamic occupancy model in a simulation study, we use it to obtain inference for the ongoing Common Myna (Acridotheres tristis) invasion in South Africa. Our results suggest that the Common Myna continues to enlarge its distribution and its spread via short distance movement, rather than long-distance dispersal. Overall, this new modeling framework provides a powerful tool for managers examining the drivers of colonization including short- vs. long-distance dispersal, habitat quality, and distance from source populations. PMID:27008788

  9. Dynamic occupancy models for explicit colonization processes

    USGS Publications Warehouse

    Broms, Kristin M.; Hooten, Mevin B.; Johnson, Devin S.; Altwegg, Res; Conquest, Loveday

    2016-01-01

    The dynamic, multi-season occupancy model framework has become a popular tool for modeling open populations with occupancies that change over time through local colonizations and extinctions. However, few versions of the model relate these probabilities to the occupancies of neighboring sites or patches. We present a modeling framework that incorporates this information and is capable of describing a wide variety of spatiotemporal colonization and extinction processes. A key feature of the model is that it is based on a simple set of small-scale rules describing how the process evolves. The result is a dynamic process that can account for complicated large-scale features. In our model, a site is more likely to be colonized if more of its neighbors were previously occupied and if it provides more appealing environmental characteristics than its neighboring sites. Additionally, a site without occupied neighbors may also become colonized through the inclusion of a long-distance dispersal process. Although similar model specifications have been developed for epidemiological applications, ours formally accounts for detectability using the well-known occupancy modeling framework. After demonstrating the viability and potential of this new form of dynamic occupancy model in a simulation study, we use it to obtain inference for the ongoing Common Myna (Acridotheres tristis) invasion in South Africa. Our results suggest that the Common Myna continues to enlarge its distribution and its spread via short distance movement, rather than long-distance dispersal. Overall, this new modeling framework provides a powerful tool for managers examining the drivers of colonization including short- vs. long-distance dispersal, habitat quality, and distance from source populations.

  10. Causally nonseparable processes admitting a causal model

    NASA Astrophysics Data System (ADS)

    Feix, Adrien; Araújo, Mateus; Brukner, Časlav

    2016-08-01

    A recent framework of quantum theory with no global causal order predicts the existence of ‘causally nonseparable’ processes. Some of these processes produce correlations incompatible with any causal order (they violate so-called ‘causal inequalities’ analogous to Bell inequalities) while others do not (they admit a ‘causal model’ analogous to a local model). Here we show for the first time that bipartite causally nonseparable processes with a causal model exist, and give evidence that they have no clear physical interpretation. We also provide an algorithm to generate processes of this kind and show that they have nonzero measure in the set of all processes. We demonstrate the existence of processes which stop violating causal inequalities but are still causally nonseparable when mixed with a certain amount of ‘white noise’. This is reminiscent of the behavior of Werner states in the context of entanglement and nonlocality. Finally, we provide numerical evidence for the existence of causally nonseparable processes which have a causal model even when extended with an entangled state shared among the parties.

  11. Implementation of the Business Process Modelling Notation (BPMN) in the modelling of anatomic pathology processes

    PubMed Central

    Rojo, Marcial García; Rolón, Elvira; Calahorra, Luis; García, Felix Óscar; Sánchez, Rosario Paloma; Ruiz, Francisco; Ballester, Nieves; Armenteros, María; Rodríguez, Teresa; Espartero, Rafael Martín

    2008-01-01

    Background Process orientation is one of the essential elements of quality management systems, including those in use in healthcare. Business processes in hospitals are very complex and variable. BPMN (Business Process Modelling Notation) is a user-oriented language specifically designed for the modelling of business (organizational) processes. Previous experiences of the use of this notation in the processes modelling within the Pathology in Spain or another country are not known. We present our experience in the elaboration of the conceptual models of Pathology processes, as part of a global programmed surgical patient process, using BPMN. Methods With the objective of analyzing the use of BPMN notation in real cases, a multidisciplinary work group was created, including software engineers from the Dep. of Technologies and Information Systems from the University of Castilla-La Mancha and health professionals and administrative staff from the Hospital General de Ciudad Real. The work in collaboration was carried out in six phases: informative meetings, intensive training, process selection, definition of the work method, process describing by hospital experts, and process modelling. Results The modelling of the processes of Anatomic Pathology is presented using BPMN. The presented subprocesses are those corresponding to the surgical pathology examination of the samples coming from operating theatre, including the planning and realization of frozen studies. Conclusion The modelling of Anatomic Pathology subprocesses has allowed the creation of an understandable graphical model, where management and improvements are more easily implemented by health professionals. PMID:18673511

  12. Stochastic differential equation model to Prendiville processes

    SciTech Connect

    Granita; Bahar, Arifah

    2015-10-22

    The Prendiville process is another variation of the logistic model which assumes linearly decreasing population growth rate. It is a continuous time Markov chain (CTMC) taking integer values in the finite interval. The continuous time Markov chain can be approximated by stochastic differential equation (SDE). This paper discusses the stochastic differential equation of Prendiville process. The work started with the forward Kolmogorov equation in continuous time Markov chain of Prendiville process. Then it was formulated in the form of a central-difference approximation. The approximation was then used in Fokker-Planck equation in relation to the stochastic differential equation of the Prendiville process. The explicit solution of the Prendiville process was obtained from the stochastic differential equation. Therefore, the mean and variance function of the Prendiville process could be easily found from the explicit solution.

  13. Agile radio resource management for proactive wireless networking (Invited Paper)

    NASA Astrophysics Data System (ADS)

    Brothers, L. Reggie; MacMullan, Samuel J.; Brown, Kevin L.; DeBardelaben, James A.

    2005-05-01

    Current military operational effectiveness can degrade rapidly with increasing communications stresses such as heavy throughput and QoS demands from disadvantaged users exposed to severe channel impairments and communications threats. This paper proposes a distributed and agile radio resource management (RRM) system to maintain mission effectiveness even under significant communications stress. Agile RRM includes a well-coordinated cross-layer design with the introduction of new OSI layer features and interactions as well as methods to incorporate communications constraints and requirements in systems controlling mission planning and execution.

  14. Cost Models for MMC Manufacturing Processes

    NASA Technical Reports Server (NTRS)

    Elzey, Dana M.; Wadley, Haydn N. G.

    1996-01-01

    The quality cost modeling (QCM) tool is intended to be a relatively simple-to-use device for obtaining a first-order assessment of the quality-cost relationship for a given process-material combination. The QCM curve is a plot of cost versus quality (an index indicating microstructural quality), which is unique for a given process-material combination. The QCM curve indicates the tradeoff between cost and performance, thus enabling one to evaluate affordability. Additionally, the effect of changes in process design, raw materials, and process conditions on the cost-quality relationship can be evaluated. Such results might indicate the most efficient means to obtain improved quality at reduced cost by process design refinements, the implementation of sensors and models for closed loop process control, or improvement in the properties of raw materials being fed into the process. QCM also allows alternative processes for producing the same or similar material to be compared in terms of their potential for producing competitively priced, high quality material. Aside from demonstrating the usefulness of the QCM concept, this is one of the main foci of the present research program, namely to compare processes for making continuous fiber reinforced, metal matrix composites (MMC's). Two processes, low pressure plasma spray deposition and tape casting are considered for QCM development. This document consists of a detailed look at the design of the QCM approach, followed by discussion of the application of QCM to each of the selected MMC manufacturing processes along with results, comparison of processes, and finally, a summary of findings and recommendations.

  15. A copula model for marked point processes.

    PubMed

    Diao, Liqun; Cook, Richard J; Lee, Ker-Ai

    2013-10-01

    Many chronic diseases feature recurring clinically important events. In addition, however, there often exists a random variable which is realized upon the occurrence of each event reflecting the severity of the event, a cost associated with it, or possibly a short term response indicating the effect of a therapeutic intervention. We describe a novel model for a marked point process which incorporates a dependence between continuous marks and the event process through the use of a copula function. The copula formulation ensures that event times can be modeled by any intensity function for point processes, and any multivariate model can be specified for the continuous marks. The relative efficiency of joint versus separate analyses of the event times and the marks is examined through simulation under random censoring. An application to data from a recent trial in transfusion medicine is given for illustration. PMID:23660874

  16. Incorporating evolutionary processes into population viability models.

    PubMed

    Pierson, Jennifer C; Beissinger, Steven R; Bragg, Jason G; Coates, David J; Oostermeijer, J Gerard B; Sunnucks, Paul; Schumaker, Nathan H; Trotter, Meredith V; Young, Andrew G

    2015-06-01

    We examined how ecological and evolutionary (eco-evo) processes in population dynamics could be better integrated into population viability analysis (PVA). Complementary advances in computation and population genomics can be combined into an eco-evo PVA to offer powerful new approaches to understand the influence of evolutionary processes on population persistence. We developed the mechanistic basis of an eco-evo PVA using individual-based models with individual-level genotype tracking and dynamic genotype-phenotype mapping to model emergent population-level effects, such as local adaptation and genetic rescue. We then outline how genomics can allow or improve parameter estimation for PVA models by providing genotypic information at large numbers of loci for neutral and functional genome regions. As climate change and other threatening processes increase in rate and scale, eco-evo PVAs will become essential research tools to evaluate the effects of adaptive potential, evolutionary rescue, and locally adapted traits on persistence.

  17. Relativistic diffusion processes and random walk models

    SciTech Connect

    Dunkel, Joern; Talkner, Peter; Haenggi, Peter

    2007-02-15

    The nonrelativistic standard model for a continuous, one-parameter diffusion process in position space is the Wiener process. As is well known, the Gaussian transition probability density function (PDF) of this process is in conflict with special relativity, as it permits particles to propagate faster than the speed of light. A frequently considered alternative is provided by the telegraph equation, whose solutions avoid superluminal propagation speeds but suffer from singular (noncontinuous) diffusion fronts on the light cone, which are unlikely to exist for massive particles. It is therefore advisable to explore other alternatives as well. In this paper, a generalized Wiener process is proposed that is continuous, avoids superluminal propagation, and reduces to the standard Wiener process in the nonrelativistic limit. The corresponding relativistic diffusion propagator is obtained directly from the nonrelativistic Wiener propagator, by rewriting the latter in terms of an integral over actions. The resulting relativistic process is non-Markovian, in accordance with the known fact that nontrivial continuous, relativistic Markov processes in position space cannot exist. Hence, the proposed process defines a consistent relativistic diffusion model for massive particles and provides a viable alternative to the solutions of the telegraph equation.

  18. Quantitative Modeling of Earth Surface Processes

    NASA Astrophysics Data System (ADS)

    Pelletier, Jon D.

    This textbook describes some of the most effective and straightforward quantitative techniques for modeling Earth surface processes. By emphasizing a core set of equations and solution techniques, the book presents state-of-the-art models currently employed in Earth surface process research, as well as a set of simple but practical research tools. Detailed case studies demonstrate application of the methods to a wide variety of processes including hillslope, fluvial, aeolian, glacial, tectonic, and climatic systems. Exercises at the end of each chapter begin with simple calculations and then progress to more sophisticated problems that require computer programming. All the necessary computer codes are available online at www.cambridge.org/9780521855976. Assuming some knowledge of calculus and basic programming experience, this quantitative textbook is designed for advanced geomorphology courses and as a reference book for professional researchers in Earth and planetary science looking for a quantitative approach to Earth surface processes.

  19. More details...
  20. A neurolinguistic model of grammatical construction processing.

    PubMed

    Dominey, Peter Ford; Hoen, Michel; Inui, Toshio

    2006-12-01

    One of the functions of everyday human language is to communicate meaning. Thus, when one hears or reads the sentence, "John gave a book to Mary," some aspect of an event concerning the transfer of possession of a book from John to Mary is (hopefully) transmitted. One theoretical approach to language referred to as construction grammar emphasizes this link between sentence structure and meaning in the form of grammatical constructions. The objective of the current research is to (1) outline a functional description of grammatical construction processing based on principles of psycholinguistics, (2) develop a model of how these functions can be implemented in human neurophysiology, and then (3) demonstrate the feasibility of the resulting model in processing languages of typologically diverse natures, that is, English, French, and Japanese. In this context, particular interest will be directed toward the processing of novel compositional structure of relative phrases. The simulation results are discussed in the context of recent neurophysiological studies of language processing.

  21. A process algebra model of QED

    NASA Astrophysics Data System (ADS)

    Sulis, William

    2016-03-01

    The process algebra approach to quantum mechanics posits a finite, discrete, determinate ontology of primitive events which are generated by processes (in the sense of Whitehead). In this ontology, primitive events serve as elements of an emergent space-time and of emergent fundamental particles and fields. Each process generates a set of primitive elements, using only local information, causally propagated as a discrete wave, forming a causal space termed a causal tapestry. Each causal tapestry forms a discrete and finite sampling of an emergent causal manifold (space-time) M and emergent wave function. Interactions between processes are described by a process algebra which possesses 8 commutative operations (sums and products) together with a non-commutative concatenation operator (transitions). The process algebra possesses a representation via nondeterministic combinatorial games. The process algebra connects to quantum mechanics through the set valued process and configuration space covering maps, which associate each causal tapestry with sets of wave functions over M. Probabilities emerge from interactions between processes. The process algebra model has been shown to reproduce many features of the theory of non-relativistic scalar particles to a high degree of accuracy, without paradox or divergences. This paper extends the approach to a semi-classical form of quantum electrodynamics.

  1. Hydrothermal processing of Hanford tank wastes: Process modeling and control

    SciTech Connect

    Currier, R.P.

    1994-10-01

    In the Los Alamos National Laboratory (LANL) hydrothermal process, waste streams are first pressurized and heated as they pass through a continuous flow tubular reactor vessel. The waste is maintained at reaction temperature of 300--550 C where organic destruction and sludge reformation occur. This report documents LANL activities in process modeling and control undertaken in FY94 to support hydrothermal process development. Key issues discussed include non-ideal flow patterns (e.g. axial dispersion) and their effect on reactor performance, the use and interpretation of inert tracer experiments, and the use of computational fluid mechanics to evaluate novel hydrothermal reactor designs. In addition, the effects of axial dispersion (and simplifications to rate expressions) on the estimated kinetic parameters are explored by non-linear regression to experimental data. Safety-related calculations are reported which estimate the explosion limits of effluent gases and the fate of hydrogen as it passes through the reactor. Development and numerical solution of a generalized one-dimensional mathematical model is also summarized. The difficulties encountered in using commercially available software to correlate the behavior of high temperature, high pressure aqueous electrolyte mixtures are summarized. Finally, details of the control system and experiments conducted to empirically determine the system response are reported.

  2. Semiempirical Model Would Control Czochralski Process

    NASA Technical Reports Server (NTRS)

    Dudukovic, M. P.; Ramachandran, P. A.; Srivastava, R. K.

    1989-01-01

    Semiempirical mathematical model proposed for control of growth of single crystals of silicon by Czochralski process. Expresses dependence of pulling rate and shape of liquid/solid interface upon important process variables; radius of growing crystal, temperature of crucible, level of melt, and height of exposed portion of crucible wall. Necessary to control shape of interface in manner consistent with other variables, to maintain radially uniform concentration of dopant, and reduce thermally induced stresses in vicinity of interface. Used to simulate complete growth cycles without requiring excessive computer time consumed by rigorous finite-element modeling.

  3. The DAB model of drawing processes

    NASA Technical Reports Server (NTRS)

    Hochhaus, Larry W.

    1989-01-01

    The problem of automatic drawing was investigated in two ways. First, a DAB model of drawing processes was introduced. DAB stands for three types of knowledge hypothesized to support drawing abilities, namely, Drawing Knowledge, Assimilated Knowledge, and Base Knowledge. Speculation concerning the content and character of each of these subsystems of the drawing process is introduced and the overall adequacy of the model is evaluated. Second, eight experts were each asked to understand six engineering drawings and to think aloud while doing so. It is anticipated that a concurrent protocol analysis of these interviews can be carried out in the future. Meanwhile, a general description of the videotape database is provided. In conclusion, the DAB model was praised as a worthwhile first step toward solution of a difficult problem, but was considered by and large inadequate to the challenge of automatic drawing. Suggestions for improvements on the model were made.

  4. Retort process modelling for Indian traditional foods.

    PubMed

    Gokhale, S V; Lele, S S

    2014-11-01

    Indian traditional staple and snack food is typically a heterogeneous recipe that incorporates varieties of vegetables, lentils and other ingredients. Modelling the retorting process of multilayer pouch packed Indian food was achieved using lumped-parameter approach. A unified model is proposed to estimate cold point temperature. Initial process conditions, retort temperature and % solid content were the significantly affecting independent variables. A model was developed using combination of vegetable solids and water, which was then validated using four traditional Indian vegetarian products: Pulav (steamed rice with vegetables), Sambar (south Indian style curry containing mixed vegetables and lentils), Gajar Halawa (carrot based sweet product) and Upama (wheat based snack product). The predicted and experimental values of temperature profile matched with ±10 % error which is a good match considering the food was a multi component system. Thus the model will be useful as a tool to reduce number of trials required to optimize retorting of various Indian traditional vegetarian foods. PMID:26396305

  5. Lisp and portability: The Process Modeling System

    SciTech Connect

    Egdorf, H.W.

    1992-09-01

    A primary mission of the Technology Modeling and Assessment group (A-7) of the Analysis Division of Los Alamos National Laboratory is to support the Department of Energy in performing analysis of both existing and future facilities that comprise the Nuclear Weapons Complex. Many of the questions to be addressed in relation to this mission involve an examination of the flow of material through a processing facility and the transformations of the material as it moves through the facility by the use of a discrete-event simulation tool. In support of these analysis tasks, a simulation tool kit has been developed that allows examination of issues related to the movement and transformation of material as it moves through a processing facility. This tool kit, The Process Modeling System, is currently the primary modeling tool used for examination of current and future DOE facilities. The flexibility of the system has led to its use in performing similar analysis on a number of nonDOE facilities under Technology Transfer initiatives. The Process Modeling System is written in Common Lisp. The purpose of this paper is to describe the structure of the modeling tool kit and discuss the advantages of Common Lisp as its implementation language.

  6. Lisp and portability: The Process Modeling System

    SciTech Connect

    Egdorf, H.W.

    1992-01-01

    A primary mission of the Technology Modeling and Assessment group (A-7) of the Analysis Division of Los Alamos National Laboratory is to support the Department of Energy in performing analysis of both existing and future facilities that comprise the Nuclear Weapons Complex. Many of the questions to be addressed in relation to this mission involve an examination of the flow of material through a processing facility and the transformations of the material as it moves through the facility by the use of a discrete-event simulation tool. In support of these analysis tasks, a simulation tool kit has been developed that allows examination of issues related to the movement and transformation of material as it moves through a processing facility. This tool kit, The Process Modeling System, is currently the primary modeling tool used for examination of current and future DOE facilities. The flexibility of the system has led to its use in performing similar analysis on a number of nonDOE facilities under Technology Transfer initiatives. The Process Modeling System is written in Common Lisp. The purpose of this paper is to describe the structure of the modeling tool kit and discuss the advantages of Common Lisp as its implementation language.

  7. Attrition and abrasion models for oil shale process modeling

    SciTech Connect

    Aldis, D.F.

    1991-10-25

    As oil shale is processed, fine particles, much smaller than the original shale are created. This process is called attrition or more accurately abrasion. In this paper, models of abrasion are presented for oil shale being processed in several unit operations. Two of these unit operations, a fluidized bed and a lift pipe are used in the Lawrence Livermore National Laboratory Hot-Recycle-Solid (HRS) process being developed for the above ground processing of oil shale. In two reports, studies were conducted on the attrition of oil shale in unit operations which are used in the HRS process. Carley reported results for attrition in a lift pipe for oil shale which had been pre-processed either by retorting or by retorting then burning. The second paper, by Taylor and Beavers, reported results for a fluidized bed processing of oil shale. Taylor and Beavers studied raw, retorted, and shale which had been retorted and then burned. In this paper, empirical models are derived, from the experimental studies conducted on oil shale for the process occurring in the HRS process. The derived models are presented along with comparisons with experimental results.

  8. SDN-Enabled Dynamic Feedback Control and Sensing in Agile Optical Networks

    NASA Astrophysics Data System (ADS)

    Lin, Likun

    Fiber optic networks are no longer just pipelines for transporting data in the long haul backbone. Exponential growth in traffic in metro-regional areas has pushed higher capacity fiber toward the edge of the network, and highly dynamic patterns of heterogeneous traffic have emerged that are often bursty, severely stressing the historical "fat and dumb pipe" static optical network, which would need to be massively over-provisioned to deal with these loads. What is required is a more intelligent network with a span of control over the optical as well as electrical transport mechanisms which enables handling of service requests in a fast and efficient way that guarantees quality of service (QoS) while optimizing capacity efficiency. An "agile" optical network is a reconfigurable optical network comprised of high speed intelligent control system fed by real-time in situ network sensing. It provides fast response in the control and switching of optical signals in response to changing traffic demands and network conditions. This agile control of optical signals is enabled by pushing switching decisions downward in the network stack to the physical layer. Implementing such agility is challenging due to the response dynamics and interactions of signals in the physical layer. Control schemes must deal with issues such as dynamic power equalization, EDFA transients and cascaded noise effects, impairments due to self-phase modulation and dispersion, and channel-to-channel cross talk. If these issues are not properly predicted and mitigated, attempts at dynamic control can drive the optical network into an unstable state. In order to enable high speed actuation of signal modulators and switches, the network controller must be able to make decisions based on predictive models. In this thesis, we consider how to take advantage of Software Defined Networking (SDN) capabilities for network reconfiguration, combined with embedded models that access updates from deployed network

  9. Modeling stroke rehabilitation processes using the Unified Modeling Language (UML).

    PubMed

    Ferrante, Simona; Bonacina, Stefano; Pinciroli, Francesco

    2013-10-01

    In organising and providing rehabilitation procedures for stroke patients, the usual need for many refinements makes it inappropriate to attempt rigid standardisation, but greater detail is required concerning workflow. The aim of this study was to build a model of the post-stroke rehabilitation process. The model, implemented in the Unified Modeling Language, was grounded on international guidelines and refined following the clinical pathway adopted at local level by a specialized rehabilitation centre. The model describes the organisation of the rehabilitation delivery and it facilitates the monitoring of recovery during the process. Indeed, a system software was developed and tested to support clinicians in the digital administration of clinical scales. The model flexibility assures easy updating after process evolution.

  10. Modeling of a thermoplastic pultrusion process

    SciTech Connect

    Astroem, B.T. ); Pipes, R.B. )

    1991-07-01

    To obtain a fundamental understanding of the effects of processing parameters and die geometry in a pultrusion process, a mathematical model is essential in order to minimize the number of trial-and-error experiments. Previous investigators have suggested a variety of more or less complete models for thermoset pultrusion, while little effort seems to have been spent modeling its less well-understood thermoplastic equivalent. Hence, a set of intricately related models to describe the temperature and pressure distributions, as well as the matrix flow, in a thermoplastic composite as it travels through the pultrusion die is presented. An approach to calculate the accumulated pulling force is also explored, and the individual mechanisms contributing to the pulling force are discussed. The pressure model incorporates a matrix viscosity that varies with shear rate, temperature, and pressure. Comparisons are made between shear-rate-dependent and Newtonian viscosity representations, indicating the necessity of including non-Newtonian fluid behavior when modeling thermoplastic pultrusion. The governing equations of the models are stated in general terms, and simplifications are implemented in order to obtain solutions without extensive numerical efforts. Pressure, temperature, cooling rate, and pulling force distributions are presented for carbon-fiber-reinforced polyetheretherketone. Pulling force predictions are compared to data obtained from preliminary experiments conducted with a model pultrusion line that was built solely for the pultrusion of thermoplastic matrix composites, and the correlation is found to be qualitatively satisfactory.

  11. Therapeutic Process During Exposure: Habituation Model

    PubMed Central

    Benito, Kristen G.; Walther, Michael

    2015-01-01

    The current paper outlines the habituation model of exposure process, which is a behavioral model emphasizing use of individually tailored functional analysis during exposures. This is a model of therapeutic process rather than one meant to explain the mechanism of change underlying exposure-based treatments. Habitation, or a natural decrease in anxiety level in the absence of anxiety-reducing behavior, might be best understood as an intermediate treatment outcome that informs therapeutic process, rather than as a mechanism of change. The habituation model purports that three conditions are necessary for optimal benefit from exposures: 1) fear activation, 2) minimization of anxiety-reducing behaviors, and 3) habituation. We describe prescribed therapist and client behaviors as those that increase or maintain anxiety level during an exposure (and therefore, facilitate habituation), and proscribed therapist and client behaviors as those that decrease anxiety during an exposure (and therefore, impede habituation). We illustrate model-consistent behaviors in the case of Monica, as well as outline the existing research support and call for additional research to further test the tenets of the habituation model as described in this paper. PMID:26258012

  12. Processing and Modeling of Porous Copper Using Sintering Dissolution Process

    NASA Astrophysics Data System (ADS)

    Salih, Mustafa Abualgasim Abdalhakam

    The growth of porous metal has produced materials with improved properties as compared to non-metals and solid metals. Porous metal can be classified as either open cell or closed cell. Open cell allows a fluid media to pass through it. Closed cell is made up of adjacent sealed pores with shared cell walls. Metal foams offer higher strength to weight ratios, increased impact energy absorption, and a greater tolerance to high temperatures and adverse environmental conditions when compared to bulk materials. Copper and its alloys are examples of these, well known for high strength and good mechanical, thermal and electrical properties. In the present study, the porous Cu was made by a powder metallurgy process, using three different space holders, sodium chloride, sodium carbonate and potassium carbonate. Several different samples have been produced, using different ratios of volume fraction. The densities of the porous metals have been measured and compared to the theoretical density calculated using an equation developed for these foams. The porous structure was determined with the removal of spacer materials through sintering process. The sintering process of each spacer material depends on the melting point of the spacer material. Processing, characterization, and mechanical properties were completed. These tests include density measurements, compression tests, computed tomography (CT) and scanning electron microscopy (SEM). The captured morphological images are utilized to generate the object-oriented finite element (OOF) analysis for the porous copper. Porous copper was formed with porosities in the range of 40-66% with density ranges from 3 to 5.2 g/cm3. A study of two different methods to measure porosity was completed. OOF (Object Oriented Finite Elements) is a desktop software application for studying the relationship between the microstructure of a material and its overall mechanical, dielectric, or thermal properties using finite element models based on

  13. Mesoscopic Modeling of Reactive Transport Processes

    NASA Astrophysics Data System (ADS)

    Kang, Q.; Chen, L.; Deng, H.

    2012-12-01

    Reactive transport processes involving precipitation and/or dissolution are pervasive in geochemical, biological and engineered systems. Typical examples include self-assembled patterns such as Liesegang rings or bands, cones of stalactites in limestones caves, biofilm growth in aqueous environment, formation of mineral deposits in boilers and heat exchangers, uptake of toxic metal ions from polluted water by calcium carbonate, and mineral trapping of CO2. Compared to experimental studies, a numerical approach enables a systematic study of the reaction kinetics, mass transport, and mechanisms of nucleation and crystal growth, and hence provides a detailed description of reactive transport processes. In this study, we enhance a previously developed lattice Boltzmann pore-scale model by taking into account the nucleation process, and develop a mesoscopic approach to simulate reactive transport processes involving precipitation and/or dissolution of solid phases. The model is then used to simulate the formation of Liesegang precipitation patterns and investigate the effects of gel on the morphology of the precipitates. It is shown that this model can capture the porous structures of the precipitates and can account for the effects of the gel concentration and material. A wide range of precipitation patterns is predicted under different gel concentrations, including regular bands, treelike patterns, and for the first time with numerical models, transition patterns from regular bands to treelike patterns. The model is also applied to study the effect of secondary precipitate on the dissolution of primary mineral. Several types of dissolution and precipitation processes are identified based on the morphology and structures of the precipitates and on the extent to which the precipitates affect the dissolution of the primary mineral. Finally the model is applied to study the formation of pseudomorph. It is demonstrated for the first time by numerical simulation that a

  14. A Generic Modeling Process to Support Functional Fault Model Development

    NASA Technical Reports Server (NTRS)

    Maul, William A.; Hemminger, Joseph A.; Oostdyk, Rebecca; Bis, Rachael A.

    2016-01-01

    Functional fault models (FFMs) are qualitative representations of a system's failure space that are used to provide a diagnostic of the modeled system. An FFM simulates the failure effect propagation paths within a system between failure modes and observation points. These models contain a significant amount of information about the system including the design, operation and off nominal behavior. The development and verification of the models can be costly in both time and resources. In addition, models depicting similar components can be distinct, both in appearance and function, when created individually, because there are numerous ways of representing the failure space within each component. Generic application of FFMs has the advantages of software code reuse: reduction of time and resources in both development and verification, and a standard set of component models from which future system models can be generated with common appearance and diagnostic performance. This paper outlines the motivation to develop a generic modeling process for FFMs at the component level and the effort to implement that process through modeling conventions and a software tool. The implementation of this generic modeling process within a fault isolation demonstration for NASA's Advanced Ground System Maintenance (AGSM) Integrated Health Management (IHM) project is presented and the impact discussed.

  15. Improving the process of process modelling by the use of domain process patterns

    NASA Astrophysics Data System (ADS)

    Koschmider, Agnes; Reijers, Hajo A.

    2015-01-01

    The use of business process models has become prevalent in a wide area of enterprise applications. But while their popularity is expanding, concerns are growing with respect to their proper creation and maintenance. An obvious way to boost the efficiency of creating high-quality business process models would be to reuse relevant parts of existing models. At this point, however, limited support exists to guide process modellers towards the usage of appropriate model content. In this paper, a set of content-oriented patterns is presented, which is extracted from a large set of process models from the order management and manufacturing production domains. The patterns are derived using a newly proposed set of algorithms, which are being discussed in this paper. The authors demonstrate how such Domain Process Patterns, in combination with information on their historic usage, can support process modellers in generating new models. To support the wider dissemination and development of Domain Process Patterns within and beyond the studied domains, an accompanying website has been set up.

  16. Dynamical modeling of laser ablation processes

    SciTech Connect

    Leboeuf, J.N.; Chen, K.R.; Donato, J.M.; Geohegan, D.B.; Liu, C.L.; Puretzky, A.A.; Wood, R.F.

    1995-09-01

    Several physics and computational approaches have been developed to globally characterize phenomena important for film growth by pulsed laser deposition of materials. These include thermal models of laser-solid target interactions that initiate the vapor plume; plume ionization and heating through laser absorption beyond local thermodynamic equilibrium mechanisms; gas dynamic, hydrodynamic, and collisional descriptions of plume transport; and molecular dynamics models of the interaction of plume particles with the deposition substrate. The complexity of the phenomena involved in the laser ablation process is matched by the diversity of the modeling task, which combines materials science, atomic physics, and plasma physics.

  17. Hencky's model for elastomer forming process

    NASA Astrophysics Data System (ADS)

    Oleinikov, A. A.; Oleinikov, A. I.

    2016-08-01

    In the numerical simulation of elastomer forming process, Henckys isotropic hyperelastic material model can guarantee relatively accurate prediction of strain range in terms of large deformations. It is shown, that this material model prolongate Hooke's law from the area of infinitesimal strains to the area of moderate ones. New representation of the fourth-order elasticity tensor for Hencky's hyperelastic isotropic material is obtained, it possesses both minor symmetries, and the major symmetry. Constitutive relations of considered model is implemented into MSC.Marc code. By calculating and fitting curves, the polyurethane elastomer material constants are selected. Simulation of equipment for elastomer sheet forming are considered.

  18. Model-based internal wave processing

    SciTech Connect

    Candy, J.V.; Chambers, D.H.

    1995-06-09

    A model-based approach is proposed to solve the oceanic internal wave signal processing problem that is based on state-space representations of the normal-mode vertical velocity and plane wave horizontal velocity propagation models. It is shown that these representations can be utilized to spatially propagate the modal (dept) vertical velocity functions given the basic parameters (wave numbers, Brunt-Vaisala frequency profile etc.) developed from the solution of the associated boundary value problem as well as the horizontal velocity components. Based on this framework, investigations are made of model-based solutions to the signal enhancement problem for internal waves.

  19. Applying Business Process Re-engineering Patterns to optimize WS-BPEL Workflows

    NASA Astrophysics Data System (ADS)

    Buys, Jonas; de Florio, Vincenzo; Blondia, Chris

    With the advent of XML-based SOA, WS-BPEL shortly turned out to become a widely accepted standard for modeling business processes. Though SOA is said to embrace the principle of business agility, BPEL process definitions are still manually crafted into their final executable version. While SOA has proven to be a giant leap forward in building flexible IT systems, this static BPEL workflow model is somewhat paradoxical to the need for real business agility and should be enhanced to better sustain continual process evolution. In this paper, we point out the potential of adding business intelligence with respect to business process re-engineering patterns to the system to allow for automatic business process optimization. Furthermore, we point out that BPR macro-rules could be implemented leveraging micro-techniques from computer science. We present some practical examples that illustrate the benefit of such adaptive process models and our preliminary findings.

  20. Hot blast stove process model and model-based controller

    SciTech Connect

    Muske, K.R.; Howse, J.W.; Hansen, G.A.; Cagliostro, D.J.; Chaubal, P.C.

    1998-12-31

    This paper describes the process model and model-based control techniques implemented on the hot blast stoves for the No. 7 Blast Furnace at the Inland Steel facility in East Chicago, Indiana. A detailed heat transfer model of the stoves is developed and verified using plant data. This model is used as part of a predictive control scheme to determine the minimum amount of fuel necessary to achieve the blast air requirements. The model is also used to predict maximum and minimum temperature constraint violations within the stove so that the controller can take corrective actions while still achieving the required stove performance.

  1. Aligning Grammatical Theories and Language Processing Models

    ERIC Educational Resources Information Center

    Lewis, Shevaun; Phillips, Colin

    2015-01-01

    We address two important questions about the relationship between theoretical linguistics and psycholinguistics. First, do grammatical theories and language processing models describe separate cognitive systems, or are they accounts of different aspects of the same system? We argue that most evidence is consistent with the one-system view. Second,…

  2. The SERIOL2 Model of Orthographic Processing

    ERIC Educational Resources Information Center

    Whitney, Carol; Marton, Yuval

    2013-01-01

    The SERIOL model of orthographic analysis proposed mechanisms for converting visual input into a serial encoding of letter order, which involved hemisphere-specific processing at the retinotopic level. As a test of SERIOL predictions, we conducted a consonant trigram-identification experiment, where the trigrams were briefly presented at various…

  3. Model Identification of Integrated ARMA Processes

    ERIC Educational Resources Information Center

    Stadnytska, Tetiana; Braun, Simone; Werner, Joachim

    2008-01-01

    This article evaluates the Smallest Canonical Correlation Method (SCAN) and the Extended Sample Autocorrelation Function (ESACF), automated methods for the Autoregressive Integrated Moving-Average (ARIMA) model selection commonly available in current versions of SAS for Windows, as identification tools for integrated processes. SCAN and ESACF can…

  4. Content, Process, and Product: Modeling Differentiated Instruction

    ERIC Educational Resources Information Center

    Taylor, Barbara Kline

    2015-01-01

    Modeling differentiated instruction is one way to demonstrate how educators can incorporate instructional strategies to address students' needs, interests, and learning styles. This article discusses how secondary teacher candidates learn to focus on content--the "what" of instruction; process--the "how" of instruction;…

  5. Mathematical Modelling of Continuous Biotechnological Processes

    ERIC Educational Resources Information Center

    Pencheva, T.; Hristozov, I.; Shannon, A. G.

    2003-01-01

    Biotechnological processes (BTP) are characterized by a complicated structure of organization and interdependent characteristics. Partial differential equations or systems of partial differential equations are used for their behavioural description as objects with distributed parameters. Modelling of substrate without regard to dispersion…

  6. Computational Process Modeling for Additive Manufacturing (OSU)

    NASA Technical Reports Server (NTRS)

    Bagg, Stacey; Zhang, Wei

    2015-01-01

    Powder-Bed Additive Manufacturing (AM) through Direct Metal Laser Sintering (DMLS) or Selective Laser Melting (SLM) is being used by NASA and the Aerospace industry to "print" parts that traditionally are very complex, high cost, or long schedule lead items. The process spreads a thin layer of metal powder over a build platform, then melts the powder in a series of welds in a desired shape. The next layer of powder is applied, and the process is repeated until layer-by-layer, a very complex part can be built. This reduces cost and schedule by eliminating very complex tooling and processes traditionally used in aerospace component manufacturing. To use the process to print end-use items, NASA seeks to understand SLM material well enough to develop a method of qualifying parts for space flight operation. Traditionally, a new material process takes many years and high investment to generate statistical databases and experiential knowledge, but computational modeling can truncate the schedule and cost -many experiments can be run quickly in a model, which would take years and a high material cost to run empirically. This project seeks to optimize material build parameters with reduced time and cost through modeling.

  7. Modeling Low-temperature Geochemical Processes

    NASA Astrophysics Data System (ADS)

    Nordstrom, D. K.

    2003-12-01

    Geochemical modeling has become a popular and useful tool for a wide number of applications from research on the fundamental processes of water-rock interactions to regulatory requirements and decisions regarding permits for industrial and hazardous wastes. In low-temperature environments, generally thought of as those in the temperature range of 0-100 °C and close to atmospheric pressure (1 atm=1.01325 bar=101,325 Pa), complex hydrobiogeochemical reactions participate in an array of interconnected processes that affect us, and that, in turn, we affect. Understanding these complex processes often requires tools that are sufficiently sophisticated to portray multicomponent, multiphase chemical reactions yet transparent enough to reveal the main driving forces. Geochemical models are such tools. The major processes that they are required to model include mineral dissolution and precipitation; aqueous inorganic speciation and complexation; solute adsorption and desorption; ion exchange; oxidation-reduction; or redox; transformations; gas uptake or production; organic matter speciation and complexation; evaporation; dilution; water mixing; reaction during fluid flow; reaction involving biotic interactions; and photoreaction. These processes occur in rain, snow, fog, dry atmosphere, soils, bedrock weathering, streams, rivers, lakes, groundwaters, estuaries, brines, and diagenetic environments. Geochemical modeling attempts to understand the redistribution of elements and compounds, through anthropogenic and natural means, for a large range of scale from nanometer to global. "Aqueous geochemistry" and "environmental geochemistry" are often used interchangeably with "low-temperature geochemistry" to emphasize hydrologic or environmental objectives.Recognition of the strategy or philosophy behind the use of geochemical modeling is not often discussed or explicitly described. Plummer (1984, 1992) and Parkhurst and Plummer (1993) compare and contrast two approaches for

  8. A model evaluation checklist for process-based environmental models

    NASA Astrophysics Data System (ADS)

    Jackson-Blake, Leah

    2015-04-01

    Mechanistic catchment-scale phosphorus models appear to perform poorly where diffuse sources dominate. The reasons for this were investigated for one commonly-applied model, the INtegrated model of CAtchment Phosphorus (INCA-P). Model output was compared to 18 months of daily water quality monitoring data in a small agricultural catchment in Scotland, and model structure, key model processes and internal model responses were examined. Although the model broadly reproduced dissolved phosphorus dynamics, it struggled with particulates. The reasons for poor performance were explored, together with ways in which improvements could be made. The process of critiquing and assessing model performance was then generalised to provide a broadly-applicable model evaluation checklist, incorporating: (1) Calibration challenges, relating to difficulties in thoroughly searching a high-dimensional parameter space and in selecting appropriate means of evaluating model performance. In this study, for example, model simplification was identified as a necessary improvement to reduce the number of parameters requiring calibration, whilst the traditionally-used Nash Sutcliffe model performance statistic was not able to discriminate between realistic and unrealistic model simulations, and alternative statistics were needed. (2) Data limitations, relating to a lack of (or uncertainty in) input data, data to constrain model parameters, data for model calibration and testing, and data to test internal model processes. In this study, model reliability could be improved by addressing all four kinds of data limitation. For example, there was insufficient surface water monitoring data for model testing against an independent dataset to that used in calibration, whilst additional monitoring of groundwater and effluent phosphorus inputs would help distinguish between alternative plausible model parameterisations. (3) Model structural inadequacies, whereby model structure may inadequately represent

  9. Atomic Layer Deposition - Process Models and Metrologies

    SciTech Connect

    Burgess, D.R. Jr.; Maslar, J.E.; Hurst, W.S.; Moore, E.F.; Kimes, W.A.; Fink, R.R.; Nguyen, N.V.

    2005-09-09

    We report on the status of a combined experimental and modeling study for atomic layer deposition (ALD) of HfO2 and Al2O3. Hafnium oxide films were deposited from tetrakis(dimethylamino)hafnium and water. Aluminum oxide films from trimethyl aluminum and water are being studied through simulations. In this work, both in situ metrologies and process models are being developed. Optically-accessible ALD reactors have been constructed for in situ, high-sensitivity Raman and infrared absorption spectroscopic measurements to monitor gas phase and surface species. A numerical model using computational fluid dynamics codes has been developed to simulate the gas flow and temperature profiles in the experimental reactor. Detailed chemical kinetic models are being developed with assistance from quantum chemical calculations to explore reaction pathways and energetics. This chemistry is then incorporated into the overall reactor models.

  10. Coal-to-Liquids Process Model

    SciTech Connect

    2006-01-01

    A comprehensive Aspen Plus model has been developed to rigorously model coal-to-liquids processes. This portion was developed under Laboratory Directed Research and Development (LDRD) funding. The model is built in a modular fashion to allow rapid reconfiguration for evaluation of process options. Aspen Plus is the framework in which the model is developed. The coal-to-liquids simulation package is an assemble of Aspen Hierarchy Blocks representing subsections of the plant. Each of these Blocks are considered individual components of the Copyright, which may be extracted and licensed as individual components, but which may be combined with one or more other components, to model general coal-conversion processes, including the following plant operations: (1) coal handling and preparation, (2) coal pyrolysis, combustion, or gasification, (3) syngas conditioning and cleanup, (4) sulfur recovery using Claus-SCOT unit operations, (5) Fischer-Tropsch liquid fuels synthesis, (6) hydrocracking of high molecular weight paraffin, (7) hydrotreating of low molecular weight paraffin and olefins, (8) gas separations, and (9) power generation representing integrated combined cycle technology.

  11. Coal-to-Liquids Process Model

    2006-01-01

    A comprehensive Aspen Plus model has been developed to rigorously model coal-to-liquids processes. This portion was developed under Laboratory Directed Research and Development (LDRD) funding. The model is built in a modular fashion to allow rapid reconfiguration for evaluation of process options. Aspen Plus is the framework in which the model is developed. The coal-to-liquids simulation package is an assemble of Aspen Hierarchy Blocks representing subsections of the plant. Each of these Blocks are consideredmore » individual components of the Copyright, which may be extracted and licensed as individual components, but which may be combined with one or more other components, to model general coal-conversion processes, including the following plant operations: (1) coal handling and preparation, (2) coal pyrolysis, combustion, or gasification, (3) syngas conditioning and cleanup, (4) sulfur recovery using Claus-SCOT unit operations, (5) Fischer-Tropsch liquid fuels synthesis, (6) hydrocracking of high molecular weight paraffin, (7) hydrotreating of low molecular weight paraffin and olefins, (8) gas separations, and (9) power generation representing integrated combined cycle technology.« less

  12. Process modeling with the regression network.

    PubMed

    van der Walt, T; Barnard, E; van Deventer, J

    1995-01-01

    A new connectionist network topology called the regression network is proposed. The structural and underlying mathematical features of the regression network are investigated. Emphasis is placed on the intricacies of the optimization process for the regression network and some measures to alleviate these difficulties of optimization are proposed and investigated. The ability of the regression network algorithm to perform either nonparametric or parametric optimization, as well as a combination of both, is also highlighted. It is further shown how the regression network can be used to model systems which are poorly understood on the basis of sparse data. A semi-empirical regression network model is developed for a metallurgical processing operation (a hydrocyclone classifier) by building mechanistic knowledge into the connectionist structure of the regression network model. Poorly understood aspects of the process are provided for by use of nonparametric regions within the structure of the semi-empirical connectionist model. The performance of the regression network model is compared to the corresponding generalization performance results obtained by some other nonparametric regression techniques.

  13. Wavelength-Agile External-Cavity Diode Laser for DWDM

    NASA Technical Reports Server (NTRS)

    Pilgrim, Jeffrey S.; Bomse, David S.

    2006-01-01

    A prototype external-cavity diode laser (ECDL) has been developed for communication systems utilizing dense wavelength- division multiplexing (DWDM). This ECDL is an updated version of the ECDL reported in Wavelength-Agile External- Cavity Diode Laser (LEW-17090), NASA Tech Briefs, Vol. 25, No. 11 (November 2001), page 14a. To recapitulate: The wavelength-agile ECDL combines the stability of an external-cavity laser with the wavelength agility of a diode laser. Wavelength is modulated by modulating the injection current of the diode-laser gain element. The external cavity is a Littman-Metcalf resonator, in which the zeroth-order output from a diffraction grating is used as the laser output and the first-order-diffracted light is retro-reflected by a cavity feedback mirror, which establishes one end of the resonator. The other end of the resonator is the output surface of a Fabry-Perot resonator that constitutes the diode-laser gain element. Wavelength is selected by choosing the angle of the diffracted return beam, as determined by position of the feedback mirror. The present wavelength-agile ECDL is distinguished by design details that enable coverage of all 60 channels, separated by 100-GHz frequency intervals, that are specified in DWDM standards.

  14. Agile Software Development Methods: A Comparative Review1

    NASA Astrophysics Data System (ADS)

    Abrahamsson, Pekka; Oza, Nilay; Siponen, Mikko T.

    Although agile software development methods have caught the attention of software engineers and researchers worldwide, scientific research still remains quite scarce. The aim of this study is to order and make sense of the different agile approaches that have been proposed. This comparative review is performed from the standpoint of using the following features as the analytical perspectives: project management support, life-cycle coverage, type of practical guidance, adaptability in actual use, type of research objectives and existence of empirical evidence. The results show that agile software development methods cover, without offering any rationale, different phases of the software development life-cycle and that most of these methods fail to provide adequate project management support. Moreover, quite a few methods continue to offer little concrete guidance on how to use their solutions or how to adapt them in different development situations. Empirical evidence after ten years of application remains quite limited. Based on the results, new directions on agile methods are outlined.

  15. A Capstone Course on Agile Software Development Using Scrum

    ERIC Educational Resources Information Center

    Mahnic, V.

    2012-01-01

    In this paper, an undergraduate capstone course in software engineering is described that not only exposes students to agile software development, but also makes it possible to observe the behavior of developers using Scrum for the first time. The course requires students to work as Scrum Teams, responsible for the implementation of a set of user…

  16. [Cellular model of blood coagulation process].

    PubMed

    Bijak, Michał; Rzeźnicka, Paulina; Saluk, Joanna; Nowak, Paweł

    2015-07-01

    Blood coagulation is a process which main objective is the prevention of blood loss when the integrity of the blood vessel is damaged. Over the years, have been presented a number of concepts characterizing the mechanism of thrombus formation. Since the 60s of last century was current cascade model of the coagulation wherein forming of the fibrin clot is determined by two pathways called extrinsic and intrinsic pathways. In the nineties of the last century Monroe and Hoffman presented his concept of blood coagulation process which complement the currently valid model of cells participation especially of blood platelets which aim is to provide a negatively charged phospholipid surface and thereby allow the coagulation enzymatic complexes formation. Developed conception they called cellular model of coagulation. The aim of this work was to present in details of this blood coagulation, including descriptions of its various phases.

  17. Modeling veterans healthcare administration disclosure processes :

    SciTech Connect

    Beyeler, Walter E; DeMenno, Mercy B.; Finley, Patrick D.

    2013-09-01

    As with other large healthcare organizations, medical adverse events at the Department of Veterans Affairs (VA) facilities can expose patients to unforeseen negative risks. VHA leadership recognizes that properly handled disclosure of adverse events can minimize potential harm to patients and negative consequences for the effective functioning of the organization. The work documented here seeks to help improve the disclosure process by situating it within the broader theoretical framework of issues management, and to identify opportunities for process improvement through modeling disclosure and reactions to disclosure. The computational model will allow a variety of disclosure actions to be tested across a range of incident scenarios. Our conceptual model will be refined in collaboration with domain experts, especially by continuing to draw on insights from VA Study of the Communication of Adverse Large-Scale Events (SCALE) project researchers.

  18. Modeling delayed processes in biological systems

    NASA Astrophysics Data System (ADS)

    Feng, Jingchen; Sevier, Stuart A.; Huang, Bin; Jia, Dongya; Levine, Herbert

    2016-09-01

    Delayed processes are ubiquitous in biological systems and are often characterized by delay differential equations (DDEs) and their extension to include stochastic effects. DDEs do not explicitly incorporate intermediate states associated with a delayed process but instead use an estimated average delay time. In an effort to examine the validity of this approach, we study systems with significant delays by explicitly incorporating intermediate steps. We show that such explicit models often yield significantly different equilibrium distributions and transition times as compared to DDEs with deterministic delay values. Additionally, different explicit models with qualitatively different dynamics can give rise to the same DDEs revealing important ambiguities. We also show that DDE-based predictions of oscillatory behavior may fail for the corresponding explicit model.

  19. An ecological process model of systems change.

    PubMed

    Peirson, Leslea J; Boydell, Katherine M; Ferguson, H Bruce; Ferris, Lorraine E

    2011-06-01

    In June 2007 the American Journal of Community Psychology published a special issue focused on theories, methods and interventions for systems change which included calls from the editors and authors for theoretical advancement in this field. We propose a conceptual model of systems change that integrates familiar and fundamental community psychology principles (succession, interdependence, cycling of resources, adaptation) and accentuates a process orientation. To situate our framework we offer a definition of systems change and a brief review of the ecological perspective and principles. The Ecological Process Model of Systems Change is depicted, described and applied to a case example of policy driven systems level change in publicly funded social programs. We conclude by identifying salient implications for thinking and action which flow from the Model. PMID:21203829

  20. Impact of flow unsteadiness on maneuvers and loads of agile aircraft

    NASA Technical Reports Server (NTRS)

    Jarrah, M. Ameen; Ashley, Holt

    1989-01-01

    A program of airload measurements on a family of low-aspect-ratio delta wings with sharp leading edges, subjected to large amplitude pitch transients with angles of attack up to 90 deg, is reviewed. Even for small values of the pitch-rate parameter, representative of maneuvers anticipated for agile aircraft, the force and moment overshoots can exceed by 50 percent their steady-state values. This is explained in terms of the hysteretic behavior of the breakdown locations of leading-edge vortices. An approximate theoretical model is proposed which includes the breakdown hysteresis as part of a three-term representation of the unsteady chordwise load distribution.

  1. Development and evaluation of an inverse solution technique for studying helicopter maneuverability and agility

    NASA Technical Reports Server (NTRS)

    Whalley, Matthew S.

    1991-01-01

    An inverse solution technique for determining the maximum maneuvering performance of a helicopter using smooth, pilotlike control inputs is presented. Also described is a pilot simulation experiment performed to investigate the accuracy of the solution resulting from this technique. The maneuverability and agility capability of the helicopter math model was varied by varying the pitch and roll damping, the maximum pitch and roll rate, and the maximum load-factor capability. Three maneuvers were investigated: a 180-deg turn, a longitudinal pop-up, and a lateral jink. The inverse solution technique yielded accurate predictions of pilot-in-the-loop maneuvering performance for two of the three maneuvers.

  2. Agile hardware and software systems engineering for critical military space applications

    NASA Astrophysics Data System (ADS)

    Huang, Philip M.; Knuth, Andrew A.; Krueger, Robert O.; Garrison-Darrin, Margaret A.

    2012-06-01

    The Multi Mission Bus Demonstrator (MBD) is a successful demonstration of agile program management and system engineering in a high risk technology application where utilizing and implementing new, untraditional development strategies were necessary. MBD produced two fully functioning spacecraft for a military/DOD application in a record breaking time frame and at dramatically reduced costs. This paper discloses the adaptation and application of concepts developed in agile software engineering to hardware product and system development for critical military applications. This challenging spacecraft did not use existing key technology (heritage hardware) and created a large paradigm shift from traditional spacecraft development. The insertion of new technologies and methods in space hardware has long been a problem due to long build times, the desire to use heritage hardware, and lack of effective process. The role of momentum in the innovative process can be exploited to tackle ongoing technology disruptions and allowing risk interactions to be mitigated in a disciplined manner. Examples of how these concepts were used during the MBD program will be delineated. Maintaining project momentum was essential to assess the constant non recurring technological challenges which needed to be retired rapidly from the engineering risk liens. Development never slowed due to tactical assessment of the hardware with the adoption of the SCRUM technique. We adapted this concept as a representation of mitigation of technical risk while allowing for design freeze later in the program's development cycle. By using Agile Systems Engineering and Management techniques which enabled decisive action, the product development momentum effectively was used to produce two novel space vehicles in a fraction of time with dramatically reduced cost.

  3. Sperm transport and storage in the agile antechinus (Antechinus agilis).

    PubMed

    Shimmin, G A; Jones, M; Taggart, D A; Temple-Smith, P D

    1999-06-01

    This study was an examination of the timing of ejaculation and the dynamics of sperm transport in the female reproductive tract of the agile Antechinus (Antechinus agilis) and the relationship of these parameters to single and multiple matings. Mating in this species is characteristically long compared with that of other mammals, lasting for up to 8-12 h during which time the pair remain locked together. After the first hour of mating, only approximately 40% of males had ejaculated, but by the third hour all males had ejaculated. The total number of spermatozoa extracted from the female tract remained at approximately 30 x 10(3) spermatozoa per side over the next 9 h of copulation. After completion of male/female access (12 h), approximately 56% of spermatozoa extracted were located in the lower isthmic region of the oviduct where specialized sperm storage crypts are located. The number of spermatozoa extracted from the female reproductive tract did not decline over the next 3 days, but there was a change in the distribution of spermatozoa with an increase in the proportion of extracted spermatozoa stored in the lower isthmus (approximately 76%). However, 7 to 14 days after mating, only approximately 30% of the stored spermatozoa ( approximately 9.4 x 10(3) spermatozoa per side) were still present in the isthmus. When females were mated with a second male on a consecutive day, the sperm numbers extracted from the tract were about twice that deposited during single mating, with sperm transport to the lower isthmus occurring over a similar time frame. Although the occurrence of extended copulations in the wild has not yet been confirmed, these laboratory results suggest that similar periods of copulation are likely, since completion of the ejaculation process requires at least 3 h. The extended copulation in A. agilis reduces the possibility of an early second mating, which might interfere with the normal transport and crypt colonization of spermatozoa through

  4. Modelling Of Manufacturing Processes With Membranes

    NASA Astrophysics Data System (ADS)

    Crăciunean, Daniel Cristian; Crăciunean, Vasile

    2015-07-01

    The current objectives to increase the standards of quality and efficiency in manufacturing processes can be achieved only through the best combination of inputs, independent of spatial distance between them. This paper proposes modelling production processes based on membrane structures introduced in [4]. Inspired from biochemistry, membrane computation [4] is based on the concept of membrane represented in its formalism by the mathematical concept of multiset. The manufacturing process is the evolution of a super cell system from its initial state according to the given actions of aggregation. In this paper we consider that the atomic production unit of the process is the action. The actions and the resources on which the actions are produced, are distributed in a virtual network of companies working together. The destination of the output resources is specified by corresponding output events.

  5. Development of a comprehensive weld process model

    SciTech Connect

    Radhakrishnan, B.; Zacharia, T.; Paul, A.

    1997-05-01

    This cooperative research and development agreement (CRADA) between Concurrent Technologies Corporation (CTC) and Lockheed Martin Energy Systems (LMES) combines CTC`s expertise in the welding area and that of LMES to develop computer models and simulation software for welding processes. This development is of significant impact to the industry, including materials producers and fabricators. The main thrust of the research effort was to develop a comprehensive welding simulation methodology. A substantial amount of work has been done by several researchers to numerically model several welding processes. The primary drawback of most of the existing models is the lack of sound linkages between the mechanistic aspects (e.g., heat transfer, fluid flow, and residual stress) and the metallurgical aspects (e.g., microstructure development and control). A comprehensive numerical model which can be used to elucidate the effect of welding parameters/conditions on the temperature distribution, weld pool shape and size, solidification behavior, and microstructure development, as well as stresses and distortion, does not exist. It was therefore imperative to develop a comprehensive model which would predict all of the above phenomena during welding. The CRADA built upon an already existing three-dimensional (3-D) welding simulation model which was developed by LMES which is capable of predicting weld pool shape and the temperature history in 3-d single-pass welds. However, the model does not account for multipass welds, microstructural evolution, distortion and residual stresses. Additionally, the model requires large resources of computing time, which limits its use for practical applications. To overcome this, CTC and LMES have developed through this CRADA the comprehensive welding simulation model described above.

  6. The effect of acute stretching on agility performance.

    PubMed

    Van Gelder, Leonard H; Bartz, Shari D

    2011-11-01

    Static stretching (SS) has shown decreases in many areas including strength, anaerobic power, and sprinting time. Dynamic stretching (DS) has shown increases in anaerobic power and decreases in sprinting time. Research on the effects of stretching on agility performance is limited. The purpose of this study was to determine the effect of SS and DS on performance time of a sport agility test. Sixty male subjects consisting of collegiate (n = 18) and recreational (n = 42) basketball athletes volunteered for the study. Subjects were randomly assigned to 1 of 3 intervention groups: SS, DS, or no stretching (NS). All groups completed a 10-minute warm-up jog followed by a 3-minute rest. The SS and DS groups then completed an 8.5-minute stretching intervention. Next, all subjects completed 3 trials of the 505 agility test with 2-5 minutes of rest between trials. A 2-way repeated-measure analysis of variance (Stretch group, athlete category, group × athlete interaction) was used to determine statistical significance (p < 0.05). A Tukey post hoc test was performed to determine differences between groups. For all athletes, the DS group produced significantly faster times on the agility test (2.22 ± 0.12 seconds, mean ± SD) in comparison to both the SS group (2.33 ± 0.15 seconds, p = 0.013) and NS group (2.32 ± 0.12 seconds, p = 0.026). Differences between the SS and NS groups revealed no significance (p = 0.962). There was a significant difference in mean times for the type of athlete (p = 0.002); however, interaction between the type of athlete and stretching group was not significant (p = 0.520). These results indicate that in comparison to SS or NS, DS significantly improves performance on closed agility skills involving a 180° change of direction.

  7. Reversibility in Quantum Models of Stochastic Processes

    NASA Astrophysics Data System (ADS)

    Gier, David; Crutchfield, James; Mahoney, John; James, Ryan

    Natural phenomena such as time series of neural firing, orientation of layers in crystal stacking and successive measurements in spin-systems are inherently probabilistic. The provably minimal classical models of such stochastic processes are ɛ-machines, which consist of internal states, transition probabilities between states and output values. The topological properties of the ɛ-machine for a given process characterize the structure, memory and patterns of that process. However ɛ-machines are often not ideal because their statistical complexity (Cμ) is demonstrably greater than the excess entropy (E) of the processes they represent. Quantum models (q-machines) of the same processes can do better in that their statistical complexity (Cq) obeys the relation Cμ >= Cq >= E. q-machines can be constructed to consider longer lengths of strings, resulting in greater compression. With code-words of sufficiently long length, the statistical complexity becomes time-symmetric - a feature apparently novel to this quantum representation. This result has ramifications for compression of classical information in quantum computing and quantum communication technology.

  8. Organizational Agility and Complex Enterprise System Innovations: A Mixed Methods Study of the Effects of Enterprise Systems on Organizational Agility

    ERIC Educational Resources Information Center

    Kharabe, Amol T.

    2012-01-01

    Over the last two decades, firms have operated in "increasingly" accelerated "high-velocity" dynamic markets, which require them to become "agile." During the same time frame, firms have increasingly deployed complex enterprise systems--large-scale packaged software "innovations" that integrate and automate…

  9. An agile acquisition decision-support workbench for evaluating ISR effectiveness

    NASA Astrophysics Data System (ADS)

    Stouch, Daniel W.; Champagne, Valerie; Mow, Christopher; Rosenberg, Brad; Serrin, Joshua

    2011-06-01

    The U.S. Air Force is consistently evolving to support current and future operations through the planning and execution of intelligence, surveillance and reconnaissance (ISR) missions. However, it is a challenge to maintain a precise awareness of current and emerging ISR capabilities to properly prepare for future conflicts. We present a decisionsupport tool for acquisition managers to empirically compare ISR capabilities and approaches to employing them, thereby enabling the DoD to acquire ISR platforms and sensors that provide the greatest return on investment. We have developed an analysis environment to perform modeling and simulation-based experiments to objectively compare alternatives. First, the analyst specifies an operational scenario for an area of operations by providing terrain and threat information; a set of nominated collections; sensor and platform capabilities; and processing, exploitation, and dissemination (PED) capacities. Next, the analyst selects and configures ISR collection strategies to generate collection plans. The analyst then defines customizable measures of effectiveness or performance to compute during the experiment. Finally, the analyst empirically compares the efficacy of each solution and generates concise reports to document their conclusions, providing traceable evidence for acquisition decisions. Our capability demonstrates the utility of using a workbench environment for analysts to design and run experiments. Crafting impartial metrics enables the acquisition manager to focus on evaluating solutions based on specific military needs. Finally, the metric and collection plan visualizations provide an intuitive understanding of the suitability of particular solutions. This facilitates a more agile acquisition strategy that handles rapidly changing technology in response to current military needs.

  10. Modeling Manufacturing Processes to Mitigate Technological Risk

    SciTech Connect

    Allgood, G.O.; Manges, W.W.

    1999-10-24

    An economic model is a tool for determining the justifiable cost of new sensors and subsystems with respect to value and operation. This process balances the R and D costs against the expense of maintaining current operations and allows for a method to calculate economic indices of performance that can be used as control points in deciding whether to continue development or suspend actions. The model can also be used as an integral part of an overall control loop utilizing real-time process data from the sensor groups to make production decisions (stop production and repair machine, continue and warn of anticipated problems, queue for repairs, etc.). This model has been successfully used and deployed in the CAFE Project. The economic model was one of seven (see Fig. 1) elements critical in developing an investment strategy. It has been successfully used in guiding the R and D activities on the CAFE Project, suspending activities on three new sensor technologies, and continuing development o f two others. The model has also been used to justify the development of a new prognostic approach for diagnosing machine health using COTS equipment and a new algorithmic approach. maintaining current operations and allows for a method to calculate economic indices of performance that can be used as control points in deciding whether to continue development or suspend actions. The model can also be used as an integral part of an overall control loop utilizing real-time process data from the sensor groups to make production decisions (stop production and repair machine, continue and warn of anticipated problems, queue for repairs, etc.).

  11. Glacier lake outburst floods - modelling process chains

    NASA Astrophysics Data System (ADS)

    Schaub, Yvonne; Huggel, Christian; Haeberli, Wilfried

    2013-04-01

    New lakes are forming in high-mountain areas all over the world due to glacier recession. Often they will be located below steep, destabilized flanks and are therefore exposed to impacts from rock-/ice-avalanches. Several events worldwide are known, where an outburst flood has been triggered by such an impact. In regions such as in the European Alps or in the Cordillera Blanca in Peru, where valley bottoms are densely populated, these far-travelling, high-magnitude events can result in major disasters. For appropriate integral risk management it is crucial to gain knowledge on how the processes (rock-/ice-avalanches - impact waves in lake - impact on dam - outburst flood) interact and how the hazard potential related to corresponding process chains can be assessed. Research in natural hazards so far has mainly concentrated on describing, understanding, modeling or assessing single hazardous processes. Some of the above mentioned individual processes are quite well understood in their physical behavior and some of the process interfaces have also been investigated in detail. Multi-hazard assessments of the entire process chain, however, have only recently become subjects of investigations. Our study aims at closing this gap and providing suggestions on how to assess the hazard potential of the entire process chain in order to generate hazard maps and support risk assessments. We analyzed different types of models (empirical, analytical, physically based) for each process regarding their suitability for application in hazard assessments of the entire process chain based on literature. Results show that for rock-/ice-avalanches, dam breach and outburst floods, only numerical, physically based models are able to provide the required information, whereas the impact wave can be estimated by means of physically based or empirical assessments. We demonstrate how the findings could be applied with the help of a case study of a recent glacier lake outburst event at Laguna

  12. Models of plasticity in spatial auditory processing.

    PubMed

    Shinn-Cunningham, B

    2001-01-01

    Both psychophysical and physiological studies have examined plasticity of spatial auditory processing. While there is a great deal known about how the system computes basic cues that influence spatial perception, less is known about how these cues are integrated to form spatial percepts and how the auditory system adapts and calibrates in order to maintain accurate spatial perception. After summarizing evidence for plasticity in the spatial auditory pathway, this paper reviews a statistical, decision-theory model of short-term plasticity and a system-level model of the spatial auditory pathway that may help elucidate how long- and short-term experiences influence the computations underlying spatial hearing.

  13. Mathematical model on a desalination process

    SciTech Connect

    Al-Samawi, A.A. )

    1994-05-01

    Mathematical models on the desalination of brackish water using EDR process are formulated. The product desalinated water variable is hypothesized as being dependent upon the following independent variables: total dissolved solids of the feed water, total dissolved solids of the product water, the rate of feed water, the temperature of feed water, the number of stages of membranes, and the energy consumption. The final model which is selected on statistical basis is considered appropriated for both prediction purposes and for the purpose of quantifying the separate effects of each significant variable upon the rate of production of desalted water variable. Results of the analysis are reported herein. 6 refs., 4 figs., 5 tabs.

  14. Expert Models and Modeling Processes Associated with a Computer-Modeling Tool

    ERIC Educational Resources Information Center

    Zhang, BaoHui; Liu, Xiufeng; Krajcik, Joseph S.

    2006-01-01

    Holding the premise that the development of expertise is a continuous process, this study concerns expert models and modeling processes associated with a modeling tool called Model-It. Five advanced Ph.D. students in environmental engineering and public health used Model-It to create and test models of water quality. Using "think aloud" technique…

  15. Near Field Environment Process Model Report

    SciTech Connect

    R.A. Wagner

    2000-11-14

    Waste emplacement and activities associated with construction of a repository system potentially will change environmental conditions within the repository system. These environmental changes principally result from heat generated by the decay of the radioactive waste, which elevates temperatures within the repository system. Elevated temperatures affect distribution of water, increase kinetic rates of geochemical processes, and cause stresses to change in magnitude and orientation from the stresses resulting from the overlying rock and from underground construction activities. The recognition of this evolving environment has been reflected in activities, studies and discussions generally associated with what has been termed the Near-Field Environment (NFE). The NFE interacts directly with waste packages and engineered barriers as well as potentially changing the fluid composition and flow conditions within the mountain. As such, the NFE defines the environment for assessing the performance of a potential Monitored Geologic Repository at Yucca Mountain, Nevada. The NFe evolves over time, and therefore is not amenable to direct characterization or measurement in the ambient system. Analysis or assessment of the NFE must rely upon projections based on tests and models that encompass the long-term processes of the evolution of this environment. This NFE Process Model Report (PMR) describes the analyses and modeling based on current understanding of the evolution of the near-field within the rock mass extending outward from the drift wall.

  16. Software Model Of Software-Development Process

    NASA Technical Reports Server (NTRS)

    Lin, Chi Y.; Synott, Debra J.; Levary, Reuven R.

    1990-01-01

    Collection of computer programs constitutes software tool for simulation of medium- to large-scale software-development projects. Necessary to include easily identifiable and more-readily quantifiable characteristics like costs, times, and numbers of errors. Mathematical model incorporating these and other factors of dynamics of software-development process implemented in the Software Life Cycle Simulator (SLICS) computer program. Simulates dynamics of software-development process. In combination with input and output expert software systems and knowledge-based management software system, develops information for use in managing large software-development project. Intended to aid managers in planning, managing, and controlling software-development processes by reducing uncertainties in budgets, required personnel, and schedules.

  17. Modeling biomedical experimental processes with OBI

    PubMed Central

    2010-01-01

    Background Experimental descriptions are typically stored as free text without using standardized terminology, creating challenges in comparison, reproduction and analysis. These difficulties impose limitations on data exchange and information retrieval. Results The Ontology for Biomedical Investigations (OBI), developed as a global, cross-community effort, provides a resource that represents biomedical investigations in an explicit and integrative framework. Here we detail three real-world applications of OBI, provide detailed modeling information and explain how to use OBI. Conclusion We demonstrate how OBI can be applied to different biomedical investigations to both facilitate interpretation of the experimental process and increase the computational processing and integration within the Semantic Web. The logical definitions of the entities involved allow computers to unambiguously understand and integrate different biological experimental processes and their relevant components. Availability OBI is available at http://purl.obolibrary.org/obo/obi/2009-11-02/obi.owl PMID:20626927

  18. The impact of working memory and the "process of process modelling" on model quality: Investigating experienced versus inexperienced modellers.

    PubMed

    Martini, Markus; Pinggera, Jakob; Neurauter, Manuel; Sachse, Pierre; Furtner, Marco R; Weber, Barbara

    2016-05-09

    A process model (PM) represents the graphical depiction of a business process, for instance, the entire process from online ordering a book until the parcel is delivered to the customer. Knowledge about relevant factors for creating PMs of high quality is lacking. The present study investigated the role of cognitive processes as well as modelling processes in creating a PM in experienced and inexperienced modellers. Specifically, two working memory (WM) functions (holding and processing of information and relational integration) and three process of process modelling phases (comprehension, modelling, and reconciliation) were related to PM quality. Our results show that the WM function of relational integration was positively related to PM quality in both modelling groups. The ratio of comprehension phases was negatively related to PM quality in inexperienced modellers and the ratio of reconciliation phases was positively related to PM quality in experienced modellers. Our research reveals central cognitive mechanisms in process modelling and has potential practical implications for the development of modelling software and teaching the craft of process modelling.

  19. The impact of working memory and the "process of process modelling" on model quality: Investigating experienced versus inexperienced modellers.

    PubMed

    Martini, Markus; Pinggera, Jakob; Neurauter, Manuel; Sachse, Pierre; Furtner, Marco R; Weber, Barbara

    2016-01-01

    A process model (PM) represents the graphical depiction of a business process, for instance, the entire process from online ordering a book until the parcel is delivered to the customer. Knowledge about relevant factors for creating PMs of high quality is lacking. The present study investigated the role of cognitive processes as well as modelling processes in creating a PM in experienced and inexperienced modellers. Specifically, two working memory (WM) functions (holding and processing of information and relational integration) and three process of process modelling phases (comprehension, modelling, and reconciliation) were related to PM quality. Our results show that the WM function of relational integration was positively related to PM quality in both modelling groups. The ratio of comprehension phases was negatively related to PM quality in inexperienced modellers and the ratio of reconciliation phases was positively related to PM quality in experienced modellers. Our research reveals central cognitive mechanisms in process modelling and has potential practical implications for the development of modelling software and teaching the craft of process modelling. PMID:27157858

  20. Reinventing The Design Process: Teams and Models

    NASA Technical Reports Server (NTRS)

    Wall, Stephen D.

    1999-01-01

    The future of space mission designing will be dramatically different from the past. Formerly, performance-driven paradigms emphasized data return with cost and schedule being secondary issues. Now and in the future, costs are capped and schedules fixed-these two variables must be treated as independent in the design process. Accordingly, JPL has redesigned its design process. At the conceptual level, design times have been reduced by properly defining the required design depth, improving the linkages between tools, and managing team dynamics. In implementation-phase design, system requirements will be held in crosscutting models, linked to subsystem design tools through a central database that captures the design and supplies needed configuration management and control. Mission goals will then be captured in timelining software that drives the models, testing their capability to execute the goals. Metrics are used to measure and control both processes and to ensure that design parameters converge through the design process within schedule constraints. This methodology manages margins controlled by acceptable risk levels. Thus, teams can evolve risk tolerance (and cost) as they would any engineering parameter. This new approach allows more design freedom for a longer time, which tends to encourage revolutionary and unexpected improvements in design.

  1. Inbreeding avoidance, patch isolation and matrix permeability influence dispersal and settlement choices by male agile antechinus in a fragmented landscape.

    PubMed

    Banks, Sam C; Lindenmayer, David B

    2014-03-01

    Animal dispersal is highly non-random and has important implications for the dynamics of populations in fragmented habitat. We identified interpatch dispersal events from genetic tagging, parentage analyses and assignment tests and modelled the factors associated with apparent emigration and post-dispersal settlement choices by individual male agile antechinus (Antechinus agilis, a marsupial carnivore of south-east Australian forests). Emigration decisions were best modelled with on data patch isolation and inbreeding risk. The choice of dispersal destination by males was influenced by inbreeding risk, female abundance, patch size, patch quality and matrix permeability (variation in land cover). Males were less likely to settle in patches without highly unrelated females. Our findings highlight the importance of individual-level dispersal data for understanding how multiple processes drive non-randomness in dispersal in modified landscapes. Fragmented landscapes present novel environmental, demographic and genetic contexts in which dispersal decisions are made, so the major factors affecting dispersal decisions in fragmented habitat may differ considerably from unfragmented landscapes. We show that the spatial scale of genetic neighbourhoods can be large in fragmented habitat, such that dispersing males can potentially settle in the presence of genetically similar females after moving considerable distances, thereby necessitating both a choice to emigrate and a choice of where to settle to avoid inbreeding.

  2. Modeling Dynamic Regulatory Processes in Stroke.

    SciTech Connect

    McDermott, Jason E.; Jarman, Kenneth D.; Taylor, Ronald C.; Lancaster, Mary J.; Shankaran, Harish; Vartanian, Keri B.; Stevens, S.L.; Stenzel-Poore, Mary; Sanfilippo, Antonio P.

    2012-10-11

    The ability to examine in silico the behavior of biological systems can greatly accelerate the pace of discovery in disease pathologies, such as stroke, where in vivo experimentation is lengthy and costly. In this paper we describe an approach to in silico examination of blood genomic responses to neuroprotective agents and subsequent stroke through the development of dynamic models of the regulatory processes observed in the experimental gene expression data. First, we identified functional gene clusters from these data. Next, we derived ordinary differential equations (ODEs) relating regulators and functional clusters from the data. These ODEs were used to develop dynamic models that simulate the expression of regulated functional clusters using system dynamics as the modeling paradigm. The dynamic model has the considerable advantage of only requiring an initial starting state, and does not require measurement of regulatory influences at each time point in order to make accurate predictions. The manipulation of input model parameters, such as changing the magnitude of gene expression, made it possible to assess the behavior of the networks through time under varying conditions. We report that an optimized dynamic model can provide accurate predictions of overall system behavior under several different preconditioning paradigms.

  3. Computational models of natural language processing

    SciTech Connect

    Bara, B.G.; Guida, G.

    1984-01-01

    The main concern in this work is the illustration of models for natural language processing, and the discussion of their role in the development of computational studies of language. Topics covered include the following: competence and performance in the design of natural language systems; planning and understanding speech acts by interpersonal games; a framework for integrating syntax and semantics; knowledge representation and natural language: extending the expressive power of proposition nodes; viewing parsing as word sense discrimination: a connectionist approach; a propositional language for text representation; from topic and focus of a sentence to linking in a text; language generation by computer; understanding the Chinese language; semantic primitives or meaning postulates: mental models of propositional representations; narrative complexity based on summarization algorithms; using focus to constrain language generation; and towards an integral model of language competence.

  4. Universal process modeling with VTRE for OPC

    NASA Astrophysics Data System (ADS)

    Granik, Yuri; Cobb, Nicolas B.; Do, Thuy

    2002-07-01

    In previous work, Cobb and Zakhor (SPIE, 2726, pp.208-222, 1996) introduced the VTR (Variable Threshold Resist) model and demonstrated its accuracy for fitting empirical data for 365 nm illumination (SPIE, 3051, pp. 458-468, 1997). The original work showed how EPE can be modeled as a function of a peak local image intensity and the slope of the adjacent cutline. Since then, authors such as J. Randall, et al., (Microel. Engineering, 46, pp. 59-63, 1999) have analyzed the VTR model including other parameters such as dose. In the current approach, the original VTR has been enhanced to the VTR-Enhanced (or VTRE) in 1999, and VT-5 models in 2002, for production in OPC applications, which include other image intensity parameters. Here we present a comprehensive report on VT (Variable Threshold) process modeling. It has the demonstrated ability to accurately capture resist and etching responses, alone or in the combination with experimental VEB (Variable Etch Bias, SPIE, 4346, p. 98, 2001) model, for a wide range of process conditions used in the contemporary IC manufacturing. We analyzed 14 different semiconductor company processes experimental setups totaling 3000 CD measurements to prove this point. We considered 248, 193, and 157 nm annular and standard illumination sources for poly, metal, and active layers. We report an accuracy of VT family models under a wide range of conditions, show usage methodology, and introduce a novel method for calculating VTRE wafer predictions on a dense image intensity grid. We use multiple regression method to fit VT models and discuss methods for calculating regression coefficients. It is shown that models with too many eigenvectors exhibit a tendency to overfit CD curves. Sub-sample cross-validation and overfitting criteria are derived to avoid this problem. The section on test pattern and usage methodology describes practical issues needed for VT usage in OPC modeling. Particularly we discuss the effects of metrology errors on

  5. MODELING PAVEMENT DETERIORATION PROCESSES BY POISSON HIDDEN MARKOV MODELS

    NASA Astrophysics Data System (ADS)

    Nam, Le Thanh; Kaito, Kiyoyuki; Kobayashi, Kiyoshi; Okizuka, Ryosuke

    In pavement management, it is important to estimate lifecycle cost, which is composed of the expenses for repairing local damages, including potholes, and repairing and rehabilitating the surface and base layers of pavements, including overlays. In this study, a model is produced under the assumption that the deterioration process of pavement is a complex one that includes local damages, which occur frequently, and the deterioration of the surface and base layers of pavement, which progresses slowly. The variation in pavement soundness is expressed by the Markov deterioration model and the Poisson hidden Markov deterioration model, in which the frequency of local damage depends on the distribution of pavement soundness, is formulated. In addition, the authors suggest a model estimation method using the Markov Chain Monte Carlo (MCMC) method, and attempt to demonstrate the applicability of the proposed Poisson hidden Markov deterioration model by studying concrete application cases.

  6. TUNS/TCIS information model/process model

    NASA Technical Reports Server (NTRS)

    Wilson, James

    1992-01-01

    An Information Model is comprised of graphical and textual notation suitable for describing and defining the problem domain - in our case, TUNS or TCIS. The model focuses on the real world under study. It identifies what is in the problem and organizes the data into a formal structure for documentation and communication purposes. The Information Model is composed of an Entity Relationship Diagram (ERD) and a Data Dictionary component. The combination of these components provide an easy to understand methodology for expressing the entities in the problem space, the relationships between entities and the characteristics (attributes) of the entities. This approach is the first step in information system development. The Information Model identifies the complete set of data elements processed by TUNS. This representation provides a conceptual view of TUNS from the perspective of entities, data, and relationships. The Information Model reflects the business practices and real-world entities that users must deal with.

  7. Multiphase Flow Modeling of Biofuel Production Processes

    SciTech Connect

    D. Gaston; D. P. Guillen; J. Tester

    2011-06-01

    As part of the Idaho National Laboratory's (INL's) Secure Energy Initiative, the INL is performing research in areas that are vital to ensuring clean, secure energy supplies for the future. The INL Hybrid Energy Systems Testing (HYTEST) Laboratory is being established to develop and test hybrid energy systems with the principal objective to safeguard U.S. Energy Security by reducing dependence on foreign petroleum. HYTEST involves producing liquid fuels in a Hybrid Energy System (HES) by integrating carbon-based (i.e., bio-mass, oil-shale, etc.) with non-carbon based energy sources (i.e., wind energy, hydro, geothermal, nuclear, etc.). Advances in process development, control and modeling are the unifying vision for HES. This paper describes new modeling tools and methodologies to simulate advanced energy processes. Needs are emerging that require advanced computational modeling of multiphase reacting systems in the energy arena, driven by the 2007 Energy Independence and Security Act, which requires production of 36 billion gal/yr of biofuels by 2022, with 21 billion gal of this as advanced biofuels. Advanced biofuels derived from microalgal biomass have the potential to help achieve the 21 billion gal mandate, as well as reduce greenhouse gas emissions. Production of biofuels from microalgae is receiving considerable interest due to their potentially high oil yields (around 600 gal/acre). Microalgae have a high lipid content (up to 50%) and grow 10 to 100 times faster than terrestrial plants. The use of environmentally friendly alternatives to solvents and reagents commonly employed in reaction and phase separation processes is being explored. This is accomplished through the use of hydrothermal technologies, which are chemical and physical transformations in high-temperature (200-600 C), high-pressure (5-40 MPa) liquid or supercritical water. Figure 1 shows a simplified diagram of the production of biofuels from algae. Hydrothermal processing has significant

  8. Agile supply chain capabilities: emerging patterns as a determinant of competitive objectives

    NASA Astrophysics Data System (ADS)

    Yusuf, Yahaya Y.; Adeleye, E. O.; Sivayoganathan, K.

    2001-10-01

    Turbulent change caused by factors such as changing customer and technological requirements threatens manufacturers through lower product life cycles, profits and bleak survival prospects. Therefore, several companies are stressing flexibility and agility in order to respond, real time, to the unique needs of customers and markets. However, the resource competencies required are often difficult to mobilise and retain by single companies. It is therefore imperative for companies to co-operate and leverage complementary competencies. To this end, legally separate and spatially distributed companies are becoming integrated through Internet-based technologies. The paper reviews emerging patterns in supply chain integration. It also explores the relationship between the emerging patterns and attainment of competitive objectives. The results reported in the paper are based on data from a survey by questionnaire. The survey involved 600 companies in the UK, as part of a larger study of agile manufacturing. The study was driven by a conceptual model, which relates supply chain practices to competitive objectives. The analysis involves the use of factor analysis to reduce research variables to a few principal components. Subsequently, multiple regression was conducted to study the relationship amongst the reduced variables. The results validate the proposed conceptual model and lend credence to current thinking that supply chain integration is a vital tool for competitive advantage.

  9. The Comprehensive Process Model of Engagement

    PubMed Central

    Cohen-Mansfield, Jiska; Marx, Marcia S.; Freedman, Laurence S.; Murad, Havi; Regier, Natalie G.; Thein, Khin; Dakheel-Ali, Maha

    2010-01-01

    Background Engagement refers to the act of being occupied or involved with an external stimulus. In dementia, engagement is the antithesis of apathy. Objective The Comprehensive Process Model of Engagement was examined, in which environmental, person, and stimulus characteristics impact the level of engagement of persons with dementia. Methods Participants were 193 residents of 7 Maryland nursing homes. All participants had a diagnosis of dementia. Stimulus engagement was assessed via the Observational Measure of Engagement. Engagement was measured by duration, attention, and attitude to the stimulus. 25 stimuli were presented, which were categorized as live human social stimuli, simulated social stimuli, inanimate social stimuli, a reading stimulus, manipulative stimuli, a music stimulus, task and work-related stimuli, and two different self-identity stimuli. Results All stimuli elicited significantly greater engagement in comparison to the control stimulus. In the multivariate model, music significantly increased engagement duration, while all other stimuli significantly increased duration, attention, and attitude. Significant environmental variables in the multivariate model that increased engagement were: use of the long introduction with modeling (relative to minimal introduction), any level of sound (most especially moderate sound), and the presence of between 2 to 24 people in the room. Significant personal attributes included MMSE scores, ADL performance and clarity of speech, which were positively associated with higher engagement scores. Conclusions Results are consistent with the Comprehensive Process Model of Engagement. Person attributes, environmental factors, and stimulus characteristics all contribute to the level and nature of engagement, with a secondary finding being that exposure to any stimulus elicits engagement in persons with dementia. PMID:21946802

  10. Process-Based Modeling of Constructed Wetlands

    NASA Astrophysics Data System (ADS)

    Baechler, S.; Brovelli, A.; Rossi, L.; Barry, D. A.

    2007-12-01

    Constructed wetlands (CWs) are widespread facilities for wastewater treatment. In subsurface flow wetlands, contaminated wastewater flows through a porous matrix, where oxidation and detoxification phenomena occur. Despite the large number of working CWs, system design and optimization are still mainly based upon empirical equations or simplified first-order kinetics. This results from an incomplete understanding of the system functioning, and may in turn hinder the performance and effectiveness of the treatment process. As a result, CWs are often considered not suitable to meet high water quality-standards, or to treat water contaminated with recalcitrant anthropogenic contaminants. To date, only a limited number of detailed numerical models have been developed and successfully applied to simulate constructed wetland behavior. Among these, one of the most complete and powerful is CW2D, which is based on Hydrus2D. The aim of this work is to develop a comprehensive simulator tailored to model the functioning of horizontal flow constructed wetlands and in turn provide a reliable design and optimization tool. The model is based upon PHWAT, a general reactive transport code for saturated flow. PHWAT couples MODFLOW, MT3DMS and PHREEQC-2 using an operator-splitting approach. The use of PHREEQC to simulate reactions allows great flexibility in simulating biogeochemical processes. The biogeochemical reaction network is similar to that of CW2D, and is based on the Activated Sludge Model (ASM). Kinetic oxidation of carbon sources and nutrient transformations (nitrogen and phosphorous primarily) are modeled via Monod-type kinetic equations. Oxygen dissolution is accounted for via a first-order mass-transfer equation. While the ASM model only includes a limited number of kinetic equations, the new simulator permits incorporation of an unlimited number of both kinetic and equilibrium reactions. Changes in pH, redox potential and surface reactions can be easily incorporated

  11. Time models and cognitive processes: a review

    PubMed Central

    Maniadakis, Michail; Trahanias, Panos

    2014-01-01

    The sense of time is an essential capacity of humans, with a major role in many of the cognitive processes expressed in our daily lifes. So far, in cognitive science and robotics research, mental capacities have been investigated in a theoretical and modeling framework that largely neglects the flow of time. Only recently there has been a rather limited, but constantly increasing interest in the temporal aspects of cognition, integrating time into a range of different models of perceptuo-motor capacities. The current paper aims to review existing works in the field and suggest directions for fruitful future work. This is particularly important for the newly developed field of artificial temporal cognition that is expected to significantly contribute in the development of sophisticated artificial agents seamlessly integrated into human societies. PMID:24578690

  12. Development of a dynamic thermal model process

    SciTech Connect

    Smith, F. R.

    1996-04-01

    A dynamic electrical-thermal modeling simulation technique was developed to allow up-front design of thermal and electronic packaging with a high degree of accuracy and confidence. We are developing a hybrid multichip module output driver which controls with power MOSFET driver circuits. These MOSFET circuits will dissipate from 13 to 26 watts per driver in a physical package less than two square inches. The power dissipation plus an operating temperature range of -55{degrees} C to 100{degrees} C makes an accurate thermal package design critical. The project goal was to develop a simulation process to dynamically model the electrical/thermal characteristics of the power MOSFETS using the SABER analog simulator and the ABAQUS finite element simulator. SABER would simulate the electrical characteristics of the multi-chip module design while co-simulation is being done with ABAQUS simulating the solid model thermal characteristics of the MOSFET package. The dynamic parameters, MOSFET power and chip temperature, would be actively passed between simulators to effect a coupled simulator modelling technique. The project required a development of a SABER late for the analog ASIC controller circuit, a dynamic electrical/thermal template for the IRF150 and IRF9130 power MOSFETs, a solid model of the multi-chip module package, FORTRAN code to handle I/Q between and HP755 workstation and SABER, and I/O between CRAY J90 computer and ABAQUS. The simulation model was certified by measured electrical characteristics of the circuits and real time thermal imaging of the output multichip module.

  13. A nursing information model process for interoperability.

    PubMed

    Chow, Marilyn; Beene, Murielle; O'Brien, Ann; Greim, Patricia; Cromwell, Tim; DuLong, Donna; Bedecarré, Diane

    2015-05-01

    The ability to share nursing data across organizations and electronic health records is a key component of improving care coordination and quality outcomes. Currently, substantial organizational and technical barriers limit the ability to share and compare essential patient data that inform nursing care. Nursing leaders at Kaiser Permanente and the U.S. Department of Veterans Affairs collaborated on the development of an evidence-based information model driven by nursing practice to enable data capture, re-use, and sharing between organizations and disparate electronic health records. This article describes a framework with repeatable steps and processes to enable the semantic interoperability of relevant and contextual nursing data. Hospital-acquired pressure ulcer prevention was selected as the prototype nurse-sensitive quality measure to develop and test the model. In a Health 2.0 Developer Challenge program from the Office of the National Coordinator for Health, mobile applications implemented the model to help nurses assess the risk of hospital-acquired pressure ulcers and reduce their severity. The common information model can be applied to other nurse-sensitive measures to enable data standardization supporting patient transitions between care settings, quality reporting, and research.

  14. Observer-participant models of neural processing.

    PubMed

    Fry, R L

    1995-01-01

    A model is proposed in which the neuron serves as an information channel. Channel distortion occurs through the channel since the mapping from input Boolean codes to output codes are many-to-one in that neuron outputs consist of just two distinguished states. Within the described model, the neuron performs a decision-making function. Decisions are made regarding the validity of a question passively posed by the neuron. This question becomes defined through learning hence learning is viewed as the process of determining an appropriate question based on supplied input ensembles. An application of the Shannon information measures of entropy and mutual information taken together in the context of the proposed model lead to the Hopfield neuron model with conditionalized Hebbian learning rules. Neural decisions are shown to be based on a sigmoidal transfer characteristic or, in the limit as computational temperature tends to zero, a maximum likelihood decision rule. The described work is contrasted with the information-theoretic approach of Linsker. PMID:18263380

  15. Sprint, agility, strength and endurance capacity in wheelchair basketball players

    PubMed Central

    Granados, C; Otero, M; Badiola, A; Olasagasti, J; Bidaurrazaga-Letona, I; Iturricastillo, A; Gil, SM

    2014-01-01

    The aims of the present study were, firstly, to determine the reliability and reproducibility of an agility T-test and Yo-Yo 10 m recovery test; and secondly, to analyse the physical characteristics measured by sprint, agility, strength and endurance field tests in wheelchair basketball (WB) players. 16 WB players (33.06 ± 7.36 years, 71.89 ± 21.71 kg and sitting body height 86.07 ± 6.82 cm) belonging to the national WB league participated in this study. Wheelchair sprint (5 and 20 m without ball, and 5 and 20 m with ball) agility (T-test and pick-up test) strength (handgrip and maximal pass) and endurance (Yo-Yo 10 m recovery test) were performed. T-test and Yo-Yo 10 m recovery test showed good reproducibility values (intraclass correlation coefficient, ICC = 0.74-0.94). The WB players’ results in 5 and 20 m sprints without a ball were 1.87 ± 0.21 s and 5.70 ± 0.43 s and with a ball 2.10 ± 0.30 s and 6.59 ± 0.61 s, being better than those reported in the literature. Regarding the pick-up test results (16.05 ± 0.52 s) and maximal pass (8.39 ± 1.77 m), players showed worse values than those obtained in elite players. The main contribution of the present study is the characterization of the physical performance profile of WB players using a field test battery. Furthermore, we demonstrated that the agility T-test and the aerobic Yo-Yo 10 m recovery test are reliable; consequently they may be appropriate instruments for measuring physical fitness in WB. PMID:25729153

  16. Sprint, agility, strength and endurance capacity in wheelchair basketball players.

    PubMed

    Yanci, J; Granados, C; Otero, M; Badiola, A; Olasagasti, J; Bidaurrazaga-Letona, I; Iturricastillo, A; Gil, Sm

    2015-03-01

    The aims of the present study were, firstly, to determine the reliability and reproducibility of an agility T-test and Yo-Yo 10 m recovery test; and secondly, to analyse the physical characteristics measured by sprint, agility, strength and endurance field tests in wheelchair basketball (WB) players. 16 WB players (33.06 ± 7.36 years, 71.89 ± 21.71 kg and sitting body height 86.07 ± 6.82 cm) belonging to the national WB league participated in this study. Wheelchair sprint (5 and 20 m without ball, and 5 and 20 m with ball) agility (T-test and pick-up test) strength (handgrip and maximal pass) and endurance (Yo-Yo 10 m recovery test) were performed. T-test and Yo-Yo 10 m recovery test showed good reproducibility values (intraclass correlation coefficient, ICC = 0.74-0.94). The WB players' results in 5 and 20 m sprints without a ball were 1.87 ± 0.21 s and 5.70 ± 0.43 s and with a ball 2.10 ± 0.30 s and 6.59 ± 0.61 s, being better than those reported in the literature. Regarding the pick-up test results (16.05 ± 0.52 s) and maximal pass (8.39 ± 1.77 m), players showed worse values than those obtained in elite players. The main contribution of the present study is the characterization of the physical performance profile of WB players using a field test battery. Furthermore, we demonstrated that the agility T-test and the aerobic Yo-Yo 10 m recovery test are reliable; consequently they may be appropriate instruments for measuring physical fitness in WB. PMID:25729153

  17. Measuring The Variability Of Gamma-Ray Sources With AGILE

    SciTech Connect

    Chen, Andrew W.; Vercellone, Stefano; Pellizzoni, Alberto; Tavani, Marco

    2005-02-21

    Variability in the gamma-ray flux above 100 MeV at various time scales is one of the primary characteristics of the sources detected by EGRET, both allowing the identification of individual sources and constraining the unidentified source classes. We present a detailed simulation of the capacity of AGILE to characterize the variability of gamma-ray sources, discussing the implications for source population studies.

  18. Laser agile illumination for object tracking and classification - Feasibility study

    NASA Technical Reports Server (NTRS)

    Scholl, Marija S.; Vanzyl, Jakob J.; Meinel, Aden B.; Meinel, Marjorie P.; Scholl, James W.

    1988-01-01

    The 'agile illumination' concept for discrimination between ICBM warheads and decoys involves a two-aperture illumination with coherent light, diffraction of light by propagation, and a resulting interference pattern on the object surface. A scanning two-beam interference pattern illuminates one object at a time; depending on the shape, momentum, spinning, and tumbling characteristics of the interrogated object, different temporal signals will be obtained for different classes of objects.

  19. AGILE Observations of the Gravitational-wave Event GW150914

    NASA Astrophysics Data System (ADS)

    Tavani, M.; Pittori, C.; Verrecchia, F.; Bulgarelli, A.; Giuliani, A.; Donnarumma, I.; Argan, A.; Trois, A.; Lucarelli, F.; Marisaldi, M.; Del Monte, E.; Evangelista, Y.; Fioretti, V.; Zoli, A.; Piano, G.; Munar-Adrover, P.; Antonelli, L. A.; Barbiellini, G.; Caraveo, P.; Cattaneo, P. W.; Costa, E.; Feroci, M.; Ferrari, A.; Longo, F.; Mereghetti, S.; Minervini, G.; Morselli, A.; Pacciani, L.; Pellizzoni, A.; Picozza, P.; Pilia, M.; Rappoldi, A.; Sabatini, S.; Vercellone, S.; Vittorini, V.; Giommi, P.; Colafrancesco, S.; Cardillo, M.; Galli, M.; Fuschino, F.

    2016-07-01

    We report the results of an extensive search through the AGILE data for a gamma-ray counterpart to the LIGO gravitational-wave (GW) event GW150914. Currently in spinning mode, AGILE has the potential of cover 80% of the sky with its gamma-ray instrument, more than 100 times a day. It turns out that AGILE came within a minute of the event time of observing the accessible GW150914 localization region. Interestingly, the gamma-ray detector exposed ∼65% of this region during the 100 s time intervals centered at ‑100 and +300 s from the event time. We determine a 2σ flux upper limit in the band 50 MeV–10 GeV, UL = 1.9 × 10‑8 erg cm‑2 s‑1, obtained ∼300 s after the event. The timing of this measurement is the fastest ever obtained for GW150914, and significantly constrains the electromagnetic emission of a possible high-energy counterpart. We also carried out a search for a gamma-ray precursor and delayed emission over five timescales ranging from minutes to days: in particular, we obtained an optimal exposure during the interval ‑150/‑30 s. In all these observations, we do not detect a significant signal associated with GW150914. We do not reveal the weak transient source reported by Fermi-GBM 0.4 s after the event time. However, even though a gamma-ray counterpart of the GW150914 event was not detected, the prospects for future AGILE observations of GW sources are decidedly promising.

  20. Sprint, agility, strength and endurance capacity in wheelchair basketball players.

    PubMed

    Yanci, J; Granados, C; Otero, M; Badiola, A; Olasagasti, J; Bidaurrazaga-Letona, I; Iturricastillo, A; Gil, Sm

    2015-03-01

    The aims of the present study were, firstly, to determine the reliability and reproducibility of an agility T-test and Yo-Yo 10 m recovery test; and secondly, to analyse the physical characteristics measured by sprint, agility, strength and endurance field tests in wheelchair basketball (WB) players. 16 WB players (33.06 ± 7.36 years, 71.89 ± 21.71 kg and sitting body height 86.07 ± 6.82 cm) belonging to the national WB league participated in this study. Wheelchair sprint (5 and 20 m without ball, and 5 and 20 m with ball) agility (T-test and pick-up test) strength (handgrip and maximal pass) and endurance (Yo-Yo 10 m recovery test) were performed. T-test and Yo-Yo 10 m recovery test showed good reproducibility values (intraclass correlation coefficient, ICC = 0.74-0.94). The WB players' results in 5 and 20 m sprints without a ball were 1.87 ± 0.21 s and 5.70 ± 0.43 s and with a ball 2.10 ± 0.30 s and 6.59 ± 0.61 s, being better than those reported in the literature. Regarding the pick-up test results (16.05 ± 0.52 s) and maximal pass (8.39 ± 1.77 m), players showed worse values than those obtained in elite players. The main contribution of the present study is the characterization of the physical performance profile of WB players using a field test battery. Furthermore, we demonstrated that the agility T-test and the aerobic Yo-Yo 10 m recovery test are reliable; consequently they may be appropriate instruments for measuring physical fitness in WB.

  1. Development of a Comprehensive Weld Process Model

    SciTech Connect

    Radhakrishnan, B.; Zacharia, T.

    1997-05-01

    This cooperative research and development agreement (CRADA) between Concurrent Technologies Corporation (CTC) and Lockheed Martin Energy Systems (LMES) combines CTC's expertise in the welding area and that of LMES to develop computer models and simulation software for welding processes. This development is of significant impact to the industry, including materials producers and fabricators. The main thrust of the research effort was to develop a comprehensive welding simulation methodology. A substantial amount of work has been done by several researchers to numerically model several welding processes. The primary drawback of most of the existing models is the lack of sound linkages between the mechanistic aspects (e.g., heat transfer, fluid flow, and residual stress) and the metallurgical aspects (e.g., microstructure development and control). A comprehensive numerical model which can be used to elucidate the effect of welding parameters/conditions on the temperature distribution, weld pool shape and size, solidification behavior, and microstructure development, as well as stresses and distortion, does not exist. It was therefore imperative to develop a comprehensive model which would predict all of the above phenomena during welding. The CRADA built upon an already existing three- dimensional (3-D) welding simulation model which was developed by LMES which is capable of predicting weld pool shape and the temperature history in 3-d single-pass welds. However, the model does not account for multipass welds, microstructural evolution, distortion and residual stresses. Additionally, the model requires large resources of computing time, which limits its use for practical applications. To overcome this, CTC and LMES have developed through this CRADA the comprehensive welding simulation model described above. The following technical tasks have been accomplished as part of the CRADA. 1. The LMES welding code has been ported to the Intel Paragon parallel computer at ORNL

  2. Clustering-based urbanisation to improve enterprise information systems agility

    NASA Astrophysics Data System (ADS)

    Imache, Rabah; Izza, Said; Ahmed-Nacer, Mohamed

    2015-11-01

    Enterprises are daily facing pressures to demonstrate their ability to adapt quickly to the unpredictable changes of their dynamic in terms of technology, social, legislative, competitiveness and globalisation. Thus, to ensure its place in this hard context, enterprise must always be agile and must ensure its sustainability by a continuous improvement of its information system (IS). Therefore, the agility of enterprise information systems (EISs) can be considered today as a primary objective of any enterprise. One way of achieving this objective is by the urbanisation of the EIS in the context of continuous improvement to make it a real asset servicing enterprise strategy. This paper investigates the benefits of EISs urbanisation based on clustering techniques as a driver for agility production and/or improvement to help managers and IT management departments to improve continuously the performance of the enterprise and make appropriate decisions in the scope of the enterprise objectives and strategy. This approach is applied to the urbanisation of a tour operator EIS.

  3. Agile Science Operations: A New Approach for Primitive Exploration Bodies

    NASA Technical Reports Server (NTRS)

    Chien, Steve A.; Thompson, David R.; Castillo-Rogez, Julie C.; Doyle, Richard; Estlin, Tara; Mclaren, David

    2012-01-01

    Primitive body exploration missions such as potential Comet Surface Sample Return or Trojan Tour and Rendezvous would challenge traditional operations practices. Earth-based observations would provide only basic understanding before arrival and many science goals would be defined during the initial rendezvous. It could be necessary to revise trajectories and observation plans to quickly characterize the target for safe, effective observations. Detection of outgassing activity and monitoring of comet surface activity are even more time constrained, with events occurring faster than round-trip light time. "Agile science operations" address these challenges with contingency plans that recognize the intrinsic uncertainty in the operating environment and science objectives. Planning for multiple alternatives can significantly improve the time required to repair and validate spacecraft command sequences. When appropriate, time-critical decisions can be automated and shifted to the spacecraft for immediate access to instrument data. Mirrored planning systems on both sides of the light-time gap permit transfer of authority back and forth as needed. We survey relevant science objectives, identifying time bottlenecks and the techniques that could be used to speed missions' reaction to new science data. Finally, we discuss the results of a trade study simulating agile observations during flyby and comet rendezvous scenarios. These experiments quantify instrument coverage of key surface features as a function of planning turnaround time. Careful application of agile operations techniques can play a significant role in realizing the Decadal Survey plan for primitive body exploration

  4. Observing peculiar γ-ray pulsars with AGILE

    NASA Astrophysics Data System (ADS)

    Pilia, M.; Pellizzoni, A.

    2011-08-01

    The AGILE γ-ray satellite provides large sky exposure levels (>=109 cm2 s per year on the Galactic Plane) with sensitivity peaking at E ~100 MeV where the bulk of pulsar energy output is typically released. Its ~1 μs absolute time tagging capability makes it perfectly suited for the study of γ-ray pulsars. AGILE collected a large number of γ-ray photons from EGRET pulsars (>=40,000 pulsed counts for Vela) in two years of observations unveiling new interesting features at sub-millisecond level in the pulsars' high-energy light-curves, γ-ray emission from pulsar glitches and Pulsar Wind Nebulae. AGILE detected about 20 nearby and energetic pulsars with good confidence through timing and/or spatial analysis. Among the newcomers we find pulsars with very high rotational energy losses, such as the remarkable PSR B1509-58 with a magnetic field in excess of 1013 Gauss, and PSR J2229+6114 providing a reliable identification for the previously unidentified EGRET source 3EG2227+6122. Moreover, the powerful millisecond pulsar B1821-24, in the globular cluster M28, is detected during a fraction of the observations.

  5. Computer vision challenges and technologies for agile manufacturing

    NASA Astrophysics Data System (ADS)

    Molley, Perry A.

    1996-02-01

    applicable to commercial production processes and applications. Computer vision will play a critical role in the new agile production environment for automation of processes such as inspection, assembly, welding, material dispensing and other process control tasks. Although there are many academic and commercial solutions that have been developed, none have had widespread adoption considering the huge potential number of applications that could benefit from this technology. The reason for this slow adoption is that the advantages of computer vision for automation can be a double-edged sword. The benefits can be lost if the vision system requires an inordinate amount of time for reprogramming by a skilled operator to account for different parts, changes in lighting conditions, background clutter, changes in optics, etc. Commercially available solutions typically require an operator to manually program the vision system with features used for the recognition. In a recent survey, we asked a number of commercial manufacturers and machine vision companies the question, 'What prevents machine vision systems from being more useful in factories?' The number one (and unanimous) response was that vision systems require too much skill to set up and program to be cost effective.

  6. Computer modeling of complete IC fabrication process

    NASA Astrophysics Data System (ADS)

    Dutton, Robert W.

    1987-05-01

    The development of fundamental algorithms for process and device modeling as well as novel integration of the tools for advanced Integrated Circuit (IC) technology design is discussed. The development of the first complete 2D process simulator, SUPREM 4, is reported. The algorithms are discussed as well as application to local-oxidation and extrinsic diffusion conditions which occur in CMOS AND BiCMOS technologies. The evolution of 1D (SEDAN) and 2D (PISCES) device analysis is discussed. The application of SEDAN to a variety of non-silicon technologies (GaAs and HgCdTe) are considered. A new multi-window analysis capability for PISCES which exploits Monte Carlo analysis of hot carriers has been demonstrated and used to characterize a variety of silicon MOSFET and GaAs MESFET effects. A parallel computer implementation of PISCES has been achieved using a Hypercube architecture. The PISCES program has been used for a range of important device studies including: latchup, analog switch analysis, MOSFET capacitance studies and bipolar transient device for ECL gates. The program is broadly applicable to RAM and BiCMOS technology analysis and design. In the analog switch technology area this research effort has produced a variety of important modeling and advances.

  7. Anisotropic model-based SAR processing

    NASA Astrophysics Data System (ADS)

    Knight, Chad; Gunther, Jake; Moon, Todd

    2013-05-01

    Synthetic aperture radar (SAR) collections that integrate over a wide range of aspect angles hold the potentional for improved resolution and fosters improved scene interpretability and target detection. However, in practice it is difficult to realize the potential due to the anisotropic scattering of objects in the scene. The radar cross section (RCS) of most objects changes as a function of aspect angle. The isotropic assumption is tacitly made for most common image formation algorithms (IFA). For wide aspect scenarios one way to account for anistropy would be to employ a piecewise linear model. This paper focuses on such a model but it incorporates aspect and spatial magnitude filters in the image formation process. This is advantageous when prior knowledge is available regarding the desired targets' RCS signature spatially and in aspect. The appropriate filters can be incorporated into the image formation processing so that specific targets are emphasized while other targets are suppressed. This is demonstrated on the Air Force Research Laboratory (AFRL) GOTCHA1 data set to demonstrate the utility of the proposed approach.

  8. Simulation model of clastic sedimentary processes

    SciTech Connect

    Tetzlaff, D.M.

    1987-01-01

    This dissertation describes SEDSIM, a computer model that simulates erosion, transport, and deposition of clastic sediments by free-surface flow in natural environments. SEDSIM is deterministic and is applicable to sedimentary processes in rivers, deltas, continental shelves, submarine canyons, and turbidite fans. The model is used to perform experiments in clastic sedimentation. Computer experimentation is limited by computing power available, but is free from scaling problems associated with laboratory experiments. SEDSIM responds to information provided to it at the outset of a simulation experiment, including topography, subsurface configuration, physical parameters of fluid and sediment, and characteristics of sediment sources. Extensive computer graphics are incorporated in SEDSIM. The user can display the three-dimensional geometry of simulated deposits in the form of successions of contour maps, perspective diagrams, vector plots of current velocities, and vertical sections of any azimuth orientation. The sections show both sediment age and composition. SEDSIM works realistically with processes involving channel shifting and topographic changes. Example applications include simulation of an ancient submarine canyon carved into a Cretaceous sequence in the National Petroleum Reserve in Alaska, known mainly from seismic sections and a sequence of Tertiary age in the Golden Meadow oil field of Louisiana, known principally from well logs.

  9. Migration and Marriage: Modeling the Joint Process

    PubMed Central

    Jang, Joy Bohyun; Casterline, John B; Snyder, Anastasia

    2016-01-01

    Background Previous research on inter-relations between migration and marriage has relied on overly simplistic assumptions about the structure of dependency between the two events. However, there is good reason to posit that each of the two transitions has an impact on the likelihood of the other, and that unobserved common factors may affect both migration and marriage, leading to a distorted impression of the causal impact of one on the other. Objective We will investigate relationships between migration and marriage in the United States using data from the National Longitudinal Survey of Youth 1979. We allow for inter-dependency between the two events and examine whether unobserved common factors affect the estimates of both migration and marriage. Methods We estimate a multi-process model in which migration and marriage are considered simultaneously in regression analysis and there is allowance for correlation between disturbances; the latter feature accounts for possible endogeneity between shared unobserved determinants. The model also includes random effects for persons, exploiting the fact that many people experience both events multiple times throughout their lives. Results Unobserved factors appear to significantly influence both migration and marriage, resulting in upward bias in estimates of the effects of each on the other when these shared common factors are not accounted for. Estimates from the multi-process model indicate that marriage significantly increases the hazard of migration while migration does not affect the hazard of marriage. Conclusions Omitting inter-dependency between life course events can lead to a mistaken impression of the direct effects of certain features of each event on the other. PMID:27182198

  10. Towards a Framework for Using Agile Approaches in Global Software Development

    NASA Astrophysics Data System (ADS)

    Hossain, Emam; Ali Babar, Muhammad; Verner, June

    As agile methods and Global Software Development (GSD) are become increasingly popular, GSD project managers have been exploring the viability of using agile approaches in their development environments. Despite the expected benefits of using an agile approach with a GSD project, the overall combining mechanisms of the two approaches are not clearly understood. To address this challenge, we propose a conceptual framework, based on the research literature. This framework is expected to aid a project manager in deciding what agile strategies are effective for a particular GSD project, taking into account project context. We use an industry-based case study to explore the components of our conceptual framework. Our case study is planned and conducted according to specific published case study guidelines. We identify the agile practices and agile supporting practices used by a GSD project manager in our case study and conclude with future research directions.

  11. Modelling infiltration processes in frozen soils

    NASA Astrophysics Data System (ADS)

    Ireson, A. M.; Barbour, L. S.

    2014-12-01

    Understanding the hydrological processes in soils subject to significant freeze-thaw is fraught by "experimental vagaries and theoretical imponderables" (Miller 1980, Applications of soil physics). The infiltration of snowmelt water and the subsequent transmission of unfrozen water during thawing, is governed by hydraulic conductivity values which are changing with both ice and unfrozen water content. Water held within pores is subject to capillary forces, which results in a freezing point depression (i.e. water remains in the liquid state slightly below 0°C). As the temperature drops below zero, water freezes first in the larger pores, and then in progressively smaller pores. Since the larger pores also are the first to empty by drainage, these pores may be air filled during freezing, while smaller water filled pores freeze. This explains why an unsaturated, frozen soil may still have a considerable infiltration capacity. Infiltration into frozen soil is a critical phenomena related to the risk of flooding in the Canadian prairies, controlling the partitioning of snowmelt into either infiltration or runoff. We propose a new model, based on conceptualizing the pore space as a bundle of capillary tubes (with significant differences to the capillary bundle model of Wannatabe and Flury, 2008, WRR, doi:10.1029/2008WR007102) which allows any air-filled macropores to contribute to the potential infiltration capacity of the soil. The patterns of infiltration and water movement during freeze-thaw from the model are compared to field observations from the Canadian prairies and Boreal Plains.

  12. Organizational Leadership Process for University Education

    ERIC Educational Resources Information Center

    Llamosa-Villalba, Ricardo; Delgado, Dario J.; Camacho, Heidi P.; Paéz, Ana M.; Valdivieso, Raúl F.

    2014-01-01

    This paper relates the "Agile School", an emerging archetype of the enterprise architecture: "Processes of Organizational Leadership" for leading and managing strategies, tactics and operations of forming in Higher Education Institutions. Agile School is a system for innovation and deep transformation of University Institutions…

  13. The Agile Approach with Doctoral Dissertation Supervision

    ERIC Educational Resources Information Center

    Tengberg, Lars Göran Wallgren

    2015-01-01

    Several research findings conclude that many doctoral students fail to complete their studies within the allowable time frame, in part because of problems related to the research and supervision process. Surveys show that most doctoral students are generally satisfied with their dissertation supervision. However, these surveys also reveal some…

  14. Mechanical-mathematical modeling for landslide process

    NASA Astrophysics Data System (ADS)

    Svalova, V.

    2009-04-01

    500 m and displacement of a landslide in the plan over 1 m. Last serious activization of a landslide has taken place in 2002 with a motion on 53 cm. Catastrophic activization of the deep blockglide landslide in the area of Khoroshevo in Moscow took place in 2006-2007. A crack of 330 m long appeared in the old sliding circus, along which a new 220 m long creeping block was separated from the plateau and began sinking with a displaced surface of the plateau reaching to 12 m. Such activization of the landslide process was not observed in Moscow since mid XIX century. The sliding area of Khoroshevo was stable during long time without manifestations of activity. Revealing of the reasons of deformation and development of ways of protection from deep landslide motions is extremely actual and difficult problem which decision is necessary for preservation of valuable historical monuments and modern city constructions. The reasons of activization and protective measures are discussed. Structure of monitoring system for urban territories is elaborated. Mechanical-mathematical model of high viscous fluid was used for modeling of matter behavior on landslide slopes. Equation of continuity and an approximated equation of the Navier-Stockes for slow motions in a thin layer were used. The results of modelling give possibility to define the place of highest velocity on landslide surface, which could be the best place for monitoring post position. Model can be used for calibration of monitoring equipment and gives possibility to investigate some fundamental aspects of matter movement on landslide slope.

  15. Team-based work and work system balance in the context of agile manufacturing.

    PubMed

    Yauch, Charlene A

    2007-01-01

    Manufacturing agility is the ability to prosper in an environment characterized by constant and unpredictable change. The purpose of this paper is to analyze team attributes necessary to facilitate agile manufacturing, and using Balance Theory as a framework, it evaluates the potential positive and negative impacts related to these team attributes that could alter the balance of work system elements and resulting "stress load" experienced by persons working on agile teams. Teams operating within the context of agile manufacturing are characterized as multifunctional, dynamic, cooperative, and virtual. A review of the literature relevant to each of these attributes is provided, as well as suggestions for future research.

  16. Relationship Between Reactive Agility and Change of Direction Speed in Amateur Soccer Players.

    PubMed

    Matlák, János; Tihanyi, József; Rácz, Levente

    2016-06-01

    Matlák, J, Tihanyi, J, and Rácz, L. Relationship between reactive agility and change of direction speed in amateur soccer players. J Strength Cond Res 30(6): 1547-1552, 2016-The aim of the study was to assess the relationship between reactive agility and change of direction speed (CODS) among amateur soccer players using running tests with four directional changes. Sixteen amateur soccer players (24.1 ± 3.3 years; 72.4 ± 7.3 kg; 178.7 ± 6 cm) completed CODS and reactive agility tests with four changes of direction using the SpeedCourt™ system (Globalspeed GmbH, Hemsbach, Germany). Countermovement jump (CMJ) height and maximal foot tapping count (completed in 3 seconds) were also measured with the same device. In the reactive agility test, participants had to react to a series of light stimuli projected onto a screen. Total time was shorter in the CODS test than in the reactive agility test (p < 0.001). Nonsignificant correlations were found among variables measured in the CODS, reactive agility, and CMJ tests. Low common variance (r = 0.03-0.18) was found between CODS and reactive agility test variables. The results of this study underscore the importance of cognitive factors in reactive agility performance and suggest that specific methods may be required for training and testing reactive agility and CODS. PMID:26562713

  17. A 'Common Information Model' for the climate modelling process

    NASA Astrophysics Data System (ADS)

    Treshansky, Allyn; Devine, Gerard

    2010-05-01

    The Common Information Model (CIM), developed by the EU-funded METAFOR project (http://metaforclimate.eu), is a formal model of the climate modeling process. It provides a rich structured description of not only climate data but also the "provenance" of that data: the software models and tools used to generate that data, the simulations those models implement, the experiments those simulations conform to, etc.. This formal metadata model is expected to add value to those datasets by firstly codifying what is currently found only in the heads of climate experts (the aforementioned provenance of climate datasets), and secondly by allowing tools to be developed that make searching for and analysing climate datasets a much more intuitive process than it has been in the past. This paper will describe the structure of the CIM, concentrating on how it works with and what it adds to other metadata standards. As alluded to above, current metadata standards concentrate on the contents of a climate dataset. Scientific detail and relevance of the model components that generated that data as well as the context for why it was run are missing. The CIM addresses this gap. However, it does not aim to replace existing standards. Rather, wherever possible it re-uses them. It also attempts to standardise our understanding of climate modeling at a very high level, at a conceptual level. This results in a UML description of climate modeling, the CONCIM. METAFOR extracts from this high-level UML the bits of the CIM that we want to use in our applications; These bits get converted into a set of XSD application schemas, the APPCIM. Other user groups may derive a different APPCIM (in a different format) that suits them from the same CONCIM. Thus there is a common understanding of the concepts used in climate modeling even if the implementation differs. In certain key places the CIM describes a general structure over which a specific Controlled Vocabulary (CV) can be applied. For example

  18. Nanowire growth process modeling and reliability models for nanodevices

    NASA Astrophysics Data System (ADS)

    Fathi Aghdam, Faranak

    Nowadays, nanotechnology is becoming an inescapable part of everyday life. The big barrier in front of its rapid growth is our incapability of producing nanoscale materials in a reliable and cost-effective way. In fact, the current yield of nano-devices is very low (around 10 %), which makes fabrications of nano-devices very expensive and uncertain. To overcome this challenge, the first and most important step is to investigate how to control nano-structure synthesis variations. The main directions of reliability research in nanotechnology can be classified either from a material perspective or from a device perspective. The first direction focuses on restructuring materials and/or optimizing process conditions at the nano-level (nanomaterials). The other direction is linked to nano-devices and includes the creation of nano-electronic and electro-mechanical systems at nano-level architectures by taking into account the reliability of future products. In this dissertation, we have investigated two topics on both nano-materials and nano-devices. In the first research work, we have studied the optimization of one of the most important nanowire growth processes using statistical methods. Research on nanowire growth with patterned arrays of catalyst has shown that the wire-to-wire spacing is an important factor affecting the quality of resulting nanowires. To improve the process yield and the length uniformity of fabricated nanowires, it is important to reduce the resource competition between nanowires during the growth process. We have proposed a physical-statistical nanowire-interaction model considering the shadowing effect and shared substrate diffusion area to determine the optimal pitch that would ensure the minimum competition between nanowires. A sigmoid function is used in the model, and the least squares estimation method is used to estimate the model parameters. The estimated model is then used to determine the optimal spatial arrangement of catalyst arrays

  19. What triggers the rising of an intraspecific biodiversity hotspot? Hints from the agile frog.

    PubMed

    Canestrelli, Daniele; Bisconti, Roberta; Sacco, Florinda; Nascetti, Giuseppe

    2014-01-01

    Hotspots of genetic diversity are regions of utmost importance for species survival and conservation, and their intimate link with the geographic location of glacial refugia has been well established. Nonetheless, the microevolutionary processes underlying the generation of hotspots in such regions have only recently become a fervent field of research. We investigated the phylogeographic and population genetic structure of the agile frog, Rana dalmatina, within its putative refugium in peninsular Italy. We found this region to harbour far more diversity, phylogeographic structure, and lineages of ancient origin than that by the rest of the species' range in Europe. This pattern appeared to be well explained by climate-driven microevolutionary processes that occurred during both glacial and interglacial epochs. Therefore, the inferred evolutionary history of R. dalmatina in Italy supports a view of glacial refugia as 'factories' rather than as repositories of genetic diversity, with significant implications for conservation strategies for hotspots.

  20. Model development for naphthenic acids ozonation process.

    PubMed

    Al Jibouri, Ali Kamel H; Wu, Jiangning

    2015-02-01

    Naphthenic acids (NAs) are toxic constituents of oil sands process-affected water (OSPW) which is generated during the extraction of bitumen from oil sands. NAs consist mainly of carboxylic acids which are generally biorefractory. For the treatment of OSPW, ozonation is a very beneficial method. It can significantly reduce the concentration of NAs and it can also convert NAs from biorefractory to biodegradable. In this study, a factorial design (2(4)) was used for the ozonation of OSPW to study the influences of the operating parameters (ozone concentration, oxygen/ozone flow rate, pH, and mixing) on the removal of a model NAs in a semi-batch reactor. It was found that ozone concentration had the most significant effect on the NAs concentration compared to other parameters. An empirical model was developed to correlate the concentration of NAs with ozone concentration, oxygen/ozone flow rate, and pH. In addition, a theoretical analysis was conducted to gain the insight into the relationship between the removal of NAs and the operating parameters. PMID:25189805

  1. How rolling forecasting facilitates dynamic, agile planning.

    PubMed

    Miller, Debra; Allen, Michael; Schnittger, Stephanie; Hackman, Theresa

    2013-11-01

    Rolling forecasting may be used to replace or supplement the annual budget process. The rolling forecast typically builds on the organization's strategic financial plan, focusing on the first three years of plan projections and comparing the strategic financial plan assumptions with the organization's expected trajectory. Leaders can then identify and respond to gaps between the rolling forecast and the strategic financial plan on an ongoing basis.

  2. Application of simulation models for the optimization of business processes

    NASA Astrophysics Data System (ADS)

    Jašek, Roman; Sedláček, Michal; Chramcov, Bronislav; Dvořák, Jiří

    2016-06-01

    The paper deals with the applications of modeling and simulation tools in the optimization of business processes, especially in solving an optimization of signal flow in security company. As a modeling tool was selected Simul8 software that is used to process modeling based on discrete event simulation and which enables the creation of a visual model of production and distribution processes.

  3. Developing Friction Stir Welding Process Model for ICME Application

    NASA Astrophysics Data System (ADS)

    Yang, Yu-Ping

    2015-01-01

    A framework for developing a product involving manufacturing processes was developed with integrated computational materials engineering approach. The key component in the framework is a process modeling tool which includes a thermal model, a microstructure model, a thermo-mechanical, and a property model. Using friction stir welding (FSW) process as an example, development of the process modeling tool was introduced in detail. The thermal model and the microstructure model of FSW of steels were validated with the experiment data. The model can predict reasonable temperature and hardness distributions as observed in the experiment. The model was applied to predict residual stress and joint strength of a pipe girth weld.

  4. The evaluation of several agility metrics for fighter aircraft using optimal trajectory analysis

    NASA Technical Reports Server (NTRS)

    Ryan, George W., III; Downing, David R.

    1993-01-01

    Several functional agility metrics, including the combat cycle time metric, dynamic speed turn plots, and relative energy state metric, are used to compare turning performance for generic F-18, X-29, and X-31-type aircraft models. These three-degree-of-freedom models have characteristics similar to the real aircraft. The performance comparisons are made using data from optimal test trajectories to reduce sensitivities to different pilot input techniques and to reduce the effects of control system limiters. The turn performance for all three aircraft is calculated for simulated minimum time 180 deg heading captures from simulation data. Comparisons of the three aircraft give more insight into turn performance than would be available from traditional measures of performance. Using the optimal test technique yields significant performance improvements as measured by the metrics. These performance improvements were found without significant increases in turn radius.

  5. Stochastic model updating utilizing Bayesian approach and Gaussian process model

    NASA Astrophysics Data System (ADS)

    Wan, Hua-Ping; Ren, Wei-Xin

    2016-03-01

    Stochastic model updating (SMU) has been increasingly applied in quantifying structural parameter uncertainty from responses variability. SMU for parameter uncertainty quantification refers to the problem of inverse uncertainty quantification (IUQ), which is a nontrivial task. Inverse problem solved with optimization usually brings about the issues of gradient computation, ill-conditionedness, and non-uniqueness. Moreover, the uncertainty present in response makes the inverse problem more complicated. In this study, Bayesian approach is adopted in SMU for parameter uncertainty quantification. The prominent strength of Bayesian approach for IUQ problem is that it solves IUQ problem in a straightforward manner, which enables it to avoid the previous issues. However, when applied to engineering structures that are modeled with a high-resolution finite element model (FEM), Bayesian approach is still computationally expensive since the commonly used Markov chain Monte Carlo (MCMC) method for Bayesian inference requires a large number of model runs to guarantee the convergence. Herein we reduce computational cost in two aspects. On the one hand, the fast-running Gaussian process model (GPM) is utilized to approximate the time-consuming high-resolution FEM. On the other hand, the advanced MCMC method using delayed rejection adaptive Metropolis (DRAM) algorithm that incorporates local adaptive strategy with global adaptive strategy is employed for Bayesian inference. In addition, we propose the use of the powerful variance-based global sensitivity analysis (GSA) in parameter selection to exclude non-influential parameters from calibration parameters, which yields a reduced-order model and thus further alleviates the computational burden. A simulated aluminum plate and a real-world complex cable-stayed pedestrian bridge are presented to illustrate the proposed framework and verify its feasibility.

  6. Heterogeneous processes: Laboratory, field, and modeling studies

    NASA Technical Reports Server (NTRS)

    Poole, Lamont R.; Kurylo, Michael J.; Jones, Rod L.; Wahner, Andreas; Calvert, Jack G.; Leu, M.-T.; Fried, A.; Molina, Mario J.; Hampson, Robert F.; Pitts, M. C.

    1991-01-01

    The efficiencies of chemical families such as ClO(x) and NO(x) for altering the total abundance and distribution of stratospheric ozone are controlled by a partitioning between reactive (active) and nonreactive (reservoir) compounds within each family. Gas phase thermodynamics, photochemistry, and kinetics would dictate, for example, that only about 1 percent of the chlorine resident in the lower stratosphere would be in the form of active Cl or ClO, the remainder existing in the reservoir compounds HCl and ClONO2. The consistency of this picture was recently challenged by the recognition that important chemical transformations take place on polar regions: the Airborne Antarctic Ozone Experiment (AAOE) and the Airborne Arctic Stratospheric Expedition (AASA). Following the discovery of the Antarctic ozone hole, Solomon et al. suggested that the heterogeneous chemical reaction: ClONO2(g)+HCl(s) yields Cl2(g)+HNO3(s) could play a key role in converting chlorine from inactive forms into a species (Cl2) that would rapidly dissociate in sunlight to liberate atomic chlorine and initiate ozone depletion. The symbols (s) and (g) denote solid phase, or adsorbed onto a solid surface, and gas phase, respectively, and represent the approach by which such a reaction is modeled rather than the microscopic details of the reaction. The reaction was expected to be most important at altitudes where PSC's were most prevalent (10 to 25 km), thereby extending the altitude range over which chlorine compounds can efficiently destroy ozone from the 35 to 45 km region (where concentrations of active chlorine are usually highest) to lower altitudes where the ozone concentration is at its peak. This chapter will briefly review the current state of knowledge of heterogeneous processes in the stratosphere, emphasizing those results obtained since the World Meteorological Organization (WMO) conference. Sections are included on laboratory investigations of heterogeneous reactions, the

  7. Reactive Agility Performance in Handball; Development and Evaluation of a Sport-Specific Measurement Protocol

    PubMed Central

    Spasic, Miodrag; Krolo, Ante; Zenic, Natasa; Delextrat, Anne; Sekulic, Damir

    2015-01-01

    There is no current study that examined sport-specific tests of reactive-agility and change-of-direction-speed (CODS) to replicate real-sport environment in handball (team-handball). This investigation evaluated the reliability and validity of two novel tests designed to assess reactive-agility and CODS of handball players. Participants were female (25.14 ± 3.71 years of age; 1.77 ± 0.09 m and 74.1 ± 6.1 kg) and male handball players (26.9 ± 4.1 years of age; 1.90 ± 0.09 m and 93.90±4.6 kg). Variables included body height, body mass, body mass index, broad jump, 5-m sprint, CODS and reactive-agility tests. Results showed satisfactory reliability for reactive-agility-test and CODS-test (ICC of 0.85-0.93, and CV of 2.4-4.8%). The reactive-agility and CODS shared less than 20% of the common variance. The calculated index of perceptual and reactive capacity (P&RC; ratio between reactive-agility- and CODS-performance) is found to be valid measure in defining true-game reactive-agility performance in handball in both genders. Therefore, the handball athletes’ P&RC should be used in the evaluation of real-game reactive-agility performance. Future studies should explore other sport-specific reactive-agility tests and factors associated to such performance in sports involving agile maneuvers. Key points Reactive agility and change-of-direction-speed should be observed as independent qualities, even when tested over the same course and similar movement template The reactive-agility-performance of the handball athletes involved in defensive duties is closer to their non-reactive-agility-score than in their peers who are not involved in defensive duties The handball specific “true-game” reactive-agility-performance should be evaluated as the ratio between reactive-agility and corresponding CODS performance. PMID:26336335

  8. Reactive Agility Performance in Handball; Development and Evaluation of a Sport-Specific Measurement Protocol.

    PubMed

    Spasic, Miodrag; Krolo, Ante; Zenic, Natasa; Delextrat, Anne; Sekulic, Damir

    2015-09-01

    There is no current study that examined sport-specific tests of reactive-agility and change-of-direction-speed (CODS) to replicate real-sport environment in handball (team-handball). This investigation evaluated the reliability and validity of two novel tests designed to assess reactive-agility and CODS of handball players. Participants were female (25.14 ± 3.71 years of age; 1.77 ± 0.09 m and 74.1 ± 6.1 kg) and male handball players (26.9 ± 4.1 years of age; 1.90 ± 0.09 m and 93.90±4.6 kg). Variables included body height, body mass, body mass index, broad jump, 5-m sprint, CODS and reactive-agility tests. Results showed satisfactory reliability for reactive-agility-test and CODS-test (ICC of 0.85-0.93, and CV of 2.4-4.8%). The reactive-agility and CODS shared less than 20% of the common variance. The calculated index of perceptual and reactive capacity (P&RC; ratio between reactive-agility- and CODS-performance) is found to be valid measure in defining true-game reactive-agility performance in handball in both genders. Therefore, the handball athletes' P&RC should be used in the evaluation of real-game reactive-agility performance. Future studies should explore other sport-specific reactive-agility tests and factors associated to such performance in sports involving agile maneuvers. Key pointsReactive agility and change-of-direction-speed should be observed as independent qualities, even when tested over the same course and similar movement templateThe reactive-agility-performance of the handball athletes involved in defensive duties is closer to their non-reactive-agility-score than in their peers who are not involved in defensive dutiesThe handball specific "true-game" reactive-agility-performance should be evaluated as the ratio between reactive-agility and corresponding CODS performance.

  9. Speed and agility of 12- and 14-year-old elite male basketball players.

    PubMed

    Jakovljevic, Sasa T; Karalejic, Milivoje S; Pajic, Zoran B; Macura, Marija M; Erculj, Frane F

    2012-09-01

    The aims of this study were (a) to identify and compare the speed and agility of 12- and 14-year-old elite male basketball players and (b) to investigate relations between speed and agility for both age groups of basketball players, to help coaches to improve their work. Sixty-four players aged 12 (M = 11.98 years, SD = 0.311) and 54 players aged 14 (M = 14.092 years, SD = 0.275) were tested. Three agility tests: agility t-test, zigzag agility drill, and agility run 4 × 15 m and 3 speed tests: 20-m run, 30-m run, and 50-m run were applied. Fourteen-year-old players achieved significantly better results in all speed and agility tests compared with 12-year-old players. The correlation coefficient (r = 0.81, p = 0.001) showed that 12-year-old players have the same ability in the 30- and 50-m runs. The other correlation coefficient (r = 0.59, p = 0.001) indicated that 20- and 30-m runs had inherently different qualities. The correlation coefficients between agility tests were <0.71, and therefore, each test in this group represents a specific task. In 14-year-old players, the correlation coefficients between the speed test results were <0.71. In contrast, the correlation coefficients between the agility tests were >0.71, which means that all the 3 tests represent the same quality. During the speed training of 12-year-old players, it is advisable to focus on shorter running distances, up to 30 m. During the agility training of the same players, it is useful to apply exercises with various complexities. In speed training of the 14-year-old players, the 30- and 50-m runs should be applied, and agility training should include more specific basketball movements and activities.

  10. Wired Widgets: Agile Visualization for Space Situational Awareness

    NASA Astrophysics Data System (ADS)

    Gerschefske, K.; Witmer, J.

    2012-09-01

    Continued advancement in sensors and analysis techniques have resulted in a wealth of Space Situational Awareness (SSA) data, made available via tools and Service Oriented Architectures (SOA) such as those in the Joint Space Operations Center Mission Systems (JMS) environment. Current visualization software cannot quickly adapt to rapidly changing missions and data, preventing operators and analysts from performing their jobs effectively. The value of this wealth of SSA data is not fully realized, as the operators' existing software is not built with the flexibility to consume new or changing sources of data or to rapidly customize their visualization as the mission evolves. While tools like the JMS user-defined operational picture (UDOP) have begun to fill this gap, this paper presents a further evolution, leveraging Web 2.0 technologies for maximum agility. We demonstrate a flexible Web widget framework with inter-widget data sharing, publish-subscribe eventing, and an API providing the basis for consumption of new data sources and adaptable visualization. Wired Widgets offers cross-portal widgets along with a widget communication framework and development toolkit for rapid new widget development, giving operators the ability to answer relevant questions as the mission evolves. Wired Widgets has been applied in a number of dynamic mission domains including disaster response, combat operations, and noncombatant evacuation scenarios. The variety of applications demonstrate that Wired Widgets provides a flexible, data driven solution for visualization in changing environments. In this paper, we show how, deployed in the Ozone Widget Framework portal environment, Wired Widgets can provide an agile, web-based visualization to support the SSA mission. Furthermore, we discuss how the tenets of agile visualization can generally be applied to the SSA problem space to provide operators flexibility, potentially informing future acquisition and system development.

  11. Opponent process model and psychostimulant addiction.

    PubMed

    Koob, G F; Caine, S B; Parsons, L; Markou, A; Weiss, F

    1997-07-01

    There are many sources of reinforcement in the spectrum of cocaine dependence that contribute to the compulsive cocaine self-administration or loss of control of cocaine intake that constitutes the core of modern definitions of dependence. The development of withdrawal has long been considered an integral part of drug addiction but has lost its impact in the theorization of drug dependence because of new emphasis on the neurobiological substrates for the positive-reinforcing properties of drugs. The present treatise reviews the neurobiological substrates for the acute positive reinforcing effects of cocaine and what is beginning to be known about the neurobiological substrates of cocaine withdrawal. The concept of motivational or affective withdrawal is reintroduced, which reemphasizes opponent process theory as a model for the motivational effects of cocaine dependence. The same neural substrates hypothesized to be involved in the acute reinforcing properties of drugs (basal forebrain regions of nucleus accumbens and amygdala) are hypothesized to be altered during chronic drug treatment to produce the negative motivational states characterizing drug withdrawal. Within these brain regions, both the neurochemical system(s) on which the drug has its primary actions and other neurochemical systems may undergo adaptations to chronic presence of the drug. An understanding of the adaptations of the motivational systems of the brain accompanying cocaine dependence leads to important predictions not only about the etiology, treatment, and prevention of cocaine addiction but also about the vulnerability of these motivational systems in non-drug-induced psychopathology.

  12. Development of a reburning boiler process model

    SciTech Connect

    Wu, K.T.

    1992-01-30

    The overall objective of this program is to integrate EER's expertise in boiler reburning performance evaluation into a package of analytical computer tools. Specific objectives of the program are to develop a computational capability with the following features: (1) can be used to predict the impact of gas reburning application on thermal conditions in the boiler radiant furnace, and on overall boiler performance; (2) can estimate gas reburning NO{sub x} reduction effectiveness based on specific reburning configurations and furnace/boiler configurations; (3) can be used as an analytical tool to evaluate the impact of boiler process parameters (e.g., fuel switching and changes in boiler operating conditions) on boiler thermal performance; (4) is adaptable to most boiler designs (tangential and wall fire boilers) and a variety of fuels (solid, liquid, gaseous and slurried fuels); (5) is sufficiently user friendly to be exercisable by engineers with a reasonable knowledge of boilers, and with reasonable computer skills. Here, user friendly'' means that the user will be guided by computer codes during the course of setting up individual input files for the boiler performance model.

  13. Signal Processing Model for Radiation Transport

    SciTech Connect

    Chambers, D H

    2008-07-28

    This note describes the design of a simplified gamma ray transport model for use in designing a sequential Bayesian signal processor for low-count detection and classification. It uses a simple one-dimensional geometry to describe the emitting source, shield effects, and detector (see Fig. 1). At present, only Compton scattering and photoelectric absorption are implemented for the shield and the detector. Other effects may be incorporated in the future by revising the expressions for the probabilities of escape and absorption. Pair production would require a redesign of the simulator to incorporate photon correlation effects. The initial design incorporates the physical effects that were present in the previous event mode sequence simulator created by Alan Meyer. The main difference is that this simulator transports the rate distributions instead of single photons. Event mode sequences and other time-dependent photon flux sequences are assumed to be marked Poisson processes that are entirely described by their rate distributions. Individual realizations can be constructed from the rate distribution using a random Poisson point sequence generator.

  14. Agents: An approach for dynamic process modelling

    NASA Astrophysics Data System (ADS)

    Grohmann, Axel; Kopetzky, Roland; Lurk, Alexander

    1999-03-01

    With the growing amount of distributed and heterogeneous information and services, conventional information systems have come to their limits. This gave rise to the development of a Multi-Agent System (the "Logical Client") which can be used in complex information systems as well as in other advanced software systems. Computer agents are proactive, reactive and social. They form a community of independent software components that can communicate and co-operate in order to accomplish complex tasks. Thus the agent-oriented paradigm provides a new and powerful approach to programming distributed systems. The communication framework developed is based on standards like CORBA, KQML and KIF. It provides an embedded rule based system to find adequate reactions to incoming messages. The macro-architecture of the Logical Client consists of independent agents and uses artificial intelligence to cope with complex patterns of communication and actions. A set of system agents is also provided, including the Strategy Service as a core component for modelling processes at runtime, the Computer Supported Cooperative Work (CSCW) Component for supporting remote co-operation between human users and the Repository for managing and hiding the file based data flow in heterogeneous networks. This architecture seems to be capable of managing complexity in information systems. It is also being implemented in a complex simulation system that monitors and simulates the environmental radioactivity in the country Baden-Württemberg.

  15. Modeling the Communication Process: The Map Is not the Territory.

    ERIC Educational Resources Information Center

    Bowman, Joel P.; Targowski, Andrew S.

    1987-01-01

    Presents a brief overview of the most significant models of the communication process, evaluates the communication models of the greatest relevance to business communication, and establishes a foundation for a new conception of that process. (JC)

  16. Muscle directly meets the vast power demands in agile lizards.

    PubMed

    Curtin, Nancy A; Woledge, Roger C; Aerts, Peter

    2005-03-22

    Level locomotion in small, agile lizards is characterized by intermittent bursts of fast running. These require very large accelerations, often reaching several times g. The power input required to increase kinetic energy is calculated to be as high as 214 W kg(-1) muscle (+/-20 W kg(-1) s.e.; averaged over the complete locomotor cycle) and 952 W kg(-1) muscle (+/-89 W kg(-1) s.e.; instantaneous peak power). In vitro muscle experiments prove that these exceptional power requirements can be met directly by the lizard's muscle fibres alone; there is no need for mechanical power amplifying mechanisms.

  17. Perspectives on Industrial Innovation from Agilent, HP, and Bell Labs

    NASA Astrophysics Data System (ADS)

    Hollenhorst, James

    2014-03-01

    Innovation is the life blood of technology companies. I will give perspectives gleaned from a career in research and development at Bell Labs, HP Labs, and Agilent Labs, from the point of view of an individual contributor and a manager. Physicists bring a unique set of skills to the corporate environment, including a desire to understand the fundamentals, a solid foundation in physical principles, expertise in applied mathematics, and most importantly, an attitude: namely, that hard problems can be solved by breaking them into manageable pieces. In my experience, hiring managers in industry seldom explicitly search for physicists, but they want people with those skills.

  18. Agile Delivery of Protein Therapeutics to CNS

    PubMed Central

    Yi, Xiang; Manickam, Devika S.; Brynskikh, Anna; Kabanov, Alexander V.

    2014-01-01

    A variety of therapeutic proteins have shown potential to treat central nervous system (CNS) disorders. Challenge to deliver these protein molecules to the brain is well known. Proteins administered through parenteral routes are often excluded from the brain because of their poor bioavailability and the existence of the blood-brain barrier (BBB). Barriers also exist to proteins administered through non-parenteral routes that bypass the BBB. Several strategies have shown promise in delivering proteins to the brain. This review, first, describes the physiology and pathology of the BBB that underscore the rationale and needs of each strategy to be applied. Second, major classes of protein therapeutics along with some key factors that affect their delivery outcomes are presented. Third, different routes of protein administration (parenteral, central intracerebroventricular and intraparenchymal, intranasal and intrathecal) are discussed along with key barriers to CNS delivery associated with each route. Finally, current delivery strategies involving chemical modification of proteins and use of particle-based carriers are overviewed using examples from literature and our own work. Whereas most of these studies are in the early stage, some provide proof of mechanism of increased protein delivery to the brain in relevant models of CNS diseases, while in few cases proof of concept had been attained in clinical studies. This review will be useful to broad audience of students, academicians and industry professionals who consider critical issues of protein delivery to the brain and aim developing and studying effective brain delivery systems for protein therapeutics. PMID:24956489

  19. Agile delivery of protein therapeutics to CNS.

    PubMed

    Yi, Xiang; Manickam, Devika S; Brynskikh, Anna; Kabanov, Alexander V

    2014-09-28

    A variety of therapeutic proteins have shown potential to treat central nervous system (CNS) disorders. Challenge to deliver these protein molecules to the brain is well known. Proteins administered through parenteral routes are often excluded from the brain because of their poor bioavailability and the existence of the blood-brain barrier (BBB). Barriers also exist to proteins administered through non-parenteral routes that bypass the BBB. Several strategies have shown promise in delivering proteins to the brain. This review, first, describes the physiology and pathology of the BBB that underscore the rationale and needs of each strategy to be applied. Second, major classes of protein therapeutics along with some key factors that affect their delivery outcomes are presented. Third, different routes of protein administration (parenteral, central intracerebroventricular and intraparenchymal, intranasal and intrathecal) are discussed along with key barriers to CNS delivery associated with each route. Finally, current delivery strategies involving chemical modification of proteins and use of particle-based carriers are overviewed using examples from literature and our own work. Whereas most of these studies are in the early stage, some provide proof of mechanism of increased protein delivery to the brain in relevant models of CNS diseases, while in few cases proof of concept had been attained in clinical studies. This review will be useful to broad audience of students, academicians and industry professionals who consider critical issues of protein delivery to the brain and aim developing and studying effective brain delivery systems for protein therapeutics.

  20. Process modeling - It's history, current status, and future

    NASA Astrophysics Data System (ADS)

    Duttweiler, Russell E.; Griffith, Walter M.; Jain, Sulekh C.

    1991-04-01

    The development of process modeling is reviewed to examine the potential of process applications to prevent and solve problems associated with the aerospace industry. The business and global environments is assessed, and the traditional approach to product/process design is argued to be obsolete. A revised engineering process is described which involves planning and prediction before production by means of process simulation. Process simulation can permit simultaneous engineering of unit processes and complex processes, and examples are given in the cross-coupling of forging-process variance. The implementation of process modeling, CAE, and computer simulations are found to reduce costs and time associated with technological development when incorporated judiciously.

  1. ISS Double-Gimbaled CMG Subsystem Simulation Using the Agile Development Method

    NASA Technical Reports Server (NTRS)

    Inampudi, Ravi

    2016-01-01

    This paper presents an evolutionary approach in simulating a cluster of 4 Control Moment Gyros (CMG) on the International Space Station (ISS) using a common sense approach (the agile development method) for concurrent mathematical modeling and simulation of the CMG subsystem. This simulation is part of Training systems for the 21st Century simulator which will provide training for crew members, instructors, and flight controllers. The basic idea of how the CMGs on the space station are used for its non-propulsive attitude control is briefly explained to set up the context for simulating a CMG subsystem. Next different reference frames and the detailed equations of motion (EOM) for multiple double-gimbal variable-speed control moment gyroscopes (DGVs) are presented. Fixing some of the terms in the EOM becomes the special case EOM for ISS's double-gimbaled fixed speed CMGs. CMG simulation development using the agile development method is presented in which customer's requirements and solutions evolve through iterative analysis, design, coding, unit testing and acceptance testing. At the end of the iteration a set of features implemented in that iteration are demonstrated to the flight controllers thus creating a short feedback loop and helping in creating adaptive development cycles. The unified modeling language (UML) tool is used in illustrating the user stories, class designs and sequence diagrams. This incremental development approach of mathematical modeling and simulating the CMG subsystem involved the development team and the customer early on, thus improving the quality of the working CMG system in each iteration and helping the team to accurately predict the cost, schedule and delivery of the software.

  2. Expert models and modeling processes associated with a computer-modeling tool

    NASA Astrophysics Data System (ADS)

    Zhang, Baohui; Liu, Xiufeng; Krajcik, Joseph S.

    2006-07-01

    Holding the premise that the development of expertise is a continuous process, this study concerns expert models and modeling processes associated with a modeling tool called Model-It. Five advanced Ph.D. students in environmental engineering and public health used Model-It to create and test models of water quality. Using think aloud technique and video recording, we captured their computer screen modeling activities and thinking processes. We also interviewed them the day following their modeling sessions to further probe the rationale of their modeling practices. We analyzed both the audio-video transcripts and the experts' models. We found the experts' modeling processes followed the linear sequence built in the modeling program with few instances of moving back and forth. They specified their goals up front and spent a long time thinking through an entire model before acting. They specified relationships with accurate and convincing evidence. Factors (i.e., variables) in expert models were clustered, and represented by specialized technical terms. Based on the above findings, we made suggestions for improving model-based science teaching and learning using Model-It.

  3. Two Undergraduate Process Modeling Courses Taught Using Inductive Learning Methods

    ERIC Educational Resources Information Center

    Soroush, Masoud; Weinberger, Charles B.

    2010-01-01

    This manuscript presents a successful application of inductive learning in process modeling. It describes two process modeling courses that use inductive learning methods such as inquiry learning and problem-based learning, among others. The courses include a novel collection of multi-disciplinary complementary process modeling examples. They were…

  4. The Impacts of Agile Development Methodology Use on Project Success: A Contingency View

    ERIC Educational Resources Information Center

    Tripp, John F.

    2012-01-01

    Agile Information Systems Development Methods have emerged in the past decade as an alternative manner of managing the work and delivery of information systems development teams, with a large number of organizations reporting the adoption & use of agile methods. The practitioners of these methods make broad claims as to the benefits of their…

  5. Renewed gamma-ray activity of the Blazar 3C 454.3 detected by AGILE

    NASA Astrophysics Data System (ADS)

    Bulgarelli, A.; Parmiggiani, N.; Fioretti, V.; Zoli, A.; Lucarelli, F.; Verrecchia, F.; Pittori, C.; Vercellone, S.; Piano, G.; Munar-Adrover, P.; Tavani, M.; Donnarumma, I.; Striani, E.; Cardillo, M.; Gianotti, F.; Trifoglio, M.; Giuliani, A.; Mereghetti, S.; Caraveo, P.; Perotti, F.; Chen, A.; Argan, A.; Costa, E.; Del Monte, E.; Evangelista, Y.; Feroci, M.; Lazzarotto, F.; Lapshov, I.; Pacciani, L.; Soffitta, P.; Sabatini, S.; Vittorini, V.; Pucella, G.; Rapisarda, M.; Di Cocco, G.; Fuschino, F.; Galli, M.; Labanti, C.; Marisaldi, M.; Pellizzoni, A.; Pilia, M.; Trois, A.; Barbiellini, G.; Vallazza, E.; Longo, F.; Morselli, A.; Picozza, P.; Prest, M.; Lipari, P.; Zanello, D.; Cattaneo, P. W.; Rappoldi, A.; Colafrancesco, S.; Ferrari, A.; Antonelli, A.; Giommi, P.; Salotti, L.; Valentini, G.; D'Amico, F.

    2016-06-01

    The AGILE satellite is detecting a significant enhancement in gamma-ray activity from the FSRQ 3C 454.3 (known as 1AGLR J2254+1609) since the recent AGILE ATel #9157, and the optical activity reported in ATel #9150.

  6. Project-Method Fit: Exploring Factors That Influence Agile Method Use

    ERIC Educational Resources Information Center

    Young, Diana K.

    2013-01-01

    While the productivity and quality implications of agile software development methods (SDMs) have been demonstrated, research concerning the project contexts where their use is most appropriate has yielded less definitive results. Most experts agree that agile SDMs are not suited for all project contexts. Several project and team factors have been…

  7. Impact of Business Intelligence and IT Infrastructure Flexibility on Competitive Advantage: An Organizational Agility Perspective

    ERIC Educational Resources Information Center

    Chen, Xiaofeng

    2012-01-01

    There is growing use of business intelligence (BI) for better management decisions in industry. However, empirical studies on BI are still scarce in academic research. This research investigates BI from an organizational agility perspective. Organizational agility is the ability to sense and respond to market opportunities and threats with speed,…

  8. Renewed gamma-ray activity of the Blazar 3C 454.3 detected by AGILE

    NASA Astrophysics Data System (ADS)

    Verrecchia, F.; Fioretti, V.; Lucarelli, F.; Pittori, C.; Bulgarelli, A.; Tavani, M.; Vercellone, S.; Piano, G.; Donnarumma, I.; Striani, E.; Cardillo, M.; Gianotti, F.; Trifoglio, M.; Giuliani, A.; Mereghetti, S.; Caraveo, P.; Perotti, F.; Chen, A.; Colafrancesco, S.; Del Monte, E.; Evangelista, Y.; Feroci, M.; Lazzarotto, F.; Pacciani, L.; Soffitta, P.; Costa, E.; Lapshov, I.; Rapisarda, M.; Argan, A.; Pucella, G.; Sabatini, S.; Trois, A.; Vittorini, V.; Fuschino, F.; Galli, M.; Labanti, C.; Marisaldi, M.; Di Cocco, G.; Pellizzoni, A.; Pilia, M.; Barbiellini, G.; Vallazza, E.; Longo, F.; Morselli, A.; Picozza, P.; Prest, M.; Lipari, P.; Zanello, D.; Cattaneo, P. W.; Rappoldi, A.; Giommi, P.; Salotti, L.; Valentini, G.

    2014-06-01

    The AGILE satellite detects a significant enhancement in gamma-ray activity from the FSRQ 3C 454.3 (known as 1AGLR J2254+1609 and 2FGL J2253.9+1609) since the recent AGILE ATel #6182, and the following NIR flare reported by Carrasco et al. ...

  9. A Speeded Item Response Model with Gradual Process Change

    ERIC Educational Resources Information Center

    Goegebeur, Yuri; De Boeck, Paul; Wollack, James A.; Cohen, Allan S.

    2008-01-01

    An item response theory model for dealing with test speededness is proposed. The model consists of two random processes, a problem solving process and a random guessing process, with the random guessing gradually taking over from the problem solving process. The involved change point and change rate are considered random parameters in order to…

  10. Modeling the evolutionary demographic processes for geomedicine

    NASA Astrophysics Data System (ADS)

    Lushnikov, A. A.; Kagan, A. I.; Gvishiani, A. D.; Lyubovtseva, Yu. S.

    2014-12-01

    We describe the principles for constructing evolutionary demographic models for geomedical statistics. Several variants of evolutionary models are proposed: (1) a model of the evolution of a closed population taking into account distribution by age, (2) a model that takes into account the morbidity and difference in mortality for groups of patients and healthy individuals, (3) a model that takes into account the distribution of different age groups by fertile ability, (4) a migration model that takes into account the population exchange between several localities, and (5) a model of the propagation of infectious diseases. Each model depends on a group of parameters determined from the medical and demographic state of the population. We discuss the possible application of the proposed evolutionary models to geomedical statistics.

  11. Gender-specific influences of balance, speed, and power on agility performance.

    PubMed

    Sekulic, Damir; Spasic, Miodrag; Mirkov, Dragan; Cavar, Mile; Sattler, Tine

    2013-03-01

    The quick change of direction (i.e., agility) is an important athletic ability in numerous sports. Because of the diverse and therefore hardly predictable manifestations of agility in sports, studies noted that the improvement in speed, power, and balance should result in an improvement of agility. However, there is evident lack of data regarding the influence of potential predictors on different agility manifestations. The aim of this study was to determine the gender-specific influence of speed, power, and balance on different agility tests. A total of 32 college-aged male athletes and 31 college-aged female athletes (age 20.02 ± 1.89 years) participated in this study. The subjects were mostly involved in team sports (soccer, team handball, basketball, and volleyball; 80% of men, and 75% of women), martial arts, gymnastics, and dance. Anthropometric variables consisted of body height, body weight, and the body mass index. Five agility tests were used: a t-test (T-TEST), zig-zag test, 20-yard shuttle test, agility test with a 180-degree turn, and forward-backward running agility test (FWDBWD). Other tests included 1 jumping ability power test (squat jump, SQJ), 2 balance tests to determine the overall stability index and an overall limit of stability score (both measured by Biodex Balance System), and 2 running speed tests using a straight sprint for 10 and 20 m (S10 and S20, respectively). A reliability analysis showed that all the agility tests were reliable. Multiple regression and correlation analysis found speed and power (among women), and balance (among men), as most significant predictors of agility. The highest Pearson's correlation in both genders is found between the results of the FWDBWD and S10M tests (0.77 and 0.81 for men and women, respectively; p < 0.05). Power, measured using the SQJ, is significantly (p < 0.05) related to FWDBWD and T-TEST results but only for women (-0.44; -0.41). The balance measures were significantly related to the agility

  12. Optimum detection of multiple vapor materials with frequency-agile lidar

    NASA Astrophysics Data System (ADS)

    Warren, Russell E.

    1996-07-01

    Differential absorption lidar (DIAL) is a well-established technology for estimating the concentration and its path integral CL of vapor materials using two closely spaced wavelengths. The recent development of frequency-agile lasers (FAL's) with as many as 60 wavelengths that can be rapidly scanned motivates the need for detection and estimation algorithms that are optimal for lidar employing these new sources. I derive detection and multimaterial CL estimation algorithms for FAL applications using the likelihood ratio test methodology of multivariate statistical inference theory. Three model sets of assumptions are considered with regard to the spectral properties of the backscatter from either topographic or aerosol targets. The calculations are illustrated through both simulated and actual lidar data.

  13. Research on rapid agile metrology for manufacturing based on real-time multitask operating system

    NASA Astrophysics Data System (ADS)

    Chen, Jihong; Song, Zhen; Yang, Daoshan; Zhou, Ji; Buckley, Shawn

    1996-10-01

    Rapid agile metrology for manufacturing (RAMM) using multiple non-contact sensors is likely to remain a growing trend in manufacturing. High speed inspecting systems for manufacturing is characterized by multitasks implemented in parallel and real-time events which occur simultaneously. In this paper, we introduce a real-time operating system into RAMM research. A general task model of a class-based object- oriented technology is proposed. A general multitask frame of a typical RAMM system using OPNet is discussed. Finally, an application example of a machine which inspects parts held on a carrier strip is described. With RTOS and OPNet, this machine can measure two dimensions of the contacts at 300 parts/second.

  14. A Model of Process-Based Automation: Cost and Quality Implications in the Medication Management Process

    ERIC Educational Resources Information Center

    Spaulding, Trent Joseph

    2011-01-01

    The objective of this research is to understand how a set of systems, as defined by the business process, creates value. The three studies contained in this work develop the model of process-based automation. The model states that complementarities among systems are specified by handoffs in the business process. The model also provides theory to…

  15. Model medication management process in Australian nursing homes using business process modeling.

    PubMed

    Qian, Siyu; Yu, Ping

    2013-01-01

    One of the reasons for end user avoidance or rejection to use health information systems is poor alignment of the system with healthcare workflow, likely causing by system designers' lack of thorough understanding about healthcare process. Therefore, understanding the healthcare workflow is the essential first step for the design of optimal technologies that will enable care staff to complete the intended tasks faster and better. The often use of multiple or "high risk" medicines by older people in nursing homes has the potential to increase medication error rate. To facilitate the design of information systems with most potential to improve patient safety, this study aims to understand medication management process in nursing homes using business process modeling method. The paper presents study design and preliminary findings from interviewing two registered nurses, who were team leaders in two nursing homes. Although there were subtle differences in medication management between the two homes, major medication management activities were similar. Further field observation will be conducted. Based on the data collected from observations, an as-is process model for medication management will be developed.

  16. Model medication management process in Australian nursing homes using business process modeling.

    PubMed

    Qian, Siyu; Yu, Ping

    2013-01-01

    One of the reasons for end user avoidance or rejection to use health information systems is poor alignment of the system with healthcare workflow, likely causing by system designers' lack of thorough understanding about healthcare process. Therefore, understanding the healthcare workflow is the essential first step for the design of optimal technologies that will enable care staff to complete the intended tasks faster and better. The often use of multiple or "high risk" medicines by older people in nursing homes has the potential to increase medication error rate. To facilitate the design of information systems with most potential to improve patient safety, this study aims to understand medication management process in nursing homes using business process modeling method. The paper presents study design and preliminary findings from interviewing two registered nurses, who were team leaders in two nursing homes. Although there were subtle differences in medication management between the two homes, major medication management activities were similar. Further field observation will be conducted. Based on the data collected from observations, an as-is process model for medication management will be developed. PMID:23920835

  17. Tokamak plasma modelling and atomic processes

    NASA Astrophysics Data System (ADS)

    Kawamura, T.

    1986-06-01

    Topics addressed include: particle control in a tokomak device; ionizing and recombining plasmas; effects of data accuracy on tokamak impurity transport modeling; plasma modeling of tokamaks; and ultraviolet and X-ray spectroscopy of tokamak plasmas.

  18. Modeling microbial processes in porous media

    NASA Astrophysics Data System (ADS)

    Murphy, Ellyn M.; Ginn, Timothy R.

    The incorporation of microbial processes into reactive transport models has generally proceeded along two separate lines of investigation: (1) transport of bacteria as inert colloids in porous media, and (2) the biodegradation of dissolved contaminants by a stationary phase of bacteria. Research over the last decade has indicated that these processes are closely linked. This linkage may occur when a change in metabolic activity alters the attachment/detachment rates of bacteria to surfaces, either promoting or retarding bacterial transport in a groundwater-contaminant plume. Changes in metabolic activity, in turn, are controlled by the time of exposure of the microbes to electron acceptors/donor and other components affecting activity. Similarly, metabolic activity can affect the reversibility of attachment, depending on the residence time of active microbes. Thus, improvements in quantitative analysis of active subsurface biota necessitate direct linkages between substrate availability, metabolic activity, growth, and attachment/detachment rates. This linkage requires both a detailed understanding of the biological processes and robust quantitative representations of these processes that can be tested experimentally. This paper presents an overview of current approaches used to represent physicochemical and biological processes in porous media, along with new conceptual approaches that link metabolic activity with partitioning of the microorganism between the aqueous and solid phases. Résumé L'introduction des processus microbiologiques dans des modèles de transport réactif a généralement suivi deux voies différentes de recherches: (1) le transport de bactéries sous forme de colloïdes inertes en milieu poreux, et (2) la biodégradation de polluants dissous par une phase stationnaire de bactéries. Les recherches conduites au cours des dix dernières années indiquent que ces processus sont intimement liés. Cette liaison peut intervenir lorsqu

  19. Modeling microbial processes in porous media

    NASA Astrophysics Data System (ADS)

    Murphy, Ellyn M.; Ginn, Timothy R.

    The incorporation of microbial processes into reactive transport models has generally proceeded along two separate lines of investigation: (1) transport of bacteria as inert colloids in porous media, and (2) the biodegradation of dissolved contaminants by a stationary phase of bacteria. Research over the last decade has indicated that these processes are closely linked. This linkage may occur when a change in metabolic activity alters the attachment/detachment rates of bacteria to surfaces, either promoting or retarding bacterial transport in a groundwater-contaminant plume. Changes in metabolic activity, in turn, are controlled by the time of exposure of the microbes to electron acceptors/donor and other components affecting activity. Similarly, metabolic activity can affect the reversibility of attachment, depending on the residence time of active microbes. Thus, improvements in quantitative analysis of active subsurface biota necessitate direct linkages between substrate availability, metabolic activity, growth, and attachment/detachment rates. This linkage requires both a detailed understanding of the biological processes and robust quantitative representations of these processes that can be tested experimentally. This paper presents an overview of current approaches used to represent physicochemical and biological processes in porous media, along with new conceptual approaches that link metabolic activity with partitioning of the microorganism between the aqueous and solid phases. Résumé L'introduction des processus microbiologiques dans des modèles de transport réactif a généralement suivi deux voies différentes de recherches: (1) le transport de bactéries sous forme de colloïdes inertes en milieu poreux, et (2) la biodégradation de polluants dissous par une phase stationnaire de bactéries. Les recherches conduites au cours des dix dernières années indiquent que ces processus sont intimement liés. Cette liaison peut intervenir lorsqu

  20. DESCRIPTION OF ATMOSPHERIC TRANSPORT PROCESSES IN EULERIAN AIR QUALITY MODELS

    EPA Science Inventory

    Key differences among many types of air quality models are the way atmospheric advection and turbulent diffusion processes are treated. Gaussian models use analytical solutions of the advection-diffusion equations. Lagrangian models use a hypothetical air parcel concept effecti...

  1. Macro Level Simulation Model Of Space Shuttle Processing

    NASA Technical Reports Server (NTRS)

    2000-01-01

    The contents include: 1) Space Shuttle Processing Simulation Model; 2) Knowledge Acquisition; 3) Simulation Input Analysis; 4) Model Applications in Current Shuttle Environment; and 5) Model Applications for Future Reusable Launch Vehicles (RLV's). This paper is presented in viewgraph form.

  2. Enhanced detection of Terrestrial Gamma-Ray Flashes by AGILE

    NASA Astrophysics Data System (ADS)

    Marisaldi, M.; Argan, A.; Ursi, A.; Gjesteland, T.; Fuschino, F.; Labanti, C.; Galli, M.; Tavani, M.; Pittori, C.; Verrecchia, F.; D'Amico, F.; Ostgaard, N.; Mereghetti, S.; Campana, R.; Cattaneo, P.; Bulgarelli, A.; Colafrancesco, S.; Dietrich, S.; Longo, F.; Gianotti, F.; Giommi, P.; Rappoldi, A.; Trifoglio, M.; Trois, A.

    2015-12-01

    At the end of March 2015 the onboard configuration of the AGILE satellite was modified in order to disable the veto signal of the anticoincidence shield for the minicalorimeter instrument. The motivation for such a change was the understanding that the dead time induced by the anticoincidence prevented the detection of a large fraction of Terrestrial Gamma-ray Flashes (TGFs), especially the short duration ones. We present here the characteristics of the new TGF sample after several months of stable operations with the new configuration. The configuration change was highly successful resulting in the detection of about 100 TGFs/month, an increase of a factor about 11 in TGFs detection rate with respect to previous configuration. As expected, the largest fraction of the new events has short duration, with a median duration of 80 microseconds. We also obtain a sample of events with simultaneous association, within 100 microseconds, with lightning sferics detected by the World Wide Lightning Location Network (WWLLN), confirming previous results reported by the Fermi mission. Given the high detection rate and the AGILE very low (+/-2.5°) orbital inclination, the new configuration provides the largest TGF detection rate surface density (TGFs / km2 / year) to date, opening space for correlation studies with lightning and atmospheric parameters on short spatial and temporal scales along the equatorial region. Eventually, the events with associated simultaneous WWLLN sferics provide a highly reliable sample to probe the long-standing issue of the TGF maximal energy.

  3. Frequency-agile microwave components using ferroelectric materials

    NASA Astrophysics Data System (ADS)

    Colom-Ustariz, Jose G.; Rodriguez-Solis, Rafael; Velez, Salmir; Rodriguez-Acosta, Snaider

    2003-04-01

    The non-linear electric field dependence of ferroelectric thin films can be used to design frequency and phase agile components. Tunable components have traditionally been developed using mechanically tuned resonant structures, ferrite components, or semiconductor-based voltage controlled electronics, but they are limited by their frequency performance, high cost, hgih losses, and integration into larger systems. In contrast, the ferroelectric-based tunable microwave component can easily be integrated into conventional microstrip circuits and attributes such as small size, light weight, and low-loss make these components attractive for broadband and multi-frequency applications. Components that are essential elements in the design of a microwave sensor can be fabricated with ferroelectric materials to achieve tunability over a broad frequency range. It has been reported that with a thin ferroelectric film placed between the top conductor layer and the dielectric material of a microstrip structure, and the proper DC bias scheme, tunable components above the Ku band can be fabricated. Components such as phase shifters, coupled line filters, and Lange couplers have been reported in the literature using this technique. In this wokr, simulated results from a full wave electromagnetic simulator are obtained to show the tunability of a matching netowrk typically used in the design of microwave amplifiers and antennas. In addition, simulated results of a multilayer Lange coupler, and a patch antenna are also presented. The results show that typical microstrip structures can be easily modified to provide frequency agile capabilities.

  4. Caffeine supplementation and reactive agility in elite youth soccer players.

    PubMed

    Jordan, J Bradley; Korgaokar, Ajit; Farley, Richard S; Coons, John M; Caputo, Jennifer L

    2014-05-01

    This study examined the effects of caffeine supplementation (6 mg·kg-1) on performance of a reactive agility test (RAT) in 17 elite, male, youth (M = 14 y) soccer players. Using a double-blind, repeated-measures design, players completed 4 days of testing on the RAT after a standardized warm-up. On day 1, anthropometric measurements were taken and players were accommodated to the RAT. On day 2, baseline performance was established. Caffeine or placebo conditions were randomly assigned on day 3 and the condition was reversed on day 4. Players completed 3 randomized trials of the RAT on days 2, 3, and 4 with at least 1 trial to the players' dominant and nondominant sides. There were no significant differences among conditions in reaction time (RT) to the dominant side, heart rates at any point of measurement, or ratings of perceived exertion (RPE) after completion of the warm-up. Caffeine produced faster RT to the nondominant side (P = .041) and higher RPE at the conclusion of the RAT (P = .013). The effect on the total time (TT) to complete the agility test to the nondominant side approached significance (P = .051). Sprint time and TT to either side did not differ. Caffeine supplementation may provide ergogenic benefit to elite, male, youth soccer players.

  5. MULTIWAVELENGTH OBSERVATIONS OF 3C 454.3. II. THE AGILE 2007 DECEMBER CAMPAIGN

    SciTech Connect

    Donnarumma, I.; Pucella, G.; Vittorini, V.; D'Ammando, F.; Tavani, M.; Chen, W. P.; Kataoka, J.; Kawai, N.; Mori, Y.; Tosti, G.; Impiombato, D.; Takahashi, T.; Sato, R.

    2009-12-20

    We report on the second Astrorivelatore Gamma a Immagini Leggero (AGILE) multiwavelength campaign of the blazar 3C 454.3 during the first half of 2007 December. This campaign involved AGILE, Spitzer, Swift, Suzaku, the Whole Earth Blazar Telescope (WEBT) consortium, the Rapid Eye Mount (REM), and the Multicolor Imaging Telescopes for Survey and Monstrous Explosions (MITSuME) telescopes, offering a broadband coverage that allowed for a simultaneous sampling of the synchrotron and inverse Compton (IC) emissions. The two-week AGILE monitoring was accompanied by radio to optical monitoring by WEBT and REM, and by sparse observations in mid-infrared and soft/hard X-ray energy bands performed by means of Target of Opportunity observations by Spitzer, Swift, and Suzaku, respectively. The source was detected with an average flux of approx250 x 10{sup -8} photons cm{sup -2} s{sup -1} above 100 MeV, typical of its flaring states. The simultaneous optical and gamma-ray monitoring allowed us to study the time lag associated with the variability in the two energy bands, resulting in a possible approxmodel with an external Compton on seed photons by a standard disk and reprocessed by the broad-line regions does not describe in a satisfactory way the SEDs of 2007 December 5, 13, and 15. An additional contribution, possibly from the hot corona with T = 10{sup 6} K surrounding the jet, is required to account simultaneously for the softness of the synchrotron and the hardness of the IC emissions during those epochs.

  6. Maintaining Compliance in Customizable Process Models

    NASA Astrophysics Data System (ADS)

    Schleicher, Daniel; Anstett, Tobias; Leymann, Frank; Mietzner, Ralph

    Compliance of business processes has gained importance during the last years. The growing number of internal and external regulations that companies need to obey has led to this state. This paper presents a practical concept of ensuring compliance during design time of customizable business processes.

  7. Computer modeling of lung cancer diagnosis-to-treatment process

    PubMed Central

    Ju, Feng; Lee, Hyo Kyung; Osarogiagbon, Raymond U.; Yu, Xinhua; Faris, Nick

    2015-01-01

    We introduce an example of a rigorous, quantitative method for quality improvement in lung cancer care-delivery. Computer process modeling methods are introduced for lung cancer diagnosis, staging and treatment selection process. Two types of process modeling techniques, discrete event simulation (DES) and analytical models, are briefly reviewed. Recent developments in DES are outlined and the necessary data and procedures to develop a DES model for lung cancer diagnosis, leading up to surgical treatment process are summarized. The analytical models include both Markov chain model and closed formulas. The Markov chain models with its application in healthcare are introduced and the approach to derive a lung cancer diagnosis process model is presented. Similarly, the procedure to derive closed formulas evaluating the diagnosis process performance is outlined. Finally, the pros and cons of these methods are discussed. PMID:26380181

  8. Verification of image processing based visibility models

    SciTech Connect

    Larson, S.M.; Cass, G.R.; Hussey, K.J.; Luce, F.

    1988-06-01

    Methods are presented for testing visibility models that use simulated photographs to display results of model calculations. An experimental protocol is developed and used to obtain input data including standard photographs of chosen scenes on a clear day and during a smog event at Pasadena, CA. With clear day photograph as a substrate, pollutant properties measured on the smoggy day are introduced into the visibility model, and results of the model calculations are displayed as a synthetic photograph of the expected appearance of the smog event. Quantitative comparisons are made between the predicted and actual appearance of the smog event. Diagnostic techniques developed are applied to the visibility modeling procedure proposed by Malm et al. That model is shown to reproduce the contrast reduction characteristic of urban air pollution but produces synthetic photographs with sky elements that differ substantially from a real photograph of the actual smog event.

  9. Modelling the Active Hearing Process in Mosquitoes

    NASA Astrophysics Data System (ADS)

    Avitabile, Daniele; Homer, Martin; Jackson, Joe; Robert, Daniel; Champneys, Alan

    2011-11-01

    A simple microscopic mechanistic model is described of the active amplification within the Johnston's organ of the mosquito species Toxorhynchites brevipalpis. The model is based on the description of the antenna as a forced-damped oscillator coupled to a set of active threads (ensembles of scolopidia) that provide an impulsive force when they twitch. This twitching is in turn controlled by channels that are opened and closed if the antennal oscillation reaches a critical amplitude. The model matches both qualitatively and quantitatively with recent experiments. New results are presented using mathematical homogenization techniques to derive a mesoscopic model as a simple oscillator with nonlinear force and damping characteristics. It is shown how the results from this new model closely resemble those from the microscopic model as the number of threads approach physiologically correct values.

  10. Modeling of Heating During Food Processing

    NASA Astrophysics Data System (ADS)

    Zheleva, Ivanka; Kamburova, Veselka

    Heat transfer processes are important for almost all aspects of food preparation and play a key role in determining food safety. Whether it is cooking, baking, boiling, frying, grilling, blanching, drying, sterilizing, or freezing, heat transfer is part of the processing of almost every food. Heat transfer is a dynamic process in which thermal energy is transferred from one body with higher temperature to another body with lower temperature. Temperature difference between the source of heat and the receiver of heat is the driving force in heat transfer.

  11. Performance, Agility and Cost of Cloud Computing Services for NASA GES DISC Giovanni Application

    NASA Astrophysics Data System (ADS)

    Pham, L.; Chen, A.; Wharton, S.; Winter, E. L.; Lynnes, C.

    2013-12-01

    The NASA Goddard Earth Science Data and Information Services Center (GES DISC) is investigating the performance, agility and cost of Cloud computing for GES DISC applications. Giovanni (Geospatial Interactive Online Visualization ANd aNalysis Infrastructure), one of the core applications at the GES DISC for online climate-related Earth science data access, subsetting, analysis, visualization, and downloading, was used to evaluate the feasibility and effort of porting an application to the Amazon Cloud Services platform. The performance and the cost of running Giovanni on the Amazon Cloud were compared to similar parameters for the GES DISC local operational system. A Giovanni Time-Series analysis of aerosol absorption optical depth (388nm) from OMI (Ozone Monitoring Instrument)/Aura was selected for these comparisons. All required data were pre-cached in both the Cloud and local system to avoid data transfer delays. The 3-, 6-, 12-, and 24-month data were used for analysis on the Cloud and local system respectively, and the processing times for the analysis were used to evaluate system performance. To investigate application agility, Giovanni was installed and tested on multiple Cloud platforms. The cost of using a Cloud computing platform mainly consists of: computing, storage, data requests, and data transfer in/out. The Cloud computing cost is calculated based on the hourly rate, and the storage cost is calculated based on the rate of Gigabytes per month. Cost for incoming data transfer is free, and for data transfer out, the cost is based on the rate in Gigabytes. The costs for a local server system consist of buying hardware/software, system maintenance/updating, and operating cost. The results showed that the Cloud platform had a 38% better performance and cost 36% less than the local system. This investigation shows the potential of cloud computing to increase system performance and lower the overall cost of system management.

  12. The frequency-agile radar: A multifunctional approach to remote sensing of the ionosphere

    NASA Astrophysics Data System (ADS)

    Tsunoda, R. T.; Livingston, R. C.; Buonocore, J. J.; McKinley, A. V.

    1995-09-01

    We introduce a new kind of diagnostic sensor that combines multifunctional measurement capabilities for ionospheric research. Multifunctionality is realized through agility in frequency selection over an extended band (1.5 to 50 MHz), system modularity, complete system control by software written in C, and a user-friendly computer interface. This sensor, which we call the frequency-agile radar (FAR), incorporates dual radar channels and an arbitrary waveform synthesizer that allows creative design of sophisticated waveforms as a means of increasing its sensitivity to weak signals while minimizing loss in radar resolution. The sensitivity of the FAR is determined by two sets of power amplifier modules: four 4-kW solid-state broadband amplifiers, and four 30-kW vacuum tube amplifiers. FAR control is by an AT-bus personal computer with on-line processing by a programmable array processor. The FAR does not simply house the separate functions of most radio sensors in use today, it provides convenient and flexible access to those functions as elements to be used in any combination. Some of the first new results obtained with the FAR during recent field campaigns are presented to illustrate its versatility. These include (1) the first detection of anomalous high-frequency (HF) reflections from a barium ion cloud, (2) the first evidence of unexpectedly large drifts and a shear north of the equatorial electrojet, (3) the first HF radar signature of a developing equatorial plasma bubble, and (4) the first measurements by a portable radar of altitude-extended, quasi-periodic backscatter from midlatitude sporadic E. We also mention the potential of the FAR for atmospheric remote sensing.

  13. Modeling fixation locations using spatial point processes.

    PubMed

    Barthelmé, Simon; Trukenbrod, Hans; Engbert, Ralf; Wichmann, Felix

    2013-10-01

    Whenever eye movements are measured, a central part of the analysis has to do with where subjects fixate and why they fixated where they fixated. To a first approximation, a set of fixations can be viewed as a set of points in space; this implies that fixations are spatial data and that the analysis of fixation locations can be beneficially thought of as a spatial statistics problem. We argue that thinking of fixation locations as arising from point processes is a very fruitful framework for eye-movement data, helping turn qualitative questions into quantitative ones. We provide a tutorial introduction to some of the main ideas of the field of spatial statistics, focusing especially on spatial Poisson processes. We show how point processes help relate image properties to fixation locations. In particular we show how point processes naturally express the idea that image features' predictability for fixations may vary from one image to another. We review other methods of analysis used in the literature, show how they relate to point process theory, and argue that thinking in terms of point processes substantially extends the range of analyses that can be performed and clarify their interpretation.

  14. Capability Maturity Model (CMM) for Software Process Improvements

    NASA Technical Reports Server (NTRS)

    Ling, Robert Y.

    2000-01-01

    This slide presentation reviews the Avionic Systems Division's implementation of the Capability Maturity Model (CMM) for improvements in the software development process. The presentation reviews the process involved in implementing the model and the benefits of using CMM to improve the software development process.

  15. Toward a Model for Picture and Word Processing.

    ERIC Educational Resources Information Center

    Snodgrass, Joan Gay

    A model was developed to account for similarities and differences between picture and word processing in a variety of semantic and episodic memory tasks. The model contains three levels of processing: low-level processing of the physical characteristics of externally presented pictures and words; an intermediate level where the low-level processor…

  16. Dynamic tumor tracking using the Elekta Agility MLC

    SciTech Connect

    Fast, Martin F. Nill, Simeon Bedford, James L.; Oelfke, Uwe

    2014-11-01

    Purpose: To evaluate the performance of the Elekta Agility multileaf collimator (MLC) for dynamic real-time tumor tracking. Methods: The authors have developed a new control software which interfaces to the Agility MLC to dynamically program the movement of individual leaves, the dynamic leaf guides (DLGs), and the Y collimators (“jaws”) based on the actual target trajectory. A motion platform was used to perform dynamic tracking experiments with sinusoidal trajectories. The actual target positions reported by the motion platform at 20, 30, or 40 Hz were used as shift vectors for the MLC in beams-eye-view. The system latency of the MLC (i.e., the average latency comprising target device reporting latencies and MLC adjustment latency) and the geometric tracking accuracy were extracted from a sequence of MV portal images acquired during irradiation for the following treatment scenarios: leaf-only motion, jaw + leaf motion, and DLG + leaf motion. Results: The portal imager measurements indicated a clear dependence of the system latency on the target position reporting frequency. Deducting the effect of the target frequency, the leaf adjustment latency was measured to be 38 ± 3 ms for a maximum target speed v of 13 mm/s. The jaw + leaf adjustment latency was 53 ± 3 at a similar speed. The system latency at a target position frequency of 30 Hz was in the range of 56–61 ms for the leaves (v ≤ 31 mm/s), 71–78 ms for the jaw + leaf motion (v ≤ 25 mm/s), and 58–72 ms for the DLG + leaf motion (v ≤ 59 mm/s). The tracking accuracy showed a similar dependency on the target position frequency and the maximum target speed. For the leaves, the root-mean-squared error (RMSE) was between 0.6–1.5 mm depending on the maximum target speed. For the jaw + leaf (DLG + leaf) motion, the RMSE was between 0.7–1.5 mm (1.9–3.4 mm). Conclusions: The authors have measured the latency and geometric accuracy of the Agility MLC, facilitating its future use for clinical

  17. Biomedical Simulation Models of Human Auditory Processes

    NASA Technical Reports Server (NTRS)

    Bicak, Mehmet M. A.

    2012-01-01

    Detailed acoustic engineering models that explore noise propagation mechanisms associated with noise attenuation and transmission paths created when using hearing protectors such as earplugs and headsets in high noise environments. Biomedical finite element (FE) models are developed based on volume Computed Tomography scan data which provides explicit external ear, ear canal, middle ear ossicular bones and cochlea geometry. Results from these studies have enabled a greater understanding of hearing protector to flesh dynamics as well as prioritizing noise propagation mechanisms. Prioritization of noise mechanisms can form an essential framework for exploration of new design principles and methods in both earplug and earcup applications. These models are currently being used in development of a novel hearing protection evaluation system that can provide experimentally correlated psychoacoustic noise attenuation. Moreover, these FE models can be used to simulate the effects of blast related impulse noise on human auditory mechanisms and brain tissue.

  18. Modeling of dynamical processes in laser ablation

    SciTech Connect

    Leboeuf, J.N.; Chen, K.R.; Donato, J.M.; Geohegan, D.B.; Liu, C.L.; Puretzky, A.A.; Wood, R.F.

    1995-12-31

    Various physics and computational approaches have been developed to globally characterize phenomena important for film growth by pulsed-laser deposition of materials. These include thermal models of laser-solid target interactions that initiate the vapor plume, plume ionization and heating through laser absorption beyond local thermodynamic equilibrium mechanisms, hydrodynamic and collisional descriptions of plume transport, and molecular dynamics models of the interaction of plume particles with the deposition substrate.

  19. A Model for the Composing Process. Occasional Paper No. 3.

    ERIC Educational Resources Information Center

    Myers, Miles

    A synthesis of three different writing theories into one model for the writing process is given in this report and evidence is provided for various parts of the model from studies using different methods of investigation. The model developed contains three phases: (1) processing or whiting as discovery, including propositionalizing, subchunking,…

  20. Computational models of cortical visual processing.

    PubMed Central

    Heeger, D J; Simoncelli, E P; Movshon, J A

    1996-01-01

    The visual responses of neurons in the cerebral cortex were first adequately characterized in the 1960s by D. H. Hubel and T. N. Wiesel [(1962) J. Physiol. (London) 160, 106-154; (1968) J. Physiol. (London) 195, 215-243] using qualitative analyses based on simple geometric visual targets. Over the past 30 years, it has become common to consider the properties of these neurons by attempting to make formal descriptions of these transformations they execute on the visual image. Most such models have their roots in linear-systems approaches pioneered in the retina by C. Enroth-Cugell and J. R. Robson [(1966) J. Physiol. (London) 187, 517-552], but it is clear that purely linear models of cortical neurons are inadequate. We present two related models: one designed to account for the responses of simple cells in primary visual cortex (V1) and one designed to account for the responses of pattern direction selective cells in MT (or V5), an extrastriate visual area thought to be involved in the analysis of visual motion. These models share a common structure that operates in the same way on different kinds of input, and instantiate the widely held view that computational strategies are similar throughout the cerebral cortex. Implementations of these models for Macintosh microcomputers are available and can be used to explore the models' properties. PMID:8570605

  1. A Comprehensive and Harmonized Digital Forensic Investigation Process Model.

    PubMed

    Valjarevic, Aleksandar; Venter, Hein S

    2015-11-01

    Performing a digital forensic investigation (DFI) requires a standardized and formalized process. There is currently neither an international standard nor does a global, harmonized DFI process (DFIP) exist. The authors studied existing state-of-the-art DFIP models and concluded that there are significant disparities pertaining to the number of processes, the scope, the hierarchical levels, and concepts applied. This paper proposes a comprehensive model that harmonizes existing models. An effort was made to incorporate all types of processes proposed by the existing models, including those aimed at achieving digital forensic readiness. The authors introduce a novel class of processes called concurrent processes. This is a novel contribution that should, together with the rest of the model, enable more efficient and effective DFI, while ensuring admissibility of digital evidence. Ultimately, the proposed model is intended to be used for different types of DFI and should lead to standardization.

  2. A Comprehensive and Harmonized Digital Forensic Investigation Process Model.

    PubMed

    Valjarevic, Aleksandar; Venter, Hein S

    2015-11-01

    Performing a digital forensic investigation (DFI) requires a standardized and formalized process. There is currently neither an international standard nor does a global, harmonized DFI process (DFIP) exist. The authors studied existing state-of-the-art DFIP models and concluded that there are significant disparities pertaining to the number of processes, the scope, the hierarchical levels, and concepts applied. This paper proposes a comprehensive model that harmonizes existing models. An effort was made to incorporate all types of processes proposed by the existing models, including those aimed at achieving digital forensic readiness. The authors introduce a novel class of processes called concurrent processes. This is a novel contribution that should, together with the rest of the model, enable more efficient and effective DFI, while ensuring admissibility of digital evidence. Ultimately, the proposed model is intended to be used for different types of DFI and should lead to standardization. PMID:26258644

  3. Statistical properties of several models of fractional random point processes

    NASA Astrophysics Data System (ADS)

    Bendjaballah, C.

    2011-08-01

    Statistical properties of several models of fractional random point processes have been analyzed from the counting and time interval statistics points of view. Based on the criterion of the reduced variance, it is seen that such processes exhibit nonclassical properties. The conditions for these processes to be treated as conditional Poisson processes are examined. Numerical simulations illustrate part of the theoretical calculations.

  4. Enhanced detection of terrestrial gamma-ray flashes by AGILE

    NASA Astrophysics Data System (ADS)

    Marisaldi, M.; Argan, A.; Ursi, A.; Gjesteland, T.; Fuschino, F.; Labanti, C.; Galli, M.; Tavani, M.; Pittori, C.; Verrecchia, F.; D'Amico, F.; Østgaard, N.; Mereghetti, S.; Campana, R.; Cattaneo, P. W.; Bulgarelli, A.; Colafrancesco, S.; Dietrich, S.; Longo, F.; Gianotti, F.; Giommi, P.; Rappoldi, A.; Trifoglio, M.; Trois, A.

    2015-11-01

    At the end of March 2015 the onboard software configuration of the Astrorivelatore Gamma a Immagini Leggero (AGILE) satellite was modified in order to disable the veto signal of the anticoincidence shield for the minicalorimeter instrument. The motivation for such a change was the understanding that the dead time induced by the anticoincidence prevented the detection of a large fraction of Terrestrial Gamma-Ray Flashes (TGFs). The configuration change was highly successful resulting in an increase of one order of magnitude in TGF detection rate. As expected, the largest fraction of the new events has short duration (<100 μs), and part of them has simultaneous association with lightning sferics detected by the World Wide Lightning Location Network. The new configuration provides the largest TGF detection rate surface density (TGFs/km2/yr) to date, opening prospects for improved correlation studies with lightning and atmospheric parameters on short spatial and temporal scales along the equatorial region.

  5. From vision to reality: strategic agility in complex times.

    PubMed

    Soule, Barbara M

    2002-04-01

    Health care is experiencing turbulent times. Change has become the constant. Complexity and sometimes chaos are common characteristics. Within this context, infection control professionals strive to maintain their practices, achieve excellence, and plan for the future. As demands shift and expectations increase, professionals in infection surveillance, prevention, and control (ISPC) programs must develop strategic agility. This article describes the rationale for strategic thinking and action set within a framework of 6 thought-provoking questions. It also describes a number of techniques to use for thinking strategically, such as designing visions, becoming entrepreneurial, and engaging in creative and futuristic exercises to evaluate possibilities for program direction. These techniques can guide individual professionals or ISPC programs in strategic decision-making that will increase the ability to survive and succeed in the future.

  6. Agile and dexterous robot for inspection and EOD operations

    NASA Astrophysics Data System (ADS)

    Handelman, David A.; Franken, Gordon H.; Komsuoglu, Haldun

    2010-04-01

    The All-Terrain Biped (ATB) robot is an unmanned ground vehicle with arms, legs and wheels designed to drive, crawl, walk and manipulate objects for inspection and explosive ordnance disposal tasks. This paper summarizes on-going development of the ATB platform. Control technology for semi-autonomous legged mobility and dual-arm dexterity is described as well as preliminary simulation and hardware test results. Performance goals include driving on flat terrain, crawling on steep terrain, walking on stairs, opening doors and grasping objects. Anticipated benefits of the adaptive mobility and dexterity of the ATB platform include increased robot agility and autonomy for EOD operations, reduced operator workload and reduced operator training and skill requirements.

  7. Frequency Agile Transceiver for Advanced Vehicle Data Links

    NASA Technical Reports Server (NTRS)

    Freudinger, Lawrence C.; Macias, Filiberto; Cornelius, Harold

    2009-01-01

    Emerging and next-generation test instrumentation increasingly relies on network communication to manage complex and dynamic test scenarios, particularly for uninhabited autonomous systems. Adapting wireless communication infrastructure to accommodate challenging testing needs can benefit from reconfigurable radio technology. Frequency agility is one characteristic of reconfigurable radios that to date has seen only limited progress toward programmability. This paper overviews an ongoing project to validate a promising chipset that performs conversion of RF signals directly into digital data for the wireless receiver and, for the transmitter, converts digital data into RF signals. The Software Configurable Multichannel Transceiver (SCMT) enables four transmitters and four receivers in a single unit, programmable for any frequency band between 1 MHz and 6 GHz.

  8. Frequency-agile bandpass filter for direct detection lidar receivers.

    PubMed

    Gittins, C M; Lawrence, W G; Marinelli, W J

    1998-12-20

    We discuss the development of a frequency-agile receiver for CO(2) laser-based differential absorption lidar (DIAL) systems. The receiver is based on the insertion of a low-order tunable etalon into the detector field of view. The incorporation of the etalon into the receiver reduces system noise by decreasing the instantaneous spectral bandwidth of the IR detector to a narrow wavelength range centered on the transmitted CO(2) laser line, thereby improving the overall D* of the detection system. A consideration of overall lidar system performance results in a projected factor of a 2-7 reduction in detector system noise, depending on the characteristics of the environment being probed. These improvements can play a key role in extending the ability of DIAL systems to monitor chemical releases from long standoff distances.

  9. A new algorithm for agile satellite-based acquisition operations

    NASA Astrophysics Data System (ADS)

    Bunkheila, Federico; Ortore, Emiliano; Circi, Christian

    2016-06-01

    Taking advantage of the high manoeuvrability and the accurate pointing of the so-called agile satellites, an algorithm which allows efficient management of the operations concerning optical acquisitions is described. Fundamentally, this algorithm can be subdivided into two parts: in the first one the algorithm operates a geometric classification of the areas of interest and a partitioning of these areas into stripes which develop along the optimal scan directions; in the second one it computes the succession of the time windows in which the acquisition operations of the areas of interest are feasible, taking into consideration the potential restrictions associated with these operations and with the geometric and stereoscopic constraints. The results and the performances of the proposed algorithm have been determined and discussed considering the case of the Periodic Sun-Synchronous Orbits.

  10. Muscle directly meets the vast power demands in agile lizards

    PubMed Central

    Curtin, Nancy A; Woledge, Roger C; Aerts, Peter

    2005-01-01

    Level locomotion in small, agile lizards is characterized by intermittent bursts of fast running. These require very large accelerations, often reaching several times g. The power input required to increase kinetic energy is calculated to be as high as 214 W kg−1 muscle (±20 W kg−1 s.e.; averaged over the complete locomotor cycle) and 952 W kg−1 muscle (±89 W kg−1 s.e.; instantaneous peak power). In vitro muscle experiments prove that these exceptional power requirements can be met directly by the lizard's muscle fibres alone; there is no need for mechanical power amplifying mechanisms. PMID:15817432

  11. AGILE follow-up of the neutrino ICECUBE-160731 event

    NASA Astrophysics Data System (ADS)

    Lucarelli, F.; Pittori, C.; Verrecchia, F.; Piano, G.; Munar-Adrover, P.; Bulgarelli, A.; Fioretti, V.; Zoli, A.; Tavani, M.; Donnarumma, I.; Vercellone, S.; Minervini, G.; Striani, E.; Cardillo, M.; Gianotti, F.; Trifoglio, M.; Giuliani, A.; Mereghetti, S.; Caraveo, P.; Perotti, F.; Chen, A.; Argan, A.; Costa, E.; Del Monte, E.; Evangelista, Y.; Feroci, M.; Lazzarotto, F.; Lapshov, I.; Pacciani, L.; Soffitta, P.; Sabatini, S.; Vittorini, V.; Pucella, G.; Rapisarda, M.; Di Cocco, G.; Fuschino, F.; Galli, M.; Labanti, C.; Marisaldi, M.; Pellizzoni, A.; Pilia, M.; Trois, A.; Barbiellini, G.; Vallazza, E.; Longo, F.; Morselli, A.; Picozza, P.; Prest, M.; Lipari, P.; Zanello, D.; Cattaneo, P. W.; Rappoldi, A.; Colafrancesco, S.; Parmiggiani, N.; Ferrari, A.; Antonelli, A.; Giommi, P.; Salotti, L.; Valentini, G.; D'Amico, F.

    2016-08-01

    Following the GCN notice posted by the ICECUBE Collaboration on July 31, 2016, reporting the detection at T0=16/07/31 01:55:04 UT of a very high energy neutrino with reconstructed arrival direction pointing at RA, DEC (J2000)=(214.5440, -0.3347 [deg]) with a 90% containement radius of 45.00 arcmin (stat+sys), we searched for transient gamma-ray emission in the AGILE data above 100 MeV. Integrating over the 48 hours from 2016-07-29 02:00 UT to 2016-07-31 02:00 UT a maximum likelihood analysis yields a possible detection at a significance level of about 3 sigma with a flux F(E > 100 MeV)=(1.5 +/- 0.7)x 10^-06 ph/cm^2/s within the GCN/AMON_ICECUBE_HESE notice error region.

  12. LISP based simulation generators for modeling complex space processes

    NASA Technical Reports Server (NTRS)

    Tseng, Fan T.; Schroer, Bernard J.; Dwan, Wen-Shing

    1987-01-01

    The development of a simulation assistant for modeling discrete event processes is presented. Included are an overview of the system, a description of the simulation generators, and a sample process generated using the simulation assistant.

  13. Sensory processing and world modeling for an active ranging device

    NASA Technical Reports Server (NTRS)

    Hong, Tsai-Hong; Wu, Angela Y.

    1991-01-01

    In this project, we studied world modeling and sensory processing for laser range data. World Model data representation and operation were defined. Sensory processing algorithms for point processing and linear feature detection were designed and implemented. The interface between world modeling and sensory processing in the Servo and Primitive levels was investigated and implemented. In the primitive level, linear features detectors for edges were also implemented, analyzed and compared. The existing world model representations is surveyed. Also presented is the design and implementation of the Y-frame model, a hierarchical world model. The interfaces between the world model module and the sensory processing module are discussed as well as the linear feature detectors that were designed and implemented.

  14. Process models: analytical tools for managing industrial energy systems

    SciTech Connect

    Howe, S O; Pilati, D A; Balzer, C; Sparrow, F T

    1980-01-01

    How the process models developed at BNL are used to analyze industrial energy systems is described and illustrated. Following a brief overview of the industry modeling program, the general methodology of process modeling is discussed. The discussion highlights the important concepts, contents, inputs, and outputs of a typical process model. A model of the US pulp and paper industry is then discussed as a specific application of process modeling methodology. Applications addressed with the case study results include projections of energy demand, conservation technology assessment, energy-related tax policies, and sensitivity analysis. A subsequent discussion of these results supports the conclusion that industry process models are versatile and powerful tools for managing industrial energy systems.

  15. A Review on Mathematical Modeling for Textile Processes

    NASA Astrophysics Data System (ADS)

    Chattopadhyay, R.

    2015-10-01

    Mathematical model is a powerful tool in engineering for studying variety of problems related to design and development of products and processes, optimization of manufacturing process, understanding a phenomenon and predicting product's behaviour in actual use. An insight of the process and use of appropriate mathematical tools are necessary for developing models. In the present paper, a review of types of model, procedure followed in developing them and their limitations have been discussed. Modeling techniques being used in few textile processes available in the literature have been cited as examples.

  16. Survey of Bayesian Models for Modelling of Stochastic Temporal Processes

    SciTech Connect

    Ng, B

    2006-10-12

    This survey gives an overview of popular generative models used in the modeling of stochastic temporal systems. In particular, this survey is organized into two parts. The first part discusses the discrete-time representations of dynamic Bayesian networks and dynamic relational probabilistic models, while the second part discusses the continuous-time representation of continuous-time Bayesian networks.

  17. Physics and Process Modeling (PPM) and Other Propulsion R and T. Volume 1; Materials Processing, Characterization, and Modeling; Lifting Models

    NASA Technical Reports Server (NTRS)

    1997-01-01

    This CP contains the extended abstracts and presentation figures of 36 papers presented at the PPM and Other Propulsion R&T Conference. The focus of the research described in these presentations is on materials and structures technologies that are parts of the various projects within the NASA Aeronautics Propulsion Systems Research and Technology Base Program. These projects include Physics and Process Modeling; Smart, Green Engine; Fast, Quiet Engine; High Temperature Engine Materials Program; and Hybrid Hyperspeed Propulsion. Also presented were research results from the Rotorcraft Systems Program and work supported by the NASA Lewis Director's Discretionary Fund. Authors from NASA Lewis Research Center, industry, and universities conducted research in the following areas: material processing, material characterization, modeling, life, applied life models, design techniques, vibration control, mechanical components, and tribology. Key issues, research accomplishments, and future directions are summarized in this publication.

  18. Verifying and Validating Proposed Models for FSW Process Optimization

    NASA Technical Reports Server (NTRS)

    Schneider, Judith

    2008-01-01

    This slide presentation reviews Friction Stir Welding (FSW) and the attempts to model the process in order to optimize and improve the process. The studies are ongoing to validate and refine the model of metal flow in the FSW process. There are slides showing the conventional FSW process, a couple of weld tool designs and how the design interacts with the metal flow path. The two basic components of the weld tool are shown, along with geometries of the shoulder design. Modeling of the FSW process is reviewed. Other topics include (1) Microstructure features, (2) Flow Streamlines, (3) Steady-state Nature, and (4) Grain Refinement Mechanisms

  19. Agent-Based Modeling of Growth Processes

    ERIC Educational Resources Information Center

    Abraham, Ralph

    2014-01-01

    Growth processes abound in nature, and are frequently the target of modeling exercises in the sciences. In this article we illustrate an agent-based approach to modeling, in the case of a single example from the social sciences: bullying.

  20. A comparison of linear speed, closed-skill agility, and open-skill agility qualities between backcourt and frontcourt adult semiprofessional male basketball players.

    PubMed

    Scanlan, Aaron T; Tucker, Patrick S; Dalbo, Vincent J

    2014-05-01

    The measurement of fitness qualities relevant to playing position is necessary to inform basketball coaching and conditioning staff of role-related differences in playing groups. To date, sprinting and agility performance have not been compared between playing positions in adult male basketball players. Therefore, the purpose of this study was to describe and compare linear speed, closed-skill agility, and open-skill agility qualities between backcourt (point guard and shooting guard positions) and frontcourt (small forward, power forward, and center positions) semiprofessional basketball players. Six backcourt (mean ± SD: age, 24.3 ± 7.9 years; stature, 183.4 ± 4.0 cm; body mass, 85.5 ± 12.3 kg; VO2max, 51.9 ± 4.8 ml·kg(-1)·min(-1)) and 6 frontcourt (mean ± SD: age, 27.5 ± 5.5 years; stature, 194.4 ± 7.1 cm; body mass, 109.4 ± 8.8 kg; VO2max, 47.1 ± 5.0 ml·kg(-1)·min(-1)) adult male basketball players completed 20-m sprint, closed-skill agility, and open-skill agility performance tests. Magnitude-based inferences revealed that backcourt players (5 m, 1.048 ± 0.027 seconds; 10 m, 1.778 ± 0.048 seconds; 20 m, 3.075 ± 0.121 seconds) possessed likely quicker linear sprint times than frontcourt players (5 m, 1.095 ± 0.085 seconds; 10 m, 1.872 ± 0.127 seconds; 20 m, 3.242 ± 0.221 seconds). Conversely, frontcourt players (1.665 ± 0.096 seconds) held possible superior closed-skill agility performance than backcourt players (1.613 ± 0.111 seconds). In addition, unclear positional differences were apparent for open-skill agility qualities. These findings indicate that linear speed and change of direction speed might be differently developed across playing positions. Furthermore, position-related functions might similarly depend on the aspects of open-skill agility performance across backcourt and frontcourt players. Basketball coaching and conditioning staff should consider the development of position-targeted training drills to improve speed, agility