Cooper, Jason G.; Pauley, Keith A.
Software assurance is a rigorous, lifecycle phase-independent set of activities which ensure completeness, safety, and reliability of software processes and products. This is accomplished by guaranteeing conformance to all requirements, standards, procedures, and regulations. These assurance processes are even more important when coupled with healthcare software systems, embedded software in medical instrumentation, and other healthcare-oriented life-critical systems. The current Food and Drug Administration (FDA) regulatory requirements and guidance documentation do not address certain aspects of complete software assurance activities. In addition, the FDA’s software oversight processes require enhancement to include increasingly complex healthcare systems such as Hospital Information Systems (HIS). The importance of complete software assurance is introduced, current regulatory requirements and guidance discussed, and the necessity for enhancements to the current processes shall be highlighted. PMID:17238324
Cooper, Jason G; Pauley, Keith A
Software assurance is a rigorous, lifecycle phase-independent set of activities which ensure completeness, safety, and reliability of software processes and products. This is accomplished by guaranteeing conformance to all requirements, standards, procedures, and regulations. These assurance processes are even more important when coupled with healthcare software systems, embedded software in medical instrumentation, and other healthcare-oriented life-critical systems. The current Food and Drug Administration (FDA) regulatory requirements and guidance documentation do not address certain aspects of complete software assurance activities. In addition, the FDA's software oversight processes require enhancement to include increasingly complex healthcare systems such as Hospital Information Systems (HIS). The importance of complete software assurance is introduced, current regulatory requirements and guidance discussed, and the necessity for enhancements to the current processes shall be highlighted.
This standard specifies the software assurance program for the provider of software. It also delineates the assurance activities for the provider and the assurance data that are to be furnished by the provider to the acquirer. In any software development effort, the provider is the entity or individual that actually designs, develops, and implements the software product, while the acquirer is the entity or individual who specifies the requirements and accepts the resulting products. This standard specifies at a high level an overall software assurance program for software developed for and by NASA. Assurance includes the disciplines of quality assurance, quality engineering, verification and validation, nonconformance reporting and corrective action, safety assurance, and security assurance. The application of these disciplines during a software development life cycle is called software assurance. Subsequent lower-level standards will specify the specific processes within these disciplines.
McRae, Kalindra A.
Software Quality Assurance (SQA) is a planned and systematic set of activities that ensures conformance of software life cycle processes and products conform to requirements, standards and procedures. In software development, software quality means meeting requirements and a degree of excellence and refinement of a project or product. Software Quality is a set of attributes of a software product by which its quality is described and evaluated. The set of attributes includes functionality, reliability, usability, efficiency, maintainability, and portability. Software Metrics help us understand the technical process that is used to develop a product. The process is measured to improve it and the product is measured to increase quality throughout the life cycle of software. Software Metrics are measurements of the quality of software. Software is measured to indicate the quality of the product, to assess the productivity of the people who produce the product, to assess the benefits derived from new software engineering methods and tools, to form a baseline for estimation, and to help justify requests for new tools or additional training. Any part of the software development can be measured. If Software Metrics are implemented in software development, it can save time, money, and allow the organization to identify the caused of defects which have the greatest effect on software development. The summer of 2004, I worked with Cynthia Calhoun and Frank Robinson in the Software Assurance/Risk Management department. My task was to research and collect, compile, and analyze SQA Metrics that have been used in other projects that are not currently being used by the SA team and report them to the Software Assurance team to see if any metrics can be implemented in their software assurance life cycle process.
technologies and processes to achieve a required level of confidence that software systems and services function in the intended manner. 1.3 Security Example...that took three high-voltage lines out of service and a software fail- ure (a race condition3) that disabled the computing service that notified the... service had failed. Instead of analyzing the details of the alarm server failure, the reviewers asked why the following software assurance claim had
CONGRESSIONAL ) HIGH ASSURANCE SOFTWARE WILLIAM MAHONEY UNIVERSITY OF NEBRASKA 10/22/2013 Final Report DISTRIBUTION A: Distribution approved for ...0704-0188 Public reporting burden for this collection of information is estimated to average 1 hour per response, including the time for reviewing...Send comments regarding this burden estimate or any other aspect of this collection of information, including suggestions for reducing this burden to
The growth in cost and importance of software to NASA has caused NASA to address the improvement of software development across the agency. One of the products of this program is a series of guidebooks that define a NASA concept of the assurance processes that are used in software development. The Software Assurance Guidebook, NASA-GB-A201, issued in September, 1989, provides an overall picture of the NASA concepts and practices in software assurance. Second level guidebooks focus on specific activities that fall within the software assurance discipline, and provide more detailed information for the manager and/or practitioner. This is the second level Software Quality Assurance Audits Guidebook that describes software quality assurance audits in a way that is compatible with practices at NASA Centers.
NASA defines software assurance as: the planned and systematic set of activities that ensure conformance of software life cycle processes and products to requirements, standards, and procedures via quality, safety, reliability, and independent verification and validation. NASA's implementation of this approach to the quality, safety, reliability, security and verification and validation of software is brought together in one discipline, software assurance. Organizationally, NASA has software assurance at each NASA center, a Software Assurance Manager at NASA Headquarters, a Software Assurance Technical Fellow (currently the same person as the SA Manager), and an Independent Verification and Validation Organization with its own facility. An umbrella risk mitigation strategy for safety and mission success assurance of NASA's software, software assurance covers a wide area and is better structured to address the dynamic changes in how software is developed, used, and managed, as well as it's increasingly complex functionality. Being flexible, risk based, and prepared for challenges in software at NASA is essential, especially as much of our software is unique for each mission.
COTS) software , and software as a service ( SaaS ). L2: Define and analyze risks in the acquisition of contracted software , COTS software , and SaaS ...2010a]: Application of technologies and processes to achieve a required level of confidence that software systems and services function in the...
The customers satisfaction depends not only on functional performance, it also depends on the quality characteristics of the software products. An examination of this quality aspect of software products will provide a clear, well defined framework for quality assurance functions, which improve the life-cycle activities of software development. Software developers must be aware of the following aspects which have been expressed by many quality experts: quality cannot be added on; the level of quality built into a program is a function of the quality attributes employed during the development process; and finally, quality must be managed. These concepts have guided our development of the following definition for a Software Quality Assurance function: Software Quality Assurance is a formal, planned approach of actions designed to evaluate the degree of an identifiable set of quality attributes present in all software systems and their products. This paper is an explanation of how this definition was developed and how it is used.
product The chicken#. (a.k.a. Process Focused Assessment ) – Management Systems ( ISO 9001, ISO 27001 , ISO 2000) – Capability Maturity Models (CMMI...Assurance PRM, RMM, Assurance for CMMI)) – Lifecycle Processes ( ISO /IEEE 15288, ISO /IEEE 12207) – COBIT, ITIL, MS SDL, OSAMM, BSIMM 5 The egg...a.k.a Product Focused Assessments) – SCAP - NIST-SCAP – ISO /OMG W3C – KDM, BPMN, RIF, XMI, RDF – OWASP Top 10 – SANS TOP 25 – Secure Code Check Lists
Measure was removed: "Sufficient level of detail in the requirements to develop test cases." ; This control measure was removed since the sufficient level of detail needed to develop test cases is recorded for all test cases. (Note: This is mandatory for applications graded with a High Quality Assurance
Berg, Jonathan Charles
This document describes the software development practice areas and processes which contribute to the ability of SWiFT software developers to provide quality software. These processes are designed to satisfy the requirements set forth by the Sandia Software Quality Assurance Program (SSQAP). APPROVALS SWiFT Software Quality Assurance Plan (SAND2016-0765) approved by: Department Manager SWiFT Site Lead Dave Minster (6121) Date Jonathan White (6121) Date SWiFT Controls Engineer Jonathan Berg (6121) Date CHANGE HISTORY Issue Date Originator(s) Description A 2016/01/27 Jon Berg (06121) Initial release of the SWiFT Software Quality Assurance Plan
Programmable Logic (PLC, FPGA, ASIC) devices are hybrids - hardware devices that are designed and programmed like software. As such, they fall in an assurance gray area. Programmable Logic is usually tested and verified as hardware, and the software aspects are ignored, potentially leading to safety or mission success concerns. The objective of this proposal is to first determine where and how Programmable Logic (PL) is used within NASA and document the current methods of assurance. Once that is known, raise awareness of the PL software aspects within the NASA engineering community and provide guidance for the use and assurance of PL form a software perspective.
Shigeta, Denise; Port, Dan; Nikora, Allen P.; Wilf, Joel
As space mission software systems become larger and more complex, it is increasingly important for the software assurance effort to have the ability to effectively assess both the artifacts produced during software system development and the development process itself. Conceptually, assurance is a straightforward idea - it is the result of activities carried out by an organization independent of the software developers to better inform project management of potential technical and programmatic risks, and thus increase management's confidence in the decisions they ultimately make. In practice, effective assurance for large, complex systems often entails assessing large, complex software artifacts (e.g., requirements specifications, architectural descriptions) as well as substantial amounts of unstructured information (e.g., anomaly reports resulting from testing activities during development). In such an environment, assurance engineers can benefit greatly from appropriate tool support. In order to do so, an assurance organization will need accurate and timely information on the tool support available for various types of assurance activities. In this paper, we investigate the current use of tool support for assurance organizations within NASA, and describe on-going work at JPL for providing assurance organizations with the information about tools they need to use them effectively.
Duncan, Stephen E.; Bailey, Elizabeth K.
The software quality assurance (SQA) function for the Guidance and Control Software (GCS) project which is part of a software error studies research program is described. The SQA plan outlines all of the procedures, controls, and audits to be carried out by the SQA organization to ensure adherence to the policies, procedures, and standards for the GCS project.
56k Modem • Cisco 1OS 12.0 operation at the MTFs, the Team • Cisco 3640 Router Configuration Fileo 24- 56k Modems recommended the Cisco 3600 series...temporarily substituted by the vendor pending availability of ordered components (e.g., the modem circuit board for the router). " Each site created a...control for software / hardware modifications and upgrades? 3.5 Is there a process for introducing new equipment (such as hosts, printers, or modems
What is IV&V? Independent Verification and Validation (IV&V) is an objective examination of safety and mission critical software processes and products. Independence: 3 Key parameters: Technical Independence; Managerial Independence; Financial Independence. NASA IV&V perspectives: Will the system's software: Do what it is supposed to do?; Not do what it is not supposed to do?; Respond as expected under adverse conditions?. Systems Engineering: Determines if the right system has been built and that it has been built correctly. IV&V Technical Approaches: Aligned with IEEE 1012; Captured in a Catalog of Methods; Spans the full project lifecycle. IV&V Assurance Strategy: The IV&V Project's strategy for providing mission assurance; Assurance Strategy is driven by the specific needs of an individual project; Implemented via an Assurance Design; Communicated via Assurance Statements.
Feather, Martin; Cornford, Steven; Menzies, Tim
Software assurance is the planned and systematic set of activities that ensures that software processes and products conform to requirements, standards, and procedures. Examples of such activities are the following: code inspections, unit tests, design reviews, performance analyses, construction of traceability matrices, etc. In practice, software development projects have only limited resources (e.g., schedule, budget, and availability of personnel) to cover the entire development effort, of which assurance is but a part. Projects must therefore select judiciously from among the possible assurance activities. At its heart, this can be viewed as an optimization problem; namely, to determine the allocation of limited resources (time, money, and personnel) to minimize risk or, alternatively, to minimize the resources needed to reduce risk to an acceptable level. The end result of the work reported here is a means to optimize quality-assurance processes used in developing software.
Burns, Luanne; George, Richard; Linger, Richard C
Modern society is dependent on software systems of remarkable scope and complexity. Yet methods for assuring their security and functionality have not kept pace. The result is persistent compromises and failures despite best efforts. Cybersecurity methods must work together for situational awareness, attack prevention and detection, threat attribution, minimization of consequences, and attack recovery. Because defective software cannot be secure, assurance technologies must play a central role in cybersecurity approaches. There is increasing recognition of the need for rigorous methods for cybersecurity and software assurance. The goal of this minitrack is to develop science foundations, technologies, and practices that canmore » improve the security and dependability of complex systems.« less
del Real Núñez, R; Contreras Puertas, P I; Moreno Ortega, E; Mena Bares, L M; Maza Muret, F R; Latre Romero, J M
The need for quality assurance of all technical aspects of nuclear medicine studies is widely recognised. However, little attention has been paid to the quality assurance of the applications software. Our work reported here aims at verifying the analysis software for processing of renal nuclear medicine studies (renograms). The software tools were used to build a synthetic dynamic model of renal system. The model consists of two phases: perfusion and function. The organs of interest (kidneys, bladder and aortic artery) were simple geometric forms. The uptake of the renal structures was described by mathematic functions. Curves corresponding to normal or pathological conditions were simulated for kidneys, bladder and aortic artery by appropriate selection of parameters. There was no difference between the parameters of the mathematic curves and the quantitative data produced by the renal analysis program. Our test procedure is simple to apply, reliable, reproducible and rapid to verify the renal applications software.
Plastow, Richard A.
Complex Electronics (CE) are now programmed to perform tasks that were previously handled in software, such as communication protocols. Many of the methods used to develop software bare a close resemblance to CE development. For instance, Field Programmable Gate Arrays (FPGAs) can have over a million logic gates while system-on-chip (SOC) devices can combine a microprocessor, input and output channels, and sometimes an FPGA for programmability. With this increased intricacy, the possibility of software-like bugs such as incorrect design, logic, and unexpected interactions within the logic is great. Since CE devices are obscuring the hardware/software boundary, we propose that mature software methodologies may be utilized with slight modifications in the development of these devices. Software Process Assurance for Complex Electronics (SPACE) is a research project that looks at using standardized S/W Assurance/Engineering practices to provide an assurance framework for development activities. Tools such as checklists, best practices and techniques can be used to detect missing requirements and bugs earlier in the development cycle creating a development process for CE that will be more easily maintained, consistent and configurable based on the device used.
Plastow, Richard A.
Complex Electronics (CE) now perform tasks that were previously handled in software, such as communication protocols. Many methods used to develop software bare a close resemblance to CE development. Field Programmable Gate Arrays (FPGAs) can have over a million logic gates while system-on-chip (SOC) devices can combine a microprocessor, input and output channels, and sometimes an FPGA for programmability. With this increased intricacy, the possibility of software-like bugs such as incorrect design, logic, and unexpected interactions within the logic is great. With CE devices obscuring the hardware/software boundary, we propose that mature software methodologies may be utilized with slight modifications in the development of these devices. Software Process Assurance for Complex Electronics (SPACE) is a research project that used standardized S/W Assurance/Engineering practices to provide an assurance framework for development activities. Tools such as checklists, best practices and techniques were used to detect missing requirements and bugs earlier in the development cycle creating a development process for CE that was more easily maintained, consistent and configurable based on the device used.
Software Quality Assurance d C t l St d dan on ro s an ar Sue Carroll Principal Software Quality Analyst, SAS John Wal z VP Technology and...for Information Operations and Reports, 1215 Jefferson Davis Highway, Suite 1204, Arlington VA 22202-4302. Respondents should be aware that...Cycle (SLC) process? • What is in a SQA Process? • Where are SQA Controls? • What is the SQA standards history? Wh t i h i i SQA?• a s c ang ng n
and Capability dEtermination-SPICE, ISO /IEC 15504, 1998. [ ISO 2007] International Organization for Standardization. " ISO /IEC 27001 & 27002 ...Implementing the Process Areas 6.2.7 Differences Between the CMMI and Software CMM Process Areas 6.3 The CMMI Appraisal Process 6.4 Adapting ISO 15504 to...Secure Software Assurance 6.4.1 Assessment and the Secure Life Cycle 6.4.2 ISO 15504 Capability Levels 6.5 Adapting the ISOIIEC 21287 Standard Approach to
This report contains viewgraphs on operational excellence philosophy of six sigma applied to software quality assurance. This report outlines the following: goal of six sigma; six sigma tools; manufacturing vs administrative processes; Software quality assurance document inspections; map software quality assurance requirements document; failure mode effects analysis for requirements document; measuring the right response variables; and questions.
of ISO 12207 -2008, both internationally and in the United States . That standard documents a comprehensive set of activities and supporting...grows, organizations must ensure that their procurement agents acquire high quality, secure software. ISO 12207 and the Software Assurance Competency...cyberattacks grows, organizations must ensure that their procurement agents acquire high quality, secure software. ISO 12207 and the Software Assurance
Previously, several research tasks have been conducted, some observations were obtained, and several possible suggestions have been contemplated involving software quality assurance engineering at NASA Johnson. These research tasks are briefly described. Also, a brief discussion is given on the role of software quality assurance in software engineering along with some observations and suggestions. A brief discussion on a training program for software quality assurance engineers is provided. A list of assurance factors as well as quality factors are also included. Finally, a process model which can be used for searching and collecting software quality assurance tools is presented.
The MPSS Software Quality Assurance (SQAP) describes the tools and strategy used in the development of the MPSS software. The document also describes the methodology for controlling and managing changes to the software.
As EPA’s environmental research expands into new areas that involve the development of software, quality assurance concepts and procedures that were originally developed for environmental data collection may not be appropriate. Fortunately, software quality assurance is a ...
This document provides a survey of software assurance techniques for highly reliable systems including a discussion of relevant safety standards for various industries in the United States and Europe, as well as examples of methods used during software development projects. It contains one section for each industry surveyed: Aerospace, Defense, Nuclear Power, Medical Devices and Transportation. Each section provides an overview of applicable standards and examples of a mission or software development project, software assurance techniques used and reliability achieved.
Perry, Sandra; Jordan, Leon; Decker, William; Page, Gerald; Mcgarry, Frank E.; Valett, Jon
The product assurance policies and procedures necessary to support flight dynamics software development projects for Goddard Space Flight Center are presented. The quality assurance and configuration management methods and tools for each phase of the software development life cycles are described, from requirements analysis through acceptance testing; maintenance and operation are not addressed.
Contents Acknowledgments iii Abstract v 1 An Undergraduate Curriculum Focus on Software Assurance 1 2 Computer Science I 7 3 Computer Science II...confidence that can be integrated into traditional software development and acquisition process models . Thus, in addition to a technology focus...testing throughout the software development life cycle ( SDLC ) AP Security and complexity—system development challenges: security failures
Feather, M. S.; Sigal, B.; Cornford, S. L.; Hutchinson, P.
The objective is to use cost-benefit analyses to identify, for a given project, optimal sets of software assurance activities. Towards this end we have incorporated cost-benefit calculations into a risk management framework.
This slide presentation reviews the interactions between the scientist and engineers doing research and technology and the software developers and others who are doing software assurance. There is a discussion of the role of the Safety and Mission Assurance (SMA) in developing software to be used for research and technology, and the importance of this role as the technology moves to the higher levels of the technology readiness levels (TRLs). There is also a call to change the way the development of software is developed.
Pressburger, Thomas; DiVito, Ben; Feather, Martin S.; Hinchey, Michael; Markosian, Lawrence; Trevino, Luis C.
Research in the software engineering community continues to lead to new development techniques that encompass processes, methods and tools. However, a number of obstacles impede their infusion into software development practices. These are the recurring obstacles common to many forms of research. Practitioners cannot readily identify the emerging techniques that may benefit them, and cannot afford to risk time and effort evaluating and trying one out while there remains uncertainty about whether it will work for them. Researchers cannot readily identify the practitioners whose problems would be amenable to their techniques, and, lacking feedback from practical applications, are hard-pressed to gauge the where and in what ways to evolve their techniques to make them more likely to be successful. This paper describes an ongoing effort conducted by a software engineering research infusion team established by NASA s Software Engineering Initiative to overcome these obstacles. .
Mathematical expression for the Total Sum of Squares which measures the variability that results when all values are treated as a combined sample coming from...primarily interested in background on software design and high-assurance computing, research in software architecture generation or evaluation...respectively. Those readers solely interested in the validation of a software design approach should at the minimum read Chapter 6 followed by Chapter
This report is concerned with the application of software quality and evaluation measures to AI software and, more broadly, with the question of quality assurance for AI software. Considered are not only the metrics that attempt to measure some aspect of software quality, but also the methodologies and techniques (such as systematic testing) that attempt to improve some dimension of quality, without necessarily quantifying the extent of the improvement. The report is divided into three parts Part 1 reviews existing software quality measures, i.e., those that have been developed for, and applied to, conventional software. Part 2 considers the characteristics of AI software, the applicability and potential utility of measures and techniques identified in the first part, and reviews those few methods developed specifically for AI software. Part 3 presents an assessment and recommendations for the further exploration of this important area.
Hihn, Jarius; Tregre, Grant
The software industry repeatedly observes cost growth of well over 100% even after decades of cost estimation research and well-known best practices, so "What's the problem?" In this paper we will provide an overview of the current state oj software cost estimation best practice. We then explore whether applying some of the methods used in software assurance might improve the quality of software cost estimates. This paper especially focuses on issues associated with model calibration, estimate review, and the development and documentation of estimates as part alan integrated plan.
Yildiz, Ozkan; Demirörs, Onur
In this study, we investigated the adoptability of software quality measures for healthcare process measurement. Quality measures of ISO/IEC 9126 are redefined from a process perspective to build a generic healthcare process quality measurement model. Case study research method is used, and the model is applied to a public hospital's Entry to Care process. After the application, weak and strong aspects of the process can be easily observed. Access audibility, fault removal, completeness of documentation, and machine utilization are weak aspects and these aspects are the candidates for process improvement. On the other hand, functional completeness, fault ratio, input validity checking, response time, and throughput time are the strong aspects of the process.
Koga, Dennis (Technical Monitor); Penix, John; Markosian, Lawrence Z.
In this paper, we describe our experience with the challenges thar we are currently facing in our effort to develop advanced software verification and validation tools. We categorize these challenges into several areas: cost benefits modeling, tool usability, customer application domain, and organizational issues. We provide examples of challenges in each area and identrfj, open research issues in areas which limit our ability to transfer high-assurance software engineering tools into practice.
Cuyno, Patrick; Malnick, Kathy D.; Schaeffer, Chad E.
This paper will provide a description of some of the challenges NASA is facing in providing software assurance within the new commercial space services paradigm, namely with the Commercial Crew Program (CCP). The CCP will establish safe, reliable, and affordable access to the International Space Station (ISS) by purchasing a ride from commercial companies. The CCP providers have varying experience with software development in safety-critical space systems. NASA's role in providing effective software assurance support to the CCP providers is critical to the success of CCP. These challenges include funding multiple vehicles that execute in parallel and have different rules of engagement, multiple providers with unique proprietary concerns, providing equivalent guidance to all providers, permitting alternates to NASA standards, and a large number of diverse stakeholders. It is expected that these challenges will exist in future programs, especially if the CCP paradigm proves successful. The proposed CCP approach to address these challenges includes a risk-based assessment with varying degrees of engagement and a distributed assurance model. This presentation will describe NASA IV&V Program's software assurance support and responses to these challenges.
Savarino, Shirley; Fitz, Rhonda; Fesq, Lorraine; Whitman, Gerek
Fault Management (FM) is focused on safety, the preservation of assets, and maintaining the desired functionality of the system. How FM is implemented varies among missions. Common to most missions is system complexity due to a need to establish a multi-dimensional structure across hardware, software and spacecraft operations. FM is necessary to identify and respond to system faults, mitigate technical risks and ensure operational continuity. Generally, FM architecture, implementation, and software assurance efforts increase with mission complexity. Because FM is a systems engineering discipline with a distributed implementation, providing efficient and effective verification and validation (V&V) is challenging. A breakout session at the 2012 NASA Independent Verification & Validation (IV&V) Annual Workshop titled "V&V of Fault Management: Challenges and Successes" exposed this issue in terms of V&V for a representative set of architectures. NASA's Software Assurance Research Program (SARP) has provided funds to NASA IV&V to extend the work performed at the Workshop session in partnership with NASA's Jet Propulsion Laboratory (JPL). NASA IV&V will extract FM architectures across the IV&V portfolio and evaluate the data set, assess visibility for validation and test, and define software assurance methods that could be applied to the various architectures and designs. This SARP initiative focuses efforts on FM architectures from critical and complex projects within NASA. The identification of particular FM architectures and associated V&V/IV&V techniques provides a data set that can enable improved assurance that a system will adequately detect and respond to adverse conditions. Ultimately, results from this activity will be incorporated into the NASA Fault Management Handbook providing dissemination across NASA, other agencies and the space community. This paper discusses the approach taken to perform the evaluations and preliminary findings from the research.
Savarino, Shirley; Fitz, Rhonda; Fesq, Lorraine; Whitman, Gerek
The satellite systems Fault Management (FM) is focused on safety, the preservation of assets, and maintaining the desired functionality of the system. How FM is implemented varies among missions. Common to most is system complexity due to a need to establish a multi-dimensional structure across hardware, software and operations. This structure is necessary to identify and respond to system faults, mitigate technical risks and ensure operational continuity. These architecture, implementation and software assurance efforts increase with mission complexity. Because FM is a systems engineering discipline with a distributed implementation, providing efficient and effective verification and validation (VV) is challenging. A breakout session at the 2012 NASA Independent Verification Validation (IVV) Annual Workshop titled VV of Fault Management: Challenges and Successes exposed these issues in terms of VV for a representative set of architectures. NASA's IVV is funded by NASA's Software Assurance Research Program (SARP) in partnership with NASA's Jet Propulsion Laboratory (JPL) to extend the work performed at the Workshop session. NASA IVV will extract FM architectures across the IVV portfolio and evaluate the data set for robustness, assess visibility for validation and test, and define software assurance methods that could be applied to the various architectures and designs. This work focuses efforts on FM architectures from critical and complex projects within NASA. The identification of particular FM architectures, visibility, and associated VVIVV techniques provides a data set that can enable higher assurance that a satellite system will adequately detect and respond to adverse conditions. Ultimately, results from this activity will be incorporated into the NASA Fault Management Handbook providing dissemination across NASA, other agencies and the satellite community. This paper discusses the approach taken to perform the evaluations and preliminary findings from the
Kandt, Ronald Kirk
The basic idea underlying this paper is that the conventional understanding of the role of a Software Quality Assurance (SQA) engineer is unduly limited. This is because few have asked who the customers of a SQA engineer are. Once you do this, you can better define what tasks a SQA engineer should perform, as well as identify the knowledge and skills that such a person should have. The consequence of doing this is that a SQA engineer can provide greater value to his or her customers. It is the position of this paper that a SQA engineer providing significant value to his or her customers must not only assume the role of an auditor, but also that of a software and systems engineer. This is because software engineers and their managers particularly value contributions that directly impact products and their development. These ideas are summarized as lessons learned, based on my experience at Jet Propulsion Laboratory (JPL).
Keegstra, P. B.; Smoot, G. F.; Bennett, C. L.; Aymon, J.; Backus, C.; Deamici, G.; Hinshaw, G.; Jackson, P. D.; Kogut, A.; Lineweaver, C.
Six Differential Microwave Radiometers (DMR) on COBE (Cosmic Background Explorer) measure the large-angular-scale isotropy of the cosmic microwave background (CMB) at 31.5, 53, and 90 GHz. Quality assurance software analyzes the daily telemetry from the spacecraft to ensure that the instrument is operating correctly and that the data are not corrupted. Quality assurance for DMR poses challenging requirements. The data are differential, so a single bad point can affect a large region of the sky, yet the CMB isotropy requires lengthy integration times (greater than 1 year) to limit potential CMB anisotropies. Celestial sources (with the exception of the moon) are not, in general, visible in the raw differential data. A 'quicklook' software system was developed that, in addition to basic plotting and limit-checking, implements a collection of data tests as well as long-term trending. Some of the key capabilities include the following: (1) stability analysis showing how well the data RMS averages down with increased data; (2) a Fourier analysis and autocorrelation routine to plot the power spectrum and confirm the presence of the 3 mK 'cosmic' dipole signal; (3) binning of the data against basic spacecraft quantities such as orbit angle; (4) long-term trending; and (5) dipole fits to confirm the spacecraft attitude azimuth angle.
Pendrill, L. R.
Is it realistic to aspire to the same kind of quality-assurance of measurement in person-centred care, currently being implemented in healthcare globally, as is established in the physical sciences and engineering? Ensuring metrological comparability (‘traceability’) and reliably declaring measurement uncertainty when assessing patient ability or increased social capital are however challenging for subjective measurements often characterised by large dispersion. Drawing simple analogies between ‘instruments’ in the social sciences—questionnaires, ability tests, etc—and engineering instruments such as thermometers does not go far enough. A possible way forward, apparently equally applicable to both physical and social measurement, seems to be to model inferences in terms of performance metrics of a measurement system. Person-centred care needs person-centred measurement and a full picture of the measurement process when man acts as a measurement instrument is given in the present paper. This complements previous work by presenting the process, step by step, from the observed indication (e.g. probability of success, P success, of achieving a task), through restitution with Rasch measurement theory, to the measurand (e.g. task difficulty). Rasch invariant measure theory can yield quantities—‘latent’ (or ‘explanatory’) variables such as task challenge or person ability—with characteristics akin to those of physical quantities. Metrological references for comparability via traceability and reliable estimates of uncertainty and decision risks are then in reach even for perceptive measurements (and other qualitative properties). As a case study, the person-centred measurement of cognitive ability is examined, as part of the EU project EMPIR 15HLT04 NeuroMet, for Alzheimer’s, where better analysis of correlations with brain atrophy is enabled thanks to the Rasch metrological approach.
Hinchey, Michael G.; Pressburger, Thomas; Markosian, Lawrence; Feather, Martin S.
New processes, methods and tools are constantly appearing in the field of software engineering. Many of these augur great potential in improving software development processes, resulting in higher quality software with greater levels of assurance. However, there are a number of obstacles that impede their infusion into software development practices. These are the recurring obstacles common to many forms of research. Practitioners cannot readily identify the emerging techniques that may most benefit them, and cannot afford to risk time and effort in evaluating and experimenting with them while there is still uncertainty about whether they will have payoff in this particular context. Similarly, researchers cannot readily identify those practitioners whose problems would be amenable to their techniques and lack the feedback from practical applications necessary to help them to evolve their techniques to make them more likely to be successful. This paper describes an ongoing effort conducted by a software engineering research infusion team, and the NASA Research Infusion Initiative, established by NASA s Software Engineering Initiative, to overcome these obstacles.
Bush, Marilyn W.
The findings are reported of the Jet Propulsion Laboratory (JPL)/Software Product Assurance (SPA) Metrics Study, conducted as part of a larger JPL effort to improve software quality and productivity. Until recently, no comprehensive data had been assembled on how JPL manages and develops software-intensive systems. The first objective was to collect data on software development from as many projects and for as many years as possible. Results from five projects are discussed. These results reflect 15 years of JPL software development, representing over 100 data points (systems and subsystems), over a third of a billion dollars, over four million lines of code and 28,000 person months. Analysis of this data provides a benchmark for gauging the effectiveness of past, present and future software development work. In addition, the study is meant to encourage projects to record existing metrics data and to gather future data. The SPA long term goal is to integrate the collection of historical data and ongoing project data with future project estimations.
in applications is software assurance. There are many subtle variations to the software assurance definition (Goertzel, et al ., 2007), but the DoD...Gary McGraw (2006), and Thorsten 18 Schneider (2006). Goertzel, et al . (2007), lists and compares several security-enhanced software development...detailed by Goertzel, et al ., is the Microsoft Trustworthy Computing Security Development Lifecycle (SDL), shown in the following figure: Figure 6
Gaebel, W; Großimlinghaus, I; Heun, R; Janssen, B; Johnson, B; Kurimay, T; Montellano, P; Muijen, M; Munk-Jorgensen, P; Rössler, W; Ruggeri, M; Thornicroft, G; Zielasek, J
To advance the quality of mental healthcare in Europe by developing guidance on implementing quality assurance. We performed a systematic literature search on quality assurance in mental healthcare and the 522 retrieved documents were evaluated by two independent reviewers (B.J. and J.Z.). Based on these evaluations, evidence tables were generated. As it was found that these did not cover all areas of mental healthcare, supplementary hand searches were performed for selected additional areas. Based on these findings, fifteen graded recommendations were developed and consented by the authors. Review by the EPA Guidance Committee and EPA Board led to two additional recommendations (on immigrant mental healthcare and parity of mental and physical healthcare funding). Although quality assurance (measures to keep a certain degree of quality), quality control and monitoring (applying quality indicators to the current degree of quality), and quality management (coordinated measures and activities with regard to quality) are conceptually distinct, in practice they are frequently used as if identical and hardly separable. There is a dearth of controlled trials addressing ways to optimize quality assurance in mental healthcare. Altogether, seventeen recommendations were developed addressing a range of aspects of quality assurance in mental healthcare, which appear usable across Europe. These were divided into recommendations about structures, processes and outcomes. Each recommendation was assigned to a hierarchical level of analysis (macro-, meso- and micro-level). There was a lack of evidence retrievable by a systematic literature search about quality assurance of mental healthcare. Therefore, only after further topics and search had been added it was possible to develop recommendations with mostly medium evidence levels. Evidence-based graded recommendations for quality assurance in mental healthcare were developed which should next be implemented and evaluated for
May 2007 www.stsc.hill.af.mil 17 2 • address security concerns in the software development life cycle ( SDLC )? • Are there formal software quality...What threat modeling process, if any, is used when designing the software ? What analysis, design, and construction tools are used by your software design...the-shelf (COTS), government off-the-shelf (GOTS), open- source, embedded, and legacy software . Attackers exploit unintentional vulnerabil- ities or
This advisory circular (AC) provides an acceptable means, but not the only means, to show compliance with the quality assurance requirements of Federal Aviation Regulations (FAR) Part 21, Certification Procedures for Products and Parts, as applicable...
... Machines (IBM), Software Group Business Unit, Quality Assurance Group, San Jose, California; Notice of... workers of International Business Machines (IBM), Software Group Business Unit, Optim Data Studio Tools QA... February 2, 2011 (76 FR 5832). The subject worker group supplies acceptance testing services, design...
frameworks 4 CMMI-DEV IEEE / ISO / IEC 15288 / 12207 Quality Assurance ©2011 Walz IEEE Life Cycle Processes & Artifacts • Systems Life Cycle Processes...TAG to ISO TC 176 Quality Management • Quality: ASQ, work experience • Software: three books, consulting, work experience • Systems: Telecom & DoD...and IEEE 730 SQA need to align. The P730 IEEE standards working group has expanded the scope of the SQA process standard to align with IS 12207
management, available at <http://www.iso.org/ iso /en/CatalogueDetailPage.CatalogueDetail?CSNUMBER=39612&ICS1=35&ICS2=40 &ICS3=>. ISO /IEC 27001 . Information...Management of the Systems Engineering Process. [ ISO /IEC 27001 ] ISO /IEC 27001 :2005. Information technology -- Security techniques -- Information security...software life cycles [ ISO /IEC 15026]. Software assurance is a key element of national security and homeland security. It is critical because dramatic
foes. Subversion or sabotage of a software-based temperature control in a nuclear power plant could result in a meltdown, while sabotage of avionic... resulting from research and development activities, and provide a comprehensive assessment of IA technologies. Topic areas for SOARs are solicited...243 18.104.22.168 Recent Research Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 244 6.2
Hayhurst, Kelly J. (Editor)
The Guidance and Control Software (GCS) project was the last in a series of software reliability studies conducted at Langley Research Center between 1977 and 1994. The technical results of the GCS project were recorded after the experiment was completed. Some of the support documentation produced as part of the experiment, however, is serving an unexpected role far beyond its original project context. Some of the software used as part of the GCS project was developed to conform to the RTCA/DO-178B software standard, "Software Considerations in Airborne Systems and Equipment Certification," used in the civil aviation industry. That standard requires extensive documentation throughout the software development life cycle, including plans, software requirements, design and source code, verification cases and results, and configuration management and quality control data. The project documentation that includes this information is open for public scrutiny without the legal or safety implications associated with comparable data from an avionics manufacturer. This public availability has afforded an opportunity to use the GCS project documents for DO-178B training. This report provides a brief overview of the GCS project, describes the 4-volume set of documents and the role they are playing in training, and includes configuration management and quality assurance documents from the GCS project. Volume 4 contains six appendices: A. Software Accomplishment Summary for the Guidance and Control Software Project; B. Software Configuration Index for the Guidance and Control Software Project; C. Configuration Management Records for the Guidance and Control Software Project; D. Software Quality Assurance Records for the Guidance and Control Software Project; E. Problem Report for the Pluto Implementation of the Guidance and Control Software Project; and F. Support Documentation Change Reports for the Guidance and Control Software Project.
activities; 2. Cryptologic activities related to national security; 3. Command and control of military forces; 4. Equipment that is an integral part of a...Testing and Integration , and Performance or Operation (6). Figure 3 shows the software life cycle and the key outputs of the phases. The first phase to...defects. This procedure is considered the Checkout (13:09-91). Once coding is complete, the Testing and Integration Phase begins. Here the developed
based information retrieval system, we might say that this program works like a hive of bees , going out for pollen and bringing it back to the hive...developers ® Six Siqma is reqistered in the U. S. Patent and Trademark Office by Motorola ^_ 33 @ AEROSPACE Major Areas in a Typical Software...requirements - Capturing and evaluating quality metrics, identifying common problem areas **» Despite its positive impact on quality, pair programming
Kristianti, Y.; Prabawanto, S.; Suhendra, S.
This study aims to examine the ability of critical thinking and students who attain learning mathematics with learning model ASSURE assisted Autograph software. The design of this study was experimental group with pre-test and post-test control group. The experimental group obtained a mathematics learning with ASSURE-assisted model Autograph software and the control group acquired the mathematics learning with the conventional model. The data are obtained from the research results through critical thinking skills tests. This research was conducted at junior high school level with research population in one of junior high school student in Subang Regency of Lesson Year 2016/2017 and research sample of class VIII student in one of junior high school in Subang Regency for 2 classes. Analysis of research data is administered quantitatively. Quantitative data analysis was performed on the normalized gain level between the two sample groups using a one-way anova test. The results show that mathematics learning with ASSURE assisted model Autograph software can improve the critical thinking ability of junior high school students. Mathematical learning using ASSURE-assisted model Autograph software is significantly better in improving the critical thinking skills of junior high school students compared with conventional models.
Macdonald, John M; Lloyd, Jane A; Turner, Cameron J
Ideally, quality is designed into software, just as quality is designed into hardware. However, when dealing with legacy systems, demonstrating that the software meets required quality standards may be difficult to achieve. As the need to demonstrate the quality of existing software was recognized at Los Alamos National Laboratory (LANL), an effort was initiated to uncover and demonstrate that legacy software met the required quality standards. This effort led to the development of a reverse engineering approach referred to as software archaeology. This paper documents the software archaeology approaches used at LANL to document legacy software systems. A case studymore » for the Robotic Integrated Packaging System (RIPS) software is included.« less
Ichu, Emmanuel A.
Software quality is perhaps one of the most sought-after attributes in product development, however; this goal is unattained. Problem factors in software development and how these have affected the maintainability of the delivered software systems requires a thorough investigation. It was, therefore, very important to understand software…
Salomons, Greg J; Kelly, Diane
This paper reports on a survey of medical physicists who write and use in-house written software as part of their professional work. The goal of the survey was to assess the extent of in-house software usage and the desire or need for related software quality guidelines. The survey contained eight multiple-choice questions, a ranking question, and seven free text questions. The survey was sent to medical physicists associated with cancer centers across Canada. The respondents to the survey expressed interest in having guidelines to help them in their software-related work, but also demonstrated extensive skills in the area of testing, safety, and communication. These existing skills form a basis for medical physicists to establish a set of software quality guidelines.
Guarro, Sergio B.
This report validates and documents the detailed features and practical application of the framework for software intensive digital systems risk assessment and risk-informed safety assurance presented in the NASA PRA Procedures Guide for Managers and Practitioner. This framework, called herein the "Context-based Software Risk Model" (CSRM), enables the assessment of the contribution of software and software-intensive digital systems to overall system risk, in a manner which is entirely compatible and integrated with the format of a "standard" Probabilistic Risk Assessment (PRA), as currently documented and applied for NASA missions and applications. The CSRM also provides a risk-informed path and criteria for conducting organized and systematic digital system and software testing so that, within this risk-informed paradigm, the achievement of a quantitatively defined level of safety and mission success assurance may be targeted and demonstrated. The framework is based on the concept of context-dependent software risk scenarios and on the modeling of such scenarios via the use of traditional PRA techniques - i.e., event trees and fault trees - in combination with more advanced modeling devices such as the Dynamic Flowgraph Methodology (DFM) or other dynamic logic-modeling representations. The scenarios can be synthesized and quantified in a conditional logic and probabilistic formulation. The application of the CSRM method documented in this report refers to the MiniAERCam system designed and developed by the NASA Johnson Space Center.
Agnew, Christina E; McGarry, Conor K
To provide a tool to enable gamma analysis software algorithms to be included in a quality assurance (QA) program. Four image sets were created comprising two geometric images to independently test the distance to agreement (DTA) and dose difference (DD) elements of the gamma algorithm, a clinical step and shoot IMRT field and a clinical VMAT arc. The images were analysed using global and local gamma analysis with 2 in-house and 8 commercially available software encompassing 15 software versions. The effect of image resolution on gamma pass rates was also investigated. All but one software accurately calculated the gamma passing rate for the geometric images. Variation in global gamma passing rates of 1% at 3%/3mm and over 2% at 1%/1mm was measured between software and software versions with analysis of appropriately sampled images. This study provides a suite of test images and the gamma pass rates achieved for a selection of commercially available software. This image suite will enable validation of gamma analysis software within a QA program and provide a frame of reference by which to compare results reported in the literature from various manufacturers and software versions. Copyright © 2015. Published by Elsevier Ireland Ltd.
This paper reports on a survey of medical physicists who write and use in‐house written software as part of their professional work. The goal of the survey was to assess the extent of in‐house software usage and the desire or need for related software quality guidelines. The survey contained eight multiple‐choice questions, a ranking question, and seven free text questions. The survey was sent to medical physicists associated with cancer centers across Canada. The respondents to the survey expressed interest in having guidelines to help them in their software‐related work, but also demonstrated extensive skills in the area of testing, safety, and communication. These existing skills form a basis for medical physicists to establish a set of software quality guidelines. PACS number: 87.55.Qr PMID:25679168
Schwantje, Robert; Smith, Claude
This document defines the responsibilities of Software Quality Assurance (SOA) for the development of the flight software installed in EOS/AMSU-A instruments, and the ground support software used in the test and integration of the EOS/AMSU-A instruments.
Assure ; CASRN 76578 - 14 - 8 Human health assessment information on a chemical substance is included in the IRIS database only after a comprehensive review of toxicity data , as outlined in the IRIS assessment development process . Sections I ( Health Hazard Assessments for Noncarcinogenic Effects
Sowunmi, Olaperi Yeside; Misra, Sanjay; Fernandez-Sanz, Luis; Crawford, Broderick; Soto, Ricardo
The importance of quality assurance in the software development process cannot be overemphasized because its adoption results in high reliability and easy maintenance of the software system and other software products. Software quality assurance includes different activities such as quality control, quality management, quality standards, quality planning, process standardization and improvement amongst others. The aim of this work is to further investigate the software quality assurance practices of practitioners in Nigeria. While our previous work covered areas on quality planning, adherence to standardized processes and the inherent challenges, this work has been extended to include quality control, software process improvement and international quality standard organization membership. It also makes comparison based on a similar study carried out in Turkey. The goal is to generate more robust findings that can properly support decision making by the software community. The qualitative research approach, specifically, the use of questionnaire research instruments was applied to acquire data from software practitioners. In addition to the previous results, it was observed that quality assurance practices are quite neglected and this can be the cause of low patronage. Moreover, software practitioners are neither aware of international standards organizations or the required process improvement techniques; as such their claimed standards are not aligned to those of accredited bodies, and are only limited to their local experience and knowledge, which makes it questionable. The comparison with Turkey also yielded similar findings, making the results typical of developing countries. The research instrument used was tested for internal consistency using the Cronbach's alpha, and it was proved reliable. For the software industry in developing countries to grow strong and be a viable source of external revenue, software assurance practices have to be taken seriously
Amalia, A.; Gunawan, D.; Hardi, S. M.; Rachmawati, D.
The Internal Quality Assurance System (in Indonesian: SPMI (Sistem Penjaminan Mutu Internal) is a systemic activity of quality assurance of higher education in Indonesia. SPMI should be done by all higher education or universities in Indonesia based on the Regulation of the Minister of Research, Technology and Higher Education of the Republic of Indonesia Number 62 of 2016. Implementation of SPMI must refer to the principle of SPMI that is independent, standardize, accurate, well planned and sustainable, documented and systematic. To assist the SPMI cycle properly, universities need a supporting software to monitor all the activities of SPMI. But in reality, many universities are not optimal in building this SPMI monitoring system. One of the obstacles is the determination of system requirements in support of SPMI principles is difficult to achieve. In this paper, we observe the initial phase of the engineering requirements elicitation. Unlike other methods that collect system requirements from users and stakeholders, we find the system requirements of the SPMI principles from SPMI guideline book. The result of this paper can be used as a choice in determining SPMI software requirements. This paper can also be used by developers and users to understand the scenario of SPMI so that could overcome the problems of understanding between this two parties.
Isambert, A; Beaudré, A; Ferreira, I; Lefkopoulos, D
Virtual simulation process is often used to prepare three dimensional conformal radiation therapy treatments. As the quality of the treatment is widely dependent on this step, it is mandatory to perform extensive controls on this software before clinical use. The tests presented in this work have been carried out on the treatment planning system ISOgray (DOSIsoft), including the delineation module IMAgo and the virtual simulation module SIMAgo. According to our experience, the most relevant controls of international protocols have been selected. These tests mainly focused on measuring and delineation tools, virtual simulation functionalities, and have been performed with three phantoms: the Quasar Multi-Purpose Body Phantom, the Quasar MLC Beam Geometry Phantom (Modus Medical Devices Inc.) and a phantom developed at Hospital Tenon. No major issues have been identified while performing the tests. These controls have emphasized the necessity for the user to consider with a critical eye the results displayed by a virtual simulation software. The contrast of visualisation, the slice thickness, the calculation and display mode of 3D structures used by the software are many factors of uncertainties. A virtual simulation software quality assurance procedure has been written and applied on a set of CT images. Similar tests have to be performed periodically and at minimum at each change of major version.
Liu, Yuxuan; Williams, Mark L.; Wiarda, Dorothea
This report fulfills the requirements for the Consortium for the Advanced Simulation of Light-Water Reactors (CASL) milestone L2:RTM.P14.02, “SQA and Verification for MPACT Library Generation,” by documenting the current status of the software quality, verification, and acceptance testing of nuclear data libraries for MPACT. It provides a brief overview of the library generation process, from general-purpose evaluated nuclear data files (ENDF/B) to a problem-dependent cross section library for modeling of light-water reactors (LWRs). The software quality assurance (SQA) programs associated with each of the software used to generate the nuclear data libraries are discussed; specific tests within the SCALE/AMPX andmore » VERA/XSTools repositories are described. The methods and associated tests to verify the quality of the library during the generation process are described in detail. The library generation process has been automated to a degree to (1) ensure that it can be run without user intervention and (2) to ensure that the library can be reproduced. Finally, the acceptance testing process that will be performed by representatives from the Radiation Transport Methods (RTM) Focus Area prior to the production library’s release is described in detail.« less
Aghazadeh, Sakineh; Pirnejad, Habibollah; Moradkhani, Alireza; Aliev, Alvosat
The aim of this study was to discover the effect of software quality characteristics on healthcare quality and efficiency indicators. Through a systematic literature review, we selected and analyzed 37 original research papers to investigate the impact of the software indicators (coming from the standard ISO 9126 quality characteristics and sub-characteristics) on some of healthcare important outcome indicators and finally ranked these software indicators. The results showed that the software characteristics usability, reliability and efficiency were mostly favored in the studies, indicating their importance. On the other hand, user satisfaction, quality of patient care, clinical workflow efficiency, providers' communication and information exchange, patient satisfaction and care costs were among the healthcare outcome indicators frequently evaluated in relation to the mentioned software characteristics. Regression Logistic Method was the most common assessment methodology, and Confirmatory Factor Analysis and Structural Equation Modeling were performed to test the structural model's fit. The software characteristics were considered to impact the healthcare outcome indicators through other intermediate factors (variables).
Nardon, Fabiane Bizinella; de A Moura, Lincoln
São Paulo is the largest city in Brazil and one of the largest cities in the world. In 2004, São Paulo City Department of Health decided to implement a Healthcare Information System to support managing healthcare services and provide an ambulatory health record. The resulting information system is one of the largest public healthcare information systems ever built, with more than 2 million lines of code. Although statistics shows that most software projects fail, and the risks for the São Paulo initiative were enormous, the information system was completed on-time and on-budget. In this paper, we discuss the software engineering principles adopted that allowed to accomplish that project's goals, hoping that sharing the experience of this project will help other healthcare information systems initiatives to succeed.
EPA benchmark dose software (BMDS) issued to evaluate chemical dose-response data in support of Agency risk assessments, and must therefore be dependable. Quality assurance testing methods developed for BMDS were designed to assess model dependability with respect to curve-fitt...
Changes to the software used in general practice could improve the collection of the Aboriginal and Torres Strait Islander status of all patients, and boost access to healthcare measures specifically for Aboriginal and Torres Strait Islander peoples provided directly or indirectly by general practitioners (GPs). Despite longstanding calls for improvements to general practice software to better support Aboriginal and Torres Strait Islander health, little change has been made. The aim of this article is to promote software improvements by identifying desirable software attributes and encouraging GPs to promote their adoption. Establishing strong links between collecting Aboriginal and Torres Strait Islander status, clinical decision supports, and uptake of GP-mediated health measures specifically for Aboriginal and Torres Strait Islander peoples - and embedding these links in GP software - is a long overdue reform. In the absence of government initiatives in this area, GPs are best placed to advocate for software changes, using the model described here as a starting point for action.
Koru, Günes; El Emam, Khaled; Neisa, Angelica; Umarji, Medha
Open source (OS) software is continuously gaining recognition and use in the biomedical domain, for example, in health informatics and bioinformatics. Given the mission critical nature of applications in this domain and their potential impact on patient safety, it is important to understand to what degree and how effectively biomedical OS developers perform standard quality assurance (QA) activities such as peer reviews and testing. This would allow the users of biomedical OS software to better understand the quality risks, if any, and the developers to identify process improvement opportunities to produce higher quality software. A survey of developers working on biomedical OS projects was conducted to examine the QA activities that are performed. We took a descriptive approach to summarize the implementation of QA activities and then examined some of the factors that may be related to the implementation of such practices. Our descriptive results show that 63% (95% CI, 54-72) of projects did not include peer reviews in their development process, while 82% (95% CI, 75-89) did include testing. Approximately 74% (95% CI, 67-81) of developers did not have a background in computing, 80% (95% CI, 74-87) were paid for their contributions to the project, and 52% (95% CI, 43-60) had PhDs. A multivariate logistic regression model to predict the implementation of peer reviews was not significant (likelihood ratio test = 16.86, 9 df, P = .051) and neither was a model to predict the implementation of testing (likelihood ratio test = 3.34, 9 df, P = .95). Less attention is paid to peer review than testing. However, the former is a complementary, and necessary, QA practice rather than an alternative. Therefore, one can argue that there are quality risks, at least at this point in time, in transitioning biomedical OS software into any critical settings that may have operational, financial, or safety implications. Developers of biomedical OS applications should invest more effort
The Software Quality Assurance (SQA) regression test suite for DYNA3D (Zywicz and Lin, 2015) and ParaDyn (DeGroot, et al., 2015) currently contains approximately 600 problems divided into 21 suites, and is a required component of ParaDyn’s SQA plan (Ferencz and Oliver, 2013). The regression suite allows developers to ensure that software modifications do not unintentionally alter the code response. The entire regression suite is run prior to permanently incorporating any software modification or addition. When code modifications alter test problem results, the specific cause must be determined and fully understood before the software changes and revised test answers can bemore » incorporated. The regression suite is executed on LLNL platforms using a Python script and an associated data file. The user specifies the DYNA3D or ParaDyn executable, number of processors to use, test problems to run, and other options to the script. The data file details how each problem and its answer extraction scripts are executed. For each problem in the regression suite there exists an input deck, an eight-processor partition file, an answer file, and various extraction scripts. These scripts assemble a temporary answer file in a specific format from the simulation results. The temporary and stored answer files are compared to a specific level of numerical precision, and when differences are detected the test problem is flagged as failed. Presently, numerical results are stored and compared to 16 digits. At this accuracy level different processor types, compilers, number of partitions, etc. impact the results to various degrees. Thus, for consistency purposes the regression suite is run with ParaDyn using 8 processors on machines with a specific processor type (currently the Intel Xeon E5530 processor). For non-parallel regression problems, i.e., the two XFEM problems, DYNA3D is used instead. When environments or platforms change, executables using the current source code and
Gillingham, Wayne; Holt, Alec; Gillies, John
The technology sector of healthcare is entering a new evolutionary phase. The medical community has an obligation to the public to provide the safest, most effective healthcare possible. This is more achievable with the use of computer technology at the point of care, and small, portable devices could fulfil this role. A PriceWaterhouse Coopers 2001 survey on information technology in physician practices found that 60% of respondents say that physicians in their organisation use personal digital assistants (PDAs), compared with 26% in the 2000 technology survey. This trend is expected to continue to the point where these devices will have their position on a physician s desk next to the stethoscope. Once this electronic evolution occurs, doctors will be able to practice medicine with greater ease and safety. In our opinion, the new generation of PDA mobile devices will be the tools to enable a transformation of healthcare to a paperless, wireless world. This article focuses on uses of PDAs in healthcare, whether by the registrar, consultant, nurse, student, teacher, patient, medical or surgical director. Current PDA healthcare software is categorised and discussed in the following five groups: 1) reference/text book; 2) calculator; 3) patient management/logbook; 4) personal clinical/study notebook; 5) utility software.
Lin, Jen-Chiun; Su, Mei-Ju; Cheng, Po-Hsun; Weng, Yung-Chien; Chen, Sao-Jie; Lai, Jin-Shin; Lai, Feipei
This paper illustrates a sustained conceptual service quality improvement process for the management of software development within a healthcare enterprise. Our proposed process is revised from Niland's healthcare quality information system (HQIS). This process includes functions to survey the satisfaction of system functions, describe the operation bylaws on-line, and provide on-demand training. To achieve these goals, we integrate five information systems in National Taiwan University Hospital, including healthcare information systems, health quality information system, requirement management system, executive information system, and digital learning system, to form a full Deming cycle. A preliminary user satisfaction survey showed that our outpatient information system scored an average of 71.31 in 2006.
Software inspections are a set of formal technical review procedures held at selected key points during software development in order to find defects in software documents--is described in terms of history, participants, tools, procedures, statistics, and database analysis.
Kumar, Ajit; Maskara, Reena; Maskara, Sanjeev; Chiang, I-Jen
The aim of this study is to conceptualize a novel approach, which facilitates us to design prototype interfaces for healthcare software. Concepts and techniques from various disciplines were used to conceptualize an interface design approach named MORTARS (Map Original Rhetorical To Adapted Rhetorical Situation). The concepts and techniques included in this approach are (1) rhetorical situation - a concept of philosophy provided by Bitzer (1968); (2) move analysis - an applied linguistic technique provided by Swales (1990) and Bhatia (1993); (3) interface design guidelines - a cognitive and computer science concept provided by Johnson (2010); (4) usability evaluation instrument - an interface evaluation questionnaire provided by Lund (2001); (5) user modeling via stereotyping - a cognitive and computer science concept provided by Rich (1979). A prototype interface for outpatient clinic software was designed to introduce the underlying concepts of MORTARS. The prototype interface was evaluated by thirty-two medical informaticians. The medical informaticians found the designed prototype interface to be useful (73.3%), easy to use (71.9%), easy to learn (93.1%), and satisfactory (53.2%). MORTARS approach was found to be effective in designing the prototype user interface for the outpatient clinic software. This approach might be further used to design interfaces for various software pertaining to healthcare and other domains. Copyright © 2014 Elsevier Inc. All rights reserved.
Griffin, Ann; Viney, Rowena; Welland, Trevor; Gafson, Irene
Objectives We present a national evaluation of the impact of independent verification visits (IVVs) performed by National Health Service (NHS) England as part of quality assuring medical revalidation. Organisational visits are central to NHS quality assurance. They are costly, yet little empirical research evidence exists concerning their impact, and what does exist is conflicting. Setting The focus was on healthcare providers in the NHS (in secondary care) and private sector across England, who were designated bodies (DBs). DBs are healthcare organisations that have a statutory responsibility, via the lead clinician, the responsible officer (RO), to implement medical revalidation. Participants All ROs who had undergone an IVV in England in 2014 and 2015 were invited to participate. 46 ROs were interviewed. Ethnographic data were gathered at 18 observations of the IVVs and 20 IVV post visit reports underwent documentary analysis. Primary and secondary outcome measures Primary outcomes were the findings pertaining to the effectiveness of the IVV system in supporting the revalidation processes at the DBs. Secondary outcomes were methodological, relating to the Model for Understanding Success in Quality (MUSIQ) and how its application to the IVV reveals the relevance of contextual factors described in the model. Results The impact of the IVVs varied by DB according to three major themes: the personal context of the RO; the organisational context of the DB; and the visit and its impact. ROs were largely satisfied with visits which raised the status of appraisal within their organisations. Inadequate or untimely feedback was associated with dissatisfaction. Conclusions Influencing teams whose prime responsibility is establishing processes and evaluating progress was crucial for internal quality improvement. Visits acted as a nudge, generating internal quality review, which was reinforced by visit teams with relevant expertise. Diverse team membership, knowledge transfer
Griffin, Ann; McKeown, Alex; Viney, Rowena; Rich, Antonia; Welland, Trevor; Gafson, Irene; Woolf, Katherine
We present a national evaluation of the impact of independent verification visits (IVVs) performed by National Health Service (NHS) England as part of quality assuring medical revalidation. Organisational visits are central to NHS quality assurance. They are costly, yet little empirical research evidence exists concerning their impact, and what does exist is conflicting. The focus was on healthcare providers in the NHS (in secondary care) and private sector across England, who were designated bodies (DBs). DBs are healthcare organisations that have a statutory responsibility, via the lead clinician, the responsible officer (RO), to implement medical revalidation. All ROs who had undergone an IVV in England in 2014 and 2015 were invited to participate. 46 ROs were interviewed. Ethnographic data were gathered at 18 observations of the IVVs and 20 IVV post visit reports underwent documentary analysis. Primary outcomes were the findings pertaining to the effectiveness of the IVV system in supporting the revalidation processes at the DBs. Secondary outcomes were methodological, relating to the Model for Understanding Success in Quality (MUSIQ) and how its application to the IVV reveals the relevance of contextual factors described in the model. The impact of the IVVs varied by DB according to three major themes: the personal context of the RO; the organisational context of the DB; and the visit and its impact. ROs were largely satisfied with visits which raised the status of appraisal within their organisations. Inadequate or untimely feedback was associated with dissatisfaction. Influencing teams whose prime responsibility is establishing processes and evaluating progress was crucial for internal quality improvement. Visits acted as a nudge, generating internal quality review, which was reinforced by visit teams with relevant expertise. Diverse team membership, knowledge transfer and timely feedback made visits more impactful. Published by the BMJ Publishing Group
Summarized: The Issue: 6/2/20118 Software defects are currently a fact of life Software defects are avenues of security vulnerabilities that cyber ... criminals , terrorists, or hostile nations can exploit. We (THE ENTIRE INDUSTY) need to change the way we build systems Decrease the number of defects
Bata, Seth A; Richardson, Terry
Health systems across the nation are recovering from massive financial and resource investments in electronic health record applications. In the midst of these recovery efforts, implementations of new care models, including accountable care organizations and population health initiatives, are underway. The shift from fee-for-service to fee-for-outcomes and fee-for-value payment models calls for care providers to work in new ways. It also changes how physicians are compensated and reimbursed. These changes necessitate that healthcare systems further invest in information technology solutions. Selecting which information technology (IT) projects are of most value is vital, especially in light of recent expenditures. Return-on-investment analysis is a powerful tool used in various industries to select the most appropriate IT investments. It has proven vital in selecting, justifying, and implementing software projects. Other financial metrics, such as net present value, economic value added, and total economic impact, also quantify the success of expenditures on information systems. This paper extends the concept of quantifying project value to include clinical outcomes and nonfinancial value as investment returns, applying a systematic approach to healthcare software projects. We term this inclusive approach Value of Investment. It offers a necessary extension for application in clinical settings where a strictly financial view may fall short in providing a complete picture of important benefits. This paper outlines the Value of Investment process and its attributes, and uses illustrative examples to explore the efficacy of this methodology within a midsized health system.
Bata, Seth A.; Richardson, Terry
Health systems across the nation are recovering from massive financial and resource investments in electronic health record applications. In the midst of these recovery efforts, implementations of new care models, including accountable care organizations and population health initiatives, are underway. The shift from fee-for-service to fee-for-outcomes and fee-for-value payment models calls for care providers to work in new ways. It also changes how physicians are compensated and reimbursed. These changes necessitate that healthcare systems further invest in information technology solutions. Selecting which information technology (IT) projects are of most value is vital, especially in light of recent expenditures. Return-on-investment analysis is a powerful tool used in various industries to select the most appropriate IT investments. It has proven vital in selecting, justifying, and implementing software projects. Other financial metrics, such as net present value, economic value added, and total economic impact, also quantify the success of expenditures on information systems. This paper extends the concept of quantifying project value to include clinical outcomes and nonfinancial value as investment returns, applying a systematic approach to healthcare software projects. We term this inclusive approach Value of Investment. It offers a necessary extension for application in clinical settings where a strictly financial view may fall short in providing a complete picture of important benefits. This paper outlines the Value of Investment process and its attributes, and uses illustrative examples to explore the efficacy of this methodology within a midsized health system. PMID:29618963
Kerns, J; Yaldo, D
Purpose: Routine linac quality assurance (QA) tests have become complex enough to require automation of most test analyses. A new data analysis software library was built that allows physicists to automate routine linear accelerator quality assurance tests. The package is open source, code tested, and benchmarked. Methods: Images and data were generated on a TrueBeam linac for the following routine QA tests: VMAT, starshot, CBCT, machine logs, Winston Lutz, and picket fence. The analysis library was built using the general programming language Python. Each test was analyzed with the library algorithms and compared to manual measurements taken at the timemore » of acquisition. Results: VMAT QA results agreed within 0.1% between the library and manual measurements. Machine logs (dynalogs & trajectory logs) were successfully parsed; mechanical axis positions were verified for accuracy and MLC fluence agreed well with EPID measurements. CBCT QA measurements were within 10 HU and 0.2mm where applicable. Winston Lutz isocenter size measurements were within 0.2mm of TrueBeam’s Machine Performance Check. Starshot analysis was within 0.2mm of the Winston Lutz results for the same conditions. Picket fence images with and without a known error showed that the library was capable of detecting MLC offsets within 0.02mm. Conclusion: A new routine QA software library has been benchmarked and is available for use by the community. The library is open-source and extensible for use in larger systems.« less
Annis, S-L; Zeng, G; Wu, X; Macpherson, M
A program has been developed in MATLAB for use in quality assurance of treatment planning of radiation therapy. It analyzes patient DVH files and compiles dose volume data for review, trending, comparison and analysis. Patient DVH files are exported from the Eclipse treatment planning system and saved according to treatment sites and date. Currently analysis is available for 4 treatment sites; Prostate, Prostate Bed, Lung, and Upper GI, with two functions for data report and analysis: patient-specific and organ-specific. The patient-specific function loads one patient DVH file and reports the user-specified dose volume data of organs and targets. These data can be compiled to an external file for a third party analysis. The organ-specific function extracts a requested dose volume of an organ from the DVH files of a patient group and reports the statistics over this population. A graphical user interface is utilized to select clinical sites, function and structures, and input user's requests. We have implemented this program in planning quality assurance at our center. The program has tracked the dosimetric improvement in GU sites after VMAT was implemented clinically. It has generated dose volume statistics for different groups of patients associated with technique or time range. This program allows reporting and statistical analysis of DVH files. It is an efficient tool for the planning quality control in radiation therapy. © 2012 American Association of Physicists in Medicine.
Simon, Christian; Caballero, Carmela
It is without question in the best interest of our patients, if we can identify ways to improve the quality of care we deliver to them. Great progress has been made within the last 25 years in terms of development and implementation of quality-assurance (QA) platforms and quality improvement programs for surgery in general, and within this context for head and neck surgery. As of now, we have successfully identified process indicators that impact outcome of our patients and the quality of care we deliver as surgeons. We have developed risk calculators to determine the risk for complications of individual surgical patients. We have created perioperative guidelines for complex head and neck procedures. We have in Europe and North America created audit registries that can gather and analyze data from institutions across the world to better understand which processes need change to obtain good outcomes and improve quality of care. QA platforms can be tested within the clearly defined environment of prospective clinical trials. If positive, such programs could be rolled out within national healthcare systems, if feasible. Testing quality programs in clinical trials could be a versatile tool to help head neck cancer patients benefit directly from such initiatives on a global level.
31 4.5.1 Market Analysis...competitive market analysis to assess the tool potential. The final transition targets were selected and expressed along with our research on the topic...public release milestones. Details of our testing methodology is in our Software Test Plan deliv- erable, CP- STP -0001. A summary of this approach is
proof in mathematics. For example, consider the proof of the Pythagorean Theorem illustrated at: http://www.cut-the-knot.org/ pythagoras / where 112...methods and tools have made significant progress in their ability to model software designs and prove correctness theorems about the systems modeled...assumption criticality” or “ theorem root set size” SITAPS detects potentially brittle verification cases. SITAPS provides tools and techniques that
Thomas, V; Daly, M K; Cakiner-Egilmez, T; Baker, E
Given the Veterans Affairs Boston Healthcare System's recent introduction of single-use Tonosafe disposable tonometer prisms as an alternative to Goldmann applanation tonometers (GATs), this study had two aims: to conduct a large-scale quality assurance trial to assess the reliability of intraocular pressure (IOP) measurements of the Tonosafe disposable tonometer compared with GAT, particularly at extremes of pressure; to evaluate the suitability of Tonosafe disposable tonometer prisms as an acceptable substitute for GATs and for clinic-wide implementation in an academic tertiary referral setting. Ophthalmology resident physicians measured the IOPs of patients in general and specialty eye clinics with the Tonosafe disposable tonometer and GAT. Tonosafe test-retest reliability data were also collected. A retrospective review of patient charts and data analysis were performed to determine the reliability of measurements. The IOPs of 652 eyes (326 patients) were measured with both GAT and Tonosafe, with a range of 3-34 mm Hg. Linear regression analysis showed R=0.93, slope=0.91, both of which supported the proposed hypothesis, and the y-intercept=-1.05 was significantly different from the hypothesized value. The Tonosafe test-retest repeatability (40 eyes of 40 patients), r=0.977, was very high, which was further supported by linear regression slope=0.993, y-intercept=0.118, and a Tonosafe repeatability coefficient of 2.06, similar to GAT repeatability. The IOP measurements by Tonosafe disposable prisms correlated closely with Goldmann measurements, with similar repeated measurement variability to GAT. This suggests that the Tonosafe is an acceptable substitute for GAT to measure IOP in ophthalmology clinic settings.
Thomas, V; Daly, M K; Cakiner-Egilmez, T; Baker, E
Purpose Given the Veterans Affairs Boston Healthcare System's recent introduction of single-use Tonosafe disposable tonometer prisms as an alternative to Goldmann applanation tonometers (GATs), this study had two aims: to conduct a large-scale quality assurance trial to assess the reliability of intraocular pressure (IOP) measurements of the Tonosafe disposable tonometer compared with GAT, particularly at extremes of pressure; to evaluate the suitability of Tonosafe disposable tonometer prisms as an acceptable substitute for GATs and for clinic-wide implementation in an academic tertiary referral setting. Methods Ophthalmology resident physicians measured the IOPs of patients in general and specialty eye clinics with the Tonosafe disposable tonometer and GAT. Tonosafe test–retest reliability data were also collected. A retrospective review of patient charts and data analysis were performed to determine the reliability of measurements. Results The IOPs of 652 eyes (326 patients) were measured with both GAT and Tonosafe, with a range of 3–34 mm Hg. Linear regression analysis showed R=0.93, slope=0.91, both of which supported the proposed hypothesis, and the y-intercept=−1.05 was significantly different from the hypothesized value. The Tonosafe test–retest repeatability (40 eyes of 40 patients), r=0.977, was very high, which was further supported by linear regression slope=0.993, y-intercept=0.118, and a Tonosafe repeatability coefficient of 2.06, similar to GAT repeatability. Conclusions The IOP measurements by Tonosafe disposable prisms correlated closely with Goldmann measurements, with similar repeated measurement variability to GAT. This suggests that the Tonosafe is an acceptable substitute for GAT to measure IOP in ophthalmology clinic settings. PMID:21455241
Carnevale, Anthony P.; Smith, Nicole; Gulish, Artem; Beach, Bennett H.
This report, provides detailed analyses and projections of occupations in healthcare fields, and wages earned. In addition, the important skills and work values associated with workers in those fields of healthcare are discussed. Finally, the authors analyze the implications of research findings for the racial, ethnic, and class diversity of the…
Narayanasamy, Ganesh; Zalman, Travis; Ha, Chul S; Papanikolaou, Niko; Stathakis, Sotirios
The purpose of this study is to evaluate the use of the Dosimetry Check system for patient-specific IMRT QA. Typical QA methods measure the dose in an array dosimeter surrounded by homogenous medium for which the treatment plan has been recomputed. With the Dosimetry Check system, fluence measurements acquired on a portal dosimeter is applied to the patient's CT scans. Instead of making dose comparisons in a plane, Dosimetry Check system produces isodose lines and dose-volume histograms based on the planning CT images. By exporting the dose distribution from the treatment planning system into the Dosimetry Check system, one is able to make a direct comparison between the calculated dose and the planned dose. The versatility of the software is evaluated with respect to the two IMRT techniques - step and shoot and volumetric arc therapy. The system analyzed measurements made using EPID, PTW seven29, and IBA MatriXX, and an intercomparison study was performed. Plans from patients previously treated at our institution with treated anatomical site on brain, head & neck, liver, lung, and prostate were analyzed using Dosimetry Check system for any anatomical site dependence. We have recommendations and possible precautions that may be necessary to ensure proper QA with the Dosimetry Check system.
Kitzmiller, Rebecca; Hunt, Eleanor; Sproat, Sara Breckenridge
It is time for a change in mindset in how nurses operationalize system implementations and manage projects. Computers and systems have evolved over time from unwieldy mysterious machines of the past to ubiquitous computer use in every aspect of daily lives and work sites. Yet, disconcertingly, the process used to implement these systems has not evolved. Technology implementation does not need to be a struggle. It is time to adapt traditional plan-driven implementation methods to incorporate agile techniques. Agility is a concept borrowed from software development and is presented here because it encourages flexibility, adaptation, and continuous learning as part of the implementation process. Agility values communication and harnesses change to an advantage, which facilitates the natural evolution of an adaptable implementation process. Specific examples of agility in an implementation are described, and plan-driven implementation stages are adapted to incorporate relevant agile techniques. This comparison demonstrates how an agile approach enhances traditional implementation techniques to meet the demands of today's complex healthcare environments.
Jabłoński, Michał; Starčuková, Jana; Starčuk, Zenon
Proton magnetic resonance spectroscopy is a non-invasive measurement technique which provides information about concentrations of up to 20 metabolites participating in intracellular biochemical processes. In order to obtain any metabolic information from measured spectra a processing should be done in specialized software, like jMRUI. The processing is interactive and complex and often requires many trials before obtaining a correct result. This paper proposes a jMRUI enhancement for efficient and unambiguous history tracking and file identification. A database storing all processing steps, parameters and files used in processing was developed for jMRUI. The solution was developed in Java, authors used a SQL database for robust storage of parameters and SHA-256 hash code for unambiguous file identification. The developed system was integrated directly in jMRUI and it will be publically available. A graphical user interface was implemented in order to make the user experience more comfortable. The database operation is invisible from the point of view of the common user, all tracking operations are performed in the background. The implemented jMRUI database is a tool that can significantly help the user to track the processing history performed on data in jMRUI. The created tool is oriented to be user-friendly, robust and easy to use. The database GUI allows the user to browse the whole processing history of a selected file and learn e.g. what processing lead to the results, where the original data are stored, to obtain the list of all processing actions performed on spectra.
Lamsweerde, "On an Extension of Dijkstra’s Semaphore Primitives ," Information Processing Letters, 1, North Holland Publishing Co., New York, October...definitions of the allowed operations, and synchronize cooperating tasks through Delay and Continue operations. ) 152 We will now give an example...that the monitor has been successful in solving the problem of exclusive access to the search return data set and at the same time has synchronized
Chen, G; Ahunbay, E; Li, X
Purpose: To develop a quality assurance software tool, ArtQA, capable of automatically checking radiation treatment plan parameters, verifying plan data transfer from treatment planning system (TPS) to record and verify (R&V) system, performing a secondary MU calculation considering the effect of magnetic field from MR-Linac, and verifying the delivery and plan consistency, for online replanning. Methods: ArtQA was developed by creating interfaces to TPS (e.g., Monaco, Elekta), R&V system (Mosaiq, Elekta), and secondary MU calculation system. The tool obtains plan parameters from the TPS via direct file reading, and retrieves plan data both transferred from TPS and recorded during themore » actual delivery in the R&V system database via open database connectivity and structured query language. By comparing beam/plan datasets in different systems, ArtQA detects and outputs discrepancies between TPS, R&V system and secondary MU calculation system, and delivery. To consider the effect of 1.5T transverse magnetic field from MR-Linac in the secondary MU calculation, a method based on modified Clarkson integration algorithm was developed and tested for a series of clinical situations. Results: ArtQA is capable of automatically checking plan integrity and logic consistency, detecting plan data transfer errors, performing secondary MU calculations with or without a transverse magnetic field, and verifying treatment delivery. The tool is efficient and effective for pre- and post-treatment QA checks of all available treatment parameters that may be impractical with the commonly-used visual inspection. Conclusion: The software tool ArtQA can be used for quick and automatic pre- and post-treatment QA check, eliminating human error associated with visual inspection. While this tool is developed for online replanning to be used on MR-Linac, where the QA needs to be performed rapidly as the patient is lying on the table waiting for the treatment, ArtQA can be used as a general
Deftereos, S; Lambrinoudakis, C; Andriopoulos, P; Farmakis, D; Aessopos, A
Beta-thalassaemia is a hereditary disease, the prevalence of which is high in persons of Mediterranean, African, and Southeast Asian ancestry. In Greece it constitutes an important public health problem. Beta-thalassaemia necessitates continuous and complicated health care procedures such as daily chelation; biweekly transfusions; and periodic cardiology, endocrinology, and hepatology evaluations. Typically, different care items are offered in different, often-distant, health care units, which leads to increased patient mobility. This is especially true in rural areas. Medical records of patients suffering from beta-thalassaemia are inevitably complex and grow in size very fast. They are currently paper-based, scattered over all units involved in the care process. This hinders communication of information between health care professionals and makes processing of the medical records difficult, thus impeding medical research. Our objective is to provide an electronic means for recording, communicating, and processing all data produced in the context of the care process of patients suffering from beta-thalassaemia. We have developed - and we present in this paper - Java-based Electronic Healthcare Record (EHCR) software, called JAnaemia. JAnaemia is a general-purpose EHCR application, which can be customized for use in all medical specialties. Customization for beta-thalassaemia has been performed in collaboration with 4 Greek hospitals. To be capable of coping with patient record diversity, JAnaemia has been based on the EHCR architecture proposed in the ENV 13606:1999 standard, published by the CEN/TC251 committee. Compliance with the CEN architecture also ensures that several additional requirements are fulfilled in relation to clinical comprehensiveness; to record sharing and communication; and to ethical, medico-legal, and computational issues. Special care has been taken to provide a user-friendly, form-based interface for data entry and processing. The
Wallasch, T-M; Bek, J; Pabel, R; Modahl, M; Demir, M; Straube, A
This paper examines a new approach to managed healthcare where a network of care providers exchanges patient information through the internet. Integrating networks of clinical specialists and general care providers promises to achieve qualitative and economic improvements in the German healthcare system. In practice, problems related to patient management and data exchange between the managing clinic and assorted caregivers arise. The implementation and use of a cross-spectrum computerized solution for the management of patients and their care is the key for a successful managed healthcare system. This paper documents the managed healthcare of chronic headache patients and the development of an IT-solution capable of providing distributed patient care and case management.
Oh, Sungyoung; Cha, Jieun; Ji, Myungkyu; Kang, Hyekyung; Kim, Seok; Heo, Eunyoung; Han, Jong Soo; Kang, Hyunggoo; Chae, Hoseok; Hwang, Hee
Objectives To design a cloud computing-based Healthcare Software-as-a-Service (SaaS) Platform (HSP) for delivering healthcare information services with low cost, high clinical value, and high usability. Methods We analyzed the architecture requirements of an HSP, including the interface, business services, cloud SaaS, quality attributes, privacy and security, and multi-lingual capacity. For cloud-based SaaS services, we focused on Clinical Decision Service (CDS) content services, basic functional services, and mobile services. Microsoft's Azure cloud computing for Infrastructure-as-a-Service (IaaS) and Platform-as-a-Service (PaaS) was used. Results The functional and software views of an HSP were designed in a layered architecture. External systems can be interfaced with the HSP using SOAP and REST/JSON. The multi-tenancy model of the HSP was designed as a shared database, with a separate schema for each tenant through a single application, although healthcare data can be physically located on a cloud or in a hospital, depending on regulations. The CDS services were categorized into rule-based services for medications, alert registration services, and knowledge services. Conclusions We expect that cloud-based HSPs will allow small and mid-sized hospitals, in addition to large-sized hospitals, to adopt information infrastructures and health information technology with low system operation and maintenance costs. PMID:25995962
Oh, Sungyoung; Cha, Jieun; Ji, Myungkyu; Kang, Hyekyung; Kim, Seok; Heo, Eunyoung; Han, Jong Soo; Kang, Hyunggoo; Chae, Hoseok; Hwang, Hee; Yoo, Sooyoung
To design a cloud computing-based Healthcare Software-as-a-Service (SaaS) Platform (HSP) for delivering healthcare information services with low cost, high clinical value, and high usability. We analyzed the architecture requirements of an HSP, including the interface, business services, cloud SaaS, quality attributes, privacy and security, and multi-lingual capacity. For cloud-based SaaS services, we focused on Clinical Decision Service (CDS) content services, basic functional services, and mobile services. Microsoft's Azure cloud computing for Infrastructure-as-a-Service (IaaS) and Platform-as-a-Service (PaaS) was used. The functional and software views of an HSP were designed in a layered architecture. External systems can be interfaced with the HSP using SOAP and REST/JSON. The multi-tenancy model of the HSP was designed as a shared database, with a separate schema for each tenant through a single application, although healthcare data can be physically located on a cloud or in a hospital, depending on regulations. The CDS services were categorized into rule-based services for medications, alert registration services, and knowledge services. We expect that cloud-based HSPs will allow small and mid-sized hospitals, in addition to large-sized hospitals, to adopt information infrastructures and health information technology with low system operation and maintenance costs.
White, Jonathan; Kirwan, Paul; Lai, Krista; Walton, Jennifer; Ross, Shelley
The use of social networking software has become ubiquitous in our society. The aim of this study was to explore the attitudes and experiences of healthcare professional students using Facebook at our school, to determine if there is a need for development of policy to assist students in this area. A mixed-methods approach was employed, using semistructured interviews to identify themes which were explored using an online survey. A combination of descriptive statistics and thematic analysis was used for analysis. Healthcare professions education programmes at a large Canadian university. Students of medicine, nursing, pharmacy, speech and language pathology, occupational therapy, physical therapy, dentistry, dental hygiene and medical laboratory Science were invited to participate. 14 participants were interviewed, and 682 participants responded to an online survey; the female:male balance was 3 : 1. 14 interviews were analysed in-depth, and 682 students responded to the survey (17% response rate). 93% reported current Facebook use. Themes identified included patterns of use and attitudes to friendship, attitudes to online privacy, breaches of professional behaviour on Facebook and attitudes to guidelines relating to Facebook use. A majority considered posting of the following material unprofessional: use of alcohol/drugs, crime, obscenity/nudity/sexual content, patient/client information, criticism of others. 44% reported seeing such material posted by a colleague, and 27% reported posting such material themselves. A majority of participants agreed that guidelines for Facebook use would be beneficial. Social networking software use, specifically Facebook use, was widespread among healthcare students at our school who responded to our survey. Our results highlight some of the challenges which can accompany the use of this new technology and offer potential insights to help understand the pedagogy and practices of Facebook use in this population, and to help
Xu, Shen; Rogers, Toby; Fairweather, Elliot; Glenn, Anthony; Curran, James; Curcin, Vasa
Data provenance is a technique that describes the history of digital objects. In health data settings, it can be used to deliver auditability and transparency, and to achieve trust in a software system. However, implementing data provenance in analytics software at an enterprise level presents a different set of challenges from the research environments where data provenance was originally devised. In this paper, the challenges of reporting provenance information to the user is presented. Provenance captured from analytics software can be large and complex and visualizing a series of tasks over a long period can be overwhelming even for a domain expert, requiring visual aggregation mechanisms that fit with complex human cognitive activities involved in the process. This research studied how provenance-based reporting can be integrated into a health data analytics software, using the example of Atmolytics visual reporting tool. PMID:29888084
ECLIPSE, an Emerging Standardized Modular, Secure and Affordable Software Toolset in Support of Product Assurance, Quality Assurance and Project Management for the Entire European Space Industry (from Innovative SMEs to Primes and Institutions)
Bennetti, Andrea; Ansari, Salim; Dewhirst, Tori; Catanese, Giuseppe
The development of satellites and ground systems (and the technologies that support them) is complex and demands a great deal of rigor in the management of both the information it relies upon and the information it generates via the performance of well established processes. To this extent for the past fifteen years Sapienza Consulting has been supporting the European Space Agency (ESA) in the management of this information and provided ESA with ECSS (European Cooperation for Space Standardization) Standards based Project Management (PM), Product Assurance (PA) and Quality Assurance (QA) software applications. In 2009 Sapienza recognised the need to modernize, standardizing and integrate its core ECSS-based software tools into a single yet modularised suite of applications named ECLIPSE aimed at: • Fulfilling a wider range of historical and emerging requirements, • Providing a better experience for users, • Increasing the value of the information it collects and manages • Lowering the cost of ownership and operation • Increasing collaboration within and between space sector organizations • Aiding in the performance of several PM, PA, QA, and configuration management tasks in adherence to ECSS standards. In this paper, Sapienza will first present the toolset, and a rationale for its development, describing and justifying its architecture, and basic modules composition. Having defined the toolset architecture, this paper will address the current status of the individual applications. A compliance assessment will be presented for each module in the toolset with respect to the ECSS standard it addresses. Lastly experience from early industry and Institutional users will be presented.
Journal of Chemical Education, 1989
Presented are reviews of two computer software packages for Apple II computers; "Organic Spectroscopy," and "Videodisc Display Program" for use with "The Periodic Table Videodisc." A sample spectrograph from "Organic Spectroscopy" is included. (CW)
The safe and reliable operation of software within civil aviation systems and equipment has historically been assured through the application of rigorous design assurance applied during the software development process. Increasingly, manufacturers ar...
Torfeh, Tarraf; Hammoud, Rabih; McGarry, Maeve; Al-Hammadi, Noora; Perkins, Gregory
To develop and validate a large field of view phantom and quality assurance software tool for the assessment and characterization of geometric distortion in MRI scanners commissioned for radiation therapy planning. A purpose built phantom was developed consisting of 357 rods (6mm in diameter) of polymethyl-methacrylat separated by 20mm intervals, providing a three dimensional array of control points at known spatial locations covering a large field of view up to a diameter of 420mm. An in-house software module was developed to allow automatic geometric distortion assessment. This software module was validated against a virtual dataset of the phantom that reproduced the exact geometry of the physical phantom, but with known translational and rotational displacements and warping. For validation experiments, clinical MRI sequences were acquired with and without the application of a commercial 3D distortion correction algorithm (Gradwarp™). The software module was used to characterize and assess system-related geometric distortion in the sequences relative to a benchmark CT dataset, and the efficacy of the vendor geometric distortion correction algorithms (GDC) was also assessed. Results issued from the validation of the software against virtual images demonstrate the algorithm's ability to accurately calculate geometric distortion with sub-pixel precision by the extraction of rods and quantization of displacements. Geometric distortion was assessed for the typical sequences used in radiotherapy applications and over a clinically relevant 420mm field of view (FOV). As expected and towards the edges of the field of view (FOV), distortion increased with increasing FOV. For all assessed sequences, the vendor GDC was able to reduce the mean distortion to below 1mm over a field of view of 5, 10, 15 and 20cm radius respectively. Results issued from the application of the developed phantoms and algorithms demonstrate a high level of precision. The results indicate that this
Background According to the latest amendment of the Medical Device Directive standalone software qualifies as a medical device when intended by the manufacturer to be used for medical purposes. In this context, the EN 62304 standard is applicable which defines the life-cycle requirements for the development and maintenance of medical device software. A pilot project was launched to acquire skills in implementing this standard in a hospital-based environment (in-house manufacture). Methods The EN 62304 standard outlines minimum requirements for each stage of the software life-cycle, defines the activities and tasks to be performed and scales documentation and testing according to its criticality. The required processes were established for the pre-existent decision-support software FlashDumpComparator (FDC) used during the quality assurance of treatment-relevant beam parameters. As the EN 62304 standard implicates compliance with the EN ISO 14971 standard on the application of risk management to medical devices, a risk analysis was carried out to identify potential hazards and reduce the associated risks to acceptable levels. Results The EN 62304 standard is difficult to implement without proper tools, thus open-source software was selected and integrated into a dedicated development platform. The control measures yielded by the risk analysis were independently implemented and verified, and a script-based test automation was retrofitted to reduce the associated test effort. After all documents facilitating the traceability of the specified requirements to the corresponding tests and of the control measures to the proof of execution were generated, the FDC was released as an accessory to the HIT facility. Conclusions The implementation of the EN 62304 standard was time-consuming, and a learning curve had to be overcome during the first iterations of the associated processes, but many process descriptions and all software tools can be re-utilized in follow-up projects
Höss, Angelika; Lampe, Christian; Panse, Ralf; Ackermann, Benjamin; Naumann, Jakob; Jäkel, Oliver
According to the latest amendment of the Medical Device Directive standalone software qualifies as a medical device when intended by the manufacturer to be used for medical purposes. In this context, the EN 62304 standard is applicable which defines the life-cycle requirements for the development and maintenance of medical device software. A pilot project was launched to acquire skills in implementing this standard in a hospital-based environment (in-house manufacture). The EN 62304 standard outlines minimum requirements for each stage of the software life-cycle, defines the activities and tasks to be performed and scales documentation and testing according to its criticality. The required processes were established for the pre-existent decision-support software FlashDumpComparator (FDC) used during the quality assurance of treatment-relevant beam parameters. As the EN 62304 standard implicates compliance with the EN ISO 14971 standard on the application of risk management to medical devices, a risk analysis was carried out to identify potential hazards and reduce the associated risks to acceptable levels. The EN 62304 standard is difficult to implement without proper tools, thus open-source software was selected and integrated into a dedicated development platform. The control measures yielded by the risk analysis were independently implemented and verified, and a script-based test automation was retrofitted to reduce the associated test effort. After all documents facilitating the traceability of the specified requirements to the corresponding tests and of the control measures to the proof of execution were generated, the FDC was released as an accessory to the HIT facility. The implementation of the EN 62304 standard was time-consuming, and a learning curve had to be overcome during the first iterations of the associated processes, but many process descriptions and all software tools can be re-utilized in follow-up projects. It has been demonstrated that a
Russo, P L; Shaban, R Z; Macbeth, D; Carter, A; Mitchell, B G
Surveillance of healthcare-associated infections is fundamental for infection prevention. The methods and practices for surveillance have evolved as technology becomes more advanced. The availability of electronic surveillance software (ESS) has increased, and yet adoption of ESS is slow. It is argued that ESS delivers savings through automation, particularly in terms of human resourcing and infection prevention (IP) staff time. To describe the findings of a systematic review on the impact of ESS on IP resources. A systematic search was conducted of electronic databases Medline and the Cumulative Index to Nursing and Allied Health Literature published between January 1 st , 2006 and December 31 st , 2016 with analysis using the Newcastle-Ottawa Scale. In all, 2832 articles were reviewed, of which 16 studies met the inclusion criteria. IP resources were identified as time undertaken on surveillance. A reduction in IP staff time to undertake surveillance was demonstrated in 13 studies. The reduction proportion ranged from 12.5% to 98.4% (mean: 73.9%). The remaining three did not allow for any estimation of the effect in terms of IP staff time. None of the studies demonstrated an increase in IP staff time. The results of this review demonstrate that adopting ESS yields considerable dividends in IP staff time relating to data collection and case ascertainment while maintaining high levels of sensitivity and specificity. This has the potential to enable reinvestment into other components of IP to maximize efficient use of scarce IP resources. Copyright © 2017 The Healthcare Infection Society. Published by Elsevier Ltd. All rights reserved.
Chen, Guang-Pei, E-mail: email@example.com; Ahunbay, Ergun; Li, X. Allen
Purpose: To develop an integrated quality assurance (QA) software tool for online replanning capable of efficiently and automatically checking radiation treatment (RT) planning parameters and gross plan quality, verifying treatment plan data transfer from treatment planning system (TPS) to record and verify (R&V) system, performing a secondary monitor unit (MU) calculation with or without a presence of a magnetic field from MR-Linac, and validating the delivery record consistency with the plan. Methods: The software tool, named ArtQA, was developed to obtain and compare plan and treatment parameters from both the TPS and the R&V system database. The TPS data aremore » accessed via direct file reading and the R&V data are retrieved via open database connectivity and structured query language. Plan quality is evaluated with both the logical consistency of planning parameters and the achieved dose–volume histograms. Beams in between the TPS and R&V system are matched based on geometry configurations. To consider the effect of a 1.5 T transverse magnetic field from MR-Linac in the secondary MU calculation, a method based on modified Clarkson integration algorithm was developed and tested for a series of clinical situations. Results: ArtQA has been used in their clinic and can quickly detect inconsistencies and deviations in the entire RT planning process. With the use of the ArtQA tool, the efficiency for plan check including plan quality, data transfer, and delivery check can be improved by at least 60%. The newly developed independent MU calculation tool for MR-Linac reduces the difference between the plan and calculated MUs by 10%. Conclusions: The software tool ArtQA can be used to perform a comprehensive QA check from planning to delivery with conventional Linac or MR-Linac and is an essential tool for online replanning where the QA check needs to be performed rapidly.« less
Chen, Guang-Pei; Ahunbay, Ergun; Li, X Allen
To develop an integrated quality assurance (QA) software tool for online replanning capable of efficiently and automatically checking radiation treatment (RT) planning parameters and gross plan quality, verifying treatment plan data transfer from treatment planning system (TPS) to record and verify (R&V) system, performing a secondary monitor unit (MU) calculation with or without a presence of a magnetic field from MR-Linac, and validating the delivery record consistency with the plan. The software tool, named ArtQA, was developed to obtain and compare plan and treatment parameters from both the TPS and the R&V system database. The TPS data are accessed via direct file reading and the R&V data are retrieved via open database connectivity and structured query language. Plan quality is evaluated with both the logical consistency of planning parameters and the achieved dose-volume histograms. Beams in between the TPS and R&V system are matched based on geometry configurations. To consider the effect of a 1.5 T transverse magnetic field from MR-Linac in the secondary MU calculation, a method based on modified Clarkson integration algorithm was developed and tested for a series of clinical situations. ArtQA has been used in their clinic and can quickly detect inconsistencies and deviations in the entire RT planning process. With the use of the ArtQA tool, the efficiency for plan check including plan quality, data transfer, and delivery check can be improved by at least 60%. The newly developed independent MU calculation tool for MR-Linac reduces the difference between the plan and calculated MUs by 10%. The software tool ArtQA can be used to perform a comprehensive QA check from planning to delivery with conventional Linac or MR-Linac and is an essential tool for online replanning where the QA check needs to be performed rapidly.
The growth in cost and importance of software to NASA has caused NASA to address the improvement of software development across the agency. One of the products of this program is a series of guidebooks that define a NASA concept of the assurance processes which are used in software development. The Software Assurance Guidebook, SMAP-GB-A201, issued in September, 1989, provides an overall picture of the concepts and practices of NASA in software assurance. Lower level guidebooks focus on specific activities that fall within the software assurance discipline, and provide more detailed information for the manager and/or practitioner. This is the Software Configuration Management Guidebook which describes software configuration management in a way that is compatible with practices in industry and at NASA Centers. Software configuration management is a key software development process, and is essential for doing software assurance.
Franchetti, Franz; Sandryhaila, Aliaksei; Johnson, Jeremy R.
In this paper we introduce High Assurance SPIRAL to solve the last mile problem for the synthesis of high assurance implementations of controllers for vehicular systems that are executed in today's and future embedded and high performance embedded system processors. High Assurance SPIRAL is a scalable methodology to translate a high level specification of a high assurance controller into a highly resource-efficient, platform-adapted, verified control software implementation for a given platform in a language like C or C++. High Assurance SPIRAL proves that the implementation is equivalent to the specification written in the control engineer's domain language. Our approach scales to problems involving floating-point calculations and provides highly optimized synthesized code. It is possible to estimate the available headroom to enable assurance/performance trade-offs under real-time constraints, and enables the synthesis of multiple implementation variants to make attacks harder. At the core of High Assurance SPIRAL is the Hybrid Control Operator Language (HCOL) that leverages advanced mathematical constructs expressing the controller specification to provide high quality translation capabilities. Combined with a verified/certified compiler, High Assurance SPIRAL provides a comprehensive complete solution to the efficient synthesis of verifiable high assurance controllers. We demonstrate High Assurance SPIRALs capability by co-synthesizing proofs and implementations for attack detection and sensor spoofing algorithms and deploy the code as ROS nodes on the Landshark unmanned ground vehicle and on a Synthetic Car in a real-time simulator.
Lehmann, Eldon D; Tatti, Patrizio
The purpose of this article is to document some recommended training requirements for health-carers planning to teach using the AIDA interactive educational diabetes simulator. AIDA is a diabetes computer program that permits the interactive simulation of plasma insulin and blood glucose profiles for teaching, demonstration, and self-learning purposes. It has been made freely available, without charge, on the Internet as a noncommercial contribution to continuing diabetes education. Since its launch in 1996 over 200,000 visits have been logged at the AIDA Website--www.2aida.org--and over 40,000 copies of the AIDA program have been downloaded free-of-charge. This report describes various training requirements that are recommended for health-care professionals who are interested in teaching with the software. Intended goals of this article are to answer possible questions from teachers using the program, highlight some minimum recommended training requirements for the software, suggest some "hints and tips" for teaching ideas, explain the importance of performing more studies/trials with the program, overview randomised controlled trial usage of the software, and highlight the importance of obtaining feedback from lesson participants. The recommendations seem to be straightforward and should help in formalising training with the program, as well as in the development of a network of teachers "accredited" to give lessons using the software. This report, together with the previous article (Part 1, Diabetes Technol Ther 2002;4:401-414), highlights the utility of providing guidelines and suggesting recommended training requirements for health-carers planning to make use of educational medical/diabetes software.
Software Quality Metrics: A Software Management Monitoring Method for Air Force Logistics Command in Its Software Quality Assurance Program for the Quantitative Assessment of the System Development Life Cycle under Configuration Management.
pilot systems. Magnitude of the mutant error is classified as: o Program does not compute. o Program computes but does not run test data. o Program...14 Test and Integration ... ............ .. 105 15 The Mapping of SQM to the SDLC ........ ... 108 16 ADS Development .... .............. . 224 17...and funds. While the test phase concludes the normal development cycle, one should realize that with software the development continues in the
White, Jonathan; Kirwan, Paul; Lai, Krista; Walton, Jennifer; Ross, Shelley
Objective The use of social networking software has become ubiquitous in our society. The aim of this study was to explore the attitudes and experiences of healthcare professional students using Facebook at our school, to determine if there is a need for development of policy to assist students in this area. Design A mixed-methods approach was employed, using semistructured interviews to identify themes which were explored using an online survey. A combination of descriptive statistics and thematic analysis was used for analysis. Setting Healthcare professions education programmes at a large Canadian university. Participants Students of medicine, nursing, pharmacy, speech and language pathology, occupational therapy, physical therapy, dentistry, dental hygiene and medical laboratory Science were invited to participate. 14 participants were interviewed, and 682 participants responded to an online survey; the female:male balance was 3 : 1. Results 14 interviews were analysed in-depth, and 682 students responded to the survey (17% response rate). 93% reported current Facebook use. Themes identified included patterns of use and attitudes to friendship, attitudes to online privacy, breaches of professional behaviour on Facebook and attitudes to guidelines relating to Facebook use. A majority considered posting of the following material unprofessional: use of alcohol/drugs, crime, obscenity/nudity/sexual content, patient/client information, criticism of others. 44% reported seeing such material posted by a colleague, and 27% reported posting such material themselves. A majority of participants agreed that guidelines for Facebook use would be beneficial. Conclusions Social networking software use, specifically Facebook use, was widespread among healthcare students at our school who responded to our survey. Our results highlight some of the challenges which can accompany the use of this new technology and offer potential insights to help understand the pedagogy and
This report discusses the design, fabrication, test and evaluation of a Headway Separation Assurance Subsystem (HSAS) capable of reliable, failsafe performance in PRT systems. The items designed include both hardware and software packages. These pack...
Massachusetts Career Development Inst., Springfield.
This booklet is one of six texts from a workplace literacy curriculum designed to assist learners in facing the increased demands of the workplace. The booklet contains five sections that cover the following topics: (1) importance of reliability; (2) meaning of quality assurance; (3) historical development of quality assurance; (4) statistical…
Connell, John; Wenneson, Greg
The Software Engineering Guidebook describes SEPG (Software Engineering Process Group) supported processes and techniques for engineering quality software in NASA environments. Three process models are supported: structured, object-oriented, and evolutionary rapid-prototyping. The guidebook covers software life-cycles, engineering, assurance, and configuration management. The guidebook is written for managers and engineers who manage, develop, enhance, and/or maintain software under the Computer Software Services Contract.
Rushby, John; Xu, Xidong; Rangarajan, Murali; Weaver, Thomas L.
Assurance cases are a method for providing assurance for a system by giving an argument to justify a claim about the system, based on evidence about its design, development, and tested behavior. In comparison with assurance based on guidelines or standards (which essentially specify only the evidence to be produced), the chief novelty in assurance cases is provision of an explicit argument. In principle, this can allow assurance cases to be more finely tuned to the specific circumstances of the system, and more agile than guidelines in adapting to new techniques and applications. The first part of this report (Sections 1-4) provides an introduction to assurance cases. Although this material should be accessible to all those with an interest in these topics, the examples focus on software for airborne systems, traditionally assured using the DO-178C guidelines and its predecessors. A brief survey of some existing assurance cases is provided in Section 5. The second part (Section 6) considers the criteria, methods, and tools that may be used to evaluate whether an assurance case provides sufficient confidence that a particular system or service is fit for its intended use. An assurance case cannot provide unequivocal "proof" for its claim, so much of the discussion focuses on the interpretation of such less-than-definitive arguments, and on methods to counteract confirmation bias and other fallibilities in human reasoning.
Mueller, David S.
The software program, QRev applies common and consistent computational algorithms combined with automated filtering and quality assessment of the data to improve the quality and efficiency of streamflow measurements and helps ensure that U.S. Geological Survey streamflow measurements are consistent, accurate, and independent of the manufacturer of the instrument used to make the measurement. Software from different manufacturers uses different algorithms for various aspects of the data processing and discharge computation. The algorithms used by QRev to filter data, interpolate data, and compute discharge are documented and compared to the algorithms used in the manufacturers’ software. QRev applies consistent algorithms and creates a data structure that is independent of the data source. QRev saves an extensible markup language (XML) file that can be imported into databases or electronic field notes software. This report is the technical manual for version 2.8 of QRev.
... not using it. Contact your doctor and home healthcare team often to review your health condition. * Check ... assurance of their safety and effectiveness. A home healthcare medical device is any product or equipment used ...
The NASA Software Documentation Standard (hereinafter referred to as "Standard") is designed to support the documentation of all software developed for NASA; its goal is to provide a framework and model for recording the essential information needed throughout the development life cycle and maintenance of a software system. The NASA Software Documentation Standard can be applied to the documentation of all NASA software. The Standard is limited to documentation format and content requirements. It does not mandate specific management, engineering, or assurance standards or techniques. This Standard defines the format and content of documentation for software acquisition, development, and sustaining engineering. Format requirements address where information shall be recorded and content requirements address what information shall be recorded. This Standard provides a framework to allow consistency of documentation across NASA and visibility into the completeness of project documentation. The basic framework consists of four major sections (or volumes). The Management Plan contains all planning and business aspects of a software project, including engineering and assurance planning. The Product Specification contains all technical engineering information, including software requirements and design. The Assurance and Test Procedures contains all technical assurance information, including Test, Quality Assurance (QA), and Verification and Validation (V&V). The Management, Engineering, and Assurance Reports is the library and/or listing of all project reports.
Mueller, David S.
The software program, QRev computes the discharge from moving-boat acoustic Doppler current profiler measurements using data collected with any of the Teledyne RD Instrument or SonTek bottom tracking acoustic Doppler current profilers. The computation of discharge is independent of the manufacturer of the acoustic Doppler current profiler because QRev applies consistent algorithms independent of the data source. In addition, QRev automates filtering and quality checking of the collected data and provides feedback to the user of potential quality issues with the measurement. Various statistics and characteristics of the measurement, in addition to a simple uncertainty assessment are provided to the user to assist them in properly rating the measurement. QRev saves an extensible markup language file that can be imported into databases or electronic field notes software. The user interacts with QRev through a tablet-friendly graphical user interface. This report is the manual for version 2.8 of QRev.
The NASA Software Documentation Standard (hereinafter referred to as Standard) can be applied to the documentation of all NASA software. This Standard is limited to documentation format and content requirements. It does not mandate specific management, engineering, or assurance standards or techniques. This Standard defines the format and content of documentation for software acquisition, development, and sustaining engineering. Format requirements address where information shall be recorded and content requirements address what information shall be recorded. This Standard provides a framework to allow consistency of documentation across NASA and visibility into the completeness of project documentation. This basic framework consists of four major sections (or volumes). The Management Plan contains all planning and business aspects of a software project, including engineering and assurance planning. The Product Specification contains all technical engineering information, including software requirements and design. The Assurance and Test Procedures contains all technical assurance information, including Test, Quality Assurance (QA), and Verification and Validation (V&V). The Management, Engineering, and Assurance Reports is the library and/or listing of all project reports.
Schulz, Erich; Barrett, James W.; Price, Colin
As controlled clinical vocabularies assume an increasing role in modern clinical information systems, so the issue of their quality demands greater attention. In order to meet the resulting stringent criteria for completeness and correctness, a quality assurance system comprising a database of more than 500 rules is being developed and applied to the Read Thesaurus. The authors discuss the requirement to apply quality assurance processes to their dynamic editing database in order to ensure the quality of exported products. Sources of errors include human, hardware, and software factors as well as new rules and transactions. The overall quality strategy includes prevention, detection, and correction of errors. The quality assurance process encompasses simple data specification, internal consistency, inspection procedures and, eventually, field testing. The quality assurance system is driven by a small number of tables and UNIX scripts, with “business rules” declared explicitly as Structured Query Language (SQL) statements. Concurrent authorship, client-server technology, and an initial failure to implement robust transaction control have all provided valuable lessons. The feedback loop for error management needs to be short. PMID:9670131
RT-119: Systemic Assurance Technical Report SERC -2015-TR-019-1 31 July 2015 Principal Investigator: Dr. William Scherlis, Carnegie...Contract No. HQ0034-13-D-0004 Task Order 0019, RT 119 Report No. SERC -2015-TR-019-RT119 July 15, 2015 Report Documentation Page Form...Technology The Systems Engineering Research Center ( SERC ) is a federally funded University Affiliated Research Center managed by Stevens Institute
Introduction of Syphilis Point-of-Care Tests, from Pilot Study to National Programme Implementation in Zambia: A Qualitative Study of Healthcare Workers' Perspectives on Testing, Training and Quality Assurance.
Ansbro, Éimhín M; Gill, Michelle M; Reynolds, Joanna; Shelley, Katharine D; Strasser, Susan; Sripipatana, Tabitha; Tshaka Ncube, Alexander; Tembo Mumba, Grace; Terris-Prestholt, Fern; Peeling, Rosanna W; Mabey, David
Syphilis affects 1.4 million pregnant women globally each year. Maternal syphilis causes congenital syphilis in over half of affected pregnancies, leading to early foetal loss, pregnancy complications, stillbirth and neonatal death. Syphilis is under-diagnosed in pregnant women. Point-of-care rapid syphilis tests (RST) allow for same-day treatment and address logistical barriers to testing encountered with standard Rapid Plasma Reagin testing. Recent literature emphasises successful introduction of new health technologies requires healthcare worker (HCW) acceptance, effective training, quality monitoring and robust health systems. Following a successful pilot, the Zambian Ministry of Health (MoH) adopted RST into policy, integrating them into prevention of mother-to-child transmission of HIV clinics in four underserved Zambian districts. We compare HCW experiences, including challenges encountered in scaling up from a highly supported NGO-led pilot to a large-scale MoH-led national programme. Questionnaires were administered through structured interviews of 16 HCWs in two pilot districts and 24 HCWs in two different rollout districts. Supplementary data were gathered via stakeholder interviews, clinic registers and supervisory visits. Using a conceptual framework adapted from health technology literature, we explored RST acceptance and usability. Quantitative data were analysed using descriptive statistics. Key themes in qualitative data were explored using template analysis. Overall, HCWs accepted RST as learnable, suitable, effective tools to improve antenatal services, which were usable in diverse clinical settings. Changes in training, supervision and quality monitoring models between pilot and rollout may have influenced rollout HCW acceptance and compromised testing quality. While quality monitoring was integrated into national policy and training, implementation was limited during rollout despite financial support and mentorship. We illustrate that new
Introduction of Syphilis Point-of-Care Tests, from Pilot Study to National Programme Implementation in Zambia: A Qualitative Study of Healthcare Workers’ Perspectives on Testing, Training and Quality Assurance
Ansbro, Éimhín M.; Gill, Michelle M.; Reynolds, Joanna; Shelley, Katharine D.; Strasser, Susan; Sripipatana, Tabitha; Ncube, Alexander Tshaka; Tembo Mumba, Grace; Terris-Prestholt, Fern; Peeling, Rosanna W.; Mabey, David
Syphilis affects 1.4 million pregnant women globally each year. Maternal syphilis causes congenital syphilis in over half of affected pregnancies, leading to early foetal loss, pregnancy complications, stillbirth and neonatal death. Syphilis is under-diagnosed in pregnant women. Point-of-care rapid syphilis tests (RST) allow for same-day treatment and address logistical barriers to testing encountered with standard Rapid Plasma Reagin testing. Recent literature emphasises successful introduction of new health technologies requires healthcare worker (HCW) acceptance, effective training, quality monitoring and robust health systems. Following a successful pilot, the Zambian Ministry of Health (MoH) adopted RST into policy, integrating them into prevention of mother-to-child transmission of HIV clinics in four underserved Zambian districts. We compare HCW experiences, including challenges encountered in scaling up from a highly supported NGO-led pilot to a large-scale MoH-led national programme. Questionnaires were administered through structured interviews of 16 HCWs in two pilot districts and 24 HCWs in two different rollout districts. Supplementary data were gathered via stakeholder interviews, clinic registers and supervisory visits. Using a conceptual framework adapted from health technology literature, we explored RST acceptance and usability. Quantitative data were analysed using descriptive statistics. Key themes in qualitative data were explored using template analysis. Overall, HCWs accepted RST as learnable, suitable, effective tools to improve antenatal services, which were usable in diverse clinical settings. Changes in training, supervision and quality monitoring models between pilot and rollout may have influenced rollout HCW acceptance and compromised testing quality. While quality monitoring was integrated into national policy and training, implementation was limited during rollout despite financial support and mentorship. We illustrate that new
Ardagna, Claudio A.; Banzi, Massimo; Damiani, Ernesto; El Ioini, Nabil; Frati, Fulvio
Software Assurance (SwA) is a complex concept that involves different stages of a software development process and may be defined differently depending on its focus, as for instance software quality, security, or dependability. In Computer Science, the term assurance is referred to all activities necessary to provide enough confidence that a software product will satisfy its users’ functional and non-functional requirements.
generation process. 4.1 Parameterized Safety Analysis We can do a qualitative analysis as well and ask questions like “ what has contributed to this...the probability of interception p1 in the previous example, we can determine what impact those parameters have on the probability of hazardous...assumed that the AEG is traversed top-down and left-to-right and only once to produce a particular event trace Randomized decisions about what
Being heavily visually-oriented, I am a firm believer in communication and conveying emotions through the art of color, motion, and transformation. A four-part online training course was created in PowerPoint and needed to be translated over into a Flash format. Issues with the Powerpoint were that the size of the files caused noticeable delays when placed online, there were compatibility issues, and from a composition and design perspective, color schemes and layout left much to be desired. High contrast, pixilated yellow text spiraling and flying on to a background of overly rich hues of blue with cheesy gradient patterns was just not appeasing to my eye, along with the menu directory buttons located at the top resembling blue pills of NyQuil on top of a stale gray border that had nothing to do with the background. The course itself is extremely broad and verbose, and will get monotonous very soon after starting. Moving about the course was very cumbersome as well. efficient by drastically reducing the size (The file size of all four parts of the actual course combined will ultimately not even be a fifth as big as one part of the original PowerPoint alone!); along with that, the course was made to be more interactive and user-friendly, as well as pleasing to the eye. Upon being viewed by fellow co-workers, nothing but positive feedback has been received. When beginning the presentation, onscreen comes a 3' 2" chubby, balding professor, who is a master in his knowledge of Programmable Logic Controllers (PLCs). He introduces himself and presents all of his vast knowledge over PLCs in a fun and innovative manner, making it much easier to acquire the information presented. A Scene Selection feature has been added making it a lot easier to jump from part to part and the back and forth arrows are much easier to utilize, and they are both less obtrusive than its PowerPoint predecessor. The user can also go at their pace, as the presentation pauses after at the end of each statement. computer animation, it was.. .still. ..animation done on a computer. I was able to incorporate my artistic talent and intuitive creativity into it, one thing I am very proficient at doing when it comes to what I do and what I will do in my profession as an artist/computer animator. At first, I felt that there was no place for an artist within a faculty of scientists, engineers, chemists, mathematicians, and programmers, but I managed to fit in quite successfully. My task was to convert the course into a Flash format, which would make it much more My project surprisingly somewhat dealt with my field of interest-though it was not
In 1996, the Joint Commission International (JCI), which is a partnership between the Joint Commission on Accreditation of Healthcare Organizations and Quality Healthcare Resources, Inc., became one of the contractors of the Quality Assurance Project (QAP). JCI recognizes the link between accreditation and quality, and uses a collaborative approach to help a country develop national quality standards that will improve patient care, satisfy patient-centered objectives, and serve the interest of all affected parties. The implementation of good standards provides support for the good performance of professionals, introduces new ideas for improvement, enhances the quality of patient care, reduces costs, increases efficiency, strengthens public confidence, improves management, and enhances the involvement of the medical staff. Such good standards are objective and measurable; achievable with current resources; adaptable to different institutions and cultures; and demonstrate autonomy, flexibility, and creativity. The QAP offers the opportunity to approach accreditation through research efforts, training programs, and regulatory processes. QAP work in the area of accreditation has been targeted for Zambia, where the goal is to provide equal access to cost-effective, quality health care; Jordan, where a consensus process for the development of standards, guidelines, and policies has been initiated; and Ecuador, where JCI has been asked to help plan an approach to the evaluation and monitoring of the health care delivery system.
Sallis, E.; Hingley, P.
This paper discusses the increasing interest in quality assurance in British education including its measurement and management through the introduction of a quality assurance system. The reasons and benefits of beginning a quality assurance system are discussed, and questions of what constitutes quality, whether it is quality in fact…
Papers are presented on the following topics: measurement of software technology, recent studies of the Software Engineering Lab, software management tools, expert systems, error seeding as a program validation technique, software quality assurance, software engineering environments (including knowledge-based environments), the Distributed Computing Design System, and various Ada experiments.
For many years, software quality assurance lagged behind hardware quality assurance in terms of methods, metrics, and successful results. New approaches such as Quality Function Deployment (QFD) the ISO 9000-9004 standards, the SEI maturity levels, and Total Quality Management (TQM) are starting to attract wide attention, and in some cases to bring software quality levels up to a parity with manufacturing quality levels. Since software is on the critical path for many engineered products, and for internal business systems as well, the new approaches are starting to affect global competition and attract widespread international interest. It can be hypothesized thatmore » success in mastering software quality will be a key strategy for dominating global software markets in the 21st century.« less
requirement for increased emphasis on software quality assurance has lead to the creation of various methods of verification and validation. Experience...result was a vast array of methods , systems, languages and automated tools to assist in the process. Given that the primary role of quality assurance is...Unfortunately, there is no single method , tool or technique that can insure accurate, reliable and cost effective software. Therefore, government and industry
electromechanical (EEE) parts, software, design and workmanship, work instructions, manufacturing and tooling, cleanrooms, electrostatic discharge ...T9001B. An external group, called the Evaluation and Assessment Team, made up of product assurance subject matter experts from NSWC Corona performs...NSWC, Corona and SSP Technical Branch(es). The FTPE, performed every 3 years, is an objective evaluation of facility performance to assure proper
The Healthcare Cost and Utilization Project is a family of health care databases and related software tools and products developed through a Federal-State-Industry partnership and sponsored by the Agency for Healthcare Research and Quality.
Baker, D. B.; Masys, D. R.; Jones, R. L.; Barnhart, R. M.
The need for security protection in Internet-based healthcare applications is generally acknowledged. Most healthcare applications that use the Internet have at least implemented some kind of encryption. Most applications also enforce user authentication and access control policies, and many audit user actions. However, most fall short on providing strong assurances that the security mechanisms are behaving as expected and that they cannot be subverted. While no system can claim to be totally "bulletproof," PCASSO provides assurance of correct operation through formal, disciplined design and development methodologies, as well as through functional and penetration testing. Through its security mechanisms, backed by strong system assurances, PCASSO is demonstrating "safe" use of public data networks for health care. PMID:10566443
Dunn, William R.; Corliss, Lloyd D.
Software safety assurance philosophy and practices at the NASA Ames are discussed. It is shown that, to be safe, software must be error-free. Software developments on two digital flight control systems and two ground facility systems are examined, including the overall system and software organization and function, the software-safety issues, and their resolution. The effectiveness of safety assurance methods is discussed, including conventional life-cycle practices, verification and validation testing, software safety analysis, and formal design methods. It is concluded (1) that a practical software safety technology does not yet exist, (2) that it is unlikely that a set of general-purpose analytical techniques can be developed for proving that software is safe, and (3) that successful software safety-assurance practices will have to take into account the detailed design processes employed and show that the software will execute correctly under all possible conditions.
obtained from Dr. Barry Boehm’s Software 5650, Contract No. F19628-86-C-O001, Engineering Economics  and from T. J. ESD/MITRE Software Center Acquisition...of References 1. Boehm, Barry W., SoJtware Engineering 3. Halstead, M. H., Elements of SoJhtare Economics, Englewood Cliffs, New Science, New York...1983, pp. 639-648. 35 35 - Bibliography Beizer, B., Software System Testing and Pressman , Roger S., Software Engineering:QualtyO Assurance, New York: Van
Palmer, Regina; Labaugh, Modenna
This paper will present data related to software development processes and personnel involvement from the perspective of software quality assurance. We examine eight years of data collected from six projects. Data collected varied by project but usually included defect and fault density with limited use of code metrics, schedule adherence, and budget growth information. The data are a blend of AFSCP 800-14 and suggested productivity measures in Software Metrics: A Practioner's Guide to Improved Product Development. A software quality assurance database tool, SQUID, was used to store and tabulate the data.
To ensure that the NASA Safety and Mission Assurance (SMA) community remains in a position to perform reliable Software Assurance (SA) on NASAs critical software (SW) systems with the software industry rapidly transitioning from waterfall to Agile processes, Terry Wilcutt, Chief, Safety and Mission Assurance, Office of Safety and Mission Assurance (OSMA) established the Agile Benchmarking Team (ABT). The Team's tasks were: 1. Research background literature on current Agile processes, 2. Perform benchmark activities with other organizations that are involved in software Agile processes to determine best practices, 3. Collect information on Agile-developed systems to enable improvements to the current NASA standards and processes to enhance their ability to perform reliable software assurance on NASA Agile-developed systems, 4. Suggest additional guidance and recommendations for updates to those standards and processes, as needed. The ABT's findings and recommendations for software management, engineering and software assurance are addressed herein.
Dana Hausman and Suzanne Woolf were crucial to the successful completion of this edition of the directory. Their teamwork, energy, and dedication...for this directory began in the summer of 1986 with a questionnaire mailed to schools selected from Peterson’s Graduate Programs in Engineering and...Christoper, and Siegel, Stan Software Cost Estimation and Life-Cycle Control by Putnam, Lawrence H. Software Quality Assurance: A Practical Approach by
Payne, Karl Fb; Goodson, Alexander Mc; Tahim, Arpan; Wharrad, Heather J; Fan, Kathleen
The recently launched iBooks 2 from Apple has created a new genre of 'interactive multimedia eBook'. This article aims to dscribe the benefit of the iBook in a medical education and healthcare setting. We discuss the attributes of an iBook as compared with the requirements of the conventional web-based Reusable Learning Object. The structure and user interface within an iBook is highlighted, and the iBook-creating software iBooks Author is discussed in detail. A report of personal experience developing and distributing an iBook for junior trainees in oral and maxillofacial surgery is provided, with discussion of the limitations of this approach and the need for further evidence-based studies.
constrained to the objects it manages A local reference monitor can be maintained, updated, and replaced with minimal effect on the rest of the system...Compositional assurance is the path towards the goal of JIT Assurance Construct individual assurance case for each trusted tcomponen Provide argument that...local policies combine to enforce the overall system policy Composability enables JIT Assurance A component can be patched, upgraded, refreshed
Multinational colleges and universities pose numerous challenges to the traditional models of quality assurance that are designed to validate domestic higher education. When institutions cross international borders, at least two quality assurance protocols are involved. To guard against fraud and abuse, quality assurance in the host country is…
activity by providing a check on the relevance and currency of the process used to develop the MSwA2010 curriculum content. Figure 2 is an expansion of...random oracle model, symmetric crypto primitives, modes of operations, asymmetric crypto primitives (Chapter 5)  Detailed design...encryption, public key encryption, digital signatures, message authentication codes, crypto protocols, cryptanalysis, and further detailed crypto
and International Electrotechnical Commission ( ISO /IEC). ISO /IEC 27002 :2005 Information Technology – Security Techniques – Code of Practice for...Compliance and Policy (CP) practice • [ ISO 2008] ISO 27002 Section 15 Research and identify (or develop) an example of policy language that...Microsoft SDL • [Merkow 2010] Chapters 5, 6, 8 • [ ISO 2008] ISO 27002 Sections 12.1-12.5 Identify practices to mitigate selected risks for sample
Patel, Vikram; Parikh, Rachana; Nandraj, Sunil; Balasubramaniam, Priya; Narayan, Kavita; Paul, Vinod K; Kumar, A K Shiva; Chatterjee, Mirai; Reddy, K Srinath
Successive Governments of India have promised to transform India's unsatisfactory health-care system, culminating in the present government's promise to expand health assurance for all. Despite substantial improvements in some health indicators in the past decade, India contributes disproportionately to the global burden of disease, with health indicators that compare unfavourably with other middle-income countries and India's regional neighbours. Large health disparities between states, between rural and urban populations, and across social classes persist. A large proportion of the population is impoverished because of high out-of-pocket health-care expenditures and suffers the adverse consequences of poor quality of care. Here we make the case not only for more resources but for a radically new architecture for India's health-care system. India needs to adopt an integrated national health-care system built around a strong public primary care system with a clearly articulated supportive role for the private and indigenous sectors. This system must address acute as well as chronic health-care needs, offer choice of care that is rational, accessible, and of good quality, support cashless service at point of delivery, and ensure accountability through governance by a robust regulatory framework. In the process, several major challenges will need to be confronted, most notably the very low levels of public expenditure; the poor regulation, rapid commercialisation of and corruption in health care; and the fragmentation of governance of health care. Most importantly, assuring universal health coverage will require the explicit acknowledgment, by government and civil society, of health care as a public good on par with education. Only a radical restructuring of the health-care system that promotes health equity and eliminates impoverishment due to out-of-pocket expenditures will assure health for all Indians by 2022--a fitting way to mark the 75th year of India
As healthcare organizations look for new and improved ways to reduce costs and still offer quality healthcare, many are turning to the Toyota Production System of doing business. Rather than focusing on cutting personnel and assets, "lean healthcare" looks to improve patient satisfaction through improved actions and processes.
Ampudia-Blasco, Francisco Javier; García-Soidán, Francisco Javier; Rubio Sánchez, Manuela; Phan, Tra-Mi
DiaScope ® is a software to help in individualized prescription of antidiabetic treatment in type 2 diabetes. This study assessed its value and acceptability by different professionals. DiaScope ® was developed based on the ADA-EASD 2012 algorithm and on the recommendation of 12 international diabetes experts using the RAND/UCLA appropriateness method. The current study was performed at a single session. In the first phase, 5 clinical scenarios were evaluated, selecting the most appropriated therapeutic option among 4 possibilities (initial test). In a second phase, the same clinical cases were evaluated with DiaScope ® (final test).Opinion surveys on DiaScope ® were also performed (questionnaire). DiaScope ® changed the selected option 1 or more times in 70.5% of cases. Among 275 evaluated questionnaires, 54.0% strongly agree that DiaScope ® allowed finding easily a similar therapeutic scenario to the corresponding patient, and 52.5 among the obtained answers were clinically plausible. Up to 58.3% will recommend it to a colleague. In particular, primary care physicians with >20 years of professional dedication found with DiaScope ® the most appropriate option for a particular situation against specialists or those with less professional dedication (p<.05). DiaScope ® is an easy to use tool for antidiabetic drug prescription that provides plausible solutions and is especially useful for primary care physicians with more years of professional practice. Copyright © 2017 SEEN. Publicado por Elsevier España, S.L.U. All rights reserved.
This paper summarizes the method of the Computer Sciences Corporation Flight Dynamics Operation (FDO) quality assurance approach to support the National Aeronautics and Space Administration Goddard Space Flight Center Flight Dynamics Support Branch. Historically, a strong need has existed for developing systematic quality assurance using methods that account for the unique nature and environment of satellite Flight Dynamics mission support. Over the past few years FDO has developed and implemented proactive quality assurance processes applied to each of the six phases of the Flight Dynamics mission support life cycle: systems and operations concept, system requirements and specifications, software development support, operations planing and training, launch support, and on-orbit mission operations. Rather than performing quality assurance as a final step after work is completed, quality assurance has been built in as work progresses in the form of process assurance. Process assurance activities occur throughout the Flight Dynamics mission support life cycle. The FDO Product Assurance Office developed process checklists for prephase process reviews, mission team orientations, in-progress reviews, and end-of-phase audits. This paper will outline the evolving history of FDO quality assurance approaches, discuss the tailoring of Computer Science Corporations's process assurance cycle procedures, describe some of the quality assurance approaches that have been or are being developed, and present some of the successful results.
Demontlivault, J.; Cadelec, Jacques
In order to increase the interest in small satellites, their cost must be reduced; reducing product assurance costs induced by quality requirements is a major objective. For a logical approach, small satellites are classified in three main categories: satellites for experimental operations with a short lifetime, operational satellites manufactured in small mass with long lifetime requirements, operational satellites (long lifetime required), of which only a few models are produced. The various requirements as regards the product assurance are examined for each satellite category: general requirements for space approach, reliability, electronic components, materials and processes, quality assurance, documentation, tests, and management. Ideal product assurance system integrates quality teams and engineering teams.
Christensen, Larry C.; Bodey, Michael R.
Addresses the issue of assuring quality software for use in computer-aided instruction and presents a structure by which developers can create quality courseware. Differences between courseware and computer-aided instruction software are discussed, methods for testing software are described, and human factors issues as well as instructional design…
Chaki, Sagar; Gurfinkel, Arie; Wallnau, Kurt; Weinstock, Charles
Proof-carrying code (PCC) provides a 'gold standard' for establishing formal and objective confidence in program behavior. However, in order to extend the benefits of PCC - and other formal certification techniques - to realistic systems, we must establish the correspondence of a mathematical proof of a program's semantics and its actual behavior. In this paper, we argue that assurance cases are an effective means of establishing such a correspondence. To this end, we present an assurance case pattern for arguing that a proof is free from various proof hazards. We also instantiate this pattern for a proof-based mechanism to provide evidence about a generic medical device software.
Ramsay, Christopher M.
NASA (National Aeronautics and Space Administration) relies more and more on software to control, monitor, and verify its safety critical systems, facilities and operations. Since the 1960's there has hardly been a spacecraft (manned or unmanned) launched that did not have a computer on board that provided vital command and control services. Despite this growing dependence on software control and monitoring, there has been no consistent application of software safety practices and methodology to NASA's projects with safety critical software. Led by the NASA Headquarters Office of Safety and Mission Assurance, the NASA Software Safety Standard (STD-18l9.13B) has recently undergone a significant update in an attempt to provide that consistency. This paper will discuss the key features of the new NASA Software Safety Standard. It will start with a brief history of the use and development of software in safety critical applications at NASA. It will then give a brief overview of the NASA Software Working Group and the approach it took to revise the software engineering process across the Agency.
Tausworthe, R. C.
This monograph contains standards for software development and engineering. The book sets forth rules for design, specification, coding, testing, documentation, and quality assurance audits of software; it also contains detailed outlines for the documentation to be produced.
Blumenfeld, S N
This paper outlines the early approaches to quality assurance, and its transition from business to health care. It then describes the development of the more recent trends in quality assurance of Total Quality Management and Continuous Quality Improvement and discusses the strengths and weaknesses of these approaches. The paper then goes on to show how these approaches have been modified for application to peripheral health services in developing countries through the work of the Primary Health Care Operations Research Project and the Quality Assurance Project.
... system (all its elements including hardware and software) must be designed to assure safe operation with... unsafe errors in the software due to human error in the software specification, design, or coding phases... (hardware or software, or both) are used in combination to ensure safety. If a common mode failure exists...
Mallory, S R
The software engineering and quality assurance disciplines are a requisite to the design of safe and effective software-based medical devices. It is in the areas of software methodology and process that the most beneficial application of these disciplines to software development can be made. Software is a product of complex operations and methodologies and is not amenable to the traditional electromechanical quality assurance processes. Software quality must be built in by the developers, with the software verification and validation engineers acting as the independent instruments for ensuring compliance with performance objectives and with development and maintenance standards. The implementation of a software quality assurance program is a complex process involving management support, organizational changes, and new skill sets, but the benefits are profound. Its rewards provide safe, reliable, cost-effective, maintainable, and manageable software, which may significantly speed the regulatory review process and therefore potentially shorten the overall time to market. The use of a trial project can greatly facilitate the learning process associated with the first-time application of a software quality assurance program.
Ramsay, Christopher M.
NASA relies more and more on software to control, monitor, and verify its safety critical systems, facilities and operations. Since the 1960's there has hardly been a spacecraft launched that does not have a computer on board that will provide command and control services. There have been recent incidents where software has played a role in high-profile mission failures and hazardous incidents. For example, the Mars Orbiter, Mars Polar Lander, the DART (Demonstration of Autonomous Rendezvous Technology), and MER (Mars Exploration Rover) Spirit anomalies were all caused or contributed to by software. The Mission Control Centers for the Shuttle, ISS, and unmanned programs are highly dependant on software for data displays, analysis, and mission planning. Despite this growing dependence on software control and monitoring, there has been little to no consistent application of software safety practices and methodology to NASA's projects with safety critical software. Meanwhile, academia and private industry have been stepping forward with procedures and standards for safety critical systems and software, for example Dr. Nancy Leveson's book Safeware: System Safety and Computers. The NASA Software Safety Standard, originally published in 1997, was widely ignored due to its complexity and poor organization. It also focused on concepts rather than definite procedural requirements organized around a software project lifecycle. Led by NASA Headquarters Office of Safety and Mission Assurance, the NASA Software Safety Standard has recently undergone a significant update. This new standard provides the procedures and guidelines for evaluating a project for safety criticality and then lays out the minimum project lifecycle requirements to assure the software is created, operated, and maintained in the safest possible manner. This update of the standard clearly delineates the minimum set of software safety requirements for a project without detailing the implementation for those
Loewy, Erich H.
Despite the fact that most American physicians, at least until around the 1970s, stood in the way of developing a universal healthcare system, most are generally not happy with the current state of healthcare – or its lack thereof – today. The primary reasons for this general unhappiness are that insurance companies and managed care have successfully conspired to remove much of the physician's autonomy (via imposed time constraints, burdensome paperwork, the time-consuming chore of having to defend going against stringent treatment algorithms that are often inappropriate for some patients) and the satisfaction of knowing their patients. Few physicians in managed care organizations (MCOs) are able to practice without constant and blindly algorithmic interference concerning the diagnostic tests and therapeutic interventions they order. As copayments have increased, they often find that patients, even though “covered,” cannot afford the therapy they deem necessary. While physicians expect to earn sufficient to pay back their not insignificant educational debts, provide their children with help through college, and assure retirements sufficient for themselves and their spouses, these should not be considered unreasonable expectations. Most physicians today do favor universal healthcare – to the point of having included such language in their various professional codes of ethics (which, perversely enough, bioethicists as a group have failed to do). Contrary to the claims of our colleagues, Altom and Churchill, physicians seem to be genuinely frustrated as to what else they can do to change the current inequitable system. PMID:17435646
Cogliati, Joshua Joseph
This report discusses the quality assurance activities needed to raise the Quality Level of Risk Analysis in a Virtual Environment (RAVEN) from Quality Level 3 to Quality Level 2. This report also describes the general RAVEN quality assurance activities. For improving the quality, reviews of code changes have been instituted, more parts of testing have been automated, and improved packaging has been created. For upgrading the quality level, requirements have been created and the workflow has been improved.
Purpose The purpose of this research study was to obtain healthcare executives' perspectives on diversity in executive healthcare leadership. The study focused on identifying perspectives about diversity and its potential impact on the access of healthcare services by people of color. The study also identified perspectives about factors that influence the attainment of executive healthcare roles by people of color. Design/methodology/approach A convenience sample of healthcare executives was obtained. The executives identified themselves as belonging to one of two subgroups, White healthcare executives or executives of color. Participants were interviewed telephonically in a semi-structured format. The interviews were transcribed and entered into a qualitative software application. The data were codified and important themes were identified. Findings The majority of the study participants perceive that diversity of the executive healthcare leadership team is important. There were differences in perspective among the subgroups as it relates to solutions to improve access to healthcare by people of color. There were also differences in perspective among the subgroups, as it relates to explaining the underrepresentation of people of color in executive healthcare leadership roles. Research limitations/implications This research effort benefited from the subject matter expertise of 24 healthcare executives from two states. Expansion of the number of survey participants and broadening the geographical spread of where participants were located may have yielded more convergence and/or more divergence in perspectives about key topics. Practical implications The findings from this research study serve to add to the existing body of literature on diversity in executive healthcare leadership. The findings expand on the importance of key elements in contemporary literature such as diversity, cultural competency and perspectives about the need for representation of people of
... 38 Pensions, Bonuses, and Veterans' Relief 1 2014-07-01 2014-07-01 false Access to quality assurance records and documents within the agency. 17.508 Section 17.508 Pensions, Bonuses, and Veterans' Relief DEPARTMENT OF VETERANS AFFAIRS MEDICAL Confidentiality of Healthcare Quality Assurance Review...
... 38 Pensions, Bonuses, and Veterans' Relief 1 2011-07-01 2011-07-01 false Access to quality assurance records and documents within the agency. 17.508 Section 17.508 Pensions, Bonuses, and Veterans' Relief DEPARTMENT OF VETERANS AFFAIRS MEDICAL Confidentiality of Healthcare Quality Assurance Review...
... 38 Pensions, Bonuses, and Veterans' Relief 1 2010-07-01 2010-07-01 false Access to quality assurance records and documents within the agency. 17.508 Section 17.508 Pensions, Bonuses, and Veterans' Relief DEPARTMENT OF VETERANS AFFAIRS MEDICAL Confidentiality of Healthcare Quality Assurance Review...
... 38 Pensions, Bonuses, and Veterans' Relief 1 2012-07-01 2012-07-01 false Access to quality assurance records and documents within the agency. 17.508 Section 17.508 Pensions, Bonuses, and Veterans' Relief DEPARTMENT OF VETERANS AFFAIRS MEDICAL Confidentiality of Healthcare Quality Assurance Review...
... 38 Pensions, Bonuses, and Veterans' Relief 1 2013-07-01 2013-07-01 false Access to quality assurance records and documents within the agency. 17.508 Section 17.508 Pensions, Bonuses, and Veterans' Relief DEPARTMENT OF VETERANS AFFAIRS MEDICAL Confidentiality of Healthcare Quality Assurance Review...
Denney, Ewen; Pai, Ganesh
The Assurance Case approach is being adopted in a number of safety-mission-critical application domains in the U.S., e.g., medical devices, defense aviation, automotive systems, and, lately, civil aviation. This paradigm refocuses traditional, process-based approaches to assurance on demonstrating explicitly stated assurance goals, emphasizing the use of structured rationale, and concrete product-based evidence as the means for providing justified confidence that systems and software are fit for purpose in safely achieving mission objectives. NASA has also been embracing assurance cases through the concepts of Risk Informed Safety Cases (RISCs), as documented in the NASA System Safety Handbook, and Objective Hierarchies (OHs) as put forth by the Agency's Office of Safety and Mission Assurance (OSMA). This talk will give an overview of the work being performed by the SGT team located at NASA Ames Research Center, in developing technologies and tools to engineer and apply assurance cases in customer projects pertaining to aviation safety. We elaborate how our Assurance Case Automation Toolset (AdvoCATE) has not only extended the state-of-the-art in assurance case research, but also demonstrated its practical utility. We have successfully developed safety assurance cases for a number of Unmanned Aircraft Systems (UAS) operations, which underwent, and passed, scrutiny both by the aviation regulator, i.e., the FAA, as well as the applicable NASA boards for airworthiness and flight safety, flight readiness, and mission readiness. We discuss our efforts in expanding AdvoCATE capabilities to support RISCs and OHs under a project recently funded by OSMA under its Software Assurance Research Program. Finally, we speculate on the applicability of our innovations beyond aviation safety to such endeavors as robotic, and human spaceflight.
Rarick, Heather L.; Godfrey, Sara H.; Kelly, John C.; Crumbley, Robert T.; Wifl, Joel M.
To identify best practices for the improvement of software engineering on projects, NASA's Offices of Chief Engineer (OCE) and Safety and Mission Assurance (OSMA) formed a team led by Heather Rarick and Sally Godfrey to conduct this benchmarking study. The primary goals of the study are to identify best practices that: Improve the management and technical development of software intensive systems; Have a track record of successful deployment by aerospace industries, universities [including research and development (R&D) laboratories], and defense services, as well as NASA's own component Centers; and Identify candidate solutions for NASA's software issues. Beginning in the late fall of 2010, focus topics were chosen and interview questions were developed, based on the NASA top software challenges. Between February 2011 and November 2011, the Benchmark Team interviewed a total of 18 organizations, consisting of five NASA Centers, five industry organizations, four defense services organizations, and four university or university R and D laboratory organizations. A software assurance representative also participated in each of the interviews to focus on assurance and software safety best practices. Interviewees provided a wealth of information on each topic area that included: software policy, software acquisition, software assurance, testing, training, maintaining rigor in small projects, metrics, and use of the Capability Maturity Model Integration (CMMI) framework, as well as a number of special topics that came up in the discussions. NASA's software engineering practices compared favorably with the external organizations in most benchmark areas, but in every topic, there were ways in which NASA could improve its practices. Compared to defense services organizations and some of the industry organizations, one of NASA's notable weaknesses involved communication with contractors regarding its policies and requirements for acquired software. One of NASA's strengths
Tompkins, F. G.
The Office of Management and Budget (OMB) Circular A-71, transmittal Memorandum No. 1, requires that each agency establish a management control process to assure that appropriate administrative, physical and technical safeguards are incorporated into all new computer applications. In addition to security specifications, the management control process should assure that the safeguards are adequate for the application. The security activities that should be integral to the system development process are examined. The software quality assurance process to assure that adequate and appropriate controls are incorporated into sensitive applications is also examined. Security for software packages is also discussed.
de Jong, K.; Trauth, S.L.
The concept of software quality can represent a significant shock to an individual who has been developing software for many years and who believes he or she has been doing a high quality job. The very idea that software includes lines of code and associated documentation is foreign and difficult to grasp, at best. Implementation of a software quality program hinges on the concept that software is a product whose quality needs improving. When this idea is introduced into a technical community that is largely ''self-taught'' and has been producing ''good'' software for some time, a fundamental understanding of themore » concepts associated with software is often weak. Software developers can react as if to say, ''What are you talking about. What do you mean I'm not doing a good job. I haven't gotten any complaints about my code yetexclamation'' Coupling such surprise and resentment with the shock that software really is a product and software quality concepts do exist, can fuel the volatility of these emotions. In this paper, we demonstrate that the concept of software quality can indeed pose a culture shock to developers. We also show that a ''typical'' quality assurance approach, that of imposing a standard and providing inspectors and auditors to assure its adherence, contributes to this shock and detracts from the very goal the approach should achieve. We offer an alternative, adopted through experience, to implement a software quality program: cooperative assistance. We show how cooperation, education, consultation and friendly assistance can overcome this culture shock. 3 refs.« less
Lauenroth, Kim; Metzger, Andreas; Pohl, Klaus
Software Product Line Engineering (SPLE) is a reuse-driven development paradigm that has been applied successfully in information system engineering and other domains. Quality assurance of the reusable artifacts of the product line (e.g. requirements, design, and code artifacts) is essential for successful product line engineering. As those artifacts are reused in several products, a defect in a reusable artifact can affect several products of the product line. A central challenge for quality assurance in product line engineering is how to consider product line variability. Since the reusable artifacts contain variability, quality assurance techniques from single-system engineering cannot directly be applied to those artifacts. Therefore, different strategies and techniques have been developed for quality assurance in the presence of variability. In this chapter, we describe those strategies and discuss in more detail one of those strategies, the so called comprehensive strategy. The comprehensive strategy aims at checking the quality of all possible products of the product line and thus offers the highest benefits, since it is able to uncover defects in all possible products of the product line. However, the central challenge for applying the comprehensive strategy is the complexity that results from the product line variability and the large number of potential products of a product line. In this chapter, we present one concrete technique that we have developed to implement the comprehensive strategy that addresses this challenge. The technique is based on model checking technology and allows for a comprehensive verification of domain artifacts against temporal logic properties.
Ellenbecker, Carol Hall; Porell, Frank W; Samia, Linda; Byleckie, James J; Milburn, Michael
To examine the level of job satisfaction and test a theoretical model of the direct and indirect effects of job satisfaction, and individual nurse and agency characteristics, on intent to stay and retention for home healthcare nurses. A descriptive correlation study of home healthcare nurses in six New England states. Home healthcare nurse job satisfaction self-report data was collected with the HHNJS survey questionnaire & Retention Survey Questionnaire. Based on a structural equation model, job tenure and job satisfaction were the strongest predictors of nurse retention. Understanding the variables associated with home healthcare nurse retention can help agencies retain nurses in a time of severe nurse shortages and increased patient demand. Predicted nursing shortages and increasing demand have made the retention of experienced, qualified nursing staff essential to assure access to high-quality home healthcare services in the future.
Long, John C
Lean Thinking is an integrated approach to designing, doing and improving the work of people that have come together to produce and deliver goods, services and information. Healthcare Lean is based on the Toyota production system and applies concepts and techniques of Lean Thinking to hospitals and physician practices.
Kauk, Justin; Hill, Austin D; Althausen, Peter L
In order for a trauma surgeon to have an intelligent discussion with hospital administrators, healthcare plans, policymakers, or any other physicians, a basic understanding of the fundamentals of healthcare is paramount. It is truly shocking how many surgeons are unable to describe the difference between Medicare and Medicaid or describe how hospitals and physicians get paid. These topics may seem burdensome but they are vital to all business decision making in the healthcare field. The following chapter provides further insight about what we call "the basics" of providing medical care today. Most of the topics presented can be applied to all specialties of medicine. It is broken down into 5 sections. The first section is a brief overview of government programs, their influence on care delivery and reimbursement, and past and future legislation. Section 2 focuses on the compliance, care provision, and privacy statutes that regulate physicians who care for Medicare/Medicaid patient populations. With a better understanding of these obligations, section 3 discusses avenues by which physicians can stay informed of current and pending health policy and provides ways that they can become involved in shaping future legislation. The fourth section changes gears slightly by explaining how the concepts of trade restraint, libel, antitrust legislation, and indemnity relate to physician practice. The fifth, and final, section ties all of components together by describing how physician-hospital alignment can be mutually beneficial in providing patient care under current healthcare policy legislation.
Cheung, Peter P. T.; Tsui, Cecilia B. S.
For higher education reform, most decision-makers aspire to achieving a higher participation rate and a respectable degree of excellence with diversity at the same time. But very few know exactly how. External quality assurance is a fair basis for differentiation but there can be doubt and resistance in some quarters. Stakeholder interests differ…
Subtitle I of the Resource Conservation and Recovery Act, as amended by the Hazardous Waste Disposal Act of 1984, brought underground storage tanks (USTs) under federal regulation. As part of that regulation, Congress directed EPA to develop financial responsibility regulations for UST owners and operators. Congress wanted owners and operators of underground storage tanks (USTs) to show that they have the financial resources to clean up a site if a release occurs, correct environmental damage, and compensate third parties for injury to their property or themselves.Owners and operators have several options: obtain insurance coverage from an insurer or a risk retention group; demonstrate self-insurance using a financial test; obtain corporate guarantees, surety bonds, or letters of credit; place the required amount into a trust fund administered by a third party; or rely on coverage provided by a state financial assurance fund.Information in this data asset includes state documentation to support this requirement. Many states have developed financial assurance funds to help owners and operators meet financial responsibility requirements and to help cover the costs of cleanups. State financial assurance fund programs, which supplement or are a substitute for private insurance, have been especially useful for small-to-medium sized petroleum marketers.EPA requires its Regional Offices to conduct annual reviews of state financial assurance funds. Data is provided by s
Bögel, K; Stöhr, K
The definition of terms and connotation of "Quality", "Quality Assurance" and "Integration" lead to an analysis and understanding of inhibiting and fostering factors of the "Health Triad" of people, animals and environment. Although "Quality" is largely or ultimately determined by the consumer, there are considerable differences as this term is applied by (a) the individual consumer, (b) the dynamic producer defending or gaining markets, (c) those engaged in traditional product manufacturing, or (d) governments setting (minimum) requirements for the sake of free trade. "Quality Assurance" offers cooperation of partners all along the food chain from "pasture to table". The managerial process turned into a continuum of responsibility and agreement on processes and product characteristics. This overcomes the disadvantages of strategies stressing distinct defense barriers. In practice this philosophy of a predominant role of defence barriers proved largely partnership destructive, in that it permitted to shift responsibilities for failures and to claim administrative competence according to momentary situations and interests. "Integrated Quality Assurance" means mutual agreement of two or more partners along the food chain (e. g. feed producers, farmers, animal health industry, veterinarians and food processors) on product characteristics and production methods. It involves essential system elements including facilities, materials, manpower, information, transport, management etc. Different principles and procedures of quality assurance have been introduced in practice, including agriculture and food processing. These different approaches are not mutually exclusive but largely of complementary nature.(ABSTRACT TRUNCATED AT 250 WORDS)
Halford, Vaughn Edward; Ryder, Ann Marie
Effective May 1, 2017, led by a new executive leadership team, Sandia began operating within a new organizational structure. National Technology and Engineering Solutions of Sandia (Sandia’s) Quality Assurance Program (QAP) was established to assign responsibilities and authorities, define workflow policies and requirements, and provide for the performance and assessment of work.
Riches, E; Robson, B
Clinical engagement can achieve lasting change in the delivery of healthcare. In October 2011, Healthcare Improvement Scotland formulated a clinical engagement strategy to ensure that a progressive and sustainable approach to engaging healthcare professionals is firmly embedded in its health improvement and public assurance activities. The strategy was developed using a 90-day process, combining an evidence base of best practice and feedback from semi-structured interviews and focus groups. The strategy aims to create a culture where clinicians view working with Healthcare Improvement Scotland as a worthwhile venture, which offers a number of positive benefits such as training, career development and research opportunities. The strategy works towards developing a respectful partnership between Healthcare Improvement Scotland, the clinical community and key stakeholders whereby clinicians' contributions are recognised in a non-financial reward system. To do this, the organisation needs a sustainable infrastructure and an efficient, cost-effective approach to clinical engagement. There are a number of obstacles to achieving successful clinical engagement and these must be addressed as key drivers in its implementation. The implementation of the strategy is supported by an action and resource plan, and its impact will be monitored by a measurement plan to ensure the organisation reviews its approaches towards clinical engagement.
Ellis, Roger, Ed.
This book, written from a British perspective, presents 17 papers on quality assurance in teaching at the university level. The first eight papers address issues of assuring quality and include: (1) "Quality Assurance for University Teaching; Issues and Approaches" (Roger Ellis); (2) "A British Standard for University…
in-vitro decision to incubate a startup, Lexumo , which is developing a commercial Software as a Service ( SaaS ) vulnerability assessment...LTS Label Transition System MUSE Mining and Understanding Software Enclaves RTEMS Real-Time Executive for Multi-processor Systems SaaS Software ...as a Service SSA Static Single Assignment SWE Software Epistemology UD/DU Def-Use/Use-Def Chains (Dataflow Graph)
Witulski, A.; Austin, R.; Evans, J.; Mahadevan, N.; Karsai, G.; Sierawski, B.; LaBel, K.; Reed, R.; Schrimpf, R.
A systematic approach is presented to constructing a radiation assurance case using Goal Structuring Notation (GSN) for spacecraft containing COTS parts. The GSN paradigm is applied to an SRAM single-event upset experiment board designed to fly on a CubeSat in January 2017. A custom software language for development of a GSN assurance case is under development at Vanderbilt. Construction of a radiation assurance case without use of hardened parts or extensive radiation testing is discussed.
The purpose of this NASA Software Management Guidebook is twofold. First, this document defines the core products and activities required of NASA software projects. It defines life-cycle models and activity-related methods but acknowledges that no single life-cycle model is appropriate for all NASA software projects. It also acknowledges that the appropriate method for accomplishing a required activity depends on characteristics of the software project. Second, this guidebook provides specific guidance to software project managers and team leaders in selecting appropriate life cycles and methods to develop a tailored plan for a software engineering project.
Marnock, M. J.
The protection of intellectual property by a patent, a copyright, or trade secrets is reviewed. The present and future use of computers and software are discussed, along with the governmental uses of software. The popularity of contractual agreements for sale or lease of computer programs and software services is also summarized.
Zygielbaum, Arthur I.
An external review of the Integrated Technology Plan for the Civil Space Program is presented. This review is specifically concerned with the Software Engineering Program. The goals of the Software Engineering Program are as follows: (1) improve NASA's ability to manage development, operation, and maintenance of complex software systems; (2) decrease NASA's cost and risk in engineering complex software systems; and (3) provide technology to assure safety and reliability of software in mission critical applications.
Gendron, Gerald; Roberts, David; Poole, Donold; Aquino, Anna
This paper proposes a cyber security modeling and simulation roadmap to enhance mission assurance governance and establish risk reduction processes within constrained budgets. The term mission assurance stems from risk management work by Carnegie Mellon's Software Engineering Institute in the late 19905. By 2010, the Defense Information Systems Agency revised its cyber strategy and established the Program Executive Officer-Mission Assurance. This highlights a shift from simply protecting data to balancing risk and begins a necessary dialogue to establish a cyber security roadmap. The Military Operations Research Society has recommended a cyber community of practice, recognizing there are too few professionals having both cyber and analytic experience. The authors characterize the limited body of knowledge in this symbiotic relationship. This paper identifies operational and research requirements for mission assurance M&S supporting defense and homeland security. M&S techniques are needed for enterprise oversight of cyber investments, test and evaluation, policy, training, and analysis.
Fowler, J.; Martin, G.
The Healthcare Administrator's Associate is a collection of portable tools designed to support analysis of data retrieved via the Internet from diverse distributed healthcare information systems by means of the InfoSleuth system of distributed software agents. Development of these tools is part of an effort to enhance access to diverse and geographically distributed healthcare data in order to improve the basis upon which administrative and clinical decisions are made. PMID:9357686
Gilliam, David; Powell, John; Haugh, Eric; Bishop, Matt
The NASA Office of Safety and Mission Assurance (OSMA) has funded the Jet Propulsion Laboratory (JPL) with a Center Initiative, 'Reducing Software Security Risk through an Integrated Approach' (RSSR), to address this need. The Initiative is a formal approach to addressing software security in the life cycle through the instantiation of a Software Security Assessment Instrument (SSAI) for the development and maintenance life cycles.
Klepper, Robert; McKenna, Edward G.
Suggests an approach that organizations can take to provide centralized support services for quality assurance in end-user information systems, based on the experiences of a support group at Citicorp Mortgage, Inc. The functions of the support group include user education, software selection, and assistance in testing, implementation, and support…
Dimou, Helen; Kameas, Achilles
Purpose: This paper aims to present a model for the quality assurance of digital educational material that is appropriate for adult education. The proposed model adopts the software quality standard ISO/IEC 9126 and takes into account adult learning theories, Bloom's taxonomy of learning objectives and two instructional design models: Kolb's model…
Rosenthal, L E
Software is the component in a computer system that permits the hardware to perform the various functions that a computer system is capable of doing. The history of software and its development can be traced to the early nineteenth century. All computer systems are designed to utilize the "stored program concept" as first developed by Charles Babbage in the 1850s. The concept was lost until the mid-1940s, when modern computers made their appearance. Today, because of the complex and myriad tasks that a computer system can perform, there has been a differentiation of types of software. There is software designed to perform specific business applications. There is software that controls the overall operation of a computer system. And there is software that is designed to carry out specialized tasks. Regardless of types, software is the most critical component of any computer system. Without it, all one has is a collection of circuits, transistors, and silicone chips.
Bolivar, S. L.; Day, J. L.
The role of the quality assurance liaison, the responsibilities of this position, and the evolutionary changes in duties over the last six years are described. The role of the quality assurance liaison has had a very positive impact on the Los Alamos Yucca Mountain Site Characterization (YW) quality assurance program. Having both technical and quality assurance expertise, the quality assurance liaisons are able to facilitate communications with scientists on quality assurance issues and requirements, thereby generating greater productivity in scientific investigations. The quality assurance liaisons help ensure that the scientific community knows and implements existing requirements, is aware of new or changing regulations, and is able to conduct scientific work within Project requirements. The influence of the role of the quality assurance liaison can be measured by an overall improvement in attitude of the staff regarding quality assurance requirements and improved job performance, as well as a decrease in deficiencies identified during both internal and external audits and surveillances. This has resulted in a more effective implementation of quality assurance requirements.
Wetherholt, Martha S.; Radley, Charles F.
Until recently, NASA did not consider allowing computers total control of flight systems. Human operators, via hardware, have constituted the ultimate safety control. In an attempt to reduce costs, NASA has come to rely more and more heavily on computers and software to control space missions. (For example. software is now planned to control most of the operational functions of the International Space Station.) Thus the need for systematic software safety programs has become crucial for mission success. Concurrent engineering principles dictate that safety should be designed into software up front, not tested into the software after the fact. 'Cost of Quality' studies have statistics and metrics to prove the value of building quality and safety into the development cycle. Unfortunately, most software engineers are not familiar with designing for safety, and most safety engineers are not software experts. Software written to specifications which have not been safety analyzed is a major source of computer related accidents. Safer software is achieved step by step throughout the system and software life cycle. It is a process that includes requirements definition, hazard analyses, formal software inspections, safety analyses, testing, and maintenance. The greatest emphasis is placed on clearly and completely defining system and software requirements, including safety and reliability requirements. Unfortunately, development and review of requirements are the weakest link in the process. While some of the more academic methods, e.g. mathematical models, may help bring about safer software, this paper proposes the use of currently approved software methodologies, and sound software and assurance practices to show how, to a large degree, safety can be designed into software from the start. NASA's approach today is to first conduct a preliminary system hazard analysis (PHA) during the concept and planning phase of a project. This determines the overall hazard potential of
External development of software has oftc n led to unsatisfactory results and great frustration for the assurE 7ce community. Contracts frequently omit critical assuranc 4 processes or the right to oversee software development activitie: At a time when NASA depends more and more on software to in plement critical system functions, combination of three factors ex; cerbate this problem: I ) the ever-increasing trend to acquire rather than develop software in-house, 2) the trend toward performance based contracts, and 3) acquisition vehicles that only state softwar 2 requirements while leaving development standards and assur! ince methodologies up to the contractor. We propose to identify specific methods at d tools that NASA projects can use to mitigate the adverse el ects of the three problems. TWO broad classes of methoddt ols will be explored. The first will be those that provide NASA p ojects with insight and oversight into contractors' activities. The st cond will be those that help projects objectively assess, and thus i nprwe, their software acquisition capability. Of particular interest is the Software Engineering Institute's (SEI) Software Acqt isition Capability Maturity Model (SA-CMMO).
Housch, Helen; Godfrey, Sally
The presentation reviews Agency process requirements and the purpose, benefits, and experiences or seven software engineering processes. The processes include: product integration, configuration management, verification, software assurance, measurement and analysis, requirements management, and planning and monitoring.
Kouloulias, V E
In 1999, the European Organisation for Research and Treatment of Cancer (EORTC), being a European pioneer in the field of cancer research as well as in quality assurance (QA), launched an Emmanuel van der Schueren fellowship for QA in radiotherapy. In this paper, the work that has been done during the first E. van der Schueren fellowship is reported, focusing on four phase III EORTC clinical trials: 22921 for rectal cancer, 22961 and 22991 for prostate cancer and 22922 for breast cancer. A historical review of the QA programme of the EORTC Radiotherapy group during the past 20 years is included.
Software safety and its relationship to other qualities are discussed. It is shown that standard reliability and fault tolerance techniques will not solve the safety problem for the present. A new attitude requires: looking at what you do NOT want software to do along with what you want it to do; and assuming things will go wrong. New procedures and changes to entire software development process are necessary: special software safety analysis techniques are needed; and design techniques, especially eliminating complexity, can be very helpful.
Software assurance is the planned and systematic set of activities that ensures that software processes and products conform to requirements, standards, and procedures. Examples of such activities are the following: code inspections, unit tests, design reviews, performance analyses, construction of traceability matrices, etc. In practice, software development projects have only limited resources (e.g., schedule, budget, and availability of personnel) to cover the entire development effort, of which assurance is but a part. Projects must therefore select judiciously from among the possible assurance activities. At its heart, this can be viewed as an optimization problem; namely, to determine the allocation of limited resources (time, money, and personnel) to minimize risk or, alternatively, to minimize the resources needed to reduce risk to an acceptable level. The end result of the work reported here is a means to optimize quality-assurance processes used in developing software. This is achieved by combining two prior programs in an innovative manner
Jacobs, J.; Deprez, T.; Marchal, G.; Bosmans, H.
MoniQA ("Monitor Quality Assurance") is a new, non-commercial, independent quality assurance software application developed in our medical physics team. It is a complete Java TM - based modular environment for the evaluation of radiological viewing devices and it thus fits in the global quality assurance network of our (film less) radiology department. The purpose of the software tool is to guide the medical physicist through an acceptance protocol and the radiologist through a constancy check protocol by presentation of the necessary test patterns and by automated data collection. Data are then sent to a central management system for further analysis. At the moment more than 55 patterns have been implemented, which can be grouped in schemes to implement protocols (i.e. AAPMtg18, DIN and EUREF). Some test patterns are dynamically created and 'drawn' on the viewing device with random parameters as is the case in a recently proposed new pattern for constancy testing. The software is installed on 35 diagnostic stations (70 monitors) in a film less radiology department. Learning time was very limited. A constancy check -with the new pattern that assesses luminance decrease, resolution problems and geometric distortion- takes only 2 minutes and 28 seconds per monitor. The modular approach of the software allows the evaluation of new or emerging test patterns. We will report on the software and its usability: practicality of the constancy check tests in our hospital and on the results from acceptance tests of viewing stations for digital mammography.
Wild, Christian W.
This work focused on capturing, using, and evolving a qualitative decision support structure across the life cycle of a project. The particular application of this study was towards business process reengineering and the representation of the business process in a set of Business Rules (BR). In this work, we defined a decision model which captured the qualitative decision deliberation process. It represented arguments both for and against proposed alternatives to a problem. It was felt that the subjective nature of many critical business policy decisions required a qualitative modeling approach similar to that of Lee and Mylopoulos. While previous work was limited almost exclusively to the decision capture phase, which occurs early in the project life cycle, we investigated the use of such a model during the later stages as well. One of our significant developments was the use of the decision model during the operational phase of a project. By operational phase, we mean the phase in which the system or set of policies which were earlier decided are deployed and put into practice. By making the decision model available to operational decision makers, they would have access to the arguments pro and con for a variety of actions and can thus make a more informed decision which balances the often conflicting criteria by which the value of action is measured. We also developed the concept of a 'monitored decision' in which metrics of performance were identified during the decision making process and used to evaluate the quality of that decision. It is important to monitor those decision which seem at highest risk of not meeting their stated objectives. Operational decisions are also potentially high risk decisions. Finally, we investigated the use of performance metrics for monitored decisions and audit logs of operational decisions in order to feed an evolutionary phase of the the life cycle. During evolution, decisions are revisisted, assumptions verified or refuted, and possible reassessments resulting in new policy are made. In this regard we implemented a machine learning algorithm which automatically defined business rules based on expert assessment of the quality of operational decisions as recorded during deployment.
errors are not found in unit testing . The rework effort to correct requirement and design problems in later phases can be as high as 300 to 1,000...Literature 31 Appendix B: Quality Cannot Be Tested In 35 Bibliography 38 CMU/SEI-2014-TN-026 | ii CMU/SEI-2014-TN-026 | iii List of Figures...Removal Densities During Development 10 Figure 8: Quality and Security-Focused Workflow 14 Figure 9: Testing Reliability Results for the Largest Project
Versioning Connectors (MVC) Representative of runtime monitoring gauges are multiversioning gauges, which monitor and analyze different versions of...multiple versions of the same component must be merged by the connector before they are forwarded to their target components. The multiversioning
no previous programming or computer science experience expected) • Precalculus -ready (that is, proficiency sufficient to enter college-level... precalculus course) • English Composition I-ready (that is, proficiency sufficient to enter college-level English I course) Co-Requisite Discrete
Introduction 1 1 Process Management (Category 1) 6 1.1 Process Definition (Area 1.1) 6 1.2 Infrastructure Standards (Area 1.2) 6 1.3 Resources (Area 1.3) 7...1.4 Training (Area 1.4) 8 2 Project Management (Category 2) 9 2.1 Project Plans (Area 2.1) 9 2.2 Project Infrastructure (Area 2.2) 10 2.3 Project...Monitoring (Area 2.3) 10 2.4 Project Risk Management (Area 2.4) 11 2.5 Supplier Management (Area 2.5) 11 3 Engineering (Category 3) 13 3.1 Product
is one of utility; each factor identified could be applied to :t production environment. The interaction of support groups within an operational...ship-support functio2s. It is comprised of aoproxi- mately 250 people and is i furctionally oriented department. The Financial Sistems Desian and...that t h-, following specific conditions ?xist: 1. Poorly Defined Regu-rements/Speci-fica:-4ons a) FM1SO design procedures/practices tend to be appli
virtual machine is much lighter than the workstation. The virtual machine doesn’t need to run anti- virus , firewalls, intrusion preven- tion systems...34] Maiorca, D., Corona , I., and Giacinto, G. Looking at the bag is not enough to find the bomb: An evasion of structural methods for malicious PDF...CCS ’13, ACM, pp. 119–130.  Maiorca, D., Giacinto, G., and Corona , I. A pattern recognition system for malicious PDF files detection. In
A JPL-created set of standard processes is to be used throughout the lifecycle of software development. These SDSPs cover a range of activities, from management and engineering activities, to assurance and support activities. These processes must be applied to software tasks per a prescribed set of procedures. JPL s Software Quality Improvement Project is currently working at the behest of the JPL Software Process Owner to ensure that all applicable software tasks follow these procedures. The SDSPs are captured as a set of 22 standards in JPL s software process domain. They were developed in-house at JPL by a number of Subject Matter Experts (SMEs) residing primarily within the Engineering and Science Directorate, but also from the Business Operations Directorate and Safety and Mission Success Directorate. These practices include not only currently performed best practices, but also JPL-desired future practices in key thrust areas like software architecting and software reuse analysis. Additionally, these SDSPs conform to many standards and requirements to which JPL projects are beholden.
Presents reviews of six computer software programs for teaching science. Provides the publisher, grade level, cost, and descriptions of software, including: (1) "Recycling Logic"; (2) "Introduction to Biochemistry"; (3) "Food for Thought"; (4) "Watts in a Home"; (5) "Geology in Action"; and (6)…
Merwarth, P., D.
The Common Software Module Repository (CSMR) is computerized library system with high product and service visibility to potential users. Online capabilities of system allow both librarian and user to interact with library. Librarian is responsible for maintaining information in CSMR library. User searches library to locate software modules that meet his or her current needs.
Giacomozzi, Claudia; Martelli, Francesco
The increasing spread of software components in the healthcare context renders explanatory guides relevant and mandatory to interpret laws and standards, and to support safe management of software products in healthcare. In 2012 a working group has been settled for the above purposes at Italian Electrotechnical Committee (CEI), made of experts from Italian National Institute of Health (ISS), representatives of industry, and representatives of the healthcare organizations. As a first outcome of the group activity, Guide CEI 62-237 was published in February 2015. The Guide incorporates an innovative approach based on the proper contextualization of software products, either medical devices or not, to the specific healthcare scenario, and addresses the risk management of IT systems. The Guide provides operators and manufacturers with an interpretative support with many detailed examples to facilitate the proper contextualization and management of health software, in compliance with related European and international regulations and standards.
Public Release; Distribution Unlimited. PA# 88ABW-2017-5388 Date Cleared: 30 OCT 2017 13. SUPPLEMENTARY NOTES 14. ABSTRACT Cyber- physical systems... physical processes that interact in intricate manners. This makes verification of the software complex and unwieldy. In this report, an approach towards...resulting implementations. 15. SUBJECT TERMS Cyber- physical systems, Formal guarantees, Code generation 16. SECURITY CLASSIFICATION OF: 17
Clark, G B
Quality has become the hallmark of industrial excellence. Many diverse factors have heightened national concern about managing quality control throughout the health-care industry, including laboratory services. Industry-wide focus on quality control has created a need for an administrative program to evaluate its effectiveness. That program is medical quality assurance. Because of national and industry-wide concern, development of quality assurance theory has gained increasing importance in medical accreditation and management circles. Scrutiny of the application of quality assurance has become particularly prominent during accreditation inspections. Implementing quality assurance programs now demands more of already finite resources. The professional laboratory manager should understand how quality assurance has developed in the United States during the past 150 years. The well-informed manager should recognize why the health-care industry only recently began to develop its own expertise in quality assurance. It is also worthwhile to understand how heavily health care has relied on the lessons learned in the non-health-care sector. This three-part series will present information that will help in applying quality assurance more effectively as a management tool in the medical laboratory. This first part outlines the early industrial, socioeconomic, and medicolegal background of quality assurance. Terminology is defined with some distinction made between the terms management and administration. The second part will address current accreditation requirements. Special emphasis will be placed on the practical application of accreditation guidelines, providing a template for quality assurance methods in the medical laboratory. The third part will provide an overview of quality assurance as a total management tool with some suggestions for developing and implementing a quality assurance program.
Stone, D. A.; Craig, J. W.; Drone, B.; Gerlach, R. H.; Williams, R. J.
The developmental status is discussed regarding the 'lifeboat' vehicle to enhance the safety of the crew on the Space Station Freedom (SSF). NASA's Assured Crew Return Vehicle (ACRV) is intended to provide a means for returning the SSF crew to earth at all times. The 'lifeboat' philosophy is the key to managing the development of the ACRV which further depends on matrixed support and total quality management for implementation. The risk of SSF mission scenarios are related to selected ACRV mission requirements, and the system and vehicle designs are related to these precepts. Four possible ACRV configurations are mentioned including the lifting-body, Apollo shape, Discoverer shape, and a new lift-to-drag concept. The SCRAM design concept is discussed in detail with attention to the 'lifeboat' philosophy and requirements for implementation.
This report provides a summary of the construction quality assurance (CQA) observation and test results, including: The results of the geosynthetic and soil materials conformance testing. The observation and testing results associates with the installation of the soil liners. The observation and testing results associated with the installation of the HDPE geomembrane liner systems. The observation and testing results associated with the installation of the leachate collection and removal systems. The observation and testing results associated with the installation of the working surfaces. The observation and testing results associated with in-plant manufacturing process. Summary of submittal reviews by Golder Constructionmore » Services, Inc. The submittal and certification of the piping material specifications. The observation and verification associated of the Acceptance Test Procedure results of the operational equipment functions. Summary of the ECNs which are incorporated into the project.« less
Dalthorp, Daniel; Huso, Manuela M. P.; Dail, David; Kenyon, Jessica
Evidence of Absence software (EoA) is a user-friendly application used for estimating bird and bat fatalities at wind farms and designing search protocols. The software is particularly useful in addressing whether the number of fatalities has exceeded a given threshold and what search parameters are needed to give assurance that thresholds were not exceeded. The software is applicable even when zero carcasses have been found in searches. Depending on the effectiveness of the searches, such an absence of evidence of mortality may or may not be strong evidence that few fatalities occurred. Under a search protocol in which carcasses are detected with nearly 100 percent certainty, finding zero carcasses would be convincing evidence that overall mortality rate was near zero. By contrast, with a less effective search protocol with low probability of detecting a carcass, finding zero carcasses does not rule out the possibility that large numbers of animals were killed but not detected in the searches. EoA uses information about the search process and scavenging rates to estimate detection probabilities to determine a maximum credible number of fatalities, even when zero or few carcasses are observed.
Price-Whelan, Adrian M.
Now more than ever, scientific results are dependent on sophisticated software and analysis. Why should we trust code written by others? How do you ensure your own code produces sensible results? How do you make sure it continues to do so as you update, modify, and add functionality? Software testing is an integral part of code validation and writing tests should be a requirement for any software project. I will talk about Python-based tools that make managing and running tests much easier and explore some statistics for projects hosted on GitHub that contain tests.
Under an SBIR (Small Business Innovative Research) contract with Johnson Space Center, Knowledge Based Systems Inc. (KBSI) developed an intelligent software environment for modeling and analyzing mission planning activities, simulating behavior, and, using a unique constraint propagation mechanism, updating plans with each change in mission planning activities. KBSI developed this technology into a commercial product, PROJECTLINK, a two-way bridge between PROSIm, KBSI's process modeling and simulation software and leading project management software like Microsoft Project and Primavera's SureTrak Project Manager.
software development scene is often charac- c. SPQR Model-Jones terized by: * schedule and cost estimates that are gross-d. COPMO-Thebaut ly inaccurate, SEI...time c. SPQR Model-Jones (in seconds) is simply derived from E by dividing T. Capers Jones has developed a software cost by the Stroud number, S...estimation model called the Software Produc- T=E/S tivity, Quality, and Reliability ( SPQR ) model. The basic approach is similar to that of Boehm’s The value
PCB Assurance Summary: PCB assurance actives are informed by risk in context of the Project. Lessons are being applied across Projects for continuous improvements. Newer component technologies, smaller/high pitch devices: tighter and more demanding PCB designs: Identifying new research areas. New materials, designs, structures and test methods.
Amaral, Alberto; Rosa, Maria Joao
In this paper we present a brief description of the evolution of quality assurance in Europe, paying particular attention to its relationship to the rising loss of trust in higher education institutions. We finalise by analysing the role of the European Commission in the setting up of new quality assurance mechanisms that tend to promote…
Ermakova, Tatiana; Fabian, Benjamin; Zarnekow, Rüdiger
Cloud computing promises to essentially improve healthcare delivery performance. However, shifting sensitive medical records to third-party cloud providers could create an adoption hurdle because of security and privacy concerns. This study examines the effect of confidentiality assurance in a cloud-computing environment on individuals' willingness to accept the infrastructure for inter-organizational sharing of medical data. We empirically investigate our research question by a survey with over 260 full responses. For the setting with a high confidentiality assurance, we base on a recent multi-cloud architecture which provides very high confidentiality assurance through a secret-sharing mechanism: Health information is cryptographically encoded and distributed in a way that no single and no small group of cloud providers is able to decode it. Our results indicate the importance of confidentiality assurance in individuals' acceptance of health clouds for sensitive medical data. Specifically, this finding holds for a variety of practically relevant circumstances, i.e., in the absence and despite the presence of conventional offline alternatives and along with pseudonymization. On the other hand, we do not find support for the effect of confidentiality assurance in individuals' acceptance of health clouds for non-sensitive medical data. These results could support the process of privacy engineering for health-cloud solutions.
Fabian, Benjamin; Zarnekow, Rüdiger
Summary Background Cloud computing promises to essentially improve healthcare delivery performance. However, shifting sensitive medical records to third-party cloud providers could create an adoption hurdle because of security and privacy concerns. Objectives This study examines the effect of confidentiality assurance in a cloud-computing environment on individuals’ willingness to accept the infrastructure for inter-organizational sharing of medical data. Methods We empirically investigate our research question by a survey with over 260 full responses. For the setting with a high confidentiality assurance, we base on a recent multi-cloud architecture which provides very high confidentiality assurance through a secret-sharing mechanism: Health information is cryptographically encoded and distributed in a way that no single and no small group of cloud providers is able to decode it. Results Our results indicate the importance of confidentiality assurance in individuals’ acceptance of health clouds for sensitive medical data. Specifically, this finding holds for a variety of practically relevant circumstances, i.e., in the absence and despite the presence of conventional offline alternatives and along with pseudonymization. On the other hand, we do not find support for the effect of confidentiality assurance in individuals’ acceptance of health clouds for non-sensitive medical data. These results could support the process of privacy engineering for health-cloud solutions. PMID:27781238
Intelligent Documentation Management System, IDMS, computer program developed to assist project managers in implementing information system documentation standard known as NASA-STD-2100-91, NASA STD, COS-10300, of NASA's Software Management and Assurance Program. Standard consists of data-item descriptions or templates, each of which governs particular component of software documentation. IDMS helps program manager in tailoring documentation standard to project. Written in C language.
systems 4. Ultrahigh software system assurance 5. Legacy maintenance and Brownfield development 6. Agile and Lean/ Kanban development. This paper...refined as the design of the maintenance modifications or Brownfield re-engineering is determined. VII. 6. AGILE AND LEAN/ KANBAN DEVELOPMENT The...difficulties of software maintenance estimation can often be mitigated by using lean workflow management techniques such as Kanban . In Kanban
Dutle, Aaron M.; Munoz, Cesar A.; Narkawicz, Anthony J.; Butler, Ricky W.
This paper explores a new approach to validating software implementations that have been produced from formally-verified algorithms. Although visual inspection gives some confidence that the implementations faithfully reflect the formal models, it does not provide complete assurance that the software is correct. The proposed approach, which is based on animation of formal specifications, compares the outputs computed by the software implementations on a given suite of input values to the outputs computed by the formal models on the same inputs, and determines if they are equal up to a given tolerance. The approach is illustrated on a prototype air traffic management system that computes simple kinematic trajectories for aircraft. Proofs for the mathematical models of the system's algorithms are carried out in the Prototype Verification System (PVS). The animation tool PVSio is used to evaluate the formal models on a set of randomly generated test cases. Output values computed by PVSio are compared against output values computed by the actual software. This comparison improves the assurance that the translation from formal models to code is faithful and that, for example, floating point errors do not greatly affect correctness and safety properties.
Mathematics and Computer Education, 1988
Presents reviews of six software packages. Includes (1) "Plain Vanilla Statistics"; (2) "MathCAD 2.0"; (3) "GrFx"; (4) "Trigonometry"; (5) "Algebra II"; (6) "Algebra Drill and Practice I, II, and III." (PK)
Wulfson, Stephen, Ed.
Provides a review of four science software programs. Includes topics such as plate tectonics, laboratory experiment simulations, the human body, and light and temperature. Contains information on ordering and reviewers' comments. (ML)
Wulfson, Eugene T., Ed.
Presents reviews by classroom teachers of software for teaching science. Includes material on the work of geologists, genetics, earth science, classification of living things, astronomy, endangered species, skeleton, drugs, and heartbeat. Provides information on availability and equipment needed. (RT)
Davis, Shelly J., Ed.; Knaupp, Jon, Ed.
Reviewed is computer software on: (1) classification of living things, a tutorial program for grades 5-10; and (2) polynomial practice using tiles, a drill-and-practice program for algebra students. (MNS)
Wulfson, Stephen, Ed.
Reviews seven computer software programs that can be used in science education programs. Describes courseware which deals with muscles and bones, terminology, classifying animals without backbones, molecular structures, drugs, genetics, and shaping the earth's surface. (TW)
Kimball, Jeffrey P.; And Others
Describes a variety of computer software. The packages reviewed include a variety of simulations, a spread sheet, a printer driver and an alternative operating system for DBM.PCs and compatible programs. (BSR)
Classroom Computer Learning, 1988
Provides reviews of three software packages including "MusicShapes,""For Comment," and "Colortrope," which were developed for music, writing, and science, respectively. Includes information on grade levels, publishers, hardware needed, and cost. (TW)
Wulfson, Stephen, Ed.
Reviewed are six computer software packages including "Lunar Greenhouse,""Dyno-Quest,""How Weather Works,""Animal Trackers,""Personal Science Laboratory," and "The Skeletal and Muscular Systems." Availability, functional, and hardware requirements are discussed. (CW)
Wulfson, Stephen, Ed.
Provides reviews of six computer software programs designed for use in elementary science education programs. Provides the title, publisher, grade level, and descriptions of courseware on ant farms, drugs, genetics, beachcombing, matter, and test generation. (TW)
Dwyer, Donna; And Others
Reviewed are seven software packages for Apple and IBM computers. Included are: "Toxicology"; "Science Corner: Space Probe"; "Alcohol and Pregnancy"; "Science Tool Kit Plus"; Computer Investigations: Plant Growth"; "Climatrolls"; and "Animal Watch: Whales." (CW)
Miller, Anne, Ed.; Radziemski, Cathy, Ed.
Three pieces of computer software are described and reviewed: HyperCard, to build and use varied applications; Iggy's Gnees, for problem solving with shapes in grades kindergarten-two; and Algebra Shop, for practicing skills and problem solving. (MNS)
Classroom Computer Learning, 1990
Reviewed are three computer software packages including "Martin Luther King, Jr.: Instant Replay of History,""Weeds to Trees," and "The New Print Shop, School Edition." Discussed are hardware requirements, costs, grade levels, availability, emphasis, strengths, and weaknesses. (CW)
Wulfson, Stephen, Ed.
Reviews seven instructional software packages covering a variety of topics. Includes: "Science Square-Off"; "The Desert"; "Science Courseware: Physical Science"; "Odell Lake"; "Safety First"; "An Experience in Artificial Intelligence"; and "Master Mapper." (TW)
Fault Management (FM) systems are ranked high in risk-based assessment of criticality within flight software, emphasizing the importance of establishing highly competent domain expertise to provide assurance for NASA projects, especially as spaceflight systems continue to increase in complexity. Insight into specific characteristics of FM architectures seen embedded within safety- and mission-critical software systems analyzed by the NASA Independent Verification Validation (IVV) Program has been enhanced with an FM Technical Reference (TR) suite. Benefits are aimed beyond the IVV community to those that seek ways to efficiently and effectively provide software assurance to reduce the FM risk posture of NASA and other space missions. The identification of particular FM architectures, visibility, and associated IVV techniques provides a TR suite that enables greater assurance that critical software systems will adequately protect against faults and respond to adverse conditions. The role FM has with regard to overall asset protection of flight software systems is being addressed with the development of an adverse condition (AC) database encompassing flight software vulnerabilities.Identification of potential off-nominal conditions and analysis to determine how a system responds to these conditions are important aspects of hazard analysis and fault management. Understanding what ACs the mission may face, and ensuring they are prevented or addressed is the responsibility of the assurance team, which necessarily should have insight into ACs beyond those defined by the project itself. Research efforts sponsored by NASAs Office of Safety and Mission Assurance defined terminology, categorized data fields, and designed a baseline repository that centralizes and compiles a comprehensive listing of ACs and correlated data relevant across many NASA missions. This prototype tool helps projects improve analysis by tracking ACs, and allowing queries based on project, mission
Software Bisque's TheSky, SkyPro and Remote Astronomy Software incorporate technology developed for the Hubble Space Telescope. TheSky and SkyPro work together to orchestrate locating, identifying and acquiring images of deep sky objects. With all three systems, the user can directly control computer-driven telescopes and charge coupled device (CCD) cameras through serial ports. Through the systems, astronomers and students can remotely operate a telescope at the Mount Wilson Observatory Institute.
Hinchey, Mike; Pressburger, Tom; Markosian, Lawrence; Feather, Martin
A viewgraph presentation describing the NASA Software Assurance Research Program (SARP) research infusion projects is shown. The topics include: 1) Background/Motivation; 2) Proposal Solicitation Process; 3) Proposal Evaluation Process; 4) Overview of Some Projects to Date; and 5) Lessons Learned.
The health insurance business in India has seen a growth of over 25% per annum in the last few years with the expansion of the private health insurance sector. The premium incomes of health insurance have crossed the Rs 8,000 crore mark with the share of private companies increasing to over 41%. This is despite the fact that from the perspective of patients, health insurance is not a good deal, especially when they need it most. This raises a number of ethical issues regarding how the health insurance business runs and how medical practice adjusts to it for profiteering. This article uses the personal experience of the author to argue that health insurance in an unregulated environment can only lead to unethical practices, further victimising the patient. Further, publicly financed healthcare which operates in an environment regulating both public and private healthcare provisioning is the only way to assure access to ethical and equitable healthcare to people.
Fridge, Ernest M., III
Today's software systems generally use obsolete technology, are not integrated properly with other software systems, and are difficult and costly to maintain. The discipline of reverse engineering is becoming prominent as organizations try to move their systems up to more modern and maintainable technology in a cost effective manner. JSC created a significant set of tools to develop and maintain FORTRAN and C code during development of the Space Shuttle. This tool set forms the basis for an integrated environment to re-engineer existing code into modern software engineering structures which are then easier and less costly to maintain and which allow a fairly straightforward translation into other target languages. The environment will support these structures and practices even in areas where the language definition and compilers do not enforce good software engineering. The knowledge and data captured using the reverse engineering tools is passed to standard forward engineering tools to redesign or perform major upgrades to software systems in a much more cost effective manner than using older technologies. A beta vision of the environment was released in Mar. 1991. The commercial potential for such re-engineering tools is very great. CASE TRENDS magazine reported it to be the primary concern of over four hundred of the top MIS executives.
Fridge, Ernest M., III; Hiott, Jim; Golej, Jim; Plumb, Allan
Today's software systems generally use obsolete technology, are not integrated properly with other software systems, and are difficult and costly to maintain. The discipline of reverse engineering is becoming prominent as organizations try to move their systems up to more modern and maintainable technology in a cost effective manner. The Johnson Space Center (JSC) created a significant set of tools to develop and maintain FORTRAN and C code during development of the space shuttle. This tool set forms the basis for an integrated environment to reengineer existing code into modern software engineering structures which are then easier and less costly to maintain and which allow a fairly straightforward translation into other target languages. The environment will support these structures and practices even in areas where the language definition and compilers do not enforce good software engineering. The knowledge and data captured using the reverse engineering tools is passed to standard forward engineering tools to redesign or perform major upgrades to software systems in a much more cost effective manner than using older technologies. The latest release of the environment was in Feb. 1992.
Fitz, Rhonda; Whitman, Gerek
Research into complexities of software systems Fault Management (FM) and how architectural design decisions affect safety, preservation of assets, and maintenance of desired system functionality has coalesced into a technical reference (TR) suite that advances the provision of safety and mission assurance. The NASA Independent Verification and Validation (IV&V) Program, with Software Assurance Research Program support, extracted FM architectures across the IV&V portfolio to evaluate robustness, assess visibility for validation and test, and define software assurance methods applied to the architectures and designs. This investigation spanned IV&V projects with seven different primary developers, a wide range of sizes and complexities, and encompassed Deep Space Robotic, Human Spaceflight, and Earth Orbiter mission FM architectures. The initiative continues with an expansion of the TR suite to include Launch Vehicles, adding the benefit of investigating differences intrinsic to model-based FM architectures and insight into complexities of FM within an Agile software development environment, in order to improve awareness of how nontraditional processes affect FM architectural design and system health management. The identification of particular FM architectures, visibility, and associated IV&V techniques provides a TR suite that enables greater assurance that critical software systems will adequately protect against faults and respond to adverse conditions. Additionally, the role FM has with regard to strengthened security requirements, with potential to advance overall asset protection of flight software systems, is being addressed with the development of an adverse conditions database encompassing flight software vulnerabilities. Capitalizing on the established framework, this TR suite provides assurance capability for a variety of FM architectures and varied development approaches. Research results are being disseminated across NASA, other agencies, and the
Clark, David A.
In light of the escalation of terrorism, the Department of Defense spearheaded the development of new antiterrorist software for all Government agencies by issuing a Broad Agency Announcement to solicit proposals. This Government-wide competition resulted in a team that includes NASA Lewis Research Center's Computer Services Division, who will develop the graphical user interface (GUI) and test it in their usability lab. The team launched a program entitled Joint Sphere of Security (JSOS), crafted a design architecture (see the following figure), and is testing the interface. This software system has a state-ofthe- art, object-oriented architecture, with a main kernel composed of the Dynamic Information Architecture System (DIAS) developed by Argonne National Laboratory. DIAS will be used as the software "breadboard" for assembling the components of explosions, such as blast and collapse simulations.
Real-Time Innovations, Inc. (RTI) collaborated with Ames Research Center, the Jet Propulsion Laboratory and Stanford University to leverage NASA research to produce ControlShell software. RTI is the first "graduate" of Ames Research Center's Technology Commercialization Center. The ControlShell system was used extensively on a cooperative project to enhance the capabilities of a Russian-built Marsokhod rover being evaluated for eventual flight to Mars. RTI's ControlShell is complex, real-time command and control software, capable of processing information and controlling mechanical devices. One ControlShell tool is StethoScope. As a real-time data collection and display tool, StethoScope allows a user to see how a program is running without changing its execution. RTI has successfully applied its software savvy in other arenas, such as telecommunications, networking, video editing, semiconductor manufacturing, automobile systems, and medical imaging.
For the radiation shielding quality assurance, the validity and reliability of the neutron transport code MCNP, which is now one of the most widely used radiation shielding analysis codes, were checked with lot of benchmark experiments. And also as a practical example, follows were performed in this thesis. One integral neutron transport experiment to measure the effect of neutron streaming in iron and void was performed with Dog-Legged Void Assembly in Knolls Atomic Power Laboratory in 1991. Neutron flux was measured six different places with the methane detectors and a BF-3 detector. The main purpose of the measurements was to provide benchmark against which various neutron transport calculation tools could be compared. Those data were used in verification of Monte Carlo Neutron & Photon Transport Code, MCNP, with the modeling for that. Experimental results and calculation results were compared in both ways, as the total integrated value of neutron fluxes along neutron energy range from 10 KeV to 2 MeV and as the neutron spectrum along with neutron energy range. Both results are well matched with the statistical error +/-20%. MCNP results were also compared with those of TORT, a three dimensional discrete ordinates code which was developed by Oak Ridge National Laboratory. MCNP results are superior to the TORT results at all detector places except one. This means that MCNP is proved as a very powerful tool for the analysis of neutron transport through iron & air and further it could be used as a powerful tool for the radiation shielding analysis. For one application of the analysis of variance (ANOVA) to neutron and gamma transport problems, uncertainties for the calculated values of critical K were evaluated as in the ANOVA on statistical data.
Riggs, D G
Although initially applied within the manufacturing sectors of industry the benefits of quality assurance to both specifiers and customers alike are rapidly being realised by all sectors of business and industry.
Over the past two decades the role of SHA has shifted from quality control (QC) of materials and : placement techniques to quality assurance (QA) and acceptance. The role of the Office of Materials : Technology (OMT) has been shifting towards assuran...
The papr describes how, rather than building "bridges" across centuries, quality assurance (QA) personnel have the opportunity to build bridges across technical disciplines, between public and private organizations, and between different QA groups. As reviewers and auditors of a...
Yilmaz, Levent; Wang, Shuo
Proper design analysis is indispensable to assure quality and reduce emergent costs due to faulty software. Teaching proper design verification skills early during pedagogical development is crucial, as such analysis is the only tractable way of resolving software problems early when they are easy to fix. The premise of the presented strategy is…
Science Teacher, 1988
Reviews two software programs for Apple series computers. Includes "Orbital Mech," a basic planetary orbital simulation for the Macintosh, and "START: Stimulus and Response Tools for Experiments in Memory, Learning, Cognition, and Perception," a program that demonstrates basic psychological principles and experiments. (CW)
Science and Children, 1988
Reviews six software packages for use with school age children ranging from grade 3 to grade 12. Includes "The Microcomputer Based Lab Project: Motion, Sound"; "Genetics"; "Geologic History"; "The Microscope Simulator"; and "Wiz Works" all for Apple II and "Reading for Information: Level…
McGrath, Diane, Ed.
Reviewed is a computer software package entitled "Audubon Wildlife Adventures: Grizzly Bears" for Apple II and IBM microcomputers. Included are availability, hardware requirements, cost, and a description of the program. The murder-mystery flavor of the program is stressed in this program that focuses on illegal hunting and game…
Science and Children, 1990
Reviewed are seven computer software packages for IBM and/or Apple Computers. Included are "Windows on Science: Volume 1--Physical Science"; "Science Probe--Physical Science"; "Wildlife Adventures--Grizzly Bears"; "Science Skills--Development Programs"; "The Clean Machine"; "Rock Doctor";…
Science Software Quarterly, 1984
Provides extensive reviews of computer software, examining documentation, ease of use, performance, error handling, special features, and system requirements. Includes statistics, problem-solving (TK Solver), label printing, database management, experimental psychology, Encyclopedia Britannica biology, and DNA-sequencing programs. A program for…
Fridge, Ernest M., III
Programs in use today generally have all of the function and information processing capabilities required to do their specified job. However, older programs usually use obsolete technology, are not integrated properly with other programs, and are difficult to maintain. Reengineering is becoming a prominent discipline as organizations try to move their systems to more modern and maintainable technologies. The Johnson Space Center (JSC) Software Technology Branch (STB) is researching and developing a system to support reengineering older FORTRAN programs into more maintainable forms that can also be more readily translated to a modern languages such as FORTRAN 8x, Ada, or C. This activity has led to the development of maintenance strategies for design recovery and reengineering. These strategies include a set of standards, methodologies, and the concepts for a software environment to support design recovery and reengineering. A brief description of the problem being addressed and the approach that is being taken by the STB toward providing an economic solution to the problem is provided. A statement of the maintenance problems, the benefits and drawbacks of three alternative solutions, and a brief history of the STB experience in software reengineering are followed by the STB new FORTRAN standards, methodology, and the concepts for a software environment.
Science and Children, 1988
Reviews five software packages for use with school age children. Includes "Science Toolkit Module 2: Earthquake Lab"; "Adaptations and Identification"; "Geoworld"; "Body Systems II Series: The Blood System: A Liquid of Life," all for Apple II, and "Science Courseware: Life Science/Biology" for…
Science Teacher, 1988
Reviews two computer software packages for use in physical science, physics, and chemistry classes. Includes "Physics of Model Rocketry" for Apple II, and "Black Box" for Apple II and IBM compatible computers. "Black Box" is designed to help students understand the concept of indirect evidence. (CW)
Kinnaman, Daniel E.; And Others
Reviews four educational software packages for Apple, IBM, and Tandy computers. Includes "How the West was One + Three x Four,""Mavis Beacon Teaches Typing,""Math and Me," and "Write On." Reviews list hardware requirements, emphasis, levels, publisher, purchase agreements, and price. Discusses the strengths…
Wulfson, Stephen, Ed.
Reviewed are six software packages for Apple and/or IBM computers. Included are "Autograph,""The New Game Show,""Science Probe-Earth Science,""Pollution Patrol,""Investigating Plant Growth," and "AIDS: The Investigation." Discussed are the grade level, function, availability, cost, and hardware requirements of each. (CW)
Bitter, Gary G., Ed.
Reviews three software packages: (1) "Physics," tutorial, grades 11-12, Macintosh; (2) "Hands On Math: Volume I," interactive math exploration/simulation of manipulatives use, grades K-7, Apple II; and (3) "A.I.: An Experience with Artificial Intelligence," simulation, grades 5-12, Apple II. (MVL)
Science and Children, 1989
Reviews of seven software packages are presented including "The Environment I: Habitats and EcoSystems; II Cycles and Interactions"; "Super Sign Maker"; "The Great Knowledge Race: Substance Abuse"; "Exploring Science: Temperature"; "Fast Food Calculator and RD Aide"; "The Human Body:…
Burke, Edmund B.
Outlines basic patent law information that pertains to computer software programs. Topics addressed include protection in other countries; how to obtain patents; kinds of patents; duration; classes of patentable subject matter, including machines and processes; patentability searches; experimental use prior to obtaining a patent; and patent…
Bitter, Gary G., Ed.
Describes three software packages: (1) "MacMendeleev"--database/graphic display for chemistry, grades 10-12, Macintosh; (2) "Geometry One: Foundations"--geometry tutorial, grades 7-12, IBM; (3) "Mathematics Exploration Toolkit"--algebra and calculus tutorial, grades 8-12, IBM. (MVL)
Mathematics and Computer Education, 1987
Presented are reviews of several microcomputer software programs. Included are reviews of: (1) Microstat (Zenith); (2) MathCAD (MathSoft); (3) Discrete Mathematics (True Basic); (4) CALCULUS (True Basic); (5) Linear-Kit (John Wiley); and (6) Geometry Sensei (Broderbund). (RH)
Teles, Elizabeth, Ed.; And Others
Reviewed are two computer software packages for Macintosh microcomputers including "Phase Portraits," an exploratory graphics tool for studying first-order planar systems; and "MacMath," a set of programs for exploring differential equations, linear algebra, and other mathematical topics. Features, ease of use, cost, availability, and hardware…
McGrath, Diane, Ed.
Reviewed are two computer software programs for Apple II computers on weather for upper elementary and middle school grades. "Weather" introduces the major factors (temperature, humidity, wind, and air pressure) affecting weather. "How Weather Works" uses simulation and auto-tutorial formats on sun, wind, fronts, clouds, and…
A chart of 40 alumni-development database systems provides information on vendor/Web site, address, contact/phone, software name, price range, minimum suggested workstation/suggested server, standard reports/reporting tools, minimum/maximum record capacity, and number of installed sites/client type. (DB)
Mackenzie, Norma N.; And Others
Reviews four computer software packages including: "The Physical Science Series: Sound" which demonstrates making waves, speed of sound, doppler effect, and human hearing; "Andromeda" depicting celestial motions in any direction; "Biology Quiz: Humans" covering chemistry, cells, viruses, and human biology; and…
This presentation provides a view of software quality for 1997 - what works and what doesn`t. For many years, software quality assurance lagged behind hardware quality assurance in terms of methods, metrics, and successful results. New approaches such as Quality Function Development (WFD) the ISO 9000-9004 standards, the SEI maturity levels, and Total Quality Management (TQM) are starting to attract wide attention, and in some cases to bring software quality levels up to a parity with manufacturing quality levels.
Lalli, Vincent R.; Packard, Michael H.; Ziemianski, Tom
The application of assurance engineering techniques improves the duration of failure-free performance of software. The totality of features and characteristics of a software product are what determine its ability to satisfy customer needs. Software in safety-critical systems is very important to NASA. We follow the System Safety Working Groups definition for system safety software as: 'The optimization of system safety in the design, development, use and maintenance of software and its integration with safety-critical systems in an operational environment. 'If it is not safe, say so' has become our motto. This paper goes over methods that have been used by NASA to make software design improvements by focusing on software quality and the design and inspection process.
American School Health Association (NJ3), 2005
This resource sets the standard for school nurses concerning the formulation of individualized healthcare plans designed to fit the unique health needs of students. Eighteen chapters focus on special issues and school nursing concepts. Computer software, which accompanies the manual, assists in the development and creation of individualized…
Eckhause, Tobias; Al-Hallaq, Hania; Ritter, Timothy; DeMarco, John; Farrey, Karl; Pawlicki, Todd; Kim, Gwe-Ya; Popple, Richard; Sharma, Vijeshwar; Perez, Mario; Park, SungYong; Booth, Jeremy T; Thorwarth, Ryan; Moran, Jean M
The purpose of this study was 2-fold. One purpose was to develop an automated, streamlined quality assurance (QA) program for use by multiple centers. The second purpose was to evaluate machine performance over time for multiple centers using linear accelerator (Linac) log files and electronic portal images. The authors sought to evaluate variations in Linac performance to establish as a reference for other centers. The authors developed analytical software tools for a QA program using both log files and electronic portal imaging device (EPID) measurements. The first tool is a general analysis tool which can read and visually represent data in the log file. This tool, which can be used to automatically analyze patient treatment or QA log files, examines the files for Linac deviations which exceed thresholds. The second set of tools consists of a test suite of QA fields, a standard phantom, and software to collect information from the log files on deviations from the expected values. The test suite was designed to focus on the mechanical tests of the Linac to include jaw, MLC, and collimator positions during static, IMRT, and volumetric modulated arc therapy delivery. A consortium of eight institutions delivered the test suite at monthly or weekly intervals on each Linac using a standard phantom. The behavior of various components was analyzed for eight TrueBeam Linacs. For the EPID and trajectory log file analysis, all observed deviations which exceeded established thresholds for Linac behavior resulted in a beam hold off. In the absence of an interlock-triggering event, the maximum observed log file deviations between the expected and actual component positions (such as MLC leaves) varied from less than 1% to 26% of published tolerance thresholds. The maximum and standard deviations of the variations due to gantry sag, collimator angle, jaw position, and MLC positions are presented. Gantry sag among Linacs was 0.336 ± 0.072 mm. The standard deviation in MLC
Quality assurance of higher education is made up of two parts: internal quality assurance (IQA) and external quality assurance (EQA). Both belong to a union of the coexistence and balance of yin and yang. But in reality there exists a paradox of "confusion of quality assurance (QA) subject consciousness, singularity of social QA and lack of QA…
... 10 Energy 2 2010-01-01 2010-01-01 false Quality assurance. 71.37 Section 71.37 Energy NUCLEAR... Package Approval § 71.37 Quality assurance. (a) The applicant shall describe the quality assurance program... quality assurance program that are applicable to the particular package design under consideration...
Bangerth, W.; Heister, T.; Hwang, L.; Kellogg, L. H.
Developing sustainable open source software packages for the sciences appears at first to be primarily a technical challenge: How can one create stable and robust algorithms, appropriate software designs, sufficient documentation, quality assurance strategies such as continuous integration and test suites, or backward compatibility approaches that yield high-quality software usable not only by the authors, but also the broader community of scientists? However, our experience from almost two decades of leading the development of the deal.II software library (http://www.dealii.org, a widely-used finite element package) and the ASPECT code (http://aspect.dealii.org, used to simulate convection in the Earth's mantle) has taught us that technical aspects are not the most difficult ones in scientific open source software. Rather, it is the social challenge of building and maintaining a community of users and developers interested in answering questions on user forums, contributing code, and jointly finding solutions to common technical and non-technical challenges. These problems are posed in an environment where project leaders typically have no resources to reward the majority of contributors, where very few people are specifically paid for the work they do on the project, and with frequent turnover of contributors as project members rotate into and out of jobs. In particular, much software work is done by graduate students who may become fluent enough in a software only a year or two before they leave academia. We will discuss strategies we have found do and do not work in maintaining and growing communities around the scientific software projects we lead. Specifically, we will discuss the management style necessary to keep contributors engaged, ways to give credit where credit is due, and structuring documentation to decrease reliance on forums and thereby allow user communities to grow without straining those who answer questions.
MathSoft Plus 5.0 is a calculation software package for electrical engineers and computer scientists who need advanced math functionality. It incorporates SmartMath, an expert system that determines a strategy for solving difficult mathematical problems. SmartMath was the result of the integration into Mathcad of CLIPS, a NASA-developed shell for creating expert systems. By using CLIPS, MathSoft, Inc. was able to save the time and money involved in writing the original program.
machine transitions. 2. We developed the TraceContract API for trace analysis in the Scala programming language. TraceContract combines a high-level...DATE (DD-MM-YYYY) 2. REPORT TYPE 3. DATES COVERED (From - To) 4. TITLE AND SUBTITLE 5a. CONTRACT NUMBER 5b. GRANT NUMBER 5c. PROGRAM ELEMENT NUMBER...awarded within the Software and Systems program . The original Pro- gram Manager was David Luginbuhl. Bob Bonneau took over as PM in March 2011. The award
Lakshminarayanan, Subitha; Kar, Sitanshu Sekhar; Gupta, Rajeev; Xavier, Denis; Bhaskar Reddy, S. Vijaya
Background: Diabetes registries monitor the population prevalence and incidence of diabetes, monitor diabetes control program, provide information of quality of care to health service providers, and provide a sampling frame for interventional studies. This study documents the process of establishing a prospective diabetes registry in a primary health-care setting in Puducherry. Methods: This is a facility-based prospective registry conducted in six randomly selected urban health centers in Puducherry, with enrollment of all known patients with diabetes attending chronic disease clinics. Administrative approvals were obtained from Government Health Services. Manuals for training of medical officers, health-care workers, and case report forms were developed. Diabetes registry was prepared using Epi Info software. Results: In the first phase, demographic characteristics, risk factors, complications, coexisting chronic conditions, lifestyle and medical management, and clinical outcomes were recorded. Around 2177 patients with diabetes have been registered in six Primary Health Centres out of a total of 2948 participants seeking care from chronic disease clinic. Registration coverage ranges from 61% to 105% in these centers. Conclusion: This study has documented methodological details, and learning experiences gained while developing a diabetes registry at the primary health care level and the scope for upscaling to a Management Information System for Diabetes and a State-wide Registry. Improvement in patient care through needs assessment and quality assurance in service delivery is an important theme envisioned by this registry. PMID:28553589
Weerasinghe, Dasun; Rajarajan, Muttukrishnan; Rakocevic, Veselin
The rapid growth in mobile technology makes the delivery of healthcare data and services on mobile phones a reality. However, the healthcare data is very sensitive and has to be protected against unauthorized access. While most of the development work on security of mobile healthcare today focuses on the data encryption and secure authentication in remote servers, protection of data on the mobile device itself has gained very little attention. This paper analyses the requirements and the architecture for a secure mobile capsule, specially designed to protect the data that is already on the device. The capsule is a downloadable software agent with additional functionalities to enable secure external communication with healthcare service providers, network operators and other relevant communication parties.
Shull, Forrest; Feldmann, Raimund L.; Seaman, Carolyn; Regardie, Myrna; Godfrey, Sally
Software inspections provide a proven approach to quality assurance for software products of all kinds, including requirements, design, code, test plans, among others. Common to all inspections is the aim of finding and fixing defects as early as possible, and thereby providing cost savings by minimizing the amount of rework necessary later in the lifecycle. Measurement data, such as the number and type of found defects and the effort spent by the inspection team, provide not only direct feedback about the software product to the project team but are also valuable for process improvement activities. In this paper, we discuss NASA's use of software inspections and the rich set of data that has resulted. In particular, we present results from analysis of inspection data that illustrate the benefits of fully utilizing that data for process improvement at several levels. Examining such data across multiple inspections or projects allows team members to monitor and trigger cross project improvements. Such improvements may focus on the software development processes of the whole organization as well as improvements to the applied inspection process itself.
Voigt, S. (Editor); Beskenis, S. (Editor)
Issues in the development of software for the Space Station are discussed. Software acquisition and management, software development environment, standards, information system support for software developers, and a future software advisory board are addressed.
Ibarra, A.; Gabriel, C.
Nowadays hardware and system infrastructure evolve on time scales much shorter than the typical duration of space astronomy missions. Data processing software capabilities have to evolve to preserve the scientific return during the entire experiment life time. Software preservation is a key issue that has to be tackled before the end of the project to keep the data usable over many years. We present RISA (Remote Interface to Science Analysis) as a solution to decouple data processing software and infrastructure life-cycles, using JAVA applications and web-services wrappers to existing software. This architecture employs embedded SAS in virtual machines assuring a homogeneous job execution environment. We will also present the first studies to reactivate the data processing software of the EXOSAT mission, the first ESA X-ray astronomy mission launched in 1983, using the generic RISA approach.
Mallasch, Paul G.
This volume contains the complete software system documentation for the Federal Communications Commission (FCC) Transponder Loading Data Conversion Software (FIX-FCC). This software was written to facilitate the formatting and conversion of FCC Transponder Occupancy (Loading) Data before it is loaded into the NASA Geosynchronous Satellite Orbital Statistics Database System (GSOSTATS). The information that FCC supplies NASA is in report form and must be converted into a form readable by the database management software used in the GSOSTATS application. Both the User's Guide and Software Maintenance Manual are contained in this document. This volume of documentation passed an independent quality assurance review and certification by the Product Assurance and Security Office of the Planning Research Corporation (PRC). The manuals were reviewed for format, content, and readability. The Software Management and Assurance Program (SMAP) life cycle and documentation standards were used in the development of this document. Accordingly, these standards were used in the review. Refer to the System/Software Test/Product Assurance Report for the Geosynchronous Satellite Orbital Statistics Database System (GSOSTATS) for additional information.
... at follow-up appointments by talking with your healthcare team about your concerns, asking questions and getting ... from the time you spend with all your healthcare providers, not just your doctor. Use the skills ...
Every day, individuals around the world retrieve, share, and exchange information on the Internet. We interact online to share personal information, find answers to questions, make financial transactions, play social games, and maintain professional and personal relationships. Sometimes our online interactions take place between two or more humans. In other cases, we rely on computers to manage information on our behalf. In each scenario, risk and uncertainty are essential for determining possible actions and outcomes. This essay highlights common deficiencies in our understanding of key concepts such as trust, trustworthiness, cooperation, and assurance in online environments. Empirical evidence from experimental work in computer-mediated environments underscores the promises and perils of overreliance on security and assurance structures as replacements for interpersonal trust. These conceptual distinctions are critical because the future shape of the Internet will depend on whether we build assurance structures to limit and control ambiguity or allow trust to emerge in the presence of risk and uncertainty.
Chandola, Varun; Schryver, Jack C; Sukumar, Sreenivas R
We discuss the problem of fraud detection in healthcare in this chapter. Given the recent scrutiny of the ineciencies in the US healthcare system, identifying fraud has been on the forefront of the eorts towards reducing the healthcare costs. In this chapter we will focus on understanding the issue of healthcare fraud in detail, and review methods that have been proposed in the literature to combat this issue using data driven approach.
Del Fiol, Guilherme; Hanseler, Haley; Crouch, Barbara Insley; Cummins, Mollie R.
Summary Background Health information exchange (HIE) between Poison Control Centers (PCCs) and Emergency Departments (EDs) could improve care of poisoned patients. However, PCC information systems are not designed to facilitate HIE with EDs; therefore, we are developing specialized software to support HIE within the normal workflow of the PCC using user-centered design and rapid prototyping. Objective To describe the design of an HIE dashboard and the refinement of user requirements through rapid prototyping. Methods Using previously elicited user requirements, we designed low-fidelity sketches of designs on paper with iterative refinement. Next, we designed an interactive high-fidelity prototype and conducted scenario-based usability tests with end users. Users were asked to think aloud while accomplishing tasks related to a case vignette. After testing, the users provided feedback and evaluated the prototype using the System Usability Scale (SUS). Results Survey results from three users provided useful feedback that was then incorporated into the design. After achieving a stable design, we used the prototype itself as the specification for development of the actual software. Benefits of prototyping included having 1) subject-matter experts heavily involved with the design; 2) flexibility to make rapid changes, 3) the ability to minimize software development efforts early in the design stage; 4) rapid finalization of requirements; 5) early visualization of designs; 6) and a powerful vehicle for communication of the design to the programmers. Challenges included 1) time and effort to develop the prototypes and case scenarios; 2) no simulation of system performance; 3) not having all proposed functionality available in the final product; and 4) missing needed data elements in the PCC information system. PMID:27081404
Various NASA Small Business Innovation Research grants from Marshall Space Flight Center, Langley Research Center and Ames Research Center were used to develop the 'kernel' of COMCO's modeling and simulation software, the PHLEX finite element code. NASA needed it to model designs of flight vehicles; one of many customized commercial applications is UNISIM, a PHLEX-based code for analyzing underground flows in oil reservoirs for Texaco, Inc. COMCO's products simulate a computational mechanics problem, estimate the solution's error and produce the optimal hp-adapted mesh for the accuracy the user chooses. The system is also used as a research or training tool in universities and in mechanical design in industrial corporations.
General Purpose Boundary Element Solution Technology (GPBEST) software employs the boundary element method of mechanical engineering analysis, as opposed to finite element. It is, according to one of its developers, 10 times faster in data preparation and more accurate than other methods. Its use results in less expensive products because the time between design and manufacturing is shortened. A commercial derivative of a NASA-developed computer code, it is marketed by Best Corporation to solve problems in stress analysis, heat transfer, fluid analysis and yielding and cracking of solids. Other applications include designing tractor and auto parts, household appliances and acoustic analysis.
Carnevale, Anthony P.; Smith, Nicole; Gulish, Artem; Beach, Bennett H.
This executive summary highlights several findings about healthcare. These are: (1) Healthcare is 18 percent of the U.S. economy, twice as high as in other countries; (2) There are two labor markets in healthcare: high-skill, high-wage professional and technical jobs and low-skill, low-wage support jobs; (3) Demand for postsecondary education in…
Strategic Studies Quarterly ♦ Spring 2009 [ 43 ] On Nuclear Deterrence and Assurance Keith B. Payne Weakness is provocative. —Donald Rumsfeld...Strategic Studies Quarterly ♦ Spring 2009 Keith B. Payne [ 44 ] No Deterrence Value for Nuclear Weapons...Payne.indd 44 2/2/09 2:31:00 PM On Nuclear Deterrence and Assurance Strategic Studies Quarterly ♦ Spring 2009 [ 45 ] of nuclear weapons were to be
Greenberg, D S; Welcker, B
With cost pressures brought to the forefront by the growth of managed care, the integration of healthcare information systems is more important than ever. Providers of healthcare information are under increasing pressure to provide timely information to end users in a cost effective manner. Organizations have had to decide between the strong functionality that a multi-vendor 'best of breed' architecture provides and the strong integration provided by a single-vendor solution. As connectivity between systems increased, these interfaces were migrated to work across serial and eventually, network, connections. In addition, the content of the information became standardized through efforts like HL7 and ANSI X12 and Edifact. Although content-based standards go a long way towards facilitating interoperability, there is also quite a bit of work required to connect two systems even when they both adhere to the standard. A key to accomplishing this goal is increasing the connectivity between disparate systems in the healthcare environment. Microsoft is working with healthcare organizations and independent software vendors to bring Microsoft's powerful enterprise object technology, ActiveX, to the healthcare industry. Whilst object orientation has been heralded as the 'next big thing' in computer applications development, Microsoft believe that, in fact, component software is the technology which will provide the greatest benefit to end users.
Miller, Douglas R.
Software reliability research is briefly described. General research topics are reliability growth models, quality of software reliability prediction, the complete monotonicity property of reliability growth, conceptual modelling of software failure behavior, assurance of ultrahigh reliability, and analysis techniques for fault-tolerant systems.
Uber, James G.
Software itself is not hazardous, but since software and hardware share common interfaces there is an opportunity for software to create hazards. Further, these software systems are complex, and proven methods for the design, analysis, and measurement of software safety are not yet available. Some past software failures, future NASA software trends, software engineering methods, and tools and techniques for various software safety analyses are reviewed. Recommendations to NASA are made based on this review.
Schulz, E B; Barrett, J W; Price, C
As controlled clinical vocabularies assume an increasing role in modern clinical information systems, so the issue of their quality demands greater attention. In order to meet the resulting stringent criteria for completeness and correctness, a quality assurance system comprising a database of more than 500 rules is being developed and applied to the Read Thesaurus. The authors discuss the requirement to apply quality assurance processes to their dynamic editing database in order to ensure the quality of exported products. Sources of errors include human, hardware, and software factors as well as new rules and transactions. The overall quality strategy includes prevention, detection, and correction of errors. The quality assurance process encompasses simple data specification, internal consistency, inspection procedures and, eventually, field testing. The quality assurance system is driven by a small number of tables and UNIX scripts, with "business rules" declared explicitly as Structured Query Language (SQL) statements. Concurrent authorship, client-server technology, and an initial failure to implement robust transaction control have all provided valuable lessons. The feedback loop for error management needs to be short.
JFQ), Summer 1995.  Alberts , C.J. & Dorofee, A.J., “Mission Assurance Analysis Protocol (MAAP): Assessing Risk in Complex Environments... CAMUS : Automatically Mapping Cyber Assets to Missions and Users,” Proc. of the 2010 Military Communications Conference (MILCOM 2009), 2009. [23
Tyszko, Jason A.
Recent criticism of higher education accreditation has prompted calls for reform and sparked interest in piloting alternative quality assurance methods that better address student learning and employment outcomes. Although this debate has brought much needed attention to improving the outcomes of graduates and safeguarding federal investment in…
As the author's contribution to a series marking the Golden Jubilee of the Association of University Administrators, he reflects on changes in quality assurance over the past twenty years and speculates on what the future may hold for quality as the association moves into a new and very different competitive regime. He begins by discussing the…
Eicke, W. G.; Riley, J. P.; Riley, K. J.
The 5300.4 series of NASA Handbooks for Reliability and Quality Assurance Programs have provisions for the establishment and utilization of a documented metrology system to control measurement processes and to provide objective evidence of quality conformance. The intent of these provisions is to assure consistency and conformance to specifications and tolerances of equipment, systems, materials, and processes procured and/or used by NASA, its international partners, contractors, subcontractors, and suppliers. This Measurement Assurance Program (MAP) guideline has the specific objectives to: (1) ensure the quality of measurements made within NASA programs; (2) establish realistic measurement process uncertainties; (3) maintain continuous control over the measurement processes; and (4) ensure measurement compatibility among NASA facilities. The publication addresses MAP methods as applied within and among NASA installations and serves as a guide to: control measurement processes at the local level (one facility); conduct measurement assurance programs in which a number of field installations are joint participants; and conduct measurement integrity (round robin) experiments in which a number of field installations participate to assess the overall quality of particular measurement processes at a point in time.
Trochim, William M. K.; Visco, Ronald J.
A number of quality assurance educational evaluation methods are illustrated. Evaluation data obtained from the Providence, Rhode Island, school district are used. The methods are: (1) from auditing, internal control; (2) from accounting, double bookkeeping; and (3) from industrial quality control, acceptance sampling and cumulative percentage…
Bocock, Jean; Edwards, Judith
This bulletin is intended to help British further education colleges clarify their rationale for entering into collaborative programs, assess prospective partners, define and implement good practice at all stages of provision, and establish rigorous quality assurance procedures. Following an introduction, Further Education Funding Council…
Quality assurance in clinical studies aiming at approval of pharmaceutical products is submitted to strict rules, controls and auditing regulations. Comparative instruments to ensure quality in diagnostic and therapeutic procedures are not available in interventional cardiology, likewise in other fields of cardiovascular medicine. Quality assurance simply consists of "quality registers" with basic data not externally controlled. Based on the experiences of clinical studies and their long history of standardization it is assumed that these data may be severely flawed thus being inappropriate to set standards for diagnostic and therapeutic strategies. The precondition for quality assurance are quality data. In invasive coronary angiography and intervention medical indications, the decision making process interventional versus surgical revascularization, technical performance and after - care are essential aspects affecting quality of diagnostics and therapy. Quality data are externally controlled data. To collect quality data an appropriate infrastructure is a necessary precondition which is not existent. For an appropriate infrastructure investments have to be done both to build up as well as to sustain the necessary preconditions. As long as there are no infrastructure and no investments there will be no "quality data". There exist simply registers of data which are not proved to be a basis for significant assurance and enhancement in quality in interventional coronary cardiology. Georg Thieme Verlag KG Stuttgart, New York.
Newell, Susan; Schoenike, Sumner L.; Lisko, Elaine A.
School nurses need to become more influential administrators, managers, and entrepreneurs. They must learn to lead and collaborate effectively in designing, implementing, and evaluating coordinated school health programs. Quality assurance is an essential ingredient in this process that requires accurate, timely, and confidential incident…
Robins, J L
To meet the current standards of excellence necessary for blood establishments, we have learned from industry that a movement toward organization-wide quality assurance/total quality management must be made. Everyone in the organization must accept responsibility for participating in providing the highest quality products and services. Quality must be built into processes and design systems to support these quality processes. Quality assurance has been redefined to include a quality planning function described as the most effective way of designing quality into processes. A formalized quality planning process must be part of quality assurance. Continuous quality improvement has been identified as the strategy every blood establishment must support while striving for error-free processing as the long-term objective. The auditing process has been realigned to support and facilitate this same objective. Implementing organization-wide quality assurance/total quality management is one proven plan for guaranteeing the quality of the 20 million products that are transfused into 4 million patients each year and for moving toward the new order.
Quality assurance or quality control is a term and concept coming from the industry. Here it is most important. All products must have a minimum quality and variation in size, for example, must be kept within certain strict limits. There must be a system to control this. May be not every single product is controlled, but spot tests must be taken. Measures must be taken to improve the quality if it is not good enough. This concept has been transferred to medicine, odontology, and consequently also to forensic odontology. These areas have in common with industry the production of that certain products. However, they are usually handmade and not produced in an industrial process. In addition, dentistry is a great deal of art and judgement and quality control of these factors may be difficult. In this paper, I will focus on forensic odontology. What are the problems? What can we do and cannot do? In addition, how can we assure the quality of the work, the assessment and conclusion, and the report? I have some personal opinions on that and I will give some suggestions. Quality assurance on an international level is difficult. Conditions and juridical systems are different in different countries. Especially forensic odontologists are different and have different opinions. This presentation will be relevant to the ongoing discussion and attempts at revising the IOFOS' guidelines for quality assurance.
Ristevski, Blagoj; Chen, Ming
This paper surveys big data with highlighting the big data analytics in medicine and healthcare. Big data characteristics: value, volume, velocity, variety, veracity and variability are described. Big data analytics in medicine and healthcare covers integration and analysis of large amount of complex heterogeneous data such as various - omics data (genomics, epigenomics, transcriptomics, proteomics, metabolomics, interactomics, pharmacogenomics, diseasomics), biomedical data and electronic health records data. We underline the challenging issues about big data privacy and security. Regarding big data characteristics, some directions of using suitable and promising open-source distributed data processing software platform are given.
Many recent nonproliferation and arms control software projects include a software authentication regime. These include U.S. Government-sponsored projects both in the United States and in the Russian Federation (RF). This trend toward requiring software authentication is only accelerating. Demonstrating assurance that software performs as expected without hidden ''backdoors'' is crucial to a project's success. In this context, ''authentication'' is defined as determining that a software package performs only its intended purpose and performs said purpose correctly and reliably over the planned duration of an agreement. In addition to visual inspections by knowledgeable computer scientists, automated tools are needed to highlightmore » suspicious code constructs, both to aid visual inspection and to guide program development. While many commercial tools are available for portions of the authentication task, they are proprietary and not extensible. An open-source, extensible tool can be customized to the unique needs of each project (projects can have both common and custom rules to detect flaws and security holes). Any such extensible tool has to be based on a complete language compiler. ROSE is precisely such a compiler infrastructure developed within the Department of Energy (DOE) and targeted at the optimization of scientific applications and user-defined libraries within large-scale applications (typically applications of a million lines of code). ROSE is a robust, source-to-source analysis and optimization infrastructure currently addressing large, million-line DOE applications in C and C++ (handling the full C, C99, C++ languages and with current collaborations to support Fortran90). We propose to extend ROSE to address a number of security-specific requirements, and apply it to software authentication for nonproliferation and arms control projects.« less
Just as much as healthcare information systems, medical devices need protection against cybersecurity threats. Anti-malware software can help safeguard the devices in your facility-but it has limitations and even risks. Find out what steps you can take to manage anti-malware applications in your devices.
The Institute for Water Resources (IWR) prepared this white paper on financial assurance for mitigation project success to provide a reference resource for Corps district staff involved with establishing and overseeing financial assurances.
Topics concerning the Common Lunar Lander for the Space Exploration Initiative are covered and include the following: product assurance tools and supports; project goals; and product assurance structured for optimal payback.
This Technical Support Document (TSD) supports EPA’s determination that the final Transport Rule’s assurance provision penalty requirement provides sufficient deterrence against a state exceeding its assurance level.
With the shift from quality control (QC) of materials and placement techniques : to quality assurance (QA) and acceptance over the years, the role of the Office : of Materials Technology (OMT) has been shifting towards assurance of : material quality...
White, Allan L.; Sjogren, Jon A.
Preliminary design developed for validation experiment that addresses problems unique to assuring extremely high quality of multiple-version programs in process-control software. Approach takes into account inertia of controlled system in sense it takes more than one failure of control program to cause controlled system to fail. Verification procedure consists of two steps: experimentation (numerical simulation) and computation, with Markov model for each step.
Hird, Geoffrey R.
The use of formal methods in software development achieves levels of quality assurance unobtainable by other means. The Larch approach to specification is described, and the specification of avionics software designed to implement the logic of a flight control system is given as an example. Penelope is described which is an Ada-verification environment. The Penelope user inputs mathematical definitions, Larch-style specifications and Ada code and performs machine-assisted proofs that the code obeys its specifications. As an example, the verification of a binary search function is considered. Emphasis is given to techniques assisting the reuse of a verification effort on modified code.
This paper provides an update and extension of a 2006 paper, “Some Future Trends and Implications for Systems and Software Engineering Processes,” Systems Engineering, Spring 2006. Some of its challenges and opportunities are similar, such as the need to simultaneously achieve high levels of both agility and assurance. Others have emerged as increasingly important, such as the challenges of dealing with ultralarge volumes of data, with multicore chips, and with software as a service. The paper is organized around eight relatively surprise-free trends and two “wild cards” whose trends and implications are harder to foresee. The eight surprise-free trends are:
Rutledge, E O
All healthcare providers, both institutional and individual, must make every effort to ensure that every person who seeks their medical care is offered competent, sincere, and equal treatment options. Unfortunately, this ideal scenario does not take into account the lack of diversity among care providers and the lack of culturally competent policies within healthcare delivery settings. As a result, many care providers continue to follow racially biased treatment practices and many organizations continue to ignore their public trust of providing fair treatment to everyone, regardless of skin color, gender, economic capabilities, etc. Although developing and implementing a diversity plan and culturally competent policies is very complex practically, politically, and programmatically for traditional institutional care providers, it must be done. The key ingredient to this effort is the absolute commitment and support of the organization's governing bodies and executive management. Institutions can certainly volunteer and begin to develop such programs that foster recruitment, selection, and retention of culturally competent care providers to ensure that equal healthcare is received by their patient populations. However, many institutions are already besieged by too many healthcare challenges to volunteer for such an effort. The Joint Commission on Accreditation of Healthcare Organizations and the National Council of Quality Assurance can certainly help jumpstart this effort by establishing an accreditation standard that requires all healthcare providers to establish and practice culturally competent care within their organizations. Providers must also embrace the diversity that is a part of our society and must not let race or ethnicity be a determining factor in offering treatment options.
Hayhurst, Kelly J.
Software is becoming increasingly significant in today's critical avionics systems. To achieve safe, reliable software, government regulatory agencies such as the Federal Aviation Administration (FAA) and the Department of Defense mandate the use of certain software development methods. However, little scientific evidence exists to show a correlation between software development methods and product quality. Given this lack of evidence, a series of experiments has been conducted to understand why and how software fails. The Guidance and Control Software (GCS) project is the latest in this series. The GCS project is a case study of the Requirements and Technical Concepts for Aviation RTCA/DO-178B guidelines, Software Considerations in Airborne Systems and Equipment Certification. All civil transport airframe and equipment vendors are expected to comply with these guidelines in building systems to be certified by the FAA for use in commercial aircraft. For the case study, two implementations of a guidance and control application were developed to comply with the DO-178B guidelines for Level A (critical) software. The development included the requirements, design, coding, verification, configuration management, and quality assurance processes. This paper discusses the details of the GCS project and presents the results of the case study.
... 10 Energy 2 2013-01-01 2013-01-01 false Quality assurance. 76.93 Section 76.93 Energy NUCLEAR REGULATORY COMMISSION (CONTINUED) CERTIFICATION OF GASEOUS DIFFUSION PLANTS Safety § 76.93 Quality assurance. The Corporation shall establish, maintain, and execute a quality assurance program satisfying each of...
... 10 Energy 2 2012-01-01 2012-01-01 false Quality assurance. 76.93 Section 76.93 Energy NUCLEAR REGULATORY COMMISSION (CONTINUED) CERTIFICATION OF GASEOUS DIFFUSION PLANTS Safety § 76.93 Quality assurance. The Corporation shall establish, maintain, and execute a quality assurance program satisfying each of...
... 10 Energy 2 2014-01-01 2014-01-01 false Quality assurance. 76.93 Section 76.93 Energy NUCLEAR REGULATORY COMMISSION (CONTINUED) CERTIFICATION OF GASEOUS DIFFUSION PLANTS Safety § 76.93 Quality assurance. The Corporation shall establish, maintain, and execute a quality assurance program satisfying each of...
... 10 Energy 2 2011-01-01 2011-01-01 false Quality assurance. 76.93 Section 76.93 Energy NUCLEAR REGULATORY COMMISSION (CONTINUED) CERTIFICATION OF GASEOUS DIFFUSION PLANTS Safety § 76.93 Quality assurance. The Corporation shall establish, maintain, and execute a quality assurance program satisfying each of...
... 14 Aeronautics and Space 5 2012-01-01 2012-01-01 false Assurances required. 1251.104 Section 1251.104 Aeronautics and Space NATIONAL AERONAUTICS AND SPACE ADMINISTRATION NONDISCRIMINATION ON BASIS OF HANDICAP General Provisions § 1251.104 Assurances required. (a) Assurances. An applicant for Federal...
... 14 Aeronautics and Space 5 2013-01-01 2013-01-01 false Assurances required. 1251.104 Section 1251.104 Aeronautics and Space NATIONAL AERONAUTICS AND SPACE ADMINISTRATION NONDISCRIMINATION ON BASIS OF HANDICAP General Provisions § 1251.104 Assurances required. (a) Assurances. An applicant for Federal...
... 14 Aeronautics and Space 5 2011-01-01 2010-01-01 true Assurances required. 1251.104 Section 1251.104 Aeronautics and Space NATIONAL AERONAUTICS AND SPACE ADMINISTRATION NONDISCRIMINATION ON BASIS OF HANDICAP General Provisions § 1251.104 Assurances required. (a) Assurances. An applicant for Federal...
... 14 Aeronautics and Space 5 2010-01-01 2010-01-01 false Assurances required. 1251.104 Section 1251.104 Aeronautics and Space NATIONAL AERONAUTICS AND SPACE ADMINISTRATION NONDISCRIMINATION ON BASIS OF HANDICAP General Provisions § 1251.104 Assurances required. (a) Assurances. An applicant for Federal...
... 30 Mineral Resources 1 2014-07-01 2014-07-01 false Quality assurance. 74.9 Section 74.9 Mineral... DUST SAMPLING DEVICES Requirements for Continuous Personal Dust Monitors § 74.9 Quality assurance. (a) General requirements. The applicant shall establish and maintain a quality control system that assures...
... 50 Wildlife and Fisheries 12 2013-10-01 2013-10-01 false Security assurances. 600.240 Section 600... ADMINISTRATION, DEPARTMENT OF COMMERCE MAGNUSON-STEVENS ACT PROVISIONS Council Membership § 600.240 Security assurances. (a) DOC Office of Security will issue security assurances to Council members following completion...
... 50 Wildlife and Fisheries 12 2012-10-01 2012-10-01 false Security assurances. 600.240 Section 600... ADMINISTRATION, DEPARTMENT OF COMMERCE MAGNUSON-STEVENS ACT PROVISIONS Council Membership § 600.240 Security assurances. (a) DOC Office of Security will issue security assurances to Council members following completion...
... 50 Wildlife and Fisheries 10 2011-10-01 2011-10-01 false Security assurances. 600.240 Section 600... ADMINISTRATION, DEPARTMENT OF COMMERCE MAGNUSON-STEVENS ACT PROVISIONS Council Membership § 600.240 Security assurances. (a) DOC Office of Security will issue security assurances to Council members following completion...
... 40 Protection of Environment 1 2010-07-01 2010-07-01 false Quality assurance. 30.54 Section 30.54... NON-PROFIT ORGANIZATIONS Post-Award Requirements Reports and Records § 30.54 Quality assurance. If the... data generation, the grantee shall develop and implement quality assurance practices consisting of...
... 10 Energy 2 2010-01-01 2010-01-01 false Quality assurance. 76.93 Section 76.93 Energy NUCLEAR REGULATORY COMMISSION (CONTINUED) CERTIFICATION OF GASEOUS DIFFUSION PLANTS Safety § 76.93 Quality assurance. The Corporation shall establish, maintain, and execute a quality assurance program satisfying each of...
... 40 Protection of Environment 24 2010-07-01 2010-07-01 false Quality assurance. 194.22 Section 194... General Requirements § 194.22 Quality assurance. (a)(1) As soon as practicable after April 9, 1996, the Department shall adhere to a quality assurance program that implements the requirements of ASME NQA-1-1989...
... 40 Protection of Environment 1 2010-07-01 2010-07-01 false Quality assurance. 31.45 Section 31.45... Requirements Reports, Records, Retention, and Enforcement § 31.45 Quality assurance. If the grantee's project... quality assurance practices consisting of policies, procedures, specifications, standards, and...
software license, software usage, ELA, Software as a Service , SaaS , Software Asset...PaaS Platform as a Service SaaS Software as a Service SAM Software Asset Management SMS System Management Server SEWP Solutions for Enterprise Wide...delivery of full Cloud Services , we will see the transition of the Cloud Computing service model from Iaas to SaaS , or Software as a Service . Software
Sijmons, Jaap G; Winter, Heinrich B; Hubben, Joep H
In 2013 the Dutch Individual Healthcare Professions Act (known as the BIG law) was evaluated for the second time. The research showed that patients have limited awareness of the registration of healthcare professionals and that the system of reserved procedures is almost unknown. On the other hand, healthcare institutions (especially hospitals) frequently check the register, as do healthcare insurance companies when contracting institutions. Knowledge of the reserved procedures system is moderate amongst professionals too, while the organisation of care is to a great extent based on this system. Since the change of system in 2006 quality assurance in professional practice has been much more rooted in the internal structure of care; in this way, the BIG law did not go the way the legislator intended. According to the researchers, this has not prevented the BIG law from still playing an essential function. Indeed, the BIG law has not reached its final destination, but it may reach its goal via another route.
Albrecht, Martin; Loos, Stefan; Otten, Marcus
Overcoming rigid sectoral segmentation in healthcare has also become a health policy target in quality assurance. With the Act to Enhance Competition in Statutory Health Insurance (GKV-WSG) coming into effect, quality assurance measures are to be designed in a cross-sectoral fashion for in- and outpatient sectors equally. An independent institution is currently mandated to develop specific quality indicators for eleven indications. For three of these operating tests have already been commissioned by the Federal Joint Committee. This article depicts the major results of a feasibility study, including a compliance cost estimate, for the aforementioned indications of cross-sectoral quality assurance (cQA). In conclusion, a number of both practical and conceptual basic challenges are still to be resolved prior to the full implementation of cQA, such as a sufficient specification to activate documentation requirements and an inspection system capable of separating actual quality problems from documentary deficits. So far, a comprehensive cost-utility analysis of cQA has not been provided, in particular with comparison to existing QA systems. In order to optimise cost and utility of cQA an evidence-based approach is required for both the extension of cQA areas and for QA provisions. Copyright © 2013. Published by Elsevier GmbH.
ESA has allocated funds for an initiative to evaluate Dependability and Safety methods of Software. The objectives of this initiative are; · More extensive validation of Safety and Dependability techniques for Software · Provide valuable results to improve the quality of the Software thus promoting the application of Dependability and Safety methods and techniques. ESA space systems are being developed according to defined PA requirement specifications. These requirements may be implemented through various design concepts, e.g. redundancy, diversity etc. varying from project to project. Analysis methods (FMECA. FTA, HA, etc) are frequently used during requirements analysis and design activities to assure the correct implementation of system PA requirements. The criticality level of failures, functions and systems is determined and by doing that the critical sub-systems are identified, on which dependability and safety techniques are to be applied during development. Proper performance of the software development requires the development of a technical specification for the products at the beginning of the life cycle. Such technical specification comprises both functional and non-functional requirements. These non-functional requirements address characteristics of the product such as quality, dependability, safety and maintainability. Software in space systems is more and more used in critical functions. Also the trend towards more frequent use of COTS and reusable components pose new difficulties in terms of assuring reliable and safe systems. Because of this, its dependability and safety must be carefully analysed. ESA identified and documented techniques, methods and procedures to ensure that software dependability and safety requirements are specified and taken into account during the design and development of a software system and to verify/validate that the implemented software systems comply with these requirements [R1].
Grant, Simon; Wheatley, R J
To create and implement a system through which pre-hospital healthcare activity across an entire operational theatre could be made available in real-time to support healthcare delivery, governance and assurance activity. An IT-based system was created that could display, manage and integrate the pre-hospital healthcare activity on Op HERRICK 18. The system was based on the Defence Medical Services Common Assurance Framework and run through Microsoft Office SharePoint. Pre-hospital healthcare activity was made available and visible across an operational theatre. This supported delivery, assurance and governance at any time. Activity from each medical facility could be integrated and display automatically improving theatre wide situational awareness. The availability of information resulted in a shift towards a more continuous process of assurance and governance rather than reliance on inherently threatening and increasingly intermittent inspection regimes. The ability to review healthcare activity remotely at anytime significantly improves the validity of assurance possible for a deployed force. Governance activity can be more responsive and less reliant on the fixed timescale and datasets of reports from outlying medical facilities. However, assurance and governance authorities must not allow such a wealth of information to impact local leadership and innovation through a perception of, or actual, micro-management. Published by the BMJ Publishing Group Limited. For permission to use (where not already granted under a licence) please go to http://group.bmj.com/group/rights-licensing/permissions.
Bailey, P H
Many nurse-researchers using qualitative strategies have been concerned with assuring quality in their work. The early literature reveals that the concepts of validity and reliability, as understood from the positivist perspective, are somehow inappropriate and inadequate when applied to interpretive research. More recent literature suggests that because of the positivist and interpretive paradigms are epistemologically divergent, the transfer of quality criteria from one perspective to the other is not automatic or even reasonable. The purpose of this article, therefore, is to clarify what the terms quality, trustworthiness, credibility, authenticity, and goodness mean in qualitative research findings. The process of assuring quality, validation, in qualitative research will be discussed within the context of the interpretive method, narrative analysis. A brief review of quality in narrative analysis nursing research will also be presented.
Alexander, Michael G.
Requirement Assurance is an act of requirement verification which assures the stakeholder or customer that a product requirement has produced its "as realized product" and has been verified with conclusive evidence. Product requirement verification answers the question, "did the product meet the stated specification, performance, or design documentation?". In order to ensure the system was built correctly, the practicing system engineer must verify each product requirement using verification methods of inspection, analysis, demonstration, or test. The products of these methods are the "verification artifacts" or "closure artifacts" which are the objective evidence needed to prove the product requirements meet the verification success criteria. Institutional direction is given to the System Engineer in NPR 7123.1A NASA Systems Engineering Processes and Requirements with regards to the requirement verification process. In response, the verification methodology offered in this report meets both the institutional process and requirement verification best practices.
Kabara, Joseph; Krishnamurthy, Prashant; Tipper, David
Emerging wireless networks will contain a hybrid infrastructure based on fixed, mobile and ad hoc topologies and technologies. In such a dynamic architecture, we define information assurance as the provisions for both information security and information availability. The implications of this definition are that the wireless network architecture must (a) provide sufficient security measures, (b) be survivable under node or link attack or failure and (c) be designed such that sufficient capacity remains for all critical services (and preferably most other services) in the event of attack or component failure. We have begun a research project to investigate the provision of information assurance for wireless networks viz. survivability, security and availability and here discuss the issues and challenges therein.
Chamis, Christos c.
A computational simulation method is presented to assure life in composite systems by using dynamic buckling of smart composite shells as an example. The combined use of composite mechanics, finite element computer codes, and probabilistic analysis enable the effective assessment of the dynamic buckling load of smart composite shells. A universal plot is generated to estimate the dynamic buckling load of composite shells at various load rates and probabilities. The shell structure is also evaluated with smart fibers embedded in the plies right below the outer plies. The results show that, on the average, the use of smart fibers improved the shell buckling resistance by about 9% at different probabilities and delayed the buckling occurrence time. The probabilistic sensitivities results indicate that uncertainties in the fiber volume ratio and ply thickness have major effects on the buckling load. The uncertainties in the electric field strength and smart material volume fraction have moderate effects and thereby in the assured life of the shell.
Kaufman, James H; Eiron, Iris; Deen, Glenn; Ford, Dan A; Smith, Eishay; Knoop, Sarah; Nelken, H; Kol, Tomer; Mesika, Yossi; Witting, Karen; Julier, Kevin; Bennett, Craig; Rapp, Bill; Carmeli, Boaz; Cohen, Simona
Recently there has been increased focus on the need to modernize the healthcare information infrastructure in the United States.1–4 The U.S. healthcare industry is by far the largest in the world in both absolute dollars and in percentage of GDP (more than $1.5 trillion, or 15 percent of GDP). It is also fragmented and complex. These difficulties, coupled with an antiquated infrastructure for the collection of and access to medical data, lead to enormous inefficiencies and sources of error. Consumer, regulatory, and governmental pressure drive a growing consensus that the time has come to modernize the U.S. healthcare information infrastructure (HII). While such transformation may be disruptive in the short term, it will, in the future, significantly improve the quality, expediency, efficiency, and successful delivery of healthcare while decreasing costs to patients and payers and improving the overall experiences of consumers and providers. The launch of a national health infrastructure initiative in the United States in May 2004-with the goal of providing an electronic health record for every American within the next decade-will eventually transform the healthcare industry in general, just as information technology (IT) has transformed other industries in the past. The key to this successful outcome will be based on the way we apply IT to healthcare data and the services delivered through IT. This must be accomplished in a way that protects individuals and allows competition but gives caregivers reliable and efficient access to the data required to treat patients and to improve the practice of medical science. This paper describes key IT solutions and technologies that address the challenges of creating a nation-wide healthcare IT infrastructure. Furthermore we discuss the emergence of new electronic healthcare services and the current efforts of IBM Research, Software Group, and Healthcare Life Sciences to realize this new vision for healthcare. PMID:18066378
As spaceflight systems increase in complexity, Fault Management (FM) systems are ranked high in risk-based assessment of software criticality, emphasizing the importance of establishing highly competent domain expertise to provide assurance. Adverse conditions (ACs) and specific vulnerabilities encountered by safety- and mission-critical software systems have been identified through efforts to reduce the risk posture of software-intensive NASA missions. Acknowledgement of potential off-nominal conditions and analysis to determine software system resiliency are important aspects of hazard analysis and FM. A key component of assuring FM is an assessment of how well software addresses susceptibility to failure through consideration of ACs. Focus on significant risk predicted through experienced analysis conducted at the NASA Independent Verification Validation (IVV) Program enables the scoping of effective assurance strategies with regard to overall asset protection of complex spaceflight as well as ground systems. Research efforts sponsored by NASA's Office of Safety and Mission Assurance defined terminology, categorized data fields, and designed a baseline repository that centralizes and compiles a comprehensive listing of ACs and correlated data relevant across many NASA missions. This prototype tool helps projects improve analysis by tracking ACs and allowing queries based on project, mission type, domaincomponent, causal fault, and other key characteristics. Vulnerability in off-nominal situations, architectural design weaknesses, and unexpected or undesirable system behaviors in reaction to faults are curtailed with the awareness of ACs and risk-significant scenarios modeled for analysts through this database. Integration within the Enterprise Architecture at NASA IVV enables interfacing with other tools and datasets, technical support, and accessibility across the Agency. This paper discusses the development of an improved workflow process utilizing this
As spaceflight systems increase in complexity, Fault Management (FM) systems are ranked high in risk-based assessment of software criticality, emphasizing the importance of establishing highly competent domain expertise to provide assurance. Adverse conditions (ACs) and specific vulnerabilities encountered by safety- and mission-critical software systems have been identified through efforts to reduce the risk posture of software-intensive NASA missions. Acknowledgement of potential off-nominal conditions and analysis to determine software system resiliency are important aspects of hazard analysis and FM. A key component of assuring FM is an assessment of how well software addresses susceptibility to failure through consideration of ACs. Focus on significant risk predicted through experienced analysis conducted at the NASA Independent Verification & Validation (IV&V) Program enables the scoping of effective assurance strategies with regard to overall asset protection of complex spaceflight as well as ground systems. Research efforts sponsored by NASAs Office of Safety and Mission Assurance (OSMA) defined terminology, categorized data fields, and designed a baseline repository that centralizes and compiles a comprehensive listing of ACs and correlated data relevant across many NASA missions. This prototype tool helps projects improve analysis by tracking ACs and allowing queries based on project, mission type, domain/component, causal fault, and other key characteristics. Vulnerability in off-nominal situations, architectural design weaknesses, and unexpected or undesirable system behaviors in reaction to faults are curtailed with the awareness of ACs and risk-significant scenarios modeled for analysts through this database. Integration within the Enterprise Architecture at NASA IV&V enables interfacing with other tools and datasets, technical support, and accessibility across the Agency. This paper discusses the development of an improved workflow process utilizing
Clark, G B
This is the fourth and final part of a series of articles on laboratory quality surveillance. Part I addressed the historical background of medical quality assurance. Part II covered surveillance guidelines of the Joint Commission on Accreditation of Healthcare Organizations (JCAHO) and the College of American Pathologists with emphasis on quality assurance (QA) and the ten-step process. Part III focused on the JCAHO transition from QA to quality assessment and improvement. Part IV concludes the series by discussing the systematic identification of quality indicators in the total quality management and continuous quality improvement environment.
... Personnel PPE Training Infection Control Assessment Tools Water Management Programs Map: HAI Prevention Activities Research CDC Supported Projects Prevention Epicenters (PE) Healthcare Safety Research (SHEPheRD) Environmental ...
Postma, Jeroen; Roos, Anne-Fleur
In many OECD countries, healthcare sectors have become increasingly concentrated as a result of mergers. However, detailed empirical insight into why healthcare providers merge is lacking. Also, we know little about the influence of national healthcare policies on mergers. We fill this gap in the literature by conducting a survey study on mergers among 848 Dutch healthcare executives, of which 35% responded (resulting in a study sample of 239 executives). A total of 65% of the respondents was involved in at least one merger between 2005 and 2012. During this period, Dutch healthcare providers faced a number of policy changes, including increasing competition, more pressure from purchasers, growing financial risks, de-institutionalisation of long-term care and decentralisation of healthcare services to municipalities. Our empirical study shows that healthcare providers predominantly merge to improve the provision of healthcare services and to strengthen their market position. Also efficiency and financial reasons are important drivers of merger activity in healthcare. We find that motives for merger are related to changes in health policies, in particular to the increasing pressure from competitors, insurers and municipalities.
Ibbott, Geoffrey S.; Haworth, Annette; Followill, David S.
Cooperative groups, of which the Radiation Therapy Oncology Group is one example, conduct national clinical trials that often involve the use of radiation therapy. In preparation for such a trial, the cooperative group prepares a protocol to define the goals of the trial, the rationale for its design, and the details of the treatment procedure to be followed. The Radiological Physics Center (RPC) is one of several quality assurance (QA) offices that is charged with assuring that participating institutions deliver doses that are clinically consistent and comparable. The RPC does this by conducting a variety of independent audits and credentialing processes. The RPC has compiled data showing that credentialing can help institutions comply with the requirements of a cooperative group clinical protocol. Phantom irradiations have been demonstrated to exercise an institution’s procedures for planning and delivering advanced external beam techniques (1–3). Similarly, RPC data indicate that a rapid review of patient treatment records or planning procedures can improve compliance with clinical trials (4). The experiences of the RPC are presented as examples of the contributions that a national clinical trials QA center can make to cooperative group trials. These experiences illustrate the critical need for comprehensive QA to assure that clinical trials are successful and cost-effective. The RPC is supported by grants CA 10953 and CA 81647 from the National Cancer Institute, NIH, DHHS. PMID:24392352
... quality assurance actions. 246.470 Section 246.470 Federal Acquisition Regulations System DEFENSE ACQUISITION REGULATIONS SYSTEM, DEPARTMENT OF DEFENSE CONTRACT MANAGEMENT QUALITY ASSURANCE Government Contract Quality Assurance 246.470 Government contract quality assurance actions. ...
Describes types of software including system software, application software, spreadsheets, accounting software, graphics packages, desktop publishing software, database, desktop and personal information management software, project and records management software, groupware, and shareware. (JOW)
Addy, Edward A.
Verification and Validation (V&V) is used to increase the level of assurance of critical software, particularly that of safety-critical and mission-critical software. V&V is a systems engineering discipline that evaluates the software in a systems context, and is currently applied during the development of a specific application system. In order to bring the effectiveness of V&V to bear within reuse-based software engineering, V&V must be incorporated within the domain engineering process.
Callender, E. David; Steinbacher, Jody
This is the fourth of five volumes on Information System Life-Cycle and Documentation Standards. This volume provides a well organized, easily used standard for assurance documentation for information systems and software, hardware, and operational procedures components, and related processes. The specifications are developed in conjunction with the corresponding management plans specifying the assurance activities to be performed.
Ko, Inseok; Chang, Hyejung
Big data analysis is receiving increasing attention in many industries, including healthcare. Visualization plays an important role not only in intuitively showing the results of data analysis but also in the whole process of collecting, cleaning, analyzing, and sharing data. This paper presents a procedure for the interactive visualization and analysis of healthcare data using Tableau as a business intelligence tool. Starting with installation of the Tableau Desktop Personal version 10.3, this paper describes the process of understanding and visualizing healthcare data using an example. The example data of colon cancer patients were obtained from health insurance claims in years 2012 and 2013, provided by the Health Insurance Review and Assessment Service. To explore the visualization of healthcare data using Tableau for beginners, this paper describes the creation of a simple view for the average length of stay of colon cancer patients. Since Tableau provides various visualizations and customizations, the level of analysis can be increased with small multiples, view filtering, mark cards, and Tableau charts. Tableau is a software that can help users explore and understand their data by creating interactive visualizations. The software has the advantages that it can be used in conjunction with almost any database, and it is easy to use by dragging and dropping to create an interactive visualization expressing the desired format.
Thomas, David J.; Weiner, Jayne; Lippincott, Ronald C.
SYN·OP·SYS™ is a computerized management information system for quality assurance and risk management. Computer software for the efficient collection and analysis of “occurrences” and the clinical data associated with these kinds of patient events is described. The system is evaluated according to certain computer design criteria, and the system's implementation is assessed.
Carnevale, Anthony P.; Smith, Nicole; Gulish, Artem; Beach, Bennett H.
This report projects education requirements linked to forecasted job growth in healthcare by state and the District of Columbia from 2010 through 2020. It complements a larger national report which projects educational demand for healthcare for the same time period. The national report shows that with or without Obamacare, the United States will…
Griner, Carolyn S.; Schneider, Michelle
The challenge of a global operations capability led to the Telescience Resource Kit (TReK) project, an in-house software development project of the Mission Operations Laboratory (MOL) at NASA's Marshall Space Flight Center (MSFC). The TReK system is being developed as an inexpensive comprehensive personal computer- (PC-) based ground support system that can be used by payload users from their home sites to interact with their payloads on board the International Space Station (ISS). The TReK project is currently using a combination of the spiral lifecycle model and the incremental lifecycle model. As with any software development project, there are four activities that can be very time consuming: Software design and development, project documentation, testing, and umbrella activities, such as quality assurance and configuration management. In order to produce a quality product, it is critical that each of these activities receive the appropriate amount of attention. For TReK, the challenge was to lay out a lifecycle and project plan that provides full support for these activities, is flexible, provides a way to deal with changing risks, can accommodate unknowns, and can respond to changes in the environment quickly. This paper will provide an overview of the TReK lifecycle, a description of the project's environment, and a general overview of project activities.
This Software Guide (SG) describes the software developed to test the Assessment Environment for Complex Systems (AECS) by the West Virginia High Technology Consortium (WVHTC) Foundation's Mission Systems Group (MSG) for the National Aeronautics and Space Administration (NASA) Aeronautics Research Mission Directorate (ARMD). This software is referred to as the AECS Test Project throughout the remainder of this document. AECS provides a framework for developing, simulating, testing, and analyzing modern avionics systems within an Integrated Modular Avionics (IMA) architecture. The purpose of the AECS Test Project is twofold. First, it provides a means to test the AECS hardware and system developed by MSG. Second, it provides an example project upon which future AECS research may be based. This Software Guide fully describes building, installing, and executing the AECS Test Project as well as its architecture and design. The design of the AECS hardware is described in the AECS Hardware Guide. Instructions on how to configure, build and use the AECS are described in the User's Guide. Sample AECS software, developed by the WVHTC Foundation, is presented in the AECS Software Guide. The AECS Hardware Guide, AECS User's Guide, and AECS Software Guide are authored by MSG. The requirements set forth for AECS are presented in the Statement of Work for the Assessment Environment for Complex Systems authored by NASA Dryden Flight Research Center (DFRC). The intended audience for this document includes software engineers, hardware engineers, project managers, and quality assurance personnel from WVHTC Foundation (the suppliers of the software), NASA (the customer), and future researchers (users of the software). Readers are assumed to have general knowledge in the field of real-time, embedded computer software development.
Griffith, J R
Stakeholders will put increasing pressure on integrated health systems (IHS) for measured performance, demanding data on quality and patient satisfaction, while simultaneously pressing for lower cost. The changes to Joint Commission on Accreditation of Healthcare Organizations (Joint Commission) and the growing importance of the National Committee on Quality Assurance (NCQA) are simply forerunners of an intensifying trend. Quality of care in particular will face increasing scrutiny. Achieving competitive targets in these areas will also require measures addressing demand and worker satisfaction. "Balanced scorecard" approaches will allow IHS and their accountable work groups to track performance on several dimensions and establish integrated goals or targets. Those with consistently good scores will be labeled "champions." Champions will support the multidimensional measures with improved decision processes. About eight major processes will be central--governance/strategic management, clinical quality, clinical organization, financial planning, planning and marketing, information services, human resources, and plant services. It is possible to map these processes to the criteria of the Joint Commission, NCQA, and Malcolm Baldrige Quality Award. The processes themselves can be measured and common weaknesses identified and corrected. Champions share some common characteristics that seem to arise from the combination of processes and measures. Among these characteristics are service line orientation, extensive partnering with other organizations, and the possibility of outsourcing organizational components.
Pashkov, Vitalii; Gutorova, Nataliya; Harkusha, Andrii
one of the areas of significant growth in medical devices has been the role of software - as an integral component of a medical device, as a standalone device and more recently as applications on mobile devices. The risk related to a malfunction of the standalone software used within healthcare is in itself not a criterion for its qualification or not as a medical device. It is therefore, necessary to clarify some criteria for the qualification of stand-alone software as medical devices Materials and methods: Ukrainian, European Union, United States of America legislation, Guidelines developed by European Commission and Food and Drug Administration's, recommendations represented by international voluntary group and scientific works. This article is based on dialectical, comparative, analytic, synthetic and comprehensive research methods. the legal regulation of software which is used for medical purpose in Ukraine limited to one definition. In European Union and United States of America were developed and applying special guidelines that help developers, manufactures and end users to difference software on types standing on medical purpose criteria. Software becomes more and more incorporated into medical devices. Developers and manufacturers may not have initially appreciated potential risks to patients and users such situation could have dangerous results for patients or users. It is necessary to develop and adopt the legislation that will intend to define the criteria for the qualification of medical device software and the application of the classification criteria to such software, provide some illustrative examples and step by step recommendations to qualify software as medical device.
Rose, P. L.; Willis, B. G.
The proposed software guidelines to be followed by the European Space Research Organization in the development of software for the Spacelab being developed for use as a payload for the space shuttle are documented. Concepts, techniques, and tools needed to assure the success of a programming project are defined as they relate to operation of the data management subsystem, support of experiments and space applications, use with ground support equipment, and for integration testing.
Kadono, Yasuo; Tsubaki, Hiroe; Tsuruho, Seishiro
The purpose of this study is to clarity the mechanism of how software engineering capabilities relate to the business performance of IT vendors in Japan. To do this, we developed a structural model using factors related to software engineering, business performance and competitive environment. By analyzing the data collected from 78 major IT vendors in Japan, we found that superior deliverables and business performance were correlated with the effort expended particularly on human resource development, quality assurance, research and development and process improvement.
DeGruy, K B
Many healthcare leaders find themselves overwhelmed with data, but lack the information they need to make informed decisions. Knowledge discovery in databases (KDD) can help organizations turn their data into information. KDD is the process of finding complex patterns and relationships in data. The tools and techniques of KDD have achieved impressive results in other industries, and healthcare needs to take advantage of advances in this exciting field. Recent advances in the KDD field have brought it from the realm of research institutions and large corporations to many smaller companies. Software and hardware advances enable small organizations to tap the power of KDD using desktop PCs. KDD has been used extensively for fraud detection and focused marketing. There is a wealth of data available within the healthcare industry that would benefit from the application of KDD tools and techniques. Providers and payers have a vast quantity of data (such as, charges and claims), but not effective way to analyze the data to accurately determine relationships and trends. Organizations that take advantage of KDD techniques will find that they offer valuable assistance in the quest to lower healthcare costs while improving healthcare quality.
Tahmasbi, Arezoo; Adabi, Sahar; Rezaee, Ali
The emergence of mobile healthcare systems is an important outcome of application of pervasive computing concepts for medical care purposes. These systems provide the facilities and infrastructure required for automatic and ubiquitous sharing of medical information. Healthcare systems have a dynamic structure and configuration, therefore having an architecture is essential for future development of these systems. The need for increased response rate, problem limited storage, accelerated processing and etc. the tendency toward creating a new generation of healthcare system architecture highlight the need for further focus on cloud-based solutions for transfer data and data processing challenges. Integrity and reliability of healthcare systems are of critical importance, as even the slightest error may put the patients' lives in danger; therefore acquiring a behavioral model for these systems and developing the tools required to model their behaviors are of significant importance. The high-level designs may contain some flaws, therefor the system must be fully examined for different scenarios and conditions. This paper presents a software architecture for development of healthcare systems based on pervasive computing concepts, and then models the behavior of described system. A set of solutions are then proposed to improve the design's qualitative characteristics including, availability, interoperability and performance.
Schneider, Simone M; Popic, Tamara
Knowing the public opinion of healthcare is essential when assessing healthcare system performance; but little research has focussed on the links between the public's general attitude to the healthcare system and its perceptions and expectations of specific healthcare-related aspects. Using data from the fourth round of the European Social Survey 2008/09, we explore the cognitive determinants of global evaluations of the healthcare system in 12 Eastern and 16 Western European countries. We find that healthcare evaluations follow a coherent cognitive reasoning. They are associated with (i) perceptions of the performance of healthcare systems (i.e. efficiency, equality of treatment, health outcomes), (ii) expectations of the government's role in providing healthcare, and (iii) reflections on demographic pressures (i.e. aging populations). Contrary to the general assumption that normative expectations are responsible for differences in healthcare evaluations between Eastern and Western Europe, our results suggest that regional differences are largely due to a more negative perception of the performance of healthcare systems within Eastern Europe. To enhance the public opinion of healthcare, policy makers should improve the efficiency of healthcare systems and take measures to assure equality in health treatment. Copyright © 2018 The Authors. Published by Elsevier B.V. All rights reserved.
On reverse aide it neceeary aid Identify by block number) Software Library; Program Library; Reusability; Generator 20 ABSTRACT (Cmlnue on revere... Software Library. A particular example of the Software Library, the Program Library, is described as a prototype of a reusable library. A hierarchical... programming libraries are described. Finally, non code products in the Software Library are discussed. Accesson Fo NTIS R~jS DrrC TA Availability Codes 0
This report documents a prototype tool developed to investigate the use of visualization and virtual reality technologies for improving software surety confidence. The tool is utilized within the execution phase of the software life cycle. It provides a capability to monitor an executing program against prespecified requirements constraints provided in a program written in the requirements specification language SAGE. The resulting Software Attribute Visual Analysis Tool (SAVAnT) also provides a technique to assess the completeness of a software specification.
Hammond, Kevin L; Jurkus, Anthony F
The article explores marketing ethics considerations in the application of marketing to healthcare. While we realize that acceptance of healthcare marketing by all stakeholders is important for successful marketing, we emphasize its level of acceptance by healthcare professionals. The high levels of resistance to advertising and other forms of healthcare marketing by healthcare professionals has been largely based on the grounds that the practices are unethical. The nature of the resistance thus invites this exploration of healthcare marketing (and the marketing concept), marketing ethics, and the acceptance (rejection) by healthcare professionals of healthcare marketing.
This report contains information about the tasks I have completed and the valuable experience I have gained at NASA. The report is divided into two different sections followed by a program summary sheet. The first section describes the two reports I have completed for the Office of Mission Assurance (OMA). I describe the approach and the resources and facilities used to complete each report. The second section describes my experience working in the Receipt Inspection/Quality Assurance Lab (RI/QA). The first report described is a Product Assurance Plan for the Gas Permeable Polymer Materials (GPPM) mission. The purpose of the Product Assurance Plan is to define the various requirements which are to be met through completion of the GPPM mission. The GPPM experiment is a space payload which will be flown in the shuttle's SPACEHAB module. The experiment will use microgravity to enable production of complex polymeric gas permeable materials. The second report described in the first section is a Fracture Analysis for the Mir Environmental Effects Payload (MEEP). The Fracture Analysis report is a summary of the fracture control classifications for all structural elements of the MEEP. The MEEP hardware consists of four experiment carriers, each of which contains an experiment container holding a passive experiment. The MEEP hardware will be attached to the cargo bay of the space shuttle. It will be transferred by Extravehicular Activity and mounted on the Mir space station. The second section of this report describes my experiences in the RVQA lab. I listed the different equipment I used at the lab and their functions. I described the extensive inspection process that must be completed for spaceflight hardware. Included, at the end of this section, are pictures of most of the equipment used in the lab. There is a summary sheet located at the end of this report. It briefly describes the valuable experience I have gained at NASA this summer and what I will be able to take
This document may be of assistance in applying the New Source Review (NSR) air permitting regulations including the Prevention of Significant Deterioration (PSD) requirements. This document is part of the NSR Policy and Guidance Database. Some documents in the database are a scanned or retyped version of a paper photocopy of the original. Although we have taken considerable effort to quality assure the documents, some may contain typographical errors. Contact the office that issued the document if you need a copy of the original.
Campola, Michael J.
Radiation Hardness Assurance (RHA) consists of all activities undertaken to ensure that the electronics and materials of a space system perform to their design specifications after exposure to the mission space environment. The subset of interests for NEPP and the REAG, are EEE parts. It is important to register that all of these undertakings are in a feedback loop and require constant iteration and updating throughout the mission life. More detail can be found in the reference materials on applicable test data for usage on parts.
Tien, James M; Goldschmidt-Clermont, Pascal J
Engineering has and will continue to have a critical impact on healthcare; the application of technology-based techniques to biological problems can be defined to be technobiology applications. This paper is primarily focused on applying the technobiology approach of systems engineering to the development of a healthcare service system that is both integrated and adaptive. In general, healthcare services are carried out with knowledge-intensive agents or components which work together as providers and consumers to create or co-produce value. Indeed, the engineering design of a healthcare system must recognize the fact that it is actually a complex integration of human-centered activities that is increasingly dependent on information technology and knowledge. Like any service system, healthcare can be considered to be a combination or recombination of three essential components - people (characterized by behaviors, values, knowledge, etc.), processes (characterized by collaboration, customization, etc.) and products (characterized by software, hardware, infrastructures, etc.). Thus, a healthcare system is an integrated and adaptive set of people, processes and products. It is, in essence, a system of systems which objectives are to enhance its efficiency (leading to greater interdependency) and effectiveness (leading to improved health). Integration occurs over the physical, temporal, organizational and functional dimensions, while adaptation occurs over the monitoring, feedback, cybernetic and learning dimensions. In sum, such service systems as healthcare are indeed complex, especially due to the uncertainties associated with the human-centered aspects of these systems. Moreover, the system complexities can only be dealt with methods that enhance system integration and adaptation.
Allaert, Francois-André; Mazen, Noël-Jean; Legrand, Louis; Quantin, Catherine
The market for Connected Health Devices (CHD) with healthcare applications is growing fast and should be worth several billion euros in turnover in the coming years. Their development will completely transform the organisation of our healthcare system, profoundly change the way patients are managed and revolutionizes disease prevention. The CHD with healthcare applications is a tidal wave that has societal impact calling into question the privacy of patients' personal and healthcare information and its protection in secure systems. Rather than trying to stop the use of CHD, we must channel the wave by clearly examining the advantages versus the risks and threats to the patients, and find counter-measures for implementation. The main difficulty is channeling the wave in a way that is acceptable to CHD developers who otherwise will bypass the rules, even if they can be sued for it. Therefore, it appears necessary to implement guidelines that can be used by all developers, defining the minimum requirement for assuring the security of patient privacy and healthcare management. In European Healthcare Systems, there is an imperative need for establishing security guidelines that CHD producers could use to ensure compliance, so that patient privacy and healthcare management is safeguarded. The aim would be to implement the guidelines a posteriori rather than a priori control so as not to hamper innovation.
... Vaccination Resources for Healthcare Professionals Recommended Vaccines for Healthcare Workers Recommend on Facebook Tweet Share Compartir On ... for More Information Resources for Those Vaccinating HCWs Healthcare workers (HCWs) are at risk for exposure to ...
Wolf, Bernhard; Herzog, Karolin
Microelectronics and microsystem technology have changed our daily lives considerably in the past 50 years. Countless everyday objects contain microelectronic components. In healthcare up to the present, however, it has not been possible to make major alterations in introducing electronics and information technology that would lead to innovative improvements and greater transparency. This paper describes initial steps in diagnostics and oncological therapy including telematic healthcare systems which can, for example, assist patients with cardiovascular diseases and shows, through these areas, how electronics and microsystems technology can contribute to better healthcare.
Healthcare requires careful coordination of several occupations. In order to attain the best possible result, including effectiveness and cost-efficiency, the specific expertise of each of these occupations must be clearly defined. Healthcare occupations, physicians and nurses, are indeed professions as opposed to mere "jobs". They are concerned with living but ill human beings and not with things. Reliance on a personal capacity of judgment is a decisive aspect of professions. Healthcare professionals perform best if they are granted specific independence relative to their work.
Wood, Ben D
Turnover among healthcare professionals is a costly consequence. The existing body of knowledge on healthcare professional turnover is correlated with job satisfaction levels. A landmark study differentiated 2 areas of job satisfaction categories: satisfiers and dissatisfiers (intrinsic and extrinsic motivators). The aim of this article is to examine existing research on precursors of turnover, such as burnout behaviors experienced by healthcare professionals, job satisfaction levels, employee organizational commitment, health complications which precede turnover, some current strategies to reduce turnover, and some effects CEO turnover has on employee turnover intentions.
Discusses resources and strategies that libraries can use to evaluate competing database management software for purchase. Needs assessments, types of software available, features of good software, evaluation aids, shareware, and marketing and product trends are covered. (KRN)
Borman, Stuart A.
Discusses various aspects of scientific software, including evaluation and selection of commercial software products; program exchanges, catalogs, and other information sources; major data analysis packages; statistics and chemometrics software; and artificial intelligence. (JN)
Jones, E.D.; Banks, W.W.; Fischer, L.E.
This report describes and summarizes the results of a quality assurance (QA) study of the Gamma Knife, a nuclear medical device used for the gamma irradiation of intracranial lesions. Focus was on the physical aspects of QA and did not address issues that are essentially medical, such as patient selection or prescription of dose. A risk-based QA assessment approach was used. Sample programs for quality control and assurance are included. The use of the Gamma Knife was found to conform to existing standards and guidelines concerning radiation safety and quality control of external beam therapies (shielding, safety reviews, radiation surveys,more » interlock systems, exposure monitoring, good medical physics practices, etc.) and to be compliant with NRC teletherapy regulations. There are, however, current practices for the Gamma Knife not covered by existing, formalized regulations, standards, or guidelines. These practices have been adopted by Gamma Knife users and continue to be developed with further experience. Some of these have appeared in publications or presentations and are slowly finding their way into recommendations of professional organizations.« less
Allen, John I
This review concerns quality assurance for gastrointestinal endoscopic procedures, especially colonoscopy and will emphasize research and guidelines published since January 2011. Important articles from previous years have been included for background. Critical lapses in endoscope processing and administration of intravenous sedation alerted us to the infection risk of endoscopy. Increases in cost of colonoscopy, evidence for overuse and studies demonstrating missed cancers have led some to question the value of endoscopy. Despite these setbacks, the National Polyp Study (NPS) consortium published their long-term follow-up of the original NPS patients and confirmed that colonoscopy with polyp removal can reduce the risk of colorectal cancer for an extended period. In this article, we will focus on ways to improve the value of outpatient colonoscopy. The United States national quality improvement agenda recently became organized into a more coordinated effort spearheaded by several public and private entities. They comprise the infrastructure by which performance measures are developed and implemented as accountability standards. Understanding wherein a gastroenterology (GI) practice fits into this infrastructure and learning ways we can improve our endoscopic practice is important for physicians who provide this vital service to patients. This article will provide a roadmap for developing a quality assurance program for endoscopic practice.
Oginni, B. M.; Odero, D. O.
Recent developments in radiation therapy have immensely impacted the way the radiation dose is delivered to patients undergoing radiation treatments. However, the fundamental quality assurance (QA) issues underlying the radiation therapy still remain the accuracy of the radiation dose and the radiation safety. One of the major duties of clinical medical physicists in the radiation therapy departments still revolves around ensuring the accuracy of dose delivery to the planning target volume (PTV), the reduction of unintended radiation to normal organs and minimization of the radiation exposure to the medical personnel based on ALARA (as low as reasonably achievable) principle. Many of the errors in radiation therapy can be minimized through a comprehensive program of periodic checks. One of the QA procedures on the patient comes in the form of chart reviews which could be in either electronic or paper-based format. We present the quality assurance procedures that have to be performed on the patient records from the beginning and periodically to the end of the treatment, based on the guidelines from the American Association of Physicists in Medicine (AAPM) and American College of Physicians (ACP).
Lee, Sang Young
Healthcare environment is growing to include not only the traditional information systems, but also a business intelligence platform. For executive leaders, consultants, and analysts, there is no longer a need to spend hours in design and develop of typical reports or charts, the entire solution can be completed through using Business Intelligence software. The current paper highlights the advantages of big data analytics and business intelligence in the healthcare industry. In this paper, In this paper we focus our discussion around intelligent techniques and methodologies which are recently used for business intelligence in healthcare.
Trevino, Luis C.; Crumbley, Tim
The slide presentation examines the Marshall Space Flight Center Flight Software Branch, including software development projects, mission critical space flight software development, software technical insight, advanced software development technologies, and continuous improvement in the software development processes and methods.
This paper reports on the experience at the European Southern Observatory on the application of software engineering techniques to a 200 man-year control software project for the Very Large Telescope (VLT). This shall provide astronomers, before the end of the century, with one of the most powerful telescopes in the world. From the definition of the general model, described in the software management plan, specific activities have been and will be defined: standards for documents and for code development, design approach using a CASE tool, the process of reviewing both documentation and code, quality assurance, test strategy, etc. The initial choices, the current implementation and the future planned activities are presented and, where feedback is already available, pros and cons are discussed.
Trammell, Carmen J.; Poore, Jesse H.
There is growing interest in software 'certification', i.e., confirmation that software has performed satisfactorily under a defined certification protocol. Regulatory agencies, customers, and prospective reusers all want assurance that a defined product standard has been met. In other industries, products are typically certified under protocols in which random samples of the product are drawn, tests characteristic of operational use are applied, analytical or statistical inferences are made, and products meeting a standard are 'certified' as fit for use. A warranty statement is often issued upon satisfactory completion of a certification protocol. This paper outlines specific engineering practices that must be used to preserve the validity of the statistical certification testing protocol. The assumptions associated with a statistical experiment are given, and their implications for statistical testing of software are described.
Kelly, Michael P.
This viewgraph presentation reviews NASA Goddard Space Flight Center's approach to safety and mission assurance. The contents include: 1) NASA GSFC Background; 2) Safety and Mission Assurance Directorate; 3) The Role of SMA-D and the Technical Authority; 4) GSFC Mission assurance Requirements; 5) GSFC Systems Review Office (SRO); 6) GSFC Supply Chain Management Program; and 7) GSFC ISO9001/AS9100 Status Brief.
Product assurance technology topics addressed include: wafer acceptance procedures, test chips, test structures, test chip methodology, fault models, and the Combined Release and Radiation Effects Satellite test chip.
... assurances. (a) General policy. An institution with PHS supported biomedical or behavioral research, research... biomedical and behavioral research, research training, or activities related to that research or research...
... assurances. (a) General policy. An institution with PHS supported biomedical or behavioral research, research... biomedical and behavioral research, research training, or activities related to that research or research...
... assurances. (a) General policy. An institution with PHS supported biomedical or behavioral research, research... biomedical and behavioral research, research training, or activities related to that research or research...
fighters’ ability to execute the mission.” Computing Services 4 We run IT Systems that: provide medical care pay the warfighters manage maintenance...users • 1,400 applications • 18 facilities • 180 software vendors • 18,000+ copies of executive software products • Virtually every type of mainframe and... chocs electriques, de branchez les deux cordons d’al imentation avant de faire le depannage P R IM A R Y SD A S B 1 2 PowerHub 7000 RST U L 00- 00
Batalden, Maren; Batalden, Paul; Margolis, Peter; Seid, Michael; Armstrong, Gail; Opipari-Arrigan, Lisa; Hartung, Hans
Efforts to ensure effective participation of patients in healthcare are called by many names—patient centredness, patient engagement, patient experience. Improvement initiatives in this domain often resemble the efforts of manufacturers to engage consumers in designing and marketing products. Services, however, are fundamentally different than products; unlike goods, services are always ‘coproduced’. Failure to recognise this unique character of a service and its implications may limit our success in partnering with patients to improve health care. We trace a partial history of the coproduction concept, present a model of healthcare service coproduction and explore its application as a design principle in three healthcare service delivery innovations. We use the principle to examine the roles, relationships and aims of this interdependent work. We explore the principle's implications and challenges for health professional development, for service delivery system design and for understanding and measuring benefit in healthcare services. PMID:26376674
Austin, Rebekah; Label, Ken A.; Sampson, Mike J.; Evans, John; Witulski, Art; Sierawski, Brian; Karsai, Gabor; Mahadevan, Nag; Schrimpf, Ron; Reed, Robert A.
This presentation will cover why modeling is useful for radiation hardness assurance cases, and also provide information on Model-Based Assurance Case+ (MBAC+), NASAs Reliability Maintainability Template, and Fault Propagation Modeling.
Kushniruk, Andre; Senathirajah, Yalini; Borycki, Elizabeth
Persistent problems with healthcare IT that is unusable and unsafe have been reported worldwide. In this paper we present our vision for deploying usability engineering in healthcare in a more substantive way in order to improve the current situation. The argument will be made that stronger and more substantial efforts need to be made to bring multiple usability engineering methods to bear on points in both system design and deployment (and not just as a one-time effort restricted to software product development). In addition, improved processes for ensuring the usability of commercial vendor-based systems being implemented in healthcare organizations need to be addressed. A discussion will also be provided on challenges and barriers that will need to be overcome to ensure that the heatlhcare IT that is released is both usable and safe.
Heller, Günther; Szecsenyi, Joachim; Willms, Gerald; Broge, Björn
For several years, the use of administrative data in mandatory quality measurement has been requested by several stakeholders in Germany. Main advantages of using administrative data include the reduction of documentary expenditures and the possibility to perform longitudinal quality analyses across different healthcare units. After a short introduction, a brief overview of the current use of administrative data for mandatory quality assurance as well as current developments is given, which will then be further exemplified by decubital ulcer prophylaxis. By using administrative data coding expenditures in this clinical area could be reduced by nine million data fields. At the same time the population analysed was expanded resulting in a more than tenfold increase in potentially quality-relevant events. Finally, perspectives, further developments, possibilities as well as limits of quality measurement with administrative data are discussed. Copyright © 2014. Published by Elsevier GmbH.
implementation of the baselines, audits , and technical reviews within an overarching systems engineering process (SEP; Defense Acquisition University...warfighters’ needs. This management and metrics effort supplements and supports the system’s technical development through the baselines, audits and...other areas that could be researched and added into the nine-tier model. Areas including software metrics, quality assurance , software-oriented
Bown, Rodney L.
The following subject areas are covered: a DOD development system - the Army Secure Operating System; non-development commercial computer systems; security, integrity, and assurance of service (SI and A); post delivery SI and A and malicious software; computer system unique attributes; positive feedback to commercial computer systems vendors; and NDI (Non-Development Item) computers and software safety.
multiversion software subject to coincident errors. IEEE Trans. Software Eng. SE-11:1511-1517. Eckhardt, D.E., A.K Caglayan, J.C. Knight, L.D. Lee, D.F...J.C. and N.G. Leveson. 1986. Experimental evaluation of the assumption of independence in multiversion software. IEEE Trans. Software
Kruse, Ted, Comp.
Provides a list of evaluation services currently producing critical reviews of educational software. Includes information about The Apple K-12 Curriculum Software Reference, The Educational Software Preview, The Educational Software Selector, MicroSIFT, and Only The Best: The Discriminating Guide for Preschool-Grade 12. (TW)
Biju, Soly Mathew
Many software development firms are now adopting the agile software development method. This method involves the customer at every level of software development, thus reducing the impact of change in the requirement at a later stage. In this article, the principles of the agile method for software development are explored and there is a focus on…
and Engineering (CMSC 735) Codes: GPEV2 * Textiooks: IEEE Tutoria on Models and Metrics for Software Management and Engameeing by Basi, Victor R...Software Engineering (Comp 227) Codes: GPRY5 Textbooks: IEEE Tutoria on Software Design Techniques by Freeman, Peter and Wasserman, Anthony 1. Software
Presents the first year's results of a continuing project to monitor the availability of software of relevance for literacy education purposes. Concludes there is an enormous amount of software available for use by teachers of reading and literacy--whereas drill-and-practice software is the largest category of software available, large numbers of…
... assurance program information that duplicates language in quality assurance regulatory guides and quality... 10 Energy 2 2013-01-01 2013-01-01 false Quality assurance program change. 63.144 Section 63.144... REPOSITORY AT YUCCA MOUNTAIN, NEVADA Quality Assurance § 63.144 Quality assurance program change. Changes to...
... assurance program information that duplicates language in quality assurance regulatory guides and quality... 10 Energy 2 2014-01-01 2014-01-01 false Quality assurance program change. 63.144 Section 63.144... REPOSITORY AT YUCCA MOUNTAIN, NEVADA Quality Assurance § 63.144 Quality assurance program change. Changes to...
... assurance program information that duplicates language in quality assurance regulatory guides and quality... 10 Energy 2 2012-01-01 2012-01-01 false Quality assurance program change. 63.144 Section 63.144... REPOSITORY AT YUCCA MOUNTAIN, NEVADA Quality Assurance § 63.144 Quality assurance program change. Changes to...
Housch, Helen; Godfrey, Sally
This slide presentation reviews seven processes that NASA uses to ensure that software is developed, acquired and maintained as specified in the NPR 7150.2A requirement. The requirement is to ensure that all software be appraised for the Capability Maturity Model Integration (CMMI). The enumerated processes are: (7) Product Integration, (6) Configuration Management, (5) Verification, (4) Software Assurance, (3) Measurement and Analysis, (2) Requirements Management and (1) Planning & Monitoring. Each of these is described and the group(s) that are responsible is described.
... 42 Public Health 4 2011-10-01 2011-10-01 false State assurances. 441.302 Section 441.302 Public Health CENTERS FOR MEDICARE & MEDICAID SERVICES, DEPARTMENT OF HEALTH AND HUMAN SERVICES (CONTINUED... may terminate a waiver already granted: (a) Health and Welfare—Assurance that necessary safeguards...
... 42 Public Health 4 2014-10-01 2014-10-01 false State assurances. 441.302 Section 441.302 Public Health CENTERS FOR MEDICARE & MEDICAID SERVICES, DEPARTMENT OF HEALTH AND HUMAN SERVICES (CONTINUED... may terminate a waiver already granted: (a) Health and Welfare—Assurance that necessary safeguards...
... 42 Public Health 4 2010-10-01 2010-10-01 false State assurances. 441.302 Section 441.302 Public Health CENTERS FOR MEDICARE & MEDICAID SERVICES, DEPARTMENT OF HEALTH AND HUMAN SERVICES (CONTINUED... may terminate a waiver already granted: (a) Health and Welfare—Assurance that necessary safeguards...
...] 2013 Assuring Radiation Protection AGENCY: Food and Drug Administration, HHS. ACTION: Notice. SUMMARY... of the Center for Devices and Radiological Health (CDRH) radiation protection program. The goal of the 2013 Assuring Radiation Protection will be to coordinate Federal, State, and Tribal activities to...
... 49 Transportation 1 2010-10-01 2010-10-01 false Assurance required. 27.9 Section 27.9 Transportation Office of the Secretary of Transportation NONDISCRIMINATION ON THE BASIS OF DISABILITY IN PROGRAMS OR ACTIVITIES RECEIVING FEDERAL FINANCIAL ASSISTANCE General § 27.9 Assurance required. (a) General...
Materu, Peter; Righetti, Petra
This article assesses the status and practice of higher education quality assurance in sub-Saharan Africa, focusing on degree-granting tertiary institutions. A main finding is that structured national-level quality assurance processes in African higher education are a very recent phenomenon and that most countries face major capacity constraints.…
... 48 Federal Acquisition Regulations System 4 2011-10-01 2011-10-01 false Assurances. 370.402 Section 370.402 Federal Acquisition Regulations System HEALTH AND HUMAN SERVICES HHS SUPPLEMENTATIONS... 20892, as early as possible to secure the necessary assurances. (c) A contractor providing animal care...
... 48 Federal Acquisition Regulations System 4 2014-10-01 2014-10-01 false Assurances. 370.402 Section 370.402 Federal Acquisition Regulations System HEALTH AND HUMAN SERVICES HHS SUPPLEMENTATIONS... 20892, as early as possible to secure the necessary assurances. (c) A contractor providing animal care...
... 48 Federal Acquisition Regulations System 4 2013-10-01 2013-10-01 false Assurances. 370.402 Section 370.402 Federal Acquisition Regulations System HEALTH AND HUMAN SERVICES HHS SUPPLEMENTATIONS... 20892, as early as possible to secure the necessary assurances. (c) A contractor providing animal care...
... 48 Federal Acquisition Regulations System 4 2010-10-01 2010-10-01 false Assurances. 370.402 Section 370.402 Federal Acquisition Regulations System HEALTH AND HUMAN SERVICES HHS SUPPLEMENTATIONS... 20892, as early as possible to secure the necessary assurances. (c) A contractor providing animal care...
... 48 Federal Acquisition Regulations System 4 2012-10-01 2012-10-01 false Assurances. 370.402 Section 370.402 Federal Acquisition Regulations System HEALTH AND HUMAN SERVICES HHS SUPPLEMENTATIONS... 20892, as early as possible to secure the necessary assurances. (c) A contractor providing animal care...
This article examines an external evaluation of the quality assurance system at the Norwegian University of Science and Technology (NTNU) conducted by The Norwegian Agency for Quality Assurance in Higher Education (NOKUT). The external audit report along with internal reports provided by the seven faculties of the university served as the major…
Stephenson, Sandra L.
The word 'quality' has become a 'central term in the lexicon of contemporary higher education and a major point of interest to various interest groups'. As quality assurance systems have developed around the world, certain assumptions are becoming alarmingly widespread: for example that quality assurance is new to higher education, that quality…
... accompanied by, an assurance that the program will be conducted or the facility operated in compliance with... to provide or is in the form of personal property, or real property or interest therein or structures thereon, or improvement of real property or structures, the assurance shall obligate the recipient, or, in...
... accompanied by, an assurance that the program will be conducted or the facility operated in compliance with... to provide or is in the form of personal property, or real property or interest therein or structures thereon, or improvement of real property or structures, the assurance shall obligate the recipient, or, in...
... accompanied by, an assurance that the program will be conducted or the facility operated in compliance with... to provide or is in the form of personal property, or real property or interest therein or structures thereon, or improvement of real property or structures, the assurance shall obligate the recipient, or, in...
Quality assurance has been integrated into the fabric of higher education in China, with the issue of quality in higher education--how to evaluate it and how to enhance it--now taking centre stage in Chinese higher education. In the past decade, the development of quality assurance in Chinese higher education has covered a broad spectrum of…
... 18 Conservation of Power and Water Resources 2 2012-04-01 2012-04-01 false Assurances required. 1307.7 Section 1307.7 Conservation of Power and Water Resources TENNESSEE VALLEY AUTHORITY NONDISCRIMINATION WITH RESPECT TO HANDICAP § 1307.7 Assurances required. (a) TVA contributes financial assistance...
... 18 Conservation of Power and Water Resources 2 2013-04-01 2012-04-01 true Assurances required. 1307.7 Section 1307.7 Conservation of Power and Water Resources TENNESSEE VALLEY AUTHORITY NONDISCRIMINATION WITH RESPECT TO HANDICAP § 1307.7 Assurances required. (a) TVA contributes financial assistance...
... 18 Conservation of Power and Water Resources 2 2014-04-01 2014-04-01 false Assurances required. 1307.7 Section 1307.7 Conservation of Power and Water Resources TENNESSEE VALLEY AUTHORITY NONDISCRIMINATION WITH RESPECT TO HANDICAP § 1307.7 Assurances required. (a) TVA contributes financial assistance...
... 18 Conservation of Power and Water Resources 2 2011-04-01 2011-04-01 false Assurances required. 1307.7 Section 1307.7 Conservation of Power and Water Resources TENNESSEE VALLEY AUTHORITY NONDISCRIMINATION WITH RESPECT TO HANDICAP § 1307.7 Assurances required. (a) TVA contributes financial assistance...
... 18 Conservation of Power and Water Resources 2 2010-04-01 2010-04-01 false Assurances required. 1307.7 Section 1307.7 Conservation of Power and Water Resources TENNESSEE VALLEY AUTHORITY NONDISCRIMINATION WITH RESPECT TO HANDICAP § 1307.7 Assurances required. (a) TVA contributes financial assistance...
Stoten, David William
Purpose: This paper aims to focus on the changing nature of quality assurance systems within the sixth form college sector. Design/methodology/approach: Ten sixth form colleges were surveyed across England and staff from varying levels within college hierarchies questioned about how quality assurance systems were implemented. Research involved…
... RECEIVING FEDERAL FINANCIAL ASSISTANCE Introduction § 19.115 Assurance required. (a) General. Either at the... 44 Emergency Management and Assistance 1 2010-10-01 2010-10-01 false Assurance required. 19.115 Section 19.115 Emergency Management and Assistance FEDERAL EMERGENCY MANAGEMENT AGENCY, DEPARTMENT OF...
Sadler, D. Royce
Quality assurance processes have been applied to many aspects of higher education, including teaching, learning and assessment. At least in the latter domain, quality assurance needs its fundamental tenets critically scrutinised. A common but inadequate approach has been to identify and promote learning environment changes "likely to…
... 30 Mineral Resources 1 2010-07-01 2010-07-01 false Quality assurance. 14.8 Section 14.8 Mineral Resources MINE SAFETY AND HEALTH ADMINISTRATION, DEPARTMENT OF LABOR TESTING, EVALUATION, AND APPROVAL OF... Quality assurance. Applicants granted an approval or an extension of approval under this Part must: (a) In...
... 30 Mineral Resources 1 2010-07-01 2010-07-01 false Quality assurance. 7.7 Section 7.7 Mineral Resources MINE SAFETY AND HEALTH ADMINISTRATION, DEPARTMENT OF LABOR TESTING, EVALUATION, AND APPROVAL OF MINING PRODUCTS TESTING BY APPLICANT OR THIRD PARTY General § 7.7 Quality assurance. Applicants granted...
... 30 Mineral Resources 1 2010-07-01 2010-07-01 false Quality assurance. 15.8 Section 15.8 Mineral Resources MINE SAFETY AND HEALTH ADMINISTRATION, DEPARTMENT OF LABOR TESTING, EVALUATION, AND APPROVAL OF... § 15.8 Quality assurance. (a) Applicants granted an approval or an extension of approval under this...
... 7 Agriculture 6 2010-01-01 2010-01-01 false Quality assurance. 652.7 Section 652.7 Agriculture... AGRICULTURE SUPPORT ACTIVITIES TECHNICAL SERVICE PROVIDER ASSISTANCE General Provisions § 652.7 Quality assurance. (a) NRCS will review, in consultation with the Farm Service Agency, as appropriate, the quality...
... 48 Federal Acquisition Regulations System 6 2010-10-01 2010-10-01 true Quality Assurance. 2453.246 Section 2453.246 Federal Acquisition Regulations System DEPARTMENT OF HOUSING AND URBAN DEVELOPMENT CLAUSES AND FORMS FORMS Prescription of Forms 2453.246 Quality Assurance. ...
... Self-Directed Personal Assistance Services Program § 441.464 State assurances. A State must assure that... problems that might be associated with budget underutilization. (2) These safeguards may include the... that budget problems are identified on a timely basis so that corrective action may be taken, if...
Ontario's Quality Assurance Framework (QAF) is reviewed and found not to meet all five criteria proposed for a strong quality assurance system focused on student learning. The QAF requires a statement of student learning outcomes and a method and means of assessing those outcomes, but it does not require that data on achievement of intended…
... of the United States for the desegregation of such school or school system, and provides an assurance... other participants. Any such assurance shall include provisions which give the United States a right to... system which the responsible official of the Department of Health, Education, and Welfare determines is...
This Software Formal Inspections Standard (hereinafter referred to as Standard) is applicable to NASA software. This Standard defines the requirements that shall be fulfilled by the software formal inspections process whenever this process is specified for NASA software. The objective of this Standard is to define the requirements for a process that inspects software products to detect and eliminate defects as early as possible in the software life cycle. The process also provides for the collection and analysis of inspection data to improve the inspection process as well as the quality of the software.
Nikora, Allen; Antczak, Thomas M.; Lyu, Michael
Computer Aided Software Reliability Estimation (CASRE) computer program developed for use in measuring reliability of other software. Easier for non-specialists in reliability to use than many other currently available programs developed for same purpose. CASRE incorporates mathematical modeling capabilities of public-domain Statistical Modeling and Estimation of Reliability Functions for Software (SMERFS) computer program and runs in Windows software environment. Provides menu-driven command interface; enabling and disabling of menu options guides user through (1) selection of set of failure data, (2) execution of mathematical model, and (3) analysis of results from model. Written in C language.
Gilliam, D. P.; Powell, J. D.
This paper presents a portion of an overall research project on the generation of the network security assessment instrument to aid developers in assessing and assuring the security of software in the development and maintenance lifecycles.
Addy, Edward A.
Verification and validation (V&V) is used to increase the level of assurance of critical software, particularly that of safety-critical and mission critical software. This paper describes the working group's success in identifying V&V tasks that could be performed in the domain engineering and transition levels of reuse-based software engineering. The primary motivation for V&V at the domain level is to provide assurance that the domain requirements are correct and that the domain artifacts correctly implement the domain requirements. A secondary motivation is the possible elimination of redundant V&V activities at the application level. The group also considered the criteria and motivation for performing V&V in domain engineering.
Having robust policies and procedures in place for infection control is fundamentally important. However, each organization has to go a step beyond this; evidence has to be provided that these policies and procedures are followed. As of 1 April 2009, with the introduction of the Care Quality Commission and The Health and Social Care Act 2008 Code of Practice for the NHS on the Prevention and Control of Healthcare-Associated Infections and Related Guidance, the assurance of robust infection control measures within any UK provider of health care became an even higher priority. Also, the commissioning of any service by the NHS must provide evidence that the provider has in place robust procedures for infection control. This article demonstrates how the clinical audit team at the Douglas Macmillan Hospice in North Staffordshire, UK, have used audit to assure high rates of compliance with the current national standards for infection control. Prior to the audit, hospice staff had assumed that the rates of compliance for infection control approached 100%. This article shows that a good quality audit tool can be used to identify areas of shortfall in infection control and the effectiveness of putting in place an action plan followed by re-audit.
Clough, A. J.
The paper describes a software inspection process that can be used to evaluate the quality of software components. Quality criteria, process application, independent testing of the process and proposed associated tool support are covered. Early results indicate that this technique is well suited for assessing software component quality in a standardized fashion. With automated machine assistance to facilitate both the evaluation and selection of software components, such a technique should promote effective reuse of software components.
Siau, Keng; Shen, Zixing
Advances in wireless technology give pace to the rapid development of mobile applications. The coming mobile revolution will bring dramatic and fundamental changes to our daily life. It will influence the way we live, the way we do things, and the way we take care of our health. For the healthcare industry, mobile applications provide a new frontier in offering better care and services to patients, and a more flexible and mobile way of communicating with suppliers and patients. Mobile applications will provide important real time data for patients, physicians, insurers, and suppliers. In addition, it will revolutionalize the way information is managed in the healthcare industry and redefine the doctor - patient communication. This paper discusses different aspects of mobile healthcare. Specifically, it presents mobile applications in healthcare, and discusses possible challenges facing the development of mobile applications. Obstacles in developing mobile healthcare applications include mobile device limitations, wireless networking problems, infrastructure constraints, security concerns, and user distrust. Research issues in resolving or alleviating these problems are also discussed in the paper.
Guidotti, Tee L; Ragain, Lisa
Studies of risk communication have identified healthcare providers, especially physicians, as the source of information most trusted by the public on issues of environmental health. Nothing in medical, nursing or most healthcare provider training actually prepares practitioners to play this role and healthcare providers are generally more oriented toward treatment and medical care than prevention and public health. Healthcare providers require education in order to play this role but rarely seek it. Gaps in the knowledge of professional on the issue of Cryptosporidium illustrate the problem. For members of the professional water community, communicating with healthcare providers is best done when messages are delivered in familiar settings, such as hospital Grand Rounds (a universal format for teaching conferences) and provided in a narrative (case-based) form but gaining access is difficult if the topic is not obviously clinical in nature. In addition to being a critically important target group itself, public health professionals are easier to reach and may mediate good working relationships with medical practitioners. We suggest a strategy for water utilities based on partnerships with academic public health and providing education through well-recognized formats in continuing medical and nursing education.
Fitz, Rhonda; Whitman, Gerek
Research into complexities of software systems Fault Management (FM) and how architectural design decisions affect safety, preservation of assets, and maintenance of desired system functionality has coalesced into a technical reference (TR) suite that advances the provision of safety and mission assurance. The NASA Independent Verification and Validation (IVV) Program, with Software Assurance Research Program support, extracted FM architectures across the IVV portfolio to evaluate robustness, assess visibility for validation and test, and define software assurance methods applied to the architectures and designs. This investigation spanned IVV projects with seven different primary developers, a wide range of sizes and complexities, and encompassed Deep Space Robotic, Human Spaceflight, and Earth Orbiter mission FM architectures. The initiative continues with an expansion of the TR suite to include Launch Vehicles, adding the benefit of investigating differences intrinsic to model-based FM architectures and insight into complexities of FM within an Agile software development environment, in order to improve awareness of how nontraditional processes affect FM architectural design and system health management.
Occupational health and safety research rarely makes use of data on employee healthcare utilization to gain insight into the physical and mental health of healthcare staff. This paper aims to fill this gap by examining the prevalence of two relevant types of healthcare utilization among staff working in healthcare organizations: physical therapy and mental healthcare utilization. The paper furthermore explores what role employee and organizational characteristics play in explaining differences in healthcare utilization between organizations. A Dutch healthcare insurance company provided healthcare utilization records for a sample of 417 organizations employing 136,804 healthcare workers in the Netherlands. The results showed that there are large differences between and within healthcare industries when it comes to employee healthcare utilization. Multivariate regression analyses revealed that employee characteristics such as age and gender distributions, and healthcare industry, explain some of the variance between healthcare organizations. Nevertheless, the results of the analyses showed that for all healthcare utilization indicators there is still a large amount of unexplained variance. Further research into the subject of organizational differences in employee healthcare utilization is needed, as finding possibilities to influence employee health and subsequent healthcare utilization is beneficial to employees, employers and society as a whole.
Hayhurst, Kelly J. (Editor)
The Guidance and Control Software (GCS) project was the last in a series of software reliability studies conducted at Langley Research Center between 1977 and 1994. The technical results of the GCS project were recorded after the experiment was completed. Some of the support documentation produced as part of the experiment, however, is serving an unexpected role far beyond its original project context. Some of the software used as part of the GCS project was developed to conform to the RTCA/DO-178B software standard, "Software Considerations in Airborne Systems and Equipment Certification," used in the civil aviation industry. That standard requires extensive documentation throughout the software development life cycle, including plans, software requirements, design and source code, verification cases and results, and configuration management and quality control data. The project documentation that includes this information is open for public scrutiny without the legal or safety implications associated with comparable data from an avionics manufacturer. This public availability has afforded an opportunity to use the GCS project documents for DO-178B training. This report provides a brief overview of the GCS project, describes the 4-volume set of documents and the role they are playing in training, and includes the planning documents from the GCS project. Volume 1 contains five appendices: A. Plan for Software Aspects of Certification for the Guidance and Control Software Project; B. Software Development Standards for the Guidance and Control Software Project; C. Software Verification Plan for the Guidance and Control Software Project; D. Software Configuration Management Plan for the Guidance and Control Software Project; and E. Software Quality Assurance Activities.
As NASA embarks upon developing the Next-Generation Extra Vehicular Activity (EVA) Radio for deep space exploration, the demands on EVA battery life will substantially increase. The number of modes and frequency bands required will continue to grow in order to enable efficient and complex multi-mode operations including communications, navigation, and tracking applications. Whether conducting astronaut excursions, communicating to soldiers, or first responders responding to emergency hazards, NASA has developed an innovative, affordable, miniaturized, power-efficient software defined radio that offers unprecedented power-efficient flexibility. This lightweight, programmable, S-band, multi-service, frequency- agile EVA software defined radio (SDR) supports data, telemetry, voice, and both standard and high-definition video. Features include a modular design, an easily scalable architecture, and the EVA SDR allows for both stationary and mobile battery powered handheld operations. Currently, the radio is equipped with an S-band RF section. However, its scalable architecture can accommodate multiple RF sections simultaneously to cover multiple frequency bands. The EVA SDR also supports multiple network protocols. It currently implements a Hybrid Mesh Network based on the 802.11s open standard protocol. The radio targets RF channel data rates up to 20 Mbps and can be equipped with a real-time operating system (RTOS) that can be switched off for power-aware applications. The EVA SDR's modular design permits implementation of the same hardware at all Network Nodes concept. This approach assures the portability of the same software into any radio in the system. It also brings several benefits to the entire system including reducing system maintenance, system complexity, and development cost.
Palvannan, R Kannapiran; Teow, Kiok Liang
Patient queues are prevalent in healthcare and wait time is one measure of access to care. We illustrate Queueing Theory-an analytical tool that has provided many insights to service providers when designing new service systems and managing existing ones. This established theory helps us to quantify the appropriate service capacity to meet the patient demand, balancing system utilization and the patient's wait time. It considers four key factors that affect the patient's wait time: average patient demand, average service rate and the variation in both. We illustrate four basic insights that will be useful for managers and doctors who manage healthcare delivery systems, at hospital or department level. Two examples from local hospitals are shown where we have used queueing models to estimate the service capacity and analyze the impact of capacity configurations, while considering the inherent variation in healthcare.
This paper offers a short commentary on the editorial by Mannion and Exworthy. The paper highlights the positive insights offered by their analysis into the tensions between the competing institutional logics of standardization and customization in healthcare, in part manifested in the conflict between managers and professionals, and endorses the plea of the authors for further research in this field. However, the editorial is criticized for its lack of a strong societal reference point, the comparative absence of focus on hybridization, and its failure to highlight structural factors impinging on the opposing logics in a broader neo-institutional framework. With reference to the Procrustean metaphor, it is argued that greater stress should be placed on the healthcare user in future health policy. Finally, the case of complementary and alternative medicine is set out which – while not explicitly mentioned in the editorial – most effectively concretizes the tensions at the heart of this analysis of healthcare. PMID:29626406
The purpose of this document is to provide experience-based guidance in implementing a software process improvement program in any NASA software development or maintenance community. This guidebook details how to define, operate, and implement a working software process improvement program. It describes the concept of the software process improvement program and its basic organizational components. It then describes the structure, organization, and operation of the software process improvement program, illustrating all these concepts with specific NASA examples. The information presented in the document is derived from the experiences of several NASA software organizations, including the SEL, the SEAL, and the SORCE. Their experiences reflect many of the elements of software process improvement within NASA. This guidebook presents lessons learned in a form usable by anyone considering establishing a software process improvement program within his or her own environment. This guidebook attempts to balance general and detailed information. It provides material general enough to be usable by NASA organizations whose characteristics do not directly match those of the sources of the information and models presented herein. It also keeps the ideas sufficiently close to the sources of the practical experiences that have generated the models and information.
This software implements software-defined radio procession over multicore, multi-CPU systems in a way that maximizes the use of CPU resources in the system. The software treats each processing step in either a communications or navigation modulator or demodulator system as an independent, threaded block. Each threaded block is defined with a programmable number of input or output buffers; these buffers are implemented using POSIX pipes. In addition, each threaded block is assigned a unique thread upon block installation. A modulator or demodulator system is built by assembly of the threaded blocks into a flow graph, which assembles the processing blocks to accomplish the desired signal processing. This software architecture allows the software to scale effortlessly between single CPU/single-core computers or multi-CPU/multi-core computers without recompilation. NASA spaceflight and ground communications systems currently rely exclusively on ASICs or FPGAs. This software allows low- and medium-bandwidth (100 bps to approx.50 Mbps) software defined radios to be designed and implemented solely in C/C++ software, while lowering development costs and facilitating reuse and extensibility.
This software implements software-defined radio procession over multi-core, multi-CPU systems in a way that maximizes the use of CPU resources in the system. The software treats each processing step in either a communications or navigation modulator or demodulator system as an independent, threaded block. Each threaded block is defined with a programmable number of input or output buffers; these buffers are implemented using POSIX pipes. In addition, each threaded block is assigned a unique thread upon block installation. A modulator or demodulator system is built by assembly of the threaded blocks into a flow graph, which assembles the processing blocks to accomplish the desired signal processing. This software architecture allows the software to scale effortlessly between single CPU/single-core computers or multi-CPU/multi-core computers without recompilation. NASA spaceflight and ground communications systems currently rely exclusively on ASICs or FPGAs. This software allows low- and medium-bandwidth (100 bps to .50 Mbps) software defined radios to be designed and implemented solely in C/C++ software, while lowering development costs and facilitating reuse and extensibility.
Programmatic requirements for the advancement of software technology are identified for meeting the space flight requirements in the 1980 to 1990 time period. The development items are described, and software technology item derivation worksheets are presented along with the cost/time/priority assessments.
He, Lulu; Shull, Forrest
Software Quality Assurance (SQA) is an important component of the software development process. SQA processes provide assurance that the software products and processes in the project life cycle conform to their specified requirements by planning, enacting, and performing a set of activities to provide adequate confidence that quality is being built into the software. Typical techniques include: (1) Testing (2) Simulation (3) Model checking (4) Symbolic execution (5) Management reviews (6) Technical reviews (7) Inspections (8) Walk-throughs (9) Audits (10) Analysis (complexity analysis, control flow analysis, algorithmic analysis) (11) Formal method Our work over the last few years has resulted in substantial knowledge about SQA techniques, especially the areas of technical reviews and inspections. But can we apply the same QA techniques to the system development process? If yes, what kind of tailoring do we need before applying them in the system engineering context? If not, what types of QA techniques are actually used at system level? And, is there any room for improvement.) After a brief examination of the system engineering literature (especially focused on NASA and DoD guidance) we found that: (1) System and software development process interact with each other at different phases through development life cycle (2) Reviews are emphasized in both system and software development. (Figl.3). For some reviews (e.g. SRR, PDR, CDR), there are both system versions and software versions. (3) Analysis techniques are emphasized (e.g. Fault Tree Analysis, Preliminary Hazard Analysis) and some details are given about how to apply them. (4) Reviews are expected to use the outputs of the analysis techniques. In other words, these particular analyses are usually conducted in preparation for (before) reviews. The goal of our work is to explore the interaction between the Quality Assurance (QA) techniques at the system level and the software level.
Sittig, Dean F.; Singh, Hardeep
Conceptual models have been developed to address challenges inherent in studying health information technology (HIT). This manuscript introduces an 8-dimensional model specifically designed to address the socio-technical challenges involved in design, development, implementation, use, and evaluation of HIT within complex adaptive healthcare systems. The 8 dimensions are not independent, sequential, or hierarchical, but rather are interdependent and interrelated concepts similar to compositions of other complex adaptive systems. Hardware and software computing infrastructure refers to equipment and software used to power, support, and operate clinical applications and devices. Clinical content refers to textual or numeric data and images that constitute the “language” of clinical applications. The human computer interface includes all aspects of the computer that users can see, touch, or hear as they interact with it. People refers to everyone who interacts in some way with the system, from developer to end-user, including potential patient-users. Workflow and communication are the processes or steps involved in assuring that patient care tasks are carried out effectively. Two additional dimensions of the model are internal organizational features (e.g., policies, procedures, and culture) and external rules and regulations, both of which may facilitate or constrain many aspects of the preceding dimensions. The final dimension is measurement and monitoring, which refers to the process of measuring and evaluating both intended and unintended consequences of HIT implementation and use. We illustrate how our model has been successfully applied in real-world complex adaptive settings to understand and improve HIT applications at various stages of development and implementation. PMID:20959322
Label, Kenneth A.; Sampson, Michael J.
This presentation will cover NASA Electrical, Electronic and Electromechanical (EEE) Parts Assurance Structure, NASA Electronic Parts and Packaging (NEPP) Program, NASA Electronic Parts Assurance Group (NEPAG), examples of assurance challenges, and future challenges.
... NUCLEAR REGULATORY COMMISSION [NRC-2013-0021] Quality Assurance Program Requirements (Operations...), DG-1300, ``Quality Assurance Program Requirements (Operations).'' DATES: Submit comments by April 1... CFR Part 50, Appendix B, ``Quality Assurance Criteria for Nuclear power Plants and Fuel Reprocessing...
Voigt, S. (Editor)
Four panels of invited experts and NASA representatives focused on the following topics: software management, software development environment, languages, and software standards. Each panel deliberated in private, held two open sessions with audience participation, and developed recommendations for the NASA Space Station Program. The major thrusts of the recommendations were as follows: (1) The software management plan should establish policies, responsibilities, and decision points for software acquisition; (2) NASA should furnish a uniform modular software support environment and require its use for all space station software acquired (or developed); (3) The language Ada should be selected for space station software, and NASA should begin to address issues related to the effective use of Ada; and (4) The space station software standards should be selected (based upon existing standards where possible), and an organization should be identified to promulgate and enforce them. These and related recommendations are described in detail in the conference proceedings.
Governments are increasingly concerned to compare the quality and effectiveness of healthcare interventions but find this a complex matter. Crude hospital statistics can be dangerously misleading and need adjusting for case mix, but identifying and weighting the patient characteristics which affect prognosis are problematical for conceptual, methodological, and practical reasons. These include the inherently uncertain nature of prognosis itself and the practical difficulties of collecting and quantifying data on the outcomes of interest for specific healthcare interventions and known risk factors such as severity. Images p1494-a PMID:8019285
In this article, we make the etiologic diagnosis for a sick patient named Healthcare: the cancer of greed. When we explore the two forms of this cancer--corporate and bureaucratic--we find the latter is the greater danger to We the Patients. The "treatments" applied to patient Healthcare by the Congressional "doctors" have consistently made the patient worse, not better. At the core of healthcare's woes is the government's diversion of money from healthcare services to healthcare bureaucracy. As this is the root cause, it is what we must address in order to cure, not sedate or palliate, patient Healthcare.
Thakur, Vikas; Ramesh, A
The importance of healthcare waste management in preserving the environment and protecting the public cannot be denied. Past research has dealt with various issues in healthcare waste management and disposal, which spreads over various journals, pipeline research disciplines and research communities. Hence, this article analyses this scattered knowledge in a systematic manner, considering the period between January 2005 and July 2014. The purpose of this study is to: (i) identify the trends in healthcare waste management literature regarding journals published; (ii) main topics of research in healthcare waste management; (iii) methodologies used in healthcare waste management research; (iv) areas most frequently researched by researchers; and (v) determine the scope of future research in healthcare waste management. To this end, the authors conducted a systematic review of 176 articles on healthcare waste management taken from the following eight esteemed journals: International Journal of Environmental Health Research, International Journal of Healthcare Quality Assurance, Journal of Environmental Management, Journal of Hazardous Material, Journal of Material Cycles and Waste Management, Resources, Conservations and Recycling, Waste Management, and Waste Management & Research. The authors have applied both quantitative and qualitative approaches for analysis, and results will be useful in the following ways: (i) results will show importance of healthcare waste management in healthcare operations; (ii) findings will give a comparative view of the various publications; (c) study will shed light on future research areas. © The Author(s) 2015.
Villamor Ordozgoiti, Alberto; Delgado Hito, Pilar; Guix Comellas, Eva María; Fernandez Sanchez, Carlos Manuel; Garcia Hernandez, Milagros; Lluch Canut, Teresa
Information and Communications Technologies in healthcare has increased the need to consider quality criteria through standardised processes. The aim of this study was to analyse the software quality evaluation models applicable to healthcare from the perspective of ICT-purchasers. Through a systematic literature review with the keywords software, product, quality, evaluation and health, we selected and analysed 20 original research papers published from 2005-2016 in health science and technology databases. The results showed four main topics: non-ISO models, software quality evaluation models based on ISO/IEC standards, studies analysing software quality evaluation models, and studies analysing ISO standards for software quality evaluation. The models provide cost-efficiency criteria for specific software, and improve use outcomes. The ISO/IEC25000 standard is shown as the most suitable for evaluating the quality of ICTs for healthcare use from the perspective of institutional acquisition.
Raab, Stephen S
Anatomic pathology laboratories use several quality assurance tools to detect errors and to improve patient safety. To review some of the anatomic pathology laboratory patient safety quality assurance practices. Different standards and measures in anatomic pathology quality assurance and patient safety were reviewed. Frequency of anatomic pathology laboratory error, variability in the use of specific quality assurance practices, and use of data for error reduction initiatives. Anatomic pathology error frequencies vary according to the detection method used. Based on secondary review, a College of American Pathologists Q-Probes study showed that the mean laboratory error frequency was 6.7%. A College of American Pathologists Q-Tracks study measuring frozen section discrepancy found that laboratories improved the longer they monitored and shared data. There is a lack of standardization across laboratories even for governmentally mandated quality assurance practices, such as cytologic-histologic correlation. The National Institutes of Health funded a consortium of laboratories to benchmark laboratory error frequencies, perform root cause analysis, and design error reduction initiatives, using quality assurance data. Based on the cytologic-histologic correlation process, these laboratories found an aggregate nongynecologic error frequency of 10.8%. Based on gynecologic error data, the laboratory at my institution used Toyota production system processes to lower gynecologic error frequencies and to improve Papanicolaou test metrics. Laboratory quality assurance practices have been used to track error rates, and laboratories are starting to use these data for error reduction initiatives.
Koles, G.; Hitchcock, R.; Sherman, M.
This report documents part of the work performed in phase I of a Laboratory Directors Research and Development (LDRD) funded project entitled Building Performance Assurances (BPA). The focus of the BPA effort is to transform the way buildings are built and operated in order to improve building performance by facilitating or providing tools, infrastructure, and information. The efforts described herein focus on the development of metrics with which to evaluate building performance and for which information and optimization tools need to be developed. The classes of building performance metrics reviewed are (1) Building Services (2) First Costs, (3) Operating Costs,more » (4) Maintenance Costs, and (5) Energy and Environmental Factors. The first category defines the direct benefits associated with buildings; the next three are different kinds of costs associated with providing those benefits; the last category includes concerns that are broader than direct costs and benefits to the building owner and building occupants. The level of detail of the various issues reflect the current state of knowledge in those scientific areas and the ability of the to determine that state of knowledge, rather than directly reflecting the importance of these issues; it intentionally does not specifically focus on energy issues. The report describes work in progress and is intended as a resource and can be used to indicate the areas needing more investigation. Other reports on BPA activities are also available.« less
A demonstration of field portable/mobile technologies for measuring trace elements in soil and sediments was conducted under the U.S. Environmental Protection Agency Superfund Innovative Technology Evaluation (SITE) Program. The demonstration took place from January 24 to 28, 2005, at the Kennedy Athletic, Recreational and Social Park at Kennedy Space Center on Merritt Island, Florida. The purpose of the demonstration was to verify the performance of various instruments that employ X-ray fluorescence (XRF) measurement technologies for the determination of 13 toxic elements in a variety of soil and sediment samples. Instruments from the technology developers listed below were demonstrated. o Innov-X Systems, Inc.o NITON LLC (2 instruments ) o Oxford Instruments Portable Division (formerly Metorex, Inc.) .Oxford Instruments Analytical .Rigaku, Inc.o RONTEC USA Inc.o Xcalibur XRF Services Inc. (Division of Elvatech Ltd. ) This demonstration plan describes the procedures that will be used to verify the performance and cost of the XRF instruments provided by these technology developers. The plan incorporates the quality assurance and quality control elements needed to generate data of sufficient quality to perform this verification. A separate innovative technology verification report (ITVR) will be prepared for each instrument. The objective of this program is to promote the acceptance and use of innovative field technologies by providing well-documented perfor
The Demonstration of innovative field devices for the measurement of mercury in soil and sediment is being conducted under the EPA's SITE Program in February 2003 at the United States Department of Energy's (DOE) Oak Ridge National Laboratory (ORNL) in Oak Ridge, Tennessee and the Tennessee Department of Environment and Conservation's Department of Energy Oversight facility in Oak Ridge, Tennessee. The primary purpose of the Demonstration is to evaluate innovative field devices for the measurement of mercury in soil and sediment based on their performance and cost as compared to a conventional, off-site laboratory analytical method. The five field measurement devices listed below will be demonstrated: .Metorex's X-M ET 2000 Metal Master Analyzer, X-Ray Fluorescence Analyzer .Milestone Inc.'s Direct Mercury Analyzer (DMA-80), Thermal Decomposition Instrument.NITON's XL-700 Series Multi-Element Analyzer, X-Ray Fluorescence Analyzer .Ohio Lumex's RA-915+ Portable Mercury Analyzer, Atomic Absorption Spectrometer, Thermal Decompostion Attachment RP 91C .MTI, Inc.'s PDV 5000 Hand Held Instrument, Anodic Stripping Voltamm eter<1). This Demonstration Plan describes the procedures that will be used to verify the performance and cost of each field measurement device. The plan incorporates the quality assurance and quality control elements needed to generate data of sufficient quality to document each device's performance and cost. A separate Innovative Technology Verifica
The demonstration of technologies for determining the presence of dioxin in soil and sediment is being conducted under the U.S. Environmental Protection Agency Superfund Innovative Technology Evaluation Program in Saginaw, Michigan, at Green Point Environmental Learning Center from approximately April 26 to May 6, 2004. The primary purpose of the demonstration is to evaluate innovative monitoring technologies. The technologies listed below will be demonstrated. .AhRC PCRTM Kit, Hybrizyme Corporation .Ah-IMMUNOASSY@ Kit, Paralsian, Inc. .Coplanar PCB Immunoassay Kit, Abraxis LLC .DF-l Dioxin/Furan Immunoassay Kit, CAPE Technologies L.L.C. .CALUX@ by Xenobiotic Detection Systems, Inc- .Dioxin ELISA Kit, Wako Pure Chemical Industries LTD. This demonstration plan describes the procedures that will be used to verify the performance and cost of these technologies. The plan incorporates the quality assurance and quality control elements needed to generate data of sufficient quality to document each technology's performance and cost. A separate innovative technology verification report (ITVR) will.be prepared for each technology. The ITVRs will present the demonstration findings associated with the demonstration objectives. The objective of this program is to promote the acceptance and use of innovative field technologies by providing well-documented performance and cost data obtained from field demonstrations.
Li, H Harold; Wu, Yu; Yang, Deshan; Mutic, Sasa
Physics chart check has long been a central quality assurance (QC) measure in radiation oncology. The purpose of this work is to describe a software tool that aims to accomplish simplification, standardization, automation, and forced functions in the process. Nationally recognized guidelines, including American College of Radiology and American Society for Radiation Oncology guidelines and technical standards, and the American Association of Physicists in Medicine Task Group reports were identified, studied, and summarized. Meanwhile, the reported events related to physics chart check service were analyzed using an event reporting and learning system. A number of shortfalls in the chart check process were identified. To address these problems, a software tool was designed and developed under Microsoft. Net in C# to hardwire as many components as possible at each stage of the process. The software consists of the following 4 independent modules: (1) chart check management; (2) pretreatment and during treatment chart check assistant; (3) posttreatment chart check assistant; and (4) quarterly peer-review management. The users were a large group of physicists in the author's radiation oncology clinic. During over 1 year of use the tool has proven very helpful in chart checking management, communication, documentation, and maintaining consistency. The software tool presented in this work aims to assist physicists at each stage of the physics chart check process. The software tool is potentially useful for any radiation oncology clinics that are either in the process of pursuing or maintaining the American College of Radiology accreditation.
Dorsey, Cheryl A.; Dorsey, Timothy A.
DO-178, Software Considerations in Airborne Systems and Equipment Certification, is the well-known international standard dealing with the assurance of software used in airborne systems [1,2]. Insights into the DO-178 experiences, strengths and weaknesses can benefit the international space community. As DO-178 is an excellent standard for safe software development when used appropriately, this paper provides lessons learned and suggestions for using it effectively.
Goodloe, Alwyn E.
Safety-critical systems are growing more complex and becoming increasingly autonomous. Runtime Verification (RV) has the potential to provide protections when a system cannot be assured by conventional means, but only if the RV itself can be trusted. In this paper, we proffer a number of challenges to realizing high-assurance RV and illustrate how we have addressed them in our research. We argue that high-assurance RV provides a rich target for automated verification tools in hope of fostering closer collaboration among the communities.
Hoover, Marcey L.; Kolb, Rachel R.
In high-consequence engineering organizations, such as Sandia, quality assurance may be heavily dependent on staff competency. Competency-dependent quality assurance models are at risk when the environment changes, as it has with increasing attrition rates, budget and schedule cuts, and competing program priorities. Risks in Sandia's competency-dependent culture can be mitigated through changes to hiring, training, and customer engagement approaches to manage people, partners, and products. Sandia's technical quality engineering organization has been able to mitigate corporate-level risks by driving changes that benefit all departments, and in doing so has assured Sandia's commitment to excellence in high-consequence engineering and national service.
Dalthorp, Daniel; Huso, Manuela; Dail, David
Evidence of Absence software (EoA) is a user-friendly software application for estimating bird and bat fatalities at wind farms and for designing search protocols. The software is particularly useful in addressing whether the number of fatalities is below a given threshold and what search parameters are needed to give assurance that thresholds were not exceeded. The software also includes tools (1) for estimating carcass persistence distributions and searcher efficiency parameters ( and ) from field trials, (2) for projecting future mortality based on past monitoring data, and (3) for exploring the potential consequences of various choices in the design of long-term incidental take permits for protected species. The software was designed specifically for cases where tolerance for mortality is low and carcass counts are small or even 0, but the tools also may be used for mortality estimates when carcass counts are large.
Nanda, Upali; Pati, Debajyoti; McCurry, Katie
While there is a growing consciousness about the importance of visually pleasing environments in healthcare design, little is known about the key underlying mechanisms that enable aesthetics to play an instrumental role in the caregiving process. Hence it is often one of the first items to be value engineered. Aesthetics has (rightfully) been provided preferential consideration in such pleasure settings such as museums and recreational facilities; but in healthcare settings it is often considered expendable. Should it be? In this paper the authors share evidence that visual stimuli undergo an aesthetic evaluation process in the human brain by default, even when not prompted; that responses to visual stimuli may be immediate and emotional; and that aesthetics can be a source of pleasure, a fundamental perceptual reward that can help mitigate the stress of a healthcare environment. The authors also provide examples of studies that address the role of specific visual elements and visual principles in aesthetic evaluations and emotional responses. Finally, they discuss the implications of these findings for the design of art and architecture in healthcare.
India is undergoing a rapid transformation in terms of governance, administrative reforms, newer policy develoment, and social movements. India is also considered one of the most vibrant economies in the world. The current discourse in public space is dominated by issues such as economic development, security, corruption free governance, gender equity, and women safety. Healthcare though remains a pressing need of population; seems to have taken a backseat. In the era of decreasing subsidies and cautious investment in social sectors, the 2nd National Conference on Family Medicine and Primary Care 2015 (FMPC) brought a focus on “healthcare” in India. The theme of this conference was “Healthcare is Primary.” The conference participants discussed on the theme of why healthcare should be a national priority and why strong primary care should remain at the center of healthcare delivery system. The experts recommended that India needs to strengthen the “general health system” instead of focusing on disease based vertical programs. Public health system should have capacity and skill pool to be able to deliver person centered comprehensive health services to the community. Proactive implementation of policies towards human resource in health is the need of the hour. As the draft National Health Policy 2015 is being debated, “family medicine” (academic primary care), the unfinished agenda of National Health Policy 2002, remains a priority area of implementation. PMID:26985402
Cocksedge, Simon; Barr, Nicky; Deakin, Corinne
In UK health policy ‘sharing good information is pivotal to improving care quality, safety, and effectiveness. Nevertheless, educators often neglect this vital communication skill. The consequences of brief communication education interventions for healthcare workers are not yet established. This study investigated a three-hour interprofessional experiential workshop (group work, theoretical input, rehearsal) training healthcare staff in sharing information using a clear structure (PARSLEY). Staff in one UK hospital participated. Questionnaires were completed before, immediately after, and eight weeks after training, with semistructured interviews seven weeks after training. Participants (n=76) were from assorted healthcare occupations (26% non-clinical). Knowledge significantly increased immediately after training. Self-efficacy, outcome expectancy, and motivation to use the structure taught were significantly increased immediately following training and at eight weeks. Respondents at eight weeks (n=35) reported their practice in sharing information had changed within seven days of training. Seven weeks after training, most interviewees (n=13) reported confidently using the PARSLEY structure regularly in varied settings. All had re-evaluated their communication practice. Brief training altered self-reported communication behaviour of healthcare staff, with sustained changes in everyday work. As sharing information is central to communication curricula, health policy, and shared decision-making, the effectiveness of brief teaching interventions has economic and educational implications.
Graber, D R; Johnson, J A
In recent years, the place of spirituality in organizations has become increasingly discussed and advocated. On a personal level, this may involve achieving personal fulfillment or spiritual growth in the workplace. In the broader sense, spirituality is considered by many to be essential in an organization's interactions with employees, customers, and the community. This article describes a possible role for greater spirituality in healthcare organizations, whose cultures in recent decades have largely excluded spirituality or religiousness. This is the consequence of an analytical, scientific perspective on human health; a reductionist paradigm in biomedical research; and the inevitable bureaucratization occurring in large healthcare organizations. However, in recent decades, numerous scientific articles supporting a connection between faith or religiousness and positive health outcomes have been published. Because individuals seek meaning when experiencing severe illnesses, and humans universally respond to compassion and caring, spirituality among healthcare workers and managers appears highly appropriate. The article describes organizational barriers to the greater inclusion of spirituality in healthcare and presents several approaches to developing a more caring organization. These include eliciting extensive input from all staff and clinicians in identifying core or common values, ethics, and a philosophy of caring. Programs should ensure that the views of nonreligious staff and patients are respected and that clear guidelines are established for the extent and nature of affective or spiritual support for patients.
Batalden, Maren; Batalden, Paul; Margolis, Peter; Seid, Michael; Armstrong, Gail; Opipari-Arrigan, Lisa; Hartung, Hans
Efforts to ensure effective participation of patients in healthcare are called by many names-patient centredness, patient engagement, patient experience. Improvement initiatives in this domain often resemble the efforts of manufacturers to engage consumers in designing and marketing products. Services, however, are fundamentally different than products; unlike goods, services are always 'coproduced'. Failure to recognise this unique character of a service and its implications may limit our success in partnering with patients to improve health care. We trace a partial history of the coproduction concept, present a model of healthcare service coproduction and explore its application as a design principle in three healthcare service delivery innovations. We use the principle to examine the roles, relationships and aims of this interdependent work. We explore the principle's implications and challenges for health professional development, for service delivery system design and for understanding and measuring benefit in healthcare services. Published by the BMJ Publishing Group Limited. For permission to use (where not already granted under a licence) please go to http://www.bmj.com/company/products-services/rights-and-licensing/
Helm, Virginia M.
Issues related to the illegal copying or piracy of educational software in the schools and its potential effect on quality software availability are discussed. Copyright violation is examined as a reason some software producers may be abandoning the school software market. An explanation of what the copyright allows and prohibits in terms of…