Diagnosis - Using automatic test equipment and artificial intelligence expert systems
NASA Astrophysics Data System (ADS)
Ramsey, J. E., Jr.
Three expert systems (ATEOPS, ATEFEXPERS, and ATEFATLAS), which were created to direct automatic test equipment (ATE), are reviewed. The purpose of the project was to develop an expert system to troubleshoot the converter-programmer power supply card for the F-15 aircraft and have that expert system direct the automatic test equipment. Each expert system uses a different knowledge base or inference engine, basing the testing on the circuit schematic, test requirements document, or ATLAS code. Implementing generalized modules allows the expert systems to be used for any different unit under test. Using converted ATLAS to LISP code allows the expert system to direct any ATE using ATLAS. The constraint propagated frame system allows for the expansion of control by creating the ATLAS code, checking the code for good software engineering techniques, directing the ATE, and changing the test sequence as needed (planning).
Signal Processing Expert Code (SPEC)
DOE Office of Scientific and Technical Information (OSTI.GOV)
Ames, H.S.
1985-12-01
The purpose of this paper is to describe a prototype expert system called SPEC which was developed to demonstrate the utility of providing an intelligent interface for users of SIG, a general purpose signal processing code. The expert system is written in NIL, runs on a VAX 11/750 and consists of a backward chaining inference engine and an English-like parser. The inference engine uses knowledge encoded as rules about the formats of SIG commands and about how to perform frequency analyses using SIG. The system demonstrated that expert system can be used to control existing codes.
An Embedded Rule-Based Diagnostic Expert System in Ada
NASA Technical Reports Server (NTRS)
Jones, Robert E.; Liberman, Eugene M.
1992-01-01
Ada is becoming an increasingly popular programming language for large Government-funded software projects. Ada with it portability, transportability, and maintainability lends itself well to today's complex programming environment. In addition, expert systems have also assumed a growing role in providing human-like reasoning capability expertise for computer systems. The integration is discussed of expert system technology with Ada programming language, especially a rule-based expert system using an ART-Ada (Automated Reasoning Tool for Ada) system shell. NASA Lewis was chosen as a beta test site for ART-Ada. The test was conducted by implementing the existing Autonomous Power EXpert System (APEX), a Lisp-based power expert system, in ART-Ada. Three components, the rule-based expert systems, a graphics user interface, and communications software make up SMART-Ada (Systems fault Management with ART-Ada). The rules were written in the ART-Ada development environment and converted to Ada source code. The graphics interface was developed with the Transportable Application Environment (TAE) Plus, which generates Ada source code to control graphics images. SMART-Ada communicates with a remote host to obtain either simulated or real data. The Ada source code generated with ART-Ada, TAE Plus, and communications code was incorporated into an Ada expert system that reads the data from a power distribution test bed, applies the rule to determine a fault, if one exists, and graphically displays it on the screen. The main objective, to conduct a beta test on the ART-Ada rule-based expert system shell, was achieved. The system is operational. New Ada tools will assist in future successful projects. ART-Ada is one such tool and is a viable alternative to the straight Ada code when an application requires a rule-based or knowledge-based approach.
Expert system for maintenance management of a boiling water reactor power plant
DOE Office of Scientific and Technical Information (OSTI.GOV)
Hong Shen; Liou, L.W.; Levine, S.
1992-01-01
An expert system code has been developed for the maintenance of two boiling water reactor units in Berwick, Pennsylvania, that are operated by the Pennsylvania Power and Light Company (PP and L). The objective of this expert system code, where the knowledge of experienced operators and engineers is captured and implemented, is to support the decisions regarding which components can be safely and reliably removed from service for maintenance. It can also serve as a query-answering facility for checking the plant system status and for training purposes. The operating and maintenance information of a large number of support systems, whichmore » must be available for emergencies and/or in the event of an accident, is stored in the data base of the code. It identifies the relevant technical specifications and management rules for shutting down any one of the systems or removing a component from service to support maintenance. Because of the complexity and time needed to incorporate a large number of systems and their components, the first phase of the expert system develops a prototype code, which includes only the reactor core isolation coolant system, the high-pressure core injection system, the instrument air system, the service water system, and the plant electrical system. The next phase is scheduled to expand the code to include all other systems. This paper summarizes the prototype code and the design concept of the complete expert system code for maintenance management of all plant systems and components.« less
Probabilistic structural analysis methods for select space propulsion system components
NASA Technical Reports Server (NTRS)
Millwater, H. R.; Cruse, T. A.
1989-01-01
The Probabilistic Structural Analysis Methods (PSAM) project developed at the Southwest Research Institute integrates state-of-the-art structural analysis techniques with probability theory for the design and analysis of complex large-scale engineering structures. An advanced efficient software system (NESSUS) capable of performing complex probabilistic analysis has been developed. NESSUS contains a number of software components to perform probabilistic analysis of structures. These components include: an expert system, a probabilistic finite element code, a probabilistic boundary element code and a fast probability integrator. The NESSUS software system is shown. An expert system is included to capture and utilize PSAM knowledge and experience. NESSUS/EXPERT is an interactive menu-driven expert system that provides information to assist in the use of the probabilistic finite element code NESSUS/FEM and the fast probability integrator (FPI). The expert system menu structure is summarized. The NESSUS system contains a state-of-the-art nonlinear probabilistic finite element code, NESSUS/FEM, to determine the structural response and sensitivities. A broad range of analysis capabilities and an extensive element library is present.
Automated encoding of clinical documents based on natural language processing.
Friedman, Carol; Shagina, Lyudmila; Lussier, Yves; Hripcsak, George
2004-01-01
The aim of this study was to develop a method based on natural language processing (NLP) that automatically maps an entire clinical document to codes with modifiers and to quantitatively evaluate the method. An existing NLP system, MedLEE, was adapted to automatically generate codes. The method involves matching of structured output generated by MedLEE consisting of findings and modifiers to obtain the most specific code. Recall and precision applied to Unified Medical Language System (UMLS) coding were evaluated in two separate studies. Recall was measured using a test set of 150 randomly selected sentences, which were processed using MedLEE. Results were compared with a reference standard determined manually by seven experts. Precision was measured using a second test set of 150 randomly selected sentences from which UMLS codes were automatically generated by the method and then validated by experts. Recall of the system for UMLS coding of all terms was .77 (95% CI.72-.81), and for coding terms that had corresponding UMLS codes recall was .83 (.79-.87). Recall of the system for extracting all terms was .84 (.81-.88). Recall of the experts ranged from .69 to .91 for extracting terms. The precision of the system was .89 (.87-.91), and precision of the experts ranged from .61 to .91. Extraction of relevant clinical information and UMLS coding were accomplished using a method based on NLP. The method appeared to be comparable to or better than six experts. The advantage of the method is that it maps text to codes along with other related information, rendering the coded output suitable for effective retrieval.
Translating expert system rules into Ada code with validation and verification
NASA Technical Reports Server (NTRS)
Becker, Lee; Duckworth, R. James; Green, Peter; Michalson, Bill; Gosselin, Dave; Nainani, Krishan; Pease, Adam
1991-01-01
The purpose of this ongoing research and development program is to develop software tools which enable the rapid development, upgrading, and maintenance of embedded real-time artificial intelligence systems. The goals of this phase of the research were to investigate the feasibility of developing software tools which automatically translate expert system rules into Ada code and develop methods for performing validation and verification testing of the resultant expert system. A prototype system was demonstrated which automatically translated rules from an Air Force expert system was demonstrated which detected errors in the execution of the resultant system. The method and prototype tools for converting AI representations into Ada code by converting the rules into Ada code modules and then linking them with an Activation Framework based run-time environment to form an executable load module are discussed. This method is based upon the use of Evidence Flow Graphs which are a data flow representation for intelligent systems. The development of prototype test generation and evaluation software which was used to test the resultant code is discussed. This testing was performed automatically using Monte-Carlo techniques based upon a constraint based description of the required performance for the system.
An Expert System for the Development of Efficient Parallel Code
NASA Technical Reports Server (NTRS)
Jost, Gabriele; Chun, Robert; Jin, Hao-Qiang; Labarta, Jesus; Gimenez, Judit
2004-01-01
We have built the prototype of an expert system to assist the user in the development of efficient parallel code. The system was integrated into the parallel programming environment that is currently being developed at NASA Ames. The expert system interfaces to tools for automatic parallelization and performance analysis. It uses static program structure information and performance data in order to automatically determine causes of poor performance and to make suggestions for improvements. In this paper we give an overview of our programming environment, describe the prototype implementation of our expert system, and demonstrate its usefulness with several case studies.
NESSUS/EXPERT - An expert system for probabilistic structural analysis methods
NASA Technical Reports Server (NTRS)
Millwater, H.; Palmer, K.; Fink, P.
1988-01-01
An expert system (NESSUS/EXPERT) is presented which provides assistance in using probabilistic structural analysis methods. NESSUS/EXPERT is an interactive menu-driven expert system that provides information to assist in the use of the probabilistic finite element code NESSUS/FEM and the fast probability integrator. NESSUS/EXPERT was developed with a combination of FORTRAN and CLIPS, a C language expert system tool, to exploit the strengths of each language.
Expert systems built by the Expert: An evaluation of OPS5
NASA Technical Reports Server (NTRS)
Jackson, Robert
1987-01-01
Two expert systems were written in OPS5 by the expert, a Ph.D. astronomer with no prior experience in artificial intelligence or expert systems, without the use of a knowledge engineer. The first system was built from scratch and uses 146 rules to check for duplication of scientific information within a pool of prospective observations. The second system was grafted onto another expert system and uses 149 additional rules to estimate the spacecraft and ground resources consumed by a set of prospective observations. The small vocabulary, the IF this occurs THEN do that logical structure of OPS5, and the ability to follow program execution allowed the expert to design and implement these systems with only the data structures and rules of another OPS5 system as an example. The modularity of the rules in OPS5 allowed the second system to modify the rulebase of the system onto which it was grafted without changing the code or the operation of that system. These experiences show that experts are able to develop their own expert systems due to the ease of programming and code reusability in OPS5.
NASA Technical Reports Server (NTRS)
Muratore, John F.
1991-01-01
Lessons learned from operational real time expert systems are examined. The basic system architecture is discussed. An expert system is any software that performs tasks to a standard that would normally require a human expert. An expert system implies knowledge contained in data rather than code. And an expert system implies the use of heuristics as well as algorithms. The 15 top lessons learned by the operation of a real time data system are presented.
System and method for creating expert systems
NASA Technical Reports Server (NTRS)
Hughes, Peter M. (Inventor); Luczak, Edward C. (Inventor)
1998-01-01
A system and method provides for the creation of a highly graphical expert system without the need for programming in code. An expert system is created by initially building a data interface, defining appropriate Mission, User-Defined, Inferred, and externally-generated GenSAA (EGG) data variables whose data values will be updated and input into the expert system. Next, rules of the expert system are created by building appropriate conditions of the rules which must be satisfied and then by building appropriate actions of rules which are to be executed upon corresponding conditions being satisfied. Finally, an appropriate user interface is built which can be highly graphical in nature and which can include appropriate message display and/or modification of display characteristics of a graphical display object, to visually alert a user of the expert system of varying data values, upon conditions of a created rule being satisfied. The data interface building, rule building, and user interface building are done in an efficient manner and can be created without the need for programming in code.
MacRae, Jayden; Love, Tom; Baker, Michael G; Dowell, Anthony; Carnachan, Matthew; Stubbe, Maria; McBain, Lynn
2015-10-06
We designed and validated a rule-based expert system to identify influenza like illness (ILI) from routinely recorded general practice clinical narrative to aid a larger retrospective research study into the impact of the 2009 influenza pandemic in New Zealand. Rules were assessed using pattern matching heuristics on routine clinical narrative. The system was trained using data from 623 clinical encounters and validated using a clinical expert as a gold standard against a mutually exclusive set of 901 records. We calculated a 98.2 % specificity and 90.2 % sensitivity across an ILI incidence of 12.4 % measured against clinical expert classification. Peak problem list identification of ILI by clinical coding in any month was 9.2 % of all detected ILI presentations. Our system addressed an unusual problem domain for clinical narrative classification; using notational, unstructured, clinician entered information in a community care setting. It performed well compared with other approaches and domains. It has potential applications in real-time surveillance of disease, and in assisted problem list coding for clinicians. Our system identified ILI presentation with sufficient accuracy for use at a population level in the wider research study. The peak coding of 9.2 % illustrated the need for automated coding of unstructured narrative in our study.
Best Practices for Reduction of Uncertainty in CFD Results
NASA Technical Reports Server (NTRS)
Mendenhall, Michael R.; Childs, Robert E.; Morrison, Joseph H.
2003-01-01
This paper describes a proposed best-practices system that will present expert knowledge in the use of CFD. The best-practices system will include specific guidelines to assist the user in problem definition, input preparation, grid generation, code selection, parameter specification, and results interpretation. The goal of the system is to assist all CFD users in obtaining high quality CFD solutions with reduced uncertainty and at lower cost for a wide range of flow problems. The best-practices system will be implemented as a software product which includes an expert system made up of knowledge databases of expert information with specific guidelines for individual codes and algorithms. The process of acquiring expert knowledge is discussed, and help from the CFD community is solicited. Benefits and challenges associated with this project are examined.
NASA Technical Reports Server (NTRS)
Allen, Cheryl L.
1991-01-01
Enhanced engineering tools can be obtained through the integration of expert system methodologies and existing design software. The application of these methodologies to the spacecraft design and cost model (SDCM) software provides an improved technique for the selection of hardware for unmanned spacecraft subsystem design. The knowledge engineering system (KES) expert system development tool was used to implement a smarter equipment section algorithm than that which is currently achievable through the use of a standard data base system. The guidance, navigation, and control subsystems of the SDCM software was chosen as the initial subsystem for implementation. The portions of the SDCM code which compute the selection criteria and constraints remain intact, and the expert system equipment selection algorithm is embedded within this existing code. The architecture of this new methodology is described and its implementation is reported. The project background and a brief overview of the expert system is described, and once the details of the design are characterized, an example of its implementation is demonstrated.
Developments in REDES: The rocket engine design expert system
NASA Technical Reports Server (NTRS)
Davidian, Kenneth O.
1990-01-01
The Rocket Engine Design Expert System (REDES) is being developed at the NASA-Lewis to collect, automate, and perpetuate the existing expertise of performing a comprehensive rocket engine analysis and design. Currently, REDES uses the rigorous JANNAF methodology to analyze the performance of the thrust chamber and perform computational studies of liquid rocket engine problems. The following computer codes were included in REDES: a gas properties program named GASP, a nozzle design program named RAO, a regenerative cooling channel performance evaluation code named RTE, and the JANNAF standard liquid rocket engine performance prediction code TDK (including performance evaluation modules ODE, ODK, TDE, TDK, and BLM). Computational analyses are being conducted by REDES to provide solutions to liquid rocket engine thrust chamber problems. REDES is built in the Knowledge Engineering Environment (KEE) expert system shell and runs on a Sun 4/110 computer.
Developments in REDES: The Rocket Engine Design Expert System
NASA Technical Reports Server (NTRS)
Davidian, Kenneth O.
1990-01-01
The Rocket Engine Design Expert System (REDES) was developed at NASA-Lewis to collect, automate, and perpetuate the existing expertise of performing a comprehensive rocket engine analysis and design. Currently, REDES uses the rigorous JANNAF methodology to analyze the performance of the thrust chamber and perform computational studies of liquid rocket engine problems. The following computer codes were included in REDES: a gas properties program named GASP; a nozzle design program named RAO; a regenerative cooling channel performance evaluation code named RTE; and the JANNAF standard liquid rocket engine performance prediction code TDK (including performance evaluation modules ODE, ODK, TDE, TDK, and BLM). Computational analyses are being conducted by REDES to provide solutions to liquid rocket engine thrust chamber problems. REDES was built in the Knowledge Engineering Environment (KEE) expert system shell and runs on a Sun 4/110 computer.
ART-Ada design project, phase 2
NASA Technical Reports Server (NTRS)
Lee, S. Daniel; Allen, Bradley P.
1990-01-01
Interest in deploying expert systems in Ada has increased. An Ada based expert system tool is described called ART-Ada, which was built to support research into the language and methodological issues of expert systems in Ada. ART-Ada allows applications of an existing expert system tool called ART-IM (Automated Reasoning Tool for Information Management) to be deployed in various Ada environments. ART-IM, a C-based expert system tool, is used to generate Ada source code which is compiled and linked with an Ada based inference engine to produce an Ada executable image. ART-Ada is being used to implement several expert systems for NASA's Space Station Freedom Program and the U.S. Air Force.
A Rocket Engine Design Expert System
NASA Technical Reports Server (NTRS)
Davidian, Kenneth J.
1989-01-01
The overall structure and capabilities of an expert system designed to evaluate rocket engine performance are described. The expert system incorporates a JANNAF standard reference computer code to determine rocket engine performance and a state of the art finite element computer code to calculate the interactions between propellant injection, energy release in the combustion chamber, and regenerative cooling heat transfer. Rule-of-thumb heuristics were incorporated for the H2-O2 coaxial injector design, including a minimum gap size constraint on the total number of injector elements. One dimensional equilibrium chemistry was used in the energy release analysis of the combustion chamber. A 3-D conduction and/or 1-D advection analysis is used to predict heat transfer and coolant channel wall temperature distributions, in addition to coolant temperature and pressure drop. Inputting values to describe the geometry and state properties of the entire system is done directly from the computer keyboard. Graphical display of all output results from the computer code analyses is facilitated by menu selection of up to five dependent variables per plot.
A rocket engine design expert system
NASA Technical Reports Server (NTRS)
Davidian, Kenneth J.
1989-01-01
The overall structure and capabilities of an expert system designed to evaluate rocket engine performance are described. The expert system incorporates a JANNAF standard reference computer code to determine rocket engine performance and a state-of-the-art finite element computer code to calculate the interactions between propellant injection, energy release in the combustion chamber, and regenerative cooling heat transfer. Rule-of-thumb heuristics were incorporated for the hydrogen-oxygen coaxial injector design, including a minimum gap size constraint on the total number of injector elements. One-dimensional equilibrium chemistry was employed in the energy release analysis of the combustion chamber and three-dimensional finite-difference analysis of the regenerative cooling channels was used to calculate the pressure drop along the channels and the coolant temperature as it exits the coolant circuit. Inputting values to describe the geometry and state properties of the entire system is done directly from the computer keyboard. Graphical display of all output results from the computer code analyses is facilitated by menu selection of up to five dependent variables per plot.
Development of a Spacecraft Materials Selector Expert System
NASA Technical Reports Server (NTRS)
Pippin, G.; Kauffman, W. (Technical Monitor)
2002-01-01
This report contains a description of the knowledge base tool and examples of its use. A downloadable version of the Spacecraft Materials Selector (SMS) knowledge base is available through the NASA Space Environments and Effects Program. The "Spacecraft Materials Selector" knowledge base is part of an electronic expert system. The expert system consists of an inference engine that contains the "decision-making" code and the knowledge base that contains the selected body of information. The inference engine is a software package previously developed at Boeing, called the Boeing Expert System Tool (BEST) kit.
NERISK: AN EXPERT SYSTEM TO ENHANCE THE INTEGRATION OF PESTICIDES WITH ARTHROPOD BIOLOGICAL CONTROL
An expert system termed NERISK was developed to evaluate the effects of pesticides on arthropod predators and parasitoids in a variety of agroecosystems. ased on a shell system (RECOG) with minor coding modifications, the system was designed to let even a novice user access the v...
An expert system based software sizing tool, phase 2
NASA Technical Reports Server (NTRS)
Friedlander, David
1990-01-01
A software tool was developed for predicting the size of a future computer program at an early stage in its development. The system is intended to enable a user who is not expert in Software Engineering to estimate software size in lines of source code with an accuracy similar to that of an expert, based on the program's functional specifications. The project was planned as a knowledge based system with a field prototype as the goal of Phase 2 and a commercial system planned for Phase 3. The researchers used techniques from Artificial Intelligence and knowledge from human experts and existing software from NASA's COSMIC database. They devised a classification scheme for the software specifications, and a small set of generic software components that represent complexity and apply to large classes of programs. The specifications are converted to generic components by a set of rules and the generic components are input to a nonlinear sizing function which makes the final prediction. The system developed for this project predicted code sizes from the database with a bias factor of 1.06 and a fluctuation factor of 1.77, an accuracy similar to that of human experts but without their significant optimistic bias.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Rosenthal, Andrew
The DOE grant, “An Integrated Partnership to Create and Lead the Solar Codes and Standards Working Group,” to New Mexico State University created the Solar America Board for Codes and Standards (Solar ABCs). From 2007 – 2013 with funding from this grant, Solar ABCs identified current issues, established a dialogue among key stakeholders, and catalyzed appropriate activities to support the development of codes and standards that facilitated the installation of high quality, safe photovoltaic systems. Solar ABCs brought the following resources to the PV stakeholder community; Formal coordination in the planning or revision of interrelated codes and standards removing “stovemore » pipes” that have only roofing experts working on roofing codes, PV experts on PV codes, fire enforcement experts working on fire codes, etc.; A conduit through which all interested stakeholders were able to see the steps being taken in the development or modification of codes and standards and participate directly in the processes; A central clearing house for new documents, standards, proposed standards, analytical studies, and recommendations of best practices available to the PV community; A forum of experts that invites and welcomes all interested parties into the process of performing studies, evaluating results, and building consensus on standards and code-related topics that affect all aspects of the market; and A biennial gap analysis to formally survey the PV community to identify needs that are unmet and inhibiting the market and necessary technical developments.« less
CLIPS: An expert system building tool
NASA Technical Reports Server (NTRS)
Riley, Gary
1991-01-01
The C Language Integrated Production System (CLIPS) is an expert system building tool, which provides a complete environment for the development and delivery of rule and/or object based expert systems. CLIPS was specifically designed to provide a low cost option for developing and deploying expert system applications across a wide range of hardware platforms. The commercial potential of CLIPS is vast. Currently, CLIPS is being used by over 3,300 individuals throughout the public and private sector. Because the CLIPS source code is readily available, numerous groups have used CLIPS as a basis for their own expert system tools. To date, three commercially available tools have been derived from CLIPS. In general, the development of CLIPS has helped to improve the ability to deliver expert system technology throughout the public and private sectors for a wide range of applications and diverse computing environments.
Uniform emergency codes: will they improve safety?
2005-01-01
There are pros and cons to uniform code systems, according to emergency medicine experts. Uniformity can be a benefit when ED nurses and other staff work at several facilities. It's critical that your staff understand not only what the codes stand for, but what they must do when codes are called. If your state institutes a new system, be sure to hold regular drills to familiarize your ED staff.
End-to-End Modeling with the Heimdall Code to Scope High-Power Microwave Systems
2007-06-01
END-TO-END MODELING WITH THE HEIMDALL CODE TO SCOPE HIGH - POWER MICROWAVE SYSTEMS ∗ John A. Swegleξ Savannah River National Laboratory, 743A...describe the expert-system code HEIMDALL, which is used to model full high - power microwave systems using over 60 systems-engineering models, developed in...of our calculations of the mass of a Supersystem producing 500-MW, 15-ns output pulses in the X band for bursts of 1 s , interspersed with 10- s
An Expert Assistant for Computer Aided Parallelization
NASA Technical Reports Server (NTRS)
Jost, Gabriele; Chun, Robert; Jin, Haoqiang; Labarta, Jesus; Gimenez, Judit
2004-01-01
The prototype implementation of an expert system was developed to assist the user in the computer aided parallelization process. The system interfaces to tools for automatic parallelization and performance analysis. By fusing static program structure information and dynamic performance analysis data the expert system can help the user to filter, correlate, and interpret the data gathered by the existing tools. Sections of the code that show poor performance and require further attention are rapidly identified and suggestions for improvements are presented to the user. In this paper we describe the components of the expert system and discuss its interface to the existing tools. We present a case study to demonstrate the successful use in full scale scientific applications.
Techniques and implementation of the embedded rule-based expert system using Ada
NASA Technical Reports Server (NTRS)
Liberman, Eugene M.; Jones, Robert E.
1991-01-01
Ada is becoming an increasingly popular programming language for large Government-funded software projects. Ada with its portability, transportability, and maintainability lends itself well to today's complex programming environment. In addition, expert systems have also assured a growing role in providing human-like reasoning capability and expertise for computer systems. The integration of expert system technology with Ada programming language, specifically a rule-based expert system using an ART-Ada (Automated Reasoning Tool for Ada) system shell is discussed. The NASA Lewis Research Center was chosen as a beta test site for ART-Ada. The test was conducted by implementing the existing Autonomous Power EXpert System (APEX), a Lisp-base power expert system, in ART-Ada. Three components, the rule-based expert system, a graphics user interface, and communications software make up SMART-Ada (Systems fault Management with ART-Ada). The main objective, to conduct a beta test on the ART-Ada rule-based expert system shell, was achieved. The system is operational. New Ada tools will assist in future successful projects. ART-Ada is one such tool and is a viable alternative to the straight Ada code when an application requires a rule-based or knowledge-based approach.
Three CLIPS-based expert systems for solving engineering problems
NASA Technical Reports Server (NTRS)
Parkinson, W. J.; Luger, G. F.; Bretz, R. E.
1990-01-01
We have written three expert systems, using the CLIPS PC-based expert system shell. These three expert systems are rule based and are relatively small, with the largest containing slightly less than 200 rules. The first expert system is an expert assistant that was written to help users of the ASPEN computer code choose the proper thermodynamic package to use with their particular vapor-liquid equilibrium problem. The second expert system was designed to help petroleum engineers choose the proper enhanced oil recovery method to be used with a given reservoir. The effectiveness of each technique is highly dependent upon the reservoir conditions. The third expert system is a combination consultant and control system. This system was designed specifically for silicon carbide whisker growth. Silicon carbide whiskers are an extremely strong product used to make ceramic and metal composites. The manufacture of whiskers is a very complicated process. which to date. has defied a good mathematical model. The process was run by experts who had gained their expertise by trial and error. A system of rules was devised by these experts both for procedure setup and for the process control. In this paper we discuss the three problem areas of the design, development and evaluation of the CLIPS-based programs.
Expert systems for space power supply - Design, analysis, and evaluation
NASA Technical Reports Server (NTRS)
Cooper, Ralph S.; Thomson, M. Kemer; Hoshor, Alan
1987-01-01
The feasibility of applying expert systems to the conceptual design, analysis, and evaluation of space power supplies in particular, and complex systems in general is evaluated. To do this, the space power supply design process and its associated knowledge base were analyzed and characterized in a form suitable for computer emulation of a human expert. The existing expert system tools and the results achieved with them were evaluated to assess their applicability to power system design. Some new concepts for combining program architectures (modular expert systems and algorithms) with information about the domain were applied to create a 'deep' system for handling the complex design problem. NOVICE, a code to solve a simplified version of a scoping study of a wide variety of power supply types for a broad range of missions, has been developed, programmed, and tested as a concrete feasibility demonstration.
NASA Astrophysics Data System (ADS)
Friesdorf, Florian; Pangercic, Dejan; Bubb, Heiner; Beetz, Michael
In mac, an ergonomic dialog-system and algorithms will be developed that enable human experts and companions to be integrated into knowledge gathering and decision making processes of highly complex cognitive systems (e.g. Assistive Household as manifested further in the paper). In this event we propose to join algorithms and methodologies coming from Ergonomics and Artificial Intelligence that: a) make cognitive systems more congenial for non-expert humans, b) facilitate their comprehension by utilizing a high-level expandable control code for human experts and c) augment representation of such cognitive system into “deep representation” obtained through an interaction with human companions.
CLIPS/Ada: An Ada-based tool for building expert systems
NASA Technical Reports Server (NTRS)
White, W. A.
1990-01-01
Clips/Ada is a production system language and a development environment. It is functionally equivalent to the CLIPS tool. CLIPS/Ada was developed in order to provide a means of incorporating expert system technology into projects where the use of the Ada language had been mandated. A secondary purpose was to glean information about the Ada language and its compilers. Specifically, whether or not the language and compilers were mature enough to support AI applications. The CLIPS/Ada tool is coded entirely in Ada and is designed to be used by Ada systems that require expert reasoning.
Development of an expert based ICD-9-CM and ICD-10-CM map to AIS 2005 update 2008.
Loftis, Kathryn L; Price, Janet P; Gillich, Patrick J; Cookman, Kathy J; Brammer, Amy L; St Germain, Trish; Barnes, Jo; Graymire, Vickie; Nayduch, Donna A; Read-Allsopp, Christine; Baus, Katherine; Stanley, Patsye A; Brennan, Maureen
2016-09-01
This article describes how maps were developed from the clinical modifications of the 9th and 10th revisions of the International Classification of Diseases (ICD) to the Abbreviated Injury Scale 2005 Update 2008 (AIS08). The development of the mapping methodology is described, with discussion of the major assumptions used in the process to map ICD codes to AIS severities. There were many intricacies to developing the maps, because the 2 coding systems, ICD and AIS, were developed for different purposes and contain unique classification structures to meet these purposes. Experts in ICD and AIS analyzed the rules and coding guidelines of both injury coding schemes to develop rules for mapping ICD injury codes to the AIS08. This involved subject-matter expertise, detailed knowledge of anatomy, and an in-depth understanding of injury terms and definitions as applied in both taxonomies. The official ICD-9-CM and ICD-10-CM versions (injury sections) were mapped to the AIS08 codes and severities, following the rules outlined in each coding manual. The panel of experts was composed of coders certified in ICD and/or AIS from around the world. In the process of developing the map from ICD to AIS, the experts created rules to address issues with the differences in coding guidelines between the 2 schemas and assure a consistent approach to all codes. Over 19,000 ICD codes were analyzed and maps were generated for each code to AIS08 chapters, AIS08 severities, and Injury Severity Score (ISS) body regions. After completion of the maps, 14,101 (74%) of the eligible 19,012 injury-related ICD-9-CM and ICD-10-CM codes were assigned valid AIS08 severity scores between 1 and 6. The remaining 4,911 codes were assigned an AIS08 of 9 (unknown) or were determined to be nonmappable because the ICD description lacked sufficient qualifying information for determining severity according to AIS rules. There were also 15,214 (80%) ICD codes mapped to AIS08 chapter and ISS body region, which allow for ISS calculations for patient data sets. This mapping between ICD and AIS provides a comprehensive, expert-designed solution for analysts to bridge the data gap between the injury descriptions provided in hospital codes (ICD-9-CM, ICD-10-CM) and injury severity codes (AIS08). By applying consistent rules from both the ICD and AIS taxonomies, the expert panel created these definitive maps, which are the only ones endorsed by the Association for the Advancement of Automotive Medicine (AAAM). Initial validation upheld the quality of these maps for the estimation of AIS severity, but future work should include verification of these maps for MAIS and ISS estimations with large data sets. These ICD-AIS maps will support data analysis from databases with injury information classified in these 2 different systems and open new doors for the investigation of injury from traumatic events using large injury data sets.
NASA Technical Reports Server (NTRS)
Paloski, William H.; Odette, Louis L.; Krever, Alfred J.; West, Allison K.
1987-01-01
A real-time expert system is being developed to serve as the astronaut interface for a series of Spacelab vestibular experiments. This expert system is written in a version of Prolog that is itself written in Forth. The Prolog contains a predicate that can be used to execute Forth definitions; thus, the Forth becomes an embedded real-time operating system within the Prolog programming environment. The expert system consists of a data base containing detailed operational instructions for each experiment, a rule base containing Prolog clauses used to determine the next step in an experiment sequence, and a procedure base containing Prolog goals formed from real-time routines coded in Forth. In this paper, we demonstrate and describe the techniques and considerations used to develop this real-time expert system, and we conclude that Forth-based Prolog provides a viable implementation vehicle for this and similar applications.
Preliminary Classification of Army and Navy Entry-Level Occupations by the Holland Coding System.
1986-12-01
Dictionary of Holland Occupational Codes (DOHC; see Gottfredson , Holland, & Ogawa, 1982) either directly or through expert judgment. Results...publications: The Dictionary of Holland Occupational Codes (DHOC; Gottfredson , Holland, & Ogawa, 192) and The Occupations Finder (Holland, 1978). The...occupational categories ( Gottfredson et al., 1982). The agreement between the first letters codes obtained from the 1977 Occupations Finder and the
CLIPS: The C language integrated production system
NASA Technical Reports Server (NTRS)
Riley, Gary
1994-01-01
Expert systems are computer programs which emulate human expertise in well defined problem domains. The potential payoff from expert systems is high: valuable expertise can be captured and preserved, repetitive and/or mundane tasks requiring human expertise can be automated, and uniformity can be applied in decision making processes. The C Language Integrated Production System (CLIPS) is an expert system building tool, developed at the Johnson Space Center, which provides a complete environment for the development and delivery of rule and/or object based expert systems. CLIPS was specifically designed to provide a low cost option for developing and deploying expert system applications across a wide range of hardware platforms. The commercial potential of CLIPS is vast. Currently, CLIPS is being used by over 5,000 individuals throughout the public and private sector. Because the CLIPS source code is readily available, numerous groups have used CLIPS as the basis for their own expert system tools. To date, three commercially available tools have been derived from CLIPS. In general, the development of CLIPS has helped to improve the ability to deliver expert system technology throughout the public and private sectors for a wide range of applications and diverse computing environments.
Code of ethics for the national pharmaceutical system: Codifying and compilation
Salari, Pooneh; Namazi, Hamidreza; Abdollahi, Mohammad; Khansari, Fatemeh; Nikfar, Shekoufeh; Larijani, Bagher; Araminia, Behin
2013-01-01
Pharmacists as one of health-care providers face ethical issues in terms of pharmaceutical care, relationship with patients and cooperation with the health-care team. Other than pharmacy, there are pharmaceutical companies in various fields of manufacturing, importing or distributing that have their own ethical issues. Therefore, pharmacy practice is vulnerable to ethical challenges and needs special code of conducts. On feeling the need, based on a shared project between experts of the ethics from relevant research centers, all the needs were fully recognized and then specified code of conduct for each was written. The code of conduct was subject to comments of all experts involved in the pharmaceutical sector and thus criticized in several meetings. The prepared code of conduct is comprised of professional code of ethics for pharmacists, ethics guideline for pharmaceutical manufacturers, ethics guideline for pharmaceutical importers, ethics guideline for pharmaceutical distributors, and ethics guideline for policy makers. The document was compiled based on the principles of bioethics and professionalism. The compiling the code of ethics for the national pharmaceutical system is the first step in implementing ethics in pharmacy practice and further attempts into teaching the professionalism and the ethical code as the necessary and complementary effort are highly recommended. PMID:24174954
Sojda, Richard S.; Cornely, John E.; Howe, Adele E.
2002-01-01
A decision support system for the management of the Rocky Mountain Population of Trumpeter Swans (Cygnus buccinators) is being developed. As part of this, three expert systems are also in development: one for assessing the quality of Trumpeter Swan breeding habitat; one for making water level recommendations in montane, palustrine wetlands; and one for assessing the contribution a particular site can make towards meeting objectives from as flyway perspective. The focus of this paper is the development of the breeding habitat expert system, which currently consists of 157 rules. Out purpose is to provide decision support for issues that appear to be beyond the capability of a single persons to conceptualize and solve. We propose that by involving multiple experts in the development and use of the systems, management will be significantly improved. The knowledge base for the expert system has been developed using standard knowledge engineering techniques with a small team of ecological experts. Knowledge was then coded using production rules organized in decision trees using a commercial expert system development shell. The final system has been deployed on the world wide web.
An expert system executive for automated assembly of large space truss structures
NASA Technical Reports Server (NTRS)
Allen, Cheryl L.
1993-01-01
Langley Research Center developed a unique test bed for investigating the practical problems associated with the assembly of large space truss structures using robotic manipulators. The test bed is the result of an interdisciplinary effort that encompasses the full spectrum of assembly problems - from the design of mechanisms to the development of software. The automated structures assembly test bed and its operation are described, the expert system executive and its development are detailed, and the planned system evolution is discussed. Emphasis is on the expert system implementation of the program executive. The executive program must direct and reliably perform complex assembly tasks with the flexibility to recover from realistic system errors. The employment of an expert system permits information that pertains to the operation of the system to be encapsulated concisely within a knowledge base. This consolidation substantially reduced code, increased flexibility, eased software upgrades, and realized a savings in software maintenance costs.
Lumb, A.M.; McCammon, R.B.; Kittle, J.L.
1994-01-01
Expert system software was developed to assist less experienced modelers with calibration of a watershed model and to facilitate the interaction between the modeler and the modeling process not provided by mathematical optimization. A prototype was developed with artificial intelligence software tools, a knowledge engineer, and two domain experts. The manual procedures used by the domain experts were identified and the prototype was then coded by the knowledge engineer. The expert system consists of a set of hierarchical rules designed to guide the calibration of the model through a systematic evaluation of model parameters. When the prototype was completed and tested, it was rewritten for portability and operational use and was named HSPEXP. The watershed model Hydrological Simulation Program--Fortran (HSPF) is used in the expert system. This report is the users manual for HSPEXP and contains a discussion of the concepts and detailed steps and examples for using the software. The system has been tested on watersheds in the States of Washington and Maryland, and the system correctly identified the model parameters to be adjusted and the adjustments led to improved calibration.
Harvey, India; Bolgan, Samuela; Mosca, Daniel; McLean, Colin; Rusconi, Elena
2016-01-01
Studies on hacking have typically focused on motivational aspects and general personality traits of the individuals who engage in hacking; little systematic research has been conducted on predispositions that may be associated not only with the choice to pursue a hacking career but also with performance in either naïve or expert populations. Here, we test the hypotheses that two traits that are typically enhanced in autism spectrum disorders-attention to detail and systemizing-may be positively related to both the choice of pursuing a career in information security and skilled performance in a prototypical hacking task (i.e., crypto-analysis or code-breaking). A group of naïve participants and of ethical hackers completed the Autism Spectrum Quotient, including an attention to detail scale, and the Systemizing Quotient (Baron-Cohen et al., 2001, 2003). They were also tested with behavioral tasks involving code-breaking and a control task involving security X-ray image interpretation. Hackers reported significantly higher systemizing and attention to detail than non-hackers. We found a positive relation between self-reported systemizing (but not attention to detail) and code-breaking skills in both hackers and non-hackers, whereas attention to detail (but not systemizing) was related with performance in the X-ray screening task in both groups, as previously reported with naïve participants (Rusconi et al., 2015). We discuss the theoretical and translational implications of our findings.
NASA Technical Reports Server (NTRS)
Gregg, Hugh; Healey, Kathleen; Hack, Edmund; Wong, Carla
1987-01-01
Traditional expert systems, such as diagnostic and training systems, interact with users only through a keyboard and screen, and are usually symbolic in nature. Expert systems that require access to data bases, complex simulations and real-time instrumentation have both symbolic as well as algorithmic computing needs. These needs could both be met using a general purpose workstation running both symbolic and algorithmic code, or separate, specialized computers networked together. The latter approach was chosen to implement TEXSYS, the thermal expert system, developed by NASA Ames Research Center in conjunction with Johnson Space Center to demonstrate the ability of an expert system to autonomously monitor the thermal control system of the space station. TEXSYS has been implemented on a Symbolics workstation, and will be linked to a microVAX computer that will control a thermal test bed. This paper will explore the integration options, and present several possible solutions.
NASA Technical Reports Server (NTRS)
Gregg, Hugh; Healey, Kathleen; Hack, Edmund; Wong, Carla
1987-01-01
Expert systems that require access to data bases, complex simulations and real time instrumentation have both symbolic as well as algorithmic computing needs. These needs could both be met using a general computing workstation running both symbolic and algorithmic code, or separate, specialized computers networked together. The later approach was chosen to implement TEXSYS, the thermal expert system, developed to demonstrate the ability of an expert system to autonomously control the thermal control system of the space station. TEXSYS has been implemented on a Symbolics workstation, and will be linked to a microVAX computer that will control a thermal test bed. Integration options are explored and several possible solutions are presented.
Development of the Diagnostic Expert System for Tea Processing
NASA Astrophysics Data System (ADS)
Yoshitomi, Hitoshi; Yamaguchi, Yuichi
A diagnostic expert system for tea processing which can presume the cause of the defect of the processed tea was developed to contribute to the improvement of tea processing. This system that consists of some programs can be used through the Internet. The inference engine, the core of the system adopts production system which is well used on artificial intelligence, and is coded by Prolog as the artificial intelligence oriented language. At present, 176 rules for inference have been registered on this system. The system will be able to presume better if more rules are added to the system.
Harvey, India; Bolgan, Samuela; Mosca, Daniel; McLean, Colin; Rusconi, Elena
2016-01-01
Studies on hacking have typically focused on motivational aspects and general personality traits of the individuals who engage in hacking; little systematic research has been conducted on predispositions that may be associated not only with the choice to pursue a hacking career but also with performance in either naïve or expert populations. Here, we test the hypotheses that two traits that are typically enhanced in autism spectrum disorders—attention to detail and systemizing—may be positively related to both the choice of pursuing a career in information security and skilled performance in a prototypical hacking task (i.e., crypto-analysis or code-breaking). A group of naïve participants and of ethical hackers completed the Autism Spectrum Quotient, including an attention to detail scale, and the Systemizing Quotient (Baron-Cohen et al., 2001, 2003). They were also tested with behavioral tasks involving code-breaking and a control task involving security X-ray image interpretation. Hackers reported significantly higher systemizing and attention to detail than non-hackers. We found a positive relation between self-reported systemizing (but not attention to detail) and code-breaking skills in both hackers and non-hackers, whereas attention to detail (but not systemizing) was related with performance in the X-ray screening task in both groups, as previously reported with naïve participants (Rusconi et al., 2015). We discuss the theoretical and translational implications of our findings. PMID:27242491
Huber, H; Brambrink, M; Funk, R; Rieger, M
2012-10-01
The purpose of this study was to evaluate differences in the D-DRG results of a hospital case by 2 independently coding MKD raters. Calculation of the 2-inter-rater reliability was performed by examination of the coding of individual hospital cases. The reasons for the non-agreement of the expert evaluations and suggestions to improve the process are discussed. From the expert evaluation pool of the MDK-WL a random sample of 0.7% of the 57,375 expertises was taken. Distribution equality with the basic total was tested by the χ² test or, respectively, Fisher's exact test. For the total of 402 individual hospital cases, the G-DRG case sums of 2 experts of the MDK were determined independently and the results checked for each individual case for agreement or non-agreement. The corresponding confidence intervals with standard errors were analysed to test if certain major diagnosis categories (MDC) were statistically significantly more affected by differing expertise results than others. In 280 of the total 402 tested hospital cases, the 2 MDK raters independently reached the same G-DRG results; in 122 cases the G-DRG case sums determined by the 2 raters differed (agreement 70%; CI 65.2-74.1). Different DRG results between the 2 experts occurred regularly in the entire MDC spectrum. No MDC chapter in which significant differences between the 2 raters arose could be identified. The results of our study demonstrate an almost 70% agreement in the evaluation of hospital cost accounts by 2 independently operating MDK. This result leaves room for improvement. Optimisation potentials can be recognised on the basis of the results. Potential for improvement was established in combination with regular further training and the expansion of binding internal code recommendations as well as exchange of code-relevant information among experts in internal forums. The presented model is in principle suitable for cross-border examinations within the MDK system with the advantage that further trends could be uncovered by more variety and larger numbers of the randomly selected cases. © Georg Thieme Verlag KG Stuttgart · New York.
Braido, Fulvio; Santus, Pierachille; Corsico, Angelo Guido; Di Marco, Fabiano; Melioli, Giovanni; Scichilone, Nicola; Solidoro, Paolo
2018-01-01
The purposes of this study were development and validation of an expert system (ES) aimed at supporting the diagnosis of chronic obstructive lung disease (COLD). A questionnaire and a WebFlex code were developed and validated in silico. An expert panel pilot validation on 60 cases and a clinical validation on 241 cases were performed. The developed questionnaire and code validated in silico resulted in a suitable tool to support the medical diagnosis. The clinical validation of the ES was performed in an academic setting that included six different reference centers for respiratory diseases. The results of the ES expressed as a score associated with the risk of suffering from COLD were matched and compared with the final clinical diagnoses. A set of 60 patients were evaluated by a pilot expert panel validation with the aim of calculating the sample size for the clinical validation study. The concordance analysis between these preliminary ES scores and diagnoses performed by the experts indicated that the accuracy was 94.7% when both experts and the system confirmed the COLD diagnosis and 86.3% when COLD was excluded. Based on these results, the sample size of the validation set was established in 240 patients. The clinical validation, performed on 241 patients, resulted in ES accuracy of 97.5%, with confirmed COLD diagnosis in 53.6% of the cases and excluded COLD diagnosis in 32% of the cases. In 11.2% of cases, a diagnosis of COLD was made by the experts, although the imaging results showed a potential concomitant disorder. The ES presented here (COLD ES ) is a safe and robust supporting tool for COLD diagnosis in primary care settings.
NASA Technical Reports Server (NTRS)
Bochsler, Daniel C.
1988-01-01
A complete listing is given of the expert system rules for the Entry phase of the Onboard Navigation (ONAV) Ground Based Expert Trainer System for aircraft/space shuttle navigation. These source listings appear in the same format as utilized and required by the C Language Integrated Production System (CLIPS) expert system shell which is the basis for the ONAV entry system. A schematic overview is given of how the rules are organized. These groups result from a partitioning of the rules according to the overall function which a given set of rules performs. This partitioning was established and maintained according to that established in the knowledge specification document. In addition, four other groups of rules are specified. The four groups (control flow, operator inputs, output management, and data tables) perform functions that affect all the other functional rule groups. As the name implies, control flow ensures that the rule groups are executed in the order required for proper operation; operator input rules control the introduction into the CLIPS fact base of various kinds of data required by the expert system; output management rules control the updating of the ONAV expert system user display screen during execution of the system; and data tables are static information utilized by many different rule sets gathered in one convenient place.
Lee, Jin Hee; Hong, Ki Jeong; Kim, Do Kyun; Kwak, Young Ho; Jang, Hye Young; Kim, Hahn Bom; Noh, Hyun; Park, Jungho; Song, Bongkyu; Jung, Jae Yun
2013-12-01
A clinically sensible diagnosis grouping system (DGS) is needed for describing pediatric emergency diagnoses for research, medical resource preparedness, and making national policy for pediatric emergency medical care. The Pediatric Emergency Care Applied Research Network (PECARN) developed the DGS successfully. We developed the modified PECARN DGS based on the different pediatric population of South Korea and validated the system to obtain the accurate and comparable epidemiologic data of pediatric emergent conditions of the selected population. The data source used to develop and validate the modified PECARN DGS was the National Emergency Department Information System of South Korea, which was coded by the International Classification of Diseases, 10th Revision (ICD-10) code system. To develop the modified DGS based on ICD-10 code, we matched the selected ICD-10 codes with those of the PECARN DGS by the General Equivalence Mappings (GEMs). After converting ICD-10 codes to ICD-9 codes by GEMs, we matched ICD-9 codes into PECARN DGS categories using the matrix developed by PECARN group. Lastly, we conducted the expert panel survey using Delphi method for the remaining diagnosis codes that were not matched. A total of 1879 ICD-10 codes were used in development of the modified DGS. After 1078 (57.4%) of 1879 ICD-10 codes were assigned to the modified DGS by GEM and PECARN conversion tools, investigators assigned each of the remaining 801 codes (42.6%) to DGS subgroups by 2 rounds of electronic Delphi surveys. And we assigned the remaining 29 codes (4%) into the modified DGS at the second expert consensus meeting. The modified DGS accounts for 98.7% and 95.2% of diagnoses of the 2008 and 2009 National Emergency Department Information System data set. This modified DGS also exhibited strong construct validity using the concepts of age, sex, site of care, and seasons. This also reflected the 2009 outbreak of H1N1 influenza in Korea. We developed and validated clinically feasible and sensible DGS system for describing pediatric emergent conditions in Korea. The modified PECARN DGS showed good comprehensiveness and demonstrated reliable construct validity. This modified DGS based on PECARN DGS framework may be effectively implemented for research, reporting, and resource planning in pediatric emergency system of South Korea.
An evaluation of computer assisted clinical classification algorithms.
Chute, C G; Yang, Y; Buntrock, J
1994-01-01
The Mayo Clinic has a long tradition of indexing patient records in high resolution and volume. Several algorithms have been developed which promise to help human coders in the classification process. We evaluate variations on code browsers and free text indexing systems with respect to their speed and error rates in our production environment. The more sophisticated indexing systems save measurable time in the coding process, but suffer from incompleteness which requires a back-up system or human verification. Expert Network does the best job of rank ordering clinical text, potentially enabling the creation of thresholds for the pass through of computer coded data without human review.
SORTA: a system for ontology-based re-coding and technical annotation of biomedical phenotype data.
Pang, Chao; Sollie, Annet; Sijtsma, Anna; Hendriksen, Dennis; Charbon, Bart; de Haan, Mark; de Boer, Tommy; Kelpin, Fleur; Jetten, Jonathan; van der Velde, Joeri K; Smidt, Nynke; Sijmons, Rolf; Hillege, Hans; Swertz, Morris A
2015-01-01
There is an urgent need to standardize the semantics of biomedical data values, such as phenotypes, to enable comparative and integrative analyses. However, it is unlikely that all studies will use the same data collection protocols. As a result, retrospective standardization is often required, which involves matching of original (unstructured or locally coded) data to widely used coding or ontology systems such as SNOMED CT (clinical terms), ICD-10 (International Classification of Disease) and HPO (Human Phenotype Ontology). This data curation process is usually a time-consuming process performed by a human expert. To help mechanize this process, we have developed SORTA, a computer-aided system for rapidly encoding free text or locally coded values to a formal coding system or ontology. SORTA matches original data values (uploaded in semicolon delimited format) to a target coding system (uploaded in Excel spreadsheet, OWL ontology web language or OBO open biomedical ontologies format). It then semi- automatically shortlists candidate codes for each data value using Lucene and n-gram based matching algorithms, and can also learn from matches chosen by human experts. We evaluated SORTA's applicability in two use cases. For the LifeLines biobank, we used SORTA to recode 90 000 free text values (including 5211 unique values) about physical exercise to MET (Metabolic Equivalent of Task) codes. For the CINEAS clinical symptom coding system, we used SORTA to map to HPO, enriching HPO when necessary (315 terms matched so far). Out of the shortlists at rank 1, we found a precision/recall of 0.97/0.98 in LifeLines and of 0.58/0.45 in CINEAS. More importantly, users found the tool both a major time saver and a quality improvement because SORTA reduced the chances of human mistakes. Thus, SORTA can dramatically ease data (re)coding tasks and we believe it will prove useful for many more projects. Database URL: http://molgenis.org/sorta or as an open source download from http://www.molgenis.org/wiki/SORTA. © The Author(s) 2015. Published by Oxford University Press.
SORTA: a system for ontology-based re-coding and technical annotation of biomedical phenotype data
Pang, Chao; Sollie, Annet; Sijtsma, Anna; Hendriksen, Dennis; Charbon, Bart; de Haan, Mark; de Boer, Tommy; Kelpin, Fleur; Jetten, Jonathan; van der Velde, Joeri K.; Smidt, Nynke; Sijmons, Rolf; Hillege, Hans; Swertz, Morris A.
2015-01-01
There is an urgent need to standardize the semantics of biomedical data values, such as phenotypes, to enable comparative and integrative analyses. However, it is unlikely that all studies will use the same data collection protocols. As a result, retrospective standardization is often required, which involves matching of original (unstructured or locally coded) data to widely used coding or ontology systems such as SNOMED CT (clinical terms), ICD-10 (International Classification of Disease) and HPO (Human Phenotype Ontology). This data curation process is usually a time-consuming process performed by a human expert. To help mechanize this process, we have developed SORTA, a computer-aided system for rapidly encoding free text or locally coded values to a formal coding system or ontology. SORTA matches original data values (uploaded in semicolon delimited format) to a target coding system (uploaded in Excel spreadsheet, OWL ontology web language or OBO open biomedical ontologies format). It then semi- automatically shortlists candidate codes for each data value using Lucene and n-gram based matching algorithms, and can also learn from matches chosen by human experts. We evaluated SORTA’s applicability in two use cases. For the LifeLines biobank, we used SORTA to recode 90 000 free text values (including 5211 unique values) about physical exercise to MET (Metabolic Equivalent of Task) codes. For the CINEAS clinical symptom coding system, we used SORTA to map to HPO, enriching HPO when necessary (315 terms matched so far). Out of the shortlists at rank 1, we found a precision/recall of 0.97/0.98 in LifeLines and of 0.58/0.45 in CINEAS. More importantly, users found the tool both a major time saver and a quality improvement because SORTA reduced the chances of human mistakes. Thus, SORTA can dramatically ease data (re)coding tasks and we believe it will prove useful for many more projects. Database URL: http://molgenis.org/sorta or as an open source download from http://www.molgenis.org/wiki/SORTA PMID:26385205
NASA Technical Reports Server (NTRS)
Brown, David B.
1990-01-01
The results of research and development efforts are described for Task one, Phase two of a general project entitled The Development of a Program Analysis Environment for Ada. The scope of this task includes the design and development of a prototype system for testing Ada software modules at the unit level. The system is called Query Utility Environment for Software Testing of Ada (QUEST/Ada). The prototype for condition coverage provides a platform that implements expert system interaction with program testing. The expert system can modify data in the instrument source code in order to achieve coverage goals. Given this initial prototype, it is possible to evaluate the rule base in order to develop improved rules for test case generation. The goals of Phase two are the following: (1) to continue to develop and improve the current user interface to support the other goals of this research effort (i.e., those related to improved testing efficiency and increased code reliable); (2) to develop and empirically evaluate a succession of alternative rule bases for the test case generator such that the expert system achieves coverage in a more efficient manner; and (3) to extend the concepts of the current test environment to address the issues of Ada concurrency.
Müller, Sean; Vallence, Ann-Maree; Winstein, Carolee
2017-12-14
A framework is presented of how theoretical predictions can be tested across the expert athlete to disabled patient skill continuum. Common-coding theory is used as the exemplar to discuss sensory and motor system contributions to perceptual-motor behavior. Behavioral and neural studies investigating expert athletes and patients recovering from cerebral stroke are reviewed. They provide evidence of bi-directional contributions of visual and motor systems to perceptual-motor behavior. Majority of this research is focused on perceptual-motor performance or learning, with less on transfer. The field is ripe for research designed to test theoretical predictions across the expert athlete to disabled patient skill continuum. Our view has implications for theory and practice in sports science, physical education, and rehabilitation.
NASA Technical Reports Server (NTRS)
Gregg, Hugh; Healey, Kathleen; Hack, Edmund; Wong, Carla
1988-01-01
Expert systems that require access to data bases, complex simulations and real time instrumentation have both symbolic and algorithmic needs. Both of these needs could be met using a general purpose workstation running both symbolic and algorithmic codes, or separate, specialized computers networked together. The later approach was chosen to implement TEXSYS, the thermal expert system, developed by the NASA Ames Research Center in conjunction with the Johnson Space Center to demonstrate the ability of an expert system to autonomously monitor the thermal control system of the space station. TEXSYS has been implemented on a Symbolics workstation, and will be linked to a microVAX computer that will control a thermal test bed. The integration options and several possible solutions are presented.
Multiple neural network approaches to clinical expert systems
NASA Astrophysics Data System (ADS)
Stubbs, Derek F.
1990-08-01
We briefly review the concept of computer aided medical diagnosis and more extensively review the the existing literature on neural network applications in the field. Neural networks can function as simple expert systems for diagnosis or prognosis. Using a public database we develop a neural network for the diagnosis of a major presenting symptom while discussing the development process and possible approaches. MEDICAL EXPERTS SYSTEMS COMPUTER AIDED DIAGNOSIS Biomedicine is an incredibly diverse and multidisciplinary field and it is not surprising that neural networks with their many applications are finding more and more applications in the highly non-linear field of biomedicine. I want to concentrate on neural networks as medical expert systems for clinical diagnosis or prognosis. Expert Systems started out as a set of computerized " ifthen" rules. Everything was reduced to boolean logic and the promised land of computer experts was said to be in sight. It never came. Why? First the computer code explodes as the number of " ifs" increases. All the " ifs" have to interact. Second experts are not very good at reducing expertise to language. It turns out that experts recognize patterns and have non-verbal left-brain intuition decision processes. Third learning by example rather than learning by rule is the way natural brains works and making computers work by rule-learning is hideously labor intensive. Neural networks can learn from example. They learn the results
Proceedings of the Workshop on software tools for distributed intelligent control systems
DOE Office of Scientific and Technical Information (OSTI.GOV)
Herget, C.J.
1990-09-01
The Workshop on Software Tools for Distributed Intelligent Control Systems was organized by Lawrence Livermore National Laboratory for the United States Army Headquarters Training and Doctrine Command and the Defense Advanced Research Projects Agency. The goals of the workshop were to the identify the current state of the art in tools which support control systems engineering design and implementation, identify research issues associated with writing software tools which would provide a design environment to assist engineers in multidisciplinary control design and implementation, formulate a potential investment strategy to resolve the research issues and develop public domain code which can formmore » the core of more powerful engineering design tools, and recommend test cases to focus the software development process and test associated performance metrics. Recognizing that the development of software tools for distributed intelligent control systems will require a multidisciplinary effort, experts in systems engineering, control systems engineering, and compute science were invited to participate in the workshop. In particular, experts who could address the following topics were selected: operating systems, engineering data representation and manipulation, emerging standards for manufacturing data, mathematical foundations, coupling of symbolic and numerical computation, user interface, system identification, system representation at different levels of abstraction, system specification, system design, verification and validation, automatic code generation, and integration of modular, reusable code.« less
A Boundary Condition for Simulation of Flow Over Porous Surfaces
NASA Technical Reports Server (NTRS)
Frink, Neal T.; Bonhaus, Daryl L.; Vatsa, Veer N.; Bauer, Steven X. S.; Tinetti, Ana F.
2001-01-01
A new boundary condition is presented.for simulating the flow over passively porous surfaces. The model builds on the prior work of R.H. Bush to eliminate the need for constructing grid within an underlying plenum, thereby simplifying the numerical modeling of passively porous flow control systems and reducing computation cost. Code experts.for two structured-grid.flow solvers, TLNS3D and CFL3D. and one unstructured solver, USM3Dns, collaborated with an experimental porosity expert to develop the model and implement it into their respective codes. Results presented,for the three codes on a slender forebody with circumferential porosity and a wing with leading-edge porosity demonstrate a good agreement with experimental data and a remarkable ability to predict the aggregate aerodynamic effects of surface porosity with a simple boundary condition.
ERIC Educational Resources Information Center
Dondi, Marco; Messinger, Daniel; Colle, Marta; Tabasso, Alessia; Simion, Francesca; Barba, Beatrice Dalla; Fogel, Alan
2007-01-01
To better understand the form and recognizability of neonatal smiling, 32 newborns (14 girls; M = 25.6 hr) were videorecorded in the behavioral states of alertness, drowsiness, active sleep, and quiet sleep. Baby Facial Action Coding System coding of both lip corner raising (simple or non-Duchenne) and lip corner raising with cheek raising…
Expert system validation in prolog
NASA Technical Reports Server (NTRS)
Stock, Todd; Stachowitz, Rolf; Chang, Chin-Liang; Combs, Jacqueline
1988-01-01
An overview of the Expert System Validation Assistant (EVA) is being implemented in Prolog at the Lockheed AI Center. Prolog was chosen to facilitate rapid prototyping of the structure and logic checkers and since February 1987, we have implemented code to check for irrelevance, subsumption, duplication, deadends, unreachability, and cycles. The architecture chosen is extremely flexible and expansible, yet concise and complementary with the normal interactive style of Prolog. The foundation of the system is in the connection graph representation. Rules and facts are modeled as nodes in the graph and arcs indicate common patterns between rules. The basic activity of the validation system is then a traversal of the connection graph, searching for various patterns the system recognizes as erroneous. To aid in specifying these patterns, a metalanguage is developed, providing the user with the basic facilities required to reason about the expert system. Using the metalanguage, the user can, for example, give the Prolog inference engine the goal of finding inconsistent conclusions among the rules, and Prolog will search the graph intantiations which can match the definition of inconsistency. Examples of code for some of the checkers are provided and the algorithms explained. Technical highlights include automatic construction of a connection graph, demonstration of the use of metalanguage, the A* algorithm modified to detect all unique cycles, general-purpose stacks in Prolog, and a general-purpose database browser with pattern completion.
SIRE: A Simple Interactive Rule Editor for NICBES
NASA Technical Reports Server (NTRS)
Bykat, Alex
1988-01-01
To support evolution of domain expertise, and its representation in an expert system knowledge base, a user-friendly rule base editor is mandatory. The Nickel Cadmium Battery Expert System (NICBES), a prototype of an expert system for the Hubble Space Telescope power storage management system, does not provide such an editor. In the following, a description of a Simple Interactive Rule Base Editor (SIRE) for NICBES is described. The SIRE provides a consistent internal representation of the NICBES knowledge base. It supports knowledge presentation and provides a user-friendly and code language independent medium for rule addition and modification. The SIRE is integrated with NICBES via an interface module. This module provides translation of the internal representation to Prolog-type rules (Horn clauses), latter rule assertion, and a simple mechanism for rule selection for its Prolog inference engine.
Information quality measurement of medical encoding support based on usability.
Puentes, John; Montagner, Julien; Lecornu, Laurent; Cauvin, Jean-Michel
2013-12-01
Medical encoding support systems for diagnoses and medical procedures are an emerging technology that begins to play a key role in billing, reimbursement, and health policies decisions. A significant problem to exploit these systems is how to measure the appropriateness of any automatically generated list of codes, in terms of fitness for use, i.e. their quality. Until now, only information retrieval performance measurements have been applied to estimate the accuracy of codes lists as quality indicator. Such measurements do not give the value of codes lists for practical medical encoding, and cannot be used to globally compare the quality of multiple codes lists. This paper defines and validates a new encoding information quality measure that addresses the problem of measuring medical codes lists quality. It is based on a usability study of how expert coders and physicians apply computer-assisted medical encoding. The proposed measure, named ADN, evaluates codes Accuracy, Dispersion and Noise, and is adapted to the variable length and content of generated codes lists, coping with limitations of previous measures. According to the ADN measure, the information quality of a codes list is fully represented by a single point, within a suitably constrained feature space. Using one scheme, our approach is reliable to measure and compare the information quality of hundreds of codes lists, showing their practical value for medical encoding. Its pertinence is demonstrated by simulation and application to real data corresponding to 502 inpatient stays in four clinic departments. Results are compared to the consensus of three expert coders who also coded this anonymized database of discharge summaries, and to five information retrieval measures. Information quality assessment applying the ADN measure showed the degree of encoding-support system variability from one clinic department to another, providing a global evaluation of quality measurement trends. Copyright © 2013 Elsevier Ireland Ltd. All rights reserved.
A simple system for detection of EEG artifacts in polysomnographic recordings.
Durka, P J; Klekowicz, H; Blinowska, K J; Szelenberger, W; Niemcewicz, Sz
2003-04-01
We present an efficient parametric system for automatic detection of electroencephalogram (EEG) artifacts in polysomnographic recordings. For each of the selected types of artifacts, a relevant parameter was calculated for a given epoch. If any of these parameters exceeded a threshold, the epoch was marked as an artifact. Performance of the system, evaluated on 18 overnight polysomnographic recordings, revealed concordance with decisions of human experts close to the interexpert agreement and the repeatability of expert's decisions, assessed via a double-blind test. Complete software (Matlab source code) for the presented system is freely available from the Internet at http://brain.fuw.edu.pl/artifacts.
Developing Software For Monitoring And Diagnosis
NASA Technical Reports Server (NTRS)
Edwards, S. J.; Caglayan, A. K.
1993-01-01
Expert-system software shell produces executable code. Report discusses beginning phase of research directed toward development of artificial intelligence for real-time monitoring of, and diagnosis of faults in, complicated systems of equipment. Motivated by need for onboard monitoring and diagnosis of electronic sensing and controlling systems of advanced aircraft. Also applicable to such equipment systems as refineries, factories, and powerplants.
Zonfrillo, Mark R; Weaver, Ashley A; Gillich, Patrick J; Price, Janet P; Stitzel, Joel D
2015-01-01
There has been a longstanding desire for a map to convert International Classification of Diseases (ICD) injury codes to Abbreviated Injury Scale (AIS) codes to reflect the severity of those diagnoses. The Association for the Advancement of Automotive Medicine (AAAM) was tasked by European Union representatives to create a categorical map classifying diagnoses codes as serious injury (Abbreviated Injury Scale [AIS] 3+), minor/moderate injury (AIS 1/2), or indeterminate. This study's objective was to map injury-related ICD-9-CM (clinical modification) and ICD-10-CM codes to these severity categories. Approximately 19,000 ICD codes were mapped, including injuries from the following categories: amputations, blood vessel injury, burns, crushing injury, dislocations/sprains/strains, foreign body, fractures, internal organ, nerve/spinal cord injury, intracranial, laceration, open wounds, and superficial injury/contusion. Two parallel activities were completed to create the maps: (1) An in-person expert panel and (2) an electronic survey. The panel consisted of expert users of AIS and ICD from North America, the United Kingdom, and Australia. The panel met in person for 5 days, with follow-up virtual meetings to create and revise the maps. Additional qualitative data were documented to resolve potential discrepancies in mapping. The electronic survey was completed by 95 injury coding professionals from North America, Spain, Australia, and New Zealand over 12 weeks. ICD-to-AIS maps were created for: ICD-9-CM and ICD-10-CM. Both maps indicated whether the corresponding AIS 2005/Update 2008 severity score for each ICD code was AIS 3+, 1/2, or indeterminable. Though some ICD codes could be mapped to multiple AIS codes, the maximum severity of all potentially mapped injuries determined the final severity categorization. The in-person panel consisted of 13 experts, with 11 Certified AIS specialists (CAISS) with a median of 8 years and an average of 15 years of coding experience. Consensus was reached for AIS severity categorization for all injury-related ICD codes. There were 95 survey respondents, with a median of 8 years of injury coding experience. Approximately 15 survey responses were collected per ICD code. Results from the 2 activities were compared, and any discrepancies were resolved using additional qualitative and quantitative data from the in-person panel and survey results, respectively. Robust maps of ICD-9-CM and ICD-10-CM injury codes to AIS severity categories (3+ versus <3) were successfully created from an in-person panel discussion and electronic survey. These maps provide a link between the common ICD diagnostic lexicons and the AIS severity coding system and are of value to injury researchers, public health scientists, and epidemiologists using large databases without available AIS coding.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Settlemyer, S.R.
1991-09-01
The Nuclear Weapons Management System combines the strengths of an expert system with the flexibility of a database management system to assist the Weapons Officer, Security Officer, and the Personnel Reliability Program Officer in the performance of administrative duties associated with the nuclear weapons programs in the United States Navy. This thesis examines the need for, and ultimately the design of, a system that will assist the Security Officer in administrative duties associated with the Shipboard Self Defense Force. This system, designed and coded utilizing dBASE IV, can be implemented as a stand alone system. Furthermore, it interfaces with themore » expert system submodule that handles the PRP screening process.« less
MHEG Based Distance Learning System on Information Superhighway.
ERIC Educational Resources Information Center
Lee, SeiHoon; Yoon, KyungSeob; Wang, ChangJong
As the need for distance education grows, requirements for the development of high-speed network-based real-time distance learning systems increases. MHEG-5 is the fifth part of the MHEG (Multimedia and Hypermedia information coding Experts Group) standard, and it defines a final-form representation for application interchange. This paper…
A rule-based expert system applied to moisture durability of building envelopes
Boudreaux, Philip R.; Pallin, Simon B.; Accawi, Gina K.; ...
2018-01-09
The moisture durability of an envelope component such as a wall or roof is difficult to predict. Moisture durability depends on all the construction materials used, as well as the climate, orientation, air tightness, and indoor conditions. Modern building codes require more insulation and tighter construction but provide little guidance about how to ensure these energy-efficient assemblies remain moisture durable. Furthermore, as new products and materials are introduced, builders are increasingly uncertain about the long-term durability of their building envelope designs. Oak Ridge National Laboratory and the US Department of Energy’s Building America Program are applying a rule-based expert systemmore » methodology in a web tool to help designers determine whether a given wall design is likely to be moisture durable and provide expert guidance on moisture risk management specific to a wall design and climate. Finally, the expert system is populated with knowledge from both expert judgment and probabilistic hygrothermal simulation results.« less
A rule-based expert system applied to moisture durability of building envelopes
DOE Office of Scientific and Technical Information (OSTI.GOV)
Boudreaux, Philip R.; Pallin, Simon B.; Accawi, Gina K.
The moisture durability of an envelope component such as a wall or roof is difficult to predict. Moisture durability depends on all the construction materials used, as well as the climate, orientation, air tightness, and indoor conditions. Modern building codes require more insulation and tighter construction but provide little guidance about how to ensure these energy-efficient assemblies remain moisture durable. Furthermore, as new products and materials are introduced, builders are increasingly uncertain about the long-term durability of their building envelope designs. Oak Ridge National Laboratory and the US Department of Energy’s Building America Program are applying a rule-based expert systemmore » methodology in a web tool to help designers determine whether a given wall design is likely to be moisture durable and provide expert guidance on moisture risk management specific to a wall design and climate. Finally, the expert system is populated with knowledge from both expert judgment and probabilistic hygrothermal simulation results.« less
Donato, Gianluca; Bartlett, Marian Stewart; Hager, Joseph C.; Ekman, Paul; Sejnowski, Terrence J.
2010-01-01
The Facial Action Coding System (FACS) [23] is an objective method for quantifying facial movement in terms of component actions. This system is widely used in behavioral investigations of emotion, cognitive processes, and social interaction. The coding is presently performed by highly trained human experts. This paper explores and compares techniques for automatically recognizing facial actions in sequences of images. These techniques include analysis of facial motion through estimation of optical flow; holistic spatial analysis, such as principal component analysis, independent component analysis, local feature analysis, and linear discriminant analysis; and methods based on the outputs of local filters, such as Gabor wavelet representations and local principal components. Performance of these systems is compared to naive and expert human subjects. Best performances were obtained using the Gabor wavelet representation and the independent component representation, both of which achieved 96 percent accuracy for classifying 12 facial actions of the upper and lower face. The results provide converging evidence for the importance of using local filters, high spatial frequencies, and statistical independence for classifying facial actions. PMID:21188284
Vendrame, Alan; Silva, Rebeca; Xuan, Ziming; Sparks, Robert; Noel, Jonathan; Pinsky, Ilana
2015-09-01
We assessed the impact of the 2010 revisions to Brazil's self-regulatory alcohol marketing code using expert and adolescent raters. Five popular TV beer ads were selected. Ads were rated based on the 2010 Brazilian self-regulatory marketing code. The expert group (N = 31) represented health-related professions; the adolescent group (N = 110) were public high school students. At least 1 ad violated 11 of 17 guidelines included in the study. Ratings by experts and adolescents were similar. Both found violations in all sections of the self-regulatory code, but significant group differences were seen in applying the section that prohibits the promotion of excessive alcohol consumption, with experts identifying more violations than adolescents. Beer ads in the sample systematically violated the self-regulatory standards for alcohol advertising in Brazil according to both experts and youth. Public policies for more effective restrictions and prohibitions in alcohol ads should be considered. © The Author 2015. Medical Council on Alcohol and Oxford University Press. All rights reserved.
NASA Astrophysics Data System (ADS)
Vaucouleur, Sebastien
2011-02-01
We introduce code query by example for customisation of evolvable software products in general and of enterprise resource planning systems (ERPs) in particular. The concept is based on an initial empirical study on practices around ERP systems. We motivate our design choices based on those empirical results, and we show how the proposed solution helps with respect to the infamous upgrade problem: the conflict between the need for customisation and the need for upgrade of ERP systems. We further show how code query by example can be used as a form of lightweight static analysis, to detect automatically potential defects in large software products. Code query by example as a form of lightweight static analysis is particularly interesting in the context of ERP systems: it is often the case that programmers working in this field are not computer science specialists but more of domain experts. Hence, they require a simple language to express custom rules.
Aerothermal Assment Of The Expert Flap In The SCIROCCO Wind Tunnel
NASA Astrophysics Data System (ADS)
Walpot, L.; Di Clemente, M.; Vos, J.; Etchells, J.; Trifoni, E.; Thoemel, J.; Gavira, J.
2011-05-01
In the frame of the “In-Flight Test Measurement Techniques for Aerothermodynamics” activity of the EXPERT Program, the EXPERT Instrumented Open Flap Assembly experiment has the objective to verify the design/sensor integration and validate the CFD tools. Ground based measurements were made in Europe’s largest high enthalpy plasma facility, Scirocco in Italy. Two EXPERT flaps of the flight article, instrumented with 14 thermocouples, 5 pressure ports, a pyrometer and an IR camera mounted in the cavity instrumented flap will collect in-flight data. During the Scirocco experiment, an EXPERT flap model identical to the flight article was mounted at 45 deg on a holder including cavity and was subjected to a hot plasma flow at an enthalpy up to 11MJ/kg at a stagnation pressure of 7 bar. The test model sports the same pressure sensors as the flight article. Hypersonic state-of-the-art codes were then be used to perform code-to-code and wind tunnel-to-code comparisons, including thermal response of the flap as collected during the tests by the sensors and camera.
Brasil, L M; de Azevedo, F M; Barreto, J M
2001-09-01
This paper proposes a hybrid expert system (HES) to minimise some complexity problems pervasive to the artificial intelligence such as: the knowledge elicitation process, known as the bottleneck of expert systems; the model choice for knowledge representation to code human reasoning; the number of neurons in the hidden layer and the topology used in the connectionist approach; the difficulty to obtain the explanation on how the network arrived to a conclusion. Two algorithms applied to developing of HES are also suggested. One of them is used to train the fuzzy neural network and the other to obtain explanations on how the fuzzy neural network attained a conclusion. To overcome these difficulties the cognitive computing was integrated to the developed system. A case study is presented (e.g. epileptic crisis) with the problem definition and simulations. Results are also discussed.
Probabilistic Structural Analysis Methods (PSAM) for select space propulsion system components
NASA Technical Reports Server (NTRS)
1991-01-01
The fourth year of technical developments on the Numerical Evaluation of Stochastic Structures Under Stress (NESSUS) system for Probabilistic Structural Analysis Methods is summarized. The effort focused on the continued expansion of the Probabilistic Finite Element Method (PFEM) code, the implementation of the Probabilistic Boundary Element Method (PBEM), and the implementation of the Probabilistic Approximate Methods (PAppM) code. The principal focus for the PFEM code is the addition of a multilevel structural dynamics capability. The strategy includes probabilistic loads, treatment of material, geometry uncertainty, and full probabilistic variables. Enhancements are included for the Fast Probability Integration (FPI) algorithms and the addition of Monte Carlo simulation as an alternate. Work on the expert system and boundary element developments continues. The enhanced capability in the computer codes is validated by applications to a turbine blade and to an oxidizer duct.
From Novice to Expert: Problem Solving in ICD-10-PCS Procedural Coding
Rousse, Justin Thomas
2013-01-01
The benefits of converting to ICD-10-CM/PCS have been well documented in recent years. One of the greatest challenges in the conversion, however, is how to train the workforce in the code sets. The International Classification of Diseases, Tenth Revision, Procedure Coding System (ICD-10-PCS) has been described as a language requiring higher-level reasoning skills because of the system's increased granularity. Training and problem-solving strategies required for correct procedural coding are unclear. The objective of this article is to propose that the acquisition of rule-based logic will need to be augmented with self-evaluative and critical thinking. Awareness of how this process works is helpful for established coders as well as for a new generation of coders who will master the complexities of the system. PMID:23861674
Spacecraft attitude control using a smart control system
NASA Technical Reports Server (NTRS)
Buckley, Brian; Wheatcraft, Louis
1992-01-01
Traditionally, spacecraft attitude control has been implemented using control loops written in native code for a space hardened processor. The Naval Research Lab has taken this approach during the development of the Attitude Control Electronics (ACE) package. After the system was developed and delivered, NRL decided to explore alternate technologies to accomplish this same task more efficiently. The approach taken by NRL was to implement the ACE control loops using systems technologies. The purpose of this effort was to: (1) research capabilities required of an expert system in processing a classic closed-loop control algorithm; (2) research the development environment required to design and test an embedded expert systems environment; (3) research the complexity of design and development of expert systems versus a conventional approach; and (4) test the resulting systems against the flight acceptance test software for both response and accuracy. Two expert systems were selected to implement the control loops. Criteria used for the selection of the expert systems included that they had to run in both embedded systems and ground based environments. Using two different expert systems allowed a comparison of the real-time capabilities, inferencing capabilities, and the ground-based development environment. The two expert systems chosen for the evaluation were Spacecraft Command Language (SCL), and NEXTPERT Object. SCL is a smart control system produced for the NRL by Interface and Control Systems (ICS). SCL was developed to be used for real-time command, control, and monitoring of a new generation of spacecraft. NEXPERT Object is a commercially available product developed by Neuron Data. Results of the effort were evaluated using the ACE test bed. The ACE test bed had been developed and used to test the original flight hardware and software using simulators and flight-like interfaces. The test bed was used for testing the expert systems in a 'near-flight' environment. The technical approach, the system architecture, the development environments, knowledge base development, and results of this effort are detailed.
Web-Based Environment for Maintaining Legacy Software
NASA Technical Reports Server (NTRS)
Tigges, Michael; Thompson, Nelson; Orr, Mark; Fox, Richard
2007-01-01
Advanced Tool Integration Environment (ATIE) is the name of both a software system and a Web-based environment created by the system for maintaining an archive of legacy software and expertise involved in developing the legacy software. ATIE can also be used in modifying legacy software and developing new software. The information that can be encapsulated in ATIE includes experts documentation, input and output data of tests cases, source code, and compilation scripts. All of this information is available within a common environment and retained in a database for ease of access and recovery by use of powerful search engines. ATIE also accommodates the embedment of supporting software that users require for their work, and even enables access to supporting commercial-off-the-shelf (COTS) software within the flow of the experts work. The flow of work can be captured by saving the sequence of computer programs that the expert uses. A user gains access to ATIE via a Web browser. A modern Web-based graphical user interface promotes efficiency in the retrieval, execution, and modification of legacy code. Thus, ATIE saves time and money in the support of new and pre-existing programs.
Application Reuse Library for Software, Requirements, and Guidelines
NASA Technical Reports Server (NTRS)
Malin, Jane T.; Thronesbery, Carroll
1994-01-01
Better designs are needed for expert systems and other operations automation software, for more reliable, usable and effective human support. A prototype computer-aided Application Reuse Library shows feasibility of supporting concurrent development and improvement of advanced software by users, analysts, software developers, and human-computer interaction experts. Such a library expedites development of quality software, by providing working, documented examples, which support understanding, modification and reuse of requirements as well as code. It explicitly documents and implicitly embodies design guidelines, standards and conventions. The Application Reuse Library provides application modules with Demo-and-Tester elements. Developers and users can evaluate applicability of a library module and test modifications, by running it interactively. Sub-modules provide application code and displays and controls. The library supports software modification and reuse, by providing alternative versions of application and display functionality. Information about human support and display requirements is provided, so that modifications will conform to guidelines. The library supports entry of new application modules from developers throughout an organization. Example library modules include a timer, some buttons and special fonts, and a real-time data interface program. The library prototype is implemented in the object-oriented G2 environment for developing real-time expert systems.
Crowdsourcing the Measurement of Interstate Conflict
2016-01-01
Much of the data used to measure conflict is extracted from news reports. This is typically accomplished using either expert coders to quantify the relevant information or machine coders to automatically extract data from documents. Although expert coding is costly, it produces quality data. Machine coding is fast and inexpensive, but the data are noisy. To diminish the severity of this tradeoff, we introduce a method for analyzing news documents that uses crowdsourcing, supplemented with computational approaches. The new method is tested on documents about Militarized Interstate Disputes, and its accuracy ranges between about 68 and 76 percent. This is shown to be a considerable improvement over automated coding, and to cost less and be much faster than expert coding. PMID:27310427
Röhnsch, Gundula; Flick, Uwe
2015-10-01
Which representations of care can be found in migrants with alcohol or drug problems from the former Soviet Union? How do they correspond with views in the care system? Episodic interviews with 46 migrants, expert interviews with 33 service providers; analysis with thematic coding. For migrants and experts holistic care is important, which include spiritual-religious components but are also control-oriented. The cultural specificity of migrants' care representations should be acknowledged by the health care system much more. © Georg Thieme Verlag KG Stuttgart · New York.
Rodrigues, J M; Trombert-Paviot, B; Baud, R; Wagner, J; Meusnier-Carriot, F
1998-01-01
GALEN has developed a language independent common reference model based on a medically oriented ontology and practical tools and techniques for managing healthcare terminology including natural language processing. GALEN-IN-USE is the current phase which applied the modelling and the tools to the development or the updating of coding systems for surgical procedures in different national coding centers co-operating within the European Federation of Coding Centre (EFCC) to create a language independent knowledge repository for multicultural Europe. We used an integrated set of artificial intelligence terminology tools named CLAssification Manager workbench to process French professional medical language rubrics into intermediate dissections and to the Grail reference ontology model representation. From this language independent concept model representation we generate controlled French natural language. The French national coding centre is then able to retrieve the initial professional rubrics with different categories of concepts, to compare the professional language proposed by expert clinicians to the French generated controlled vocabulary and to finalize the linguistic labels of the coding system in relation with the meanings of the conceptual system structure.
An expert system for wind shear avoidance
NASA Technical Reports Server (NTRS)
Stengel, Robert F.; Stratton, D. Alexander
1990-01-01
A study of intelligent guidance and control concepts for protecting against the adverse effects of wind shear during aircraft takeoffs and landings is being conducted, with current emphasis on developing an expert system for wind shear avoidance. Principal objectives are to develop methods for assessing the likelihood of wind shear encounter (based on real-time information in the cockpit), for deciding what flight path to pursue (e.g., takeoff abort, landing go-around, or normal climbout or glide slope), and for using the aircraft's full potential for combating wind shear. This study requires the definition of both deterministic and statistical techniques for fusing internal and external information , for making go/no-go decisions, and for generating commands to the manually controlled flight. The program has begun with the development of the WindShear Safety Advisor, an expert system for pilot aiding that is based on the FAA Windshear Training Aid; a two-volume manual that presents an overview , pilot guide, training program, and substantiating data provides guidelines for this initial development. The WindShear Safety Advisor expert system currently contains over 200 rules and is coded in the LISP programming language.
1987-08-01
TO 8/87 68 6 CUPPEIFMyTARY 0A,()N Copies are available from,. the National Technical Information Service Springfield, VA 22161 *COSATI CODES 18 SUBJECT... information . The first exploratory research step was to determine the breath and depth of the construction schedule analysis domain. This step defined...ADDITIONAL INFORMATION REGARDING THIS RESEARCIH I. O’Connor, Michael J., Jesus M. Dc La Garza, and C. William Ibbs, "An Expert Systcm for Construction
Automated Classification of Power Signals
2008-06-01
determine when a transient occurs. The identification of this signal can then be determined by an expert classifier and a series of these...the manual identification and classification of system events. Once events were located, the characteristics were examined to determine if system... identification code, which varies depending on the system classifier that is specified. Figure 3-7 provides an example of a Linux directory containing
Overview of the H.264/AVC video coding standard
NASA Astrophysics Data System (ADS)
Luthra, Ajay; Topiwala, Pankaj N.
2003-11-01
H.264/MPEG-4 AVC is the latest coding standard jointly developed by the Video Coding Experts Group (VCEG) of ITU-T and Moving Picture Experts Group (MPEG) of ISO/IEC. It uses state of the art coding tools and provides enhanced coding efficiency for a wide range of applications including video telephony, video conferencing, TV, storage (DVD and/or hard disk based), streaming video, digital video creation, digital cinema and others. In this paper an overview of this standard is provided. Some comparisons with the existing standards, MPEG-2 and MPEG-4 Part 2, are also provided.
NASA Technical Reports Server (NTRS)
Jones, Jeremy; Grosvenor, Sandy; Wolf, Karl; Li, Connie; Koratkar, Anuradha; Powers, Edward I. (Technical Monitor)
2001-01-01
In the Virtual Observatory (VO), software tools will perform the functions that have traditionally been performed by physical observatories and their instruments. These tools will not be adjuncts to VO functionality but will make up the very core of the VO. Consequently, the tradition of observatory and system independent tools serving a small user base is not valid for the VO. For the VO to succeed, we must improve software collaboration and code sharing between projects and groups. A significant goal of the Scientist's Expert Assistant (SEA) project has been promoting effective collaboration and code sharing between groups. During the past three years, the SEA project has been developing prototypes for new observation planning software tools and strategies. Initially funded by the Next Generation Space Telescope, parts of the SEA code have since been adopted by the Space Telescope Science Institute. SEA has also supplied code for SOFIA, the SIRTF planning tools, and the JSky Open Source Java library. The potential benefits of sharing code are clear. The recipient gains functionality for considerably less cost. The provider gains additional developers working with their code. If enough users groups adopt a set of common code and tools, defacto standards can emerge (as demonstrated by the success of the FITS standard). Code sharing also raises a number of challenges related to the management of the code. In this talk, we will review our experiences with SEA - both successes and failures - and offer some lessons learned that may promote further successes in collaboration and re-use.
NASA Technical Reports Server (NTRS)
Korathkar, Anuradha; Grosvenor, Sandy; Jones, Jeremy; Li, Connie; Mackey, Jennifer; Neher, Ken; Obenschain, Arthur F. (Technical Monitor)
2001-01-01
In the Virtual Observatory (VO), software tools will perform the functions that have traditionally been performed by physical observatories and their instruments. These tools will not be adjuncts to VO functionality but will make up the very core of the VO. Consequently, the tradition of observatory and system independent tools serving a small user base is not valid for the VO. For the VO to succeed, we must improve software collaboration and code sharing between projects and groups. A significant goal of the Scientist's Expert Assistant (SEA) project has been promoting effective collaboration and code sharing among groups. During the past three years, the SEA project has been developing prototypes for new observation planning software tools and strategies. Initially funded by the Next Generation Space Telescope, parts of the SEA code have since been adopted by the Space Telescope Science Institute. SEA has also supplied code for the SIRTF (Space Infrared Telescope Facility) planning tools, and the JSky Open Source Java library. The potential benefits of sharing code are clear. The recipient gains functionality for considerably less cost. The provider gains additional developers working with their code. If enough users groups adopt a set of common code and tools, de facto standards can emerge (as demonstrated by the success of the FITS standard). Code sharing also raises a number of challenges related to the management of the code. In this talk, we will review our experiences with SEA--both successes and failures, and offer some lessons learned that might promote further successes in collaboration and re-use.
Babor, Thomas F; Xuan, Ziming; Damon, Donna
2013-10-01
This study evaluated the use of a modified Delphi technique in combination with a previously developed alcohol advertising rating procedure to detect content violations in the U.S. Beer Institute Code. A related aim was to estimate the minimum number of raters needed to obtain reliable evaluations of code violations in television commercials. Six alcohol ads selected for their likelihood of having code violations were rated by community and expert participants (N = 286). Quantitative rating scales were used to measure the content of alcohol advertisements based on alcohol industry self-regulatory guidelines. The community group participants represented vulnerability characteristics that industry codes were designed to protect (e.g., age <21); experts represented various health-related professions, including public health, human development, alcohol research, and mental health. Alcohol ads were rated on 2 occasions separated by 1 month. After completing Time 1 ratings, participants were randomized to receive feedback from 1 group or the other. Findings indicate that (i) ratings at Time 2 had generally reduced variance, suggesting greater consensus after feedback, (ii) feedback from the expert group was more influential than that of the community group in developing group consensus, (iii) the expert group found significantly fewer violations than the community group, (iv) experts representing different professional backgrounds did not differ among themselves in the number of violations identified, and (v) a rating panel composed of at least 15 raters is sufficient to obtain reliable estimates of code violations. The Delphi technique facilitates consensus development around code violations in alcohol ad content and may enhance the ability of regulatory agencies to monitor the content of alcoholic beverage advertising when combined with psychometric-based rating procedures. Copyright © 2013 by the Research Society on Alcoholism.
Babor, Thomas F.; Xuan, Ziming; Damon, Donna
2013-01-01
Background This study evaluated the use of a modified Delphi technique in combination with a previously developed alcohol advertising rating procedure to detect content violations in the US Beer Institute code. A related aim was to estimate the minimum number of raters needed to obtain reliable evaluations of code violations in television commercials. Methods Six alcohol ads selected for their likelihood of having code violations were rated by community and expert participants (N=286). Quantitative rating scales were used to measure the content of alcohol advertisements based on alcohol industry self-regulatory guidelines. The community group participants represented vulnerability characteristics that industry codes were designed to protect (e.g., age < 21); experts represented various health-related professions, including public health, human development, alcohol research and mental health. Alcohol ads were rated on two occasions separated by one month. After completing Time 1 ratings, participants were randomized to receive feedback from one group or the other. Results Findings indicate that (1) ratings at Time 2 had generally reduced variance, suggesting greater consensus after feedback, (2) feedback from the expert group was more influential than that of the community group in developing group consensus, (3) the expert group found significantly fewer violations than the community group, (4) experts representing different professional backgrounds did not differ among themselves in the number of violations identified; (5) a rating panel composed of at least 15 raters is sufficient to obtain reliable estimates of code violations. Conclusions The Delphi Technique facilitates consensus development around code violations in alcohol ad content and may enhance the ability of regulatory agencies to monitor the content of alcoholic beverage advertising when combined with psychometric-based rating procedures. PMID:23682927
CLIPS: A tool for the development and delivery of expert systems
NASA Technical Reports Server (NTRS)
Riley, Gary
1991-01-01
The C Language Integrated Production System (CLIPS) is a forward chaining rule-based language developed by the Software Technology Branch at the Johnson Space Center. CLIPS provides a complete environment for the construction of rule-based expert systems. CLIPS was designed specifically to provide high probability, low cost, and easy integration with external systems. Other key features of CLIPS include a powerful rule syntax, an interactive development environment, high performance, extensibility, a verification/validation tool, extensive documentation, and source code availability. The current release of CLIPS, version 4.3, is being used by over 2,500 users throughout the public and private community including: all NASA sites and branches of the military, numerous Federal bureaus, government contractors, 140 universities, and many companies.
Rule-Based Flight Software Cost Estimation
NASA Technical Reports Server (NTRS)
Stukes, Sherry A.; Spagnuolo, John N. Jr.
2015-01-01
This paper discusses the fundamental process for the computation of Flight Software (FSW) cost estimates. This process has been incorporated in a rule-based expert system [1] that can be used for Independent Cost Estimates (ICEs), Proposals, and for the validation of Cost Analysis Data Requirements (CADRe) submissions. A high-level directed graph (referred to here as a decision graph) illustrates the steps taken in the production of these estimated costs and serves as a basis of design for the expert system described in this paper. Detailed discussions are subsequently given elaborating upon the methodology, tools, charts, and caveats related to the various nodes of the graph. We present general principles for the estimation of FSW using SEER-SEM as an illustration of these principles when appropriate. Since Source Lines of Code (SLOC) is a major cost driver, a discussion of various SLOC data sources for the preparation of the estimates is given together with an explanation of how contractor SLOC estimates compare with the SLOC estimates used by JPL. Obtaining consistency in code counting will be presented as well as factors used in reconciling SLOC estimates from different code counters. When sufficient data is obtained, a mapping into the JPL Work Breakdown Structure (WBS) from the SEER-SEM output is illustrated. For across the board FSW estimates, as was done for the NASA Discovery Mission proposal estimates performed at JPL, a comparative high-level summary sheet for all missions with the SLOC, data description, brief mission description and the most relevant SEER-SEM parameter values is given to illustrate an encapsulation of the used and calculated data involved in the estimates. The rule-based expert system described provides the user with inputs useful or sufficient to run generic cost estimation programs. This system's incarnation is achieved via the C Language Integrated Production System (CLIPS) and will be addressed at the end of this paper.
The challenge of mapping between two medical coding systems.
Wojcik, Barbara E; Stein, Catherine R; Devore, Raymond B; Hassell, L Harrison
2006-11-01
Deployable medical systems patient conditions (PCs) designate groups of patients with similar medical conditions and, therefore, similar treatment requirements. PCs are used by the U.S. military to estimate field medical resources needed in combat operations. Information associated with each of the 389 PCs is based on subject matter expert opinion, instead of direct derivation from standard medical codes. Currently, no mechanisms exist to tie current or historical medical data to PCs. Our study objective was to determine whether reliable conversion between PC codes and International Classification of Diseases, 9th Revision, Clinical Modification (ICD-9-CM) diagnosis codes is possible. Data were analyzed for three professional coders assigning all applicable ICD-9-CM diagnosis codes to each PC code. Inter-rater reliability was measured by using Cohen's K statistic and percent agreement. Methods were developed to calculate kappa statistics when multiple responses could be selected from many possible categories. Overall, we found moderate support for the possibility of reliable conversion between PCs and ICD-9-CM diagnoses (mean kappa = 0.61). Current PCs should be modified into a system that is verifiable with real data.
Expert system decision support for low-cost launch vehicle operations
NASA Technical Reports Server (NTRS)
Szatkowski, G. P.; Levin, Barry E.
1991-01-01
Progress in assessing the feasibility, benefits, and risks associated with AI expert systems applied to low cost expendable launch vehicle systems is described. Part one identified potential application areas in vehicle operations and on-board functions, assessed measures of cost benefit, and identified key technologies to aid in the implementation of decision support systems in this environment. Part two of the program began the development of prototypes to demonstrate real-time vehicle checkout with controller and diagnostic/analysis intelligent systems and to gather true measures of cost savings vs. conventional software, verification and validation requirements, and maintainability improvement. The main objective of the expert advanced development projects was to provide a robust intelligent system for control/analysis that must be performed within a specified real-time window in order to meet the demands of the given application. The efforts to develop the two prototypes are described. Prime emphasis was on a controller expert system to show real-time performance in a cryogenic propellant loading application and safety validation implementation of this system experimentally, using commercial-off-the-shelf software tools and object oriented programming techniques. This smart ground support equipment prototype is based in C with imbedded expert system rules written in the CLIPS protocol. The relational database, ORACLE, provides non-real-time data support. The second demonstration develops the vehicle/ground intelligent automation concept, from phase one, to show cooperation between multiple expert systems. This automated test conductor (ATC) prototype utilizes a knowledge-bus approach for intelligent information processing by use of virtual sensors and blackboards to solve complex problems. It incorporates distributed processing of real-time data and object-oriented techniques for command, configuration control, and auto-code generation.
Probabilistic Structural Analysis Methods (PSAM) for select space propulsion system components
NASA Technical Reports Server (NTRS)
1991-01-01
This annual report summarizes the work completed during the third year of technical effort on the referenced contract. Principal developments continue to focus on the Probabilistic Finite Element Method (PFEM) which has been under development for three years. Essentially all of the linear capabilities within the PFEM code are in place. Major progress in the application or verifications phase was achieved. An EXPERT module architecture was designed and partially implemented. EXPERT is a user interface module which incorporates an expert system shell for the implementation of a rule-based interface utilizing the experience and expertise of the user community. The Fast Probability Integration (FPI) Algorithm continues to demonstrate outstanding performance characteristics for the integration of probability density functions for multiple variables. Additionally, an enhanced Monte Carlo simulation algorithm was developed and demonstrated for a variety of numerical strategies.
Sollie, Annet; Sijmons, Rolf H; Lindhout, Dick; van der Ploeg, Ans T; Rubio Gozalbo, M Estela; Smit, G Peter A; Verheijen, Frans; Waterham, Hans R; van Weely, Sonja; Wijburg, Frits A; Wijburg, Rudolph; Visser, Gepke
2013-07-01
Data sharing is essential for a better understanding of genetic disorders. Good phenotype coding plays a key role in this process. Unfortunately, the two most widely used coding systems in medicine, ICD-10 and SNOMED-CT, lack information necessary for the detailed classification and annotation of rare and genetic disorders. This prevents the optimal registration of such patients in databases and thus data-sharing efforts. To improve care and to facilitate research for patients with metabolic disorders, we developed a new coding system for metabolic diseases with a dedicated group of clinical specialists. Next, we compared the resulting codes with those in ICD and SNOMED-CT. No matches were found in 76% of cases in ICD-10 and in 54% in SNOMED-CT. We conclude that there are sizable gaps in the SNOMED-CT and ICD coding systems for metabolic disorders. There may be similar gaps for other classes of rare and genetic disorders. We have demonstrated that expert groups can help in addressing such coding issues. Our coding system has been made available to the ICD and SNOMED-CT organizations as well as to the Orphanet and HPO organizations for further public application and updates will be published online (www.ddrmd.nl and www.cineas.org). © 2013 WILEY PERIODICALS, INC.
Collision count in rugby union: A comparison of micro-technology and video analysis methods.
Reardon, Cillian; Tobin, Daniel P; Tierney, Peter; Delahunt, Eamonn
2017-10-01
The aim of our study was to determine if there is a role for manipulation of g force thresholds acquired via micro-technology for accurately detecting collisions in rugby union. In total, 36 players were recruited from an elite Guinness Pro12 rugby union team. Player movement profiles and collisions were acquired via individual global positioning system (GPS) micro-technology units. Players were assigned to a sub-category of positions in order to determine positional collision demands. The coding of collisions by micro-technology at g force thresholds between 2 and 5.5 g (0.5 g increments) was compared with collision coding by an expert video analyst using Bland-Altman assessments. The most appropriate g force threshold (smallest mean difference compared with video analyst coding) was lower for all forwards positions (2.5 g) than for all backs positions (3.5 g). The Bland-Altman 95% limits of agreement indicated that there may be a substantial over- or underestimation of collisions coded via GPS micro-technology when using expert video analyst coding as the reference comparator. The manipulation of the g force thresholds applied to data acquired by GPS micro-technology units based on incremental thresholds of 0.5 g does not provide a reliable tool for the accurate coding of collisions in rugby union. Future research should aim to investigate smaller g force threshold increments and determine the events that cause coding of false positives.
C Language Integrated Production System, Ada Version
NASA Technical Reports Server (NTRS)
Culbert, Chris; Riley, Gary; Savely, Robert T.; Melebeck, Clovis J.; White, Wesley A.; Mcgregor, Terry L.; Ferguson, Melisa; Razavipour, Reza
1992-01-01
CLIPS/Ada provides capabilities of CLIPS v4.3 but uses Ada as source language for CLIPS executable code. Implements forward-chaining rule-based language. Program contains inference engine and language syntax providing framework for construction of expert-system program. Also includes features for debugging application program. Based on Rete algorithm which provides efficient method for performing repeated matching of patterns. Written in Ada.
Hou, I-Ching; Chang, Polun; Chan, Hui-Ya; Dykes, Patricia C
2013-05-01
Standardized terminology is an important infrastructure component of the electronic health record. ICNP(®) is a systemic coding system that can support the development of nursing information systems. Translation of the standardized terminology preferred terms into local terms is an important first step in the translation process. The purpose of this case report is to describe the translation strategy used and challenges faced in translating ICNP(®) Version 2 preferred terms from English to traditional Chinese. A modified Delphi strategy using forward translation and expert consensus was conducted to facilitate semantic and cultural translation and validation of the ICNP(®) and to make the process generalizable. A nursing informatics expert completed the initial forward translation. Five nursing experts with rich clinical and academic experiences joined this process and validated the initial translation. The nursing experts' consensus was then used to finalize the traditional Chinese terms. A total of 1863 preferred terms from the ICNP(®) Version 2 were translated from English into traditional Chinese. Majority agreement from two or more nursing experts was achieved for 98.3% (n=1832) of the preferred term translations. Less than 2% (n=31) of terms had no majority agreement. Translation challenges include the following: (1) changes in code structure of preferred terms from the ICNP(®) β2 version to Verson 2, (2) inability to identify resources to complete the translation that fully met ICNP recommendations for terminology translators, (3) ambiguous preferred term descriptions, and (4) ambiguous preferred term names. Most of the ICNP(®) Version 2 preferred terms were translated from English into traditional Chinese with majority consensus. For the terms without consensus, we recommend that all synonyms be included in the ICNP(®) translation. In countries like Taiwan where nursing education occurs in English, it is recommended that English terms are displayed along with the translated official language to help nurses to interpret and use the terminology correctly. Crown Copyright © 2012. Published by Elsevier Ireland Ltd. All rights reserved.
An integrated knowledge system for wind tunnel testing - Project Engineers' Intelligent Assistant
NASA Technical Reports Server (NTRS)
Lo, Ching F.; Shi, George Z.; Hoyt, W. A.; Steinle, Frank W., Jr.
1993-01-01
The Project Engineers' Intelligent Assistant (PEIA) is an integrated knowledge system developed using artificial intelligence technology, including hypertext, expert systems, and dynamic user interfaces. This system integrates documents, engineering codes, databases, and knowledge from domain experts into an enriched hypermedia environment and was designed to assist project engineers in planning and conducting wind tunnel tests. PEIA is a modular system which consists of an intelligent user-interface, seven modules and an integrated tool facility. Hypermedia technology is discussed and the seven PEIA modules are described. System maintenance and updating is very easy due to the modular structure and the integrated tool facility provides user access to commercial software shells for documentation, reporting, or database updating. PEIA is expected to provide project engineers with technical information, increase efficiency and productivity, and provide a realistic tool for personnel training.
Piloting a Collaborative Web-Based System for Testing ICD-11.
Donada, Marc; Kostanjsek, Nenad; Della Mea, Vincenzo; Celik, Can; Jakob, Robert
2017-01-01
The 11th revision of the International Classification of Diseases (ICD-11), for the first time in ICD history, deployed web-based collaboration of experts and ICT tools. To ensure that ICD-11 is working well, it needs to be systematically field tested in different settings, across the world. This will be done by means of a number of experiments. In order to support its implementation, a web-based system (ICDfit) has been designed and developed. The present paper illustrates the current prototype of the system and its technical testing. the system has been designed according to WHO requirements, and implemented using PHP and MySQL. Then, a preliminary technical test has been designed and run in January 2016, involving 8 users. They had to carry out double coding, that is, coding case summaries with both ICD-10 and ICD-11, and answering quick questions on the coding difficulty. the 8 users coded 632 cases each, spending an average of 163 seconds per case. While we found an issue in the mechanism used to record coding times, no further issues were found. the proposed system seems to be technically adequate for supporting future ICD-11 testing.
How effective is the revised regulatory code for alcohol advertising in Australia?
Jones, Sandra C; Hall, Danika; Munro, Geoffrey
2008-01-01
Australia, like several other countries, has a self-regulatory approach to advertising. However, in recent years the effectiveness of the regulatory system has been questioned, and there have been increasing public calls for an overhaul of the system. Following a formal review in 2003, the Ministerial Council on Drug Strategy proposed a revised Alcoholic Beverages Advertising Code (ABAC), which came into operation in 2004. The purpose of the present study was to examine the effectiveness of this revised system. From May 2004 until March 2005 television and magazine advertising campaigns were monitored for alcohol products. Over this period 14 complaints against alcohol advertisements were lodged with the self-regulatory board, and the authors recruited an independent expert panel to assess the advertisements and complaints. In eight of the 14 cases a majority of the judges perceived the advertisement to be in breach of the code, and in no cases did a majority perceive no breach. Conversely, however, none of the complaints were upheld by the Advertising Standards Board (ASB) and only one by the ABAC Panel. The results of this study suggest that the decisions made by the ASB in relation to complaints against alcohol advertisements are not in harmony with the judgement of independent experts, and that the ASB may not be performing an adequate job of representing community standards or protecting the community from offensive or inappropriate advertisements. Further, it appears that the revisions to the ABAC code, and associated processes, have not reduced the problems associated with alcohol advertising in Australia.
McEvoy, Matthew D.; Smalley, Jeremy C.; Nietert, Paul J.; Field, Larry C.; Furse, Cory M.; Blenko, John W.; Cobb, Benjamin G.; Walters, Jenna L.; Pendarvis, Allen; Dalal, Nishita S.; Schaefer, John J.
2012-01-01
Introduction Defining valid, reliable, defensible, and generalizable standards for the evaluation of learner performance is a key issue in assessing both baseline competence and mastery in medical education. However, prior to setting these standards of performance, the reliability of the scores yielding from a grading tool must be assessed. Accordingly, the purpose of this study was to assess the reliability of scores generated from a set of grading checklists used by non-expert raters during simulations of American Heart Association (AHA) MegaCodes. Methods The reliability of scores generated from a detailed set of checklists, when used by four non-expert raters, was tested by grading team leader performance in eight MegaCode scenarios. Videos of the scenarios were reviewed and rated by trained faculty facilitators and by a group of non-expert raters. The videos were reviewed “continuously” and “with pauses.” Two content experts served as the reference standard for grading, and four non-expert raters were used to test the reliability of the checklists. Results Our results demonstrate that non-expert raters are able to produce reliable grades when using the checklists under consideration, demonstrating excellent intra-rater reliability and agreement with a reference standard. The results also demonstrate that non-expert raters can be trained in the proper use of the checklist in a short amount of time, with no discernible learning curve thereafter. Finally, our results show that a single trained rater can achieve reliable scores of team leader performance during AHA MegaCodes when using our checklist in continuous mode, as measures of agreement in total scoring were very strong (Lin’s Concordance Correlation Coefficient = 0.96; Intraclass Correlation Coefficient = 0.97). Discussion We have shown that our checklists can yield reliable scores, are appropriate for use by non-expert raters, and are able to be employed during continuous assessment of team leader performance during the review of a simulated MegaCode. This checklist may be more appropriate for use by Advanced Cardiac Life Support (ACLS) instructors during MegaCode assessments than current tools provided by the AHA. PMID:22863996
Georgsson, Mattias; Kushniruk, Andre
2016-01-01
The cognitive walkthrough (CW) is a task-based, expert inspection usability evaluation method involving benefits such as cost effectiveness and efficiency. A drawback of the method is that it doesn't involve the user perspective from real users but instead is based on experts' predictions about the usability of the system and how users interact. In this paper, we propose a way of involving the user in an expert evaluation method by modifying the CW with patient groups as mediators. This along with other modifications include a dual domain session facilitator, specific patient groups and three different phases: 1) a preparation phase where suitable tasks are developed by a panel of experts and patients, validated through the content validity index 2) a patient user evaluation phase including an individual and collaborative process part 3) an analysis and coding phase where all data is digitalized and synthesized making use of Qualitative Data Analysis Software (QDAS) to determine usability deficiencies. We predict that this way of evaluating will utilize the benefits of the expert methods, also providing a way of including the patient user of these self-management systems. Results from this prospective study should provide evidence of the usefulness of this method modification.
NASA Technical Reports Server (NTRS)
Fink, Pamela K.; Palmer, Karol K.
1988-01-01
The development of a probabilistic structural analysis methodology (PSAM) is described. In the near-term, the methodology will be applied to designing critical components of the next generation space shuttle main engine. In the long-term, PSAM will be applied very broadly, providing designers with a new technology for more effective design of structures whose character and performance are significantly affected by random variables. The software under development to implement the ideas developed in PSAM resembles, in many ways, conventional deterministic structural analysis code. However, several additional capabilities regarding the probabilistic analysis makes the input data requirements and the resulting output even more complex. As a result, an intelligent front- and back-end to the code is being developed to assist the design engineer in providing the input data in a correct and appropriate manner. The type of knowledge that this entails is, in general, heuristically-based, allowing the fairly well-understood technology of production rules to apply with little difficulty. However, the PSAM code, called NESSUS, is written in FORTRAN-77 and runs on a DEC VAX. Thus, the associated expert system, called NESSUS/EXPERT, must run on a DEC VAX as well, and integrate effectively and efficiently with the existing FORTRAN code. This paper discusses the process undergone to select a suitable tool, identify an appropriate division between the functions that should be performed in FORTRAN and those that should be performed by production rules, and how integration of the conventional and AI technologies was achieved.
Catching on it early: Bodily and brain anticipatory mechanisms for excellence in sport.
Abreu, Ana M; Candidi, Matteo; Aglioti, Salvatore M
2017-01-01
Programming and executing a subsequent move is inherently linked to the ability to anticipate the actions of others when interacting. Such fundamental social ability is particularly important in sport. Here, we discuss the possible mechanisms behind the highly sophisticated anticipation skills that characterize experts. We contend that prediction in sports might rely on a finely tuned perceptual system that endows experts with a fast, partially unconscious, pickup of relevant cues. Furthermore, we discuss the role of the multimodal, perceptuomotor, multiple-duty cells (mirror neurons) that play an important function in action anticipation by means of an inner motor simulation process. Finally, we suggest the role of predictive coding, interoception, and the enteric nervous system as the processual and biological support for intuition and "gut feelings" in sports-the missing link that might explain outstanding expert performance based on action anticipation. © 2017 Elsevier B.V. All rights reserved.
Diagnostics aid for mass spectrometer trouble-shooting
NASA Astrophysics Data System (ADS)
Filby, E. E.; Rankin, R. A.; Webb, G. W.
The MS Expert system provides problem diagnostics for instruments used in the Mass Spectrometry Laboratory (MSL). The most critical results generated on these mass spectrometers are the uranium concentration and isotopic content data used for process control and materials accountability at the Idaho Chemical Processing Plant. The two purposes of the system are: (1) to minimize instrument downtime and thereby provide the best possible support to the Plant, and (2) to improve long-term data quality. This system combines the knowledge of several experts on mass spectrometry to provide a diagnostic tool, and can make these skills available on a more timely basis. It integrates code written in the Pascal language with a knowledge base entered into a commercial expert system shell. The user performs some preliminary status checks, and then selects from among several broad diagnostic categories. These initial steps provide input to the rule base. The overall analysis provides the user with a set of possible solutions to the observed problems, graded as to their probabilities. Besides the trouble-shooting benefits expected from this system, it will also provide structures diagnostic training for lab personnel. In addition, development of the system knowledge base has already produced a better understanding of instrument behavior. Two key findings are that a good user interface is necessary for full acceptance of the tool, and a development system should include standard programming capabilities as well as the expert system shell.
Assessment on EXPERT Descent and Landing System Aerodynamics
NASA Astrophysics Data System (ADS)
Wong, H.; Muylaert, J.; Northey, D.; Riley, D.
2009-01-01
EXPERT is a re-entry vehicle designed for validation of aero-thermodynamic models, numerical schemes in Computational Fluid Dynamics codes and test facilities for measuring flight data under an Earth re-entry environment. This paper addresses the design for the descent and landing sequence for EXPERT. It includes the descent sequence, the choice of drogue and main parachutes, and the parachute deployment condition, which can be supersonic or subsonic. The analysis is based mainly on an engineering tool, PASDA, together with some hand calculations for parachute sizing and design. The tool consists of a detailed 6-DoF simulation performed with the aerodynamics database of the vehicle, an empirical wakes model and the International Standard Atmosphere database. The aerodynamics database for the vehicle is generated by DNW experimental data and CFD codes within the framework of an ESA contract to CIRA. The analysis will be presented in terms of altitude, velocity, accelerations, angle-of- attack, pitch angle and angle of rigging line. Discussion on the advantages and disadvantages of each parachute deployment condition is included in addition to some comparison with the available data based on a Monte-Carlo method from a Russian company, FSUE NIIPS. Sensitivity on wind speed to the performance of EXPERT is shown to be strong. Supersonic deployment of drogue shows a better performance in stability at the expense of a larger G-load than those from the subsonic deployment of drogue. Further optimization on the parachute design is necessary in order to fulfill all the EXPERT specifications.
An expert system for wind shear avoidance
NASA Technical Reports Server (NTRS)
Stengel, Robert F.; Stratton, D. Alexander
1990-01-01
The principal objectives are to develop methods for assessing the likelihood of wind shear encounter (based on real-time information in the cockpit), for deciding what flight path to pursue (e.g., takeoff abort, landing go-around, or normal climbout or glide slope), and for using the aircraft's full potential for combating wind shear. This study requires the definition of both deterministic and statistical techniques for fusing internal and external information, for making go/no-go decisions, and for generating commands to the aircraft's autopilot and flight directors for both automatic and manually controlled flight. The expert system for pilot aiding is based on the results of the FAA Windshear Training Aids Program, a two-volume manual that presents an overview, pilot guide, training program, and substantiating data that provides guidelines for this initial development. The Windshear Safety Advisor expert system currently contains over 140 rules and is coded in the LISP programming language for implementation on a Symbolics 3670 LISP Machine.
Numerical, analytical, experimental study of fluid dynamic forces in seals
NASA Technical Reports Server (NTRS)
Shapiro, William; Artiles, Antonio; Aggarwal, Bharat; Walowit, Jed; Athavale, Mahesh M.; Preskwas, Andrzej J.
1992-01-01
NASA/Lewis Research Center is sponsoring a program for providing computer codes for analyzing and designing turbomachinery seals for future aerospace and engine systems. The program is made up of three principal components: (1) the development of advanced three dimensional (3-D) computational fluid dynamics codes, (2) the production of simpler two dimensional (2-D) industrial codes, and (3) the development of a knowledge based system (KBS) that contains an expert system to assist in seal selection and design. The first task has been to concentrate on cylindrical geometries with straight, tapered, and stepped bores. Improvements have been made by adoption of a colocated grid formulation, incorporation of higher order, time accurate schemes for transient analysis and high order discretization schemes for spatial derivatives. This report describes the mathematical formulations and presents a variety of 2-D results, including labyrinth and brush seal flows. Extensions of 3-D are presently in progress.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Little, M.P.; Muirhead, C.R.; Goossens, L.H.J.
1997-12-01
The development of two new probabilistic accident consequence codes, MACCS and COSYMA, was completed in 1990. These codes estimate the consequence from the accidental releases of radiological material from hypothesized accidents at nuclear installations. In 1991, the US Nuclear Regulatory Commission and the Commission of the European Communities began cosponsoring a joint uncertainty analysis of the two codes. The ultimate objective of this joint effort was to systematically develop credible and traceable uncertainty distributions for the respective code input variables. A formal expert judgment elicitation and evaluation process was identified as the best technology available for developing a library ofmore » uncertainty distributions for these consequence parameters. This report focuses on the results of the study to develop distribution for variables related to the MACCS and COSYMA late health effects models. This volume contains appendices that include (1) a summary of the MACCS and COSYMA consequence codes, (2) the elicitation questionnaires and case structures, (3) the rationales and results for the expert panel on late health effects, (4) short biographies of the experts, and (5) the aggregated results of their responses.« less
Semantic enrichment of medical forms - semi-automated coding of ODM-elements via web services.
Breil, Bernhard; Watermann, Andreas; Haas, Peter; Dziuballe, Philipp; Dugas, Martin
2012-01-01
Semantic interoperability is an unsolved problem which occurs while working with medical forms from different information systems or institutions. Standards like ODM or CDA assure structural homogenization but in order to compare elements from different data models it is necessary to use semantic concepts and codes on an item level of those structures. We developed and implemented a web-based tool which enables a domain expert to perform semi-automated coding of ODM-files. For each item it is possible to inquire web services which result in unique concept codes without leaving the context of the document. Although it was not feasible to perform a totally automated coding we have implemented a dialog based method to perform an efficient coding of all data elements in the context of the whole document. The proportion of codable items was comparable to results from previous studies.
Composite load spectra for select space propulsion structural components
NASA Technical Reports Server (NTRS)
Newell, James F.; Ho, Hing W.
1991-01-01
This report summarizes the development for: (1) correlation fields; (2) applications to liquid oxygen post; (3) models for pressure fluctuatios and vibration loads fluctuations; (4) additions to expert systems; and (5) scaling criteria. Implementation to computer code is also described. Demonstration sample cases are included with additional applications to engine duct and pipe bend.
A real-time expert system for self-repairing flight control
NASA Technical Reports Server (NTRS)
Gaither, S. A.; Agarwal, A. K.; Shah, S. C.; Duke, E. L.
1989-01-01
An integrated environment for specifying, prototyping, and implementing a self-repairing flight-control (SRFC) strategy is described. At an interactive workstation, the user can select paradigms such as rule-based expert systems, state-transition diagrams, and signal-flow graphs and hierarchically nest them, assign timing and priority attributes, establish blackboard-type communication, and specify concurrent execution on single or multiple processors. High-fidelity nonlinear simulations of aircraft and SRFC systems can be performed off-line, with the possibility of changing SRFC rules, inference strategies, and other heuristics to correct for control deficiencies. Finally, the off-line-generated SRFC can be transformed into highly optimized application-specific real-time C-language code. An application of this environment to the design of aircraft fault detection, isolation, and accommodation algorithms is presented in detail.
Benchmarks of programming languages for special purposes in the space station
NASA Technical Reports Server (NTRS)
Knoebel, Arthur
1986-01-01
Although Ada is likely to be chosen as the principal programming language for the Space Station, certain needs, such as expert systems and robotics, may be better developed in special languages. The languages, LISP and Prolog, are studied and some benchmarks derived. The mathematical foundations for these languages are reviewed. Likely areas of the space station are sought out where automation and robotics might be applicable. Benchmarks are designed which are functional, mathematical, relational, and expert in nature. The coding will depend on the particular versions of the languages which become available for testing.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Harper, F.T.; Young, M.L.; Miller, L.A.
The development of two new probabilistic accident consequence codes, MACCS and COSYMA, completed in 1990, estimate the risks presented by nuclear installations based on postulated frequencies and magnitudes of potential accidents. In 1991, the US Nuclear Regulatory Commission (NRC) and the Commission of the European Communities (CEC) began a joint uncertainty analysis of the two codes. The objective was to develop credible and traceable uncertainty distributions for the input variables of the codes. Expert elicitation, developed independently, was identified as the best technology available for developing a library of uncertainty distributions for the selected consequence parameters. The study was formulatedmore » jointly and was limited to the current code models and to physical quantities that could be measured in experiments. To validate the distributions generated for the wet deposition input variables, samples were taken from these distributions and propagated through the wet deposition code model along with the Gaussian plume model (GPM) implemented in the MACCS and COSYMA codes. Resulting distributions closely replicated the aggregated elicited wet deposition distributions. Project teams from the NRC and CEC cooperated successfully to develop and implement a unified process for the elaboration of uncertainty distributions on consequence code input parameters. Formal expert judgment elicitation proved valuable for synthesizing the best available information. Distributions on measurable atmospheric dispersion and deposition parameters were successfully elicited from experts involved in the many phenomenological areas of consequence analysis. This volume is the second of a three-volume document describing the project and contains two appendices describing the rationales for the dispersion and deposition data along with short biographies of the 16 experts who participated in the project.« less
[Deontology of the medical expert].
Raszeja, S
1995-09-01
The authority of prosecuting organ to choose the expert, set his task and verify the following opinion is defined. The qualities of the medical expert and his duties are described, referring to: -his expertise; -his morality; -his ability to issue an independent (objective) opinion. Detailed rules, which can be ascribed to a specific medical expert's deontological code, are listed and explained.
PS: A nonprocedural language with data types and modules
NASA Technical Reports Server (NTRS)
Gokhale, M. B.
1986-01-01
The Problem Specification (PS) nonprocedural language is a very high level language for algorithm specification. PS is suitable for nonprogrammers, who can specify a problem using mathematically-oriented equations; for expert programmers, who can prototype different versions of a software system for evaluation; and for those who wish to use specifications for portions (if not all) of a program. PS has data types and modules similar to Modula-2. The compiler generates C code. PS is first shown by example, and then efficiency issues in scheduling and code generation are discussed.
Advanced software development workstation project: Engineering scripting language. Graphical editor
NASA Technical Reports Server (NTRS)
1992-01-01
Software development is widely considered to be a bottleneck in the development of complex systems, both in terms of development and in terms of maintenance of deployed systems. Cost of software development and maintenance can also be very high. One approach to reducing costs and relieving this bottleneck is increasing the reuse of software designs and software components. A method for achieving such reuse is a software parts composition system. Such a system consists of a language for modeling software parts and their interfaces, a catalog of existing parts, an editor for combining parts, and a code generator that takes a specification and generates code for that application in the target language. The Advanced Software Development Workstation is intended to be an expert system shell designed to provide the capabilities of a software part composition system.
Development of a data reduction expert assistant
NASA Technical Reports Server (NTRS)
Miller, Glenn E.
1994-01-01
This report documents the development and deployment of the Data Reduction Expert Assistant (DRACO). The system was successfully applied to two astronomical research projects. The first was the removal of cosmic ray artifacts from Hubble Space Telescope (HST) Wide Field Planetary Camera data. The second was the reduction and calibration of low-dispersion CCD spectra taken from a ground-based telescope. This has validated our basic approach and demonstrated the applicability of this technology. This work has been made available to the scientific community in two ways. First, we have published the work in the scientific literature and presented papers at relevant conferences. Secondly, we have made the entire system (including documentation and source code) available to the community via the World Wide Web.
Transformation reborn: A new generation expert system for planning HST operations
NASA Technical Reports Server (NTRS)
Gerb, Andrew
1991-01-01
The Transformation expert system (TRANS) converts proposals for astronomical observations with the Hubble Space Telescope (HST) into detailed observing plans. It encodes expert knowledge to solve problems faced in planning and commanding HST observations to enable their processing by the Science Operations Ground System (SOGS). Among these problems are determining an acceptable order of executing observations, grouping of observations to enhance efficiency and schedulability, inserting extra observations when necessary, and providing parameters for commanding HST instruments. TRANS is currently an operational system and plays a critical role in the HST ground system. It was originally designed using forward-chaining provided by the OPS5 expert system language, but has been reimplemented using a procedural knowledge base. This reimplementation was forced by the explosion in the amount of OPS5 code required to specify the increasingly complicated situations requiring expert-level intervention by the TRANS knowledge base. This problem was compounded by the difficulty of avoiding unintended interaction between rules. To support the TRANS knowledge base, XCL, a small but powerful extension to Commom Lisp was implemented. XCL allows a compact syntax for specifying assignments and references to object attributes. XCL also allows the capability to iterate over objects and perform keyed lookup. The reimplementation of TRANS has greatly diminished the effort needed to maintain and enhance it. As a result of this, its functions have been expanded to include warnings about observations that are difficult or impossible to schedule or command, providing data to aid SPIKE, an intelligent planning system used for HST long-term scheduling, and providing information to the Guide Star Selection System (GSSS) to aid in determination of the long range availability of guide stars.
Wavelet-based compression of pathological images for telemedicine applications
NASA Astrophysics Data System (ADS)
Chen, Chang W.; Jiang, Jianfei; Zheng, Zhiyong; Wu, Xue G.; Yu, Lun
2000-05-01
In this paper, we present the performance evaluation of wavelet-based coding techniques as applied to the compression of pathological images for application in an Internet-based telemedicine system. We first study how well suited the wavelet-based coding is as it applies to the compression of pathological images, since these images often contain fine textures that are often critical to the diagnosis of potential diseases. We compare the wavelet-based compression with the DCT-based JPEG compression in the DICOM standard for medical imaging applications. Both objective and subjective measures have been studied in the evaluation of compression performance. These studies are performed in close collaboration with expert pathologists who have conducted the evaluation of the compressed pathological images and communication engineers and information scientists who designed the proposed telemedicine system. These performance evaluations have shown that the wavelet-based coding is suitable for the compression of various pathological images and can be integrated well with the Internet-based telemedicine systems. A prototype of the proposed telemedicine system has been developed in which the wavelet-based coding is adopted for the compression to achieve bandwidth efficient transmission and therefore speed up the communications between the remote terminal and the central server of the telemedicine system.
Organization of Programming Knowledge of Novices and Experts.
ERIC Educational Resources Information Center
Wiedenbeck, Susan
1986-01-01
Reports on an experiment where novice and expert programmers made decisions about Fortran code segments. The results show that, although expert programmers are better able to extract and use functional information, they don't differ significantly from novices in their ability to use syntactic concepts. (Author/EM)
Trombert-Paviot, B; Rodrigues, J M; Rogers, J E; Baud, R; van der Haring, E; Rassinoux, A M; Abrial, V; Clavel, L; Idir, H
2000-09-01
Generalised architecture for languages, encyclopedia and nomenclatures in medicine (GALEN) has developed a new generation of terminology tools based on a language independent model describing the semantics and allowing computer processing and multiple reuses as well as natural language understanding systems applications to facilitate the sharing and maintaining of consistent medical knowledge. During the European Union 4 Th. framework program project GALEN-IN-USE and later on within two contracts with the national health authorities we applied the modelling and the tools to the development of a new multipurpose coding system for surgical procedures named CCAM in a minority language country, France. On one hand, we contributed to a language independent knowledge repository and multilingual semantic dictionaries for multicultural Europe. On the other hand, we support the traditional process for creating a new coding system in medicine which is very much labour consuming by artificial intelligence tools using a medically oriented recursive ontology and natural language processing. We used an integrated software named CLAW (for classification workbench) to process French professional medical language rubrics produced by the national colleges of surgeons domain experts into intermediate dissections and to the Grail reference ontology model representation. From this language independent concept model representation, on one hand, we generate with the LNAT natural language generator controlled French natural language to support the finalization of the linguistic labels (first generation) in relation with the meanings of the conceptual system structure. On the other hand, the Claw classification manager proves to be very powerful to retrieve the initial domain experts rubrics list with different categories of concepts (second generation) within a semantic structured representation (third generation) bridge to the electronic patient record detailed terminology.
Novice to expert practice via postprofessional athletic training education: a grounded theory.
Neibert, Peter J
2009-01-01
To discover the theoretic constructs that confirm, disconfirm, or extend the principles and their applications appropriate for National Athletic Trainers' Association (NATA)-accredited postprofessional athletic training education programs. Interviews at the 2003 NATA Annual Meeting & Clinical Symposia. Qualitative study using grounded theory procedures. Thirteen interviews were conducted with postprofessional graduates. Participants were purposefully selected based on theoretic sampling and availability. The transcribed interviews were analyzed using open coding, axial coding, and selective coding procedures. Member checks, reflective journaling, and triangulation were used to ensure trustworthiness. The participants' comments confirmed and extended the current principles of postprofessional athletic training education programs and offered additional suggestions for more effective practical applications. The emergence of this central category of novice to expert practice is a paramount finding. The tightly woven fabric of the 10 processes, when interlaced with one another, provides a strong tapestry supporting novice to expert practice via postprofessional athletic training education. The emergence of this theoretic position pushes postprofessional graduate athletic training education forward to the future for further investigation into the theoretic constructs of novice to expert practice.
NASA Astrophysics Data System (ADS)
The present conference discusses topics in multiwavelength network technology and its applications, advanced digital radio systems in their propagation environment, mobile radio communications, switching programmability, advancements in computer communications, integrated-network management and security, HDTV and image processing in communications, basic exchange communications radio advancements in digital switching, intelligent network evolution, speech coding for telecommunications, and multiple access communications. Also discussed are network designs for quality assurance, recent progress in coherent optical systems, digital radio applications, advanced communications technologies for mobile users, communication software for switching systems, AI and expert systems in network management, intelligent multiplexing nodes, video and image coding, network protocols and performance, system methods in quality and reliability, the design and simulation of lightwave systems, local radio networks, mobile satellite communications systems, fiber networks restoration, packet video networks, human interfaces for future networks, and lightwave networking.
Scudder, Ashley; Herschell, Amy D
2015-08-01
In order to make EBTs available to a large number of children and families, developers and expert therapists have used their experience and expertise to train community-based therapists in EBTs. Understanding current training practices of treatment experts may be one method for establishing best practices for training community-based therapists prior to comprehensive empirical examinations of training practices. A qualitative study was conducted using surveys and phone interviews to identify the specific procedures used by treatment experts to train and implement an evidence-based treatment in community settings. Twenty-three doctoral-level, clinical psychologists were identified to participate because of their expertise in conducting and training Parent-Child Interaction Therapy. Semi-structured qualitative interviews were completed by phone, later transcribed verbatim, and analyzed using thematic coding. The de-identified data were coded by two independent qualitative data researchers and then compared for consistency of interpretation. The themes that emerged following the final coding were used to construct a training protocol to be empirically tested. The goal of this paper is to not only understand the current state of training practices for training therapists in a particular EBT, Parent-Child Interaction Therapy, but to illustrate the use of expert opinion as the best available evidence in preparation for empirical evaluation.
A Graphical User-Interface Development Tool for Intelligent Computer- Assisted Instruction Systems
1993-09-01
Wesley Publishing Co., 1991 [HEND 88] Hendler, James A., Expert Systems: The User Interface, Ablex Publishing Corporation, 1988 [WALK 87] Walker, Adrian...Shimeall Code CSSm Assistant Professor, Computer Science Department Naval Postgraduate School Monterey, CA 93943-5000 5. Kepala StafUmum ABRI Mabes ABRI...KASAU Mabes TNI-AU, JI. Gatot Subroto No. 72, Jakarta Timur, Indonesia 8. Diraeroau Mabes TNI-AU, J1. Gatot Subroto No. 72, Jakarta Timur, Indonesia 9
Zeng, Xiaoming; Bell, Paul D
2011-01-01
In this study, we report on a qualitative method known as the Delphi method, used in the first part of a research study for improving the accuracy and reliability of ICD-9-CM coding. A panel of independent coding experts interacted methodically to determine that the three criteria to identify a problematic ICD-9-CM subcategory for further study were cost, volume, and level of coding confusion caused. The Medicare Provider Analysis and Review (MEDPAR) 2007 fiscal year data set as well as suggestions from the experts were used to identify coding subcategories based on cost and volume data. Next, the panelists performed two rounds of independent ranking before identifying Excisional Debridement as the subcategory that causes the most confusion among coders. As a result, they recommended it for further study aimed at improving coding accuracy and variation. This framework can be adopted at different levels for similar studies in need of a schema for determining problematic subcategories of code sets. PMID:21796264
Variation of SNOMED CT coding of clinical research concepts among coding experts.
Andrews, James E; Richesson, Rachel L; Krischer, Jeffrey
2007-01-01
To compare consistency of coding among professional SNOMED CT coders representing three commercial providers of coding services when coding clinical research concepts with SNOMED CT. A sample of clinical research questions from case report forms (CRFs) generated by the NIH-funded Rare Disease Clinical Research Network (RDCRN) were sent to three coding companies with instructions to code the core concepts using SNOMED CT. The sample consisted of 319 question/answer pairs from 15 separate studies. The companies were asked to select SNOMED CT concepts (in any form, including post-coordinated) that capture the core concept(s) reflected in the question. Also, they were asked to state their level of certainty, as well as how precise they felt their coding was. Basic frequencies were calculated to determine raw level agreement among the companies and other descriptive information. Krippendorff's alpha was used to determine a statistical measure of agreement among the coding companies for several measures (semantic, certainty, and precision). No significant level of agreement among the experts was found. There is little semantic agreement in coding of clinical research data items across coders from 3 professional coding services, even using a very liberal definition of agreement.
Composite load spectra for select space propulsion structural components
NASA Technical Reports Server (NTRS)
Newell, J. F.; Kurth, R. E.; Ho, H.
1986-01-01
A multiyear program is performed with the objective to develop generic load models with multiple levels of progressive sophistication to simulate the composite (combined) load spectra that are induced in space propulsion system components, representative of Space Shuttle Main Engines (SSME), such as transfer ducts, turbine blades, and liquid oxygen (LOX) posts. Progress of the first year's effort includes completion of a sufficient portion of each task -- probabilistic models, code development, validation, and an initial operational code. This code has from its inception an expert system philosophy that could be added to throughout the program and in the future. The initial operational code is only applicable to turbine blade type loadings. The probabilistic model included in the operational code has fitting routines for loads that utilize a modified Discrete Probabilistic Distribution termed RASCAL, a barrier crossing method and a Monte Carlo method. An initial load model was developed by Battelle that is currently used for the slowly varying duty cycle type loading. The intent is to use the model and related codes essentially in the current form for all loads that are based on measured or calculated data that have followed a slowly varying profile.
Monte Carlo simulations in Nuclear Medicine
NASA Astrophysics Data System (ADS)
Loudos, George K.
2007-11-01
Molecular imaging technologies provide unique abilities to localise signs of disease before symptoms appear, assist in drug testing, optimize and personalize therapy, and assess the efficacy of treatment regimes for different types of cancer. Monte Carlo simulation packages are used as an important tool for the optimal design of detector systems. In addition they have demonstrated potential to improve image quality and acquisition protocols. Many general purpose (MCNP, Geant4, etc) or dedicated codes (SimSET etc) have been developed aiming to provide accurate and fast results. Special emphasis will be given to GATE toolkit. The GATE code currently under development by the OpenGATE collaboration is the most accurate and promising code for performing realistic simulations. The purpose of this article is to introduce the non expert reader to the current status of MC simulations in nuclear medicine and briefly provide examples of current simulated systems, and present future challenges that include simulation of clinical studies and dosimetry applications.
Position coding effects in a 2D scenario: the case of musical notation.
Perea, Manuel; García-Chamorro, Cristina; Centelles, Arnau; Jiménez, María
2013-07-01
How does the cognitive system encode the location of objects in a visual scene? In the past decade, this question has attracted much attention in the field of visual-word recognition (e.g., "jugde" is perceptually very close to "judge"). Letter transposition effects have been explained in terms of perceptual uncertainty or shared "open bigrams". In the present study, we focus on note position coding in music reading (i.e., a 2D scenario). The usual way to display music is the staff (i.e., a set of 5 horizontal lines and their resultant 4 spaces). When reading musical notation, it is critical to identify not only each note (temporal duration), but also its pitch (y-axis) and its temporal sequence (x-axis). To examine note position coding, we employed a same-different task in which two briefly and consecutively presented staves contained four notes. The experiment was conducted with experts (musicians) and non-experts (non-musicians). For the "different" trials, the critical conditions involved staves in which two internal notes that were switched vertically, horizontally, or fully transposed--as well as the appropriate control conditions. Results revealed that note position coding was only approximate at the early stages of processing and that this encoding process was modulated by expertise. We examine the implications of these findings for models of object position encoding. Copyright © 2013 Elsevier B.V. All rights reserved.
Maljovec, D.; Liu, S.; Wang, B.; ...
2015-07-14
Here, dynamic probabilistic risk assessment (DPRA) methodologies couple system simulator codes (e.g., RELAP and MELCOR) with simulation controller codes (e.g., RAVEN and ADAPT). Whereas system simulator codes model system dynamics deterministically, simulation controller codes introduce both deterministic (e.g., system control logic and operating procedures) and stochastic (e.g., component failures and parameter uncertainties) elements into the simulation. Typically, a DPRA is performed by sampling values of a set of parameters and simulating the system behavior for that specific set of parameter values. For complex systems, a major challenge in using DPRA methodologies is to analyze the large number of scenarios generated,more » where clustering techniques are typically employed to better organize and interpret the data. In this paper, we focus on the analysis of two nuclear simulation datasets that are part of the risk-informed safety margin characterization (RISMC) boiling water reactor (BWR) station blackout (SBO) case study. We provide the domain experts a software tool that encodes traditional and topological clustering techniques within an interactive analysis and visualization environment, for understanding the structures of such high-dimensional nuclear simulation datasets. We demonstrate through our case study that both types of clustering techniques complement each other for enhanced structural understanding of the data.« less
Guidelines and rules for automated assembly by robots in space
NASA Technical Reports Server (NTRS)
Srivastava, Sadanand
1992-01-01
The development of an expert system for a 'Mechanical Design System' is discussed. Two different implementation approaches are described. One is coded in C, and the other is realized by a software package - 'Exsys.' The first method has the advantage of greater flexibility and quicker responses, while the latter one is easier to develop. This report discusses the feasible ways to establish a real mechanical intelligent design system applying artificial intelligence techniques so that the products designed by this system could best meet the requirements for space assembly.
Expert vs. novice: Problem decomposition/recomposition in engineering design
NASA Astrophysics Data System (ADS)
Song, Ting
The purpose of this research was to investigate the differences of using problem decomposition and problem recomposition among dyads of engineering experts, dyads of engineering seniors, and dyads of engineering freshmen. Fifty participants took part in this study. Ten were engineering design experts, 20 were engineering seniors, and 20 were engineering freshmen. Participants worked in dyads to complete an engineering design challenge within an hour. The entire design process was video and audio recorded. After the design session, members participated in a group interview. This study used protocol analysis as the methodology. Video and audio data were transcribed, segmented, and coded. Two coding systems including the FBS ontology and "levels of the problem" were used in this study. A series of statistical techniques were used to analyze data. Interview data and participants' design sketches also worked as supplemental data to help answer the research questions. By analyzing the quantitative and qualitative data, it was found that students used less problem decomposition and problem recomposition than engineer experts in engineering design. This result implies that engineering education should place more importance on teaching problem decomposition and problem recomposition. Students were found to spend less cognitive effort when considering the problem as a whole and interactions between subsystems than engineer experts. In addition, students were also found to spend more cognitive effort when considering details of subsystems. These results showed that students tended to use dept-first decomposition and experts tended to use breadth-first decomposition in engineering design. The use of Function (F), Behavior (B), and Structure (S) among engineering experts, engineering seniors, and engineering freshmen was compared on three levels. Level 1 represents designers consider the problem as an integral whole, Level 2 represents designers consider interactions between subsystems, and Level 3 represents designers consider details of subsystems. The results showed that students used more S on Level 1 and 3 but they used less F on Level 1 than engineering experts. The results imply that engineering curriculum should improve the teaching of problem definition in engineering design because students need to understand the problem before solving it.
VICTORIA: A mechanistic model for radionuclide behavior in the reactor coolant system
DOE Office of Scientific and Technical Information (OSTI.GOV)
Schaperow, J.H.; Bixler, N.E.
1996-12-31
VICTORIA is the U.S. Nuclear Regulatory Commission`s (NRC`s) mechanistic, best-estimate code for analysis of fission product release from the core and subsequent transport in the reactor vessel and reactor coolant system. VICTORIA requires thermal-hydraulic data (i.e., temperatures, pressures, and velocities) as input. In the past, these data have been taken from the results of calculations from thermal-hydraulic codes such as SCDAP/RELAP5, MELCOR, and MAAP. Validation and assessment of VICTORIA 1.0 have been completed. An independent peer review of VICTORIA, directed by Brookhaven National Laboratory and supported by experts in the areas of fuel release, fission product chemistry, and aerosol physics,more » has been undertaken. This peer review, which will independently assess the code`s capabilities, is nearing completion with the peer review committee`s final report expected in Dec 1996. A limited amount of additional development is expected as a result of the peer review. Following this additional development, the NRC plans to release VICTORIA 1.1 and an updated and improved code manual. Future plans mainly involve use of the code for plant calculations to investigate specific safety issues as they arise. Also, the code will continue to be used in support of the Phebus experiments.« less
Pujar, Shashikant; O’Leary, Nuala A; Farrell, Catherine M; Mudge, Jonathan M; Wallin, Craig; Diekhans, Mark; Barnes, If; Bennett, Ruth; Berry, Andrew E; Cox, Eric; Davidson, Claire; Goldfarb, Tamara; Gonzalez, Jose M; Hunt, Toby; Jackson, John; Joardar, Vinita; Kay, Mike P; Kodali, Vamsi K; McAndrews, Monica; McGarvey, Kelly M; Murphy, Michael; Rajput, Bhanu; Rangwala, Sanjida H; Riddick, Lillian D; Seal, Ruth L; Webb, David; Zhu, Sophia; Aken, Bronwen L; Bult, Carol J; Frankish, Adam; Pruitt, Kim D
2018-01-01
Abstract The Consensus Coding Sequence (CCDS) project provides a dataset of protein-coding regions that are identically annotated on the human and mouse reference genome assembly in genome annotations produced independently by NCBI and the Ensembl group at EMBL-EBI. This dataset is the product of an international collaboration that includes NCBI, Ensembl, HUGO Gene Nomenclature Committee, Mouse Genome Informatics and University of California, Santa Cruz. Identically annotated coding regions, which are generated using an automated pipeline and pass multiple quality assurance checks, are assigned a stable and tracked identifier (CCDS ID). Additionally, coordinated manual review by expert curators from the CCDS collaboration helps in maintaining the integrity and high quality of the dataset. The CCDS data are available through an interactive web page (https://www.ncbi.nlm.nih.gov/CCDS/CcdsBrowse.cgi) and an FTP site (ftp://ftp.ncbi.nlm.nih.gov/pub/CCDS/). In this paper, we outline the ongoing work, growth and stability of the CCDS dataset and provide updates on new collaboration members and new features added to the CCDS user interface. We also present expert curation scenarios, with specific examples highlighting the importance of an accurate reference genome assembly and the crucial role played by input from the research community. PMID:29126148
Novice to Expert Practice via Postprofessional Athletic Training Education: A Grounded Theory
Neibert, Peter J
2009-01-01
Objective: To discover the theoretic constructs that confirm, disconfirm, or extend the principles and their applications appropriate for National Athletic Trainers' Association (NATA)–accredited postprofessional athletic training education programs. Design: Interviews at the 2003 NATA Annual Meeting & Clinical Symposia. Setting: Qualitative study using grounded theory procedures. Patients and Other Participants: Thirteen interviews were conducted with postprofessional graduates. Participants were purposefully selected based on theoretic sampling and availability. Data Collection and Analysis: The transcribed interviews were analyzed using open coding, axial coding, and selective coding procedures. Member checks, reflective journaling, and triangulation were used to ensure trustworthiness. Results: The participants' comments confirmed and extended the current principles of postprofessional athletic training education programs and offered additional suggestions for more effective practical applications. Conclusions: The emergence of this central category of novice to expert practice is a paramount finding. The tightly woven fabric of the 10 processes, when interlaced with one another, provides a strong tapestry supporting novice to expert practice via postprofessional athletic training education. The emergence of this theoretic position pushes postprofessional graduate athletic training education forward to the future for further investigation into the theoretic constructs of novice to expert practice. PMID:19593420
Quantitative Measures for Software Independent Verification and Validation
NASA Technical Reports Server (NTRS)
Lee, Alice
1996-01-01
As software is maintained or reused, it undergoes an evolution which tends to increase the overall complexity of the code. To understand the effects of this, we brought in statistics experts and leading researchers in software complexity, reliability, and their interrelationships. These experts' project has resulted in our ability to statistically correlate specific code complexity attributes, in orthogonal domains, to errors found over time in the HAL/S flight software which flies in the Space Shuttle. Although only a prototype-tools experiment, the result of this research appears to be extendable to all other NASA software, given appropriate data similar to that logged for the Shuttle onboard software. Our research has demonstrated that a more complete domain coverage can be mathematically demonstrated with the approach we have applied, thereby ensuring full insight into the cause-and-effects relationship between the complexity of a software system and the fault density of that system. By applying the operational profile we can characterize the dynamic effects of software path complexity under this same approach We now have the ability to measure specific attributes which have been statistically demonstrated to correlate to increased error probability, and to know which actions to take, for each complexity domain. Shuttle software verifiers can now monitor the changes in the software complexity, assess the added or decreased risk of software faults in modified code, and determine necessary corrections. The reports, tool documentation, user's guides, and new approach that have resulted from this research effort represent advances in the state of the art of software quality and reliability assurance. Details describing how to apply this technique to other NASA code are contained in this document.
[Consistency and Reliability of MDK Expertise Examining the Encoding in the German DRG System].
Gaertner, T; Lehr, F; Blum, B; van Essen, J
2015-09-01
Hospital inpatient stays are reimbursed on the basis of German diagnosis-related groups (G-DRG). The G-DRG classification system is based on complex coding guidelines. The Medical Review Board of the Statutory Health Insurance Funds (MDK) examines the encoding by hospitals and delivers individual expertises on behalf of the German statutory health insurance companies in cases in which irregularities are suspected. A study was conducted on the inter-rater reliability of the MDK expertises regarding the scope of the assessment. A representative sample of 212 MDK expertises was taken from a selected pool of 1 392 MDK expertises in May 2013. This representative sample underwent a double-examination by 2 independent MDK experts using a special software based on the 3MTM G-DRG Grouper 2013 of 3M Medica, Germany. The following items encoded by the hospitals were examined: DRG, principal diagnosis, secondary diagnoses, procedures and additional payments. It was analysed whether the results of MDK expertises were consistent, reliable and correct. 202 expertises were eligible for evaluation, containing a total of 254 questions regarding one or more of the 5 items encoded by hospitals. The double-examination by 2 independent MDK experts showed matching results in 187 questions (73.6%) meaning they had been examined consistently and correctly. 59 questions (23.2%) did not show matching results, nevertheless they had been examined correctly regarding the scope of the assessment. None of the principal diagnoses was significantly affected by inconsistent or wrong judgment. A representative sample of MDK expertises examining the DRG encoding by hospitals showed a very high percentage of correct examination by the MDK experts. Identical MDK expertises cannot be achieved in all cases due to the scope of the assessment. Further improvement and simplification of codes and coding guidelines are required to reduce the scope of assessment with regard to correct DRG encoding and its examination. © Georg Thieme Verlag KG Stuttgart · New York.
Zafirah, S A; Nur, Amrizal Muhammad; Puteh, Sharifa Ezat Wan; Aljunid, Syed Mohamed
2018-01-25
The accuracy of clinical coding is crucial in the assignment of Diagnosis Related Groups (DRGs) codes, especially if the hospital is using Casemix System as a tool for resource allocations and efficiency monitoring. The aim of this study was to estimate the potential loss of income due to an error in clinical coding during the implementation of the Malaysia Diagnosis Related Group (MY-DRG ® ) Casemix System in a teaching hospital in Malaysia. Four hundred and sixty-four (464) coded medical records were selected, re-examined and re-coded by an independent senior coder (ISC). This ISC re-examined and re-coded the error code that was originally entered by the hospital coders. The pre- and post-coding results were compared, and if there was any disagreement, the codes by the ISC were considered the accurate codes. The cases were then re-grouped using a MY-DRG ® grouper to assess and compare the changes in the DRG assignment and the hospital tariff assignment. The outcomes were then verified by a casemix expert. Coding errors were found in 89.4% (415/424) of the selected patient medical records. Coding errors in secondary diagnoses were the highest, at 81.3% (377/464), followed by secondary procedures at 58.2% (270/464), principal procedures of 50.9% (236/464) and primary diagnoses at 49.8% (231/464), respectively. The coding errors resulted in the assignment of different MY-DRG ® codes in 74.0% (307/415) of the cases. From this result, 52.1% (160/307) of the cases had a lower assigned hospital tariff. In total, the potential loss of income due to changes in the assignment of the MY-DRG ® code was RM654,303.91. The quality of coding is a crucial aspect in implementing casemix systems. Intensive re-training and the close monitoring of coder performance in the hospital should be performed to prevent the potential loss of hospital income.
Ertl, Lorenz; Christ, Frank
2007-08-01
Better quality bystander first-aid could improve outcome rates for emergency victims significantly. In this case-control study, we hypothesised that expert knowledge presented step-by-step to untrained helpers using a personal digital assistant (PDA), would improve the quality of bystanders basic life support. We confronted 101 lay-helpers with two standard emergency situations. (1) An unconscious trauma victim with severe bleeding. (2) Cardiopulmonary resuscitation (CPR). Performance was assessed using an Objective Structured Clinical Examination (OSCE). One group was supported by a PDA providing visual and audio instructions, whereas the control group acted only with their current knowledge. The expert system was programmed in HTML-code and displayed on the PDA's Internet browser. The maximum score obtainable was 24 points corresponding to optimal treatment. The control group without the PDA reached 14.8+/-3.5 (mean value+/-standard deviation), whereas the PDA supported group scored significantly higher (21.9+/-2.7, p<0.01). The difference in performance was measurable in all criteria tested and particularly notable in the items: placing in recovery position, airway management and quality of CPR. The PDA based expert system increased the performance of untrained helpers supplying emergency care significantly. Since Internet compatible mobile devices have become widely available, a significant quality improvement in bystander first-aid seems possible.
NASA Astrophysics Data System (ADS)
Shahriari, D.; Zolfaghari, A.; Masoumi, F.
2011-01-01
Nondestructive evaluation is explained as nondestructive testing, nondestructive inspection, and nondestructive examination. It is a desire to determine some characteristic of the object or to determine whether the object contains irregularities, discontinuities, or flaws. Ultrasound based inspection techniques are used extensively throughout industry for detection of flaws in engineering materials. The range and variety of imperfections encountered is large, and critical assessment of location, size, orientation and type is often difficult. In addition, increasing quality requirements of new standards and codes of practice relating to fitness for purpose are placing higher demands on operators. Applying of an expert knowledge-based analysis in ultrasonic examination is a powerful tool that can help assure safety, quality, and reliability; increase productivity; decrease liability; and save money. In this research, an expert module system is coupled with ultrasonic examination (A-Scan Procedure) to determine and evaluate type and location of flaws that embedded during welding parts. The processing module of this expert system is implemented based on EN standard to classify welding defects, acceptance condition and measuring of their location via echo static pattern and image processing. The designed module introduces new system that can automate evaluating of the results of A-scan method according to EN standard. It can simultaneously recognize the number and type of defects, and determine flaw position during each scan.
Automated software development workstation
NASA Technical Reports Server (NTRS)
1986-01-01
Engineering software development was automated using an expert system (rule-based) approach. The use of this technology offers benefits not available from current software development and maintenance methodologies. A workstation was built with a library or program data base with methods for browsing the designs stored; a system for graphical specification of designs including a capability for hierarchical refinement and definition in a graphical design system; and an automated code generation capability in FORTRAN. The workstation was then used in a demonstration with examples from an attitude control subsystem design for the space station. Documentation and recommendations are presented.
SER assistant: An expert system for safety evaluation reports
DOE Office of Scientific and Technical Information (OSTI.GOV)
DeChaine, M.D.; Levine, S.H.; Feltus, M.A.
1993-01-01
The SER Assistant is an expert system that assists engineers to write safety evaluation reports (SERs). Section 50.59 of the Code of Federal Regulations allows modifications to be made to nuclear power plants without prior US Nuclear Regulatory Commission approval if two conditions are satisfied. First, the change must not affect the technical specifications of the plant. Second, the modification must not affect a part of the plant described in the final safety analysis report, or if it does, it must not create an unreviewed safety question. The purpose of an SER is to ensure that these conditions are satisfiedmore » for the proposed modification. The SER Assistant aids this process by providing relevant, but directed, questions and information as well as giving engineers an organized environment to document their thought processes.« less
Pujar, Shashikant; O'Leary, Nuala A; Farrell, Catherine M; Loveland, Jane E; Mudge, Jonathan M; Wallin, Craig; Girón, Carlos G; Diekhans, Mark; Barnes, If; Bennett, Ruth; Berry, Andrew E; Cox, Eric; Davidson, Claire; Goldfarb, Tamara; Gonzalez, Jose M; Hunt, Toby; Jackson, John; Joardar, Vinita; Kay, Mike P; Kodali, Vamsi K; Martin, Fergal J; McAndrews, Monica; McGarvey, Kelly M; Murphy, Michael; Rajput, Bhanu; Rangwala, Sanjida H; Riddick, Lillian D; Seal, Ruth L; Suner, Marie-Marthe; Webb, David; Zhu, Sophia; Aken, Bronwen L; Bruford, Elspeth A; Bult, Carol J; Frankish, Adam; Murphy, Terence; Pruitt, Kim D
2018-01-04
The Consensus Coding Sequence (CCDS) project provides a dataset of protein-coding regions that are identically annotated on the human and mouse reference genome assembly in genome annotations produced independently by NCBI and the Ensembl group at EMBL-EBI. This dataset is the product of an international collaboration that includes NCBI, Ensembl, HUGO Gene Nomenclature Committee, Mouse Genome Informatics and University of California, Santa Cruz. Identically annotated coding regions, which are generated using an automated pipeline and pass multiple quality assurance checks, are assigned a stable and tracked identifier (CCDS ID). Additionally, coordinated manual review by expert curators from the CCDS collaboration helps in maintaining the integrity and high quality of the dataset. The CCDS data are available through an interactive web page (https://www.ncbi.nlm.nih.gov/CCDS/CcdsBrowse.cgi) and an FTP site (ftp://ftp.ncbi.nlm.nih.gov/pub/CCDS/). In this paper, we outline the ongoing work, growth and stability of the CCDS dataset and provide updates on new collaboration members and new features added to the CCDS user interface. We also present expert curation scenarios, with specific examples highlighting the importance of an accurate reference genome assembly and the crucial role played by input from the research community. Published by Oxford University Press on behalf of Nucleic Acids Research 2017.
Three decades of the WHO code and marketing of infant formulas.
Forsyth, Stewart
2012-05-01
The International Code of Marketing of Breast Milk Substitutes states that governments, non-governmental organizations, experts, consumers and industry need to cooperate in activities aimed at improving infant nutrition. However, the evidence from the last three decades is that of a series of disputes, legal proceedings and boycotts. The purpose of this review is to assess the overall progress in the implementation of the Code and to examine the problematic areas of monitoring, compliance and governance. There are continuing issues of implementation, monitoring and compliance which predominantly reflect weak governance. Many Member States have yet to fully implement the Code recommendations and most States do not have adequate monitoring and reporting mechanisms. Application of the Code in developed countries may be undermined by a lack of consensus on the WHO recommendation of 6 months exclusive breastfeeding. There is evidence of continuing conflict and acrimony, especially between non-government organizations and industry. Measures need to be taken to encourage the Member States to implement the Code and to establish the governance systems that will not only ensure effective implementation and monitoring of the Code, but also deliver the Code within a spirit of participation, collaboration and trust.
NASA Astrophysics Data System (ADS)
Sullivan, Gary J.; Topiwala, Pankaj N.; Luthra, Ajay
2004-11-01
H.264/MPEG-4 AVC is the latest international video coding standard. It was jointly developed by the Video Coding Experts Group (VCEG) of the ITU-T and the Moving Picture Experts Group (MPEG) of ISO/IEC. It uses state-of-the-art coding tools and provides enhanced coding efficiency for a wide range of applications, including video telephony, video conferencing, TV, storage (DVD and/or hard disk based, especially high-definition DVD), streaming video, digital video authoring, digital cinema, and many others. The work on a new set of extensions to this standard has recently been completed. These extensions, known as the Fidelity Range Extensions (FRExt), provide a number of enhanced capabilities relative to the base specification as approved in the Spring of 2003. In this paper, an overview of this standard is provided, including the highlights of the capabilities of the new FRExt features. Some comparisons with the existing MPEG-2 and MPEG-4 Part 2 standards are also provided.
1988-03-01
A 11 7. MISSION TARGET COORDINATION a SAFETY’ CH- ECKS FIRE SUPPORT COORDINATION CENTERS S’. iN NFA jNQQ MSSIN C o CFL RqESSIN81 TARGET NO iYs YS...INITIAL DISTRIBUTION LIST No. Copies I. Defense Technical Information Center 2 Cameron Station Alexandria, Virginia 22304-6145 2. Library, Code 0142 2
Intelligent guidance and control for wind shear encounter
NASA Technical Reports Server (NTRS)
Stengel, Robert F.
1988-01-01
The principal objective is to develop methods for assessing the likelihood of wind shear encounter, for deciding what flight path to pursue, and for using the aircraft's full potential for combating wind shear. This study requires the definition of both deterministic and statistical techniques for fusing internal and external information, for making go/no-go decisions, and for generating commands to the aircraft's cockpit displays and autopilot for both manually controlled and automatic flight. The program has begun with the development of a real-time expert system for pilot aiding that is based on the results of the FAA Windshear Training Aids Program. A two-volume manual that presents an overview, pilot guide, training program, and substantiating data provides guidelines for this initial development. The Expert System to Avoid Wind Shear (ESAWS) currently contains over 140 rules and is coded in the LISP programming language for implementation on a Symbolics 3670 LISP machine.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Friedman-Hill, Ernest
Java Expert Shell System - Jess - is a rule engine and scripting environment written entirely in Sun's Java language, Jess was orginially inspired by the CLIPS expert system shell, but has grown int a complete, distinct JAVA-influenced environment of its own. Using Jess, you can build Java applets and applications that have the capacity to "reason" using knowledge you supply in the form of declarative rules. Jess is surprisingly fast, and for some problems is faster than CLIPS, in that many Jess scripts are valid CLIPS scripts and vice-versa. Like CLIPS, Jess uses the Rete algorithm to process rules,more » a very efficient mechanism for solving the difficult many-to-many matching problem. Jess adds many features to CLIPS, including backwards chaining and the ability to manipulate and directly reason about Java objects. Jess is also a powerful Java scripting environment, from which you can create Java objects and call Java methods without compiling any Java Code.« less
The Computer Aided Aircraft-design Package (CAAP)
NASA Technical Reports Server (NTRS)
Yalif, Guy U.
1994-01-01
The preliminary design of an aircraft is a complex, labor-intensive, and creative process. Since the 1970's, many computer programs have been written to help automate preliminary airplane design. Time and resource analyses have identified, 'a substantial decrease in project duration with the introduction of an automated design capability'. Proof-of-concept studies have been completed which establish 'a foundation for a computer-based airframe design capability', Unfortunately, today's design codes exist in many different languages on many, often expensive, hardware platforms. Through the use of a module-based system architecture, the Computer aided Aircraft-design Package (CAAP) will eventually bring together many of the most useful features of existing programs. Through the use of an expert system, it will add an additional feature that could be described as indispensable to entry level engineers and students: the incorporation of 'expert' knowledge into the automated design process.
Gnjidic, Danijela; Pearson, Sallie-Anne; Hilmer, Sarah N; Basilakis, Jim; Schaffer, Andrea L; Blyth, Fiona M; Banks, Emily
2015-03-30
Increasingly, automated methods are being used to code free-text medication data, but evidence on the validity of these methods is limited. To examine the accuracy of automated coding of previously keyed in free-text medication data compared with manual coding of original handwritten free-text responses (the 'gold standard'). A random sample of 500 participants (475 with and 25 without medication data in the free-text box) enrolled in the 45 and Up Study was selected. Manual coding involved medication experts keying in free-text responses and coding using Anatomical Therapeutic Chemical (ATC) codes (i.e. chemical substance 7-digit level; chemical subgroup 5-digit; pharmacological subgroup 4-digit; therapeutic subgroup 3-digit). Using keyed-in free-text responses entered by non-experts, the automated approach coded entries using the Australian Medicines Terminology database and assigned corresponding ATC codes. Based on manual coding, 1377 free-text entries were recorded and, of these, 1282 medications were coded to ATCs manually. The sensitivity of automated coding compared with manual coding was 79% (n = 1014) for entries coded at the exact ATC level, and 81.6% (n = 1046), 83.0% (n = 1064) and 83.8% (n = 1074) at the 5, 4 and 3-digit ATC levels, respectively. The sensitivity of automated coding for blank responses was 100% compared with manual coding. Sensitivity of automated coding was highest for prescription medications and lowest for vitamins and supplements, compared with the manual approach. Positive predictive values for automated coding were above 95% for 34 of the 38 individual prescription medications examined. Automated coding for free-text prescription medication data shows very high to excellent sensitivity and positive predictive values, indicating that automated methods can potentially be useful for large-scale, medication-related research.
Ranegger, Renate; Hackl, Werner O; Ammenwerth, Elske
2015-01-01
A Nursing Minimum Data Set (NMDS) aims at systematically describing nursing care in terms of patient problems, nursing activities, and patient outcomes. In an earlier Delphi study, 56 data elements were proposed to be included in an Austrian Nursing Minimum Data Set (NMDS-AT). To identify the most important data elements of this list, and to identify appropriate coding systems. Online Delphi-based survey with 88 experts. 43 data elements were rated as relevant for an NMDS-AT (strong agreement of more than half of the experts): nine data elements concerning the institution, patient demographics, and medical condition; 18 data elements concerning patient problems by using nursing diagnosis; seven data elements concerning nursing outcomes, and nine data elements concerning nursing interventions. As classification systems, national classification systems were proposed besides ICNP, NNN, and nursing-sensitive indicators. The resulting proposal for an NMDS-AT will now be tested with routine data.
Reinventing radiology reimbursement.
Marshall, John; Adema, Denise
2005-01-01
Lee Memorial Health System (LMHS), located in southwest Florida, consists of 5 hospitals, a home health agency, a skilled nursing facility, multiple outpatient centers, walk-in medical centers, and primary care physician offices. LMHS annually performs more than 300,000 imaging procedures with gross imaging revenues exceeding dollar 350 million. In fall 2002, LMHS received the results of an independent audit of its IR coding. The overall IR coding error rate was determined to be 84.5%. The projected net financial impact of these errors was an annual reimbursement loss of dollar 182,000. To address the issues of coding errors and reimbursement loss, LMHS implemented its clinical reimbursementspecialist (CRS) system in October 2003, as an extension of financial services' reimbursement division. LMHS began with CRSs in 3 service lines: emergency department, cardiac catheterization, and radiology. These 3 CRSs coordinate all facets of their respective areas' chargemaster, patient charges, coding, and reimbursement functions while serving as a resident coding expert within their clinical areas. The radiology reimbursement specialist (RRS) combines an experienced radiologic technologist, interventional technologist, medical records coder, financial auditor, reimbursement specialist, and biller into a single position. The RRS's radiology experience and technologist knowledge are key assets to resolving coding conflicts and handling complex interventional coding. In addition, performing a daily charge audit and an active code review are essential if an organization is to eliminate coding errors. One of the inherent effects of eliminating coding errors is the capturing of additional RVUs and units of service. During its first year, based on account level detail, the RRS system increased radiology productivity through the additional capture of just more than 3,000 RVUs and 1,000 additional units of service. In addition, the physicians appreciate having someone who "keeps up with all the coding changes" and looks out for the charges. By assisting a few physicians' staff with coding questions, providing coding updates, and allowing them to sit in on educational sessions, at least 2 physicians have transferred some their volume to LMHS from a competitor. The provision of a "clean account," without coding errors, allows the biller to avoid the rework and billing delays caused by coding issues. During the first quarter of the RRS system, the billers referred an average of 9 accounts per day for coding resolution. During the fourth quarter of the system, these referrals were reduced to less than one per day. Prior to the RRS system, resolving these issues took an average of 4 business days. Now the conflicts are resolved within 24 hours.
DataRocket: Interactive Visualisation of Data Structures
NASA Astrophysics Data System (ADS)
Parkes, Steve; Ramsay, Craig
2010-08-01
CodeRocket is a software engineering tool that provides cognitive support to the software engineer for reasoning about a method or procedure and for documenting the resulting code [1]. DataRocket is a software engineering tool designed to support visualisation and reasoning about program data structures. DataRocket is part of the CodeRocket family of software tools developed by Rapid Quality Systems [2] a spin-out company from the Space Technology Centre at the University of Dundee. CodeRocket and DataRocket integrate seamlessly with existing architectural design and coding tools and provide extensive documentation with little or no effort on behalf of the software engineer. Comprehensive, abstract, detailed design documentation is available early on in a project so that it can be used for design reviews with project managers and non expert stakeholders. Code and documentation remain fully synchronised even when changes are implemented in the code without reference to the existing documentation. At the end of a project the press of a button suffices to produce the detailed design document. Existing legacy code can be easily imported into CodeRocket and DataRocket to reverse engineer detailed design documentation making legacy code more manageable and adding substantially to its value. This paper introduces CodeRocket. It then explains the rationale for DataRocket and describes the key features of this new tool. Finally the major benefits of DataRocket for different stakeholders are considered.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Bragg-Sitton, Shannon Michelle
The Organization for Economic Cooperation and Development /Nuclear Energy Agency (OECD/NEA) Nuclear Science Committee approved the formation of an Expert Group on Accident Tolerant Fuel (ATF) for LWRs (EGATFL) in 2014. Chaired by Kemal Pasamehmetoglu, INL Associate Laboratory Director for Nuclear Science and Technology, the mandate for the EGATFL defines work under three task forces: (1) Systems Assessment, (2) Cladding and Core Materials, and (3) Fuel Concepts. Scope for the Systems Assessment task force (TF1) includes definition of evaluation metrics for ATF, technology readiness level definition, definition of illustrative scenarios for ATF evaluation, and identification of fuel performance and systemmore » codes applicable to ATF evaluation. The Cladding and Core Materials (TF2) and Fuel Concepts (TF3) task forces will identify gaps and needs for modeling and experimental demonstration; define key properties of interest; identify the data necessary to perform concept evaluation under normal conditions and illustrative scenarios; identify available infrastructure (internationally) to support experimental needs; and make recommendations on priorities. Where possible, considering proprietary and other export restrictions (e.g., International Traffic in Arms Regulations), the Expert Group will facilitate the sharing of data and lessons learned across the international group membership. The Systems Assessment task force is chaired by Shannon Bragg-Sitton (Idaho National Laboratory [INL], U.S.), the Cladding Task Force is chaired by Marie Moatti (Electricite de France [EdF], France), and the Fuels Task Force is chaired by a Masaki Kurata (Japan Atomic Energy Agency [JAEA], Japan). The original Expert Group mandate was established for June 2014 to June 2016. In April 2016 the Expert Group voted to extend the mandate one additional year to June 2017 in order to complete the task force deliverables; this request was subsequently approved by the Nuclear Science Committee. This report provides an update on the status Systems Assessment Task Force activities.« less
ERIC Educational Resources Information Center
Holbrook, M. Cay; MacCuspie, P. Ann
2010-01-01
Braille-reading mathematicians, scientists, and computer scientists were asked to examine the usability of the Unified English Braille Code (UEB) for technical materials. They had little knowledge of the code prior to the study. The research included two reading tasks, a short tutorial about UEB, and a focus group. The results indicated that the…
DOE Office of Scientific and Technical Information (OSTI.GOV)
Goossens, L.H.J.; Kraan, B.C.P.; Cooke, R.M.
1998-04-01
The development of two new probabilistic accident consequence codes, MACCS and COSYMA, was completed in 1990. These codes estimate the consequence from the accidental releases of radiological material from hypothesized accidents at nuclear installations. In 1991, the US Nuclear Regulatory Commission and the Commission of the European Communities began cosponsoring a joint uncertainty analysis of the two codes. The ultimate objective of this joint effort was to systematically develop credible and traceable uncertainty distributions for the respective code input variables. A formal expert judgment elicitation and evaluation process was identified as the best technology available for developing a library ofmore » uncertainty distributions for these consequence parameters. This report focuses on the results of the study to develop distribution for variables related to the MACCS and COSYMA internal dosimetry models. This volume contains appendices that include (1) a summary of the MACCS and COSYMA consequence codes, (2) the elicitation questionnaires and case structures, (3) the rationales and results for the panel on internal dosimetry, (4) short biographies of the experts, and (5) the aggregated results of their responses.« less
Assigning clinical codes with data-driven concept representation on Dutch clinical free text.
Scheurwegs, Elyne; Luyckx, Kim; Luyten, Léon; Goethals, Bart; Daelemans, Walter
2017-05-01
Clinical codes are used for public reporting purposes, are fundamental to determining public financing for hospitals, and form the basis for reimbursement claims to insurance providers. They are assigned to a patient stay to reflect the diagnosis and performed procedures during that stay. This paper aims to enrich algorithms for automated clinical coding by taking a data-driven approach and by using unsupervised and semi-supervised techniques for the extraction of multi-word expressions that convey a generalisable medical meaning (referred to as concepts). Several methods for extracting concepts from text are compared, two of which are constructed from a large unannotated corpus of clinical free text. A distributional semantic model (i.c. the word2vec skip-gram model) is used to generalize over concepts and retrieve relations between them. These methods are validated on three sets of patient stay data, in the disease areas of urology, cardiology, and gastroenterology. The datasets are in Dutch, which introduces a limitation on available concept definitions from expert-based ontologies (e.g. UMLS). The results show that when expert-based knowledge in ontologies is unavailable, concepts derived from raw clinical texts are a reliable alternative. Both concepts derived from raw clinical texts perform and concepts derived from expert-created dictionaries outperform a bag-of-words approach in clinical code assignment. Adding features based on tokens that appear in a semantically similar context has a positive influence for predicting diagnostic codes. Furthermore, the experiments indicate that a distributional semantics model can find relations between semantically related concepts in texts but also introduces erroneous and redundant relations, which can undermine clinical coding performance. Copyright © 2017. Published by Elsevier Inc.
Parallel software for lattice N = 4 supersymmetric Yang-Mills theory
NASA Astrophysics Data System (ADS)
Schaich, David; DeGrand, Thomas
2015-05-01
We present new parallel software, SUSY LATTICE, for lattice studies of four-dimensional N = 4 supersymmetric Yang-Mills theory with gauge group SU(N). The lattice action is constructed to exactly preserve a single supersymmetry charge at non-zero lattice spacing, up to additional potential terms included to stabilize numerical simulations. The software evolved from the MILC code for lattice QCD, and retains a similar large-scale framework despite the different target theory. Many routines are adapted from an existing serial code (Catterall and Joseph, 2012), which SUSY LATTICE supersedes. This paper provides an overview of the new parallel software, summarizing the lattice system, describing the applications that are currently provided and explaining their basic workflow for non-experts in lattice gauge theory. We discuss the parallel performance of the code, and highlight some notable aspects of the documentation for those interested in contributing to its future development.
Lesko, Mehdi M; Woodford, Maralyn; White, Laura; O'Brien, Sarah J; Childs, Charmaine; Lecky, Fiona E
2010-08-06
The purpose of Abbreviated Injury Scale (AIS) is to code various types of Traumatic Brain Injuries (TBI) based on their anatomical location and severity. The Marshall CT Classification is used to identify those subgroups of brain injured patients at higher risk of deterioration or mortality. The purpose of this study is to determine whether and how AIS coding can be translated to the Marshall Classification Initially, a Marshall Class was allocated to each AIS code through cross-tabulation. This was agreed upon through several discussion meetings with experts from both fields (clinicians and AIS coders). Furthermore, in order to make this translation possible, some necessary assumptions with regards to coding and classification of mass lesions and brain swelling were essential which were all approved and made explicit. The proposed method involves two stages: firstly to determine all possible Marshall Classes which a given patient can attract based on allocated AIS codes; via cross-tabulation and secondly to assign one Marshall Class to each patient through an algorithm. This method can be easily programmed in computer softwares and it would enable future important TBI research programs using trauma registry data.
2010-01-01
Background The purpose of Abbreviated Injury Scale (AIS) is to code various types of Traumatic Brain Injuries (TBI) based on their anatomical location and severity. The Marshall CT Classification is used to identify those subgroups of brain injured patients at higher risk of deterioration or mortality. The purpose of this study is to determine whether and how AIS coding can be translated to the Marshall Classification Methods Initially, a Marshall Class was allocated to each AIS code through cross-tabulation. This was agreed upon through several discussion meetings with experts from both fields (clinicians and AIS coders). Furthermore, in order to make this translation possible, some necessary assumptions with regards to coding and classification of mass lesions and brain swelling were essential which were all approved and made explicit. Results The proposed method involves two stages: firstly to determine all possible Marshall Classes which a given patient can attract based on allocated AIS codes; via cross-tabulation and secondly to assign one Marshall Class to each patient through an algorithm. Conclusion This method can be easily programmed in computer softwares and it would enable future important TBI research programs using trauma registry data. PMID:20691038
A domain specific language for performance portable molecular dynamics algorithms
NASA Astrophysics Data System (ADS)
Saunders, William Robert; Grant, James; Müller, Eike Hermann
2018-03-01
Developers of Molecular Dynamics (MD) codes face significant challenges when adapting existing simulation packages to new hardware. In a continuously diversifying hardware landscape it becomes increasingly difficult for scientists to be experts both in their own domain (physics/chemistry/biology) and specialists in the low level parallelisation and optimisation of their codes. To address this challenge, we describe a "Separation of Concerns" approach for the development of parallel and optimised MD codes: the science specialist writes code at a high abstraction level in a domain specific language (DSL), which is then translated into efficient computer code by a scientific programmer. In a related context, an abstraction for the solution of partial differential equations with grid based methods has recently been implemented in the (Py)OP2 library. Inspired by this approach, we develop a Python code generation system for molecular dynamics simulations on different parallel architectures, including massively parallel distributed memory systems and GPUs. We demonstrate the efficiency of the auto-generated code by studying its performance and scalability on different hardware and compare it to other state-of-the-art simulation packages. With growing data volumes the extraction of physically meaningful information from the simulation becomes increasingly challenging and requires equally efficient implementations. A particular advantage of our approach is the easy expression of such analysis algorithms. We consider two popular methods for deducing the crystalline structure of a material from the local environment of each atom, show how they can be expressed in our abstraction and implement them in the code generation framework.
Minozzi, Silvia; Armaroli, Paola; Espina, Carolina; Villain, Patricia; Wiseman, Martin; Schüz, Joachim; Segnan, Nereo
2015-12-01
The European Code Against Cancer is a set of recommendations to give advice on cancer prevention. Its 4th edition is an update of the 3rd edition, from 2003. Working Groups of independent experts from different fields of cancer prevention were appointed to review the recommendations, supported by a Literature Group to provide scientific and technical support in the assessment of the scientific evidence, through systematic reviews of the literature. Common procedures were developed to guide the experts in identifying, retrieving, assessing, interpreting and summarizing the scientific evidence in order to revise the recommendations. The Code strictly followed the concept of providing advice to European Union citizens based on the current best available science. The advice, if followed, would be expected to reduce cancer risk, referring both to avoiding or reducing exposure to carcinogenic agents or changing behaviour related to cancer risk and to participating in medical interventions able to avert specific cancers or their consequences. The information sources and procedures for the review of the scientific evidence are described here in detail. The 12 recommendations of the 4th edition of the European Code Against Cancer were ultimately approved by a Scientific Committee of leading European cancer and public health experts. Copyright © 2015 International Agency for Research on Cancer. Published by Elsevier Ltd. All rights reserved.
Sollie, Annet; Sijmons, Rolf H; Helsper, Charles; Numans, Mattijs E
2017-03-01
To assess quality and reusability of coded cancer diagnoses in routine primary care data. To identify factors that influence data quality and areas for improvement. A dynamic cohort study in a Dutch network database containing 250,000 anonymized electronic medical records (EMRs) from 52 general practices was performed. Coded data from 2000 to 2011 for the three most common cancer types (breast, colon and prostate cancer) was compared to the Netherlands Cancer Registry. Data quality is expressed in Standard Incidence Ratios (SIRs): the ratio between the number of coded cases observed in the primary care network database and the expected number of cases based on the Netherlands Cancer Registry. Ratios were multiplied by 100% for readability. The overall SIR was 91.5% (95%CI 88.5-94.5) and showed improvement over the years. SIRs differ between cancer types: from 71.5% for colon cancer in males to 103.9% for breast cancer. There are differences in data quality (SIRs 76.2% - 99.7%) depending on the EMR system used, with SIRs up to 232.9% for breast cancer. Frequently observed errors in routine healthcare data can be classified as: lack of integrity checks, inaccurate use and/or lack of codes, and lack of EMR system functionality. Re-users of coded routine primary care Electronic Medical Record data should be aware that 30% of cancer cases can be missed. Up to 130% of cancer cases found in the EMR data can be false-positive. The type of EMR system and the type of cancer influence the quality of coded diagnosis registry. While data quality can be improved (e.g. through improving system design and by training EMR system users), re-use should only be taken care of by appropriately trained experts. Copyright © 2016. Published by Elsevier B.V.
ERIC Educational Resources Information Center
Eells, Tracy D.; Lombart, Kenneth G.; Kendjelic, Edward M.; Turner, L. Carolyn; Lucas, Cynthia P.
2005-01-01
Sixty-five expert, experienced, and novice cognitive-behavioral and psychodynamic psychotherapists provided "think aloud" case formulations in response to 6 standardized patient vignettes varying in disorder and prototypicality. The 390 formulations were reliably transcribed, segmented into idea units, content coded, and rated on multiple…
Welcome to Fermilab Butterflies!!
, fascinating insects, and there's a lot to learn about them! Join our expert, Tom Peterson, and explore the Meet Tom Peterson, Fermilab's Butterfly Expert Go to our Butterfly Links Have fun! Graphics and Page Design: Rory Parilac, Content: Tom Peterson and Rory Parilac Database and Lasso Code: Liz Quigg Web
An efficient system for reliably transmitting image and video data over low bit rate noisy channels
NASA Technical Reports Server (NTRS)
Costello, Daniel J., Jr.; Huang, Y. F.; Stevenson, Robert L.
1994-01-01
This research project is intended to develop an efficient system for reliably transmitting image and video data over low bit rate noisy channels. The basic ideas behind the proposed approach are the following: employ statistical-based image modeling to facilitate pre- and post-processing and error detection, use spare redundancy that the source compression did not remove to add robustness, and implement coded modulation to improve bandwidth efficiency and noise rejection. Over the last six months, progress has been made on various aspects of the project. Through our studies of the integrated system, a list-based iterative Trellis decoder has been developed. The decoder accepts feedback from a post-processor which can detect channel errors in the reconstructed image. The error detection is based on the Huber Markov random field image model for the compressed image. The compression scheme used here is that of JPEG (Joint Photographic Experts Group). Experiments were performed and the results are quite encouraging. The principal ideas here are extendable to other compression techniques. In addition, research was also performed on unequal error protection channel coding, subband vector quantization as a means of source coding, and post processing for reducing coding artifacts. Our studies on unequal error protection (UEP) coding for image transmission focused on examining the properties of the UEP capabilities of convolutional codes. The investigation of subband vector quantization employed a wavelet transform with special emphasis on exploiting interband redundancy. The outcome of this investigation included the development of three algorithms for subband vector quantization. The reduction of transform coding artifacts was studied with the aid of a non-Gaussian Markov random field model. This results in improved image decompression. These studies are summarized and the technical papers included in the appendices.
Russ, Daniel E.; Ho, Kwan-Yuet; Colt, Joanne S.; Armenti, Karla R.; Baris, Dalsu; Chow, Wong-Ho; Davis, Faith; Johnson, Alison; Purdue, Mark P.; Karagas, Margaret R.; Schwartz, Kendra; Schwenn, Molly; Silverman, Debra T.; Johnson, Calvin A.; Friesen, Melissa C.
2016-01-01
Background Mapping job titles to standardized occupation classification (SOC) codes is an important step in identifying occupational risk factors in epidemiologic studies. Because manual coding is time-consuming and has moderate reliability, we developed an algorithm called SOCcer (Standardized Occupation Coding for Computer-assisted Epidemiologic Research) to assign SOC-2010 codes based on free-text job description components. Methods Job title and task-based classifiers were developed by comparing job descriptions to multiple sources linking job and task descriptions to SOC codes. An industry-based classifier was developed based on the SOC prevalence within an industry. These classifiers were used in a logistic model trained using 14,983 jobs with expert-assigned SOC codes to obtain empirical weights for an algorithm that scored each SOC/job description. We assigned the highest scoring SOC code to each job. SOCcer was validated in two occupational data sources by comparing SOC codes obtained from SOCcer to expert assigned SOC codes and lead exposure estimates obtained by linking SOC codes to a job-exposure matrix. Results For 11,991 case-control study jobs, SOCcer-assigned codes agreed with 44.5% and 76.3% of manually assigned codes at the 6- and 2-digit level, respectively. Agreement increased with the score, providing a mechanism to identify assignments needing review. Good agreement was observed between lead estimates based on SOCcer and manual SOC assignments (kappa: 0.6–0.8). Poorer performance was observed for inspection job descriptions, which included abbreviations and worksite-specific terminology. Conclusions Although some manual coding will remain necessary, using SOCcer may improve the efficiency of incorporating occupation into large-scale epidemiologic studies. PMID:27102331
Intellectual system of identification of Arabic graphics
NASA Astrophysics Data System (ADS)
Abdoullayeva, Gulchin G.; Aliyev, Telman A.; Gurbanova, Nazakat G.
2001-08-01
The studies made by using the domain of graphic images allowed creating facilities of the artificial intelligence for letters, letter combinations etc. for various graphics and prints. The work proposes a system of recognition and identification of symbols of the Arabic graphics, which has its own specificity as compared to Latin and Cyrillic ones. The starting stage of the recognition and the identification is coding with further entry of information into a computer. Here the problem of entry is one of the essentials. For entry of a large volume of information in the unit of time a scanner is usually employed. Along with the scanner the authors suggest their elaboration of technical facilities for effective input and coding of the information. For refinement of symbols not identified from the scanner mostly for a small bulk of information the developed coding devices are used directly in the process of writing. The functional design of the software is elaborated on the basis of the heuristic model of the creative activity of a researcher and experts in the description and estimation of states of the weakly formalizable systems on the strength of the methods of identification and of selection of geometric features.
The Contract Management Body of Knowledge: A Comparison of Contracting Competencies
2013-12-01
SME subject matter expert SOW statement of work TINA Truth in Negotiations Act UCC uniform commercial code WBS work breakdown structure xv...documents whose terms and condition are legally enforceable. Sources of law and guidance covered include the uniform commercial code ( UCC ), Federal...contracting including the uniform commercial code ( UCC ), Federal Acquisition Regulation (FAR), as well as various other laws pertaining to both
Milton, Constance L
2003-10-01
The American Nurses Association Code of Ethics for Nurses calls for the nurse to practice with compassion and respect for every individual. What are the ethics and challenges of practicing professional nursing with expertise and educating a new generation of nurses while incorporating the interpretive statements into practice? This column differentiates the traditional biomedical views on human dignity and respect while exploring the embedded ethics of respect and self-determination and what it truly means to be an expert of nursing from the theoretical perspective of the human becoming school of thought.
OntoBrowser: a collaborative tool for curation of ontologies by subject matter experts.
Ravagli, Carlo; Pognan, Francois; Marc, Philippe
2017-01-01
The lack of controlled terminology and ontology usage leads to incomplete search results and poor interoperability between databases. One of the major underlying challenges of data integration is curating data to adhere to controlled terminologies and/or ontologies. Finding subject matter experts with the time and skills required to perform data curation is often problematic. In addition, existing tools are not designed for continuous data integration and collaborative curation. This results in time-consuming curation workflows that often become unsustainable. The primary objective of OntoBrowser is to provide an easy-to-use online collaborative solution for subject matter experts to map reported terms to preferred ontology (or code list) terms and facilitate ontology evolution. Additional features include web service access to data, visualization of ontologies in hierarchical/graph format and a peer review/approval workflow with alerting. The source code is freely available under the Apache v2.0 license. Source code and installation instructions are available at http://opensource.nibr.com This software is designed to run on a Java EE application server and store data in a relational database. philippe.marc@novartis.com. © The Author 2016. Published by Oxford University Press.
OntoBrowser: a collaborative tool for curation of ontologies by subject matter experts
Ravagli, Carlo; Pognan, Francois
2017-01-01
Summary: The lack of controlled terminology and ontology usage leads to incomplete search results and poor interoperability between databases. One of the major underlying challenges of data integration is curating data to adhere to controlled terminologies and/or ontologies. Finding subject matter experts with the time and skills required to perform data curation is often problematic. In addition, existing tools are not designed for continuous data integration and collaborative curation. This results in time-consuming curation workflows that often become unsustainable. The primary objective of OntoBrowser is to provide an easy-to-use online collaborative solution for subject matter experts to map reported terms to preferred ontology (or code list) terms and facilitate ontology evolution. Additional features include web service access to data, visualization of ontologies in hierarchical/graph format and a peer review/approval workflow with alerting. Availability and implementation: The source code is freely available under the Apache v2.0 license. Source code and installation instructions are available at http://opensource.nibr.com. This software is designed to run on a Java EE application server and store data in a relational database. Contact: philippe.marc@novartis.com PMID:27605099
Aerosol Therapy for Obstructive Lung Diseases
2011-01-01
Inhaled aerosol therapies are the mainstay of treatment of obstructive lung diseases. Aerosol devices deliver drugs rapidly and directly into the airways, allowing high local drug concentrations while limiting systemic toxicity. While numerous clinical trials, literature reviews, and expert panel guidelines inform the choice of inhalational drugs, deciding which aerosol device (ie, metered-dose inhaler, nebulizer, or dry powder inhaler) best suits a given patient and clinical setting can seem arbitrary and confusing. Similar confusion regarding Current Procedural Terminology (CPT) coding for administration of aerosol therapies can lead to lost revenue from underbilling and wasted administrative effort handling denied claims. This article reviews the aerosol devices currently available, discusses their relative merits in various clinical settings, and summarizes appropriate CPT coding for aerosol therapy. PMID:21896522
TARGET - TASK ANALYSIS REPORT GENERATION TOOL, VERSION 1.0
NASA Technical Reports Server (NTRS)
Ortiz, C. J.
1994-01-01
The Task Analysis Report Generation Tool, TARGET, is a graphical interface tool used to capture procedural knowledge and translate that knowledge into a hierarchical report. TARGET is based on VISTA, a knowledge acquisition tool developed by the Naval Systems Training Center. TARGET assists a programmer and/or task expert organize and understand the steps involved in accomplishing a task. The user can label individual steps in the task through a dialogue-box and get immediate graphical feedback for analysis. TARGET users can decompose tasks into basic action kernels or minimal steps to provide a clear picture of all basic actions needed to accomplish a job. This method allows the user to go back and critically examine the overall flow and makeup of the process. The user can switch between graphics (box flow diagrams) and text (task hierarchy) versions to more easily study the process being documented. As the practice of decomposition continues, tasks and their subtasks can be continually modified to more accurately reflect the user's procedures and rationale. This program is designed to help a programmer document an expert's task thus allowing the programmer to build an expert system which can help others perform the task. Flexibility is a key element of the system design and of the knowledge acquisition session. If the expert is not able to find time to work on the knowledge acquisition process with the program developer, the developer and subject matter expert may work in iterative sessions. TARGET is easy to use and is tailored to accommodate users ranging from the novice to the experienced expert systems builder. TARGET is written in C-language for IBM PC series and compatible computers running MS-DOS and Microsoft Windows version 3.0 or 3.1. No source code is supplied. The executable also requires 2Mb of RAM, a Microsoft compatible mouse, a VGA display and an 80286, 386 or 486 processor machine. The standard distribution medium for TARGET is one 5.25 inch 360K MS-DOS format diskette. TARGET was developed in 1991.
Completing and Adapting Models of Biological Processes
NASA Technical Reports Server (NTRS)
Margaria, Tiziana; Hinchey, Michael G.; Raffelt, Harald; Rash, James L.; Rouff, Christopher A.; Steffen, Bernhard
2006-01-01
We present a learning-based method for model completion and adaptation, which is based on the combination of two approaches: 1) R2D2C, a technique for mechanically transforming system requirements via provably equivalent models to running code, and 2) automata learning-based model extrapolation. The intended impact of this new combination is to make model completion and adaptation accessible to experts of the field, like biologists or engineers. The principle is briefly illustrated by generating models of biological procedures concerning gene activities in the production of proteins, although the main application is going to concern autonomic systems for space exploration.
75 FR 61139 - Board of Scientific Counselors (BOSC); Request for Nominations of Experts
Federal Register 2010, 2011, 2012, 2013, 2014
2010-10-04
.... Heather Drumm, Mail Code 8104-R, Office of Science Policy, Office of Research and Development... and bioinformatics); socioeconomics; environmental justice, science policy (research, policy, and.... Acting Director, Office of Science Policy . [FR Doc. 2010-24805 Filed 10-1-10; 8:45 am] BILLING CODE 6560...
Matin, Ivan; Hadzistevic, Miodrag; Vukelic, Djordje; Potran, Michal; Brajlih, Tomaz
2017-07-01
Nowadays, the integrated CAD/CAE systems are favored solutions for the design of simulation models for casting metal substructures of metal-ceramic crowns. The worldwide authors have used different approaches to solve the problems using an expert system. Despite substantial research progress in the design of experts systems for the simulation model design and manufacturing have insufficiently considered the specifics of casting in dentistry, especially the need for further CAD, RE, CAE for the estimation of casting parameters and the control of the casting machine. The novel expert system performs the following: CAD modeling of the simulation model for casting, fast modeling of gate design, CAD eligibility and cast ability check of the model, estimation and running of the program code for the casting machine, as well as manufacturing time reduction of the metal substructure. The authors propose an integration method using common data model approach, blackboard architecture, rule-based reasoning and iterative redesign method. Arithmetic mean roughness values was determinated with constant Gauss low-pass filter (cut-off length of 2.5mm) according to ISO 4287 using Mahr MARSURF PS1. Dimensional deviation between the designed model and manufactured cast was determined using the coordinate measuring machine Zeiss Contura G2 and GOM Inspect software. The ES allows for obtaining the castings derived roughness grade number N7. The dimensional deviation between the simulation model of the metal substructure and the manufactured cast is 0.018mm. The arithmetic mean roughness values measured on the casting substructure are from 1.935µm to 2.778µm. The realized developed expert system with the integrated database is fully applicable for the observed hardware and software. Values of the arithmetic mean roughness and dimensional deviation indicate that casting substructures are surface quality, which is more than enough and useful for direct porcelain veneering. The manufacture of the substructure shows that the proposed ES allows the improvement of the design process while reducing the manufacturing time. Copyright © 2017 Elsevier B.V. All rights reserved.
Debugging Techniques Used by Experienced Programmers to Debug Their Own Code.
1990-09-01
IS. NUMBER OF PAGES code debugging 62 computer programmers 16. PRICE CODE debug programming 17. SECURITY CLASSIFICATION 18. SECURITY CLASSIFICATION 119...Davis, and Schultz (1987) also compared experts and novices, but focused on the way a computer program is represented cognitively and how that...of theories in the emerging computer programming domain (Fisher, 1987). In protocol analysis, subjects are asked to talk/think aloud as they solve
ERIC Educational Resources Information Center
Smith, Derrick; Rosenblum, L. Penny
2013-01-01
Introduction: The purpose of the study presented here was the initial validation of a comprehensive set of competencies focused solely on the Nemeth code. Methods: Using the Delphi method, 20 expert panelists were recruited to participate in the study on the basis of their past experience in teaching a university-level course in the Nemeth code.…
Raczek, Ewa
2009-01-01
On June 13, 2009, the new Family and Guardianship Code came into effect. Many important modifications were implemented to Chapter I. "Origin of a child", the issue being of special importance in the work of a forensic geneticist. Those changes are related not only to arguableness of the fatherhood of both types--the one that is judged in lawsuit of denial of the fatherhood and that in which ineffectiveness of paternity is recognized--but for the first time they also demand on maternity testing. The Code defines who--according to Polish law--is a mother to a child and on this base states motherhood. In consequence, the main legal maxim Mater semper certa est, which has existed since Ancient Rome times is now annulled. The paper presents some remarks of an expert witness on the introduced changes.
CrossTalk: The Journal of Defense Software Engineering. Volume 26, Number 6, November/December 2013
2013-12-01
requirements during sprint planning. Automated scanning, which includes automated code-review tools, allows the expert to monitor the system... sprint . This enables the validator to leverage the test results for formal validation and verification, and perform a shortened “hybrid” style of IV&V...per SPRINT (1-4 weeks) 1 week 1 Month Up to four months Ø Deliverable product to user Ø Security posture assessed Ø Accredited to field/operate
Advances in the p and h-p Versions of the Finite Element Method. A survey
1988-01-01
p versions is the code PROBE which was developed by NOETIC Technologies, St. Louis, MO [49] [60]. PROBE solves two dimensional problems of linear...p and h-p versions of the finite element method was studied in detail from various point of view. We will mention here some essential illustrative...49] PROBE - Sample Problems. Series of reports, Noetic Technologies, St. Louis, MO 63117. [50] Rank, E., Babu’ka, I., An expert system for the
DOE Office of Scientific and Technical Information (OSTI.GOV)
Goossens, L.H.J.; Kraan, B.C.P.; Cooke, R.M.
1997-12-01
The development of two new probabilistic accident consequence codes, MACCS and COSYMA, was completed in 1990. These codes estimate the consequence from the accidental releases of radiological material from hypothesized accidents at nuclear installations. In 1991, the US Nuclear Regulatory Commission and the Commission of the European Communities began cosponsoring a joint uncertainty analysis of the two codes. The ultimate objective of this joint effort was to systematically develop credible and traceable uncertainty distributions for the respective code input variables. A formal expert judgment elicitation and evaluation process was identified as the best technology available for developing a library ofmore » uncertainty distributions for these consequence parameters. This report focuses on the results of the study to develop distribution for variables related to the MACCS and COSYMA deposited material and external dose models. This volume contains appendices that include (1) a summary of the MACCS and COSYMA consequence codes, (2) the elicitation questionnaires and case structures, (3) the rationales and results for the panel on deposited material and external doses, (4) short biographies of the experts, and (5) the aggregated results of their responses.« less
Reusable rocket engine turbopump health monitoring system, part 3
NASA Technical Reports Server (NTRS)
Perry, John G.
1989-01-01
Degradation mechanisms and sensor identification/selection resulted in a list of degradation modes and a list of sensors that are utilized in the diagnosis of these degradation modes. The sensor list is divided into primary and secondary indicators of the corresponding degradation modes. The signal conditioning requirements are discussed, describing the methods of producing the Space Shuttle Main Engine (SSME) post-hot-fire test data to be utilized by the Health Monitoring System. Development of the diagnostic logic and algorithms is also presented. The knowledge engineering approach, as utilized, includes the knowledge acquisition effort, characterization of the expert's problem solving strategy, conceptually defining the form of the applicable knowledge base, and rule base, and identifying an appropriate inferencing mechanism for the problem domain. The resulting logic flow graphs detail the diagnosis/prognosis procedure as followed by the experts. The nature and content of required support data and databases is also presented. The distinction between deep and shallow types of knowledge is identified. Computer coding of the Health Monitoring System is shown to follow the logical inferencing of the logic flow graphs/algorithms.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Bragg-Sitton, Shannon Michelle
The Organization for Economic Cooperation and Development /Nuclear Energy Agency (OECD/NEA) Nuclear Science Committee approved the formation of an Expert Group on Accident Tolerant Fuel (ATF) for LWRs (EGATFL) in 2014. Chaired by Kemal Pasamehmetoglu, INL Associate Laboratory Director for Nuclear Science and Technology, the mandate for the EGATFL defines work under three task forces: (1) Systems Assessment, (2) Cladding and Core Materials, and (3) Fuel Concepts. Scope for the Systems Assessment task force includes definition of evaluation metrics for ATF, technology readiness level definition, definition of illustrative scenarios for ATF evaluation, parametric studies, and selection of system codes. Themore » Cladding and Core Materials and Fuel Concepts task forces will identify gaps and needs for modeling and experimental demonstration; define key properties of interest; identify the data necessary to perform concept evaluation under normal conditions and illustrative scenarios; identify available infrastructure (internationally) to support experimental needs; and make recommendations on priorities. Where possible, considering proprietary and other export restrictions (e.g., International Traffic in Arms Regulations), the Expert Group will facilitate the sharing of data and lessons learned across the international group membership. The Systems Assessment Task Force is chaired by Shannon Bragg-Sitton (INL), while the Cladding Task Force will be chaired by a representative from France (Marie Moatti, Electricite de France [EdF]) and the Fuels Task Force will be chaired by a representative from Japan (Masaki Kurata, Japan Atomic Energy Agency [JAEA]). This report provides an overview of the Systems Assessment Task Force charter and status of work accomplishment.« less
Progress in The Semantic Analysis of Scientific Code
NASA Technical Reports Server (NTRS)
Stewart, Mark
2000-01-01
This paper concerns a procedure that analyzes aspects of the meaning or semantics of scientific and engineering code. This procedure involves taking a user's existing code, adding semantic declarations for some primitive variables, and parsing this annotated code using multiple, independent expert parsers. These semantic parsers encode domain knowledge and recognize formulae in different disciplines including physics, numerical methods, mathematics, and geometry. The parsers will automatically recognize and document some static, semantic concepts and help locate some program semantic errors. These techniques may apply to a wider range of scientific codes. If so, the techniques could reduce the time, risk, and effort required to develop and modify scientific codes.
Simulations of pattern dynamics for reaction-diffusion systems via SIMULINK.
Wang, Kaier; Steyn-Ross, Moira L; Steyn-Ross, D Alistair; Wilson, Marcus T; Sleigh, Jamie W; Shiraishi, Yoichi
2014-04-11
Investigation of the nonlinear pattern dynamics of a reaction-diffusion system almost always requires numerical solution of the system's set of defining differential equations. Traditionally, this would be done by selecting an appropriate differential equation solver from a library of such solvers, then writing computer codes (in a programming language such as C or Matlab) to access the selected solver and display the integrated results as a function of space and time. This "code-based" approach is flexible and powerful, but requires a certain level of programming sophistication. A modern alternative is to use a graphical programming interface such as Simulink to construct a data-flow diagram by assembling and linking appropriate code blocks drawn from a library. The result is a visual representation of the inter-relationships between the state variables whose output can be made completely equivalent to the code-based solution. As a tutorial introduction, we first demonstrate application of the Simulink data-flow technique to the classical van der Pol nonlinear oscillator, and compare Matlab and Simulink coding approaches to solving the van der Pol ordinary differential equations. We then show how to introduce space (in one and two dimensions) by solving numerically the partial differential equations for two different reaction-diffusion systems: the well-known Brusselator chemical reactor, and a continuum model for a two-dimensional sheet of human cortex whose neurons are linked by both chemical and electrical (diffusive) synapses. We compare the relative performances of the Matlab and Simulink implementations. The pattern simulations by Simulink are in good agreement with theoretical predictions. Compared with traditional coding approaches, the Simulink block-diagram paradigm reduces the time and programming burden required to implement a solution for reaction-diffusion systems of equations. Construction of the block-diagram does not require high-level programming skills, and the graphical interface lends itself to easy modification and use by non-experts.
NASA Technical Reports Server (NTRS)
Berke, Laszlo; Patnaik, Surya N.; Murthy, Pappu L. N.
1993-01-01
The application of artificial neural networks to capture structural design expertise is demonstrated. The principal advantage of a trained neural network is that it requires trivial computational effort to produce an acceptable new design. For the class of problems addressed, the development of a conventional expert system would be extremely difficult. In the present effort, a structural optimization code with multiple nonlinear programming algorithms and an artificial neural network code NETS were used. A set of optimum designs for a ring and two aircraft wings for static and dynamic constraints were generated by using the optimization codes. The optimum design data were processed to obtain input and output pairs, which were used to develop a trained artificial neural network with the code NETS. Optimum designs for new design conditions were predicted by using the trained network. Neural net prediction of optimum designs was found to be satisfactory for most of the output design parameters. However, results from the present study indicate that caution must be exercised to ensure that all design variables are within selected error bounds.
Optimum Design of Aerospace Structural Components Using Neural Networks
NASA Technical Reports Server (NTRS)
Berke, L.; Patnaik, S. N.; Murthy, P. L. N.
1993-01-01
The application of artificial neural networks to capture structural design expertise is demonstrated. The principal advantage of a trained neural network is that it requires a trivial computational effort to produce an acceptable new design. For the class of problems addressed, the development of a conventional expert system would be extremely difficult. In the present effort, a structural optimization code with multiple nonlinear programming algorithms and an artificial neural network code NETS were used. A set of optimum designs for a ring and two aircraft wings for static and dynamic constraints were generated using the optimization codes. The optimum design data were processed to obtain input and output pairs, which were used to develop a trained artificial neural network using the code NETS. Optimum designs for new design conditions were predicted using the trained network. Neural net prediction of optimum designs was found to be satisfactory for the majority of the output design parameters. However, results from the present study indicate that caution must be exercised to ensure that all design variables are within selected error bounds.
High Resolution, High Frame Rate Video Technology
NASA Technical Reports Server (NTRS)
1990-01-01
Papers and working group summaries presented at the High Resolution, High Frame Rate Video (HHV) Workshop are compiled. HHV system is intended for future use on the Space Shuttle and Space Station Freedom. The Workshop was held for the dual purpose of: (1) allowing potential scientific users to assess the utility of the proposed system for monitoring microgravity science experiments; and (2) letting technical experts from industry recommend improvements to the proposed near-term HHV system. The following topics are covered: (1) State of the art in the video system performance; (2) Development plan for the HHV system; (3) Advanced technology for image gathering, coding, and processing; (4) Data compression applied to HHV; (5) Data transmission networks; and (6) Results of the users' requirements survey conducted by NASA.
The Italian Code of Medical Deontology: characterizing features of its 2014 edition.
Conti, Andrea Alberto
2015-09-14
The latest edition of the Italian Code of Medical Deontology has been released by the Italian Federation of the Registers of Physicians and Dentists in May 2014 (1). The previous edition of the Italian Code dated back to 2006 (2), and it has been integrated and updated by a multi-professional and inter-disciplinary panel involving, besides physicians, representatives of scientific societies and trade unions, jurisconsults and experts in bioethics....
2013-01-01
Background The harmonization of European health systems brings with it a need for tools to allow the standardized collection of information about medical care. A common coding system and standards for the description of services are needed to allow local data to be incorporated into evidence-informed policy, and to permit equity and mobility to be assessed. The aim of this project has been to design such a classification and a related tool for the coding of services for Long Term Care (DESDE-LTC), based on the European Service Mapping Schedule (ESMS). Methods The development of DESDE-LTC followed an iterative process using nominal groups in 6 European countries. 54 researchers and stakeholders in health and social services contributed to this process. In order to classify services, we use the minimal organization unit or “Basic Stable Input of Care” (BSIC), coded by its principal function or “Main Type of Care” (MTC). The evaluation of the tool included an analysis of feasibility, consistency, ontology, inter-rater reliability, Boolean Factor Analysis, and a preliminary impact analysis (screening, scoping and appraisal). Results DESDE-LTC includes an alpha-numerical coding system, a glossary and an assessment instrument for mapping and counting LTC. It shows high feasibility, consistency, inter-rater reliability and face, content and construct validity. DESDE-LTC is ontologically consistent. It is regarded by experts as useful and relevant for evidence-informed decision making. Conclusion DESDE-LTC contributes to establishing a common terminology, taxonomy and coding of LTC services in a European context, and a standard procedure for data collection and international comparison. PMID:23768163
Russ, Daniel E; Ho, Kwan-Yuet; Colt, Joanne S; Armenti, Karla R; Baris, Dalsu; Chow, Wong-Ho; Davis, Faith; Johnson, Alison; Purdue, Mark P; Karagas, Margaret R; Schwartz, Kendra; Schwenn, Molly; Silverman, Debra T; Johnson, Calvin A; Friesen, Melissa C
2016-06-01
Mapping job titles to standardised occupation classification (SOC) codes is an important step in identifying occupational risk factors in epidemiological studies. Because manual coding is time-consuming and has moderate reliability, we developed an algorithm called SOCcer (Standardized Occupation Coding for Computer-assisted Epidemiologic Research) to assign SOC-2010 codes based on free-text job description components. Job title and task-based classifiers were developed by comparing job descriptions to multiple sources linking job and task descriptions to SOC codes. An industry-based classifier was developed based on the SOC prevalence within an industry. These classifiers were used in a logistic model trained using 14 983 jobs with expert-assigned SOC codes to obtain empirical weights for an algorithm that scored each SOC/job description. We assigned the highest scoring SOC code to each job. SOCcer was validated in 2 occupational data sources by comparing SOC codes obtained from SOCcer to expert assigned SOC codes and lead exposure estimates obtained by linking SOC codes to a job-exposure matrix. For 11 991 case-control study jobs, SOCcer-assigned codes agreed with 44.5% and 76.3% of manually assigned codes at the 6-digit and 2-digit level, respectively. Agreement increased with the score, providing a mechanism to identify assignments needing review. Good agreement was observed between lead estimates based on SOCcer and manual SOC assignments (κ 0.6-0.8). Poorer performance was observed for inspection job descriptions, which included abbreviations and worksite-specific terminology. Although some manual coding will remain necessary, using SOCcer may improve the efficiency of incorporating occupation into large-scale epidemiological studies. Published by the BMJ Publishing Group Limited. For permission to use (where not already granted under a licence) please go to http://www.bmj.com/company/products-services/rights-and-licensing/
An iconic programming language for sensor-based robots
NASA Technical Reports Server (NTRS)
Gertz, Matthew; Stewart, David B.; Khosla, Pradeep K.
1993-01-01
In this paper we describe an iconic programming language called Onika for sensor-based robotic systems. Onika is both modular and reconfigurable and can be used with any system architecture and real-time operating system. Onika is also a multi-level programming environment wherein tasks are built by connecting a series of icons which, in turn, can be defined in terms of other icons at the lower levels. Expert users are also allowed to use control block form to define servo tasks. The icons in Onika are both shape and color coded, like the pieces of a jigsaw puzzle, thus providing a form of error control in the development of high level applications.
Plant identification credibility in ethnobotany: a closer look at Polish ethnographic studies
2010-01-01
Background This paper is an attempt to estimate the percentage of erroneously identified taxa in ethnographic studies concerning the use of plants and to propose a code for recording credibility of identification in historical ethnobotany publications. Methods A sample of Polish-language ethnobotanical literature (45 published sources from 1874-2005) and four collections of voucher specimens (from 1894-1975) were analyzed. Errors were detected in the publications by comparing the data with existing knowledge on the distribution of plant names and species ranges. The voucher specimens were re-examined. A one-letter code was invented for quick identification of the credibility of data published in lists of species compiled from historical or ethnographic sources, according to the source of identification: voucher specimen, Latin binominal, botanical expert, obvious widespread name, folk name, mode of use, range, physical description or photograph. To test the use of the code an up-to-date list of wild food plants used in Poland was made. Results A significant difference between the ratio of mistakes in the voucher specimen collections and the ratio of detectable mistakes in the studies without herbarium documentation was found. At least 2.3% of taxa in the publications were identified erroneously (mean rate was 6.2% per publication), and in half of these mistakes even the genus was not correct. As many as 10.0% of voucher specimens (on average 9.2% per collection) were originally erroneously identified, but three quarters of the identification mistakes remained within-genus. The species of the genera Thymus, Rumex and Rubus were most often confused within the genus. Not all of the invented credibility codes were used in the list of wild food plants, but they may be useful for other researchers. The most often used codes were the ones signifying identification by: voucher specimen, botanical expert and by a common name used throughout the country. Conclusions The results of this study support the rigorous use of voucher specimens in ethnobotany, although they also reveal a relatively high percentage of misidentified taxa in the specimens studied. The invented credibility coding system may become a useful tool for communication between historical ethnobotanists, particularly in creating larger databases. PMID:21167056
A brief history and technical review of the expert system research
NASA Astrophysics Data System (ADS)
Tan, Haocheng
2017-09-01
The expert system is a computer system that emulates the decision-making ability of a human expert, which aims to solve complex problems by reasoning knowledge. It is an important branch of artificial intelligence. In this paper, firstly, we briefly introduce the development and basic structure of the expert system. Then, from the perspective of the enabling technology, we classify the current expert systems and elaborate four expert systems: The Rule-Based Expert System, the Framework-Based Expert System, the Fuzzy Logic-Based Expert System and the Expert System Based on Neural Network.
Automatic translation of MPI source into a latency-tolerant, data-driven form
DOE Office of Scientific and Technical Information (OSTI.GOV)
Nguyen, Tan; Cicotti, Pietro; Bylaska, Eric
Hiding communication behind useful computation is an important performance programming technique but remains an inscrutable programming exercise even for the expert. We present Bamboo, a code transformation framework that can realize communication overlap in applications written in MPI without the need to intrusively modify the source code. We reformulate MPI source into a task dependency graph representation, which partially orders the tasks, enabling the program to execute in a data-driven fashion under the control of an external runtime system. Experimental results demonstrate that Bamboo significantly reduces communication delays while requiring only modest amounts of programmer annotation for a variety ofmore » applications and platforms, including those employing co-processors and accelerators. Moreover, Bamboo’s performance meets or exceeds that of labor-intensive hand coding. As a result, the translator is more than a means of hiding communication costs automatically; it demonstrates the utility of semantic level optimization against a well-known library.« less
Automatic translation of MPI source into a latency-tolerant, data-driven form
Nguyen, Tan; Cicotti, Pietro; Bylaska, Eric; ...
2017-03-06
Hiding communication behind useful computation is an important performance programming technique but remains an inscrutable programming exercise even for the expert. We present Bamboo, a code transformation framework that can realize communication overlap in applications written in MPI without the need to intrusively modify the source code. We reformulate MPI source into a task dependency graph representation, which partially orders the tasks, enabling the program to execute in a data-driven fashion under the control of an external runtime system. Experimental results demonstrate that Bamboo significantly reduces communication delays while requiring only modest amounts of programmer annotation for a variety ofmore » applications and platforms, including those employing co-processors and accelerators. Moreover, Bamboo’s performance meets or exceeds that of labor-intensive hand coding. As a result, the translator is more than a means of hiding communication costs automatically; it demonstrates the utility of semantic level optimization against a well-known library.« less
Automatic translation of MPI source into a latency-tolerant, data-driven form
DOE Office of Scientific and Technical Information (OSTI.GOV)
Nguyen, Tan; Cicotti, Pietro; Bylaska, Eric
Hiding communication behind useful computation is an important performance programming technique but remains an inscrutable programming exercise even for the expert. We present Bamboo, a code transformation framework that can realize communication overlap in applications written in MPI without the need to intrusively modify the source code. Bamboo reformulates MPI source into the form of a task dependency graph that expresses a partial ordering among tasks, enabling the program to execute in a data-driven fashion under the control of an external runtime system. Experimental results demonstrate that Bamboo significantly reduces communication delays while requiring only modest amounts of programmer annotationmore » for a variety of applications and platforms, including those employing co-processors and accelerators. Moreover, Bamboo's performance meets or exceeds that of labor-intensive hand coding. The translator is more than a means of hiding communication costs automatically; it demonstrates the utility of semantic level optimization against a wellknown library.« less
An advanced artificial intelligence tool for menu design.
Khan, Abdus Salam; Hoffmann, Achim
2003-01-01
The computer-assisted menu design still remains a difficult task. Usually knowledge that aids in menu design by a computer is hard-coded and because of that a computerised menu planner cannot handle the menu design problem for an unanticipated client. To address this problem we developed a menu design tool, MIKAS (menu construction using incremental knowledge acquisition system), an artificial intelligence system that allows the incremental development of a knowledge-base for menu design. We allow an incremental knowledge acquisition process in which the expert is only required to provide hints to the system in the context of actual problem instances during menu design using menus stored in a so-called Case Base. Our system incorporates Case-Based Reasoning (CBR), an Artificial Intelligence (AI) technique developed to mimic human problem solving behaviour. Ripple Down Rules (RDR) are a proven technique for the acquisition of classification knowledge from expert directly while they are using the system, which complement CBR in a very fruitful way. This combination allows the incremental improvement of the menu design system while it is already in routine use. We believe MIKAS allows better dietary practice by leveraging a dietitian's skills and expertise. As such MIKAS has the potential to be helpful for any institution where dietary advice is practised.
Expert Systems: An Overview for Teacher-Librarians.
ERIC Educational Resources Information Center
Orwig, Gary; Barron, Ann
1992-01-01
Provides an overview of expert systems for teacher librarians. Highlights include artificial intelligence and expert systems; the development of the MYCIN medical expert system; rule-based expert systems; the use of expert system shells to develop a specific system; and how to select an appropriate application for an expert system. (11 references)…
Proposal for a CLIPS software library
NASA Technical Reports Server (NTRS)
Porter, Ken
1991-01-01
This paper is a proposal to create a software library for the C Language Integrated Production System (CLIPS) expert system shell developed by NASA. Many innovative ideas for extending CLIPS were presented at the First CLIPS Users Conference, including useful user and database interfaces. CLIPS developers would benefit from a software library of reusable code. The CLIPS Users Group should establish a software library-- a course of action to make that happen is proposed. Open discussion to revise this library concept is essential, since only a group effort is likely to succeed. A response form intended to solicit opinions and support from the CLIPS community is included.
An Experiment in Scientific Code Semantic Analysis
NASA Technical Reports Server (NTRS)
Stewart, Mark E. M.
1998-01-01
This paper concerns a procedure that analyzes aspects of the meaning or semantics of scientific and engineering code. This procedure involves taking a user's existing code, adding semantic declarations for some primitive variables, and parsing this annotated code using multiple, distributed expert parsers. These semantic parser are designed to recognize formulae in different disciplines including physical and mathematical formulae and geometrical position in a numerical scheme. The parsers will automatically recognize and document some static, semantic concepts and locate some program semantic errors. Results are shown for a subroutine test case and a collection of combustion code routines. This ability to locate some semantic errors and document semantic concepts in scientific and engineering code should reduce the time, risk, and effort of developing and using these codes.
Simulations of pattern dynamics for reaction-diffusion systems via SIMULINK
2014-01-01
Background Investigation of the nonlinear pattern dynamics of a reaction-diffusion system almost always requires numerical solution of the system’s set of defining differential equations. Traditionally, this would be done by selecting an appropriate differential equation solver from a library of such solvers, then writing computer codes (in a programming language such as C or Matlab) to access the selected solver and display the integrated results as a function of space and time. This “code-based” approach is flexible and powerful, but requires a certain level of programming sophistication. A modern alternative is to use a graphical programming interface such as Simulink to construct a data-flow diagram by assembling and linking appropriate code blocks drawn from a library. The result is a visual representation of the inter-relationships between the state variables whose output can be made completely equivalent to the code-based solution. Results As a tutorial introduction, we first demonstrate application of the Simulink data-flow technique to the classical van der Pol nonlinear oscillator, and compare Matlab and Simulink coding approaches to solving the van der Pol ordinary differential equations. We then show how to introduce space (in one and two dimensions) by solving numerically the partial differential equations for two different reaction-diffusion systems: the well-known Brusselator chemical reactor, and a continuum model for a two-dimensional sheet of human cortex whose neurons are linked by both chemical and electrical (diffusive) synapses. We compare the relative performances of the Matlab and Simulink implementations. Conclusions The pattern simulations by Simulink are in good agreement with theoretical predictions. Compared with traditional coding approaches, the Simulink block-diagram paradigm reduces the time and programming burden required to implement a solution for reaction-diffusion systems of equations. Construction of the block-diagram does not require high-level programming skills, and the graphical interface lends itself to easy modification and use by non-experts. PMID:24725437
Cross-terminology mapping challenges: a demonstration using medication terminological systems.
Saitwal, Himali; Qing, David; Jones, Stephen; Bernstam, Elmer V; Chute, Christopher G; Johnson, Todd R
2012-08-01
Standardized terminological systems for biomedical information have provided considerable benefits to biomedical applications and research. However, practical use of this information often requires mapping across terminological systems-a complex and time-consuming process. This paper demonstrates the complexity and challenges of mapping across terminological systems in the context of medication information. It provides a review of medication terminological systems and their linkages, then describes a case study in which we mapped proprietary medication codes from an electronic health record to SNOMED CT and the UMLS Metathesaurus. The goal was to create a polyhierarchical classification system for querying an i2b2 clinical data warehouse. We found that three methods were required to accurately map the majority of actively prescribed medications. Only 62.5% of source medication codes could be mapped automatically. The remaining codes were mapped using a combination of semi-automated string comparison with expert selection, and a completely manual approach. Compound drugs were especially difficult to map: only 7.5% could be mapped using the automatic method. General challenges to mapping across terminological systems include (1) the availability of up-to-date information to assess the suitability of a given terminological system for a particular use case, and to assess the quality and completeness of cross-terminology links; (2) the difficulty of correctly using complex, rapidly evolving, modern terminologies; (3) the time and effort required to complete and evaluate the mapping; (4) the need to address differences in granularity between the source and target terminologies; and (5) the need to continuously update the mapping as terminological systems evolve. Copyright © 2012 Elsevier Inc. All rights reserved.
Hefelfinger, Jenny; Patty, Alice; Ussery, Ann; Young, Walter
2013-10-24
This study assessed the value of technical assistance provided by state health department expert advisors and by the staff of the National Association of Chronic Disease Directors (NACDD) to community groups that participated in the Action Communities for Health, Innovation, and Environmental Change (ACHIEVE) Program, a CDC-funded health promotion program. We analyzed quantitative and qualitative data reported by community project coordinators to assess the nature and value of technical assistance provided by expert advisors and NACDD staff and the usefulness of ACHIEVE resources in the development and implementation of community action plans. A grounded theory approach was used to analyze and categorize phrases in text data provided by community coordinators. Open coding placed conceptual labels on text phrases. Frequency distributions of the quantitative data are described and discussed. The most valuable technical assistance and program support resources were those determined to be in the interpersonal domain (ie, interactions with state expert advisors, NACDD staff, and peer-to-peer support). The most valuable technical assistance events were action institutes, coaches' meetings, webinars, and technical assistance conference calls. This analysis suggests that ACHIEVE communities valued the management and training assistance provided by expert advisors and NACDD staff. State health department expert advisors provided technical guidance and support, including such skills or knowledge-based services as best-practice strategies, review and discussion of community assessment data, sustainability planning, and identification of possible funding opportunities. NACDD staff led development and implementation of technical assistance events.
An Expert-System-Like Feedback Approach in the hp-Version of the Finite Element Method.
1986-05-01
and, besides some research codes, the authors know of only two commercial programs based on the p-version. These are the computer program PROBE ( Noetic ...assumptions. Let us first study the problem of the best approximation on the interval I=(-1,1) . Let for E < 1 (x + X - for x > 0 for x and let W(a,E,x) = (x...Comp. and Maths. with Appls., 5, pp. 99-115, 1979. [] Szabo, B., PROBE: Theoretical Manual, NOETIC Technology Corporation, 7980 Clayton Road, Suite 205
Hyshka, Elaine; Karekezi, Kamagaju; Tan, Benjamin; Slater, Linda G; Jahrig, Jesse; Wild, T Cameron
2017-03-20
A growing body of research assesses population need for substance use services. However, the extent to which survey research incorporates expert versus consumer perspectives on service need is unknown. We conducted a large, international review to (1) describe extant research on population need for substance use services, and the extent to which it incorporates expert and consumer perspectives on service need, (2) critically assess methodological and measurement approaches used to study consumer-defined need, and (3) examine the potential for existing research that prioritizes consumer perspectives to inform substance use service system planning. Systematic searches of seven databases identified 1930 peer-reviewed articles addressing population need for substance use services between January 1980 and May 2015. Empirical studies (n = 1887) were categorized according to source(s) of data used to derive population estimates of service need (administrative records, biological samples, qualitative data, and/or quantitative surveys). Quantitative survey studies (n = 1594) were categorized as to whether service need was assessed from an expert and/or consumer perspective; studies employing consumer-defined need measures (n = 217) received further in-depth quantitative coding to describe study designs and measurement strategies. Almost all survey studies (96%; n = 1534) used diagnostically-oriented measures derived from an expert perspective to assess service need. Of the small number (14%, n = 217) of survey studies that assessed consumer's perspectives, most (77%) measured perceived need for generic services (i.e. 'treatment'), with fewer (42%) examining self-assessed barriers to service use, or informal help-seeking from family and friends (10%). Unstandardized measures were commonly used, and very little research was longitudinal or tested hypotheses. Only one study used a consumer-defined need measure to estimate required service system capacity. Rhetorical calls for including consumer perspectives in substance use service system planning are belied by the empirical literature, which is dominated by expert-driven approaches to measuring population need. Studies addressing consumer-defined need for substance use services are conceptually underdeveloped, and exhibit methodological and measurement weaknesses. Further scholarship is needed to integrate multidisciplinary perspectives in this literature, and fully realize the promise of incorporating consumer perspectives into substance use service system planning.
VizieR Online Data Catalog: LAMOST-Kepler MKCLASS spectral classification (Gray+, 2016)
NASA Astrophysics Data System (ADS)
Gray, R. O.; Corbally, C. J.; De Cat, P.; Fu, J. N.; Ren, A. B.; Shi, J. R.; Luo, A. L.; Zhang, H. T.; Wu, Y.; Cao, Z.; Li, G.; Zhang, Y.; Hou, Y.; Wang, Y.
2016-07-01
The data for the LAMOST-Kepler project are supplied by the Large Sky Area Multi Object Fiber Spectroscopic Telescope (LAMOST, also known as the Guo Shou Jing Telescope). This unique astronomical instrument is located at the Xinglong observatory in China, and combines a large aperture (4 m) telescope with a 5° circular field of view (Wang et al. 1996ApOpt..35.5155W). Our role in this project is to supply accurate two-dimensional spectral types for the observed targets. The large number of spectra obtained for this project (101086) makes traditional visual classification techniques impractical, so we have utilized the MKCLASS code to perform these classifications. The MKCLASS code (Gray & Corbally 2014AJ....147...80G, v1.07 http://www.appstate.edu/~grayro/mkclass/), an expert system designed to classify blue-violet spectra on the MK Classification system, was employed to produce the spectral classifications reported in this paper. MKCLASS was designed to reproduce the steps skilled human classifiers employ in the classification process. (2 data files).
Xiao, Bo; Huang, Chewei; Imel, Zac E; Atkins, David C; Georgiou, Panayiotis; Narayanan, Shrikanth S
2016-04-01
Scaling up psychotherapy services such as for addiction counseling is a critical societal need. One challenge is ensuring quality of therapy, due to the heavy cost of manual observational assessment. This work proposes a speech technology-based system to automate the assessment of therapist empathy-a key therapy quality index-from audio recordings of the psychotherapy interactions. We designed a speech processing system that includes voice activity detection and diarization modules, and an automatic speech recognizer plus a speaker role matching module to extract the therapist's language cues. We employed Maximum Entropy models, Maximum Likelihood language models, and a Lattice Rescoring method to characterize high vs. low empathic language. We estimated therapy-session level empathy codes using utterance level evidence obtained from these models. Our experiments showed that the fully automated system achieved a correlation of 0.643 between expert annotated empathy codes and machine-derived estimations, and an accuracy of 81% in classifying high vs. low empathy, in comparison to a 0.721 correlation and 86% accuracy in the oracle setting using manual transcripts. The results show that the system provides useful information that can contribute to automatic quality insurance and therapist training.
Xiao, Bo; Huang, Chewei; Imel, Zac E.; Atkins, David C.; Georgiou, Panayiotis; Narayanan, Shrikanth S.
2016-01-01
Scaling up psychotherapy services such as for addiction counseling is a critical societal need. One challenge is ensuring quality of therapy, due to the heavy cost of manual observational assessment. This work proposes a speech technology-based system to automate the assessment of therapist empathy—a key therapy quality index—from audio recordings of the psychotherapy interactions. We designed a speech processing system that includes voice activity detection and diarization modules, and an automatic speech recognizer plus a speaker role matching module to extract the therapist's language cues. We employed Maximum Entropy models, Maximum Likelihood language models, and a Lattice Rescoring method to characterize high vs. low empathic language. We estimated therapy-session level empathy codes using utterance level evidence obtained from these models. Our experiments showed that the fully automated system achieved a correlation of 0.643 between expert annotated empathy codes and machine-derived estimations, and an accuracy of 81% in classifying high vs. low empathy, in comparison to a 0.721 correlation and 86% accuracy in the oracle setting using manual transcripts. The results show that the system provides useful information that can contribute to automatic quality insurance and therapist training. PMID:28286867
A framework for development of an intelligent system for design and manufacturing of stamping dies
NASA Astrophysics Data System (ADS)
Hussein, H. M. A.; Kumar, S.
2014-07-01
An integration of computer aided design (CAD), computer aided process planning (CAPP) and computer aided manufacturing (CAM) is required for development of an intelligent system to design and manufacture stamping dies in sheet metal industries. In this paper, a framework for development of an intelligent system for design and manufacturing of stamping dies is proposed. In the proposed framework, the intelligent system is structured in form of various expert system modules for different activities of design and manufacturing of dies. All system modules are integrated with each other. The proposed system takes its input in form of a CAD file of sheet metal part, and then system modules automate all tasks related to design and manufacturing of stamping dies. Modules are coded using Visual Basic (VB) and developed on the platform of AutoCAD software.
Helping the police with their inquiries
NASA Astrophysics Data System (ADS)
Kitson, Anthony J.
1995-09-01
The UK Home Office has held a long term interest in facial recognition. Work has concentrated upon providing the UK police with facilities to improve the use that can be made of the memory of victims and witnesses rather than automatically matching images. During the 1970s a psychological coding scheme and a search method were developed by Aberdeen University and Home Office. This has been incorporated into systems for searching prisoner photographs both experimentally and operationally. The coding scheme has also been incorporated in a facial likeness composition system. The Home Office is currenly implementing a national criminal record system (Phoenix) and work has been conducted to define and demonstrate standards for image enabled terminals for this application. Users have been consulted to establish suitable picture quality for the purpose, and a study of compression methods is in hand. Recently there has been increased use made by UK courts of expert testimony based upon the measurement of facial images. We are currently working with a group of practitioners to examine and improve the quality of such evidence and to develop a national standard.
Robot navigation research using the HERMIES mobile robot
DOE Office of Scientific and Technical Information (OSTI.GOV)
Barnett, D.L.
1989-01-01
In recent years robot navigation has attracted much attention from researchers around the world. Not only are theoretical studies being simulated on sophisticated computers, but many mobile robots are now used as test vehicles for these theoretical studies. Various algorithms have been perfected for navigation in a known static environment; but navigation in an unknown and dynamic environment poses a much more challenging problem for researchers. Many different methodologies have been developed for autonomous robot navigation, but each methodology is usually restricted to a particular type of environment. One important research focus of the Center for Engineering Systems Advanced researchmore » (CESAR) at Oak Ridge National Laboratory, is autonomous navigation in unknown and dynamic environments using the series of HERMIES mobile robots. The research uses an expert system for high-level planning interfaced with C-coded routines for implementing the plans, and for quick processing of data requested by the expert system. In using this approach, the navigation is not restricted to one methodology since the expert system can activate a rule module for the methodology best suited for the current situation. Rule modules can be added the rule base as they are developed and tested. Modules are being developed or enhanced for navigating from a map, searching for a target, exploring, artificial potential-field navigation, navigation using edge-detection, etc. This paper will report on the various rule modules and methods of navigation in use, or under development at CESAR, using the HERMIES-IIB robot as a testbed. 13 refs., 5 figs., 1 tab.« less
There is a need for new systemic sclerosis subset criteria. A content analytic approach.
Johnson, S R; Soowamber, M L; Fransen, J; Khanna, D; Van Den Hoogen, F; Baron, M; Matucci-Cerinic, M; Denton, C P; Medsger, T A; Carreira, P E; Riemekasten, G; Distler, J; Gabrielli, A; Steen, V; Chung, L; Silver, R; Varga, J; Müller-Ladner, U; Vonk, M C; Walker, U A; Wollheim, F A; Herrick, A; Furst, D E; Czirjak, L; Kowal-Bielecka, O; Del Galdo, F; Cutolo, M; Hunzelmann, N; Murray, C D; Foeldvari, I; Mouthon, L; Damjanov, N; Kahaleh, B; Frech, T; Assassi, S; Saketkoo, L A; Pope, J E
2018-01-01
Systemic sclerosis (SSc) is heterogenous. The objectives of this study were to evaluate the purpose, strengths and limitations of existing SSc subset criteria, and identify ideas among experts about subsets. We conducted semi-structured interviews with randomly sampled international SSc experts. The interview transcripts underwent an iterative process with text deconstructed to single thought units until a saturated conceptual framework with coding was achieved and respondent occurrence tabulated. Serial cross-referential analyses of clusters were developed. Thirty experts from 13 countries were included; 67% were male, 63% were from Europe and 37% from North America; median experience of 22.5 years, with a median of 55 new SSc patients annually. Three thematic clusters regarding subsetting were identified: research and communication; management; and prognosis (prediction of internal organ involvement, survival). The strength of the limited/diffuse system was its ease of use, however 10% stated this system had marginal value. Shortcomings of the diffuse/limited classification were the risk of misclassification, predictions/generalizations did not always hold true, and that the elbow or knee threshold was arbitrary. Eighty-seven percent use more than 2 subsets including: SSc sine scleroderma, overlap conditions, antibody-determined subsets, speed of progression, and age of onset (juvenile, elderly). We have synthesized an international view of the construct of SSc subsets in the modern era. We found a number of factors underlying the construct of SSc subsets. Considerations for the next phase include rate of change and hierarchal clustering (e.g. limited/diffuse, then by antibodies).
Cross-terminology mapping challenges: A demonstration using medication terminological systems
Saitwal, Himali; Qing, David; Jones, Stephen; Bernstam, Elmer; Chute, Christopher G.; Johnson, Todd R.
2015-01-01
Standardized terminological systems for biomedical information have provided considerable benefits to biomedical applications and research. However, practical use of this information often requires mapping across terminological systems—a complex and time-consuming process. This paper demonstrates the complexity and challenges of mapping across terminological systems in the context of medication information. It provides a review of medication terminological systems and their linkages, then describes a case study in which we mapped proprietary medication codes from an electronic health record to SNOMED-CT and the UMLS Metathesaurus. The goal was to create a polyhierarchical classification system for querying an i2b2 clinical data warehouse. We found that three methods were required to accurately map the majority of actively prescribed medications. Only 62.5% of source medication codes could be mapped automatically. The remaining codes were mapped using a combination of semi-automated string comparison with expert selection, and a completely manual approach. Compound drugs were especially difficult to map: only 7.5% could be mapped using the automatic method. General challenges to mapping across terminological systems include (1) the availability of up-to-date information to assess the suitability of a given terminological system for a particular use case, and to assess the quality and completeness of cross-terminology links; (2) the difficulty of correctly using complex, rapidly evolving, modern terminologies; (3) the time and effort required to complete and evaluate the mapping; (4) the need to address differences in granularity between the source and target terminologies; and (5) the need to continuously update the mapping as terminological systems evolve. PMID:22750536
A meta-model for computer executable dynamic clinical safety checklists.
Nan, Shan; Van Gorp, Pieter; Lu, Xudong; Kaymak, Uzay; Korsten, Hendrikus; Vdovjak, Richard; Duan, Huilong
2017-12-12
Safety checklist is a type of cognitive tool enforcing short term memory of medical workers with the purpose of reducing medical errors caused by overlook and ignorance. To facilitate the daily use of safety checklists, computerized systems embedded in the clinical workflow and adapted to patient-context are increasingly developed. However, the current hard-coded approach of implementing checklists in these systems increase the cognitive efforts of clinical experts and coding efforts for informaticists. This is due to the lack of a formal representation format that is both understandable by clinical experts and executable by computer programs. We developed a dynamic checklist meta-model with a three-step approach. Dynamic checklist modeling requirements were extracted by performing a domain analysis. Then, existing modeling approaches and tools were investigated with the purpose of reusing these languages. Finally, the meta-model was developed by eliciting domain concepts and their hierarchies. The feasibility of using the meta-model was validated by two case studies. The meta-model was mapped to specific modeling languages according to the requirements of hospitals. Using the proposed meta-model, a comprehensive coronary artery bypass graft peri-operative checklist set and a percutaneous coronary intervention peri-operative checklist set have been developed in a Dutch hospital and a Chinese hospital, respectively. The result shows that it is feasible to use the meta-model to facilitate the modeling and execution of dynamic checklists. We proposed a novel meta-model for the dynamic checklist with the purpose of facilitating creating dynamic checklists. The meta-model is a framework of reusing existing modeling languages and tools to model dynamic checklists. The feasibility of using the meta-model is validated by implementing a use case in the system.
Hepworth, S J; Bolton, A; Parslow, R C; van Tongeren, M; Muir, K R; McKinney, P A
2006-04-01
To compare assignment of occupational pesticide and solvent exposure using self-reported data collected by a computer assisted personal interview (CAPI) with exposure based on expert assessment of job codes. To discuss the advantages and disadvantages of using a CAPI to collect individual occupational exposure data. Between 2001 and 2004, 1495 participants were interviewed using a CAPI for a case-control study of adult brain tumours and acoustic neuromas. Two types of occupational data were collected: (1) a full history, including job title from which a job code was assigned from the Standard Occupational Classification; and (2) specific details on pesticide and solvent exposure reported by participants. Study members' experiences of using the CAPI were recorded and advantages and disadvantages summarised. Of 7192 jobs recorded, the prevalence of self-reported exposure was 1.3% for pesticides and 11.5% for solvents. Comparing this with exposure expertly assessed from job titles showed 53.6% and 45.8% concordance for pesticides and solvents respectively. Advantages of the CAPI include no data entry stage, automatic input validation, and a reduction in interviewer bias. Disadvantages include an adverse effect on study implementation as a consequence of resources required for programming and difficulties encountered with data management prior to analysis. Different methods of exposure assessment derive different exposure levels for pesticide and solvent exposure at work. Agreement between self-reported and expert assessment of exposure was greater for pesticides compared to solvents. The advantages of using a CAPI for the collection of complex data outweigh the disadvantages for interviewers and data quality but using such a method requires extra resources at the study outset.
Hepworth, S J; Bolton, A; Parslow, R C; van Tongeren, M; Muir, K R; McKinney, P A
2006-01-01
Objectives To compare assignment of occupational pesticide and solvent exposure using self‐reported data collected by a computer assisted personal interview (CAPI) with exposure based on expert assessment of job codes. To discuss the advantages and disadvantages of using a CAPI to collect individual occupational exposure data. Methods Between 2001 and 2004, 1495 participants were interviewed using a CAPI for a case‐control study of adult brain tumours and acoustic neuromas. Two types of occupational data were collected: (1) a full history, including job title from which a job code was assigned from the Standard Occupational Classification; and (2) specific details on pesticide and solvent exposure reported by participants. Study members' experiences of using the CAPI were recorded and advantages and disadvantages summarised. Results Of 7192 jobs recorded, the prevalence of self‐reported exposure was 1.3% for pesticides and 11.5% for solvents. Comparing this with exposure expertly assessed from job titles showed 53.6% and 45.8% concordance for pesticides and solvents respectively. Advantages of the CAPI include no data entry stage, automatic input validation, and a reduction in interviewer bias. Disadvantages include an adverse effect on study implementation as a consequence of resources required for programming and difficulties encountered with data management prior to analysis. Conclusions Different methods of exposure assessment derive different exposure levels for pesticide and solvent exposure at work. Agreement between self‐reported and expert assessment of exposure was greater for pesticides compared to solvents. The advantages of using a CAPI for the collection of complex data outweigh the disadvantages for interviewers and data quality but using such a method requires extra resources at the study outset. PMID:16556747
Standard terminology and labeling of ocular tissue for transplantation.
Armitage, W John; Ashford, Paul; Crow, Barbara; Dahl, Patricia; DeMatteo, Jennifer; Distler, Pat; Gopinathan, Usha; Madden, Peter W; Mannis, Mark J; Moffatt, S Louise; Ponzin, Diego; Tan, Donald
2013-06-01
To develop an internationally agreed terminology for describing ocular tissue grafts to improve the accuracy and reliability of information transfer, to enhance tissue traceability, and to facilitate the gathering of comparative global activity data, including denominator data for use in biovigilance analyses. ICCBBA, the international standards organization for terminology, coding, and labeling of blood, cells, and tissues, approached the major Eye Bank Associations to form an expert advisory group. The group met by regular conference calls to develop a standard terminology, which was released for public consultation and amended accordingly. The terminology uses broad definitions (Classes) with modifying characteristics (Attributes) to define each ocular tissue product. The terminology may be used within the ISBT 128 system to label tissue products with standardized bar codes enabling the electronic capture of critical data in the collection, processing, and distribution of tissues. Guidance on coding and labeling has also been developed. The development of a standard terminology for ocular tissue marks an important step for improving traceability and reducing the risk of mistakes due to transcription errors. ISBT 128 computer codes have been assigned and may now be used to label ocular tissues. Eye banks are encouraged to adopt this standard terminology and move toward full implementation of ISBT 128 nomenclature, coding, and labeling.
Romero-Fernández, Ma Mar; Royo-Bordonada, Miguel Angel; Rodríguez-Artalejo, Fernando
2010-07-01
To evaluate the level of compliance with the PAOS Code (Publicidad, Actividad, Obesidad y Salud), which establishes standards for the self-regulation of food marketing aimed at minors, in television advertising by food and beverage companies that have agreed to abide by the Code. The study sample consisted of food and beverage advertisements targeting children during 80 h of programming by four Spanish television networks. The level of compliance with each standard of the PAOS Code was classified into three categories: 'compliance', 'non-compliance' and 'uncertain compliance'. Overall, an advertisement was considered compliant with the PAOS Code if it met all the standards; non-compliant if it contravened one or more standards; and uncertain in all other cases. Of a total of 203 television advertisements from companies that agreed to the PAOS Code, the overall prevalence of non-compliance was 49.3% (v. 50.8% among those that did not agree to the code), with 20.7% of advertisements considered of uncertain compliance. Non-compliance was more frequent on Saturdays, in longer advertisements, in advertisements containing promotions or dairy products, and for advertisements from companies of French or US origin. Non-compliance with the PAOS Code was very high and was similar for companies that did and did not agree to the Code, casting doubt on the Code's effectiveness and oversight system. It seems the time has come to commit to statutory regulations that reduce the negative impact of advertising on children's diets, as demanded by public health experts and consumer associations.
An Experiment in Scientific Program Understanding
NASA Technical Reports Server (NTRS)
Stewart, Mark E. M.; Owen, Karl (Technical Monitor)
2000-01-01
This paper concerns a procedure that analyzes aspects of the meaning or semantics of scientific and engineering code. This procedure involves taking a user's existing code, adding semantic declarations for some primitive variables, and parsing this annotated code using multiple, independent expert parsers. These semantic parsers encode domain knowledge and recognize formulae in different disciplines including physics, numerical methods, mathematics, and geometry. The parsers will automatically recognize and document some static, semantic concepts and help locate some program semantic errors. Results are shown for three intensively studied codes and seven blind test cases; all test cases are state of the art scientific codes. These techniques may apply to a wider range of scientific codes. If so, the techniques could reduce the time, risk, and effort required to develop and modify scientific codes.
Implementation of a frame-based representation in CLIPS
NASA Technical Reports Server (NTRS)
Assal, Hisham; Myers, Leonard
1990-01-01
Knowledge representation is one of the major concerns in expert systems. The representation of domain-specific knowledge should agree with the nature of the domain entities and their use in the real world. For example, architectural applications deal with objects and entities such as spaces, walls, and windows. A natural way of representing these architectural entities is provided by frames. This research explores the potential of using the expert system shell CLIPS, developed by NASA, to implement a frame-based representation that can accommodate architectural knowledge. These frames are similar but quite different from the 'template' construct in version 4.3 of CLIPS. Templates support only the grouping of related information and the assignment of default values to template fields. In addition to these features frames provide other capabilities including definition of classes, inheritance between classes and subclasses, relation of objects of different classes with 'has-a', association of methods (demons) of different types (standard and user-defined) to fields (slots), and creation of new fields at run-time. This frame-based representation is implemented completely in CLIPS. No change to the source code is necessary.
ERIC Educational Resources Information Center
Adiga, Sadashiv
1984-01-01
Discusses: (1) the architecture of expert systems; (2) features that distinguish expert systems from conventional programs; (3) conditions necessary to select a particular application for the development of successful expert systems; (4) issues to be resolved when building expert systems; and (5) limitations. Examples of selected expert systems…
[Medico-legal autopsy--selected legal issues: the autopsy protocol].
Gaszczyk-Ozarowski, Zbigniew; Chowaniec, Czesław
2010-01-01
The majority of experts in the field of forensic medicine maintain that the minutes of the medicolegal autopsy should be taken by the forensic pathologist. The authors argue that it is the public prosecutor who is obliged to draw up the minutes, whereas the forensic pathologist issues the expert opinion. To support their stance, the authors make frequent references to several provisions of the Criminal Procedure Code of 1997. The authors also imply that due to organizational reasons and the ratio legis of the aforementioned code, the forensic pathologist should not be assigned the role of the minutes-taker, despite the lack of a specific exclusion rule governing such a case. Possible consequences caused by the lack of the properly drawn up minutes are briefly discussed as well.
Lloyd, Kelly; Cameron, Elaine; Williams, Hannah; Banister, Emma; Donmall, Michael; Higgins, Alan; French, David P
2018-04-01
Televised alcohol advertisements in the United Kingdom must abide by the Broadcast Committee of Advertising Practice Code, which provides guidelines concerning advertisements not implying, condoning or encouraging immoderate, irresponsible or antisocial drinking. Previously, 75 per cent of 373 general public respondents were shown one of seven advertisements rated a breach of at least one guideline. This study assessed whether experts in marketing ( n = 25) and alcohol treatment/public health ( n = 25) perceived the same seven television alcohol advertisements as complying with the Broadcast Committee of Advertising Practice Code. Overall, 83 per cent of advertisements were rated as breaching at least one guideline. This provides further proof that self-regulatory alcohol guidelines are not fit for purpose.
Telemetry: Summary of concept and rationale
NASA Astrophysics Data System (ADS)
1987-12-01
This report presents the concept and supporting rationale for the telemetry system developed by the Consultative Committee for Space Data Systems (CCSDS). The concepts, protocols and data formats developed for the telemetry system are designed for flight and ground data systems supporting conventional, contemporary free-flyer spacecraft. Data formats are designed with efficiency as a primary consideration, i.e., format overhead is minimized. The results reflect the consensus of experts from many space agencies. An overview of the CCSDS telemetry system introduces the notion of architectural layering to achieve transparent and reliable delivery of scientific and engineering sensor data (generated aboard space vehicles) to users located in space or on earth. The system is broken down into two major conceptual categories: a packet telemetry concept and a telemetry channel coding concept. Packet telemetry facilitates data transmission from source to user in a standardized and highly automated manner. It provides a mechanism for implementing common data structures and protocols which can enhance the development and operation of space mission systems. Telemetry channel coding is a method by which data can be sent from a source to a destination by processing it in such a way that distinct messages are created which are easily distinguishable from one another. This allows construction of the data with low error probability, thus improving performance of the channel.
Development of an occupational airborne chemical exposure matrix
Kurmi, O. P.; Chambers, H.; Lam, K. B. H.; Fishwick, D.
2016-01-01
Background Population-based studies of the occupational contribution to chronic obstructive pulmonary disease generally rely on self-reported exposures to vapours, gases, dusts and fumes (VGDF), which are susceptible to misclassification. Aims To develop an airborne chemical job exposure matrix (ACE JEM) for use with the UK Standard Occupational Classification (SOC 2000) system. Methods We developed the ACE JEM in stages: (i) agreement of definitions, (ii) a binary assignation of exposed/not exposed to VGDF, fibres or mists (VGDFFiM), for each of the individual 353 SOC codes and (iii) assignation of levels of exposure (L; low, medium and high) and (iv) the proportion of workers (P) likely to be exposed in each code. We then expanded the estimated exposures to include biological dusts, mineral dusts, metals, diesel fumes and asthmagens. Results We assigned 186 (53%) of all SOC codes as exposed to at least one category of VGDFFiM, with 23% assigned as having medium or high exposure. We assigned over 68% of all codes as not being exposed to fibres, gases or mists. The most common exposure was to dusts (22% of codes with >50% exposed); 12% of codes were assigned exposure to fibres. We assigned higher percentages of the codes as exposed to diesel fumes (14%) compared with metals (8%). Conclusions We developed an expert-derived JEM, using a strict set of a priori defined rules. The ACE JEM could also be applied to studies to assess risks of diseases where the main route of occupational exposure is via inhalation. PMID:27067914
Development of an occupational airborne chemical exposure matrix.
Sadhra, S S; Kurmi, O P; Chambers, H; Lam, K B H; Fishwick, D
2016-07-01
Population-based studies of the occupational contribution to chronic obstructive pulmonary disease generally rely on self-reported exposures to vapours, gases, dusts and fumes (VGDF), which are susceptible to misclassification. To develop an airborne chemical job exposure matrix (ACE JEM) for use with the UK Standard Occupational Classification (SOC 2000) system. We developed the ACE JEM in stages: (i) agreement of definitions, (ii) a binary assignation of exposed/not exposed to VGDF, fibres or mists (VGDFFiM), for each of the individual 353 SOC codes and (iii) assignation of levels of exposure (L; low, medium and high) and (iv) the proportion of workers (P) likely to be exposed in each code. We then expanded the estimated exposures to include biological dusts, mineral dusts, metals, diesel fumes and asthmagens. We assigned 186 (53%) of all SOC codes as exposed to at least one category of VGDFFiM, with 23% assigned as having medium or high exposure. We assigned over 68% of all codes as not being exposed to fibres, gases or mists. The most common exposure was to dusts (22% of codes with >50% exposed); 12% of codes were assigned exposure to fibres. We assigned higher percentages of the codes as exposed to diesel fumes (14%) compared with metals (8%). We developed an expert-derived JEM, using a strict set of a priori defined rules. The ACE JEM could also be applied to studies to assess risks of diseases where the main route of occupational exposure is via inhalation. © Crown copyright 2016.
Expert systems in civil engineering
DOE Office of Scientific and Technical Information (OSTI.GOV)
Kostem, C.N.; Maher, M.L.
1986-01-01
This book presents the papers given at a symposium on expert systems in civil engineering. Topics considered at the symposium included problem solving using expert system techniques, construction schedule analysis, decision making and risk analysis, seismic risk analysis systems, an expert system for inactive hazardous waste site characterization, an expert system for site selection, knowledge engineering, and knowledge-based expert systems in seismic analysis.
Rocket Engine Numerical Simulator (RENS)
NASA Technical Reports Server (NTRS)
Davidian, Kenneth O.
1997-01-01
Work is being done at three universities to help today's NASA engineers use the knowledge and experience of their Apolloera predecessors in designing liquid rocket engines. Ground-breaking work is being done in important subject areas to create a prototype of the most important functions for the Rocket Engine Numerical Simulator (RENS). The goal of RENS is to develop an interactive, realtime application that engineers can utilize for comprehensive preliminary propulsion system design functions. RENS will employ computer science and artificial intelligence research in knowledge acquisition, computer code parallelization and objectification, expert system architecture design, and object-oriented programming. In 1995, a 3year grant from the NASA Lewis Research Center was awarded to Dr. Douglas Moreman and Dr. John Dyer of Southern University at Baton Rouge, Louisiana, to begin acquiring knowledge in liquid rocket propulsion systems. Resources of the University of West Florida in Pensacola were enlisted to begin the process of enlisting knowledge from senior NASA engineers who are recognized experts in liquid rocket engine propulsion systems. Dr. John Coffey of the University of West Florida is utilizing his expertise in interviewing and concept mapping techniques to encode, classify, and integrate information obtained through personal interviews. The expertise extracted from the NASA engineers has been put into concept maps with supporting textual, audio, graphic, and video material. A fundamental concept map was delivered by the end of the first year of work and the development of maps containing increasing amounts of information is continuing. Find out more information about this work at the Southern University/University of West Florida. In 1996, the Southern University/University of West Florida team conducted a 4day group interview with a panel of five experts to discuss failures of the RL10 rocket engine in conjunction with the Centaur launch vehicle. The discussion was recorded on video and audio tape. Transcriptions of the entire proceedings and an abbreviated video presentation of the discussion highlights are under development. Also in 1996, two additional 3year grants were awarded to conduct parallel efforts that would complement the work being done by Southern University and the University of West Florida. Dr. Prem Bhalla of Jackson State University in Jackson, Mississippi, is developing the architectural framework for RENS. By employing the Rose Rational language and Booch Object Oriented Programming (OOP) technology, Dr. Bhalla is developing the basic structure of RENS by identifying and encoding propulsion system components, their individual characteristics, and cross-functionality and dependencies. Dr. Ruknet Cezzar of Hampton University, located in Hampton, Virginia, began working on the parallelization and objectification of rocket engine analysis and design codes. Dr. Cezzar will use the Turbo C++ OOP language to translate important liquid rocket engine computer codes from FORTRAN and permit their inclusion into the RENS framework being developed at Jackson State University. The Southern University/University of West Florida grant was extended by 1 year to coordinate the conclusion of all three efforts in 1999.
A CLIPS based personal computer hardware diagnostic system
NASA Technical Reports Server (NTRS)
Whitson, George M.
1991-01-01
Often the person designated to repair personal computers has little or no knowledge of how to repair a computer. Described here is a simple expert system to aid these inexperienced repair people. The first component of the system leads the repair person through a number of simple system checks such as making sure that all cables are tight and that the dip switches are set correctly. The second component of the system assists the repair person in evaluating error codes generated by the computer. The final component of the system applies a large knowledge base to attempt to identify the component of the personal computer that is malfunctioning. We have implemented and tested our design with a full system to diagnose problems for an IBM compatible system based on the 8088 chip. In our tests, the inexperienced repair people found the system very useful in diagnosing hardware problems.
A Model Code of Ethics for the Use of Computers in Education.
ERIC Educational Resources Information Center
Shere, Daniel T.; Cannings, Terence R.
Two Delphi studies were conducted by the Ethics and Equity Committee of the International Council for Computers in Education (ICCE) to obtain the opinions of experts on areas that should be covered by ethical guides for the use of computers in education and for software development, and to develop a model code of ethics for each of these areas.…
Unified English Braille in the United Kingdom: Part 1--Examination by Technical Expert Braille Users
ERIC Educational Resources Information Center
Cryer, Heather; Home, Sarah; Morley Wilkins, Sarah
2013-01-01
The Unified English Braille (UEB) code has already been adopted by various countries. To inform the decision about UK adoption, a suite of research was carried out with UK braille readers and other stakeholders. Previous research indicates that readers of technical braille codes may be most affected by a move to UEB. In this study, six technical…
An 802.11 n wireless local area network transmission scheme for wireless telemedicine applications.
Lin, C F; Hung, S I; Chiang, I H
2010-10-01
In this paper, an 802.11 n transmission scheme is proposed for wireless telemedicine applications. IEEE 802.11n standards, a power assignment strategy, space-time block coding (STBC), and an object composition Petri net (OCPN) model are adopted. With the proposed wireless system, G.729 audio bit streams, Joint Photographic Experts Group 2000 (JPEG 2000) clinical images, and Moving Picture Experts Group 4 (MPEG-4) video bit streams achieve a transmission bit error rate (BER) of 10-7, 10-4, and 103 simultaneously. The proposed system meets the requirements prescribed for wireless telemedicine applications. An essential feature of this proposed transmission scheme is that clinical information that requires a high quality of service (QoS) is transmitted at a high power transmission rate with significant error protection. For maximizing resource utilization and minimizing the total transmission power, STBC and adaptive modulation techniques are used in the proposed 802.11 n wireless telemedicine system. Further, low power, direct mapping (DM), low-error protection scheme, and high-level modulation are adopted for messages that can tolerate a high BER. With the proposed transmission scheme, the required reliability of communication can be achieved. Our simulation results have shown that the proposed 802.11 n transmission scheme can be used for developing effective wireless telemedicine systems.
Expert system application for the loading capability assessment of transmission lines
DOE Office of Scientific and Technical Information (OSTI.GOV)
Le, T.L.; Negnevitsky, M.; Piekutowski, M.
1995-11-01
This paper describes the application of an expert system for the evaluation of the short time thermal rating and temperature rise of overhead conductors. The expert system has been developed using a database and Leonardo expert system shell which is gaining popularity among commercial tools for developing expert system applications. The expert system has been found to compare well when evaluated against the site tests. A practical application is given to demonstrate the usefulness of the expert system developed.
NASA Astrophysics Data System (ADS)
Papers are presented on ISDN, mobile radio systems and techniques for digital connectivity, centralized and distributed algorithms in computer networks, communications networks, quality assurance and impact on cost, adaptive filters in communications, the spread spectrum, signal processing, video communication techniques, and digital satellite services. Topics discussed include performance evaluation issues for integrated protocols, packet network operations, the computer network theory and multiple-access, microwave single sideband systems, switching architectures, fiber optic systems, wireless local communications, modulation, coding, and synchronization, remote switching, software quality, transmission, and expert systems in network operations. Consideration is given to wide area networks, image and speech processing, office communications application protocols, multimedia systems, customer-controlled network operations, digital radio systems, channel modeling and signal processing in digital communications, earth station/on-board modems, computer communications system performance evaluation, source encoding, compression, and quantization, and adaptive communications systems.
NASA Technical Reports Server (NTRS)
Reed, John A.; Afjeh, Abdollah A.; Lewandowski, Henry; Homer, Patrick T.; Schlichting, Richard D.
1996-01-01
The NASA Numerical Propulsion System Simulation (NPSS) project is exploring the use of computer simulation to facilitate the design of new jet engines. Several key issues raised in this research are being examined in an NPSS-related research project: zooming, monitoring and control, and support for heterogeneity. The design of a simulation executive that addresses each of these issues is described. In this work, the strategy of zooming, which allows codes that model at different levels of fidelity to be integrated within a single simulation, is applied to the fan component of a turbofan propulsion system. A prototype monitoring and control system has been designed for this simulation to support experimentation with expert system techniques for active control of the simulation. An interconnection system provides a transparent means of connecting the heterogeneous systems that comprise the prototype.
Hefelfinger, Jenny; Patty, Alice; Ussery, Ann
2013-01-01
Introduction This study assessed the value of technical assistance provided by state health department expert advisors and by the staff of the National Association of Chronic Disease Directors (NACDD) to community groups that participated in the Action Communities for Health, Innovation, and Environmental Change (ACHIEVE) Program, a CDC-funded health promotion program. Methods We analyzed quantitative and qualitative data reported by community project coordinators to assess the nature and value of technical assistance provided by expert advisors and NACDD staff and the usefulness of ACHIEVE resources in the development and implementation of community action plans. A grounded theory approach was used to analyze and categorize phrases in text data provided by community coordinators. Open coding placed conceptual labels on text phrases. Frequency distributions of the quantitative data are described and discussed. Results The most valuable technical assistance and program support resources were those determined to be in the interpersonal domain (ie, interactions with state expert advisors, NACDD staff, and peer-to-peer support). The most valuable technical assistance events were action institutes, coaches’ meetings, webinars, and technical assistance conference calls. Conclusion This analysis suggests that ACHIEVE communities valued the management and training assistance provided by expert advisors and NACDD staff. State health department expert advisors provided technical guidance and support, including such skills or knowledge-based services as best-practice strategies, review and discussion of community assessment data, sustainability planning, and identification of possible funding opportunities. NACDD staff led development and implementation of technical assistance events. PMID:24157078
[The QR code in society, economy and medicine--fields of application, options and chances].
Flaig, Benno; Parzeller, Markus
2011-01-01
2D codes like the QR Code ("Quick Response") are becoming more and more common in society and medicine. The application spectrum and benefits in medicine and other fields are described. 2D codes can be created free of charge on any computer with internet access without any previous knowledge. The codes can be easily used in publications, presentations, on business cards and posters. Editors choose between contact details, text or a hyperlink as information behind the code. At expert conferences, linkage by QR Code allows the audience to download presentations and posters quickly. The documents obtained can then be saved, printed, processed etc. Fast access to stored data in the internet makes it possible to integrate additional and explanatory multilingual videos into medical posters. In this context, a combination of different technologies (printed handout, QR Code and screen) may be reasonable.
Data Collection Answers - SEER Registrars
Read clarifications to existing coding rules, which should be implemented immediately. Data collection experts from American College of Surgeons Commission on Cancer, CDC National Program of Cancer Registries, and SEER Program compiled these answers.
Choquet, Remy; Maaroufi, Meriem; Fonjallaz, Yannick; de Carrara, Albane; Vandenbussche, Pierre-Yves; Dhombres, Ferdinand; Landais, Paul
Characterizing a rare disease diagnosis for a given patient is often made through expert's networks. It is a complex task that could evolve over time depending on the natural history of the disease and the evolution of the scientific knowledge. Most rare diseases have genetic causes and recent improvements of sequencing techniques contribute to the discovery of many new diseases every year. Diagnosis coding in the rare disease field requires data from multiple knowledge bases to be aggregated in order to offer the clinician a global information space from possible diagnosis to clinical signs (phenotypes) and known genetic mutations (genotype). Nowadays, the major barrier to the coding activity is the lack of consolidation of such information scattered in different thesaurus such as Orphanet, OMIM or HPO. The Linking Open data for Rare Diseases (LORD) web portal we developed stands as the first attempt to fill this gap by offering an integrated view of 8,400 rare diseases linked to more than 14,500 signs and 3,270 genes. The application provides a browsing feature to navigate through the relationships between diseases, signs and genes, and some Application Programming Interfaces to help its integration in health information systems in routine.
Standardization of Terminology in Laboratory Medicine II
Lee, Kap No; Yoon, Jong-Hyun; Min, Won Ki; Lim, Hwan Sub; Song, Junghan; Chae, Seok Lae; Jang, Seongsoo; Ki, Chang-Seok; Bae, Sook Young; Kim, Jang Su; Kwon, Jung-Ah; Lee, Chang Kyu
2008-01-01
Standardization of medical terminology is essential in data transmission between health care institutes and in maximizing the benefits of information technology. The purpose of this study was to standardize medical terms for laboratory observations. During the second year of the study, a standard database of concept names for laboratory terms that covered those used in tertiary health care institutes and reference laboratories was developed. The laboratory terms in the Logical Observation Identifier Names and Codes (LOINC) database were adopted and matched with the electronic data interchange (EDI) codes in Korea. A public hearing and a workshop for clinical pathologists were held to collect the opinions of experts. The Korean standard laboratory terminology database containing six axial concept names, components, property, time aspect, system (specimen), scale type, and method type, was established for 29,340 test observations. Short names and mapping tables for EDI codes and UMLS were added. Synonym tables were prepared to help match concept names to common terms used in the fields. We herein described the Korean standard laboratory terminology database for test names, result description terms, and result units encompassing most of the laboratory tests in Korea. PMID:18756062
REDEX - The ranging equipment diagnostic expert system
NASA Technical Reports Server (NTRS)
Luczak, Edward C.; Gopalakrishnan, K.; Zillig, David J.
1989-01-01
REDEX, an advanced prototype expert system that diagnoses hardware failures in the Ranging Equipment (RE) at NASA's Ground Network tracking stations is described. REDEX will help the RE technician identify faulty circuit cards or modules that must be replaced, and thereby reduce troubleshooting time. It features a highly graphical user interface that uses color block diagrams and layout diagrams to illustrate the location of a fault. A semantic network knowledge representation technique was used to model the design structure of the RE. A catalog of generic troubleshooting rules was compiled to represent heuristics that are applied in diagnosing electronic equipment. Specific troubleshooting rules were identified to represent additional diagnostic knowledge that is unique to the RE. Over 50 generic and 250 specific troubleshooting rules have been derived. REDEX is implemented in Prolog on an IBM PC AT-compatible workstation. Block diagram graphics displays are color-coded to identify signals that have been monitored or inferred to have nominal values, signals that are out of tolerance, and circuit cards and functions that are diagnosed as faulty. A hypertext-like scheme is used to allow the user to easily navigate through the space of diagrams and tables. Over 50 graphic and tabular displays have been implemented. REDEX is currently being evaluated in a stand-alone mode using simulated RE fault scenarios. It will soon be interfaced to the RE and tested in an online environment. When completed and fielded, REDEX will be a concrete example of the application of expert systems technology to the problem of improving performance and reducing the lifecycle costs of operating NASA's communications networks in the 1990s.
REDEX: The ranging equipment diagnostic expert system
NASA Technical Reports Server (NTRS)
Luczak, Edward C.; Gopalakrishnan, K.; Zillig, David J.
1989-01-01
REDEX, an advanced prototype expert system that diagnoses hardware failures in the Ranging Equipment (RE) at NASA's Ground Network tracking stations is described. REDEX will help the RE technician identify faulty circuit cards or modules that must be replaced, and thereby reduce troubleshooting time. It features a highly graphical user interface that uses color block diagrams and layout diagrams to illustrate the location of a fault. A semantic network knowledge representation technique was used to model the design structure of the RE. A catalog of generic troubleshooting rules was compiled to represent heuristics that are applied in diagnosing electronic equipment. Specific troubleshooting rules were identified to represent additional diagnostic knowledge that is unique to the RE. Over 50 generic and 250 specific troubleshooting rules have been derived. REDEX is implemented in Prolog on an IBM PC AT-compatible workstation. Block diagram graphics displays are color-coded to identify signals that have been monitored or inferred to have nominal values, signals that are out of tolerance, and circuit cards and functions that are diagnosed as faulty. A hypertext-like scheme is used to allow the user to easily navigate through the space of diagrams and tables. Over 50 graphic and tabular displays have been implemented. REDEX is currently being evaluated in a stand-alone mode using simulated RE fault scenarios. It will soon be interfaced to the RE and tested in an online environment. When completed and fielded, REDEX will be a concrete example of the application of expert systems technology to the problem of improving performance and reducing the lifecycle costs of operating NASA's communications networks in the 1990's.
REDEX - The ranging equipment diagnostic expert system
NASA Astrophysics Data System (ADS)
Luczak, Edward C.; Gopalakrishnan, K.; Zillig, David J.
REDEX, an advanced prototype expert system that diagnoses hardware failures in the Ranging Equipment (RE) at NASA's Ground Network tracking stations is described. REDEX will help the RE technician identify faulty circuit cards or modules that must be replaced, and thereby reduce troubleshooting time. It features a highly graphical user interface that uses color block diagrams and layout diagrams to illustrate the location of a fault. A semantic network knowledge representation technique was used to model the design structure of the RE. A catalog of generic troubleshooting rules was compiled to represent heuristics that are applied in diagnosing electronic equipment. Specific troubleshooting rules were identified to represent additional diagnostic knowledge that is unique to the RE. Over 50 generic and 250 specific troubleshooting rules have been derived. REDEX is implemented in Prolog on an IBM PC AT-compatible workstation. Block diagram graphics displays are color-coded to identify signals that have been monitored or inferred to have nominal values, signals that are out of tolerance, and circuit cards and functions that are diagnosed as faulty. A hypertext-like scheme is used to allow the user to easily navigate through the space of diagrams and tables. Over 50 graphic and tabular displays have been implemented. REDEX is currently being evaluated in a stand-alone mode using simulated RE fault scenarios. It will soon be interfaced to the RE and tested in an online environment. When completed and fielded, REDEX will be a concrete example of the application of expert systems technology to the problem of improving performance and reducing the lifecycle costs of operating NASA's communications networks in the 1990s.
REDEX: The ranging equipment diagnostic expert system
NASA Astrophysics Data System (ADS)
Luczak, Edward C.; Gopalakrishnan, K.; Zillig, David J.
1989-04-01
REDEX, an advanced prototype expert system that diagnoses hardware failures in the Ranging Equipment (RE) at NASA's Ground Network tracking stations is described. REDEX will help the RE technician identify faulty circuit cards or modules that must be replaced, and thereby reduce troubleshooting time. It features a highly graphical user interface that uses color block diagrams and layout diagrams to illustrate the location of a fault. A semantic network knowledge representation technique was used to model the design structure of the RE. A catalog of generic troubleshooting rules was compiled to represent heuristics that are applied in diagnosing electronic equipment. Specific troubleshooting rules were identified to represent additional diagnostic knowledge that is unique to the RE. Over 50 generic and 250 specific troubleshooting rules have been derived. REDEX is implemented in Prolog on an IBM PC AT-compatible workstation. Block diagram graphics displays are color-coded to identify signals that have been monitored or inferred to have nominal values, signals that are out of tolerance, and circuit cards and functions that are diagnosed as faulty. A hypertext-like scheme is used to allow the user to easily navigate through the space of diagrams and tables. Over 50 graphic and tabular displays have been implemented. REDEX is currently being evaluated in a stand-alone mode using simulated RE fault scenarios. It will soon be interfaced to the RE and tested in an online environment. When completed and fielded, REDEX will be a concrete example of the application of expert systems technology to the problem of improving performance and reducing the lifecycle costs of operating NASA's communications networks in the 1990's.
Limited capacity in US pediatric drug trials: qualitative analysis of expert interviews.
Wasserman, Richard; Bocian, Alison; Harris, Donna; Slora, Eric
2011-04-01
The recently renewed Best Pharmaceuticals for Children and Pediatric Research Equity Acts (BPCA/PREA) have continued industry incentives and opportunities for pediatric drug trials (PDTs). However, there is no current assessment of the capacity to perform PDTs. The aim of this study was to deepen understanding of the capacity for US PDTs by assessing PDT infrastructure, present barriers to PDTs, and potential approaches and solutions to identified issues. Pediatric clinical research experts participated in semi-structured interviews on current US pediatric research capacity (February-July 2007). An initial informant list was developed using purposive sampling, and supplemented and refined to generate a group of respondents to explore emerging themes. Each phone interview included a physician researcher and two health researchers who took notes and recorded the calls. Health researchers produced detailed summaries, which were verified by the physician researcher and informants. We then undertook qualitative analysis of the summaries, employing multiple coding, with the two health researchers and the physician researcher independently coding each summary for themes and subthemes. Coding variations were resolved by physician researcher/health researcher discussion and consensus achieved on themes and subthemes. The 33 informants' primary or secondary roles included academia (n = 21), federal official (5), industry medical officer (8), pediatric research network leader (10), pediatric specialist leader (8), pediatric clinical pharmacologist (5), and practitioner/research site director (9). While most experts noted an increase in PDTs since the initial passage of BPCA/PREA, a dominant theme of insufficient US PDT capacity emerged. Subthemes included (i) lack of systems for finding, incentivizing, and/or maintaining trial sites; (ii) complexity/demands of conducting PDTs in clinical settings; (iii) inadequate numbers of qualified pediatric pharmacologists and clinician investigators trained in FDA Good Clinical Practice; and (iv) poor PDT protocol design resulting in operational and enrollment difficulties in the pediatric population. Suggested potential solutions for insufficient PDT capacity included (i) consensus-building among stakeholders to create PDT systems; (ii) initiatives to train more pediatric pharmacologists and educate clinicians in Good Clinical Practice; (iii) advocacy for PDT protocols designed by individuals sensitive to pediatric issues; and (iv) physician and public education on the importance of PDTs. Insufficient US PDT capacity may hinder the development of new drugs for children and limit studies on the safety and efficacy of drugs presently used to treat pediatric conditions. Further public policy initiatives may be needed to achieve the full promise of BPCA/PREA.
Catts, Stanley V; Frost, Aaron D J; O'Toole, Brian I; Carr, Vaughan J; Lewin, Terry; Neil, Amanda L; Harris, Meredith G; Evans, Russell W; Crissman, Belinda R; Eadie, Kathy
2011-01-01
Clinical practice improvement carried out in a quality assurance framework relies on routinely collected data using clinical indicators. Herein we describe the development, minimum training requirements, and inter-rater agreement of indicators that were used in an Australian multi-site evaluation of the effectiveness of early psychosis (EP) teams. Surveys of clinician opinion and face-to-face consensus-building meetings were used to select and conceptually define indicators. Operationalization of definitions was achieved by iterative refinement until clinicians could be quickly trained to code indicators reliably. Calculation of percentage agreement with expert consensus coding was based on ratings of paper-based clinical vignettes embedded in a 2-h clinician training package. Consensually agreed upon conceptual definitions for seven clinical indicators judged most relevant to evaluating EP teams were operationalized for ease-of-training. Brief training enabled typical clinicians to code indicators with acceptable percentage agreement (60% to 86%). For indicators of suicide risk, psychosocial function, and family functioning this level of agreement was only possible with less precise 'broad range' expert consensus scores. Estimated kappa values indicated fair to good inter-rater reliability (kappa > 0.65). Inspection of contingency tables (coding category by health service) and modal scores across services suggested consistent, unbiased coding across services. Clinicians are able to agree upon what information is essential to routinely evaluate clinical practice. Simple indicators of this information can be designed and coding rules can be reliably applied to written vignettes after brief training. The real world feasibility of the indicators remains to be tested in field trials.
DOT National Transportation Integrated Search
1987-01-01
Expert systems, a branch of artificial-intelligence studies, is introduced with a view to its relevance in transportation engineering. Knowledge engineering, the process of building expert systems or transferring knowledge from human experts to compu...
NASA Astrophysics Data System (ADS)
Rates, Christopher A.; Mulvey, Bridget K.; Feldon, David F.
2016-08-01
Components of complex systems apply across multiple subject areas, and teaching these components may help students build unifying conceptual links. Students, however, often have difficulty learning these components, and limited research exists to understand what types of interventions may best help improve understanding. We investigated 32 high school students' understandings of complex systems components and whether an agent-based simulation could improve their understandings. Pretest and posttest essays were coded for changes in six components to determine whether students showed more expert thinking about the complex system of the Chesapeake Bay watershed. Results showed significant improvement for the components Emergence ( r = .26, p = .03), Order ( r = .37, p = .002), and Tradeoffs ( r = .44, p = .001). Implications include that the experiential nature of the simulation has the potential to support conceptual change for some complex systems components, presenting a promising option for complex systems instruction.
NASA Technical Reports Server (NTRS)
Gryphon, Coranth D.; Miller, Mark D.
1991-01-01
PCLIPS (Parallel CLIPS) is a set of extensions to the C Language Integrated Production System (CLIPS) expert system language. PCLIPS is intended to provide an environment for the development of more complex, extensive expert systems. Multiple CLIPS expert systems are now capable of running simultaneously on separate processors, or separate machines, thus dramatically increasing the scope of solvable tasks within the expert systems. As a tool for parallel processing, PCLIPS allows for an expert system to add to its fact-base information generated by other expert systems, thus allowing systems to assist each other in solving a complex problem. This allows individual expert systems to be more compact and efficient, and thus run faster or on smaller machines.
A Semantic Analysis Method for Scientific and Engineering Code
NASA Technical Reports Server (NTRS)
Stewart, Mark E. M.
1998-01-01
This paper develops a procedure to statically analyze aspects of the meaning or semantics of scientific and engineering code. The analysis involves adding semantic declarations to a user's code and parsing this semantic knowledge with the original code using multiple expert parsers. These semantic parsers are designed to recognize formulae in different disciplines including physical and mathematical formulae and geometrical position in a numerical scheme. In practice, a user would submit code with semantic declarations of primitive variables to the analysis procedure, and its semantic parsers would automatically recognize and document some static, semantic concepts and locate some program semantic errors. A prototype implementation of this analysis procedure is demonstrated. Further, the relationship between the fundamental algebraic manipulations of equations and the parsing of expressions is explained. This ability to locate some semantic errors and document semantic concepts in scientific and engineering code should reduce the time, risk, and effort of developing and using these codes.
ART-Ada: An Ada-based expert system tool
NASA Technical Reports Server (NTRS)
Lee, S. Daniel; Allen, Bradley P.
1990-01-01
The Department of Defense mandate to standardize on Ada as the language for software systems development has resulted in an increased interest in making expert systems technology readily available in Ada environments. NASA's Space Station Freedom is an example of the large Ada software development projects that will require expert systems in the 1990's. Another large scale application that can benefit from Ada based expert system tool technology is the Pilot's Associate (PA) expert system project for military combat aircraft. The Automated Reasoning Tool-Ada (ART-Ada), an Ada expert system tool, is explained. ART-Ada allows applications of a C-based expert system tool called ART-IM to be deployed in various Ada environments. ART-Ada is being used to implement several prototype expert systems for NASA's Space Station Freedom program and the U.S. Air Force.
ART-Ada: An Ada-based expert system tool
NASA Technical Reports Server (NTRS)
Lee, S. Daniel; Allen, Bradley P.
1991-01-01
The Department of Defense mandate to standardize on Ada as the language for software systems development has resulted in increased interest in making expert systems technology readily available in Ada environments. NASA's Space Station Freedom is an example of the large Ada software development projects that will require expert systems in the 1990's. Another large scale application that can benefit from Ada based expert system tool technology is the Pilot's Associate (PA) expert system project for military combat aircraft. Automated Reasoning Tool (ART) Ada, an Ada Expert system tool is described. ART-Ada allow applications of a C-based expert system tool called ART-IM to be deployed in various Ada environments. ART-Ada is being used to implement several prototype expert systems for NASA's Space Station Freedom Program and the U.S. Air Force.
Development and validation of a notational system to study the offensive process in football.
Sarmento, Hugo; Anguera, Teresa; Campaniço, Jorge; Leitão, José
2010-01-01
The most striking change within football development is the application of science to its problems and in particular the use of increasingly sophisticated technology that, supported by scientific data, allows us to establish a "code of reading" the reality of the game. Therefore, this study describes the process of the development and validation of an ad hoc system of categorization, which allows the different methods of offensive game in football and the interaction to be analyzed. Therefore, through an exploratory phase of the study, we identified 10 vertebrate criteria and the respective behaviors observed for each of these criteria. We heard a panel of five experts with the purpose of a content validation. The resulting instrument is characterized by a combination of field formats and systems of categories. The reliability of the instrument was calculated by the intraobserver agreement, and values above 0.95 for all criteria were achieved. Two FC Barcelona games were coded and analyzed, which allowed the detection of various T-patterns. The results show that the instrument serves the purpose for which it was developed and can provide important information for the understanding of game interaction in football.
Integral nuclear data validation using experimental spent nuclear fuel compositions
Gauld, Ian C.; Williams, Mark L.; Michel-Sendis, Franco; ...
2017-07-19
Measurements of the isotopic contents of spent nuclear fuel provide experimental data that are a prerequisite for validating computer codes and nuclear data for many spent fuel applications. Under the auspices of the Organisation for Economic Co-operation and Development (OECD) Nuclear Energy Agency (NEA) and guidance of the Expert Group on Assay Data of Spent Nuclear Fuel of the NEA Working Party on Nuclear Criticality Safety, a new database of expanded spent fuel isotopic compositions has been compiled. The database, Spent Fuel Compositions (SFCOMPO) 2.0, includes measured data for more than 750 fuel samples acquired from 44 different reactors andmore » representing eight different reactor technologies. Measurements for more than 90 isotopes are included. This new database provides data essential for establishing the reliability of code systems for inventory predictions, but it also has broader potential application to nuclear data evaluation. Furthermore, the database, together with adjoint based sensitivity and uncertainty tools for transmutation systems developed to quantify the importance of nuclear data on nuclide concentrations, are described.« less
Non-native (exotic) snake envenomations in the U.S., 2005-2011.
Warrick, Brandon J; Boyer, Leslie V; Seifert, Steven A
2014-09-29
Non-native (exotic) snakes are a problematic source of envenomation worldwide. This manuscript describes the current demographics, outcomes and challenges of non-native snakebites in the United States (U.S.). We performed a retrospective case series of the National Poison Data System (NPDS) database between 2005 and 2011. There were 258 human exposures involving at least 61 unique exotic venomous species (average = 37 per year; range = 33-40). Males comprised 79% and females 21%. The average age was 33 years with 16% less than 20 years old. 70% of bites occurred in a private residence and 86% were treated at a healthcare facility. 35% of cases received antivenom and 10% were given antibiotics. This study is compared to our previous study (1994-2004) in which there was a substantial coding error rate. Software modifications significantly reduced coding errors. Identification and acquisition of appropriate antivenoms pose a number of logistical difficulties in the management of these envenomations. In the U.S., poison centers have valuable systems and clinical roles in the provision of expert consultation and in the management of these cases.
Non-Native (Exotic) Snake Envenomations in the U.S., 2005–2011
Warrick, Brandon J.; Boyer, Leslie V.; Seifert, Steven A.
2014-01-01
Non-native (exotic) snakes are a problematic source of envenomation worldwide. This manuscript describes the current demographics, outcomes and challenges of non-native snakebites in the United States (U.S.). We performed a retrospective case series of the National Poison Data System (NPDS) database between 2005 and 2011. There were 258 human exposures involving at least 61 unique exotic venomous species (average = 37 per year; range = 33–40). Males comprised 79% and females 21%. The average age was 33 years with 16% less than 20 years old. 70% of bites occurred in a private residence and 86% were treated at a healthcare facility. 35% of cases received antivenom and 10% were given antibiotics. This study is compared to our previous study (1994–2004) in which there was a substantial coding error rate. Software modifications significantly reduced coding errors. Identification and acquisition of appropriate antivenoms pose a number of logistical difficulties in the management of these envenomations. In the U.S., poison centers have valuable systems and clinical roles in the provision of expert consultation and in the management of these cases. PMID:25268980
Integral nuclear data validation using experimental spent nuclear fuel compositions
DOE Office of Scientific and Technical Information (OSTI.GOV)
Gauld, Ian C.; Williams, Mark L.; Michel-Sendis, Franco
Measurements of the isotopic contents of spent nuclear fuel provide experimental data that are a prerequisite for validating computer codes and nuclear data for many spent fuel applications. Under the auspices of the Organisation for Economic Co-operation and Development (OECD) Nuclear Energy Agency (NEA) and guidance of the Expert Group on Assay Data of Spent Nuclear Fuel of the NEA Working Party on Nuclear Criticality Safety, a new database of expanded spent fuel isotopic compositions has been compiled. The database, Spent Fuel Compositions (SFCOMPO) 2.0, includes measured data for more than 750 fuel samples acquired from 44 different reactors andmore » representing eight different reactor technologies. Measurements for more than 90 isotopes are included. This new database provides data essential for establishing the reliability of code systems for inventory predictions, but it also has broader potential application to nuclear data evaluation. Furthermore, the database, together with adjoint based sensitivity and uncertainty tools for transmutation systems developed to quantify the importance of nuclear data on nuclide concentrations, are described.« less
Porting a Mental Expert System to a Mainstream Programming Environment
Jao, Chiang S.; Hier, Daniel B.; Dollear, Winifred; Fu, Wenying
2001-01-01
Expert systems are increasingly being applied to problems in medical diagnosis and treatment. Initial medical expert systems were programmed in specialized “expert system” shell programming environments. As the power of mainstream programming languages has increased, it has become possible to implement medical expert systems within these mainstream languages. We originally implemented an expert system to record and score the mental status examination utilizing a specialized expert system programming environment. We have now ported that application to a mainstream programming environment without losing any functionality of an accurate and comprehensive diagnostic tool. New system supplements the need of normative consultation report and offline reference library to the traditional patient care system.
A system for environmental model coupling and code reuse: The Great Rivers Project
NASA Astrophysics Data System (ADS)
Eckman, B.; Rice, J.; Treinish, L.; Barford, C.
2008-12-01
As part of the Great Rivers Project, IBM is collaborating with The Nature Conservancy and the Center for Sustainability and the Global Environment (SAGE) at the University of Wisconsin, Madison to build a Modeling Framework and Decision Support System (DSS) designed to help policy makers and a variety of stakeholders (farmers, fish & wildlife managers, hydropower operators, et al.) to assess, come to consensus, and act on land use decisions representing effective compromises between human use and ecosystem preservation/restoration. Initially focused on Brazil's Paraguay-Parana, China's Yangtze, and the Mississippi Basin in the US, the DSS integrates data and models from a wide variety of environmental sectors, including water balance, water quality, carbon balance, crop production, hydropower, and biodiversity. In this presentation we focus on the modeling framework aspect of this project. In our approach to these and other environmental modeling projects, we see a flexible, extensible modeling framework infrastructure for defining and running multi-step analytic simulations as critical. In this framework, we divide monolithic models into atomic components with clearly defined semantics encoded via rich metadata representation. Once models and their semantics and composition rules have been registered with the system by their authors or other experts, non-expert users may construct simulations as workflows of these atomic model components. A model composition engine enforces rules/constraints for composing model components into simulations, to avoid the creation of Frankenmodels, models that execute but produce scientifically invalid results. A common software environment and common representations of data and models are required, as well as an adapter strategy for code written in e.g., Fortran or python, that still enables efficient simulation runs, including parallelization. Since each new simulation, as a new composition of model components, requires calibration of parameters (fudge factors) to produce scientifically valid results, we are also developing an autocalibration engine. Finally, visualization is a key element of this modeling framework strategy, both to convey complex scientific data effectively, and also to enable non-expert users to make full use of the relevant features of the framework. We are developing a visualization environment with a strong data model, to enable visualizations, model results, and data all to be handled similarly.
An expert system for the design of heating, ventilating, and air-conditioning systems
NASA Astrophysics Data System (ADS)
Camejo, Pedro Jose
1989-12-01
Expert systems are computer programs that seek to mimic human reason. An expert system shelf, a software program commonly used for developing expert systems in a relatively short time, was used to develop a prototypical expert system for the design of heating, ventilating, and air-conditioning (HVAC) systems in buildings. Because HVAC design involves several related knowledge domains, developing an expert system for HVAC design requires the integration of several smaller expert systems known as knowledge bases. A menu program and several auxiliary programs for gathering data, completing calculations, printing project reports, and passing data between the knowledge bases are needed and have been developed to join the separate knowledge bases into one simple-to-use program unit.
Application of Boiler Op for combustion optimization at PEPCO
DOE Office of Scientific and Technical Information (OSTI.GOV)
Maines, P.; Williams, S.; Levy, E.
1997-09-01
Title IV requires the reduction of NOx at all stations within the PEPCO system. To assist PEPCO plant personnel in achieving low heat rates while meeting NOx targets, Lehigh University`s Energy Research Center and PEPCO developed a new combustion optimization software package called Boiler Op. The Boiler Op code contains an expert system, neural networks and an optimization algorithm. The expert system guides the plant engineer through a series of parametric boiler tests, required for the development of a comprehensive boiler database. The data are then analyzed by the neural networks and optimization algorithm to provide results on the boilermore » control settings which result in the best possible heat rate at a target NOx level or produce minimum NOx. Boiler Op has been used at both Potomac River and Morgantown Stations to help PEPCO engineers optimize combustion. With the use of Boiler Op, Morgantown Station operates under low NOx restrictions and continues to achieve record heat rate values, similar to pre-retrofit conditions. Potomac River Station achieves the regulatory NOx limit through the use of Boiler Op recommended control settings and without NOx burners. Importantly, any software like Boiler Op cannot be used alone. Its application must be in concert with human intelligence to ensure unit safety, reliability and accurate data collection.« less
MARBLE: A system for executing expert systems in parallel
NASA Technical Reports Server (NTRS)
Myers, Leonard; Johnson, Coe; Johnson, Dean
1990-01-01
This paper details the MARBLE 2.0 system which provides a parallel environment for cooperating expert systems. The work has been done in conjunction with the development of an intelligent computer-aided design system, ICADS, by the CAD Research Unit of the Design Institute at California Polytechnic State University. MARBLE (Multiple Accessed Rete Blackboard Linked Experts) is a system of C Language Production Systems (CLIPS) expert system tool. A copied blackboard is used for communication between the shells to establish an architecture which supports cooperating expert systems that execute in parallel. The design of MARBLE is simple, but it provides support for a rich variety of configurations, while making it relatively easy to demonstrate the correctness of its parallel execution features. In its most elementary configuration, individual CLIPS expert systems execute on their own processors and communicate with each other through a modified blackboard. Control of the system as a whole, and specifically of writing to the blackboard is provided by one of the CLIPS expert systems, an expert control system.
Psychology of developing and designing expert systems
DOE Office of Scientific and Technical Information (OSTI.GOV)
Tonn, B.; MacGregor, D.
This paper discusses psychological problems relevant to developing and designing expert systems. With respect to the former, the psychological literature suggests that several cognitive biases may affect the elicitation of a valid knowledge base from the expert. The literature also suggests that common expert system inference engines may be quite inconsistent with reasoning heuristics employed by experts. With respect to expert system user interfaces, care should be taken when eliciting uncertainty estimates from users, presenting system conclusions, and ordering questions.
Eyles, John; Heddle, Nancy; Webert, Kathryn; Arnold, Emmy; McCurdy, Bronwen
2011-08-24
Examining professional assessments of a blood product recall/withdrawal and its implications for risk and public health, the paper introduces ideas about perceptions of minimal risk and its management. It also describes the context of publicly funded blood transfusion in Canada and the withdrawal event that is the basis of this study. Interviews with 45 experts from administration, medicine, blood supply, laboratory services and risk assessment took place using a multi-level sampling framework in the aftermath of the recall. These experts either directly dealt with the withdrawal or were involved in the management of the blood supply at the national level. Data from these interviews were coded in NVivo for analysis and interpretation. Analytically, data were interpreted to derive typifications to relate interview responses to risk management heuristics. While all those interviewed agreed on the importance of patient safety, differences in the ways in which the risk was contextualized and explicated were discerned. Risk was seen in terms of patient safety, liability or precaution. These different risk logics are illustrated by selected quotations. Expert assessments did not fully converge and it is possible that these different risk logics and discourses may affect the risk management process more generally, although not necessarily in a negative way. Patient safety is not to be compromised but management of blood risk in publicly funded systems may vary. We suggest ways of managing blood risk using formal and safety case approaches.
2011-01-01
Background Examining professional assessments of a blood product recall/withdrawal and its implications for risk and public health, the paper introduces ideas about perceptions of minimal risk and its management. It also describes the context of publicly funded blood transfusion in Canada and the withdrawal event that is the basis of this study. Methods Interviews with 45 experts from administration, medicine, blood supply, laboratory services and risk assessment took place using a multi-level sampling framework in the aftermath of the recall. These experts either directly dealt with the withdrawal or were involved in the management of the blood supply at the national level. Data from these interviews were coded in NVivo for analysis and interpretation. Analytically, data were interpreted to derive typifications to relate interview responses to risk management heuristics. Results While all those interviewed agreed on the importance of patient safety, differences in the ways in which the risk was contextualized and explicated were discerned. Risk was seen in terms of patient safety, liability or precaution. These different risk logics are illustrated by selected quotations. Conclusions Expert assessments did not fully converge and it is possible that these different risk logics and discourses may affect the risk management process more generally, although not necessarily in a negative way. Patient safety is not to be compromised but management of blood risk in publicly funded systems may vary. We suggest ways of managing blood risk using formal and safety case approaches. PMID:21864330
NASA Astrophysics Data System (ADS)
Giovanna, Vessia; Luca, Pisano; Carmela, Vennari; Mauro, Rossi; Mario, Parise
2016-01-01
This paper proposes an automated method for the selection of rainfall data (duration, D, and cumulated, E), responsible for shallow landslide initiation. The method mimics an expert person identifying D and E from rainfall records through a manual procedure whose rules are applied according to her/his judgement. The comparison between the two methods is based on 300 D-E pairs drawn from temporal rainfall data series recorded in a 30 days time-lag before the landslide occurrence. Statistical tests, employed on D and E samples considered both paired and independent values to verify whether they belong to the same population, show that the automated procedure is able to replicate the expert pairs drawn by the expert judgment. Furthermore, a criterion based on cumulated distribution functions (CDFs) is proposed to select the most related D-E pairs to the expert one among the 6 drawn from the coded procedure for tracing the empirical rainfall threshold line.
Processes in construction of failure management expert systems from device design information
NASA Technical Reports Server (NTRS)
Malin, Jane T.; Lance, Nick
1987-01-01
This paper analyzes the tasks and problem solving methods used by an engineer in constructing a failure management expert system from design information about the device to te diagnosed. An expert test engineer developed a trouble-shooting expert system based on device design information and experience with similar devices, rather than on specific expert knowledge gained from operating the device or troubleshooting its failures. The construction of the expert system was intensively observed and analyzed. This paper characterizes the knowledge, tasks, methods, and design decisions involved in constructing this type of expert system, and makes recommendations concerning tools for aiding and automating construction of such systems.
Rushton, A; White, L; Heap, A; Heneghan, N; Goodwin, P
2016-01-01
Objectives To develop an optimised 1:1 physiotherapy intervention that reflects best practice, with flexibility to tailor management to individual patients, thereby ensuring patient-centred practice. Design Mixed-methods combining evidence synthesis, expert review and focus groups. Setting Secondary care involving 5 UK specialist spinal centres. Participants A purposive panel of clinical experts from the 5 spinal centres, comprising spinal surgeons, inpatient and outpatient physiotherapists, provided expert review of the draft intervention. Purposive samples of patients (n=10) and physiotherapists (n=10) (inpatient/outpatient physiotherapists managing patients with lumbar discectomy) were invited to participate in the focus groups at 1 spinal centre. Methods A draft intervention developed from 2 systematic reviews; a survey of current practice and research related to stratified care was circulated to the panel of clinical experts. Lead physiotherapists collaborated with physiotherapy and surgeon colleagues to provide feedback that informed the intervention presented at 2 focus groups investigating acceptability to patients and physiotherapists. The focus groups were facilitated by an experienced facilitator, recorded in written and tape-recorded forms by an observer. Tape recordings were transcribed verbatim. Data analysis, conducted by 2 independent researchers, employed an iterative and constant comparative process of (1) initial descriptive coding to identify categories and subsequent themes, and (2) deeper, interpretive coding and thematic analysis enabling concepts to emerge and overarching pattern codes to be identified. Results The intervention reflected best available evidence and provided flexibility to ensure patient-centred care. The intervention comprised up to 8 sessions of 1:1 physiotherapy over 8 weeks, starting 4 weeks postsurgery. The intervention was acceptable to patients and physiotherapists. Conclusions A rigorous process informed an optimised 1:1 physiotherapy intervention post-lumbar discectomy that reflects best practice. The developed intervention was agreed on by the 5 spinal centres for implementation in a randomised controlled trial to evaluate its effectiveness. PMID:26916690
Bahl, Rachna; Murphy, Deirdre J; Strachan, Bryony
2010-06-01
Non-technical skills are cognitive and social skills required in an operational task. These skills have been identified and taught in the surgical domain but are of particular relevance to obstetrics where the patient is awake, the partner is present and the clinical circumstances are acute and often stressful. The aim of this study was to define the non-technical skills of an operative vaginal delivery (forceps or vacuum) to facilitate transfer of skills from expert obstetricians to trainee obstetricians. Qualitative study using interviews and video recordings. The study was conducted at two university teaching hospitals (St. Michael's Hospital, Bristol and Ninewells Hospital, Dundee). Participants included 10 obstetricians and eight midwives identified as experts in conducting or supporting operative vaginal deliveries. Semi-structured interviews were carried out using routine clinical scenarios. The experts were also video recorded conducting forceps and vacuum deliveries in a simulation setting. The interviews and video recordings were transcribed verbatim and analysed using thematic coding. The anonymised data were independently coded by the three researchers and then compared for consistency of interpretation. The experts reviewed the coded data for respondent validation and clarification. The themes that emerged were used to identify the non-technical skills required for conducting an operative vaginal delivery. The final skills list was classified into seven main categories. Four categories (situational awareness, decision making, task management, and team work and communication) were similar to the categories identified in surgery. Three further categories unique to obstetrics were also identified (professional relationship with the woman, maintaining professional behaviour and cross-monitoring of performance). This explicitly defined skills taxonomy could aid trainees' understanding of the non-technical skills to be considered when conducting an operative vaginal delivery and potentially reduce morbidity and improve the experience of delivery for the mother. Copyright 2010 Elsevier Ireland Ltd. All rights reserved.
Computer Aided Dosimetry and Verification of Exposure to Radiation
NASA Astrophysics Data System (ADS)
Waller, Edward; Stodilka, Robert Z.; Leach, Karen E.; Lalonde, Louise
2002-06-01
In the timeframe following the September 11th attacks on the United States, increased emphasis has been placed on Chemical, Biological, Radiological and Nuclear (CBRN) preparedness. Of prime importance is rapid field assessment of potential radiation exposure to Canadian Forces field personnel. This work set up a framework for generating an 'expert' computer system for aiding and assisting field personnel in determining the extent of radiation insult to military personnel. Data was gathered by review of the available literature, discussions with medical and health physics personnel having hands-on experience dealing with radiation accident victims, and from experience of the principal investigator. Flow charts and generic data fusion algorithms were developed. Relationships between known exposure parameters, patient interview and history, clinical symptoms, clinical work-ups, physical dosimetry, biological dosimetry, and dose reconstruction as critical data indicators were investigated. The data obtained was examined in terms of information theory. A main goal was to determine how best to generate an adaptive model (i.e. when more data becomes available, how is the prediction improved). Consideration was given to determination of predictive algorithms for health outcome. In addition. the concept of coding an expert medical treatment advisor system was developed (U)
Developing Information Power Grid Based Algorithms and Software
NASA Technical Reports Server (NTRS)
Dongarra, Jack
1998-01-01
This was an exploratory study to enhance our understanding of problems involved in developing large scale applications in a heterogeneous distributed environment. It is likely that the large scale applications of the future will be built by coupling specialized computational modules together. For example, efforts now exist to couple ocean and atmospheric prediction codes to simulate a more complete climate system. These two applications differ in many respects. They have different grids, the data is in different unit systems and the algorithms for inte,-rating in time are different. In addition the code for each application is likely to have been developed on different architectures and tend to have poor performance when run on an architecture for which the code was not designed, if it runs at all. Architectural differences may also induce differences in data representation which effect precision and convergence criteria as well as data transfer issues. In order to couple such dissimilar codes some form of translation must be present. This translation should be able to handle interpolation from one grid to another as well as construction of the correct data field in the correct units from available data. Even if a code is to be developed from scratch, a modular approach will likely be followed in that standard scientific packages will be used to do the more mundane tasks such as linear algebra or Fourier transform operations. This approach allows the developers to concentrate on their science rather than becoming experts in linear algebra or signal processing. Problems associated with this development approach include difficulties associated with data extraction and translation from one module to another, module performance on different nodal architectures, and others. In addition to these data and software issues there exists operational issues such as platform stability and resource management.
NASA Technical Reports Server (NTRS)
Chang, C. L.; Stachowitz, R. A.
1988-01-01
Software quality is of primary concern in all large-scale expert system development efforts. Building appropriate validation and test tools for ensuring software reliability of expert systems is therefore required. The Expert Systems Validation Associate (EVA) is a validation system under development at the Lockheed Artificial Intelligence Center. EVA provides a wide range of validation and test tools to check correctness, consistency, and completeness of an expert system. Testing a major function of EVA. It means executing an expert system with test cases with the intent of finding errors. In this paper, we describe many different types of testing such as function-based testing, structure-based testing, and data-based testing. We describe how appropriate test cases may be selected in order to perform good and thorough testing of an expert system.
Analyzing the Language of Therapist Empathy in Motivational Interview based Psychotherapy
Xiao, Bo; Can, Dogan; Georgiou, Panayiotis G.; Atkins, David; Narayanan, Shrikanth S.
2016-01-01
Empathy is an important aspect of social communication, especially in medical and psychotherapy applications. Measures of empathy can offer insights into the quality of therapy. We use an N-gram language model based maximum likelihood strategy to classify empathic versus non-empathic utterances and report the precision and recall of classification for various parameters. High recall is obtained with unigram while bigram features achieved the highest F1-score. Based on the utterance level models, a group of lexical features are extracted at the therapy session level. The effectiveness of these features in modeling session level annotator perceptions of empathy is evaluated through correlation with expert-coded session level empathy scores. Our combined feature set achieved a correlation of 0.558 between predicted and expert-coded empathy scores. Results also suggest that the longer term empathy perception process may be more related to isolated empathic salient events. PMID:27602411
Expert systems applications for space shuttle payload integration automation
NASA Technical Reports Server (NTRS)
Morris, Keith
1988-01-01
Expert systems technologies have been and are continuing to be applied to NASA's Space Shuttle orbiter payload integration problems to provide a level of automation previously unrealizable. NASA's Space Shuttle orbiter was designed to be extremely flexible in its ability to accommodate many different types and combinations of satellites and experiments (payloads) within its payload bay. This flexibility results in differnet and unique engineering resource requirements for each of its payloads, creating recurring payload and cargo integration problems. Expert systems provide a successful solution for these recurring problems. The Orbiter Payload Bay Cabling Expert (EXCABL) was the first expert system, developed to solve the electrical services provisioning problem. A second expert system, EXMATCH, was developed to generate a list of the reusable installation drawings available for each EXCABL solution. These successes have proved the applicability of expert systems technologies to payload integration problems and consequently a third expert system is currently in work. These three expert systems, the manner in which they resolve payload problems and how they will be integrated are described.
NASA Technical Reports Server (NTRS)
Prince, Mary Ellen
1987-01-01
The expert system is a computer program which attempts to reproduce the problem-solving behavior of an expert, who is able to view problems from a broad perspective and arrive at conclusions rapidly, using intuition, shortcuts, and analogies to previous situations. Expert systems are a departure from the usual artificial intelligence approach to problem solving. Researchers have traditionally tried to develop general modes of human intelligence that could be applied to many different situations. Expert systems, on the other hand, tend to rely on large quantities of domain specific knowledge, much of it heuristic. The reasoning component of the system is relatively simple and straightforward. For this reason, expert systems are often called knowledge based systems. The report expands on the foregoing. Section 1 discusses the architecture of a typical expert system. Section 2 deals with the characteristics that make a problem a suitable candidate for expert system solution. Section 3 surveys current technology, describing some of the software aids available for expert system development. Section 4 discusses the limitations of the latter. The concluding section makes predictions of future trends.
Expert system for the design of heating, ventilating, and air-conditioning systems. Master's thesis
DOE Office of Scientific and Technical Information (OSTI.GOV)
Camejo, P.J.
1989-12-01
Expert systems are computer programs that seek to mimic human reason. An expert system shelf, a software program commonly used for developing expert systems in a relatively short time, was used to develop a prototypical expert system for the design of heating, ventilating, and air-conditioning (HVAC) systems in buildings. Because HVAC design involves several related knowledge domains, developing an expert system for HVAC design requires the integration of several smaller expert systems known as knowledge bases. A menu program and several auxiliary programs for gathering data, completing calculations, printing project reports, and passing data between the knowledge bases are neededmore » and have been developed to join the separate knowledge bases into one simple-to-use program unit.« less
Le Moual, Nicole; Zock, Jan-Paul; Dumas, Orianne; Lytras, Theodore; Andersson, Eva; Lillienberg, Linnéa; Schlünssen, Vivi; Benke, Geza; Kromhout, Hans
2018-07-01
We aimed to update an asthmagen job exposure matrix (JEM) developed in the late 1990s. Main reasons were: the number of suspected and recognised asthmagens has since tripled; understanding of the aetiological role of irritants in asthma and methodological insights in application of JEMs have emerged in the period. For each agent of the new occupational asthma-specific JEM (OAsJEM), a working group of three experts out of eight evaluated exposure for each International Standard Classification of Occupations, 1988 (ISCO-88) job code into three categories: 'high' (high probability of exposure and moderate-to-high intensity), 'medium' (low-to-moderate probability or low intensity) and 'unexposed'. Within a working group, experts evaluated exposures independently from each other. If expert assessments were inconsistent the final decision was taken by consensus. Specificity was favoured over sensitivity, that is, jobs were classified with high exposure only if the probability of exposure was high and the intensity moderate-to-high. In the final review, all experts checked assigned exposures and proposed/improved recommendations for expert re-evaluation after default application of the JEM. The OAsJEM covers exposures to 30 sensitisers/irritants, including 12 newly recognised, classified into seven broad groups. Initial agreement between the three experts was mostly fair to moderate (κ values 0.2-0.5). Out of 506 ISCO-88 codes, the majority was classified as unexposed (from 82.6% (organic solvents) to 99.8% (persulfates)) and a minority as 'high-exposed' (0.2% (persulfates) to 2.6% (organic solvents)). The OAsJEM developed to improve occupational exposure assessment may improve evaluations of associations with asthma in epidemiological studies and contribute to assessment of the burden of work-related asthma. © Article author(s) (or their employer(s) unless otherwise stated in the text of the article) 2018. All rights reserved. No commercial use is permitted unless otherwise expressly granted.
Pan, Yi-Ling; Hwang, Ai-Wen; Simeonsson, Rune J; Lu, Lu; Liao, Hua-Fang
2015-01-01
Comprehensive description of functioning is important in providing early intervention services for infants with developmental delay/disabilities (DD). A code set of the International Classification of Functioning, Disability and Health: Children and Youth Version (ICF-CY) could facilitate the practical use of the ICF-CY in team evaluation. The purpose of this study was to derive an ICF-CY code set for infants under three years of age with early delay and disabilities (EDD Code Set) for initial team evaluation. The EDD Code Set based on the ICF-CY was developed on the basis of a Delphi survey of international professionals experienced in implementing the ICF-CY and professionals in early intervention service system in Taiwan. Twenty-five professionals completed the Delphi survey. A total of 82 ICF-CY second-level categories were identified for the EDD Code Set, including 28 categories from the domain Activities and Participation, 29 from body functions, 10 from body structures and 15 from environmental factors. The EDD Code Set of 82 ICF-CY categories could be useful in multidisciplinary team evaluations to describe functioning of infants younger than three years of age with DD, in a holistic manner. Future validation of the EDD Code Set and examination of its clinical utility are needed. The EDD Code Set with 82 essential ICF-CY categories could be useful in the initial team evaluation as a common language to describe functioning of infants less than three years of age with developmental delay/disabilities, with a more holistic view. The EDD Code Set including essential categories in activities and participation, body functions, body structures and environmental factors could be used to create a functional profile for each infant with special needs and to clarify the interaction of child and environment accounting for the child's functioning.
Engineering monitoring expert system's developer
NASA Technical Reports Server (NTRS)
Lo, Ching F.
1991-01-01
This research project is designed to apply artificial intelligence technology including expert systems, dynamic interface of neural networks, and hypertext to construct an expert system developer. The developer environment is specifically suited to building expert systems which monitor the performance of ground support equipment for propulsion systems and testing facilities. The expert system developer, through the use of a graphics interface and a rule network, will be transparent to the user during rule constructing and data scanning of the knowledge base. The project will result in a software system that allows its user to build specific monitoring type expert systems which monitor various equipments used for propulsion systems or ground testing facilities and accrues system performance information in a dynamic knowledge base.
[Sexual offences--selected cases].
Łabecka, Marzena; Jarzabek-Bielecka, Grazyna; Lorkiewicz-Muszyńska, Dorota
2013-04-01
Expert testimony on violence victims also includes victims of sexual assault. The role of an expert is to classify the injuries by their severity as defined in art. 157 156 or 217 of the Criminal Code pertaining to crimes against health and life. Also, the role of an expert opinion is to determine whether the injuries identified during the exam occurred at the time and under the circumstances stated in medical history. The examination of sexual assault victims is conducted by two experts: a gynecologist and a forensic physician. Most examinations are performed at different times and various medical centers. The conclusions are presented in an official report. Regardless of victim age, all sexual crimes are investigated ex officio by the Police Department and the Prosecutor's Office. Further legal classification of criminal offenses is the task of an appropriate legal body and the offenses are codified in accordance with the provisions of chapter XXV of the Criminal Code, articles 197 - 205. In controversial cases, i.e. when two different expert opinions appear on the same case, or if, according to the law enforcement, a medical opinion is insufficient for some reason, an appropriate expert or team of experts is appointed to resolve the problem. To present selected cases of sexual violence victims treated at the Department of Gynecology and assessed at the Department of Forensic Medicine with reference to the challenges regarding qualification of the sustained injuries and clinical diagnoses. Research material included selected forensic opinions developed for law enforcement offices that involved victims of sexual violence. The expert opinions were prepared either on the basis of submitted evidence, or both, submitted evidence and examination of the victim at the Department of Forensic Medicine. Moreover the article presents a case of a patient examined and treated at the Department of Gynecology in Poznan. Based on the selected cases, the authors conclude that a medico-legal expert cannot uncritically accept previous diagnoses. Moreover every expert is given the right and obligation to verify them. The need for complete, rapid and almost simultaneous colaboration between physicians in charge of the case, forensic doctors, police officers and prosecutors was demonstrated. Lack of cooperation may give rise to different opinions, leading to unnecessary elongation of the medico-legal procedures. It was observed that time plays a crucial role if qualification of an injury is required. The obligation of medical staff to inform the law enforcement about all cases of child abuse was also emphasized.
Artificial intelligence and space power systems automation
NASA Technical Reports Server (NTRS)
Weeks, David J.
1987-01-01
Various applications of artificial intelligence to space electrical power systems are discussed. An overview is given of completed, on-going, and planned knowledge-based system activities. These applications include the Nickel-Cadmium Battery Expert System (NICBES) (the expert system interfaced with the Hubble Space Telescope electrical power system test bed); the early work with the Space Station Experiment Scheduler (SSES); the three expert systems under development in the space station advanced development effort in the core module power management and distribution system test bed; planned cooperation of expert systems in the Core Module Power Management and Distribution (CM/PMAD) system breadboard with expert systems for the space station at other research centers; and the intelligent data reduction expert system under development.
Considerations in development of expert systems for real-time space applications
NASA Technical Reports Server (NTRS)
Murugesan, S.
1988-01-01
Over the years, demand on space systems has increased tremendously and this trend will continue for the near future. Enhanced capabilities of space systems, however, can only be met with increased complexity and sophistication of onboard and ground systems. Artificial Intelligence and expert system techniques have great potential in space applications. Expert systems could facilitate autonomous decision making, improve in-orbit fault diagnosis and repair, enhance performance and reduce reliance on ground support. However, real-time expert systems, unlike conventional off-line consultative systems, have to satisfy certain special stringent requirements before they could be used for onboard space applications. Challenging and interesting new environments are faced while developing expert system space applications. This paper discusses the special characteristics, requirements and typical life cycle issues for onboard expert systems. Further, it also describes considerations in design, development, and implementation which are particularly important to real-time expert systems for space applications.
An hierarchical approach to performance evaluation of expert systems
NASA Technical Reports Server (NTRS)
Dominick, Wayne D. (Editor); Kavi, Srinu
1985-01-01
The number and size of expert systems is growing rapidly. Formal evaluation of these systems - which is not performed for many systems - increases the acceptability by the user community and hence their success. Hierarchical evaluation that had been conducted for computer systems is applied for expert system performance evaluation. Expert systems are also evaluated by treating them as software systems (or programs). This paper reports many of the basic concepts and ideas in the Performance Evaluation of Expert Systems Study being conducted at the University of Southwestern Louisiana.
An engineering approach to the use of expert systems technology in avionics applications
NASA Technical Reports Server (NTRS)
Duke, E. L.; Regenie, V. A.; Brazee, M.; Brumbaugh, R. W.
1986-01-01
The concept of using a knowledge compiler to transform the knowledge base and inference mechanism of an expert system into a conventional program is presented. The need to accommodate real-time systems requirements in applications such as embedded avionics is outlined. Expert systems and a brief comparison of expert systems and conventional programs are reviewed. Avionics applications of expert systems are discussed before the discussions of applying the proposed concept to example systems using forward and backward chaining.
Expert systems for real-time monitoring and fault diagnosis
NASA Technical Reports Server (NTRS)
Edwards, S. J.; Caglayan, A. K.
1989-01-01
Methods for building real-time onboard expert systems were investigated, and the use of expert systems technology was demonstrated in improving the performance of current real-time onboard monitoring and fault diagnosis applications. The potential applications of the proposed research include an expert system environment allowing the integration of expert systems into conventional time-critical application solutions, a grammar for describing the discrete event behavior of monitoring and fault diagnosis systems, and their applications to new real-time hardware fault diagnosis and monitoring systems for aircraft.
Equating an expert system to a classifier in order to evaluate the expert system
NASA Technical Reports Server (NTRS)
Odell, Patrick L.
1989-01-01
A strategy to evaluate an expert system is formulated. The strategy proposed is based on finding an equivalent classifier to an expert system and evaluate that classifier with respect to an optimal classifier, a Bayes classifier. Here it is shown that for the rules considered an equivalent classifier exists. Also, a brief consideration of meta and meta-meta rules is included. Also, a taxonomy of expert systems is presented and an assertion made that an equivalent classifier exists for each type of expert system in the taxonomy with associated sets of underlying assumptions.
Small Knowledge-Based Systems in Education and Training: Something New Under the Sun.
ERIC Educational Resources Information Center
Wilson, Brent G.; Welsh, Jack R.
1986-01-01
Discusses artificial intelligence, robotics, natural language processing, and expert or knowledge-based systems research; examines two large expert systems, MYCIN and XCON; and reviews the resources required to build large expert systems and affordable smaller systems (intelligent job aids) for training. Expert system vendors and products are…
Romagnoli, Katrina M; Nelson, Scott D; Hines, Lisa; Empey, Philip; Boyce, Richard D; Hochheiser, Harry
2017-02-22
Drug information compendia and drug-drug interaction information databases are critical resources for clinicians and pharmacists working to avoid adverse events due to exposure to potential drug-drug interactions (PDDIs). Our goal is to develop information models, annotated data, and search tools that will facilitate the interpretation of PDDI information. To better understand the information needs and work practices of specialists who search and synthesize PDDI evidence for drug information resources, we conducted an inquiry that combined a thematic analysis of published literature with unstructured interviews. Starting from an initial set of relevant articles, we developed search terms and conducted a literature search. Two reviewers conducted a thematic analysis of included articles. Unstructured interviews with drug information experts were conducted and similarly coded. Information needs, work processes, and indicators of potential strengths and weaknesses of information systems were identified. Review of 92 papers and 10 interviews identified 56 categories of information needs related to the interpretation of PDDI information including drug and interaction information; study design; evidence including clinical details, quality and content of reports, and consequences; and potential recommendations. We also identified strengths/weaknesses of PDDI information systems. We identified the kinds of information that might be most effective for summarizing PDDIs. The drug information experts we interviewed had differing goals, suggesting a need for detailed information models and flexible presentations. Several information needs not discussed in previous work were identified, including temporal overlaps in drug administration, biological plausibility of interactions, and assessment of the quality and content of reports. Richly structured depictions of PDDI information may help drug information experts more effectively interpret data and develop recommendations. Effective information models and system designs will be needed to maximize the utility of this information.
Forensic mental health assessment in France: recommendations for quality improvement.
Combalbert, Nicolas; Andronikof, Anne; Armand, Marine; Robin, Cécile; Bazex, Hélène
2014-01-01
The quality of forensic mental health assessment has been a growing concern in various countries on both sides of the Atlantic, but the legal systems are not always comparable and some aspects of forensic assessment are specific to a given country. This paper describes the legal context of forensic psychological assessment in France (i.e. pre-trial investigation phase entrusted to a judge, with mental health assessment performed by preselected professionals called "experts" in French), its advantages and its pitfalls. Forensic psychiatric or psychological assessment is often an essential and decisive element in criminal cases, but since a judiciary scandal which was made public in 2005 (the Outreau case) there has been increasing criticism from the public and the legal profession regarding the reliability of clinical conclusions. Several academic studies and a parliamentary report have highlighted various faulty aspects in both the judiciary process and the mental health assessments. The heterogeneity of expert practices in France appears to be mainly related to a lack of consensus on several core notions such as mental health diagnosis or assessment methods, poor working conditions, lack of specialized training, and insufficient familiarity with the Code of Ethics. In this article we describe and analyze the French practice of forensic psychologists and psychiatrists in criminal cases and propose steps that could be taken to improve its quality, such as setting up specialized training courses, enforcing the Code of Ethics for psychologists, and calling for consensus on diagnostic and assessment methods. Copyright © 2014 Elsevier Ltd. All rights reserved.
Expert systems for C3I. Volume 1. A user's introduction
NASA Astrophysics Data System (ADS)
Clapp, J. A.; Hockett, S. M.; Prelle, M. J.; Tallant, A. M.; Triant, D. D.
1985-10-01
There has been a tremendous burgeoning of interest in artificial intelligence (AI) over the last few years. Investments of commercial and government sponsors reflect a widespread belief that AI is now ready for practical applications. The area of AI currently receiving the greatest attention and investment is expert system technology. Most major high tech corporations have begun to develop expert systems, and many software houses specializing in expert system tools and applications have recently appeared. The defense community is one of the heaviest investors in expert system technology, and within this community one of the application areas receiving greatest attention is C3I. Many ESD programs are now beginning to ask whether expert system applications for C3I are ready for incorporation into ESD-developed systems, and, if so, what are the potential benefits and risks of doing so. This report was prepared to help ESD and MITRE personnel working on acquisition programs to address these issues and to gain a better understanding of what expert systems are all about. The primary intention of this report is to investigate what expert systems are and the advances that are being made in expert system technology for C3I applications. The report begins with a brief tutorial on expert systems, emphasizing how they differ from conventional software systems and what they are best at doing.
Interfaces and Expert Systems for Online Retrieval.
ERIC Educational Resources Information Center
Kehoe, Cynthia A.
1985-01-01
This paper reviews the history of separate online system interfaces which led to efforts to develop expert systems for searching databases, particularly for end users, and introduces the research on such expert systems. Appended is a bibliography of sources on interfaces and expert systems for online retrieval. (Author/EJS)
Expert and Knowledge Based Systems.
ERIC Educational Resources Information Center
Demaid, Adrian; Edwards, Lyndon
1987-01-01
Discusses the nature and current state of knowledge-based systems and expert systems. Describes an expert system from the viewpoints of a computer programmer and an applications expert. Addresses concerns related to materials selection and forecasts future developments in the teaching of materials engineering. (ML)
Barriers Against Implementing Blunt Abdominal Trauma Guidelines in a Hospital: A Qualitative Study.
Zaboli, Rouhollah; Tofighi, Shahram; Aghighi, Ali; Shokouh, Seyyed Javad Hosaini; Naraghi, Nader; Goodarzi, Hassan
2016-08-01
Clinical practice guidelines are structured recommendations that help physicians and patients to make proper decisions when dealing with a specific clinical condition. Because blunt abdominal trauma causes a various range of mild, single-system, and multisystem injuries, early detection will help to reduce mortality and resulting disability. Emergency treatment should be initiated based on CPGs. This study aimed to determine the variables affecting implementing blunt abdominal trauma CPGs in an Iranian hospital. This study was conducted as a qualitative and phenomenology study in the Family Hospital in Tehran (Iran) in 2015. The research population included eight experts and key people in the area of blunt abdominal trauma clinical practice guidelines. Sampling was based on purposive and nonrandom methods. A semistructured interview was done for the data collection. A framework method was applied for the data analysis by using Atlas.ti software. After framework analyzing and various reviewing and deleting and combining the codes from 251 codes obtained, 15 families and five super families were extracted, including technical knowledge barriers, economical barriers, barriers related to deployment and monitoring, political will barriers, and managing barriers. Structural reform is needed for eliminating the defects available in the healthcare system. As with most of the codes, subconcepts and concepts are classified into the field of human resources; it seems that the education and knowledge will be more important than other resources such as capital and equipment.
The SERGISAI procedure for seismic risk assessment
NASA Astrophysics Data System (ADS)
Zonno, G.; Garcia-Fernandez, M.; Jimenez, M.J.; Menoni, S.; Meroni, F.; Petrini, V.
The European project SERGISAI developed a computational tool where amethodology for seismic risk assessment at different geographical scales hasbeen implemented. Experts of various disciplines, including seismologists,engineers, planners, geologists, and computer scientists, co-operated in anactual multidisciplinary process to develop this tool. Standard proceduralcodes, Geographical Information Systems (GIS), and Artificial Intelligence(AI) techniques compose the whole system, that will enable the end userto carry out a complete seismic risk assessment at three geographical scales:regional, sub-regional and local. At present, single codes or models thathave been incorporated are not new in general, but the modularity of theprototype, based on a user-friendly front-end, offers potential users thepossibility of updating or replacing any code or model if desired. Theproposed procedure is a first attempt to integrate tools, codes and methodsfor assessing expected earthquake damage, and it was mainly designedto become a useful support for civil defence and land use planning agencies.Risk factors have been treated in the most suitable way for each one, interms of level of detail, kind of parameters and units of measure.Identifying various geographical scales is not a mere question of dimension;since entities to be studied correspond to areas defined by administrativeand geographical borders. The procedure was applied in the following areas:Toscana in Italy, for the regional scale, the Garfagnana area in Toscana, forthe sub-regional scale, and a part of Barcelona city, Spain, for the localscale.
NASA Astrophysics Data System (ADS)
Terrett, D. L.
The basis of this report is 2 days spent with an AVS expert from DEC's CERN project office attempting to convert an ADAM application into an AVS module. The experiment was successful in that we succeeded in running a KAPPA application (ADD) as a module in an AVS network without modifying the applications program code in any way. We took many short cuts and it became clear that doing the job properly would be a major exercise, but we learned enough to know that the job is feasible and gained a clear idea of what the final system would look like and what it would be capable of.
How to Build an AppleSeed: A Parallel Macintosh Cluster for Numerically Intensive Computing
NASA Astrophysics Data System (ADS)
Decyk, V. K.; Dauger, D. E.
We have constructed a parallel cluster consisting of a mixture of Apple Macintosh G3 and G4 computers running the Mac OS, and have achieved very good performance on numerically intensive, parallel plasma particle-incell simulations. A subset of the MPI message-passing library was implemented in Fortran77 and C. This library enabled us to port code, without modification, from other parallel processors to the Macintosh cluster. Unlike Unix-based clusters, no special expertise in operating systems is required to build and run the cluster. This enables us to move parallel computing from the realm of experts to the main stream of computing.
Posted wait times an added advantage to multi-facility systems?
2011-04-01
Methodist Le Bonheur Healthcare in Memphis, TN, is investigating whether posting ED wait times via the internet can positively impact patient flow in the six EDs the health system operates in the Memphis region. The health system began posting wait times in August 2010, resulting in increases in ED volume ranging from 6% to 10%. The health system is monitoring ED arrivals by zip code to assess any impact on load balancing between its busy EDs. One marketing challenge is that a competitor is posting ED wait times as well, but it is posting the time it takes for a patient to be placed in a bed as opposed to the door-to-provider time that Methodist Le Bonheur is posting. The approach has the most impact on lower-acuity patients, but experts worry that in the future, payers may not be reimbursed for ED care for these patients.
Reliability and performance evaluation of systems containing embedded rule-based expert systems
NASA Technical Reports Server (NTRS)
Beaton, Robert M.; Adams, Milton B.; Harrison, James V. A.
1989-01-01
A method for evaluating the reliability of real-time systems containing embedded rule-based expert systems is proposed and investigated. It is a three stage technique that addresses the impact of knowledge-base uncertainties on the performance of expert systems. In the first stage, a Markov reliability model of the system is developed which identifies the key performance parameters of the expert system. In the second stage, the evaluation method is used to determine the values of the expert system's key performance parameters. The performance parameters can be evaluated directly by using a probabilistic model of uncertainties in the knowledge-base or by using sensitivity analyses. In the third and final state, the performance parameters of the expert system are combined with performance parameters for other system components and subsystems to evaluate the reliability and performance of the complete system. The evaluation method is demonstrated in the context of a simple expert system used to supervise the performances of an FDI algorithm associated with an aircraft longitudinal flight-control system.
Kaspar, Mathias; Fette, Georg; Güder, Gülmisal; Seidlmayer, Lea; Ertl, Maximilian; Dietrich, Georg; Greger, Helmut; Puppe, Frank; Störk, Stefan
2018-04-17
Heart failure is the predominant cause of hospitalization and amongst the leading causes of death in Germany. However, accurate estimates of prevalence and incidence are lacking. Reported figures originating from different information sources are compromised by factors like economic reasons or documentation quality. We implemented a clinical data warehouse that integrates various information sources (structured parameters, plain text, data extracted by natural language processing) and enables reliable approximations to the real number of heart failure patients. Performance of ICD-based diagnosis in detecting heart failure was compared across the years 2000-2015 with (a) advanced definitions based on algorithms that integrate various sources of the hospital information system, and (b) a physician-based reference standard. Applying these methods for detecting heart failure in inpatients revealed that relying on ICD codes resulted in a marked underestimation of the true prevalence of heart failure, ranging from 44% in the validation dataset to 55% (single year) and 31% (all years) in the overall analysis. Percentages changed over the years, indicating secular changes in coding practice and efficiency. Performance was markedly improved using search and permutation algorithms from the initial expert-specified query (F1 score of 81%) to the computer-optimized query (F1 score of 86%) or, alternatively, optimizing precision or sensitivity depending on the search objective. Estimating prevalence of heart failure using ICD codes as the sole data source yielded unreliable results. Diagnostic accuracy was markedly improved using dedicated search algorithms. Our approach may be transferred to other hospital information systems.
Executing CLIPS expert systems in a distributed environment
NASA Technical Reports Server (NTRS)
Taylor, James; Myers, Leonard
1990-01-01
This paper describes a framework for running cooperating agents in a distributed environment to support the Intelligent Computer Aided Design System (ICADS), a project in progress at the CAD Research Unit of the Design Institute at the California Polytechnic State University. Currently, the systems aids an architectural designer in creating a floor plan that satisfies some general architectural constraints and project specific requirements. At the core of ICADS is the Blackboard Control System. Connected to the blackboard are any number of domain experts called Intelligent Design Tools (IDT). The Blackboard Control System monitors the evolving design as it is being drawn and helps resolve conflicts from the domain experts. The user serves as a partner in this system by manipulating the floor plan in the CAD system and validating recommendations made by the domain experts. The primary components of the Blackboard Control System are two expert systems executed by a modified CLIPS shell. The first is the Message Handler. The second is the Conflict Resolver. The Conflict Resolver synthesizes the suggestions made by the domain experts, which can be either CLIPS expert systems, or compiled C programs. In DEMO1, the current ICADS prototype, the CLIPS domain expert systems are Acoustics, Lighting, Structural, and Thermal; the compiled C domain experts are the CAD system and the User Interface.
Expert system prototype developments for NASA-KSC business and engineering applications
NASA Technical Reports Server (NTRS)
Ragusa, James M.; Gonzalez, Avelino J.
1988-01-01
Prototype expert systems developed for a variety of NASA projects in the business/management and engineering domains are discussed. Business-related problems addressed include an assistant for simulating launch vehicle processing, a plan advisor for the acquisition of automated data processing equipment, and an expert system for the identification of customer requirements. Engineering problems treated include an expert system for detecting potential ignition sources in LOX and gaseous-oxygen transportation systems and an expert system for hazardous-gas detection.
A CLIPS expert system for maximizing alfalfa (Medicago Sativa L.) production
NASA Technical Reports Server (NTRS)
Engel, B. A.; Jones, D. D.; Rhykerd, R. L.; Rhykerd, L. M.; Rhykerd, C. L., Jr.; Rhykerd, C. L.
1990-01-01
An alfalfa management expert system originally developed by Purdue University agricultural scientists on the PC Plus expert system shell from Texas Instrument has been updated and successfully converted to CLIPS (C Language Integrated Production System). This reduces the cost and legal restrictions associated with making the expert system available to agribusiness industries, extension personnel and farm managers and operators. The expert system includes recommendations concerning soil drainage, liming, P and K fertilization, weed control, variety selection and seeding rate including pure live seeds.
Knowledge-based fault diagnosis system for refuse collection vehicle
DOE Office of Scientific and Technical Information (OSTI.GOV)
Tan, CheeFai; Juffrizal, K.; Khalil, S. N.
The refuse collection vehicle is manufactured by local vehicle body manufacturer. Currently; the company supplied six model of the waste compactor truck to the local authority as well as waste management company. The company is facing difficulty to acquire the knowledge from the expert when the expert is absence. To solve the problem, the knowledge from the expert can be stored in the expert system. The expert system is able to provide necessary support to the company when the expert is not available. The implementation of the process and tool is able to be standardize and more accurate. The knowledgemore » that input to the expert system is based on design guidelines and experience from the expert. This project highlighted another application on knowledge-based system (KBS) approached in trouble shooting of the refuse collection vehicle production process. The main aim of the research is to develop a novel expert fault diagnosis system framework for the refuse collection vehicle.« less
NASA Technical Reports Server (NTRS)
Mclean, David R.; Tuchman, Alan; Potter, William J.
1991-01-01
Recently, many expert systems were developed in a LISP environment and then ported to the real world C environment before the final system is delivered. This situation may require that the entire system be completely rewritten in C and may actually result in a system which is put together as quickly as possible with little regard for maintainability and further evolution. With the introduction of high performance UNIX and X-windows based workstations, a great deal of the advantages of developing a first system in the LISP environment have become questionable. A C-based AI development effort is described which is based on a software tools approach with emphasis on reusability and maintainability of code. The discussion starts with simple examples of how list processing can easily be implemented in C and then proceeds to the implementations of frames and objects which use dynamic memory allocation. The implementation of procedures which use depth first search, constraint propagation, context switching and a blackboard-like simulation environment are described. Techniques for managing the complexity of C-based AI software are noted, especially the object-oriented techniques of data encapsulation and incremental development. Finally, all these concepts are put together by describing the components of planning software called the Planning And Resource Reasoning (PARR) shell. This shell was successfully utilized for scheduling services of the Tracking and Data Relay Satellite System for the Earth Radiation Budget Satellite since May 1987 and will be used for operations scheduling of the Explorer Platform in November 1991.
A parallel strategy for implementing real-time expert systems using CLIPS
NASA Technical Reports Server (NTRS)
Ilyes, Laszlo A.; Villaseca, F. Eugenio; Delaat, John
1994-01-01
As evidenced by current literature, there appears to be a continued interest in the study of real-time expert systems. It is generally recognized that speed of execution is only one consideration when designing an effective real-time expert system. Some other features one must consider are the expert system's ability to perform temporal reasoning, handle interrupts, prioritize data, contend with data uncertainty, and perform context focusing as dictated by the incoming data to the expert system. This paper presents a strategy for implementing a real time expert system on the iPSC/860 hypercube parallel computer using CLIPS. The strategy takes into consideration not only the execution time of the software, but also those features which define a true real-time expert system. The methodology is then demonstrated using a practical implementation of an expert system which performs diagnostics on the Space Shuttle Main Engine (SSME). This particular implementation uses an eight node hypercube to process ten sensor measurements in order to simultaneously diagnose five different failure modes within the SSME. The main program is written in ANSI C and embeds CLIPS to better facilitate and debug the rule based expert system.
Fuel management optimization using genetic algorithms and expert knowledge
DOE Office of Scientific and Technical Information (OSTI.GOV)
DeChaine, M.D.; Feltus, M.A.
1996-09-01
The CIGARO fuel management optimization code based on genetic algorithms is described and tested. The test problem optimized the core lifetime for a pressurized water reactor with a penalty function constraint on the peak normalized power. A bit-string genotype encoded the loading patterns, and genotype bias was reduced with additional bits. Expert knowledge about fuel management was incorporated into the genetic algorithm. Regional crossover exchanged physically adjacent fuel assemblies and improved the optimization slightly. Biasing the initial population toward a known priority table significantly improved the optimization.
Explainable expert systems: A research program in information processing
NASA Technical Reports Server (NTRS)
Paris, Cecile L.
1993-01-01
Our work in Explainable Expert Systems (EES) had two goals: to extend and enhance the range of explanations that expert systems can offer, and to ease their maintenance and evolution. As suggested in our proposal, these goals are complementary because they place similar demands on the underlying architecture of the expert system: they both require the knowledge contained in a system to be explicitly represented, in a high-level declarative language and in a modular fashion. With these two goals in mind, the Explainable Expert Systems (EES) framework was designed to remedy limitations to explainability and evolvability that stem from related fundamental flaws in the underlying architecture of current expert systems.
SWAN: An expert system with natural language interface for tactical air capability assessment
NASA Technical Reports Server (NTRS)
Simmons, Robert M.
1987-01-01
SWAN is an expert system and natural language interface for assessing the war fighting capability of Air Force units in Europe. The expert system is an object oriented knowledge based simulation with an alternate worlds facility for performing what-if excursions. Responses from the system take the form of generated text, tables, or graphs. The natural language interface is an expert system in its own right, with a knowledge base and rules which understand how to access external databases, models, or expert systems. The distinguishing feature of the Air Force expert system is its use of meta-knowledge to generate explanations in the frame and procedure based environment.
DELTA: An Expert System for Diesel Electric Locomotive Repair
1984-06-01
Rules and Inference Mechanisms. AD-P003 943 The ACE (Automated Cable Expert) Exlpelient: Initial Evaluation of an Expert System for Preventive...tions. The first field prototype expert system, designated CATS -i (Computer-Aided Troubleshooting System - Version 1), was delivered in July 1983 and is
DiClemente, Carlo C; Crouch, Taylor Berens; Norwood, Amber E Q; Delahanty, Janine; Welsh, Christopher
2015-03-01
Screening, brief intervention, and referral to treatment (SBIRT) has become an empirically supported and widely implemented approach in primary and specialty care for addressing substance misuse. Accordingly, training of providers in SBIRT has increased exponentially in recent years. However, the quality and fidelity of training programs and subsequent interventions are largely unknown because of the lack of SBIRT-specific evaluation tools. The purpose of this study was to create a coding scale to assess quality and fidelity of SBIRT interactions addressing alcohol, tobacco, illicit drugs, and prescription medication misuse. The scale was developed to evaluate performance in an SBIRT residency training program. Scale development was based on training protocol and competencies with consultation from Motivational Interviewing coding experts. Trained medical residents practiced SBIRT with standardized patients during 10- to 15-min videotaped interactions. This study included 25 tapes from the Family Medicine program coded by 3 unique coder pairs with varying levels of coding experience. Interrater reliability was assessed for overall scale components and individual items via intraclass correlation coefficients. Coder pair-specific reliability was also assessed. Interrater reliability was excellent overall for the scale components (>.85) and nearly all items. Reliability was higher for more experienced coders, though still adequate for the trained coder pair. Descriptive data demonstrated a broad range of adherence and skills. Subscale correlations supported concurrent and discriminant validity. Data provide evidence that the MD3 SBIRT Coding Scale is a psychometrically reliable coding system for evaluating SBIRT interactions and can be used to evaluate implementation skills for fidelity, training, assessment, and research. Recommendations for refinement and further testing of the measure are discussed. (PsycINFO Database Record (c) 2015 APA, all rights reserved).
Adding intelligent services to an object oriented system
NASA Technical Reports Server (NTRS)
Robideaux, Bret R.; Metzler, Theodore A.
1994-01-01
As today's software becomes increasingly complex, the need grows for intelligence of one sort or another to becomes part of the application, often an intelligence that does not readily fit the paradigm of one's software development. There are many methods of developing software, but at this time, the most promising is the object oriented (OO) method. This method involves an analysis to abstract the problem into separate 'objects' that are unique in the data that describe them and the behavior that they exhibit, and eventually to convert this analysis into computer code using a programming language that was designed (or retrofitted) for OO implementation. This paper discusses the creation of three different applications that are analyzed, designed, and programmed using the Shlaer/Mellor method of OO development and C++ as the programming language. All three, however, require the use of an expert system to provide an intelligence that C++ (or any other 'traditional' language) is not directly suited to supply. The flexibility of CLIPS permitted us to make modifications to it that allow seamless integration with any of our applications that require an expert system. We illustrate this integration with the following applications: (1) an after action review (AAR) station that assists a reviewer in watching a simulated tank battle and developing an AAR to critique the performance of the participants in the battle; (2) an embedded training system and over-the-shoulder coach for howitzer crewmen; and (3) a system to identify various chemical compounds from their infrared absorption spectra.
Techniques for capturing expert knowledge - An expert systems/hypertext approach
NASA Technical Reports Server (NTRS)
Lafferty, Larry; Taylor, Greg; Schumann, Robin; Evans, Randy; Koller, Albert M., Jr.
1990-01-01
The knowledge-acquisition strategy developed for the Explosive Hazards Classification (EHC) Expert System is described in which expert systems and hypertext are combined, and broad applications are proposed. The EHC expert system is based on rapid prototyping in which primary knowledge acquisition from experts is not emphasized; the explosive hazards technical bulletin, technical guidance, and minimal interviewing are used to develop the knowledge-based system. Hypertext is used to capture the technical information with respect to four issues including procedural, materials, test, and classification issues. The hypertext display allows the integration of multiple knowlege representations such as clarifications or opinions, and thereby allows the performance of a broad range of tasks on a single machine. Among other recommendations, it is suggested that the integration of hypertext and expert systems makes the resulting synergistic system highly efficient.
Weighing In: The Taste-Engineering Frame in Obesity Expert Discourse
Zimmerman, Frederick J.; Gilliam, Franklin D.
2015-01-01
Objectives. We sought expert opinion on the problems with 2 dominant obesity-prevention discourse frames—personal responsibility and the environment—and examined alternative frames for understanding and addressing obesity. Methods. We conducted 60-minute, semistructured interviews with 15 US-based obesity experts. We manually coded and entered interview transcripts into software, generating themes and subthematic areas that captured the debate’s essence. Results. Although the environmental frame is the dominant model used in communications with the public and policymakers, several experts found that communicating key messages within this frame was difficult because of the enormity of the obesity problem. A subframe of the environmental frame—the taste-engineering frame—identifies food industry strategies to influence the overconsumption of certain foods and beverages. This emerging frame deconstructs the environmental frame so that causal attributes and responsible agents are more easily identifiable and proposed policies and public health interventions more salient. Conclusions. Expert interviews are an invaluable resource for understanding how experts use frames in discussing their work and in conversations with the public and policymakers. Future empirical studies testing the effectiveness of the taste-engineering frame on public opinion and support for structural-level health policies are needed. PMID:25602888
Weighing in: the taste-engineering frame in obesity expert discourse.
Ortiz, Selena E; Zimmerman, Frederick J; Gilliam, Franklin D
2015-03-01
We sought expert opinion on the problems with 2 dominant obesity-prevention discourse frames-personal responsibility and the environment-and examined alternative frames for understanding and addressing obesity. We conducted 60-minute, semistructured interviews with 15 US-based obesity experts. We manually coded and entered interview transcripts into software, generating themes and subthematic areas that captured the debate's essence. Although the environmental frame is the dominant model used in communications with the public and policymakers, several experts found that communicating key messages within this frame was difficult because of the enormity of the obesity problem. A subframe of the environmental frame--the taste-engineering frame--identifies food industry strategies to influence the overconsumption of certain foods and beverages. This emerging frame deconstructs the environmental frame so that causal attributes and responsible agents are more easily identifiable and proposed policies and public health interventions more salient. Expert interviews are an invaluable resource for understanding how experts use frames in discussing their work and in conversations with the public and policymakers. Future empirical studies testing the effectiveness of the taste-engineering frame on public opinion and support for structural-level health policies are needed.
Expert Systems--The New International Language of Business.
ERIC Educational Resources Information Center
Sondak, Norman E.; And Others
A discussion of expert systems, computer programs designed to simulate human reasoning and expertise, begins with the assumption that few business educators understand the impact that expert systems will have on international business. The fundamental principles of the design and development of expert systems in business are outlined, with special…
Expert Systems in Education and Training: Automated Job Aids or Sophisticated Instructional Media?
ERIC Educational Resources Information Center
Romiszowski, Alexander J.
1987-01-01
Describes the current status and limitations of expert systems, and explores the possible applications of such systems in education and training. The use of expert systems as tutors, as job aids, and as a vehicle for students to develop their own expert systems on specific topics are discussed. (40 references) (CLB)
An object oriented generic controller using CLIPS
NASA Technical Reports Server (NTRS)
Nivens, Cody R.
1990-01-01
In today's applications, the need for the division of code and data has focused on the growth of object oriented programming. This philosophy gives software engineers greater control over the environment of an application. Yet the use of object oriented design does not exclude the need for greater understanding by the application of what the controller is doing. Such understanding is only possible by using expert systems. Providing a controller that is capable of controlling an object by using rule-based expertise would expedite the use of both object oriented design and expert knowledge of the dynamic of an environment in modern controllers. This project presents a model of a controller that uses the CLIPS expert system and objects in C++ to create a generic controller. The polymorphic abilities of C++ allow for the design of a generic component stored in individual data files. Accompanying the component is a set of rules written in CLIPS which provide the following: the control of individual components, the input of sensory data from components and the ability to find the status of a given component. Along with the data describing the application, a set of inference rules written in CLIPS allows the application to make use of sensory facts and status and control abilities. As a demonstration of this ability, the control of the environment of a house is provided. This demonstration includes the data files describing the rooms and their contents as far as devices, windows and doors. The rules used for the home consist of the flow of people in the house and the control of devices by the home owner.
Expert systems and simulation models; Proceedings of the Seminar, Tucson, AZ, November 18, 19, 1985
NASA Technical Reports Server (NTRS)
1986-01-01
The seminar presents papers on modeling and simulation methodology, artificial intelligence and expert systems, environments for simulation/expert system development, and methodology for simulation/expert system development. Particular attention is given to simulation modeling concepts and their representation, modular hierarchical model specification, knowledge representation, and rule-based diagnostic expert system development. Other topics include the combination of symbolic and discrete event simulation, real time inferencing, and the management of large knowledge-based simulation projects.
Strategies for adding adaptive learning mechanisms to rule-based diagnostic expert systems
NASA Technical Reports Server (NTRS)
Stclair, D. C.; Sabharwal, C. L.; Bond, W. E.; Hacke, Keith
1988-01-01
Rule-based diagnostic expert systems can be used to perform many of the diagnostic chores necessary in today's complex space systems. These expert systems typically take a set of symptoms as input and produce diagnostic advice as output. The primary objective of such expert systems is to provide accurate and comprehensive advice which can be used to help return the space system in question to nominal operation. The development and maintenance of diagnostic expert systems is time and labor intensive since the services of both knowledge engineer(s) and domain expert(s) are required. The use of adaptive learning mechanisms to increment evaluate and refine rules promises to reduce both time and labor costs associated with such systems. This paper describes the basic adaptive learning mechanisms of strengthening, weakening, generalization, discrimination, and discovery. Next basic strategies are discussed for adding these learning mechanisms to rule-based diagnostic expert systems. These strategies support the incremental evaluation and refinement of rules in the knowledge base by comparing the set of advice given by the expert system (A) with the correct diagnosis (C). Techniques are described for selecting those rules in the in the knowledge base which should participate in adaptive learning. The strategies presented may be used with a wide variety of learning algorithms. Further, these strategies are applicable to a large number of rule-based diagnostic expert systems. They may be used to provide either immediate or deferred updating of the knowledge base.
Building bridges across electronic health record systems through inferred phenotypic topics.
Chen, You; Ghosh, Joydeep; Bejan, Cosmin Adrian; Gunter, Carl A; Gupta, Siddharth; Kho, Abel; Liebovitz, David; Sun, Jimeng; Denny, Joshua; Malin, Bradley
2015-06-01
Data in electronic health records (EHRs) is being increasingly leveraged for secondary uses, ranging from biomedical association studies to comparative effectiveness. To perform studies at scale and transfer knowledge from one institution to another in a meaningful way, we need to harmonize the phenotypes in such systems. Traditionally, this has been accomplished through expert specification of phenotypes via standardized terminologies, such as billing codes. However, this approach may be biased by the experience and expectations of the experts, as well as the vocabulary used to describe such patients. The goal of this work is to develop a data-driven strategy to (1) infer phenotypic topics within patient populations and (2) assess the degree to which such topics facilitate a mapping across populations in disparate healthcare systems. We adapt a generative topic modeling strategy, based on latent Dirichlet allocation, to infer phenotypic topics. We utilize a variance analysis to assess the projection of a patient population from one healthcare system onto the topics learned from another system. The consistency of learned phenotypic topics was evaluated using (1) the similarity of topics, (2) the stability of a patient population across topics, and (3) the transferability of a topic across sites. We evaluated our approaches using four months of inpatient data from two geographically distinct healthcare systems: (1) Northwestern Memorial Hospital (NMH) and (2) Vanderbilt University Medical Center (VUMC). The method learned 25 phenotypic topics from each healthcare system. The average cosine similarity between matched topics across the two sites was 0.39, a remarkably high value given the very high dimensionality of the feature space. The average stability of VUMC and NMH patients across the topics of two sites was 0.988 and 0.812, respectively, as measured by the Pearson correlation coefficient. Also the VUMC and NMH topics have smaller variance of characterizing patient population of two sites than standard clinical terminologies (e.g., ICD9), suggesting they may be more reliably transferred across hospital systems. Phenotypic topics learned from EHR data can be more stable and transferable than billing codes for characterizing the general status of a patient population. This suggests that EHR-based research may be able to leverage such phenotypic topics as variables when pooling patient populations in predictive models. Copyright © 2015 Elsevier Inc. All rights reserved.
Detection of counterfeit electronic components through ambient mass spectrometry and chemometrics.
Pfeuffer, Kevin P; Caldwell, Jack; Shelley, Jake T; Ray, Steven J; Hieftje, Gary M
2014-09-21
In the last several years, illicit electronic components have been discovered in the inventories of several distributors and even installed in commercial and military products. Illicit or counterfeit electronic components include a broad category of devices that can range from the correct unit with a more recent date code to lower-specification or non-working systems with altered names, manufacturers and date codes. Current methodologies for identification of counterfeit electronics rely on visual microscopy by expert users and, while effective, are very time-consuming. Here, a plasma-based ambient desorption/ionization source, the flowing atmospheric pressure afterglow (FAPA) is used to generate a mass-spectral fingerprint from the surface of a variety of discrete electronic integrated circuits (ICs). Chemometric methods, specifically principal component analysis (PCA) and the bootstrapped error-adjusted single-sample technique (BEAST), are used successfully to differentiate between genuine and counterfeit ICs. In addition, chemical and physical surface-removal techniques are explored and suggest which surface-altering techniques were utilized by counterfeiters.
Yates, Kenneth; Sullivan, Maura; Clark, Richard
2012-01-01
Cognitive task analysis (CTA) methods were used for 2 surgical procedures to determine (1) the extent that experts omitted critical information, (2) the number of experts required to capture the optimalamount of information, and (3) the effectiveness of a CTA-informed curriculum. Six expert physicians for both the central venous catheter placement and open cricothyrotomy were interviewed. The transcripts were coded, corrected, and aggregated as a "gold standard." The information captured for each surgeon was then analyzed against the gold standard. Experts omitted an average of 34% of the decisions for the central venous catheter and 77% of the decisions for the Cric. Three to 4 experts were required to capture the optimal amount of information. A significant positive effect on performance (t([21]) = 2.08, P = .050), and self-efficacy ratings (t([18]) = 2.38, P = .029) were found for the CTA-informed curriculum for cricothyrotomy. CTA is an effective method to capture expertise in surgery and a valuable component to improve surgical training. Copyright © 2012 Elsevier Inc. All rights reserved.
A neural network architecture for implementation of expert systems for real time monitoring
NASA Technical Reports Server (NTRS)
Ramamoorthy, P. A.
1991-01-01
Since neural networks have the advantages of massive parallelism and simple architecture, they are good tools for implementing real time expert systems. In a rule based expert system, the antecedents of rules are in the conjunctive or disjunctive form. We constructed a multilayer feedforward type network in which neurons represent AND or OR operations of rules. Further, we developed a translator which can automatically map a given rule base into the network. Also, we proposed a new and powerful yet flexible architecture that combines the advantages of both fuzzy expert systems and neural networks. This architecture uses the fuzzy logic concepts to separate input data domains into several smaller and overlapped regions. Rule-based expert systems for time critical applications using neural networks, the automated implementation of rule-based expert systems with neural nets, and fuzzy expert systems vs. neural nets are covered.
Carinci, F; Van Gool, K; Mainz, J; Veillard, J; Pichora, E C; Januel, J M; Arispe, I; Kim, S M; Klazinga, N S
2015-04-01
To review and update the conceptual framework, indicator content and research priorities of the Organisation for Economic Cooperation and Development's (OECD) Health Care Quality Indicators (HCQI) project, after a decade of collaborative work. A structured assessment was carried out using a modified Delphi approach, followed by a consensus meeting, to assess the suite of HCQI for international comparisons, agree on revisions to the original framework and set priorities for research and development. International group of countries participating to OECD projects. Members of the OECD HCQI expert group. A reference matrix, based on a revised performance framework, was used to map and assess all seventy HCQI routinely calculated by the OECD expert group. A total of 21 indicators were agreed to be excluded, due to the following concerns: (i) relevance, (ii) international comparability, particularly where heterogeneous coding practices might induce bias, (iii) feasibility, when the number of countries able to report was limited and the added value did not justify sustained effort and (iv) actionability, for indicators that were unlikely to improve on the basis of targeted policy interventions. The revised OECD framework for HCQI represents a new milestone of a long-standing international collaboration among a group of countries committed to building common ground for performance measurement. The expert group believes that the continuation of this work is paramount to provide decision makers with a validated toolbox to directly act on quality improvement strategies. © The Author 2015. Published by Oxford University Press in association with the International Society for Quality in Health Care; all rights reserved.
Peyrin-Biroulet, Laurent; Cieza, Alarcos; Sandborn, William J; Coenen, Michaela; Chowers, Yehuda; Hibi, Toshifumi; Kostanjsek, Nenad; Stucki, Gerold
2011-01-01
Objective The impact of inflammatory bowel disease (IBD) on disability remains poorly understood. The World Health Organization's integrative model of human functioning and disability in the International Classification of Functioning, Disability and Health (ICF) makes disability assessment possible. The ICF is a hierarchical coding system with four levels of details that includes over 1400 categories. The aim of this study was to develop the first disability index for IBD by selecting most relevant ICF categories that are affected by IBD. Methods Relevant ICF categories were identified through four preparatory studies (systematic literature review, qualitative study, expert survey and cross-sectional study), which were presented at a consensus conference. Based on the identified ICF categories, a questionnaire to be filled in by clinicians, called the ‘IBD disability index’, was developed. Results The four preparatory studies identified 138 second-level categories: 75 for systematic literature review (153 studies), 38 for qualitative studies (six focus groups; 27 patients), 108 for expert survey (125 experts; 37 countries; seven occupations) and 98 for cross-sectional study (192 patients; three centres). The consensus conference (20 experts; 17 countries) led to the selection of 19 ICF core set categories that were used to develop the IBD disability index: seven on body functions, two on body structures, five on activities and participation and five on environmental factors. Conclusions The IBD disability index is now available. It will be used in studies to evaluate the long-term effect of IBD on patient functional status and will serve as a new endpoint in disease-modification trials. PMID:21646246
What Is An Expert System? ERIC Digest.
ERIC Educational Resources Information Center
Boss, Richard W.
This digest describes and defines the various components of an expert system, e.g., a computerized tool designed to enhance the quality and availability of knowledge required by decision makers. It is noted that expert systems differ from conventional applications software in the following areas: (1) the existence of the expert systems shell, or…
Rhetorical Consequences of the Computer Society: Expert Systems and Human Communication.
ERIC Educational Resources Information Center
Skopec, Eric Wm.
Expert systems are computer programs that solve selected problems by modelling domain-specific behaviors of human experts. These computer programs typically consist of an input/output system that feeds data into the computer and retrieves advice, an inference system using the reasoning and heuristic processes of human experts, and a knowledge…
Expert Systems: A Challenge for the Reading Profession.
ERIC Educational Resources Information Center
Balajthy, Ernest
The expert systems are designed to imitate the reasoning of a human expert in a content area field. Designed to be advisors, these software systems combine the content area knowledge and decision-making ability of an expert with the user's understanding and knowledge of particular circumstances. The reading diagnosis system, the RD2P System…
Expert Systems: Implications for the Diagnosis and Treatment of Learning Disabilities.
ERIC Educational Resources Information Center
Hofmeister, Alan M.; Lubke, Margaret M.
1988-01-01
The article examines characteristics and present or potential applications of expert systems technology for diagnosis and treatment of learning disabilities. Preliminary findings indicate that expert systems can perform as well as humans in specific areas, and that the process of organizing knowledge bases for expert systems helps clarify existing…
Design of an Ada expert system shell for the VHSIC avionic modular flight processor
NASA Technical Reports Server (NTRS)
Fanning, F. Jesse
1992-01-01
The Embedded Computer System Expert System Shell (ES Shell) is an Ada-based expert system shell developed at the Avionics Laboratory for use on the VHSIC Avionic Modular Processor (VAMP) running under the Ada Avionics Real-Time Software (AARTS) Operating System. The ES Shell provides the interface between the expert system and the avionics environment, and controls execution of the expert system. Testing of the ES Shell in the Avionics Laboratory's Integrated Test Bed (ITB) has demonstrated its ability to control a non-deterministic software application executing on the VAMP's which can control the ITB's real-time closed-loop aircraft simulation. The results of these tests and the conclusions reached in the design and development of the ES Shell have played an important role in the formulation of the requirements for a production-quality expert system inference engine, an ingredient necessary for the successful use of expert systems on the VAMP embedded avionic flight processor.
The need for a comprehensive expert system development methodology
NASA Technical Reports Server (NTRS)
Baumert, John; Critchfield, Anna; Leavitt, Karen
1988-01-01
In a traditional software development environment, the introduction of standardized approaches has led to higher quality, maintainable products on the technical side and greater visibility into the status of the effort on the management side. This study examined expert system development to determine whether it differed enough from traditional systems to warrant a reevaluation of current software development methodologies. Its purpose was to identify areas of similarity with traditional software development and areas requiring tailoring to the unique needs of expert systems. A second purpose was to determine whether existing expert system development methodologies meet the needs of expert system development, management, and maintenance personnel. The study consisted of a literature search and personal interviews. It was determined that existing methodologies and approaches to developing expert systems are not comprehensive nor are they easily applied, especially to cradle to grave system development. As a result, requirements were derived for an expert system development methodology and an initial annotated outline derived for such a methodology.
Validation of an expert system intended for research in distributed artificial intelligence
NASA Technical Reports Server (NTRS)
Grossner, C.; Lyons, J.; Radhakrishnan, T.
1991-01-01
The expert system discussed in this paper is designed to function as a testbed for research on cooperating expert systems. Cooperating expert systems are members of an organization which dictates the manner in which the expert systems will interact when solving a problem. The Blackbox Expert described in this paper has been constructed using the C Language Integrated Production System (CLIPS), C++, and X windowing environment. CLIPS is embedded in a C++ program which provides objects that are used to maintain the state of the Blackbox puzzle. These objects are accessed by CLIPS rules through user-defined functions calls. The performance of the Blackbox Expert is validated by experimentation. A group of people are asked to solve a set of test cases for the Blackbox puzzle. A metric has been devised which evaluates the 'correctness' of a solution proposed for a test case of Blackbox. Using this metric and the solutions proposed by the humans, each person receives a rating for their ability to solve the Blackbox puzzle. The Blackbox Expert solves the same set of test cases and is assigned a rating for its ability. Then the rating obtained by the Blackbox Expert is compared with the ratings of the people, thus establishing the skill level of our expert system.
NASA Astrophysics Data System (ADS)
Weatherwax Scott, Caroline; Tsareff, Christopher R.
1990-06-01
One of the main goals of process engineering in the semiconductor industry is to improve wafer fabrication productivity and throughput. Engineers must work continuously toward this goal in addition to performing sustaining and development tasks. To accomplish these objectives, managers must make efficient use of engineering resources. One of the tools being used to improve efficiency is the diagnostic expert system. Expert systems are knowledge based computer programs designed to lead the user through the analysis and solution of a problem. Several photolithography diagnostic expert systems have been implemented at the Hughes Technology Center to provide a systematic approach to process problem solving. This systematic approach was achieved by documenting cause and effect analyses for a wide variety of processing problems. This knowledge was organized in the form of IF-THEN rules, a common structure for knowledge representation in expert system technology. These rules form the knowledge base of the expert system which is stored in the computer. The systems also include the problem solving methodology used by the expert when addressing a problem in his area of expertise. Operators now use the expert systems to solve many process problems without engineering assistance. The systems also facilitate the collection of appropriate data to assist engineering in solving unanticipated problems. Currently, several expert systems have been implemented to cover all aspects of the photolithography process. The systems, which have been in use for over a year, include wafer surface preparation (HMDS), photoresist coat and softbake, align and expose on a wafer stepper, and develop inspection. These systems are part of a plan to implement an expert system diagnostic environment throughout the wafer fabrication facility. In this paper, the systems' construction is described, including knowledge acquisition, rule construction, knowledge refinement, testing, and evaluation. The roles played by the process engineering expert and the knowledge engineer are discussed. The features of the systems are shown, particularly the interactive quality of the consultations and the ease of system use.
MOORE: A prototype expert system for diagnosing spacecraft problems
NASA Technical Reports Server (NTRS)
Howlin, Katherine; Weissert, Jerry; Krantz, Kerry
1988-01-01
MOORE is a rule-based, prototype expert system that assists in diagnosing operational Tracking and Data Relay Satellite (TDRS) problems. It is intended to assist spacecraft engineers at the TDRS ground terminal in trouble shooting problems that are not readily solved with routine procedures, and without expert counsel. An additional goal of the prototype system is to develop in-house expert system and knowledge engineering skills. The prototype system diagnoses antenna pointing and earth pointing problems that may occur within the TDRS Attitude Control System (ACS). Plans include expansion to fault isolation of problems in the most critical subsystems of the TDRS spacecraft. Long term benefits are anticipated with use of an expert system during future TDRS programs with increased mission support time, reduced problem solving time, and retained expert knowledge and experience. Phase 2 of the project is intended to provide NASA the necessary expertise and capability to define requirements, evaluate proposals, and monitor the development progress of a highly competent expert system for NASA's Tracking Data Relay Satellite. Phase 2 also envisions addressing two unexplored applications for expert systems, spacecraft integration and tests (I and T) and support to launch activities. The concept, goals, domain, tools, knowledge acquisition, developmental approach, and design of the expert system. It will explain how NASA obtained the knowledge and capability to develop the system in-house without assistance from outside consultants. Future plans will also be presented.
Expert systems applied to spacecraft fire safety
NASA Technical Reports Server (NTRS)
Smith, Richard L.; Kashiwagi, Takashi
1989-01-01
Expert systems are problem-solving programs that combine a knowledge base and a reasoning mechanism to simulate a human expert. The development of an expert system to manage fire safety in spacecraft, in particular the NASA Space Station Freedom, is difficult but clearly advantageous in the long-term. Some needs in low-gravity flammability characteristics, ventilating-flow effects, fire detection, fire extinguishment, and decision models, all necessary to establish the knowledge base for an expert system, are discussed.
DOE Office of Scientific and Technical Information (OSTI.GOV)
MacAllister, D.J.; Day, R.; McCormack, M.D.
This paper gives an overview of a major integrated oil company`s experience with artificial intelligence (AI) over the last 5 years, with an emphasis on expert systems. The authors chronicle the development of an AI group, including details on development tool selection, project selection strategies, potential pitfalls, and descriptions of several completed expert systems. Small expert systems produced by teams of petroleum technology experts and experienced expert system developers that are focused in well-defined technical areas have produced substantial benefits and accelerated petroleum technology transfer.
An expert systems approach to automated fault management in a regenerative life support subsystem
NASA Technical Reports Server (NTRS)
Malin, J. T.; Lance, N., Jr.
1986-01-01
This paper describes FIXER, a prototype expert system for automated fault management in a regenerative life support subsystem typical of Space Station applications. The development project provided an evaluation of the use of expert systems technology to enhance controller functions in space subsystems. The software development approach permitted evaluation of the effectiveness of direct involvement of the expert in design and development. The approach also permitted intensive observation of the knowledge and methods of the expert. This paper describes the development of the prototype expert system and presents results of the evaluation.
NASA Technical Reports Server (NTRS)
Toms, David; Hadden, George D.; Harrington, Jim
1990-01-01
The Maintenance and Diagnostic System (MDS) that is being developed at Honeywell to enhance the Fault Detection Isolation and Recovery system (FDIR) for the Attitude Determination and Control System on Space Station Freedom is described. The MDS demonstrates ways that AI-based techniques can be used to improve the maintainability and safety of the Station by helping to resolve fault anomalies that cannot be fully determined by built-in-test, by providing predictive maintenance capabilities, and by providing expert maintenance assistance. The MDS will address the problems associated with reasoning about dynamic, continuous information versus only about static data, the concerns of porting software based on AI techniques to embedded targets, and the difficulties associated with real-time response. An initial prototype was built of the MDS. The prototype executes on Sun and IBM PS/2 hardware and is implemented in the Common Lisp; further work will evaluate its functionality and develop mechanisms to port the code to Ada.
Dale L. Bartos; Kent B. Downing
1989-01-01
A knowledge acquisition program was written to aid in obtaining knowledge from the experts concerning endemic populations of mountain pine beetle in lodgepole pine forest. An application expert system is then automatically generated by the knowledge acquisition program that contains the codified base of expert knowledge. Data can then be entered into the expert system...
An SSME High Pressure Oxidizer Turbopump diagnostic system using G2 real-time expert system
NASA Technical Reports Server (NTRS)
Guo, Ten-Huei
1991-01-01
An expert system which diagnoses various seal leakage faults in the High Pressure Oxidizer Turbopump of the SSME was developed using G2 real-time expert system. Three major functions of the software were implemented: model-based data generation, real-time expert system reasoning, and real-time input/output communication. This system is proposed as one module of a complete diagnostic system for the SSME. Diagnosis of a fault is defined as the determination of its type, severity, and likelihood. Since fault diagnosis is often accomplished through the use of heuristic human knowledge, an expert system based approach has been adopted as a paradigm to develop this diagnostic system. To implement this approach, a software shell which can be easily programmed to emulate the human decision process, the G2 Real-Time Expert System, was selected. Lessons learned from this implementation are discussed.
An SSME high pressure oxidizer turbopump diagnostic system using G2(TM) real-time expert system
NASA Technical Reports Server (NTRS)
Guo, Ten-Huei
1991-01-01
An expert system which diagnoses various seal leakage faults in the High Pressure Oxidizer Turbopump of the SSME was developed using G2(TM) real-time expert system. Three major functions of the software were implemented: model-based data generation, real-time expert system reasoning, and real-time input/output communication. This system is proposed as one module of a complete diagnostic system for Space Shuttle Main Engine. Diagnosis of a fault is defined as the determination of its type, severity, and likelihood. Since fault diagnosis is often accomplished through the use of heuristic human knowledge, an expert system based approach was adopted as a paradigm to develop this diagnostic system. To implement this approach, a software shell which can be easily programmed to emulate the human decision process, the G2 Real-Time Expert System, was selected. Lessons learned from this implementation are discussed.
Rodríguez-Salvador, Marisela; Rio-Belver, Rosa María; Garechana-Anacabe, Gaizka
2017-01-01
This research proposes an innovative data model to determine the landscape of emerging technologies. It is based on a competitive technology intelligence methodology that incorporates the assessment of scientific publications and patent analysis production, and is further supported by experts' feedback. It enables the definition of the growth rate of scientific and technological output in terms of the top countries, institutions and journals producing knowledge within the field as well as the identification of main areas of research and development by analyzing the International Patent Classification codes including keyword clusterization and co-occurrence of patent assignees and patent codes. This model was applied to the evolving domain of 3D bioprinting. Scientific documents from the Scopus and Web of Science databases, along with patents from 27 authorities and 140 countries, were retrieved. In total, 4782 scientific publications and 706 patents were identified from 2000 to mid-2016. The number of scientific documents published and patents in the last five years showed an annual average growth of 20% and 40%, respectively. Results indicate that the most prolific nations and institutions publishing on 3D bioprinting are the USA and China, including the Massachusetts Institute of Technology (USA), Nanyang Technological University (Singapore) and Tsinghua University (China), respectively. Biomaterials and Biofabrication are the predominant journals. The most prolific patenting countries are China and the USA; while Organovo Holdings Inc. (USA) and Tsinghua University (China) are the institutions leading. International Patent Classification codes reveal that most 3D bioprinting inventions intended for medical purposes apply porous or cellular materials or biologically active materials. Knowledge clusters and expert drivers indicate that there is a research focus on tissue engineering including the fabrication of organs, bioinks and new 3D bioprinting systems. Our model offers a guide to researchers to understand the knowledge production of pioneering technologies, in this case 3D bioprinting.
ERIC Educational Resources Information Center
Balajthy, Ernest
1989-01-01
The article examines decision-making expert systems and discusses their implications for diagnosis and prescription of reading difficulties. A detailed description of how a reading diagnostic expert system might operate to aid classroom teachers is followed by a discussion of advantages and limitations of expert systems for educational use.…
EXSPRT: An Expert Systems Approach to Computer-Based Adaptive Testing.
ERIC Educational Resources Information Center
Frick, Theodore W.; And Others
Expert systems can be used to aid decision making. A computerized adaptive test (CAT) is one kind of expert system, although it is not commonly recognized as such. A new approach, termed EXSPRT, was devised that combines expert systems reasoning and sequential probability ratio test stopping rules. EXSPRT-R uses random selection of test items,…
Expert database system for quality control
NASA Astrophysics Data System (ADS)
Wang, Anne J.; Li, Zhi-Cheng
1993-09-01
There are more competitors today. Markets are not homogeneous they are fragmented into increasingly focused niches requiring greater flexibility in the product mix shorter manufacturing production runs and above allhigher quality. In this paper the author identified a real-time expert system as a way to improve plantwide quality management. The quality control expert database system (QCEDS) by integrating knowledge of experts in operations quality management and computer systems use all information relevant to quality managementfacts as well as rulesto determine if a product meets quality standards. Keywords: expert system quality control data base
Generating performance portable geoscientific simulation code with Firedrake (Invited)
NASA Astrophysics Data System (ADS)
Ham, D. A.; Bercea, G.; Cotter, C. J.; Kelly, P. H.; Loriant, N.; Luporini, F.; McRae, A. T.; Mitchell, L.; Rathgeber, F.
2013-12-01
This presentation will demonstrate how a change in simulation programming paradigm can be exploited to deliver sophisticated simulation capability which is far easier to programme than are conventional models, is capable of exploiting different emerging parallel hardware, and is tailored to the specific needs of geoscientific simulation. Geoscientific simulation represents a grand challenge computational task: many of the largest computers in the world are tasked with this field, and the requirements of resolution and complexity of scientists in this field are far from being sated. However, single thread performance has stalled, even sometimes decreased, over the last decade, and has been replaced by ever more parallel systems: both as conventional multicore CPUs and in the emerging world of accelerators. At the same time, the needs of scientists to couple ever-more complex dynamics and parametrisations into their models makes the model development task vastly more complex. The conventional approach of writing code in low level languages such as Fortran or C/C++ and then hand-coding parallelism for different platforms by adding library calls and directives forces the intermingling of the numerical code with its implementation. This results in an almost impossible set of skill requirements for developers, who must simultaneously be domain science experts, numericists, software engineers and parallelisation specialists. Even more critically, it requires code to be essentially rewritten for each emerging hardware platform. Since new platforms are emerging constantly, and since code owners do not usually control the procurement of the supercomputers on which they must run, this represents an unsustainable development load. The Firedrake system, conversely, offers the developer the opportunity to write PDE discretisations in the high-level mathematical language UFL from the FEniCS project (http://fenicsproject.org). Non-PDE model components, such as parametrisations, can be written as short C kernels operating locally on the underlying mesh, with no explicit parallelism. The executable code is then generated in C, CUDA or OpenCL and executed in parallel on the target architecture. The system also offers features of special relevance to the geosciences. In particular, the large scale separation between the vertical and horizontal directions in many geoscientific processes can be exploited to offer the flexibility of unstructured meshes in the horizontal direction, without the performance penalty usually associated with those methods.
Parallel processing and expert systems
NASA Technical Reports Server (NTRS)
Yan, Jerry C.; Lau, Sonie
1991-01-01
Whether it be monitoring the thermal subsystem of Space Station Freedom, or controlling the navigation of the autonomous rover on Mars, NASA missions in the 90's cannot enjoy an increased level of autonomy without the efficient use of expert systems. Merely increasing the computational speed of uniprocessors may not be able to guarantee that real time demands are met for large expert systems. Speed-up via parallel processing must be pursued alongside the optimization of sequential implementations. Prototypes of parallel expert systems have been built at universities and industrial labs in the U.S. and Japan. The state-of-the-art research in progress related to parallel execution of expert systems was surveyed. The survey is divided into three major sections: (1) multiprocessors for parallel expert systems; (2) parallel languages for symbolic computations; and (3) measurements of parallelism of expert system. Results to date indicate that the parallelism achieved for these systems is small. In order to obtain greater speed-ups, data parallelism and application parallelism must be exploited.
NASA Technical Reports Server (NTRS)
Durkin, John; Schlegelmilch, Richard; Tallo, Donald
1992-01-01
LeRC has recently completed the design of a Ka-band satellite transponder system, as part of the Advanced Communication Technology Satellite (ACTS) System. To enhance the reliability of this satellite, NASA funded the University of Akron to explore the application of an expert system to provide the transponder with an autonomous diagnosis capability. The results of this research was the development of a prototype diagnosis expert system called FIDEX (fault-isolation and diagnosis expert). FIDEX is a frame-based expert system that was developed in the NEXPERT Object development environment by Neuron Data, Inc. It is a MicroSoft Windows version 3.0 application, and was designed to operate on an Intel i80386 based personal computer system.
Victoroff, Michael S.
1985-01-01
The title is a double entendre. The discussion approaches expert systems from two directions: “What ethical hazards are created by expert systems in medicine?” and “Would it be ethical to design an expert system for solving problems in bioethics?” Computers present new ethical problems to society, some of which are unprecedented. These can be categorized under several rubrics. The paper describes a rudimentary scheme for understanding ethical issues raised by computers, in general, and medical expert systems, in particular. It focuses on bioethical implications of AI in medicine; explores norms, assumptions and taboos; and highlights certain ethical pitfalls. Principles are elucidated, for building ethically sound systems. Finally, a proposal is discussed, for the design of an expert system for moral problem solving, and the ethical implications of this notion are analyzed.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Miller, M.; Kempner, L. Jr.; Mueller, W. III
The concept of an Expert System is not new. It has been around since the days of the early computers when scientists had dreams of robot automation to do everything from washing windows to automobile design. This paper discusses an application of an expert system and addresses software development issues and various levels of expert system development form a structural engineering viewpoint. An expert system designed to aid the structural engineer in first order inelastic analysis of latticed steel transmission powers is presented. The utilization of expert systems with large numerical analysis programs is discussed along with the software developmentmore » of such a system.« less
CLEAR: Communications Link Expert Assistance Resource
NASA Technical Reports Server (NTRS)
Hull, Larry G.; Hughes, Peter M.
1987-01-01
Communications Link Expert Assistance Resource (CLEAR) is a real time, fault diagnosis expert system for the Cosmic Background Explorer (COBE) Mission Operations Room (MOR). The CLEAR expert system is an operational prototype which assists the MOR operator/analyst by isolating and diagnosing faults in the spacecraft communication link with the Tracking and Data Relay Satellite (TDRS) during periods of realtime data acquisition. The mission domain, user requirements, hardware configuration, expert system concept, tool selection, development approach, and system design were discussed. Development approach and system implementation are emphasized. Also discussed are system architecture, tool selection, operation, and future plans.
NASA Technical Reports Server (NTRS)
Liberman, Eugene M.; Manner, David B.; Dolce, James L.; Mellor, Pamela A.
1993-01-01
Expert systems are widely used in health monitoring and fault detection applications. One of the key features of an expert system is that it possesses a large body of knowledge about the application for which it was designed. When the user consults this knowledge base, it is essential that the expert system's reasoning process and its conclusions be as concise as possible. If, in addition, an expert system is part of a process monitoring system, the expert system's conclusions must be combined with current events of the process. Under these circumstances, it is difficult for a user to absorb and respond to all the available information. For example, a user can become distracted and confused if two or more unrelated devices in different parts of the system require attention. A human interface designed to integrate expert system diagnoses with process data and to focus the user's attention to the important matters provides a solution to the 'information overload' problem. This paper will discuss a user interface to the power distribution expert system for Space Station Freedom. The importance of features which simplify assessing system status and which minimize navigating through layers of information will be discussed. Design rationale and implementation choices will also be presented.
Automated Docking Screens: A Feasibility Study
2009-01-01
Molecular docking is the most practical approach to leverage protein structure for ligand discovery, but the technique retains important liabilities that make it challenging to deploy on a large scale. We have therefore created an expert system, DOCK Blaster, to investigate the feasibility of full automation. The method requires a PDB code, sometimes with a ligand structure, and from that alone can launch a full screen of large libraries. A critical feature is self-assessment, which estimates the anticipated reliability of the automated screening results using pose fidelity and enrichment. Against common benchmarks, DOCK Blaster recapitulates the crystal ligand pose within 2 Å rmsd 50−60% of the time; inferior to an expert, but respectrable. Half the time the ligand also ranked among the top 5% of 100 physically matched decoys chosen on the fly. Further tests were undertaken culminating in a study of 7755 eligible PDB structures. In 1398 cases, the redocked ligand ranked in the top 5% of 100 property-matched decoys while also posing within 2 Å rmsd, suggesting that unsupervised prospective docking is viable. DOCK Blaster is available at http://blaster.docking.org. PMID:19719084
Automated docking screens: a feasibility study.
Irwin, John J; Shoichet, Brian K; Mysinger, Michael M; Huang, Niu; Colizzi, Francesco; Wassam, Pascal; Cao, Yiqun
2009-09-24
Molecular docking is the most practical approach to leverage protein structure for ligand discovery, but the technique retains important liabilities that make it challenging to deploy on a large scale. We have therefore created an expert system, DOCK Blaster, to investigate the feasibility of full automation. The method requires a PDB code, sometimes with a ligand structure, and from that alone can launch a full screen of large libraries. A critical feature is self-assessment, which estimates the anticipated reliability of the automated screening results using pose fidelity and enrichment. Against common benchmarks, DOCK Blaster recapitulates the crystal ligand pose within 2 A rmsd 50-60% of the time; inferior to an expert, but respectrable. Half the time the ligand also ranked among the top 5% of 100 physically matched decoys chosen on the fly. Further tests were undertaken culminating in a study of 7755 eligible PDB structures. In 1398 cases, the redocked ligand ranked in the top 5% of 100 property-matched decoys while also posing within 2 A rmsd, suggesting that unsupervised prospective docking is viable. DOCK Blaster is available at http://blaster.docking.org .
The (Missed) Potential of the Patient-centered Medical Home for Disparities.
Reibling, Nadine; Rosenthal, Meredith B
2016-01-01
Disparities in health care and health outcomes are a significant problem in the United States. Delivery system reforms such as the patient-centered medical home (PCMH) could have important implications for disparities. To investigate what role disparities play in current PCMH initiatives and how their set-up might impact on disparities. We selected 4 state-based PCMH initiatives (Colorado, Massachusetts, Pennsylvania, and Rhode Island), 1 regional initiative in New Orleans, and 1 multistate initiative. We interviewed 30 key actors in these initiatives and 3 health policy experts on disparities in the context of PCMH. Interview data were coded using the constant comparative method. We find that disparities are not an explicit priority in PCMH initiatives. Nevertheless, many policymakers, providers, and initiative leaders believe that the model has the potential to reduce disparities. However, because of the funding structure of initiatives and the lack of adjustment of quality metrics, health policy experts do not share this optimism and safety-net providers report concerns and frustration. Even though disparities are currently not a priority in the PCMH community, the design of initiatives has important implications for disparities.
Developing a Web-Based Advisory Expert System for Implementing Traffic Calming Strategies
Falamarzi, Amir; Borhan, Muhamad Nazri; Rahmat, Riza Atiq O. K.
2014-01-01
Lack of traffic safety has become a serious issue in residential areas. In this paper, a web-based advisory expert system for the purpose of applying traffic calming strategies on residential streets is described because there currently lacks a structured framework for the implementation of such strategies. Developing an expert system can assist and advise engineers for dealing with traffic safety problems. This expert system is developed to fill the gap between the traffic safety experts and people who seek to employ traffic calming strategies including decision makers, engineers, and students. In order to build the expert system, examining sources related to traffic calming studies as well as interviewing with domain experts have been carried out. The system includes above 150 rules and 200 images for different types of measures. The system has three main functions including classifying traffic calming measures, prioritizing traffic calming strategies, and presenting solutions for different traffic safety problems. Verifying, validating processes, and comparing the system with similar works have shown that the system is consistent and acceptable for practical uses. Finally, some recommendations for improving the system are presented. PMID:25276861
Developing a web-based advisory expert system for implementing traffic calming strategies.
Falamarzi, Amir; Borhan, Muhamad Nazri; Rahmat, Riza Atiq O K
2014-01-01
Lack of traffic safety has become a serious issue in residential areas. In this paper, a web-based advisory expert system for the purpose of applying traffic calming strategies on residential streets is described because there currently lacks a structured framework for the implementation of such strategies. Developing an expert system can assist and advise engineers for dealing with traffic safety problems. This expert system is developed to fill the gap between the traffic safety experts and people who seek to employ traffic calming strategies including decision makers, engineers, and students. In order to build the expert system, examining sources related to traffic calming studies as well as interviewing with domain experts have been carried out. The system includes above 150 rules and 200 images for different types of measures. The system has three main functions including classifying traffic calming measures, prioritizing traffic calming strategies, and presenting solutions for different traffic safety problems. Verifying, validating processes, and comparing the system with similar works have shown that the system is consistent and acceptable for practical uses. Finally, some recommendations for improving the system are presented.
Executive system software design and expert system implementation
NASA Technical Reports Server (NTRS)
Allen, Cheryl L.
1992-01-01
The topics are presented in viewgraph form and include: software requirements; design layout of the automated assembly system; menu display for automated composite command; expert system features; complete robot arm state diagram and logic; and expert system benefits.
Kubota, Yoshie; Yano, Yoshitaka; Seki, Susumu; Takada, Kaori; Sakuma, Mio; Morimoto, Takeshi; Akaike, Akinori; Hiraide, Atsushi
2011-04-11
To determine the value of using the Roter Interaction Analysis System during objective structured clinical examinations (OSCEs) to assess pharmacy students' communication competence. As pharmacy students completed a clinical OSCE involving an interview with a simulated patient, 3 experts used a global rating scale to assess students' overall performance in the interview, and both the student's and patient's languages were coded using the Roter Interaction Analysis System (RIAS). The coders recorded the number of utterances (ie, units of spoken language) in each RIAS category. Correlations between the raters' scores and the number and types of utterances were examined. There was a significant correlation between students' global rating scores on the OSCE and the number of utterances in the RIAS socio-emotional category but not the RIAS business category. The RIAS proved to be a useful tool for assessing the socio-emotional aspect of students' interview skills.
Expert Systems in Reference Services.
ERIC Educational Resources Information Center
Roysdon, Christine, Ed.; White, Howard D., Ed.
1989-01-01
Eleven articles introduce expert systems applications in library and information science, and present design and implementation issues of system development for reference services. Topics covered include knowledge based systems, prototype development, the use of artificial intelligence to remedy current system inadequacies, and an expert system to…
1993-08-01
disclaimers should be a top priority. Contract law involves the Uniform Commercial Code (UCC). This is an agreement between all the states (except...to contract law than this, the basic issue with software is that the sup- plier is generally an expert on an arcane and sophisticated technology and
DOT National Transportation Integrated Search
1990-01-01
The validation and evaluation of an expert system for traffic control in highway work zones (TRANZ) is described. The stages in the evaluation process consisted of the following: revisit the experts, selectively distribute copies of TRANZ with docume...
Phylotastic! Making tree-of-life knowledge accessible, reusable and convenient.
Stoltzfus, Arlin; Lapp, Hilmar; Matasci, Naim; Deus, Helena; Sidlauskas, Brian; Zmasek, Christian M; Vaidya, Gaurav; Pontelli, Enrico; Cranston, Karen; Vos, Rutger; Webb, Campbell O; Harmon, Luke J; Pirrung, Megan; O'Meara, Brian; Pennell, Matthew W; Mirarab, Siavash; Rosenberg, Michael S; Balhoff, James P; Bik, Holly M; Heath, Tracy A; Midford, Peter E; Brown, Joseph W; McTavish, Emily Jane; Sukumaran, Jeet; Westneat, Mark; Alfaro, Michael E; Steele, Aaron; Jordan, Greg
2013-05-13
Scientists rarely reuse expert knowledge of phylogeny, in spite of years of effort to assemble a great "Tree of Life" (ToL). A notable exception involves the use of Phylomatic, which provides tools to generate custom phylogenies from a large, pre-computed, expert phylogeny of plant taxa. This suggests great potential for a more generalized system that, starting with a query consisting of a list of any known species, would rectify non-standard names, identify expert phylogenies containing the implicated taxa, prune away unneeded parts, and supply branch lengths and annotations, resulting in a custom phylogeny suited to the user's needs. Such a system could become a sustainable community resource if implemented as a distributed system of loosely coupled parts that interact through clearly defined interfaces. With the aim of building such a "phylotastic" system, the NESCent Hackathons, Interoperability, Phylogenies (HIP) working group recruited 2 dozen scientist-programmers to a weeklong programming hackathon in June 2012. During the hackathon (and a three-month follow-up period), 5 teams produced designs, implementations, documentation, presentations, and tests including: (1) a generalized scheme for integrating components; (2) proof-of-concept pruners and controllers; (3) a meta-API for taxonomic name resolution services; (4) a system for storing, finding, and retrieving phylogenies using semantic web technologies for data exchange, storage, and querying; (5) an innovative new service, DateLife.org, which synthesizes pre-computed, time-calibrated phylogenies to assign ages to nodes; and (6) demonstration projects. These outcomes are accessible via a public code repository (GitHub.com), a website (http://www.phylotastic.org), and a server image. Approximately 9 person-months of effort (centered on a software development hackathon) resulted in the design and implementation of proof-of-concept software for 4 core phylotastic components, 3 controllers, and 3 end-user demonstration tools. While these products have substantial limitations, they suggest considerable potential for a distributed system that makes phylogenetic knowledge readily accessible in computable form. Widespread use of phylotastic systems will create an electronic marketplace for sharing phylogenetic knowledge that will spur innovation in other areas of the ToL enterprise, such as annotation of sources and methods and third-party methods of quality assessment.
Preliminary PANSAT ground station software design and use of an expert system to analyze telemetry
NASA Astrophysics Data System (ADS)
Lawrence, Gregory W.
1994-03-01
The Petite Amateur Navy Satellite (PANSAT) is a communications satellite designed to be used by civilian amateur radio operators. A master ground station is being built at the Naval Postgraduate School. This computer system performs satellite commands, displays telemetry, trouble-shoots problems, and passes messages. The system also controls an open loop tracking antenna. This paper concentrates on the telemetry display, decoding, and interpretation through artificial intelligence (AI). The telemetry is displayed in an easily interpretable format, so that any user can understand the current health of the satellite and be cued as to any problems and possible solutions. Only the master ground station has the ability to receive all telemetry and send commands to the spacecraft; civilian ham users do not have access to this information. The telemetry data is decommutated and analyzed before it is displayed to the user, so that the raw data will not have to be interpreted by ground users. The analysis will use CLIPS imbedded in the code, and derive its inputs from telemetry decommutation. The program is an expert system using a forward chaining set of rules based on the expected operation and parameters of the satellite. By building the rules during the construction and design of the satellite, the telemetry can be well understood and interpreted after the satellite is launched and the designers may no longer be available to provide input to the problem.
Development and practice of a Telehealthcare Expert System (TES).
Lin, Hanjun; Hsu, Yeh-Liang; Hsu, Ming-Shinn; Cheng, Chih-Ming
2013-07-01
Expert systems have been widely used in medical and healthcare practice for various purposes. In addition to vital sign data, important concerns in telehealthcare include the compliance with the measurement prescription, the accuracy of vital sign measurements, and the functioning of vital sign meters and home gateways. However, few expert system applications are found in the telehealthcare domain to address these issues. This article presents an expert system application for one of the largest commercialized telehealthcare practices in Taiwan by Min-Sheng General Hospital. The main function of the Telehealthcare Expert System (TES) developed in this research is to detect and classify events based on the measurement data transmitted to the database at the call center, including abnormality of vital signs, violation of vital sign measurement prescriptions, and malfunction of hardware devices (home gateway and vital sign meter). When the expert system detects an abnormal event, it assigns an "urgent degree" and alerts the nursing team in the call center to take action, such as phoning the patient for counseling or to urge the patient to return to the hospital for further tests. During 2 years of clinical practice, from 2009 to 2011, 19,182 patients were served by the expert system. The expert system detected 41,755 events, of which 22.9% indicated abnormality of vital signs, 75.2% indicated violation of measurement prescription, and 1.9% indicated malfunction of devices. On average, the expert system reduced by 76.5% the time that the nursing team in the call center spent in handling the events. The expert system helped to reduce cost and improve quality of the telehealthcare service.
An SQL query generator for CLIPS
NASA Technical Reports Server (NTRS)
Snyder, James; Chirica, Laurian
1990-01-01
As expert systems become more widely used, their access to large amounts of external information becomes increasingly important. This information exists in several forms such as statistical, tabular data, knowledge gained by experts and large databases of information maintained by companies. Because many expert systems, including CLIPS, do not provide access to this external information, much of the usefulness of expert systems is left untapped. The scope of this paper is to describe a database extension for the CLIPS expert system shell. The current industry standard database language is SQL. Due to SQL standardization, large amounts of information stored on various computers, potentially at different locations, will be more easily accessible. Expert systems should be able to directly access these existing databases rather than requiring information to be re-entered into the expert system environment. The ORACLE relational database management system (RDBMS) was used to provide a database connection within the CLIPS environment. To facilitate relational database access a query generation system was developed as a CLIPS user function. The queries are entered in a CLlPS-like syntax and are passed to the query generator, which constructs and submits for execution, an SQL query to the ORACLE RDBMS. The query results are asserted as CLIPS facts. The query generator was developed primarily for use within the ICADS project (Intelligent Computer Aided Design System) currently being developed by the CAD Research Unit in the California Polytechnic State University (Cal Poly). In ICADS, there are several parallel or distributed expert systems accessing a common knowledge base of facts. Expert system has a narrow domain of interest and therefore needs only certain portions of the information. The query generator provides a common method of accessing this information and allows the expert system to specify what data is needed without specifying how to retrieve it.
Cataloging Expert Systems: Optimism and Frustrated Reality.
ERIC Educational Resources Information Center
Olmstadt, William J.
2000-01-01
Discusses artificial intelligence and attempts to catalog expert systems. Topics include the nature of expertise; examples of cataloging expert systems; barriers to implementation; and problems, including total automation, cataloging expertise, priorities, and system design. (LRW)
Development of an instructional expert system for hole drilling processes
NASA Technical Reports Server (NTRS)
Al-Mutawa, Souhaila; Srinivas, Vijay; Moon, Young Bai
1990-01-01
An expert system which captures the expertise of workshop technicians in the drilling domain was developed. The expert system is aimed at novice technicians who know how to operate the machines but have not acquired the decision making skills that are gained with experience. This paper describes the domain background and the stages of development of the expert system.
Rushton, A; White, L; Heap, A; Calvert, M; Heneghan, N; Goodwin, P
2016-02-25
To develop an optimised 1:1 physiotherapy intervention that reflects best practice, with flexibility to tailor management to individual patients, thereby ensuring patient-centred practice. Mixed-methods combining evidence synthesis, expert review and focus groups. Secondary care involving 5 UK specialist spinal centres. A purposive panel of clinical experts from the 5 spinal centres, comprising spinal surgeons, inpatient and outpatient physiotherapists, provided expert review of the draft intervention. Purposive samples of patients (n=10) and physiotherapists (n=10) (inpatient/outpatient physiotherapists managing patients with lumbar discectomy) were invited to participate in the focus groups at 1 spinal centre. A draft intervention developed from 2 systematic reviews; a survey of current practice and research related to stratified care was circulated to the panel of clinical experts. Lead physiotherapists collaborated with physiotherapy and surgeon colleagues to provide feedback that informed the intervention presented at 2 focus groups investigating acceptability to patients and physiotherapists. The focus groups were facilitated by an experienced facilitator, recorded in written and tape-recorded forms by an observer. Tape recordings were transcribed verbatim. Data analysis, conducted by 2 independent researchers, employed an iterative and constant comparative process of (1) initial descriptive coding to identify categories and subsequent themes, and (2) deeper, interpretive coding and thematic analysis enabling concepts to emerge and overarching pattern codes to be identified. The intervention reflected best available evidence and provided flexibility to ensure patient-centred care. The intervention comprised up to 8 sessions of 1:1 physiotherapy over 8 weeks, starting 4 weeks postsurgery. The intervention was acceptable to patients and physiotherapists. A rigorous process informed an optimised 1:1 physiotherapy intervention post-lumbar discectomy that reflects best practice. The developed intervention was agreed on by the 5 spinal centres for implementation in a randomised controlled trial to evaluate its effectiveness. Published by the BMJ Publishing Group Limited. For permission to use (where not already granted under a licence) please go to http://www.bmj.com/company/products-services/rights-and-licensing/
DOE Office of Scientific and Technical Information (OSTI.GOV)
Earl, Christopher; Might, Matthew; Bagusetty, Abhishek
This study presents Nebo, a declarative domain-specific language embedded in C++ for discretizing partial differential equations for transport phenomena on multiple architectures. Application programmers use Nebo to write code that appears sequential but can be run in parallel, without editing the code. Currently Nebo supports single-thread execution, multi-thread execution, and many-core (GPU-based) execution. With single-thread execution, Nebo performs on par with code written by domain experts. With multi-thread execution, Nebo can linearly scale (with roughly 90% efficiency) up to 12 cores, compared to its single-thread execution. Moreover, Nebo’s many-core execution can be over 140x faster than its single-thread execution.
Earl, Christopher; Might, Matthew; Bagusetty, Abhishek; ...
2016-01-26
This study presents Nebo, a declarative domain-specific language embedded in C++ for discretizing partial differential equations for transport phenomena on multiple architectures. Application programmers use Nebo to write code that appears sequential but can be run in parallel, without editing the code. Currently Nebo supports single-thread execution, multi-thread execution, and many-core (GPU-based) execution. With single-thread execution, Nebo performs on par with code written by domain experts. With multi-thread execution, Nebo can linearly scale (with roughly 90% efficiency) up to 12 cores, compared to its single-thread execution. Moreover, Nebo’s many-core execution can be over 140x faster than its single-thread execution.
A parallel expert system for the control of a robotic air vehicle
NASA Technical Reports Server (NTRS)
Shakley, Donald; Lamont, Gary B.
1988-01-01
Expert systems can be used to govern the intelligent control of vehicles, for example the Robotic Air Vehicle (RAV). Due to the nature of the RAV system the associated expert system needs to perform in a demanding real-time environment. The use of a parallel processing capability to support the associated expert system's computational requirement is critical in this application. Thus, algorithms for parallel real-time expert systems must be designed, analyzed, and synthesized. The design process incorporates a consideration of the rule-set/face-set size along with representation issues. These issues are looked at in reference to information movement and various inference mechanisms. Also examined is the process involved with transporting the RAV expert system functions from the TI Explorer, where they are implemented in the Automated Reasoning Tool (ART), to the iPSC Hypercube, where the system is synthesized using Concurrent Common LISP (CCLISP). The transformation process for the ART to CCLISP conversion is described. The performance characteristics of the parallel implementation of these expert systems on the iPSC Hypercube are compared to the TI Explorer implementation.
NASA Astrophysics Data System (ADS)
Sakaguchi, Hideharu
Do you remember an expert system? I think there are various impressions about the system. For example, some might say “It reminds me of old days”. On the other hand, some might say “It was really troublesome”. About 25 years ago, from late 1980s to the middle of 1990s, when the Showa era was about to change into the Heisei Era, artificial intelligence boomed. Research and development for an expert system which was equipped with expertise and worked as smart as expert, was advanced in various fields. Our company also picked up the system as the new system which covered weak point of conventional computer technology. We started research and development in 1984, and installed an expert system in a SCADA system, which started operating in March 1990 in the Fukuoka Integrated Control Center. In this essay, as an electric power engineer who involved in development at that time, I introduce the situation and travail story about developing an expert system which support restorative actions from the outage and overload condition of power networks.
Cooperating Expert Systems For Space Station Power Distribution Management
NASA Astrophysics Data System (ADS)
Nguyen, T. A.; Chiou, W. C.
1987-02-01
In a complex system such as the manned Space Station, it is deem necessary that many expert systems must perform tasks in a concurrent and cooperative manner. An important question arise is: what cooperative-task-performing models are appropriate for multiple expert systems to jointly perform tasks. The solution to this question will provide a crucial automation design criteria for the Space Station complex systems architecture. Based on a client/server model for performing tasks, we have developed a system that acts as a front-end to support loosely-coupled communications between expert systems running on multiple Symbolics machines. As an example, we use two ART*-based expert systems to demonstrate the concept of parallel symbolic manipulation for power distribution management and dynamic load planner/scheduler in the simulated Space Station environment. This on-going work will also explore other cooperative-task-performing models as alternatives which can evaluate inter and intra expert system communication mechanisms. It will be served as a testbed and a bench-marking tool for other Space Station expert subsystem communication and information exchange.
A software engineering approach to expert system design and verification
NASA Technical Reports Server (NTRS)
Bochsler, Daniel C.; Goodwin, Mary Ann
1988-01-01
Software engineering design and verification methods for developing expert systems are not yet well defined. Integration of expert system technology into software production environments will require effective software engineering methodologies to support the entire life cycle of expert systems. The software engineering methods used to design and verify an expert system, RENEX, is discussed. RENEX demonstrates autonomous rendezvous and proximity operations, including replanning trajectory events and subsystem fault detection, onboard a space vehicle during flight. The RENEX designers utilized a number of software engineering methodologies to deal with the complex problems inherent in this system. An overview is presented of the methods utilized. Details of the verification process receive special emphasis. The benefits and weaknesses of the methods for supporting the development life cycle of expert systems are evaluated, and recommendations are made based on the overall experiences with the methods.
NASA Technical Reports Server (NTRS)
Liebowitz, J.
1986-01-01
The development of an expert system prototype for software functional requirement determination for NASA Goddard's Command Management System, as part of its process of transforming general requests into specific near-earth satellite commands, is described. The present knowledge base was formulated through interactions with domain experts, and was then linked to the existing Knowledge Engineering Systems (KES) expert system application generator. Steps in the knowledge-base development include problem-oriented attribute hierarchy development, knowledge management approach determination, and knowledge base encoding. The KES Parser and Inspector, in addition to backcasting and analogical mapping, were used to validate the expert system-derived requirements for one of the major functions of a spacecraft, the solar Maximum Mission. Knowledge refinement, evaluation, and implementation procedures of the expert system were then accomplished.
Third CLIPS Conference Proceedings, volume 1
NASA Technical Reports Server (NTRS)
Riley, Gary (Editor)
1994-01-01
Expert systems are computed programs which emulate human expertise in well defined problem domains. The potential payoff from expert systems is high: valuable expertise can be captured and preserved, repetitive and/or mundane tasks requiring human expertise can be automated, and uniformity can be applied in decision making processes. The C Language Integrated Production Systems (CLIPS) is an expert system building tool, developed at the Johnson Space Center, which provides a complete environment for the development and delivery of rule and/or object based expert systems. CLIPS was specifically designed to provide a low cost option for developing and deploying expert system applications across a wide range of hardware platforms. The development of CLIPS has helped to improve the ability to deliver expert systems technology throughout the public and private sectors for a wide range of applications and diverse computing environments.
System of experts for intelligent data management (SEIDAM)
NASA Technical Reports Server (NTRS)
Goodenough, David G.; Iisaka, Joji; Fung, KO
1993-01-01
A proposal to conduct research and development on a system of expert systems for intelligent data management (SEIDAM) is being developed. CCRS has much expertise in developing systems for integrating geographic information with space and aircraft remote sensing data and in managing large archives of remotely sensed data. SEIDAM will be composed of expert systems grouped in three levels. At the lowest level, the expert systems will manage and integrate data from diverse sources, taking account of symbolic representation differences and varying accuracies. Existing software can be controlled by these expert systems, without rewriting existing software into an Artificial Intelligence (AI) language. At the second level, SEIDAM will take the interpreted data (symbolic and numerical) and combine these with data models. at the top level, SEIDAM will respond to user goals for predictive outcomes given existing data. The SEIDAM Project will address the research areas of expert systems, data management, storage and retrieval, and user access and interfaces.
System of Experts for Intelligent Data Management (SEIDAM)
NASA Technical Reports Server (NTRS)
Goodenough, David G.; Iisaka, Joji; Fung, KO
1992-01-01
It is proposed to conduct research and development on a system of expert systems for intelligent data management (SEIDAM). CCRS has much expertise in developing systems for integrating geographic information with space and aircraft remote sensing data and in managing large archives of remotely sensed data. SEIDAM will be composed of expert systems grouped in three levels. At the lowest level, the expert systems will manage and integrate data from diverse sources, taking account of symbolic representation differences and varying accuracies. Existing software can be controlled by these expert systems, without rewriting existing software into an Artificial Intelligence (AI) language. At the second level, SEIDAM will take the interpreted data (symbolic and numerical) and combine these with data models. At the top level, SEIDAM will respond to user goals for predictive outcomes given existing data. The SEIDAM Project will address the research areas of expert systems, data management, storage and retrieval, and user access and interfaces.
TES: A modular systems approach to expert system development for real-time space applications
NASA Technical Reports Server (NTRS)
Cacace, Ralph; England, Brenda
1988-01-01
A major goal of the Space Station era is to reduce reliance on support from ground based experts. The development of software programs using expert systems technology is one means of reaching this goal without requiring crew members to become intimately familiar with the many complex spacecraft subsystems. Development of an expert systems program requires a validation of the software with actual flight hardware. By combining accurate hardware and software modelling techniques with a modular systems approach to expert systems development, the validation of these software programs can be successfully completed with minimum risk and effort. The TIMES Expert System (TES) is an application that monitors and evaluates real time data to perform fault detection and fault isolation tasks as they would otherwise be carried out by a knowledgeable designer. The development process and primary features of TES, a modular systems approach, and the lessons learned are discussed.
Accuracy of external cause-of-injury coding in VA polytrauma patient discharge records.
Carlson, Kathleen F; Nugent, Sean M; Grill, Joseph; Sayer, Nina A
2010-01-01
Valid and efficient methods of identifying the etiology of treated injuries are critical for characterizing patient populations and developing prevention and rehabilitation strategies. We examined the accuracy of external cause-of-injury codes (E-codes) in Veterans Health Administration (VHA) administrative data for a population of injured patients. Chart notes and E-codes were extracted for 566 patients treated at any one of four VHA Polytrauma Rehabilitation Center sites between 2001 and 2006. Two expert coders, blinded to VHA E-codes, used chart notes to assign "gold standard" E-codes to injured patients. The accuracy of VHA E-coding was examined based on these gold standard E-codes. Only 382 of 517 (74%) injured patients were assigned E-codes in VHA records. Sensitivity of VHA E-codes varied significantly by site (range: 59%-91%, p < 0.001). Sensitivity was highest for combat-related injuries (81%) and lowest for fall-related injuries (60%). Overall specificity of E-codes was high (92%). E-coding accuracy was markedly higher when we restricted analyses to records that had been assigned VHA E-codes. E-codes may not be valid for ascertaining source-of-injury data for all injuries among VHA rehabilitation inpatients at this time. Enhanced training and policies may ensure more widespread, standardized use and accuracy of E-codes for injured veterans treated in the VHA.
A secured e-tendering modeling using misuse case approach
NASA Astrophysics Data System (ADS)
Mohd, Haslina; Robie, Muhammad Afdhal Muhammad; Baharom, Fauziah; Darus, Norida Muhd; Saip, Mohamed Ali; Yasin, Azman
2016-08-01
Major risk factors relating to electronic transactions may lead to destructive impacts on trust and transparency in the process of tendering. Currently, electronic tendering (e-tendering) systems still remain uncertain in issues relating to legal and security compliance and most importantly it has an unclear security framework. Particularly, the available systems are lacking in addressing integrity, confidentiality, authentication, and non-repudiation in e-tendering requirements. Thus, one of the challenges in developing an e-tendering system is to ensure the system requirements include the function for secured and trusted environment. Therefore, this paper aims to model a secured e-tendering system using misuse case approach. The modeling process begins with identifying the e-tendering process, which is based on the Australian Standard Code of Tendering (AS 4120-1994). It is followed by identifying security threats and their countermeasure. Then, the e-tendering was modelled using misuse case approach. The model can contribute to e-tendering developers and also to other researchers or experts in the e-tendering domain.
The advanced software development workstation project
NASA Technical Reports Server (NTRS)
Fridge, Ernest M., III; Pitman, Charles L.
1991-01-01
The Advanced Software Development Workstation (ASDW) task is researching and developing the technologies required to support Computer Aided Software Engineering (CASE) with the emphasis on those advanced methods, tools, and processes that will be of benefit to support all NASA programs. Immediate goals are to provide research and prototype tools that will increase productivity, in the near term, in projects such as the Software Support Environment (SSE), the Space Station Control Center (SSCC), and the Flight Analysis and Design System (FADS) which will be used to support the Space Shuttle and Space Station Freedom. Goals also include providing technology for development, evolution, maintenance, and operations. The technologies under research and development in the ASDW project are targeted to provide productivity enhancements during the software life cycle phase of enterprise and information system modeling, requirements generation and analysis, system design and coding, and system use and maintenance. On-line user's guides will assist users in operating the developed information system with knowledge base expert assistance.
Evaluating pharmacy leader development through the seven action logics.
Philip, Achsah; Desai, Avani; Nguyen, Phouc Anne; Birney, Patrick; Colavecchia, Anthony; Karralli, Rusol; Smith, Lindsey; Lorimer, Dirk; Burgess, Gwen; Munch, Kyle; Daniel, Nelvin; Lionetti, Jason; Garey, Kevin W
2016-01-15
Pharmacy leader development over time was analyzed using the seven action logics. As part of an ongoing leadership seminar series, students were required to select a visionary pharmacy leader and conduct a structured interview to evaluate pharmacy leaders' action logics. A standardized questionnaire comprising 13 questions was created by the class. Questions addressed leadership qualities during the leaders' early years, education years, and work years. Transcripts were then coded by two separate trained investigators based on the leader's stage of life to provide a score for each action logic individually over time. Kappa coefficient was used to evaluate interrater agreement. A total of 14 leaders were interviewed. All leaders were currently employed and had won national awards for their contributions to pharmacy practice. Overall, there was 82% agreement between the two evaluators' scores for the various characteristics. Action logics changed based on the leaders' life stage. Using aggregate data from all leader interviews, a progression from lower-order action logics (opportunist, diplomat, expert) to higher-order action logics (strategist, alchemist) was found. Ten leaders (71%) were diplomats during their early years. Six leaders (43%) were experts during their education years, and 4 (29%) were strategists or alchemists. During the third life stage analyzed (the work years), 6 leaders (43%) were strategists, and 2 were alchemists. During their work years, all leaders had a percentage of their answers coded as alchemist (range, 5-22%). Throughout their professional careers, pharmacy leaders continually develop skills through formal education and mentorship that follow action logics. Copyright © 2016 by the American Society of Health-System Pharmacists, Inc. All rights reserved.
Autonomously acquiring declarative and procedural knowledge for ICAT systems
NASA Technical Reports Server (NTRS)
Kovarik, Vincent J., Jr.
1993-01-01
The construction of Intelligent Computer Aided Training (ICAT) systems is critically dependent on the ability to define and encode knowledge. This knowledge engineering effort can be broadly divided into two categories: domain knowledge and expert or task knowledge. Domain knowledge refers to the physical environment or system with which the expert interacts. Expert knowledge consists of the set of procedures and heuristics employed by the expert in performing their task. Both these areas are a significant bottleneck in the acquisition of knowledge for ICAT systems. This paper presents a research project in the area of autonomous knowledge acquisition using a passive observation concept. The system observes an expert and then generalizes the observations into production rules representing the domain expert's knowledge.
NASA Technical Reports Server (NTRS)
Hruska, S. I.; Dalke, A.; Ferguson, J. J.; Lacher, R. C.
1991-01-01
Rule-based expert systems may be structurally and functionally mapped onto a special class of neural networks called expert networks. This mapping lends itself to adaptation of connectionist learning strategies for the expert networks. A parsing algorithm to translate C Language Integrated Production System (CLIPS) rules into a network of interconnected assertion and operation nodes has been developed. The translation of CLIPS rules to an expert network and back again is illustrated. Measures of uncertainty similar to those rules in MYCIN-like systems are introduced into the CLIPS system and techniques for combining and hiring nodes in the network based on rule-firing with these certainty factors in the expert system are presented. Several learning algorithms are under study which automate the process of attaching certainty factors to rules.
RAMBOT: A Connectionist Expert System That Learns by Example.
ERIC Educational Resources Information Center
Mozer, Michael C.
One solution to the problem of getting expert knowledge into expert systems would be to endow the systems with powerful learning procedures that could discover appropriate behaviors by observing an expert in action. A promising source of such learning procedures can be found in recent work on connectionist networks, which are massively parallel…
Tools and technologies for expert systems: A human factors perspective
NASA Technical Reports Server (NTRS)
Rajaram, Navaratna S.
1987-01-01
It is widely recognized that technologies based on artificial intelligence (AI), especially expert systems, can make significant contributions to the productivity and effectiveness of operations of information and knowledge intensive organizations such as NASA. At the same time, these being relatively new technologies, there is the problem of transfering technology to key personnel of such organizations. The problems of examining the potential of expert systems and of technology transfer is addressed in the context of human factors applications. One of the topics of interest was the investigation of the potential use of expert system building tools, particularly NEXPERT as a technology transfer medium. Two basic conclusions were reached in this regard. First, NEXPERT is an excellent tool for rapid prototyping of experimental expert systems, but not ideal as a delivery vehicle. Therefore, it is not a substitute for general purpose system implementation languages such a LISP or C. This assertion probably holds for nearly all such tools on the market today. Second, an effective technology transfer mechanism is to formulate and implement expert systems for problems which members of the organization in question can relate to. For this purpose, the LIghting EnGineering Expert (LIEGE) was implemented using NEXPERT as the tool for technology transfer and to illustrate the value of expert systems to the activities of the Man-System Division.
1991-09-01
Distribution system ... ......... 4 2. Architechture of an Expert system .. .............. 66 vi List of Tables Table Page 1. Prototype Component Model...expert system to properly process work requests Ln civil engineering (8:23). Electric Power Research Institute (EPRI). EPRI is a private organization ...used (51) Training Level. The level of training shop technicians receive, and the resulting proficiency, are important in all organizations . Experts 1
NASA Technical Reports Server (NTRS)
Kawamura, K.; Beale, G. O.; Schaffer, J. D.; Hsieh, B. J.; Padalkar, S.; Rodriguez-Moscoso, J. J.
1985-01-01
A reference manual is provided for NESS, a simulation expert system. This manual gives user information regarding starting and operating NASA expert simulation system (NESS). This expert system provides an intelligent interface to a generic simulation program for spacecraft attitude control problems. A menu of the functions the system can perform is provided. Control repeated returns to this menu after executing each user request.
Development of a coupled expert system for the spacecraft attitude control problem
NASA Technical Reports Server (NTRS)
Kawamura, K.; Beale, G.; Schaffer, J.; Hsieh, B.-J.; Padalkar, S.; Rodriguezmoscoso, J.; Vinz, F.; Fernandez, K.
1987-01-01
A majority of the current expert systems focus on the symbolic-oriented logic and inference mechanisms of artificial intelligence (AI). Common rule-based systems employ empirical associations and are not well suited to deal with problems often arising in engineering. Described is a prototype expert system which combines both symbolic and numeric computing. The expert system's configuration is presented and its application to a spacecraft attitude control problem is discussed.
Barriers Against Implementing Blunt Abdominal Trauma Guidelines in a Hospital: A Qualitative Study
Zaboli, Rouhollah; Tofighi, Shahram; Aghighi, Ali; Shokouh, Seyyed Javad Hosaini; Naraghi, Nader; Goodarzi, Hassan
2016-01-01
Introduction Clinical practice guidelines are structured recommendations that help physicians and patients to make proper decisions when dealing with a specific clinical condition. Because blunt abdominal trauma causes a various range of mild, single-system, and multisystem injuries, early detection will help to reduce mortality and resulting disability. Emergency treatment should be initiated based on CPGs. This study aimed to determine the variables affecting implementing blunt abdominal trauma CPGs in an Iranian hospital. Methods This study was conducted as a qualitative and phenomenology study in the Family Hospital in Tehran (Iran) in 2015. The research population included eight experts and key people in the area of blunt abdominal trauma clinical practice guidelines. Sampling was based on purposive and nonrandom methods. A semistructured interview was done for the data collection. A framework method was applied for the data analysis by using Atlas.ti software. Results After framework analyzing and various reviewing and deleting and combining the codes from 251 codes obtained, 15 families and five super families were extracted, including technical knowledge barriers, economical barriers, barriers related to deployment and monitoring, political will barriers, and managing barriers. Conclusion Structural reform is needed for eliminating the defects available in the healthcare system. As with most of the codes, subconcepts and concepts are classified into the field of human resources; it seems that the education and knowledge will be more important than other resources such as capital and equipment. PMID:27757191
An Expert System for Diagnosing Eye Diseases using Forward Chaining Method
NASA Astrophysics Data System (ADS)
Munaiseche, C. P. C.; Kaparang, D. R.; Rompas, P. T. D.
2018-02-01
Expert System is a system that seeks to adopt human knowledge to the computer, so that the computer can solve problems which are usually done by experts. The purpose of medical expert system is to support the diagnosis process of physicians. It considers facts and symptoms to provide diagnosis. This implies that a medical expert system uses knowledge about diseases and facts about the patients to suggest diagnosis. The aim of this research is to design an expert system application for diagnosing eye diseases using forward chaining method and to figure out user acceptance to this application through usability testing. Eye is selected because it is one of the five senses which is very sensitive and important. The scope of the work is extended to 16 types of eye diseases with 41 symptoms of the disease, arranged in 16 rules. The computer programming language employed was the PHP programming language and MySQL as the Relational Database Management System (RDBMS). The results obtained showed that the expert system was able to successfully diagnose eye diseases corresponding to the selected symptoms entered as query and the system evaluation through usability testing showed the expert system for diagnosis eye diseases had very good rate of usability, which includes learnability, efficiency, memorability, errors, and satisfaction so that the system can be received in the operational environment.
Expert Witness: A system for developing expert medical testimony
NASA Technical Reports Server (NTRS)
Lewandowski, Raymond; Perkins, David; Leasure, David
1994-01-01
Expert Witness in an expert system designed to assist attorneys and medical experts in determining the merit of medical malpractice claims in the area of obstetrics. It substitutes the time of the medical expert with the time of a paralegal assistant guided by the expert system during the initial investigation of the medical records and patient interviews. The product of the system is a narrative transcript containing important data, immediate conclusions from the data, and overall conclusions of the case that the attorney and medical expert use to make decisions about whether and how to proceed with the case. The transcript may also contain directives for gathering additional information needed for the case. The system is a modified heuristic classifier and is implemented using over 600 CLIPS rules together with a C-based user interface. The data abstraction and solution refinement are implemented directly using forward chaining production and matching. The use of CLIPS and C is essential to delivering a system that runs on a generic PC platform. The direct implementation in CLIPS together with locality of inference ensures that the system will scale gracefully. Two years of use has revealed no errors in the reasoning.
[Development of expert diagnostic system for common respiratory diseases].
Xu, Wei-hua; Chen, You-ling; Yan, Zheng
2014-03-01
To develop an internet-based expert diagnostic system for common respiratory diseases. SaaS system was used to build architecture; pattern of forward reasoning was applied for inference engine design; ASP.NET with C# from the tool pack of Microsoft Visual Studio 2005 was used for website-interview medical expert system.The database of the system was constructed with Microsoft SQL Server 2005. The developed expert system contained large data memory and high efficient function of data interview and data analysis for diagnosis of various diseases.The users were able to perform this system to obtain diagnosis for common respiratory diseases via internet. The developed expert system may be used for internet-based diagnosis of various respiratory diseases,particularly in telemedicine setting.
Knowledge-based systems for power management
NASA Technical Reports Server (NTRS)
Lollar, L. F.
1992-01-01
NASA-Marshall's Electrical Power Branch has undertaken the development of expert systems in support of further advancements in electrical power system automation. Attention is given to the features (1) of the Fault Recovery and Management Expert System, (2) a resource scheduler or Master of Automated Expert Scheduling Through Resource Orchestration, and (3) an adaptive load-priority manager, or Load Priority List Management System. The characteristics of an advisory battery manager for the Hubble Space Telescope, designated the 'nickel-hydrogen expert system', are also noted.
Morrison, Zoe; Fernando, Bernard; Kalra, Dipak; Cresswell, Kathrin; Sheikh, Aziz
2014-01-01
We aimed to explore stakeholder views, attitudes, needs, and expectations regarding likely benefits and risks resulting from increased structuring and coding of clinical information within electronic health records (EHRs). Qualitative investigation in primary and secondary care and research settings throughout the UK. Data were derived from interviews, expert discussion groups, observations, and relevant documents. Participants (n=70) included patients, healthcare professionals, health service commissioners, policy makers, managers, administrators, systems developers, researchers, and academics. Four main themes arose from our data: variations in documentation practice; patient care benefits; secondary uses of information; and informing and involving patients. We observed a lack of guidelines, co-ordination, and dissemination of best practice relating to the design and use of information structures. While we identified immediate benefits for direct care and secondary analysis, many healthcare professionals did not see the relevance of structured and/or coded data to clinical practice. The potential for structured information to increase patient understanding of their diagnosis and treatment contrasted with concerns regarding the appropriateness of coded information for patients. The design and development of EHRs requires the capture of narrative information to reflect patient/clinician communication and computable data for administration and research purposes. Increased structuring and/or coding of EHRs therefore offers both benefits and risks. Documentation standards within clinical guidelines are likely to encourage comprehensive, accurate processing of data. As data structures may impact upon clinician/patient interactions, new models of documentation may be necessary if EHRs are to be read and authored by patients.
Morrison, Zoe; Fernando, Bernard; Kalra, Dipak; Cresswell, Kathrin; Sheikh, Aziz
2014-01-01
Objective We aimed to explore stakeholder views, attitudes, needs, and expectations regarding likely benefits and risks resulting from increased structuring and coding of clinical information within electronic health records (EHRs). Materials and methods Qualitative investigation in primary and secondary care and research settings throughout the UK. Data were derived from interviews, expert discussion groups, observations, and relevant documents. Participants (n=70) included patients, healthcare professionals, health service commissioners, policy makers, managers, administrators, systems developers, researchers, and academics. Results Four main themes arose from our data: variations in documentation practice; patient care benefits; secondary uses of information; and informing and involving patients. We observed a lack of guidelines, co-ordination, and dissemination of best practice relating to the design and use of information structures. While we identified immediate benefits for direct care and secondary analysis, many healthcare professionals did not see the relevance of structured and/or coded data to clinical practice. The potential for structured information to increase patient understanding of their diagnosis and treatment contrasted with concerns regarding the appropriateness of coded information for patients. Conclusions The design and development of EHRs requires the capture of narrative information to reflect patient/clinician communication and computable data for administration and research purposes. Increased structuring and/or coding of EHRs therefore offers both benefits and risks. Documentation standards within clinical guidelines are likely to encourage comprehensive, accurate processing of data. As data structures may impact upon clinician/patient interactions, new models of documentation may be necessary if EHRs are to be read and authored by patients. PMID:24186957
DOE Office of Scientific and Technical Information (OSTI.GOV)
Ebert, D.
1997-07-01
This is a report on the CSNI Workshop on Transient Thermal-Hydraulic and Neutronic Codes Requirements held at Annapolis, Maryland, USA November 5-8, 1996. This experts` meeting consisted of 140 participants from 21 countries; 65 invited papers were presented. The meeting was divided into five areas: (1) current and prospective plans of thermal hydraulic codes development; (2) current and anticipated uses of thermal-hydraulic codes; (3) advances in modeling of thermal-hydraulic phenomena and associated additional experimental needs; (4) numerical methods in multi-phase flows; and (5) programming language, code architectures and user interfaces. The workshop consensus identified the following important action items tomore » be addressed by the international community in order to maintain and improve the calculational capability: (a) preserve current code expertise and institutional memory, (b) preserve the ability to use the existing investment in plant transient analysis codes, (c) maintain essential experimental capabilities, (d) develop advanced measurement capabilities to support future code validation work, (e) integrate existing analytical capabilities so as to improve performance and reduce operating costs, (f) exploit the proven advances in code architecture, numerics, graphical user interfaces, and modularization in order to improve code performance and scrutibility, and (g) more effectively utilize user experience in modifying and improving the codes.« less
Hripcsak, George; Wilcox, Adam
2002-01-01
Medical informatics systems are often designed to perform at the level of human experts. Evaluation of the performance of these systems is often constrained by lack of reference standards, either because the appropriate response is not known or because no simple appropriate response exists. Even when performance can be assessed, it is not always clear whether the performance is sufficient or reasonable. These challenges can be addressed if an evaluator enlists the help of clinical domain experts. 1) The experts can carry out the same tasks as the system, and then their responses can be combined to generate a reference standard. 2)The experts can judge the appropriateness of system output directly. 3) The experts can serve as comparison subjects with which the system can be compared. These are separate roles that have different implications for study design, metrics, and issues of reliability and validity. Diagrams help delineate the roles of experts in complex study designs.
NASA Astrophysics Data System (ADS)
Hoffman, Kenneth J.; Keithley, Hudson
1994-12-01
There are few systems which aggregate standardized pertinent clinical observations of discrete patient problems and resolutions. The systematic information supplied by clinicians is generally provided to justify reimbursement from insurers. Insurers, by their nature, and expert in modeling health care costs by diagnosis, procedures, and population risk groups. Medically, they rely on clinician generated diagnostic and coded procedure information. Clinicians will document a patient's status at a discrete point in time through narrative. Clinical notes do not support aggregate and systematic analysis of outcome. A methodology exists and has been used by the US Army Drug and Alcohol Program to model the clinical activities, associated costs, and data requirements of an outpatient clinic. This has broad applicability for a comprehensive health care system to which patient costs and data requirements can be established.
Ong, Rob S G; Post, Johan; van Rooij, Harry; de Haan, Jan
2008-02-13
Cooperatives delivering out of hours care in the Netherlands are hesitant about the use of expert systems during triage. Apart from the extra costs, cooperatives are not sure that quality of triage is sufficiently enhanced by these systems and believe that call duration will be prolonged drastically. No figures about the influence of the use of an expert system during triage on call duration and triage decisions in out of hours care in the Netherlands are available. Electronically registered data concerning call duration and triage decisions were collected in two cooperatives. One in Tilburg, a cooperative in a Southern city of the Netherlands using an expert system, and one in Groningen, a cooperative in a Northern city not using an expert system. Some other relevant information about the care process was collected additionally. Data about call duration was compared using an independent sample t-test. Data about call decisions was compared using Chi Square. The mean call time in the cooperative using the TAS expert system is 4.6 minutes, in the cooperative not using the expert system 3.9 minutes. A significant difference of 0.7 minutes (0.4 - 1.0, 95% CI) minutes. In the cooperative with an expert system a larger percentage of patients is handled by the assistant, patients are less often referred to a telephone consultation with the GP and are less likely to be offered a visit by the GP.A quick interpretation of the impact of the difference in triage decisions, show that these may be large enough to support the hypothesis that longer call duration is compensated for by less contacts with the GP (by telephone or face-to-face). There is no proof, however, that these differences are caused by the use of the triage system. The larger amount of calls handled by the assistant may be partly caused by the fact that the assistants in the cooperative with an expert system more often consult the GP during triage. And it is not likely that the larger amount of home visits in Groningen can be attributed to the absence of an expert system. The expert system only offers advice whether a GP should be seen, not in which way (by consultation in the office or by home visit). The differences in call times between a cooperative using an expert system and a cooperative not using an expert system are small; 0.4 - 1.0 min. Differences in triage decisions were found, but it is not proven that these can be contributed to the use of an expert system.
Ong, Rob SG; Post, Johan; van Rooij, Harry; de Haan, Jan
2008-01-01
Background Cooperatives delivering out of hours care in the Netherlands are hesitant about the use of expert systems during triage. Apart from the extra costs, cooperatives are not sure that quality of triage is sufficiently enhanced by these systems and believe that call duration will be prolonged drastically. No figures about the influence of the use of an expert system during triage on call duration and triage decisions in out of hours care in the Netherlands are available. Methods Electronically registered data concerning call duration and triage decisions were collected in two cooperatives. One in Tilburg, a cooperative in a Southern city of the Netherlands using an expert system, and one in Groningen, a cooperative in a Northern city not using an expert system. Some other relevant information about the care process was collected additionally. Data about call duration was compared using an independent sample t-test. Data about call decisions was compared using Chi Square. Results The mean call time in the cooperative using the TAS expert system is 4.6 minutes, in the cooperative not using the expert system 3.9 minutes. A significant difference of 0.7 minutes (0.4 – 1.0, 95% CI) minutes. In the cooperative with an expert system a larger percentage of patients is handled by the assistant, patients are less often referred to a telephone consultation with the GP and are less likely to be offered a visit by the GP. A quick interpretation of the impact of the difference in triage decisions, show that these may be large enough to support the hypothesis that longer call duration is compensated for by less contacts with the GP (by telephone or face-to-face). There is no proof, however, that these differences are caused by the use of the triage system. The larger amount of calls handled by the assistant may be partly caused by the fact that the assistants in the cooperative with an expert system more often consult the GP during triage. And it is not likely that the larger amount of home visits in Groningen can be attributed to the absence of an expert system. The expert system only offers advice whether a GP should be seen, not in which way (by consultation in the office or by home visit). Conclusion The differences in call times between a cooperative using an expert system and a cooperative not using an expert system are small; 0.4 – 1.0 min. Differences in triage decisions were found, but it is not proven that these can be contributed to the use of an expert system. PMID:18271970
The Potential of Computer-Based Expert Systems for Special Educators in Rural Settings.
ERIC Educational Resources Information Center
Parry, James D.; Ferrara, Joseph M.
Knowledge-based expert computer systems are addressing issues relevant to all special educators, but are particularly relevant in rural settings where human experts are less available because of distance and cost. An expert system is an application of artificial intelligence (AI) that typically engages the user in a dialogue resembling the…
Fire Effects, Education, and Expert Systems
Robert E. Martin
1987-01-01
Predicting the effects of fires in the year 2000 and beyond will be enhanced by the use of expert systems. Although our predictions may have broad confidence limits, expert systems should help us to improve the predictions and to focus on the areas where improved knowledge is most needed. The knowledge of experts can be incorporated into previously existing knowledge...
NASA Technical Reports Server (NTRS)
Happell, Nadine; Miksell, Steve; Carlisle, Candace
1989-01-01
A major barrier in taking expert systems from prototype to operational status involves instilling end user confidence in the operational system. The software of different life cycle models is examined and the advantages and disadvantages of each when applied to expert system development are explored. The Fault Isolation Expert System for Tracking and data relay satellite system Applications (FIESTA) is presented as a case study of development of an expert system. The end user confidence necessary for operational use of this system is accentuated by the fact that it will handle real-time data in a secure environment, allowing little tolerance for errors. How FIESTA is dealing with transition problems as it moves from an off-line standalone prototype to an on-line real-time system is discussed.
NASA Technical Reports Server (NTRS)
Happell, Nadine; Miksell, Steve; Carlisle, Candace
1989-01-01
A major barrier in taking expert systems from prototype to operational status involves instilling end user confidence in the operational system. The software of different life cycle models is examined and the advantages and disadvantages of each when applied to expert system development are explored. The Fault Isolation Expert System for Tracking and data relay satellite system Applications (FIESTA) is presented as a case study of development of an expert system. The end user confidence necessary for operational use of this system is accentuated by the fact that it will handle real-time data in a secure environment, allowing little tolerance for errors. How FIESTA is dealing with transition problems as it moves from an off-line standalone prototype to an on-line real-time system is discussed.
NASA Technical Reports Server (NTRS)
Liebowitz, Jay; Krishnamurthy, Vijaya; Rodens, Ira; Houston, Chapman; Liebowitz, Alisa; Baek, Seung; Radko, Joe; Zeide, Janet
1996-01-01
Scheduling has become an increasingly important element in today's society and workplace. Within the NASA environment, scheduling is one of the most frequently performed and challenging functions. Towards meeting NASA's scheduling needs, a research version of a generic expert scheduling system architecture and toolkit has been developed. This final report describes the development and testing of GUESS (Generically Used Expert Scheduling System).
Korošec, Peter; Eftimov, Tome; Ocke, Marga; van der Laan, Jan; Roe, Mark; Berry, Rachel; Turrini, Aida; Krems, Carolin; Slimani, Nadia; Finglas, Paul
2018-01-01
This paper identifies the requirements for computer-supported food matching, in order to address not only national and European but also international current related needs and represents an integrated research contribution of the FP7 EuroDISH project. The available classification and coding systems and the specific problems of food matching are summarized and a new concept for food matching based on optimization methods and machine-based learning is proposed. To illustrate and test this concept, a study has been conducted in four European countries (i.e., Germany, The Netherlands, Italy and the UK) using different classification and coding systems. This real case study enabled us to evaluate the new food matching concept and provide further recommendations for future work. In the first stage of the study, we prepared subsets of food consumption data described and classified using different systems, that had already been manually matched with national food composition data. Once the food matching algorithm was trained using this data, testing was performed on another subset of food consumption data. Experts from different countries validated food matching between consumption and composition data by selecting best matches from the options given by the matching algorithm without seeing the result of the previously made manual match. The evaluation of study results stressed the importance of the role and quality of the food composition database as compared to the selected classification and/or coding systems and the need to continue compiling national food composition data as eating habits and national dishes still vary between countries. Although some countries managed to collect extensive sets of food consumption data, these cannot be easily matched with food composition data if either food consumption or food composition data are not properly classified and described using any classification and coding systems. The study also showed that the level of human expertise played an important role, at least in the training stage. Both sets of data require continuous development to improve their quality in dietary assessment. PMID:29601516
ATS displays: A reasoning visualization tool for expert systems
NASA Technical Reports Server (NTRS)
Selig, William John; Johannes, James D.
1990-01-01
Reasoning visualization is a useful tool that can help users better understand the inherently non-sequential logic of an expert system. While this is desirable in most all expert system applications, it is especially so for such critical systems as those destined for space-based operations. A hierarchical view of the expert system reasoning process and some characteristics of these various levels is presented. Also presented are Abstract Time Slice (ATS) displays, a tool to visualize the plethora of interrelated information available at the host inferencing language level of reasoning. The usefulness of this tool is illustrated with some examples from a prototype potable water expert system for possible use aboard Space Station Freedom.
COMPUTERIZED RISK AND BIOACCUMULATION SYSTEM (VERSION 1.0)
CRABS is a combination of a rule-based expert system and more traditional procedural programming techniques. ule-based expert systems attempt to emulate the decision making process of human experts within a clearly defined subject area. xpert systems consist of an "inference engi...
An Expert System for Environmental Data Management.
ERIC Educational Resources Information Center
Berka, Petr; Jirku, Petr
1995-01-01
Examines the possibility of using expert system tools for environmental data management. Describes the domain-independent expert system shell SAK and Knowledge EXplorer, a system that learns rules from data. Demonstrates the functionality of Knowledge EXplorer on an example of water quality evaluation. (LZ)
Data reduction expert assistant
NASA Technical Reports Server (NTRS)
Miller, Glenn E.; Johnston, Mark D.; Hanisch, Robert J.
1991-01-01
Viewgraphs on data reduction expert assistant are presented. Topics covered include: data analysis systems; philosophy of these systems; disadvantages; expert assistant; useful goals; and implementation considerations.
Lawrence, Renée H; Tomolo, Anne M
2011-03-01
Although practice-based learning and improvement (PBLI) is now recognized as a fundamental and necessary skill set, we are still in need of tools that yield specific information about gaps in knowledge and application to help nurture the development of quality improvement (QI) skills in physicians in a proficient and proactive manner. We developed a questionnaire and coding system as an assessment tool to evaluate and provide feedback regarding PBLI self-efficacy, knowledge, and application skills for residency programs and related professional requirements. Five nationally recognized QI experts/leaders reviewed and completed our questionnaire. Through an iterative process, a coding system based on identifying key variables needed for ideal responses was developed to score project proposals. The coding system comprised 14 variables related to the QI projects, and an additional 30 variables related to the core knowledge concepts related to PBLI. A total of 86 residents completed the questionnaire, and 2 raters coded their open-ended responses. Interrater reliability was assessed by percentage agreement and Cohen κ for individual variables and Lin concordance correlation for total scores for knowledge and application. Discriminative validity (t test to compare known groups) and coefficient of reproducibility as an indicator of construct validity (item difficulty hierarchy) were also assessed. Interrater reliability estimates were good (percentage of agreements, above 90%; κ, above 0.4 for most variables; concordances for total scores were R = .88 for knowledge and R = .98 for application). Despite the residents' limited range of experiences in the group with prior PBLI exposure, our tool met our goal of differentiating between the 2 groups in our preliminary analyses. Correcting for chance agreement identified some variables that are potentially problematic. Although additional evaluation is needed, our tool may prove helpful and provide detailed information about trainees' progress and the curriculum.
Lawrence, Renée H; Tomolo, Anne M
2011-01-01
Background Although practice-based learning and improvement (PBLI) is now recognized as a fundamental and necessary skill set, we are still in need of tools that yield specific information about gaps in knowledge and application to help nurture the development of quality improvement (QI) skills in physicians in a proficient and proactive manner. We developed a questionnaire and coding system as an assessment tool to evaluate and provide feedback regarding PBLI self-efficacy, knowledge, and application skills for residency programs and related professional requirements. Methods Five nationally recognized QI experts/leaders reviewed and completed our questionnaire. Through an iterative process, a coding system based on identifying key variables needed for ideal responses was developed to score project proposals. The coding system comprised 14 variables related to the QI projects, and an additional 30 variables related to the core knowledge concepts related to PBLI. A total of 86 residents completed the questionnaire, and 2 raters coded their open-ended responses. Interrater reliability was assessed by percentage agreement and Cohen κ for individual variables and Lin concordance correlation for total scores for knowledge and application. Discriminative validity (t test to compare known groups) and coefficient of reproducibility as an indicator of construct validity (item difficulty hierarchy) were also assessed. Results Interrater reliability estimates were good (percentage of agreements, above 90%; κ, above 0.4 for most variables; concordances for total scores were R = .88 for knowledge and R = .98 for application). Conclusion Despite the residents' limited range of experiences in the group with prior PBLI exposure, our tool met our goal of differentiating between the 2 groups in our preliminary analyses. Correcting for chance agreement identified some variables that are potentially problematic. Although additional evaluation is needed, our tool may prove helpful and provide detailed information about trainees' progress and the curriculum. PMID:22379522
A Systematic Approach for Obtaining Performance on Matrix-Like Operations
NASA Astrophysics Data System (ADS)
Veras, Richard Michael
Scientific Computation provides a critical role in the scientific process because it allows us ask complex queries and test predictions that would otherwise be unfeasible to perform experimentally. Because of its power, Scientific Computing has helped drive advances in many fields ranging from Engineering and Physics to Biology and Sociology to Economics and Drug Development and even to Machine Learning and Artificial Intelligence. Common among these domains is the desire for timely computational results, thus a considerable amount of human expert effort is spent towards obtaining performance for these scientific codes. However, this is no easy task because each of these domains present their own unique set of challenges to software developers, such as domain specific operations, structurally complex data and ever-growing datasets. Compounding these problems are the myriads of constantly changing, complex and unique hardware platforms that an expert must target. Unfortunately, an expert is typically forced to reproduce their effort across multiple problem domains and hardware platforms. In this thesis, we demonstrate the automatic generation of expert level high-performance scientific codes for Dense Linear Algebra (DLA), Structured Mesh (Stencil), Sparse Linear Algebra and Graph Analytic. In particular, this thesis seeks to address the issue of obtaining performance on many complex platforms for a certain class of matrix-like operations that span across many scientific, engineering and social fields. We do this by automating a method used for obtaining high performance in DLA and extending it to structured, sparse and scale-free domains. We argue that it is through the use of the underlying structure found in the data from these domains that enables this process. Thus, obtaining performance for most operations does not occur in isolation of the data being operated on, but instead depends significantly on the structure of the data.
75 FR 19268 - Kasugamycin; Pesticide Tolerances for Emergency Exemptions
Federal Register 2010, 2011, 2012, 2013, 2014
2010-04-14
... entities may include, but are not limited to: Crop production (NAICS code 111). Animal production (NAICS... into account uncertainties inherent in the extrapolation from laboratory animal data to humans and in... recommended by public health experts to sustain the effectiveness of antibiotic materials. Field use of this...
Expertise and age differences in pilot decision making.
Morrow, Daniel G; Miller, Lisa M Soederberg; Ridolfo, Heather E; Magnor, Clifford; Fischer, Ute M; Kokayeff, Nina K; Stine-Morrow, Elizabeth A L
2009-01-01
We examined the influence of age and expertise on pilot decision making. Older and younger expert and novice pilots read at their own pace scenarios describing simpler or more complex flight situations. Then in a standard interview they discussed the scenario problem and how they would respond. Protocols were coded for identification of problem and solutions to this problem, and frequency of elaborations on problem and solution. Scenario comprehension was measured as differential reading time allocation to problem-critical information and scenario memory by the accuracy of answering questions about the scenarios after the interview. All groups accurately identified the problems, but experts elaborated problem descriptions more than novices did. Experts also spent more time reading critical information in the complex scenarios, which may reflect time needed to develop elaborate situation models of the problems. Expertise comprehension benefits were similar for older and younger pilots. Older experts were especially likely to elaborate the problem compared to younger experts, while older novices were less likely to elaborate the problem and to identify appropriate solutions compared to their younger counterparts. The findings suggest age invariance in knowledge-based comprehension relevant to pilot decision making.
iPhone app adherence to expert-recommended guidelines for pediatric obesity prevention.
Wearing, Jessica R; Nollen, Nikki; Befort, Christie; Davis, Ann M; Agemy, Carolina K
2014-04-01
Pediatric obesity is a serious and prevalent problem. Smartphone technology, which is becoming increasingly available to children of diverse backgrounds, presents a unique opportunity to instill healthy behaviors before the onset of obesity. Past studies have examined the use of smartphone applications as tools of health behavior modification for adults. The present study examines the content of children's exercise and nutrition smartphone apps. Sixty-two iPhone apps were identified and coded by two independent raters for adherence to expert-recommended behaviors (e.g., five fruits/vegetables per day) and strategies (e.g., self-monitoring diet/physical activity) for the prevention of pediatric obesity. App behavioral and strategy index scores were uniformly low. Apps were more likely to address expert-recommended behaviors for the prevention of pediatric obesity (93.5%), whereas few apps addressed recommended strategies (20.9%). The most common behaviors addressed included physical activity (53.2%) and fruit/vegetable consumption (48.3%). Other important behaviors (e.g., screen time [1.6%] and family meals together [1.6%]) were rarely addressed. Current children's diet and exercise apps could be improved with increased adherence to expert-recommended guidelines, especially expert-recommended strategies.
iPhone App Adherence to Expert-Recommended Guidelines for Pediatric Obesity Prevention
Wearing, Jessica R.; Befort, Christie; Davis, Ann M.; Agemy, Carolina K.
2014-01-01
Abstract Background: Pediatric obesity is a serious and prevalent problem. Smartphone technology, which is becoming increasingly available to children of diverse backgrounds, presents a unique opportunity to instill healthy behaviors before the onset of obesity. Past studies have examined the use of smartphone applications as tools of health behavior modification for adults. The present study examines the content of children's exercise and nutrition smartphone apps. Method: Sixty-two iPhone apps were identified and coded by two independent raters for adherence to expert-recommended behaviors (e.g., five fruits/vegetables per day) and strategies (e.g., self-monitoring diet/physical activity) for the prevention of pediatric obesity. Results: App behavioral and strategy index scores were uniformly low. Apps were more likely to address expert-recommended behaviors for the prevention of pediatric obesity (93.5%), whereas few apps addressed recommended strategies (20.9%). The most common behaviors addressed included physical activity (53.2%) and fruit/vegetable consumption (48.3%). Other important behaviors (e.g., screen time [1.6%] and family meals together [1.6%]) were rarely addressed. Conclusions: Current children's diet and exercise apps could be improved with increased adherence to expert-recommended guidelines, especially expert-recommended strategies. PMID:24655230
ERIC Educational Resources Information Center
Hankins, George.
1987-01-01
Describes the novice-to-expert model of human learning and compares it to the recent advances in the areas of artificial intelligence and expert systems. Discusses some of the characteristics of experts, proposing connections between them with expert systems and theories of left-right brain functions. (TW)
Propulsion Control Technology Development in the United States A Historical Perspective
NASA Technical Reports Server (NTRS)
Jaw, Link C.a; Garg, Sanjay
2005-01-01
This paper presents a historical perspective of the advancement of control technologies for aircraft gas turbine engines. The paper primarily covers technology advances in the United States in the last 60 years (1940 to approximately 2002). The paper emphasizes the pioneering technologies that have been tested or implemented during this period, assimilating knowledge and experience from industry experts, including personal interviews with both current and retired experts. Since the first United States-built aircraft gas turbine engine was flown in 1942, engine control technology has evolved from a simple hydro-mechanical fuel metering valve to a full-authority digital electronic control system (FADEC) that is common to all modern aircraft propulsion systems. At the same time, control systems have provided engine diagnostic functions. Engine diagnostic capabilities have also evolved from pilot observation of engine gauges to the automated on-board diagnostic system that uses mathematical models to assess engine health and assist in post-flight troubleshooting and maintenance. Using system complexity and capability as a measure, we can break the historical development of control systems down to four phases: (1) the start-up phase (1942 to 1949), (2) the growth phase (1950 to 1969), (3) the electronic phase (1970 to 1989), and (4) the integration phase (1990 to 2002). In each phase, the state-of-the-art control technology is described and the engines that have become historical landmarks, from the control and diagnostic standpoint, are identified. Finally, a historical perspective of engine controls in the last 60 years is presented in terms of control system complexity, number of sensors, number of lines of software (or embedded code), and other factors.
Expert systems in clinical microbiology.
Winstanley, Trevor; Courvalin, Patrice
2011-07-01
This review aims to discuss expert systems in general and how they may be used in medicine as a whole and clinical microbiology in particular (with the aid of interpretive reading). It considers rule-based systems, pattern-based systems, and data mining and introduces neural nets. A variety of noncommercial systems is described, and the central role played by the EUCAST is stressed. The need for expert rules in the environment of reset EUCAST breakpoints is also questioned. Commercial automated systems with on-board expert systems are considered, with emphasis being placed on the "big three": Vitek 2, BD Phoenix, and MicroScan. By necessity and in places, the review becomes a general review of automated system performances for the detection of specific resistance mechanisms rather than focusing solely on expert systems. Published performance evaluations of each system are drawn together and commented on critically.
OECD/NEA Ongoing activities related to the nuclear fuel cycle
DOE Office of Scientific and Technical Information (OSTI.GOV)
Cornet, S.M.; McCarthy, K.; Chauvin, N.
2013-07-01
As part of its role in encouraging international collaboration, the OECD Nuclear Energy Agency is coordinating a series of projects related to the Nuclear Fuel Cycle. The Nuclear Science Committee (NSC) Working Party on Scientific Issues of the Nuclear Fuel Cycle (WPFC) comprises five different expert groups covering all aspects of the fuel cycle from front to back-end. Activities related to fuels, materials, physics, separation chemistry, and fuel cycles scenarios are being undertaken. By publishing state-of-the-art reports and organizing workshops, the groups are able to disseminate recent research advancements to the international community. Current activities mainly focus on advanced nuclearmore » systems, and experts are working on analyzing results and establishing challenges associated to the adoption of new materials and fuels. By comparing different codes, the Expert Group on Advanced Fuel Cycle Scenarios is aiming at gaining further understanding of the scientific issues and specific national needs associated with the implementation of advanced fuel cycles. At the back end of the fuel cycle, separation technologies (aqueous and pyrochemical processing) are being assessed. Current and future activities comprise studies on minor actinides separation and post Fukushima studies. Regular workshops are also organized to discuss recent developments on Partitioning and Transmutation. In addition, the Nuclear Development Committee (NDC) focuses on the analysis of the economics of nuclear power across the fuel cycle in the context of changes of electricity markets, social acceptance and technological advances and assesses the availability of the nuclear fuel and infrastructure required for the deployment of existing and future nuclear power. The Expert Group on the Economics of the Back End of the Nuclear Fuel Cycle (EBENFC), in particular, is looking at assessing economic and financial issues related to the long term management of spent nuclear fuel. (authors)« less
EXPERT: An atmospheric re-entry test-bed
NASA Astrophysics Data System (ADS)
Massobrio, F.; Viotto, R.; Serpico, M.; Sansone, A.; Caporicci, M.; Muylaert, J.-M.
2007-06-01
In recognition of the importance of an independent European access to the International Space Station (ISS) and in preparation for the future needs of exploration missions, ESA is conducting parallel activities to generate flight data using atmospheric re-entry test-beds and to identify vehicle design solutions for human and cargo transportation vehicles serving the ISS and beyond. The EXPERT (European eXPErimental Re-entry Test-bed) vehicle represents the major on-going development in the first class of activities. Its results may also benefit in due time scientific missions to planets with an atmosphere and future reusable launcher programmes. The objective of EXPERT is to provide a test-bed for the validation of aerothermodynamics models, codes and ground test facilities in a representative flight environment, to improve the understanding of issues related to analysis, testing and extrapolation to flight. The vehicle will be launched on a sub-orbital trajectory using a Volna missile. The EXPERT concept is based on a symmetrical re-entry capsule whose shape is composed of simple geometrical elements. The suborbital trajectory will reach 120 km altitude and a re-entry velocity of 5 6km/s. The dimensions of the capsule are 1.6 m high and 1.3 m diameter; the overall mass is in the range of 250 350kg, depending upon the mission parameters and the payload/instrumentation complement. A consistent number of scientific experiments are foreseen on-board, from innovative air data system to shock wave/boundary layer interaction, from sharp hot structures characterisation to natural and induced regime transition. Currently the project is approaching completion of the phase B, with Alenia Spazio leading the industrial team and CIRA coordinating the scientific payload development under ESA contract.
Sheikhtaheri, Abbas; Sadoughi, Farahnaz; Hashemi Dehaghi, Zahra
2014-09-01
Complicacy of clinical decisions justifies utilization of information systems such as artificial intelligence (e.g. expert systems and neural networks) to achieve better decisions, however, application of these systems in the medical domain faces some challenges. We aimed at to review the applications of these systems in the medical domain and discuss about such challenges. Following a brief introduction of expert systems and neural networks by representing few examples, the challenges of these systems in the medical domain are discussed. We found that the applications of expert systems and artificial neural networks have been increased in the medical domain. These systems have shown many advantages such as utilization of experts' knowledge, gaining rare knowledge, more time for assessment of the decision, more consistent decisions, and shorter decision-making process. In spite of all these advantages, there are challenges ahead of developing and using such systems including maintenance, required experts, inputting patients' data into the system, problems for knowledge acquisition, problems in modeling medical knowledge, evaluation and validation of system performance, wrong recommendations and responsibility, limited domains of such systems and necessity of integrating such systems into the routine work flows. We concluded that expert systems and neural networks can be successfully used in medicine; however, there are many concerns and questions to be answered through future studies and discussions.
a Study on Satellite Diagnostic Expert Systems Using Case-Based Approach
NASA Astrophysics Data System (ADS)
Park, Young-Tack; Kim, Jae-Hoon; Park, Hyun-Soo
1997-06-01
Many research works are on going to monitor and diagnose diverse malfunctions of satellite systems as the complexity and number of satellites increase. Currently, many works on monitoring and diagnosis are carried out by human experts but there are needs to automate much of the routine works of them. Hence, it is necessary to study on using expert systems which can assist human experts routine work by doing automatically, thereby allow human experts devote their expertise more critical and important areas of monitoring and diagnosis. In this paper, we are employing artificial intelligence techniques to model human experts' knowledge and inference the constructed knowledge. Especially, case-based approaches are used to construct a knowledge base to model human expert capabilities which use previous typical exemplars. We have designed and implemented a prototype case-based system for diagnosing satellite malfunctions using cases. Our system remembers typical failure cases and diagnoses a current malfunction by indexing the case base. Diverse methods are used to build a more user friendly interface which allows human experts can build a knowledge base in as easy way.
Feurzeig, Wallace
1984-01-01
The first expert instructional system, the Socratic System, was developed in 1964. One of the earliest applications of this system was in the area of differential diagnosis in clinical medicine. The power of the underlying instructional paradigm was demonstrated and the potential of the approach for valuably supplementing medical instruction was recognized. Twenty years later, despite further educationally significant advances in expert systems technology and enormous reductions in the cost of computers, expert instructional methods have found very little application in medical schools.
Projects in an expert system class
NASA Technical Reports Server (NTRS)
Whitson, George M.
1991-01-01
Many universities now teach courses in expert systems. In these courses students study the architecture of an expert system, knowledge acquisition techniques, methods of implementing systems and verification and validation techniques. A major component of any such course is a class project consisting of the design and implementation of an expert system. Discussed here are a number of techniques that we have used at the University of Texas at Tyler to develop meaningful projects that could be completed in a semester course.
Program maintenance manual for nickel cadmium battery expert system, version 1
NASA Technical Reports Server (NTRS)
1986-01-01
The Nickel-Cadmium Battery Expert System (NICBES) is an expert system for fault diagnosis and advice of the nickel-cadmium batteries found in the Hubble Space Telescope (HST). The system application and security, equipment environment, and the program maintenance procedures are examined.
A Logic Basis for Information Retrieval.
ERIC Educational Resources Information Center
Watters, C. R.; Shepherd, M. A.
1987-01-01
Discusses the potential of recent work in artificial intelligence, especially expert systems, for the development of more effective information retrieval systems. Highlights include the role of an expert bibliographic retrieval system and a prototype expert retrieval system, PROBIB-2, that uses MicroProlog to provide deductive reasoning…
An expert system for the quantification of fault rates in construction fall accidents.
Talat Birgonul, M; Dikmen, Irem; Budayan, Cenk; Demirel, Tuncay
2016-01-01
Expert witness reports, prepared with the aim of quantifying fault rates among parties, play an important role in a court's final decision. However, conflicting fault rates assigned by different expert witness boards lead to iterative objections raised by the related parties. This unfavorable situation mainly originates due to the subjectivity of expert judgments and unavailability of objective information about the causes of accidents. As a solution to this shortcoming, an expert system based on a rule-based system was developed for the quantification of fault rates in construction fall accidents. The aim of developing DsSafe is decreasing the subjectivity inherent in expert witness reports. Eighty-four inspection reports prepared by the official and authorized inspectors were examined and root causes of construction fall accidents in Turkey were identified. Using this information, an evaluation form was designed and submitted to the experts. Experts were asked to evaluate the importance level of the factors that govern fall accidents and determine the fault rates under different scenarios. Based on expert judgments, a rule-based expert system was developed. The accuracy and reliability of DsSafe were tested with real data as obtained from finalized court cases. DsSafe gives satisfactory results.
NASA Technical Reports Server (NTRS)
Durkin, John; Schlegelmilch, Richard; Tallo, Donald
1992-01-01
A research effort was undertaken to investigate how expert system technology could be applied to a satellite communications system. The focus of the expert system is the satellite earth station. A proof of concept expert system called the Ground Terminal Expert (GTEX) was developed at the University of Akron in collaboration with the NASA Lewis Research Center. With the increasing demand for satellite earth stations, maintenance is becoming a vital issue. Vendors of such systems will be looking for cost effective means of maintaining such systems. The objective of GTEX is to aid in diagnosis of faults occurring with the digital earth station. GTEX was developed on a personal computer using the Automated Reasoning Tool for Information Management (ART-IM) developed by the Inference Corporation. Developed for the Phase 2 digital earth station, GTEX is a part of the Systems Integration Test and Evaluation (SITE) facility located at the NASA Lewis Research Center.
Threat expert system technology advisor
NASA Technical Reports Server (NTRS)
Kurrasch, E. R.; Tripp, L. R.
1987-01-01
A prototype expert system was developed to determine the feasibility of using expert system technology to enhance the performance and survivability of helicopter pilots in a combat threat environment while flying NOE (Nap of the Earth) missions. The basis for the concept is the potential of using an Expert System Advisor to reduce the extreme overloading of the pilot who flies NOE mission below treetop level at approximately 40 knots while performing several other functions. The ultimate goal is to develop a Threat Expert System Advisor which provides threat information and advice that are better than even a highly experienced copilot. The results clearly show that the NOE pilot needs all the help in decision aiding and threat situation awareness that he can get. It clearly shows that heuristics are important and that an expert system for combat NOE helicopter missions can be of great help to the pilot in complex threat situations and in making decisions.
[The application and development of artificial intelligence in medical diagnosis systems].
Chen, Zhencheng; Jiang, Yong; Xu, Mingyu; Wang, Hongyan; Jiang, Dazong
2002-09-01
This paper has reviewed the development of artificial intelligence in medical practice and medical diagnostic expert systems, and has summarized the application of artificial neural network. It explains that a source of difficulty in medical diagnostic system is the co-existence of multiple diseases--the potentially inter-related diseases. However, the difficulty of image expert systems is inherent in high-level vision. And it increases the complexity of expert system in medical image. At last, the prospect for the development of artificial intelligence in medical image expert systems is made.
Intelligent systems for human resources.
Kline, K B
1988-11-01
An intelligent system contains knowledge about some domain; it has sophisticated decision-making processes and the ability to explain its actions. The most important aspect of an intelligent system is its ability to effectively interact with humans to teach or assist complex information processing. Two intelligent systems are Intelligent Tutoring Systems (ITs) and Expert Systems. The ITSs provide instruction to a student similar to a human tutor. The ITSs capture individual performance and tutor deficiencies. These systems consist of an expert module, which contains the knowledge or material to be taught; the student module, which contains a representation of the knowledge the student knows and does not know about the domain; and the instructional or teaching module, which selects specific knowledge to teach, the instructional strategy, and provides assistance to the student to tutor deficiencies. Expert systems contain an expert's knowledge about some domain and perform specialized tasks or aid a novice in the performance of certain tasks. The most important part of an expert system is the knowledge base. This knowledge base contains all the specialized and technical knowledge an expert possesses. For an expert system to interact effectively with humans, it must have the ability to explain its actions. Use of intelligent systems can have a profound effect on human resources. The ITSs can provide better training by tutoring on an individual basis, and the expert systems can make better use of human resources through job aiding and performing complex tasks. With increasing training requirements and "doing more with less," intelligent systems can have a positive effect on human resources.
ERIC Educational Resources Information Center
Borko, Harold
1985-01-01
Defines artificial intelligence (AI) and expert systems; describes library applications utilizing AI to automate creation of document representations, request formulations, and design and modify search strategies for information retrieval systems; discusses expert system development for information services; and reviews impact of these…
Enhanced use of CLIPS at the Los Alamos National Laboratory
NASA Technical Reports Server (NTRS)
Duerre, K. H.; Parkinson, W. J.; Osowski, J. J.
1991-01-01
Early efforts for producing expert systems for engineering applications used a limited subset of C Language Integrated Production System (CLIPS) features. The implementation details of previous expert systems and of the current expert system, which is used for training operators in the control of the Isotope Separation System, are discussed.
Third CLIPS Conference Proceedings, volume 2
NASA Technical Reports Server (NTRS)
Riley, Gary (Editor)
1994-01-01
Expert systems are computer programs which emulate human expertise in well defined problem domains. The C Language Integrated Production System (CLIPS) is an expert system building tool, developed at the Johnson Space Center, which provides a complete environment for the development and delivery of rule and/or object based expert systems. CLIPS was specifically designed to provide a low cost option for developing and deploying expert system applications across a wide range of hardware platforms. The development of CLIPS has helped to improve the ability to deliver expert system technology throughout the public and private sectors for a wide range of applications and diverse computing environments. The Third Conference on CLIPS provided a forum for CLIPS users to present and discuss papers relating to CLIPS applications, uses, and extensions.
SigmaCLIPSE = presentation management + NASA CLI PS + SQL
NASA Technical Reports Server (NTRS)
Weiss, Bernard P., Jr.
1990-01-01
SigmaCLIPSE provides an expert systems and 'intelligent' data base development program for diverse systems integration environments that require support for automated reasoning and expert systems technology, presentation management, and access to 'intelligent' SQL data bases. The SigmaCLIPSE technology and and its integrated ability to access 4th generation application development and decision support tools through a portable SQL interface, comprises a sophisticated software development environment for solving knowledge engineering and expert systems development problems in information intensive commercial environments -- financial services, health care, and distributed process control -- where the expert system must be extendable -- a major architectural advantage of NASA CLIPS. SigmaCLIPSE is a research effort intended to test the viability of merging SQL data bases with expert systems technology.
A diagnostic expert system for aircraft generator control unit (GCU)
NASA Astrophysics Data System (ADS)
Ho, Ting-Long; Bayles, Robert A.; Havlicsek, Bruce L.
The modular VSCF (variable-speed constant-frequency) generator families are described as using standard modules to reduce the maintenance cost and to improve the product's testability. A general diagnostic expert system shell that guides troubleshooting of modules or line replaceable units (LRUs) is introduced. An application of the diagnostic system to a particular LRU, the generator control unit (GCU) is reported. The approach to building the diagnostic expert system is first to capture general diagnostic strategy in an expert system shell. This shell can be easily applied to different devices or LRUs by writing rules to capture only additional device-specific diagnostic information from expert repair personnel. The diagnostic system has the necessary knowledge embedded in its programs and exhibits expertise to troubleshoot the GCU.
Weindling, P
2001-01-01
The Nuremberg Code has generally been seen as arising from the Nuremberg Medical Trial. This paper examines developments prior to the Trial, involving the physiologist Andrew Conway Ivy and an inter-Allied Scientific Commission on Medical War Crimes. The paper traces the formulation of the concept of a medical war crime by the physiologist John West Thompson, as part of the background to Ivy's code on human experiments of 1 August 1946. It evaluates subsequent responses by the American Medical Association, and by other war crimes experts, notably Leo Alexander, who developed Ivy's conceptual framework. Ivy's interaction with the judges at Nuremberg alerted them to the importance of formulating ethical guidelines for clinical research.
Watterson, Dina; Cleland, Heather; Picton, Natalie; Simpson, Pam M; Gabbe, Belinda J
2011-03-01
The percentage of total body surface area burnt (%TBSA) is a critical measure of burn injury severity and a key predictor of burn injury outcome. This study evaluated the level of agreement between four sources of %TBSA using 120 cases identified through the Victorian State Trauma Registry. Expert clinician, ICD-10-AM, Abbreviated Injury Scale, and burns registry coding were compared using measures of agreement. There was near-perfect agreement (weighted Kappa statistic 0.81-1) between all sources of data, suggesting that ICD-10-AM is a valid source of %TBSA and use of ICD-10-AM codes could reduce the resource used by trauma and burns registries capturing this information.
A rule-based expert system for generating control displays at the Advanced Photon Source
NASA Astrophysics Data System (ADS)
Coulter, Karen J.
1994-12-01
The integration of a rule-based expert system for generating screen displays for controlling and monitoring instrumentation under the Experimental Physics and Industrial Control System (EPICS) is presented. The expert system is implemented using CLIPS, an expert system shell from the Software Technology Branch at Lyndon B. Johnson Space Center. The user selects the hardware input and output to be displayed and the expert system constructs a graphical control screen appropriate for the data. Such a system provides a method for implementing a common look and feel for displays created by several different users and reduces the amount of time required to create displays for new hardware configurations. Users are able to modify the displays as needed using the EPICS display editor tool.
The potential of expert systems for remote sensing application
NASA Technical Reports Server (NTRS)
Mooneyhan, D. W.
1983-01-01
An overview of the status and potential of artificial intelligence-driven expert systems in the role of image data analysis is presented. An expert system is defined and its structure is summarized. Three such systems designed for image interpretation are outlined. The use of an expert system to detect changes on the earth's surface is discussed, and the components of a knowledge-based image interpretation system and their make-up are outlined. An example of how such a system should work for an area in the tropics where deforestation has occurred is presented as a sequence of situation/action decisions.
A CLIPS-based expert system for the evaluation and selection of robots
NASA Technical Reports Server (NTRS)
Nour, Mohamed A.; Offodile, Felix O.; Madey, Gregory R.
1994-01-01
This paper describes the development of a prototype expert system for intelligent selection of robots for manufacturing operations. The paper first develops a comprehensive, three-stage process to model the robot selection problem. The decisions involved in this model easily lend themselves to an expert system application. A rule-based system, based on the selection model, is developed using the CLIPS expert system shell. Data about actual robots is used to test the performance of the prototype system. Further extensions to the rule-based system for data handling and interfacing capabilities are suggested.
Expert Systems in Clinical Microbiology
Winstanley, Trevor; Courvalin, Patrice
2011-01-01
Summary: This review aims to discuss expert systems in general and how they may be used in medicine as a whole and clinical microbiology in particular (with the aid of interpretive reading). It considers rule-based systems, pattern-based systems, and data mining and introduces neural nets. A variety of noncommercial systems is described, and the central role played by the EUCAST is stressed. The need for expert rules in the environment of reset EUCAST breakpoints is also questioned. Commercial automated systems with on-board expert systems are considered, with emphasis being placed on the “big three”: Vitek 2, BD Phoenix, and MicroScan. By necessity and in places, the review becomes a general review of automated system performances for the detection of specific resistance mechanisms rather than focusing solely on expert systems. Published performance evaluations of each system are drawn together and commented on critically. PMID:21734247
Controlling Real-Time Processes On The Space Station With Expert Systems
NASA Astrophysics Data System (ADS)
Leinweber, David; Perry, John
1987-02-01
Many aspects of space station operations involve continuous control of real-time processes. These processes include electrical power system monitoring, propulsion system health and maintenance, environmental and life support systems, space suit checkout, on-board manufacturing, and servicing of attached vehicles such as satellites, shuttles, orbital maneuvering vehicles, orbital transfer vehicles and remote teleoperators. Traditionally, monitoring of these critical real-time processes has been done by trained human experts monitoring telemetry data. However, the long duration of space station missions and the high cost of crew time in space creates a powerful economic incentive for the development of highly autonomous knowledge-based expert control procedures for these space stations. In addition to controlling the normal operations of these processes, the expert systems must also be able to quickly respond to anomalous events, determine their cause and initiate corrective actions in a safe and timely manner. This must be accomplished without excessive diversion of system resources from ongoing control activities and any events beyond the scope of the expert control and diagnosis functions must be recognized and brought to the attention of human operators. Real-time sensor based expert systems (as opposed to off-line, consulting or planning systems receiving data via the keyboard) pose particular problems associated with sensor failures, sensor degradation and data consistency, which must be explicitly handled in an efficient manner. A set of these systems must also be able to work together in a cooperative manner. This paper describes the requirements for real-time expert systems in space station control, and presents prototype implementations of space station expert control procedures in PICON (process intelligent control). PICON is a real-time expert system shell which operates in parallel with distributed data acquisition systems. It incorporates a specialized inference engine with a specialized scheduling portion specifically designed to match the allocation of system resources with the operational requirements of real-time control systems. Innovative knowledge engineering techniques used in PICON to facilitate the development of real-time sensor-based expert systems which use the special features of the inference engine are illustrated in the prototype examples.
Decision support system and medical liability.
Allaërt, F. A.; Dusserre, L.
1992-01-01
Expert systems, which are going to be an essential tool in Medicine, are evolving in terms of sophistication of both knowledge representation and types of reasoning models used. The more efficient they are, the more often they will be used and professional liability will be involved. So after giving a short survey of configuration and working of expert systems, the authors will study the liabilities of people building and the using expert systems regarding some various dysfunctions. Of course the expert systems have to be considered only for human support and they should not possess any authority themselves, therefore the doctors must keep in mind that it is their own responsibility and as such keep their judgment and criticism. However other professionals could be involved, if they have participated in the building of expert systems. The different liabilities and the burden of proof are discussed according to some possible dysfunctions. In any case the final proof is inside the expert system by itself through re-computation of data. PMID:1482972
PVDaCS - A prototype knowledge-based expert system for certification of spacecraft data
NASA Technical Reports Server (NTRS)
Wharton, Cathleen; Shiroma, Patricia J.; Simmons, Karen E.
1989-01-01
On-line data management techniques to certify spacecraft information are mandated by increasing telemetry rates. Knowledge-based expert systems offer the ability to certify data electronically without the need for time-consuming human interaction. Issues of automatic certification are explored by designing a knowledge-based expert system to certify data from a scientific instrument, the Orbiter Ultraviolet Spectrometer, on an operating NASA planetary spacecraft, Pioneer Venus. The resulting rule-based system, called PVDaCS (Pioneer Venus Data Certification System), is a functional prototype demonstrating the concepts of a larger system design. A key element of the system design is the representation of an expert's knowledge through the usage of well ordered sequences. PVDaCS produces a certification value derived from expert knowledge and an analysis of the instrument's operation. Results of system performance are presented.
Validation and verification of expert systems
NASA Technical Reports Server (NTRS)
Gilstrap, Lewey
1991-01-01
Validation and verification (V&V) are procedures used to evaluate system structure or behavior with respect to a set of requirements. Although expert systems are often developed as a series of prototypes without requirements, it is not possible to perform V&V on any system for which requirements have not been prepared. In addition, there are special problems associated with the evaluation of expert systems that do not arise in the evaluation of conventional systems, such as verification of the completeness and accuracy of the knowledge base. The criticality of most NASA missions make it important to be able to certify the performance of the expert systems used to support these mission. Recommendations for the most appropriate method for integrating V&V into the Expert System Development Methodology (ESDM) and suggestions for the most suitable approaches for each stage of ESDM development are presented.
Toward the Development of Expert Assessment Systems.
ERIC Educational Resources Information Center
Hasselbring, Ted S.
1986-01-01
The potential application of "expert systems" to the diagnosis and assessment of special-needs children is examined and existing prototype systems are reviewed. The future of this artificial intelligence technology is discussed in relation to emerging development tools designed for the creation of expert systems by the lay public. (Author)
Compiling knowledge-based systems from KEE to Ada
NASA Technical Reports Server (NTRS)
Filman, Robert E.; Bock, Conrad; Feldman, Roy
1990-01-01
The dominant technology for developing AI applications is to work in a multi-mechanism, integrated, knowledge-based system (KBS) development environment. Unfortunately, systems developed in such environments are inappropriate for delivering many applications - most importantly, they carry the baggage of the entire Lisp environment and are not written in conventional languages. One resolution of this problem would be to compile applications from complex environments to conventional languages. Here the first efforts to develop a system for compiling KBS developed in KEE to Ada (trademark). This system is called KATYDID, for KEE/Ada Translation Yields Development Into Delivery. KATYDID includes early prototypes of a run-time KEE core (object-structure) library module for Ada, and translation mechanisms for knowledge structures, rules, and Lisp code to Ada. Using these tools, part of a simple expert system was compiled (not quite automatically) to run in a purely Ada environment. This experience has given us various insights on Ada as an artificial intelligence programming language, potential solutions of some of the engineering difficulties encountered in early work, and inspiration on future system development.
Second CLIPS Conference Proceedings, volume 1
NASA Technical Reports Server (NTRS)
Giarratano, Joseph (Editor); Culbert, Christopher J. (Editor)
1991-01-01
Topics covered at the 2nd CLIPS Conference held at the Johnson Space Center, September 23-25, 1991 are given. Topics include rule groupings, fault detection using expert systems, decision making using expert systems, knowledge representation, computer aided design and debugging expert systems.
The Error Reporting in the ATLAS TDAQ System
NASA Astrophysics Data System (ADS)
Kolos, Serguei; Kazarov, Andrei; Papaevgeniou, Lykourgos
2015-05-01
The ATLAS Error Reporting provides a service that allows experts and shift crew to track and address errors relating to the data taking components and applications. This service, called the Error Reporting Service (ERS), gives to software applications the opportunity to collect and send comprehensive data about run-time errors, to a place where it can be intercepted in real-time by any other system component. Other ATLAS online control and monitoring tools use the ERS as one of their main inputs to address system problems in a timely manner and to improve the quality of acquired data. The actual destination of the error messages depends solely on the run-time environment, in which the online applications are operating. When an application sends information to ERS, depending on the configuration, it may end up in a local file, a database, distributed middleware which can transport it to an expert system or display it to users. Thanks to the open framework design of ERS, new information destinations can be added at any moment without touching the reporting and receiving applications. The ERS Application Program Interface (API) is provided in three programming languages used in the ATLAS online environment: C++, Java and Python. All APIs use exceptions for error reporting but each of them exploits advanced features of a given language to simplify the end-user program writing. For example, as C++ lacks language support for exceptions, a number of macros have been designed to generate hierarchies of C++ exception classes at compile time. Using this approach a software developer can write a single line of code to generate a boilerplate code for a fully qualified C++ exception class declaration with arbitrary number of parameters and multiple constructors, which encapsulates all relevant static information about the given type of issues. When a corresponding error occurs at run time, the program just need to create an instance of that class passing relevant values to one of the available class constructors and send this instance to ERS. This paper presents the original design solutions exploited for the ERS implementation and describes how it was used during the first ATLAS run period. The cross-system error reporting standardization introduced by ERS was one of the key points for the successful implementation of automated mechanisms for online error recovery.
Regionalization by fuzzy expert system based approach optimized by genetic algorithm
NASA Astrophysics Data System (ADS)
Chavoshi, Sattar; Azmin Sulaiman, Wan Nor; Saghafian, Bahram; Bin Sulaiman, Md. Nasir; Manaf, Latifah Abd
2013-04-01
SummaryIn recent years soft computing methods are being increasingly used to model complex hydrologic processes. These methods can simulate the real life processes without prior knowledge of the exact relationship between their components. The principal aim of this paper is perform hydrological regionalization based on soft computing concepts in the southern strip of the Caspian Sea basin, north of Iran. The basin with an area of 42,400 sq. km has been affected by severe floods in recent years that caused damages to human life and properties. Although some 61 hydrometric stations and 31 weather stations with 44 years of observed data (1961-2005) are operated in the study area, previous flood studies in this region have been hampered by insufficient and/or reliable observed rainfall-runoff records. In order to investigate the homogeneity (h) of catchments and overcome incompatibility that may occur on boundaries of cluster groups, a fuzzy expert system (FES) approach is used which incorporates physical and climatic characteristics, as well as flood seasonality and geographic location. Genetic algorithm (GA) was employed to adjust parameters of FES and optimize the system. In order to achieve the objective, a MATLAB programming code was developed which considers the heterogeneity criteria of less than 1 (H < 1) as the satisfying criteria. The adopted approach was found superior to the conventional hydrologic regionalization methods in the region because it employs greater number of homogeneity parameters and produces lower values of heterogeneity criteria.
Personalized Risk Scoring for Critical Care Prognosis Using Mixtures of Gaussian Processes.
Alaa, Ahmed M; Yoon, Jinsung; Hu, Scott; van der Schaar, Mihaela
2018-01-01
In this paper, we develop a personalized real-time risk scoring algorithm that provides timely and granular assessments for the clinical acuity of ward patients based on their (temporal) lab tests and vital signs; the proposed risk scoring system ensures timely intensive care unit admissions for clinically deteriorating patients. The risk scoring system is based on the idea of sequential hypothesis testing under an uncertain time horizon. The system learns a set of latent patient subtypes from the offline electronic health record data, and trains a mixture of Gaussian Process experts, where each expert models the physiological data streams associated with a specific patient subtype. Transfer learning techniques are used to learn the relationship between a patient's latent subtype and her static admission information (e.g., age, gender, transfer status, ICD-9 codes, etc). Experiments conducted on data from a heterogeneous cohort of 6321 patients admitted to Ronald Reagan UCLA medical center show that our score significantly outperforms the currently deployed risk scores, such as the Rothman index, MEWS, APACHE, and SOFA scores, in terms of timeliness, true positive rate, and positive predictive value. Our results reflect the importance of adopting the concepts of personalized medicine in critical care settings; significant accuracy and timeliness gains can be achieved by accounting for the patients' heterogeneity. The proposed risk scoring methodology can confer huge clinical and social benefits on a massive number of critically ill inpatients who exhibit adverse outcomes including, but not limited to, cardiac arrests, respiratory arrests, and septic shocks.
Metzger, Marie-Hélène; Tvardik, Nastassia; Gicquel, Quentin; Bouvry, Côme; Poulet, Emmanuel; Potinet-Pagliaroli, Véronique
2017-06-01
The aim of this study was to determine whether an expert system based on automated processing of electronic health records (EHRs) could provide a more accurate estimate of the annual rate of emergency department (ED) visits for suicide attempts in France, as compared to the current national surveillance system based on manual coding by emergency practitioners. A feasibility study was conducted at Lyon University Hospital, using data for all ED patient visits in 2012. After automatic data extraction and pre-processing, including automatic coding of medical free-text through use of the Unified Medical Language System, seven different machine-learning methods were used to classify the reasons for ED visits into "suicide attempts" versus "other reasons". The performance of these different methods was compared by using the F-measure. In a test sample of 444 patients admitted to the ED in 2012 (98 suicide attempts, 48 cases of suicidal ideation, and 292 controls with no recorded non-fatal suicidal behaviour), the F-measure for automatic detection of suicide attempts ranged from 70.4% to 95.3%. The random forest and naïve Bayes methods performed best. This study demonstrates that machine-learning methods can improve the quality of epidemiological indicators as compared to current national surveillance of suicide attempts. Copyright © 2016 John Wiley & Sons, Ltd.
Parallel processing and expert systems
NASA Technical Reports Server (NTRS)
Lau, Sonie; Yan, Jerry C.
1991-01-01
Whether it be monitoring the thermal subsystem of Space Station Freedom, or controlling the navigation of the autonomous rover on Mars, NASA missions in the 1990s cannot enjoy an increased level of autonomy without the efficient implementation of expert systems. Merely increasing the computational speed of uniprocessors may not be able to guarantee that real-time demands are met for larger systems. Speedup via parallel processing must be pursued alongside the optimization of sequential implementations. Prototypes of parallel expert systems have been built at universities and industrial laboratories in the U.S. and Japan. The state-of-the-art research in progress related to parallel execution of expert systems is surveyed. The survey discusses multiprocessors for expert systems, parallel languages for symbolic computations, and mapping expert systems to multiprocessors. Results to date indicate that the parallelism achieved for these systems is small. The main reasons are (1) the body of knowledge applicable in any given situation and the amount of computation executed by each rule firing are small, (2) dividing the problem solving process into relatively independent partitions is difficult, and (3) implementation decisions that enable expert systems to be incrementally refined hamper compile-time optimization. In order to obtain greater speedups, data parallelism and application parallelism must be exploited.
Fuzzy logic based expert system for the treatment of mobile tooth.
Mago, Vijay Kumar; Mago, Anjali; Sharma, Poonam; Mago, Jagmohan
2011-01-01
The aim of this research work is to design an expert system to assist dentist in treating the mobile tooth. There is lack of consistency among dentists in choosing the treatment plan. Moreover, there is no expert system currently available to verify and support such decision making in dentistry. A Fuzzy Logic based expert system has been designed to accept imprecise and vague values of dental sign-symptoms related to mobile tooth and the system suggests treatment plan(s). The comparison of predictions made by the system with those of the dentist is conducted. Chi-square Test of homogeneity is conducted and it is found that the system is capable of predicting accurate results. With this system, dentist feels more confident while planning the treatment of mobile tooth as he can verify his decision with the expert system. The authors also argue that Fuzzy Logic provides an appropriate mechanism to handle imprecise values of dental domain.
NASA ground terminal communication equipment automated fault isolation expert systems
NASA Technical Reports Server (NTRS)
Tang, Y. K.; Wetzel, C. R.
1990-01-01
The prototype expert systems are described that diagnose the Distribution and Switching System I and II (DSS1 and DSS2), Statistical Multiplexers (SM), and Multiplexer and Demultiplexer systems (MDM) at the NASA Ground Terminal (NGT). A system level fault isolation expert system monitors the activities of a selected data stream, verifies that the fault exists in the NGT and identifies the faulty equipment. Equipment level fault isolation expert systems are invoked to isolate the fault to a Line Replaceable Unit (LRU) level. Input and sometimes output data stream activities for the equipment are available. The system level fault isolation expert system compares the equipment input and output status for a data stream and performs loopback tests (if necessary) to isolate the faulty equipment. The equipment level fault isolation system utilizes the process of elimination and/or the maintenance personnel's fault isolation experience stored in its knowledge base. The DSS1, DSS2 and SM fault isolation systems, using the knowledge of the current equipment configuration and the equipment circuitry issues a set of test connections according to the predefined rules. The faulty component or board can be identified by the expert system by analyzing the test results. The MDM fault isolation system correlates the failure symptoms with the faulty component based on maintenance personnel experience. The faulty component can be determined by knowing the failure symptoms. The DSS1, DSS2, SM, and MDM equipment simulators are implemented in PASCAL. The DSS1 fault isolation expert system was converted to C language from VP-Expert and integrated into the NGT automation software for offline switch diagnoses. Potentially, the NGT fault isolation algorithms can be used for the DSS1, SM, amd MDM located at Goddard Space Flight Center (GSFC).
Expert Systems for Libraries at SCIL [Small Computers in Libraries]'88.
ERIC Educational Resources Information Center
Kochtanek, Thomas R.; And Others
1988-01-01
Six brief papers on expert systems for libraries cover (1) a knowledge-based approach to database design; (2) getting started in expert systems; (3) using public domain software to develop a business reference system; (4) a music cataloging inquiry system; (5) linguistic analysis of reference transactions; and (6) a model of a reference librarian.…
Patient Health Goals Elicited During Home Care Admission: A Categorization.
Sockolow, Paulina; Radhakrishnan, Kavita; Chou, Edgar Y; Wojciechowicz, Christine
2017-11-01
Home care agencies are initiating "patient health goal elicitation" activities as part of home care admission planning. We categorized elicited goals and identified "clinically informative" goals at a home care agency. We examined patient goals that admitting clinicians documented in the point-of-care electronic health record; conducted content analysis on patient goal data to develop a coding scheme; grouped goal themes into codes; assigned codes to each goal; and identified goals that were in the patient voice. Of the 1,763 patient records, 16% lacked a goal; only 15 goals were in a patient's voice. Nurse and physician experts identified 12 of the 20 codes as clinically important accounting for 82% of goal occurrences. The most frequent goal documented was safety/falls (23%). Training and consistent communication of the intent and operationalization of patient goal elicitation may address the absence of patient voice and the less than universal recording of home care patients' goals.
Reference Standards, Judges, and Comparison Subjects
Hripcsak, George; Wilcox, Adam
2002-01-01
Medical informatics systems are often designed to perform at the level of human experts. Evaluation of the performance of these systems is often constrained by lack of reference standards, either because the appropriate response is not known or because no simple appropriate response exists. Even when performance can be assessed, it is not always clear whether the performance is sufficient or reasonable. These challenges can be addressed if an evaluator enlists the help of clinical domain experts. 1) The experts can carry out the same tasks as the system, and then their responses can be combined to generate a reference standard. 2)The experts can judge the appropriateness of system output directly. 3) The experts can serve as comparison subjects with which the system can be compared. These are separate roles that have different implications for study design, metrics, and issues of reliability and validity. Diagrams help delineate the roles of experts in complex study designs. PMID:11751799
Counseling, Artificial Intelligence, and Expert Systems.
ERIC Educational Resources Information Center
Illovsky, Michael E.
1994-01-01
Considers the use of artificial intelligence and expert systems in counseling. Limitations are explored; candidates for counseling versus those for expert systems are discussed; programming considerations are reviewed; and techniques for dealing with rational, nonrational, and irrational thoughts and feelings are described. (Contains 46…
1989-09-01
OGT, F1EPQRTJTL4, W" - 3^ n"r-- n *ON EXPERT SYSTEMS IN DESIGN, CONSTRUCTION AND’, IWAJNTENANCE-OF STRUCTURES Arockiasamy, Sunghoon Lee Clepartrhent...based expert system applications in the areas of structural design, design standards, and construction planning. This study will aid in the development...of a comprehensive expert system for tvical hydraulic structures. Funding for this report was provided by the US Army Engineer Waterways Experiment
NASA Technical Reports Server (NTRS)
Jaworski, Allan; Lavallee, David; Zoch, David
1987-01-01
The prototype demonstrates the feasibility of using Ada for expert systems and the implementation of an expert-friendly interface which supports knowledge entry. In the Ford LISP-Ada Connection (FLAC) system LISP and Ada are used in ways which complement their respective capabilities. Future investigation will concentrate on the enhancement of the expert knowledge entry/debugging interface and on the issues associated with multitasking and real-time expert systems implementation in Ada.
Benchmarking expert system tools
NASA Technical Reports Server (NTRS)
Riley, Gary
1988-01-01
As part of its evaluation of new technologies, the Artificial Intelligence Section of the Mission Planning and Analysis Div. at NASA-Johnson has made timing tests of several expert system building tools. Among the production systems tested were Automated Reasoning Tool, several versions of OPS5, and CLIPS (C Language Integrated Production System), an expert system builder developed by the AI section. Also included in the test were a Zetalisp version of the benchmark along with four versions of the benchmark written in Knowledge Engineering Environment, an object oriented, frame based expert system tool. The benchmarks used for testing are studied.
Semi-Automated Annotation of Biobank Data Using Standard Medical Terminologies in a Graph Database.
Hofer, Philipp; Neururer, Sabrina; Goebel, Georg
2016-01-01
Data describing biobank resources frequently contains unstructured free-text information or insufficient coding standards. (Bio-) medical ontologies like Orphanet Rare Diseases Ontology (ORDO) or the Human Disease Ontology (DOID) provide a high number of concepts, synonyms and entity relationship properties. Such standard terminologies increase quality and granularity of input data by adding comprehensive semantic background knowledge from validated entity relationships. Moreover, cross-references between terminology concepts facilitate data integration across databases using different coding standards. In order to encourage the use of standard terminologies, our aim is to identify and link relevant concepts with free-text diagnosis inputs within a biobank registry. Relevant concepts are selected automatically by lexical matching and SPARQL queries against a RDF triplestore. To ensure correctness of annotations, proposed concepts have to be confirmed by medical data administration experts before they are entered into the registry database. Relevant (bio-) medical terminologies describing diseases and phenotypes were identified and stored in a graph database which was tied to a local biobank registry. Concept recommendations during data input trigger a structured description of medical data and facilitate data linkage between heterogeneous systems.
Valente, Ana Rita S; Hall, Andreia; Alvelos, Helena; Leahy, Margaret; Jesus, Luis M T
2018-04-12
The appropriate use of language in context depends on the speaker's pragmatic language competencies. A coding system was used to develop a specific and adult-focused self-administered questionnaire to adults who stutter and adults who do not stutter, The Assessment of Language Use in Social Contexts for Adults, with three categories: precursors, basic exchanges, and extended literal/non-literal discourse. This paper presents the content validity, item analysis, reliability coefficients and evidences of construct validity of the instrument. Content validity analysis was based on a two-stage process: first, 11 pragmatic questionnaires were assessed to identify items that probe each pragmatic competency and to create the first version of the instrument; second, items were assessed qualitatively by an expert panel composed by adults who stutter and controls, and quantitatively and qualitatively by an expert panel composed by clinicians. A pilot study was conducted with five adults who stutter and five controls to analyse items and calculate reliability. Construct validity evidences were obtained using the hypothesized relationships method and factor analysis with 28 adults who stutter and 28 controls. Concerning content validity, the questionnaires assessed up to 13 pragmatic competencies. Qualitative and quantitative analysis revealed ambiguities in items construction. Disagreement between experts was solved through item modification. The pilot study showed that the instrument presented internal consistency and temporal stability. Significant differences between adults who stutter and controls and different response profiles revealed the instrument's underlying construct. The instrument is reliable and presented evidences of construct validity.
MCMEG: Simulations of both PDD and TPR for 6 MV LINAC photon beam using different MC codes
NASA Astrophysics Data System (ADS)
Fonseca, T. C. F.; Mendes, B. M.; Lacerda, M. A. S.; Silva, L. A. C.; Paixão, L.; Bastos, F. M.; Ramirez, J. V.; Junior, J. P. R.
2017-11-01
The Monte Carlo Modelling Expert Group (MCMEG) is an expert network specializing in Monte Carlo radiation transport and the modelling and simulation applied to the radiation protection and dosimetry research field. For the first inter-comparison task the group launched an exercise to model and simulate a 6 MV LINAC photon beam using the Monte Carlo codes available within their laboratories and validate their simulated results by comparing them with experimental measurements carried out in the National Cancer Institute (INCA) in Rio de Janeiro, Brazil. The experimental measurements were performed using an ionization chamber with calibration traceable to a Secondary Standard Dosimetry Laboratory (SSDL). The detector was immersed in a water phantom at different depths and was irradiated with a radiation field size of 10×10 cm2. This exposure setup was used to determine the dosimetric parameters Percentage Depth Dose (PDD) and Tissue Phantom Ratio (TPR). The validation process compares the MC calculated results to the experimental measured PDD20,10 and TPR20,10. Simulations were performed reproducing the experimental TPR20,10 quality index which provides a satisfactory description of both the PDD curve and the transverse profiles at the two depths measured. This paper reports in detail the modelling process using MCNPx, MCNP6, EGSnrc and Penelope Monte Carlo codes, the source and tally descriptions, the validation processes and the results.
Using support vector machines to detect medical fraud and abuse.
Francis, Charles; Pepper, Noah; Strong, Homer
2011-01-01
This paper examines the architecture and efficacy of Quash, an automated medical bill processing system capable of bill routing and abuse detection. Quash is designed to be used in conjunction with human auditors and a standard bill review software platform to provide a complete cost containment solution for medical claims. The primary contribution of Quash is to provide a real world speed up for medical fraud detection experts in their work. There will be a discussion of implementation details and preliminary experimental results. In this paper we are entirely focused on medical data and billing patterns that occur within the United States, though these results should be applicable to any financial transaction environment in which structured coding data can be mined.
Gangs in Schools. Breaking Up Is Hard To Do.
ERIC Educational Resources Information Center
National School Safety Center, Malibu, CA.
This handbook offers the latest information on gangs and practical advice on preventing or reducing gang encroachment in schools. Gang experts believe that establishing codes of conduct, diligent awareness of gang rivalries, prevention courses, and community and parental involvement can make an impact in keeping gangs away from campus. Chapter 1,…
78 FR 14912 - International Aviation Safety Assessment (IASA) Program Change
Federal Register 2010, 2011, 2012, 2013, 2014
2013-03-08
...; and Public Expectations of IASA Categories Removal of Inactive Countries Under the IASA program, the... can put a U.S. carrier code on its flights. Public Expectations of IASA Category Ratings Members of... by a team consisting of a team leader and at least one expert in operations, maintenance, and...
PRA and Risk Informed Analysis
DOE Office of Scientific and Technical Information (OSTI.GOV)
Bernsen, Sidney A.; Simonen, Fredric A.; Balkey, Kenneth R.
2006-01-01
The Boiler and Pressure Vessel Code (BPVC) of the American Society of Mechanical Engineers (ASME) has introduced a risk based approach into Section XI that covers Rules for Inservice Inspection of Nuclear Power Plant Components. The risk based approach requires application of the probabilistic risk assessments (PRA). Because no industry consensus standard existed for PRAs, ASME has developed a standard to evaluate the quality level of an available PRA needed to support a given risk based application. The paper describes the PRA standard, Section XI application of PRAs, and plans for broader applications of PRAs to other ASME nuclear codesmore » and standards. The paper addresses several specific topics of interest to Section XI. Important consideration are special methods (surrogate components) used to overcome the lack of PRA treatments of passive components in PRAs. The approach allows calculations of conditional core damage probabilities both for component failures that cause initiating events and failures in standby systems that decrease the availability of these systems. The paper relates the explicit risk based methods of the new Section XI code cases to the implicit consideration of risk used in the development of Section XI. Other topics include the needed interactions of ISI engineers, plant operating staff, PRA specialists, and members of expert panels that review the risk based programs.« less
MATLAB Stability and Control Toolbox Trim and Static Stability Module
NASA Technical Reports Server (NTRS)
Kenny, Sean P.; Crespo, Luis
2012-01-01
MATLAB Stability and Control Toolbox (MASCOT) utilizes geometric, aerodynamic, and inertial inputs to calculate air vehicle stability in a variety of critical flight conditions. The code is based on fundamental, non-linear equations of motion and is able to translate results into a qualitative, graphical scale useful to the non-expert. MASCOT was created to provide the conceptual aircraft designer accurate predictions of air vehicle stability and control characteristics. The code takes as input mass property data in the form of an inertia tensor, aerodynamic loading data, and propulsion (i.e. thrust) loading data. Using fundamental nonlinear equations of motion, MASCOT then calculates vehicle trim and static stability data for the desired flight condition(s). Available flight conditions include six horizontal and six landing rotation conditions with varying options for engine out, crosswind, and sideslip, plus three take-off rotation conditions. Results are displayed through a unique graphical interface developed to provide the non-stability and control expert conceptual design engineer a qualitative scale indicating whether the vehicle has acceptable, marginal, or unacceptable static stability characteristics. If desired, the user can also examine the detailed, quantitative results.
NASA Technical Reports Server (NTRS)
Barth, Tim; Zapata, Edgar; Benjamin, Perakath; Graul, Mike; Jones, Doug
2005-01-01
Portfolio Analysis Tool (PAT) is a Web-based, client/server computer program that helps managers of multiple projects funded by different customers to make decisions regarding investments in those projects. PAT facilitates analysis on a macroscopic level, without distraction by parochial concerns or tactical details of individual projects, so that managers decisions can reflect the broad strategy of their organization. PAT is accessible via almost any Web-browser software. Experts in specific projects can contribute to a broad database that managers can use in analyzing the costs and benefits of all projects, but do not have access for modifying criteria for analyzing projects: access for modifying criteria is limited to managers according to levels of administrative privilege. PAT affords flexibility for modifying criteria for particular "focus areas" so as to enable standardization of criteria among similar projects, thereby making it possible to improve assessments without need to rewrite computer code or to rehire experts, and thereby further reducing the cost of maintaining and upgrading computer code. Information in the PAT database and results of PAT analyses can be incorporated into a variety of ready-made or customizable tabular or graphical displays.
An expert system to manage the operation of the Space Shuttle's fuel cell cryogenic reactant tanks
NASA Technical Reports Server (NTRS)
Murphey, Amy Y.
1990-01-01
This paper describes a rule-based expert system to manage the operation of the Space Shuttle's cryogenic fuel system. Rules are based on standard fuel tank operating procedures described in the EECOM Console Handbook. The problem of configuring the operation of the Space Shuttle's fuel tanks is well-bounded and well defined. Moreover, the solution of this problem can be encoded in a knowledge-based system. Therefore, a rule-based expert system is the appropriate paradigm. Furthermore, the expert system could be used in coordination with power system simulation software to design operating procedures for specific missions.
Expert systems as applied to bridges and pavements : an overview.
DOT National Transportation Integrated Search
1986-01-01
Expert systems is a rapidly emerging new application of computers to aid decision makers in solving problems. This report gives an overview of what expert systems are and of what use they may be to a transportation department. The focus of the applic...
GenSAA: A tool for advancing satellite monitoring with graphical expert systems
NASA Technical Reports Server (NTRS)
Hughes, Peter M.; Luczak, Edward C.
1993-01-01
During numerous contacts with a satellite each day, spacecraft analysts must closely monitor real time data for combinations of telemetry parameter values, trends, and other indications that may signify a problem or failure. As satellites become more complex and the number of data items increases, this task is becoming increasingly difficult for humans to perform at acceptable performance levels. At the NASA Goddard Space Flight Center, fault-isolation expert systems have been developed to support data monitoring and fault detection tasks in satellite control centers. Based on the lessons learned during these initial efforts in expert system automation, a new domain-specific expert system development tool named the Generic Spacecraft Analyst Assistant (GenSAA) is being developed to facilitate the rapid development and reuse of real-time expert systems to serve as fault-isolation assistants for spacecraft analysts. Although initially domain-specific in nature, this powerful tool will support the development of highly graphical expert systems for data monitoring purposes throughout the space and commercial industry.
The Generic Spacecraft Analyst Assistant (gensaa): a Tool for Developing Graphical Expert Systems
NASA Technical Reports Server (NTRS)
Hughes, Peter M.
1993-01-01
During numerous contacts with a satellite each day, spacecraft analysts must closely monitor real-time data. The analysts must watch for combinations of telemetry parameter values, trends, and other indications that may signify a problem or failure. As the satellites become more complex and the number of data items increases, this task is becoming increasingly difficult for humans to perform at acceptable performance levels. At NASA GSFC, fault-isolation expert systems are in operation supporting this data monitoring task. Based on the lessons learned during these initial efforts in expert system automation, a new domain-specific expert system development tool named the Generic Spacecraft Analyst Assistant (GenSAA) is being developed to facilitate the rapid development and reuse of real-time expert systems to serve as fault-isolation assistants for spacecraft analysts. Although initially domain-specific in nature, this powerful tool will readily support the development of highly graphical expert systems for data monitoring purposes throughout the space and commercial industry.
NASA Technical Reports Server (NTRS)
Liebowitz, Jay
1986-01-01
At NASA Goddard, the role of the command management system (CMS) is to transform general requests for spacecraft opeerations into detailed operational plans to be uplinked to the spacecraft. The CMS is part of the NASA Data System which entails the downlink of science and engineering data from NASA near-earth satellites to the user, and the uplink of command and control data to the spacecraft. Presently, it takes one to three years, with meetings once or twice a week, to determine functional requirements for CMS software design. As an alternative approach to the present technique of developing CMS software functional requirements, an expert system prototype was developed to aid in this function. Specifically, the knowledge base was formulated through interactions with domain experts, and was then linked to an existing expert system application generator called 'Knowledge Engineering System (Version 1.3).' Knowledge base development focused on four major steps: (1) develop the problem-oriented attribute hierachy; (2) determine the knowledge management approach; (3) encode the knowledge base; and (4) validate, test, certify, and evaluate the knowledge base and the expert system prototype as a whole. Backcasting was accomplished for validating and testing the expert system prototype. Knowledge refinement, evaluation, and implementation procedures of the expert system prototype were then transacted.
Qpais: A Web-Based Expert System for Assistedidentification of Quarantine Stored Insect Pests
NASA Astrophysics Data System (ADS)
Huang, Han; Rajotte, Edwin G.; Li, Zhihong; Chen, Ke; Zhang, Shengfang
Stored insect pests can seriously depredate stored products causing worldwide economic losses. Pests enter countries traveling with transported goods. Inspection and Quarantine activities are essential to prevent the invasion and spread of pests. Identification of quarantine stored insect pests is an important component of the China's Inspection and Quarantine procedure, and it is necessary not only to identify whether the species captured is an invasive species, but determine control procedures for stored insect pests. With the development of information technologies, many expert systems that aid in the identification of agricultural pests have been developed. Expert systems for the identification of quarantine stored insect pests are rare and are mainly developed for stand-alone PCs. This paper describes the development of a web-based expert system for identification of quarantine stored insect pests as part of the China 11th Five-Year National Scientific and Technological Support Project (115 Project). Based on user needs, textual knowledge and images were gathered from the literature and expert interviews. ASP.NET, C# and SQL language were used to program the system. Improvement of identification efficiency and flexibility was achieved using a new inference method called characteristic-select-based spatial distance method. The expert system can assist identifying 150 species of quarantine stored insect pests and provide detailed information for each species. The expert system has also been evaluated using two steps: system testing and identification testing. With a 85% rate of correct identification and high efficiency, the system evaluation shows that this expert system can be used in identification work of quarantine stored insect pests.
The resource envelope as a basis for space station management system scheduling
NASA Technical Reports Server (NTRS)
Bush, Joy; Critchfield, Anna
1987-01-01
The Platform Management System (PMS) Resource Envelope Scheduling System (PRESS) expert system prototype developed for space station scheduling is described. The purpose of developing the prototype was too investigate the resource envelope concept in a practical scheduling application, using a commercially available expert system shell. PRESS is being developed on an IBM PC/AT using Teknowledge, Inc.'s M.1 expert system shell.
DEVELOPMENT AND TESTING OF FAULT-DIAGNOSIS ALGORITHMS FOR REACTOR PLANT SYSTEMS
DOE Office of Scientific and Technical Information (OSTI.GOV)
Grelle, Austin L.; Park, Young S.; Vilim, Richard B.
Argonne National Laboratory is further developing fault diagnosis algorithms for use by the operator of a nuclear plant to aid in improved monitoring of overall plant condition and performance. The objective is better management of plant upsets through more timely, informed decisions on control actions with the ultimate goal of improved plant safety, production, and cost management. Integration of these algorithms with visual aids for operators is taking place through a collaboration under the concept of an operator advisory system. This is a software entity whose purpose is to manage and distill the enormous amount of information an operator mustmore » process to understand the plant state, particularly in off-normal situations, and how the state trajectory will unfold in time. The fault diagnosis algorithms were exhaustively tested using computer simulations of twenty different faults introduced into the chemical and volume control system (CVCS) of a pressurized water reactor (PWR). The algorithms are unique in that each new application to a facility requires providing only the piping and instrumentation diagram (PID) and no other plant-specific information; a subject-matter expert is not needed to install and maintain each instance of an application. The testing approach followed accepted procedures for verifying and validating software. It was shown that the code satisfies its functional requirement which is to accept sensor information, identify process variable trends based on this sensor information, and then to return an accurate diagnosis based on chains of rules related to these trends. The validation and verification exercise made use of GPASS, a one-dimensional systems code, for simulating CVCS operation. Plant components were failed and the code generated the resulting plant response. Parametric studies with respect to the severity of the fault, the richness of the plant sensor set, and the accuracy of sensors were performed as part of the validation exercise. The background and overview of the software will be presented to give an overview of the approach. Following, the verification and validation effort using the GPASS code for simulation of plant transients including a sensitivity study on important parameters will be presented« less
NASA Astrophysics Data System (ADS)
Flanagan, S.; Schachter, J. M.; Schissel, D. P.
2001-10-01
A Data Analysis Monitoring (DAM) system has been developed to monitor between pulse physics analysis at the DIII-D National Fusion Facility. The system allows for rapid detection of discrepancies in diagnostic measurements or the results from physics analysis codes. This enables problems to be detected and possibly fixed between pulses as opposed to after the experimental run has concluded thus increasing the efficiency of experimental time. An example of a consistency check is comparing the stored energy from integrating the measured kinetic profiles to that calculated from magnetic measurements by EFIT. This new system also tracks the progress of MDSplus dispatching of software for data analysis and the loading of analyzed data into MDSplus. DAM uses a Java Servlet to receive messages, Clips to implement expert system logic, and displays its results to multiple web clients via HTML. If an error is detected by DAM, users can view more detailed information so that steps can be taken to eliminate the error for the next pulse. A demonstration of this system including a simulated DIII-D pulse cycle will be presented.
An expert system for diagnostics and estimation of steam turbine components condition
NASA Astrophysics Data System (ADS)
Murmansky, B. E.; Aronson, K. E.; Brodov, Yu. M.
2017-11-01
The report describes an expert system of probability type for diagnostics and state estimation of steam turbine technological subsystems components. The expert system is based on Bayes’ theorem and permits to troubleshoot the equipment components, using expert experience, when there is a lack of baseline information on the indicators of turbine operation. Within a unified approach the expert system solves the problems of diagnosing the flow steam path of the turbine, bearings, thermal expansion system, regulatory system, condensing unit, the systems of regenerative feed-water and hot water heating. The knowledge base of the expert system for turbine unit rotors and bearings contains a description of 34 defects and of 104 related diagnostic features that cause a change in its vibration state. The knowledge base for the condensing unit contains 12 hypotheses and 15 evidence (indications); the procedures are also designated for 20 state parameters estimation. Similar knowledge base containing the diagnostic features and faults hypotheses are formulated for other technological subsystems of turbine unit. With the necessary initial information available a number of problems can be solved within the expert system for various technological subsystems of steam turbine unit: for steam flow path it is the correlation and regression analysis of multifactor relationship between the vibration parameters variations and the regime parameters; for system of thermal expansions it is the evaluation of force acting on the longitudinal keys depending on the temperature state of the turbine cylinder; for condensing unit it is the evaluation of separate effect of the heat exchange surface contamination and of the presence of air in condenser steam space on condenser thermal efficiency performance, as well as the evaluation of term for condenser cleaning and for tube system replacement and so forth. With a lack of initial information the expert system enables to formulate a diagnosis, calculating the probability of faults hypotheses, given the degree of the expert confidence in estimation of turbine components operation parameters.
Proceedings of the international conference on cybernetics and societ
DOE Office of Scientific and Technical Information (OSTI.GOV)
Not Available
1985-01-01
This book presents the papers given at a conference on artificial intelligence, expert systems and knowledge bases. Topics considered at the conference included automating expert system development, modeling expert systems, causal maps, data covariances, robot vision, image processing, multiprocessors, parallel processing, VLSI structures, man-machine systems, human factors engineering, cognitive decision analysis, natural language, computerized control systems, and cybernetics.
Bonneville Power Administration Communication Alarm Processor expert system:
DOE Office of Scientific and Technical Information (OSTI.GOV)
Goeltz, R.; Purucker, S.; Tonn, B.
This report describes the Communications Alarm Processor (CAP), a prototype expert system developed for the Bonneville Power Administration by Oak Ridge National Laboratory. The system is designed to receive and diagnose alarms from Bonneville's Microwave Communications System (MCS). The prototype encompasses one of seven branches of the communications network and a subset of alarm systems and alarm types from each system. The expert system employs a backward chaining approach to diagnosing alarms. Alarms are fed into the expert system directly from the communication system via RS232 ports and sophisticated alarm filtering and mailbox software. Alarm diagnoses are presented to operatorsmore » for their review and concurrence before the diagnoses are archived. Statistical software is incorporated to allow analysis of archived data for report generation and maintenance studies. The delivered system resides on a Digital Equipment Corporation VAX 3200 workstation and utilizes Nexpert Object and SAS for the expert system and statistical analysis, respectively. 11 refs., 23 figs., 7 tabs.« less
MixDroid: A multi-features and multi-classifiers bagging system for Android malware detection
NASA Astrophysics Data System (ADS)
Huang, Weiqing; Hou, Erhang; Zheng, Liang; Feng, Weimiao
2018-05-01
In the past decade, Android platform has rapidly taken over the mobile market for its superior convenience and open source characteristics. However, with the popularity of Android, malwares targeting on Android devices are increasing rapidly, while the conventional rule-based and expert-experienced approaches are no longer able to handle such explosive growth. In this paper, combining with the theory of natural language processing and machine learning, we not only implement the basic feature extraction of permission application features, but also propose two innovative schemes of feature extraction: Dalvik opcode features and malicious code image, and implement an automatic Android malware detection system MixDroid which is based on multi-features and multi-classifiers. According to our experiment results on 20,000 Android applications, detection accuracy of MixDroid is 98.1%, which proves our schemes' effectiveness in Android malware detection.
Using PVM to host CLIPS in distributed environments
NASA Technical Reports Server (NTRS)
Myers, Leonard; Pohl, Kym
1994-01-01
It is relatively easy to enhance CLIPS (C Language Integrated Production System) to support multiple expert systems running in a distributed environment with heterogeneous machines. The task is minimized by using the PVM (Parallel Virtual Machine) code from Oak Ridge Labs to provide the distributed utility. PVM is a library of C and FORTRAN subprograms that supports distributive computing on many different UNIX platforms. A PVM deamon is easily installed on each CPU that enters the virtual machine environment. Any user with rsh or rexec access to a machine can use the one PVM deamon to obtain a generous set of distributed facilities. The ready availability of both CLIPS and PVM makes the combination of software particularly attractive for budget conscious experimentation of heterogeneous distributive computing with multiple CLIPS executables. This paper presents a design that is sufficient to provide essential message passing functions in CLIPS and enable the full range of PVM facilities.
Calderon, Karynna; Dadisman, S.V.; Kindinger, J.L.; Flocks, J.G.; Wiese, D.S.; Kulp, Mark; Penland, Shea; Britsch, L.D.; Brooks, G.R.
2003-01-01
This archive consists of two-dimensional marine seismic reflection profile data collected in the Barataria Basin of southern Louisiana. These data were acquired in May, June, and July of 2000 aboard the R/V G.K. Gilbert. Included here are data in a variety of formats including binary, American Standard Code for Information Interchange (ASCII), Hyper-Text Markup Language (HTML), shapefiles, and Graphics Interchange Format (GIF) and Joint Photographic Experts Group (JPEG) images. Binary data are in Society of Exploration Geophysicists (SEG) SEG-Y format and may be downloaded for further processing or display. Reference maps and GIF images of the profiles may be viewed with a web browser. The Geographic Information Systems (GIS) information provided here is compatible with Environmental Systems Research Institute (ESRI) GIS software.