Sample records for machine pvm software

  1. Using PVM to host CLIPS in distributed environments

    NASA Technical Reports Server (NTRS)

    Myers, Leonard; Pohl, Kym

    1994-01-01

    It is relatively easy to enhance CLIPS (C Language Integrated Production System) to support multiple expert systems running in a distributed environment with heterogeneous machines. The task is minimized by using the PVM (Parallel Virtual Machine) code from Oak Ridge Labs to provide the distributed utility. PVM is a library of C and FORTRAN subprograms that supports distributive computing on many different UNIX platforms. A PVM deamon is easily installed on each CPU that enters the virtual machine environment. Any user with rsh or rexec access to a machine can use the one PVM deamon to obtain a generous set of distributed facilities. The ready availability of both CLIPS and PVM makes the combination of software particularly attractive for budget conscious experimentation of heterogeneous distributive computing with multiple CLIPS executables. This paper presents a design that is sufficient to provide essential message passing functions in CLIPS and enable the full range of PVM facilities.

  2. PVM Wrapper

    NASA Technical Reports Server (NTRS)

    Katz, Daniel

    2004-01-01

    PVM Wrapper is a software library that makes it possible for code that utilizes the Parallel Virtual Machine (PVM) software library to run using the message-passing interface (MPI) software library, without needing to rewrite the entire code. PVM and MPI are the two most common software libraries used for applications that involve passing of messages among parallel computers. Since about 1996, MPI has been the de facto standard. Codes written when PVM was popular often feature patterns of {"initsend," "pack," "send"} and {"receive," "unpack"} calls. In many cases, these calls are not contiguous and one set of calls may even exist over multiple subroutines. These characteristics make it difficult to obtain equivalent functionality via a single MPI "send" call. Because PVM Wrapper is written to run with MPI- 1.2, some PVM functions are not permitted and must be replaced - a task that requires some programming expertise. The "pvm_spawn" and "pvm_parent" function calls are not replaced, but a programmer can use "mpirun" and knowledge of the ranks of parent and child tasks with supplied macroinstructions to enable execution of codes that use "pvm_spawn" and "pvm_parent."

  3. Intel NX to PVM 3.2 message passing conversion library

    NASA Technical Reports Server (NTRS)

    Arthur, Trey; Nelson, Michael L.

    1993-01-01

    NASA Langley Research Center has developed a library that allows Intel NX message passing codes to be executed under the more popular and widely supported Parallel Virtual Machine (PVM) message passing library. PVM was developed at Oak Ridge National Labs and has become the defacto standard for message passing. This library will allow the many programs that were developed on the Intel iPSC/860 or Intel Paragon in a Single Program Multiple Data (SPMD) design to be ported to the numerous architectures that PVM (version 3.2) supports. Also, the library adds global operations capability to PVM. A familiarity with Intel NX and PVM message passing is assumed.

  4. Efficient Process Migration for Parallel Processing on Non-Dedicated Networks of Workstations

    NASA Technical Reports Server (NTRS)

    Chanchio, Kasidit; Sun, Xian-He

    1996-01-01

    This paper presents the design and preliminary implementation of MpPVM, a software system that supports process migration for PVM application programs in a non-dedicated heterogeneous computing environment. New concepts of migration point as well as migration point analysis and necessary data analysis are introduced. In MpPVM, process migrations occur only at previously inserted migration points. Migration point analysis determines appropriate locations to insert migration points; whereas, necessary data analysis provides a minimum set of variables to be transferred at each migration pint. A new methodology to perform reliable point-to-point data communications in a migration environment is also discussed. Finally, a preliminary implementation of MpPVM and its experimental results are presented, showing the correctness and promising performance of our process migration mechanism in a scalable non-dedicated heterogeneous computing environment. While MpPVM is developed on top of PVM, the process migration methodology introduced in this study is general and can be applied to any distributed software environment.

  5. Dust Dynamics in Protoplanetary Disks: Parallel Computing with PVM

    NASA Astrophysics Data System (ADS)

    de La Fuente Marcos, Carlos; Barge, Pierre; de La Fuente Marcos, Raúl

    2002-03-01

    We describe a parallel version of our high-order-accuracy particle-mesh code for the simulation of collisionless protoplanetary disks. We use this code to carry out a massively parallel, two-dimensional, time-dependent, numerical simulation, which includes dust particles, to study the potential role of large-scale, gaseous vortices in protoplanetary disks. This noncollisional problem is easy to parallelize on message-passing multicomputer architectures. We performed the simulations on a cache-coherent nonuniform memory access Origin 2000 machine, using both the parallel virtual machine (PVM) and message-passing interface (MPI) message-passing libraries. Our performance analysis suggests that, for our problem, PVM is about 25% faster than MPI. Using PVM and MPI made it possible to reduce CPU time and increase code performance. This allows for simulations with a large number of particles (N ~ 105-106) in reasonable CPU times. The performances of our implementation of the pa! rallel code on an Origin 2000 supercomputer are presented and discussed. They exhibit very good speedup behavior and low load unbalancing. Our results confirm that giant gaseous vortices can play a dominant role in giant planet formation.

  6. Scalable and reusable emulator for evaluating the performance of SS7 networks

    NASA Astrophysics Data System (ADS)

    Lazar, Aurel A.; Tseng, Kent H.; Lim, Koon Seng; Choe, Winston

    1994-04-01

    A scalable and reusable emulator was designed and implemented for studying the behavior of SS7 networks. The emulator design was largely based on public domain software. It was developed on top of an environment supported by PVM, the Parallel Virtual Machine, and managed by OSIMIS-the OSI Management Information Service platform. The emulator runs on top of a commercially available ATM LAN interconnecting engineering workstations. As a case study for evaluating the emulator, the behavior of the Singapore National SS7 Network under fault and unbalanced loading conditions was investigated.

  7. Distributed multitasking ITS with PVM

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Fan, W.C.; Halbleib, J.A. Sr.

    1995-12-31

    Advances in computer hardware and communication software have made it possible to perform parallel-processing computing on a collection of desktop workstations. For many applications, multitasking on a cluster of high-performance workstations has achieved performance comparable to or better than that on a traditional supercomputer. From the point of view of cost-effectiveness, it also allows users to exploit available but unused computational resources and thus achieve a higher performance-to-cost ratio. Monte Carlo calculations are inherently parallelizable because the individual particle trajectories can be generated independently with minimum need for interprocessor communication. Furthermore, the number of particle histories that can be generatedmore » in a given amount of wall-clock time is nearly proportional to the number of processors in the cluster. This is an important fact because the inherent statistical uncertainty in any Monte Carlo result decreases as the number of histories increases. For these reasons, researchers have expended considerable effort to take advantage of different parallel architectures for a variety of Monte Carlo radiation transport codes, often with excellent results. The initial interest in this work was sparked by the multitasking capability of the MCNP code on a cluster of workstations using the Parallel Virtual Machine (PVM) software. On a 16-machine IBM RS/6000 cluster, it has been demonstrated that MCNP runs ten times as fast as on a single-processor CRAY YMP. In this paper, we summarize the implementation of a similar multitasking capability for the coupled electronphoton transport code system, the Integrated TIGER Series (ITS), and the evaluation of two load-balancing schemes for homogeneous and heterogeneous networks.« less

  8. Distributed multitasking ITS with PVM

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Fan, W.C.; Halbleib, J.A. Sr.

    1995-02-01

    Advances of computer hardware and communication software have made it possible to perform parallel-processing computing on a collection of desktop workstations. For many applications, multitasking on a cluster of high-performance workstations has achieved performance comparable or better than that on a traditional supercomputer. From the point of view of cost-effectiveness, it also allows users to exploit available but unused computational resources, and thus achieve a higher performance-to-cost ratio. Monte Carlo calculations are inherently parallelizable because the individual particle trajectories can be generated independently with minimum need for interprocessor communication. Furthermore, the number of particle histories that can be generated inmore » a given amount of wall-clock time is nearly proportional to the number of processors in the cluster. This is an important fact because the inherent statistical uncertainty in any Monte Carlo result decreases as the number of histories increases. For these reasons, researchers have expended considerable effort to take advantage of different parallel architectures for a variety of Monte Carlo radiation transport codes, often with excellent results. The initial interest in this work was sparked by the multitasking capability of MCNP on a cluster of workstations using the Parallel Virtual Machine (PVM) software. On a 16-machine IBM RS/6000 cluster, it has been demonstrated that MCNP runs ten times as fast as on a single-processor CRAY YMP. In this paper, we summarize the implementation of a similar multitasking capability for the coupled electron/photon transport code system, the Integrated TIGER Series (ITS), and the evaluation of two load balancing schemes for homogeneous and heterogeneous networks.« less

  9. Performance Evaluation in Network-Based Parallel Computing

    NASA Technical Reports Server (NTRS)

    Dezhgosha, Kamyar

    1996-01-01

    Network-based parallel computing is emerging as a cost-effective alternative for solving many problems which require use of supercomputers or massively parallel computers. The primary objective of this project has been to conduct experimental research on performance evaluation for clustered parallel computing. First, a testbed was established by augmenting our existing SUNSPARCs' network with PVM (Parallel Virtual Machine) which is a software system for linking clusters of machines. Second, a set of three basic applications were selected. The applications consist of a parallel search, a parallel sort, a parallel matrix multiplication. These application programs were implemented in C programming language under PVM. Third, we conducted performance evaluation under various configurations and problem sizes. Alternative parallel computing models and workload allocations for application programs were explored. The performance metric was limited to elapsed time or response time which in the context of parallel computing can be expressed in terms of speedup. The results reveal that the overhead of communication latency between processes in many cases is the restricting factor to performance. That is, coarse-grain parallelism which requires less frequent communication between processes will result in higher performance in network-based computing. Finally, we are in the final stages of installing an Asynchronous Transfer Mode (ATM) switch and four ATM interfaces (each 155 Mbps) which will allow us to extend our study to newer applications, performance metrics, and configurations.

  10. Automated Instrumentation, Monitoring and Visualization of PVM Programs Using AIMS

    NASA Technical Reports Server (NTRS)

    Mehra, Pankaj; VanVoorst, Brian; Yan, Jerry; Lum, Henry, Jr. (Technical Monitor)

    1994-01-01

    We present views and analysis of the execution of several PVM (Parallel Virtual Machine) codes for Computational Fluid Dynamics on a networks of Sparcstations, including: (1) NAS Parallel Benchmarks CG and MG; (2) a multi-partitioning algorithm for NAS Parallel Benchmark SP; and (3) an overset grid flowsolver. These views and analysis were obtained using our Automated Instrumentation and Monitoring System (AIMS) version 3.0, a toolkit for debugging the performance of PVM programs. We will describe the architecture, operation and application of AIMS. The AIMS toolkit contains: (1) Xinstrument, which can automatically instrument various computational and communication constructs in message-passing parallel programs; (2) Monitor, a library of runtime trace-collection routines; (3) VK (Visual Kernel), an execution-animation tool with source-code clickback; and (4) Tally, a tool for statistical analysis of execution profiles. Currently, Xinstrument can handle C and Fortran 77 programs using PVM 3.2.x; Monitor has been implemented and tested on Sun 4 systems running SunOS 4.1.2; and VK uses XIIR5 and Motif 1.2. Data and views obtained using AIMS clearly illustrate several characteristic features of executing parallel programs on networked workstations: (1) the impact of long message latencies; (2) the impact of multiprogramming overheads and associated load imbalance; (3) cache and virtual-memory effects; and (4) significant skews between workstation clocks. Interestingly, AIMS can compensate for constant skew (zero drift) by calibrating the skew between a parent and its spawned children. In addition, AIMS' skew-compensation algorithm can adjust timestamps in a way that eliminates physically impossible communications (e.g., messages going backwards in time). Our current efforts are directed toward creating new views to explain the observed performance of PVM programs. Some of the features planned for the near future include: (1) ConfigView, showing the physical topology of the virtual machine, inferred using specially formatted IP (Internet Protocol) packets: and (2) LoadView, synchronous animation of PVM-program execution and resource-utilization patterns.

  11. A Systems Engineering Framework for Implementing a Security and Critical Patch Management Process in Diverse Environments (Academic Departments' Workstations)

    ERIC Educational Resources Information Center

    Mohammadi, Hadi

    2014-01-01

    Use of the Patch Vulnerability Management (PVM) process should be seriously considered for any networked computing system. The PVM process prevents the operating system (OS) and software applications from being attacked due to security vulnerabilities, which lead to system failures and critical data leakage. The purpose of this research is to…

  12. Prototype Vector Machine for Large Scale Semi-Supervised Learning

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Zhang, Kai; Kwok, James T.; Parvin, Bahram

    2009-04-29

    Practicaldataminingrarelyfalls exactlyinto the supervisedlearning scenario. Rather, the growing amount of unlabeled data poses a big challenge to large-scale semi-supervised learning (SSL). We note that the computationalintensivenessofgraph-based SSLarises largely from the manifold or graph regularization, which in turn lead to large models that are dificult to handle. To alleviate this, we proposed the prototype vector machine (PVM), a highlyscalable,graph-based algorithm for large-scale SSL. Our key innovation is the use of"prototypes vectors" for effcient approximation on both the graph-based regularizer and model representation. The choice of prototypes are grounded upon two important criteria: they not only perform effective low-rank approximation of themore » kernel matrix, but also span a model suffering the minimum information loss compared with the complete model. We demonstrate encouraging performance and appealing scaling properties of the PVM on a number of machine learning benchmark data sets.« less

  13. DOE Office of Scientific and Technical Information (OSTI.GOV)

    Painter, J.; McCormick, P.; Krogh, M.

    This paper presents the ACL (Advanced Computing Lab) Message Passing Library. It is a high throughput, low latency communications library, based on Thinking Machines Corp.`s CMMD, upon which message passing applications can be built. The library has been implemented on the Cray T3D, Thinking Machines CM-5, SGI workstations, and on top of PVM.

  14. A Systems Engineering Framework for Implementing a Security and Critical Patch Management Process in Diverse Environments (Academic Departments' Workstations)

    NASA Astrophysics Data System (ADS)

    Mohammadi, Hadi

    Use of the Patch Vulnerability Management (PVM) process should be seriously considered for any networked computing system. The PVM process prevents the operating system (OS) and software applications from being attacked due to security vulnerabilities, which lead to system failures and critical data leakage. The purpose of this research is to create and design a Security and Critical Patch Management Process (SCPMP) framework based on Systems Engineering (SE) principles. This framework will assist Information Technology Department Staff (ITDS) to reduce IT operating time and costs and mitigate the risk of security and vulnerability attacks. Further, this study evaluates implementation of the SCPMP in the networked computing systems of an academic environment in order to: 1. Meet patch management requirements by applying SE principles. 2. Reduce the cost of IT operations and PVM cycles. 3. Improve the current PVM methodologies to prevent networked computing systems from becoming the targets of security vulnerability attacks. 4. Embed a Maintenance Optimization Tool (MOT) in the proposed framework. The MOT allows IT managers to make the most practicable choice of methods for deploying and installing released patches and vulnerability remediation. In recent years, there has been a variety of frameworks for security practices in every networked computing system to protect computer workstations from becoming compromised or vulnerable to security attacks, which can expose important information and critical data. I have developed a new mechanism for implementing PVM for maximizing security-vulnerability maintenance, protecting OS and software packages, and minimizing SCPMP cost. To increase computing system security in any diverse environment, particularly in academia, one must apply SCPMP. I propose an optimal maintenance policy that will allow ITDS to measure and estimate the variation of PVM cycles based on their department's requirements. My results demonstrate that MOT optimizes the process of implementing SCPMP in academic workstations.

  15. Parallel computation and the Basis system

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Smith, G.R.

    1992-12-16

    A software package has been written that can facilitate efforts to develop powerful, flexible, and easy-to-use programs that can run in single-processor, massively parallel, and distributed computing environments. Particular attention has been given to the difficulties posed by a program consisting of many science packages that represent subsystems of a complicated, coupled system. Methods have been found to maintain independence of the packages by hiding data structures without increasing the communication costs in a parallel computing environment. Concepts developed in this work are demonstrated by a prototype program that uses library routines from two existing software systems, Basis and Parallelmore » Virtual Machine (PVM). Most of the details of these libraries have been encapsulated in routines and macros that could be rewritten for alternative libraries that possess certain minimum capabilities. The prototype software uses a flexible master-and-slaves paradigm for parallel computation and supports domain decomposition with message passing for partitioning work among slaves. Facilities are provided for accessing variables that are distributed among the memories of slaves assigned to subdomains. The software is named PROTOPAR.« less

  16. Parallel computation and the basis system

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Smith, G.R.

    1993-05-01

    A software package has been written that can facilitate efforts to develop powerful, flexible, and easy-to use programs that can run in single-processor, massively parallel, and distributed computing environments. Particular attention has been given to the difficulties posed by a program consisting of many science packages that represent subsystems of a complicated, coupled system. Methods have been found to maintain independence of the packages by hiding data structures without increasing the communications costs in a parallel computing environment. Concepts developed in this work are demonstrated by a prototype program that uses library routines from two existing software systems, Basis andmore » Parallel Virtual Machine (PVM). Most of the details of these libraries have been encapsulated in routines and macros that could be rewritten for alternative libraries that possess certain minimum capabilities. The prototype software uses a flexible master-and-slaves paradigm for parallel computation and supports domain decomposition with message passing for partitioning work among slaves. Facilities are provided for accessing variables that are distributed among the memories of slaves assigned to subdomains. The software is named PROTOPAR.« less

  17. Research in Parallel Algorithms and Software for Computational Aerosciences

    NASA Technical Reports Server (NTRS)

    Domel, Neal D.

    1996-01-01

    Phase I is complete for the development of a Computational Fluid Dynamics parallel code with automatic grid generation and adaptation for the Euler analysis of flow over complex geometries. SPLITFLOW, an unstructured Cartesian grid code developed at Lockheed Martin Tactical Aircraft Systems, has been modified for a distributed memory/massively parallel computing environment. The parallel code is operational on an SGI network, Cray J90 and C90 vector machines, SGI Power Challenge, and Cray T3D and IBM SP2 massively parallel machines. Parallel Virtual Machine (PVM) is the message passing protocol for portability to various architectures. A domain decomposition technique was developed which enforces dynamic load balancing to improve solution speed and memory requirements. A host/node algorithm distributes the tasks. The solver parallelizes very well, and scales with the number of processors. Partially parallelized and non-parallelized tasks consume most of the wall clock time in a very fine grain environment. Timing comparisons on a Cray C90 demonstrate that Parallel SPLITFLOW runs 2.4 times faster on 8 processors than its non-parallel counterpart autotasked over 8 processors.

  18. Research in Parallel Algorithms and Software for Computational Aerosciences

    NASA Technical Reports Server (NTRS)

    Domel, Neal D.

    1996-01-01

    Phase 1 is complete for the development of a computational fluid dynamics CFD) parallel code with automatic grid generation and adaptation for the Euler analysis of flow over complex geometries. SPLITFLOW, an unstructured Cartesian grid code developed at Lockheed Martin Tactical Aircraft Systems, has been modified for a distributed memory/massively parallel computing environment. The parallel code is operational on an SGI network, Cray J90 and C90 vector machines, SGI Power Challenge, and Cray T3D and IBM SP2 massively parallel machines. Parallel Virtual Machine (PVM) is the message passing protocol for portability to various architectures. A domain decomposition technique was developed which enforces dynamic load balancing to improve solution speed and memory requirements. A host/node algorithm distributes the tasks. The solver parallelizes very well, and scales with the number of processors. Partially parallelized and non-parallelized tasks consume most of the wall clock time in a very fine grain environment. Timing comparisons on a Cray C90 demonstrate that Parallel SPLITFLOW runs 2.4 times faster on 8 processors than its non-parallel counterpart autotasked over 8 processors.

  19. Scalability and Portability of Two Parallel Implementations of ADI

    NASA Technical Reports Server (NTRS)

    Phung, Thanh; VanderWijngaart, Rob F.

    1994-01-01

    Two domain decompositions for the implementation of the NAS Scalar Penta-diagonal Parallel Benchmark on MIMD systems are investigated, namely transposition and multi-partitioning. Hardware platforms considered are the Intel iPSC/860 and Paragon XP/S-15, and clusters of SGI workstations on ethernet, communicating through PVM. It is found that the multi-partitioning strategy offers the kind of coarse granularity that allows scaling up to hundreds of processors on a massively parallel machine. Moreover, efficiency is retained when the code is ported verbatim (save message passing syntax) to a PVM environment on a modest size cluster of workstations.

  20. Performance Evaluation of Communication Software Systems for Distributed Computing

    NASA Technical Reports Server (NTRS)

    Fatoohi, Rod

    1996-01-01

    In recent years there has been an increasing interest in object-oriented distributed computing since it is better quipped to deal with complex systems while providing extensibility, maintainability, and reusability. At the same time, several new high-speed network technologies have emerged for local and wide area networks. However, the performance of networking software is not improving as fast as the networking hardware and the workstation microprocessors. This paper gives an overview and evaluates the performance of the Common Object Request Broker Architecture (CORBA) standard in a distributed computing environment at NASA Ames Research Center. The environment consists of two testbeds of SGI workstations connected by four networks: Ethernet, FDDI, HiPPI, and ATM. The performance results for three communication software systems are presented, analyzed and compared. These systems are: BSD socket programming interface, IONA's Orbix, an implementation of the CORBA specification, and the PVM message passing library. The results show that high-level communication interfaces, such as CORBA and PVM, can achieve reasonable performance under certain conditions.

  1. Automated Instrumentation, Monitoring and Visualization of PVM Programs Using AIMS

    NASA Technical Reports Server (NTRS)

    Mehra, Pankaj; VanVoorst, Brian; Yan, Jerry; Tucker, Deanne (Technical Monitor)

    1994-01-01

    We present views and analysis of the execution of several PVM codes for Computational Fluid Dynamics on a network of Sparcstations, including (a) NAS Parallel benchmarks CG and MG (White, Alund and Sunderam 1993); (b) a multi-partitioning algorithm for NAS Parallel Benchmark SP (Wijngaart 1993); and (c) an overset grid flowsolver (Smith 1993). These views and analysis were obtained using our Automated Instrumentation and Monitoring System (AIMS) version 3.0, a toolkit for debugging the performance of PVM programs. We will describe the architecture, operation and application of AIMS. The AIMS toolkit contains (a) Xinstrument, which can automatically instrument various computational and communication constructs in message-passing parallel programs; (b) Monitor, a library of run-time trace-collection routines; (c) VK (Visual Kernel), an execution-animation tool with source-code clickback; and (d) Tally, a tool for statistical analysis of execution profiles. Currently, Xinstrument can handle C and Fortran77 programs using PVM 3.2.x; Monitor has been implemented and tested on Sun 4 systems running SunOS 4.1.2; and VK uses X11R5 and Motif 1.2. Data and views obtained using AIMS clearly illustrate several characteristic features of executing parallel programs on networked workstations: (a) the impact of long message latencies; (b) the impact of multiprogramming overheads and associated load imbalance; (c) cache and virtual-memory effects; and (4significant skews between workstation clocks. Interestingly, AIMS can compensate for constant skew (zero drift) by calibrating the skew between a parent and its spawned children. In addition, AIMS' skew-compensation algorithm can adjust timestamps in a way that eliminates physically impossible communications (e.g., messages going backwards in time). Our current efforts are directed toward creating new views to explain the observed performance of PVM programs. Some of the features planned for the near future include: (a) ConfigView, showing the physical topology of the virtual machine, inferred using specially formatted IP (Internet Protocol) packets; and (b) LoadView, synchronous animation of PVM-program execution and resource-utilization patterns.

  2. Porting Gravitational Wave Signal Extraction to Parallel Virtual Machine (PVM)

    NASA Technical Reports Server (NTRS)

    Thirumalainambi, Rajkumar; Thompson, David E.; Redmon, Jeffery

    2009-01-01

    Laser Interferometer Space Antenna (LISA) is a planned NASA-ESA mission to be launched around 2012. The Gravitational Wave detection is fundamentally the determination of frequency, source parameters, and waveform amplitude derived in a specific order from the interferometric time-series of the rotating LISA spacecrafts. The LISA Science Team has developed a Mock LISA Data Challenge intended to promote the testing of complicated nested search algorithms to detect the 100-1 millihertz frequency signals at amplitudes of 10E-21. However, it has become clear that, sequential search of the parameters is very time consuming and ultra-sensitive; hence, a new strategy has been developed. Parallelization of existing sequential search algorithms of Gravitational Wave signal identification consists of decomposing sequential search loops, beginning with outermost loops and working inward. In this process, the main challenge is to detect interdependencies among loops and partitioning the loops so as to preserve concurrency. Existing parallel programs are based upon either shared memory or distributed memory paradigms. In PVM, master and node programs are used to execute parallelization and process spawning. The PVM can handle process management and process addressing schemes using a virtual machine configuration. The task scheduling and the messaging and signaling can be implemented efficiently for the LISA Gravitational Wave search process using a master and 6 nodes. This approach is accomplished using a server that is available at NASA Ames Research Center, and has been dedicated to the LISA Data Challenge Competition. Historically, gravitational wave and source identification parameters have taken around 7 days in this dedicated single thread Linux based server. Using PVM approach, the parameter extraction problem can be reduced to within a day. The low frequency computation and a proxy signal-to-noise ratio are calculated in separate nodes that are controlled by the master using message and vector of data passing. The message passing among nodes follows a pattern of synchronous and asynchronous send-and-receive protocols. The communication model and the message buffers are allocated dynamically to address rapid search of gravitational wave source information in the Mock LISA data sets.

  3. High-energy physics software parallelization using database techniques

    NASA Astrophysics Data System (ADS)

    Argante, E.; van der Stok, P. D. V.; Willers, I.

    1997-02-01

    A programming model for software parallelization, called CoCa, is introduced that copes with problems caused by typical features of high-energy physics software. By basing CoCa on the database transaction paradimg, the complexity induced by the parallelization is for a large part transparent to the programmer, resulting in a higher level of abstraction than the native message passing software. CoCa is implemented on a Meiko CS-2 and on a SUN SPARCcenter 2000 parallel computer. On the CS-2, the performance is comparable with the performance of native PVM and MPI.

  4. Evaluation of Pneumonia Virus of Mice as a Possible Human Pathogen

    PubMed Central

    Brock, Linda G.; Karron, Ruth A.; Krempl, Christine D.; Collins, Peter L.

    2012-01-01

    Pneumonia virus of mice (PVM), a relative of human respiratory syncytial virus (RSV), causes respiratory disease in mice. There is serologic evidence suggesting widespread exposure of humans to PVM. To investigate replication in primates, African green monkeys (AGM) and rhesus macaques (n = 4) were inoculated with PVM by the respiratory route. Virus was shed intermittently at low levels by a subset of animals, suggesting poor permissiveness. PVM efficiently replicated in cultured human cells and inhibited the type I interferon (IFN) response in these cells. This suggests that poor replication in nonhuman primates was not due to a general nonpermissiveness of primate cells or poor control of the IFN response. Seroprevalence in humans was examined by screening sera from 30 adults and 17 young children for PVM-neutralizing activity. Sera from a single child (6%) and 40% of adults had low neutralizing activity against PVM, which could be consistent with increasing incidence of exposure following early childhood. There was no cross-reaction of human or AGM sera between RSV and PVM and no cross-protection in the mouse model. In native Western blots, human sera reacted with RSV but not PVM proteins under conditions in which AGM immune sera reacted strongly. Serum reactivity was further evaluated by flow cytometry using unfixed Vero cells infected with PVM or RSV expressing green fluorescent protein (GFP) as a measure of viral gene expression. The reactivity of human sera against RSV-infected cells correlated with GFP expression, whereas reactivity against PVM-infected cells was low and uncorrelated with GFP expression. Thus, PVM specificity was not evident. Our results indicate that the PVM-neutralizing activity of human sera is not due to RSV- or PVM-specific antibodies but may be due to low-affinity, polyreactive natural antibodies of the IgG subclass. The absence of PVM-specific antibodies and restriction in nonhuman primates makes PVM unlikely to be a human pathogen. PMID:22438539

  5. Parallelization of KENO-Va Monte Carlo code

    NASA Astrophysics Data System (ADS)

    Ramón, Javier; Peña, Jorge

    1995-07-01

    KENO-Va is a code integrated within the SCALE system developed by Oak Ridge that solves the transport equation through the Monte Carlo Method. It is being used at the Consejo de Seguridad Nuclear (CSN) to perform criticality calculations for fuel storage pools and shipping casks. Two parallel versions of the code: one for shared memory machines and other for distributed memory systems using the message-passing interface PVM have been generated. In both versions the neutrons of each generation are tracked in parallel. In order to preserve the reproducibility of the results in both versions, advanced seeds for random numbers were used. The CONVEX C3440 with four processors and shared memory at CSN was used to implement the shared memory version. A FDDI network of 6 HP9000/735 was employed to implement the message-passing version using proprietary PVM. The speedup obtained was 3.6 in both cases.

  6. Unique nonstructural proteins of Pneumonia Virus of Mice (PVM) promote degradation of interferon (IFN) pathway components and IFN-stimulated gene proteins.

    PubMed

    Dhar, Jayeeta; Barik, Sailen

    2016-12-01

    Pneumonia Virus of Mice (PVM) is the only virus that shares the Pneumovirus genus of the Paramyxoviridae family with Respiratory Syncytial Virus (RSV). A deadly mouse pathogen, PVM has the potential to serve as a robust animal model of RSV infection, since human RSV does not fully replicate the human pathology in mice. Like RSV, PVM also encodes two nonstructural proteins that have been implicated to suppress the IFN pathway, but surprisingly, they exhibit no sequence similarity with their RSV equivalents. The molecular mechanism of PVM NS function, therefore, remains unknown. Here, we show that recombinant PVM NS proteins degrade the mouse counterparts of the IFN pathway components. Proteasomal degradation appears to be mediated by ubiquitination promoted by PVM NS proteins. Interestingly, NS proteins of PVM lowered the levels of several ISG (IFN-stimulated gene) proteins as well. These results provide a molecular foundation for the mechanisms by which PVM efficiently subverts the IFN response of the murine cell. They also reveal that in spite of their high sequence dissimilarity, the two pneumoviral NS proteins are functionally and mechanistically similar.

  7. 21 CFR 872.3500 - Polyvinylmethylether maleic anhydride (PVM-MA), acid copolymer, and carboxymethylcellulose sodium...

    Code of Federal Regulations, 2013 CFR

    2013-04-01

    ... 21 Food and Drugs 8 2013-04-01 2013-04-01 false Polyvinylmethylether maleic anhydride (PVM-MA...) MEDICAL DEVICES DENTAL DEVICES Prosthetic Devices § 872.3500 Polyvinylmethylether maleic anhydride (PVM-MA.... Polyvinylmethylether maleic anhydride (PVM-MA), acid copolymer, and carboxymethylcellulose sodium (NACMC) denture...

  8. 21 CFR 872.3500 - Polyvinylmethylether maleic anhydride (PVM-MA), acid copolymer, and carboxymethylcellulose sodium...

    Code of Federal Regulations, 2014 CFR

    2014-04-01

    ... 21 Food and Drugs 8 2014-04-01 2014-04-01 false Polyvinylmethylether maleic anhydride (PVM-MA...) MEDICAL DEVICES DENTAL DEVICES Prosthetic Devices § 872.3500 Polyvinylmethylether maleic anhydride (PVM-MA.... Polyvinylmethylether maleic anhydride (PVM-MA), acid copolymer, and carboxymethylcellulose sodium (NACMC) denture...

  9. 21 CFR 872.3500 - Polyvinylmethylether maleic anhydride (PVM-MA), acid copolymer, and carboxymethylcellulose sodium...

    Code of Federal Regulations, 2010 CFR

    2010-04-01

    ... 21 Food and Drugs 8 2010-04-01 2010-04-01 false Polyvinylmethylether maleic anhydride (PVM-MA...) MEDICAL DEVICES DENTAL DEVICES Prosthetic Devices § 872.3500 Polyvinylmethylether maleic anhydride (PVM-MA.... Polyvinylmethylether maleic anhydride (PVM-MA), acid copolymer, and carboxymethylcellulose sodium (NACMC) denture...

  10. 21 CFR 872.3500 - Polyvinylmethylether maleic anhydride (PVM-MA), acid copolymer, and carboxymethylcellulose sodium...

    Code of Federal Regulations, 2012 CFR

    2012-04-01

    ... 21 Food and Drugs 8 2012-04-01 2012-04-01 false Polyvinylmethylether maleic anhydride (PVM-MA...) MEDICAL DEVICES DENTAL DEVICES Prosthetic Devices § 872.3500 Polyvinylmethylether maleic anhydride (PVM-MA.... Polyvinylmethylether maleic anhydride (PVM-MA), acid copolymer, and carboxymethylcellulose sodium (NACMC) denture...

  11. 21 CFR 872.3500 - Polyvinylmethylether maleic anhydride (PVM-MA), acid copolymer, and carboxymethylcellulose sodium...

    Code of Federal Regulations, 2011 CFR

    2011-04-01

    ... 21 Food and Drugs 8 2011-04-01 2011-04-01 false Polyvinylmethylether maleic anhydride (PVM-MA...) MEDICAL DEVICES DENTAL DEVICES Prosthetic Devices § 872.3500 Polyvinylmethylether maleic anhydride (PVM-MA.... Polyvinylmethylether maleic anhydride (PVM-MA), acid copolymer, and carboxymethylcellulose sodium (NACMC) denture...

  12. Users manual for the Chameleon parallel programming tools

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Gropp, W.; Smith, B.

    1993-06-01

    Message passing is a common method for writing programs for distributed-memory parallel computers. Unfortunately, the lack of a standard for message passing has hampered the construction of portable and efficient parallel programs. In an attempt to remedy this problem, a number of groups have developed their own message-passing systems, each with its own strengths and weaknesses. Chameleon is a second-generation system of this type. Rather than replacing these existing systems, Chameleon is meant to supplement them by providing a uniform way to access many of these systems. Chameleon`s goals are to (a) be very lightweight (low over-head), (b) be highlymore » portable, and (c) help standardize program startup and the use of emerging message-passing operations such as collective operations on subsets of processors. Chameleon also provides a way to port programs written using PICL or Intel NX message passing to other systems, including collections of workstations. Chameleon is tracking the Message-Passing Interface (MPI) draft standard and will provide both an MPI implementation and an MPI transport layer. Chameleon provides support for heterogeneous computing by using p4 and PVM. Chameleon`s support for homogeneous computing includes the portable libraries p4, PICL, and PVM and vendor-specific implementation for Intel NX, IBM EUI (SP-1), and Thinking Machines CMMD (CM-5). Support for Ncube and PVM 3.x is also under development.« less

  13. Detection of a pneumonia virus of mice (PVM) in an African hedgehog (Atelerix arbiventris) with suspected wobbly hedgehog syndrome (WHS).

    PubMed

    Madarame, Hiroo; Ogihara, Kikumi; Kimura, Moe; Nagai, Makoto; Omatsu, Tsutomu; Ochiai, Hideharu; Mizutani, Tetsyuya

    2014-09-17

    A pneumonia virus of mice (PVM) from an African hedgehog (Atelerix arbiventris) with suspected wobbly hedgehog syndrome (WHS) was detected and genetically characterized. The affected hedgehog had a nonsuppurative encephalitis with vacuolization of the white matter, and the brain samples yielded RNA reads highly homogeneous to PVM strain 15 (96.5% of full genomic sequence homology by analysis of next generation sequencing). PVM antigen was also detected in the brain and the lungs immunohistochemically. A PVM was strongly suggested as a causative agent of encephalitis of a hedgehog with suspected WHS. This is a first report of PVM infection in hedgehogs. Copyright © 2014 Elsevier B.V. All rights reserved.

  14. Reconditioning of Batteries on the International Space Station

    NASA Technical Reports Server (NTRS)

    Hajela, Gyan; Cohen, Fred; Dalton, Penni

    2004-01-01

    Primary source of electric power for the International Space Station (ISS) is the photovoltaic module (PVM). At assembly complete stage, the ISS will be served by 4 PVMs. Each PVM contains two independent power channels such that one failure will result in loss of only one power channel. During early stages of assembly, the ISS is served by only one PVM designated as P6. Solar arrays are used to convert solar flux into electrical power. Nickel hydrogen batteries are used to store electrical power for use during periods when the solar input is not adequate to support channel loads. Batteries are operated per established procedures that ensure that they are maintained within specified temperature limits, charge current is controlled to conform to a specified charge profile, and battery voltages are maintained within specified limits. Both power channels on the PVM P6 have been operating flawlessly since December 2000 with 100 percent power availability. All components, including batteries, are monitored regularly to ensure that they are operating within specified limits and to trend their wear out and age effects. The paper briefly describes the battery trend data. Batteries have started to show some effects of aging and a battery reconditioning procedure is being evaluated at this time. Reconditioning is expected to reduce cell voltage divergence and provide data that can be used to update the state of charge (SOC) computation in the software to account for battery age. During reconditioning, each battery, one at a time, will be discharged per a specified procedure and then returned to a full state of charge. The paper describes the reconditioning procedure and the expected benefits. The reconditioning procedures have been thoroughly coordinated by all affected technical teams and approved by all required boards. The reconditioning is tentatively scheduled for September 2004.

  15. Demonstration of a full volume 3D pre-stack depth migration in the Garden Banks area using massively parallel processor (MPP) technology

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Solano, M.; Chang, H.; VanDyke, J.

    1996-12-31

    This paper describes the implementation and results of portable, production-scale 3D Pre-stack Kirchhoff depth migration software. Full volume pre-stack imaging was applied to a six million trace (46.9 Gigabyte) data set from a subsalt play in the Garden Banks area in the Gulf of Mexico. The velocity model building and updating, were derived using image depth gathers and an image-driven strategy. After three velocity iterations, depth migrated sections revealed drilling targets that were not visible in the conventional 3D post-stack time migrated data set. As expected from the implementation of the migration algorithm, it was found that amplitudes are wellmore » preserved and anomalies associated with known reservoirs, conform to petrophysical predictions. Image gathers for velocity analysis and the final depth migrated volume, were generated on an 1824 node Intel Paragon at Sandia National Laboratories. The code has been successfully ported to a CRAY (T3D) and Unix workstation Parallel Virtual Machine environments (PVM).« less

  16. Brain perivascular macrophages: characterization and functional roles in health and disease.

    PubMed

    Faraco, Giuseppe; Park, Laibaik; Anrather, Josef; Iadecola, Costantino

    2017-11-01

    Perivascular macrophages (PVM) are a distinct population of resident brain macrophages characterized by a close association with the cerebral vasculature. PVM migrate from the yolk sac into the brain early in development and, like microglia, are likely to be a self-renewing cell population that, in the normal state, is not replenished by circulating monocytes. Increasing evidence implicates PVM in several disease processes, ranging from brain infections and immune activation to regulation of the hypothalamic-adrenal axis and neurovascular-neurocognitive dysfunction in the setting of hypertension, Alzheimer disease pathology, or obesity. These effects involve crosstalk between PVM and cerebral endothelial cells, interaction with circulating immune cells, and/or production of reactive oxygen species. Overall, the available evidence supports the idea that PVM are a key component of the brain-resident immune system with broad implications for the pathogenesis of major brain diseases. A better understanding of the biology and pathobiology of PVM may lead to new insights and therapeutic strategies for a wide variety of brain diseases.

  17. Thermal control system for Space Station Freedom photovoltaic power module

    NASA Technical Reports Server (NTRS)

    Hacha, Thomas H.; Howard, Laura

    1994-01-01

    The electric power for Space Station Freedom (SSF) is generated by the solar arrays of the photovoltaic power modules (PVM's) and conditioned, controlled, and distributed by a power management and distribution system. The PVM's are located outboard of the alpha gimbals of SSF. A single-phase thermal control system is being developed to provide thermal control of PVM electrical equipment and energy storage batteries. This system uses ammonia as the coolant and a direct-flow deployable radiator. The description and development status of the PVM thermal control system is presented.

  18. Thermal control system for Space Station Freedom photovoltaic power module

    NASA Technical Reports Server (NTRS)

    Hacha, Thomas H.; Howard, Laura S.

    1992-01-01

    The electric power for Space Station Freedom (SSF) is generated by the solar arrays of the photovoltaic power modules (PVM's) and conditioned, controlled, and distributed by a power management and distribution system. The PVM's are located outboard of the alpha gimbals of SSF. A single-phase thermal control system is being developed to provide thermal control of PVM electrical equipment and energy storage batteries. This system uses ammonia as the coolant and a direct-flow deployable radiator. This paper presents the description and development status of the PVM thermal control system.

  19. Laboratory Investigation of Direct Measurement of Ice Water Content, Ice Surface Area, and Effective Radius of Ice Crystals Using a Laser-Diffraction Instrument

    NASA Technical Reports Server (NTRS)

    Gerber, H.; DeMott, P. J.; Rogers, D. C.

    1995-01-01

    The aircraft microphysics probe, PVM-100A, was tested in the Colorado State University dynamic cloud chamber to establish its ability to measure ice water content (IWC), PSA, and Re in ice clouds. Its response was compared to other means of measuring those ice-cloud parameters that included using FSSP-100 and 230-X 1-D optical probes for ice-crystal concentrations, a film-loop microscope for ice-crystal habits and dimensions, and an in-situ microscope for determining ice-crystal orientation. Intercomparisons were made in ice clouds containing ice crystals ranging in size from about 10 microns to 150 microns diameter, and ice crystals with plate, columnar, dendritic, and spherical shapes. It was not possible to determine conclusively that the PVM accurately measures IWC, PSA, and Re of ice crystals, because heat from the PVM evaporated in part the crystals in its vicinity in the chamber thus affecting its measurements. Similarities in the operating principle of the FSSP and PVM, and a comparison between Re measured by both instruments, suggest, however, that the PVM can make those measurements. The resolution limit of the PVM for IWC measurements was found to be on the order of 0.001 g/cubic m. Algorithms for correcting IWC measured by FSSP and PVM were developed.

  20. Performance Measurement, Visualization and Modeling of Parallel and Distributed Programs

    NASA Technical Reports Server (NTRS)

    Yan, Jerry C.; Sarukkai, Sekhar R.; Mehra, Pankaj; Lum, Henry, Jr. (Technical Monitor)

    1994-01-01

    This paper presents a methodology for debugging the performance of message-passing programs on both tightly coupled and loosely coupled distributed-memory machines. The AIMS (Automated Instrumentation and Monitoring System) toolkit, a suite of software tools for measurement and analysis of performance, is introduced and its application illustrated using several benchmark programs drawn from the field of computational fluid dynamics. AIMS includes (i) Xinstrument, a powerful source-code instrumentor, which supports both Fortran77 and C as well as a number of different message-passing libraries including Intel's NX Thinking Machines' CMMD, and PVM; (ii) Monitor, a library of timestamping and trace -collection routines that run on supercomputers (such as Intel's iPSC/860, Delta, and Paragon and Thinking Machines' CM5) as well as on networks of workstations (including Convex Cluster and SparcStations connected by a LAN); (iii) Visualization Kernel, a trace-animation facility that supports source-code clickback, simultaneous visualization of computation and communication patterns, as well as analysis of data movements; (iv) Statistics Kernel, an advanced profiling facility, that associates a variety of performance data with various syntactic components of a parallel program; (v) Index Kernel, a diagnostic tool that helps pinpoint performance bottlenecks through the use of abstract indices; (vi) Modeling Kernel, a facility for automated modeling of message-passing programs that supports both simulation -based and analytical approaches to performance prediction and scalability analysis; (vii) Intrusion Compensator, a utility for recovering true performance from observed performance by removing the overheads of monitoring and their effects on the communication pattern of the program; and (viii) Compatibility Tools, that convert AIMS-generated traces into formats used by other performance-visualization tools, such as ParaGraph, Pablo, and certain AVS/Explorer modules.

  1. The Pedagogical Variation Model (PVM) for Work-Based Training in Virtual Classrooms: Evaluation at Kuwait University

    ERIC Educational Resources Information Center

    Rogers, Maria Susy; Aldhafeeri, Fayiz Mensher

    2015-01-01

    A collaborative research initiative was undertaken to evaluate the pedagogical variation model (PVM) for online learning and teaching at Kuwait University. Outcomes from sample populations of students--both postgraduates and undergraduates--from the Faculty of Education were analyzed for comparison. As predicted in the PVM, the findings indicate…

  2. The engine design engine. A clustered computer platform for the aerodynamic inverse design and analysis of a full engine

    NASA Technical Reports Server (NTRS)

    Sanz, J.; Pischel, K.; Hubler, D.

    1992-01-01

    An application for parallel computation on a combined cluster of powerful workstations and supercomputers was developed. A Parallel Virtual Machine (PVM) is used as message passage language on a macro-tasking parallelization of the Aerodynamic Inverse Design and Analysis for a Full Engine computer code. The heterogeneous nature of the cluster is perfectly handled by the controlling host machine. Communication is established via Ethernet with the TCP/IP protocol over an open network. A reasonable overhead is imposed for internode communication, rendering an efficient utilization of the engaged processors. Perhaps one of the most interesting features of the system is its versatile nature, that permits the usage of the computational resources available that are experiencing less use at a given point in time.

  3. Preparation of Poly-(Methyl vinyl ether-co-maleic Anhydride) Nanoparticles by Solution-Enhanced Dispersion by Supercritical CO2

    PubMed Central

    Chen, Ai-Zheng; Wang, Guang-Ya; Wang, Shi-Bin; Feng, Jian-Gang; Liu, Yuan-Gang; Kang, Yong-Qiang

    2012-01-01

    The supercritical CO2-based technologies have been widely used in the formation of drug and/or polymer particles for biomedical applications. In this study, nanoparticles of poly-(methyl vinyl ether-co-maleic anhydride) (PVM/MA) were successfully fabricated by a process of solution-enhanced dispersion by supercritical CO2 (SEDS). A 23 factorial experiment was designed to investigate and identify the significance of the processing parameters (concentration, flow and solvent/nonsolvent) for the surface morphology, particle size, and particle size distribution of the products. The effect of the concentration of PVM/MA was found to be dominant in the results regarding particle size. Decreasing the initial solution concentration of PVM/MA decreased the particle size significantly. After optimization, the resulting PVM/MA nanoparticles exhibited a good spherical shape, a smooth surface, and a narrow particle size distribution. Fourier transform infrared spectroscopy (FTIR) spectra demonstrated that the chemical composition of PVM/MA was not altered during the SEDS process and that the SEDS process was therefore a typical physical process. The absolute value of zeta potential of the obtained PVM/MA nanoparticles was larger than 40 mV, indicating the samples’ stability in aqueous suspension. Analysis of thermogravimetry-differential scanning calorimetry (TG-DSC) revealed that the effect of the SEDS process on the thermostability of PVM/MA was negligible. The results of gas chromatography (GC) analysis confirmed that the SEDS process could efficiently remove the organic residue.

  4. The use of portable video media vs standard verbal communication in the urological consent process: a multicentre, randomised controlled, crossover trial.

    PubMed

    Winter, Matthew; Kam, Jonathan; Nalavenkata, Sunny; Hardy, Ellen; Handmer, Marcus; Ainsworth, Hannah; Lee, Wai Gin; Louie-Johnsun, Mark

    2016-11-01

    To determine if portable video media (PVM) improves patient's knowledge and satisfaction acquired during the consent process for cystoscopy and insertion of a ureteric stent compared to standard verbal communication (SVC), as informed consent is a crucial component of patient care and PVM is an emerging technology that may help improve the consent process. In this multi-centre randomised controlled crossover trial, patients requiring cystoscopy and stent insertion were recruited from two major teaching hospitals in Australia over a 15-month period (July 2014-December 2015). Patient information delivery was via PVM and SVC. The PVM consisted of an audio-visual presentation with cartoon animation presented on an iPad. Patient satisfaction was assessed using the validated Client Satisfaction Questionnaire 8 (CSQ-8; maximum score 32) and knowledge was tested using a true/false questionnaire (maximum score 28). Questionnaires were completed after first intervention and after crossover. Scores were analysed using the independent samples t-test and Wilcoxon signed-rank test for the crossover analysis. In all, 88 patients were recruited. A significant 3.1 point (15.5%) increase in understanding was demonstrable favouring the use of PVM (P < 0.001). There was no difference in patient satisfaction between the groups as judged by the CSQ-8. A significant 3.6 point (17.8%) increase in knowledge score was seen when the SVC group were crossed over to the PVM arm. A total of 80.7% of patients preferred PVM and 19.3% preferred SVC. Limitations include the lack of a validated questionnaire to test knowledge acquired from the interventions. This study demonstrates patients' preference towards PVM in the urological consent process of cystoscopy and ureteric stent insertion. PVM improves patient's understanding compared with SVC and is a more effective means of content delivery to patients in terms of overall preference and knowledge gained during the consent process. © 2016 The Authors BJU International © 2016 BJU International Published by John Wiley & Sons Ltd.

  5. By Hand or Not By-Hand: A Case Study of Alternative Approaches to Parallelize CFD Applications

    NASA Technical Reports Server (NTRS)

    Yan, Jerry C.; Bailey, David (Technical Monitor)

    1997-01-01

    While parallel processing promises to speed up applications by several orders of magnitude, the performance achieved still depends upon several factors, including the multiprocessor architecture, system software, data distribution and alignment, as well as the methods used for partitioning the application and mapping its components onto the architecture. The existence of the Gorden Bell Prize given out at Supercomputing every year suggests that while good performance can be attained for real applications on general purpose multiprocessors, the large investment in man-power and time still has to be repeated for each application-machine combination. As applications and machine architectures become more complex, the cost and time-delays for obtaining performance by hand will become prohibitive. Computer users today can turn to three possible avenues for help: parallel libraries, parallel languages and compilers, interactive parallelization tools. The success of these methodologies, in turn, depends on proper application of data dependency analysis, program structure recognition and transformation, performance prediction as well as exploitation of user supplied knowledge. NASA has been developing multidisciplinary applications on highly parallel architectures under the High Performance Computing and Communications Program. Over the past six years, the transition of underlying hardware and system software have forced the scientists to spend a large effort to migrate and recede their applications. Various attempts to exploit software tools to automate the parallelization process have not produced favorable results. In this paper, we report our most recent experience with CAPTOOL, a package developed at Greenwich University. We have chosen CAPTOOL for three reasons: 1. CAPTOOL accepts a FORTRAN 77 program as input. This suggests its potential applicability to a large collection of legacy codes currently in use. 2. CAPTOOL employs domain decomposition to obtain parallelism. Although the fact that not all kinds of parallelism are handled may seem unappealing, many NASA applications in computational aerosciences as well as earth and space sciences are amenable to domain decomposition. 3. CAPTOOL generates code for a large variety of environments employed across NASA centers: MPI/PVM on network of workstations to the IBS/SP2 and CRAY/T3D.

  6. Professional Veterinary Programs' Perceptions and Experiences Pertaining to Emotional Support Animals and Service Animals, and Recommendations for Policy Development.

    PubMed

    Schoenfeld-Tacher, Regina M; Kogan, Lori R

    Given the unique nature of programs in professional veterinary medicine (PVM), the increasing numbers of students requesting accommodations for emotional support animals (ESAs) in higher education settings is of growing interest to student affairs and administrative staff in PVM settings. Since the legislation pertaining to this type of support animal differs from the laws governing disability service animals, colleges and universities now need to develop new policies and guidelines. Representatives from a sample of 28 PVM programs completed a survey about the prevalence of student requests for ESAs and service animals. PVM associate deans for academic affairs also reported their perceptions of this issue and the challenges these requests might pose within veterinary teaching laboratories and patient treatment areas. Responses indicated that approximately one third of PVM programs have received requests for ESAs (32.1%) in the last 2 years, 17.9% have had requests for psychiatric service animals, and 17.9% for other types of service animals. Despite this, most associate deans reported not having or not being aware of university or college policies pertaining to these issues. Most associate deans are interested in learning more about this topic. This paper provides general recommendations for establishing university or PVM program policies.

  7. STUDIES ON PNEUMONIA VIRUS OF MICE (PVM) IN CELL CULTURE

    PubMed Central

    Harter, Donald H.; Choppin, Purnell W.

    1967-01-01

    Pneumonia virus of mice (PVM) has been serially propagated in a line of baby hamster kidney (BHK21) cells. A maximum titer of 6.3 x 106 TCID50 per ml was obtained, and there was little variation in yield on serial passage. PVM grown in BHK21 cells was antigenically similar to virus obtained from the mouse lung, but was somewhat less virulent for the mouse after 10 serial passages in these cells. Virus produced by BHK21 cells agglutinated mouse erythrocytes without prior heating or other treatment. Sedimentation of PVM in the ultracentrifuge or precipitation by ammonium sulfate resulted in a loss in infectivity but an increase in hemagglutinating activity, presumably due to disruption of the virus particle. In a potassium tartrate density gradient, the major portion of infective virus sedimented at a density of approximately 1.15, and noninfective hemagglutinin, at a density of approximately 1.13. Stock virus preparations appear to contain a large amount of noninfective hemagglutinin. The replication of PVM was not inhibited by 5-fluoro-2'-deoxyuridine, 5-bromo-2'-deoxyuridine, or 5-iodo-2'-deoxyuridine. Infected cells contained eosinophilic cytoplasmic inclusions which showed the acridine orange staining characteristic of single-stranded RNA. Foci of viral antigen were observed in the cytoplasm of infected cells by fluorescent antibody staining. The results suggest that PVM is an RNA virus that replicates in the cytoplasm. PMID:4165740

  8. Molecular make-up of the Plasmodium parasitophorous vacuolar membrane.

    PubMed

    Spielmann, Tobias; Montagna, Georgina N; Hecht, Leonie; Matuschewski, Kai

    2012-10-01

    Plasmodium, the causative agent of malaria, is an obligate, intracellular, eukaryotic cell that invades, replicates, and differentiates within hepatocytes and erythrocytes. Inside a host cell, a second membrane delineates the developing pathogen in addition to the parasite plasma membrane, resulting in a distinct cellular compartment, termed parasitophorous vacuole (PV). The PV membrane (PVM) constitutes the parasite-host cell interface and is likely central to nutrient acquisition, host cell remodeling, waste disposal, environmental sensing, and protection from innate defense. Over the past two decades, a number of parasite-encoded PVM proteins have been identified. They include multigene families and protein complexes, such as early-transcribed membrane proteins (ETRAMPs) and the Plasmodium translocon for exported proteins (PTEX). Nearly all Plasmodium PVM proteins are restricted to this genus and display transient and stage-specific expression. Here, we provide an overview of the PVM proteins of Plasmodium blood and liver stages. Biochemical and experimental genetics data suggest that some PVM proteins are ideal targets for novel anti-malarial intervention strategies. Copyright © 2012 Elsevier GmbH. All rights reserved.

  9. DOE Office of Scientific and Technical Information (OSTI.GOV)

    Glineur, Stephanie F.; Renshaw, Randall W.; Percopo, Caroline M.

    A previous report of a novel pneumovirus (PnV) isolated from the respiratory tract of a dog described its significant homology to the rodent pathogen, pneumonia virus of mice (PVM). The original PnV–Ane4 pathogen replicated in and could be re-isolated in infectious state from mouse lung but elicited minimal mortality compared to PVM strain J3666. Here we assess phylogeny and physiologic responses to 10 new PnV isolates. The G/glycoprotein sequences of all PnVs include elongated amino-termini when compared to the characterized PVMs, and suggest division into groups A and B. While we observed significant differences in cytokine production and neutrophil recruitmentmore » to the lungs of BALB/c mice in response to survival doses (50 TCID{sub 50} units) of representative group A (114378-10-29-KY-F) and group B (7968-11-OK) PnVs, we observed no evidence for positive selection (dN>dS) among the PnV/PnV, PVM/PnV or PVM/PVM G/glycoprotein or F/fusion protein sequence pairs. - Highlights: • We consider ten novel isolates of the pneumovirus (PnV) first described by Renshaw and colleagues. • The G/glycoprotein sequences of all PnVs include elongated amino-termini when compared to PVM. • We detect cytokine production and neutrophil recruitment to the lungs of mice in response to PnV. • We observed no evidence for positive selection (dN>dS) among the gene sequence pairs.« less

  10. Interactions between Multiple Genetic Determinants in the 5′ UTR and VP1 Capsid Control Pathogenesis of Chronic Post-Viral Myopathy caused by Coxsackievirus B1

    PubMed Central

    Sandager, Maribeth M.; Nugent, Jaime L.; Schulz, Wade L.; Messner, Ronald P.; Tam, Patricia E.

    2008-01-01

    Mice infected with coxsackievirus B1 Tucson (CVB1T) develop chronic, post-viral myopathy (PVM) with clinical manifestations of hind limb muscle weakness and myositis. The objective of the current study was to establish the genetic basis of myopathogenicity in CVB1T. Using a reverse genetics approach, full attenuation of PVM could only be achieved by simultaneously mutating four sites located at C706U in the 5′ untranslated region (5′ UTR) and at Y87F, V136A, and T276A in the VP1 capsid. Engineering these four myopathic determinants into an amyopathic CVB1T variant restored the ability to cause PVM. Moreover, these same four determinants controlled PVM expression in a second strain of mice, indicating that the underlying mechanism is operational in mice of different genetic backgrounds. Modeling studies predict that C706U alters both local and long-range pairing in the 5′ UTR, and that VP1 determinants are located on the capsid surface. However, these differences did not affect viral titers, temperature stability, pH stability, or the antibody response to virus. These studies demonstrate that PVM develops from a complex interplay between viral determinants in the 5′ UTR and VP1 capsid and have uncovered intriguing similarities between genetic determinants that cause PVM and those involved in pathogenesis of other enteroviruses. PMID:18029287

  11. Pneumonia Virus of Mice Severe Respiratory Virus Infection in a Natural Host

    PubMed Central

    Rosenberg, Helene F.; Domachowske, Joseph B.

    2008-01-01

    Pneumonia virus of mice (PVM; family Paramyxoviridae, genus Pneumovirus) is a natural mouse pathogen that is closely related to the human and bovine respiratory syncytial viruses. Among the prominent features of this infection, robust replication of PVM takes place in bronchial epithelial cells in response to a minimal virus inoculum. Virus replication in situ results in local production of proinflammatory cytokines (MIP-1α, MIP-2, MCP-1 and IFNγ) and granulocyte recruitment to the lung. If left unchecked, PVM infection and the ensuing inflammatory response ultimately lead to pulmonary edema, respiratory compromise and death. In this review, we consider the recent studies using the PVM model that have provided important insights into the role of the inflammatory response in the pathogenesis of severe respiratory virus infection. We also highlight several works that have elucidated acquired immune responses to this pathogen, including T cell responses and the development of humoral immunity. Finally, we consider several immunomodulatory strategies that have been used successfully to reduce morbidity and mortality when administered to PVM infected, symptomatic mice, and thus hold promise as realistic therapeutic strategies for severe respiratory virus infections in human subjects. PMID:18471897

  12. Air Traffic Complexity Measurement Environment (ACME): Software User's Guide

    NASA Technical Reports Server (NTRS)

    1996-01-01

    A user's guide for the Air Traffic Complexity Measurement Environment (ACME) software is presented. The ACME consists of two major components, a complexity analysis tool and user interface. The Complexity Analysis Tool (CAT) analyzes complexity off-line, producing data files which may be examined interactively via the Complexity Data Analysis Tool (CDAT). The Complexity Analysis Tool is composed of three independently executing processes that communicate via PVM (Parallel Virtual Machine) and Unix sockets. The Runtime Data Management and Control process (RUNDMC) extracts flight plan and track information from a SAR input file, and sends the information to GARP (Generate Aircraft Routes Process) and CAT (Complexity Analysis Task). GARP in turn generates aircraft trajectories, which are utilized by CAT to calculate sector complexity. CAT writes flight plan, track and complexity data to an output file, which can be examined interactively. The Complexity Data Analysis Tool (CDAT) provides an interactive graphic environment for examining the complexity data produced by the Complexity Analysis Tool (CAT). CDAT can also play back track data extracted from System Analysis Recording (SAR) tapes. The CDAT user interface consists of a primary window, a controls window, and miscellaneous pop-ups. Aircraft track and position data is displayed in the main viewing area of the primary window. The controls window contains miscellaneous control and display items. Complexity data is displayed in pop-up windows. CDAT plays back sector complexity and aircraft track and position data as a function of time. Controls are provided to start and stop playback, adjust the playback rate, and reposition the display to a specified time.

  13. Parallel Event Analysis Under Unix

    NASA Astrophysics Data System (ADS)

    Looney, S.; Nilsson, B. S.; Oest, T.; Pettersson, T.; Ranjard, F.; Thibonnier, J.-P.

    The ALEPH experiment at LEP, the CERN CN division and Digital Equipment Corp. have, in a joint project, developed a parallel event analysis system. The parallel physics code is identical to ALEPH's standard analysis code, ALPHA, only the organisation of input/output is changed. The user may switch between sequential and parallel processing by simply changing one input "card". The initial implementation runs on an 8-node DEC 3000/400 farm, using the PVM software, and exhibits a near-perfect speed-up linearity, reducing the turn-around time by a factor of 8.

  14. High Performance Programming Using Explicit Shared Memory Model on the Cray T3D

    NASA Technical Reports Server (NTRS)

    Saini, Subhash; Simon, Horst D.; Lasinski, T. A. (Technical Monitor)

    1994-01-01

    The Cray T3D is the first-phase system in Cray Research Inc.'s (CRI) three-phase massively parallel processing program. In this report we describe the architecture of the T3D, as well as the CRAFT (Cray Research Adaptive Fortran) programming model, and contrast it with PVM, which is also supported on the T3D We present some performance data based on the NAS Parallel Benchmarks to illustrate both architectural and software features of the T3D.

  15. High Performance Programming Using Explicit Shared Memory Model on Cray T3D1

    NASA Technical Reports Server (NTRS)

    Simon, Horst D.; Saini, Subhash; Grassi, Charles

    1994-01-01

    The Cray T3D system is the first-phase system in Cray Research, Inc.'s (CRI) three-phase massively parallel processing (MPP) program. This system features a heterogeneous architecture that closely couples DEC's Alpha microprocessors and CRI's parallel-vector technology, i.e., the Cray Y-MP and Cray C90. An overview of the Cray T3D hardware and available programming models is presented. Under Cray Research adaptive Fortran (CRAFT) model four programming methods (data parallel, work sharing, message-passing using PVM, and explicit shared memory model) are available to the users. However, at this time data parallel and work sharing programming models are not available to the user community. The differences between standard PVM and CRI's PVM are highlighted with performance measurements such as latencies and communication bandwidths. We have found that the performance of neither standard PVM nor CRI s PVM exploits the hardware capabilities of the T3D. The reasons for the bad performance of PVM as a native message-passing library are presented. This is illustrated by the performance of NAS Parallel Benchmarks (NPB) programmed in explicit shared memory model on Cray T3D. In general, the performance of standard PVM is about 4 to 5 times less than obtained by using explicit shared memory model. This degradation in performance is also seen on CM-5 where the performance of applications using native message-passing library CMMD on CM-5 is also about 4 to 5 times less than using data parallel methods. The issues involved (such as barriers, synchronization, invalidating data cache, aligning data cache etc.) while programming in explicit shared memory model are discussed. Comparative performance of NPB using explicit shared memory programming model on the Cray T3D and other highly parallel systems such as the TMC CM-5, Intel Paragon, Cray C90, IBM-SP1, etc. is presented.

  16. Final report of PVM-6 and PVM-7 Weather Documentation, AFCRL/Minuteman Report Number 6

    DTIC Science & Technology

    1975-09-11

    Research Laboratories, Hanscon AFB, MA 01731. USAFGL ltr, 1 Aug 1983 TH!3 R~PORT HAS BEEN DELIMI~ED AND CLtARED FOR PUBLIC REL~5E UNDER DOP DiRECTIVE...5200.20 AND NO RESTniCTIONS ARE IMPDSED UPON r rs usE P.ND n 1 scu1sURI:. DISTRIBUTION STATE~ENT A APPROVED FQR PUBLIC RELEASE; DISTRIBUTION...NO AFCRL -TR-75 -04817^ F C ^ I - 1 ^T" 1 . » " 3^ 3 / *■ TlTUSfltf fidUtUlt- JINAL REPORT OFPyM-6 AND PVM-7 WEATHER DOCUMENTATION^ | AFCRl

  17. Non-Traumatic Myositis Ossificans in the Lumbosacral Paravertebral Muscle

    PubMed Central

    Jung, DaeYoung; Roh, Ji Hyeon

    2013-01-01

    Myositis ossificans (MO) is a benign condition of non-neoplastic heterotopic bone formation in the muscle or soft tissue. Trauma plays a role in the development of MO, thus, non-traumatic MO is very rare. Although MO may occur anywhere in the body, it is rarely seen in the lumbosacral paravertebral muscle (PVM). Herein, we report a case of non-traumatic MO in the lumbosacral PVM. A 42-year-old man with no history of trauma was referred to our hospital for pain in the low back, left buttock, and left thigh. On physical examination, a slightly tender, hard, and fixed mass was palpated in the left lumbosacral PVM. Computed tomography showed a calcified mass within the left lumbosacral PVM. Magnetic resonance imaging (MRI) showed heterogeneous high signal intensity in T1- and T2-weighted image, and no enhancement of the mass was found in the postcontrast T1-weighted MRI. The lack of typical imaging features required an open biopsy, and MO was confirmed. MO should be considered in the differential diagnosis when the imaging findings show a mass involving PVM. When it is difficult to distinguish MO from soft tissue or bone malignancy by radiology, it is necessary to perform a biopsy to confirm the diagnosis. PMID:23908707

  18. Test of prototype liquid-water-content meter for aircraft use

    NASA Technical Reports Server (NTRS)

    Gerber, Hermann E.

    1993-01-01

    This report describes the effort undertaken to meet the objectives of National Science Foundation Grant ATM-9207345 titled 'Test of Prototype Liquid-Water-Content Meter for Aircraft Use.' Three activities were proposed for testing the new aircraft instrument, PVM-100A: (1) Calibrate the PVM-100A in a facility where the liquid-water-content (LWC) channel, and the integrated surface area channel (PSA) could be compared to standard means for LWC and PSA measurements. Scaling constant for the channels were to be determined in this facility. The fog/wind tunnel at ECN, Petten, The Netherlands was judged the most suitable facility for this effort. (2) Expose the PVM-100A to high wind speeds similar to those expected on research aircraft, and test the anti-icing heaters on the PVM-100A under typical icing conditions expected in atmospheric clouds. The high-speed icing tunnel at NRC, Ottawa, Canada was to be utilized. (3) Operate the PVM-100A on an aircraft during cloud penetrations to determine its stability and practicality for such measurements. The C-131A aircraft of the University of Washington was the aircraft of opportunity for these-tests, which were to be conducted during the 4-week Atlantic Stratocumulus Transition Experiment (ASTEX) in June of 1992.

  19. On the anatomy and histology of the pubovisceral muscle enthesis in women.

    PubMed

    Kim, Jinyong; Ramanah, Rajeev; DeLancey, John O L; Ashton-Miller, James A

    2011-09-01

    The origin of the pubovisceral muscle (PVM) from the pubic bone is known to be at elevated risk for injury during difficult vaginal births. We examined the anatomy and histology of its enthesial origin to classify its type and see if it differs from appendicular entheses. Parasagittal sections of the pubic bone, PVM enthesis, myotendinous junction, and muscle proper were harvested from five female cadavers (51-98 years). Histological sections were prepared with hematoxylin and eosin, Masson's trichrome, and Verhoeff-Van Gieson stains. The type of enthesis was identified according to a published enthesial classification scheme. Quantitative imaging analysis was performed in sampling bands 2 mm apart along the enthesis to determine its cross-sectional area and composition. The PVM enthesis can be classified as a fibrous enthesis. The PVM muscle fibers terminated in collagenous fibers that insert tangentially onto the periosteum of the pubic bone for the most part. Sharpey's fibers were not observed. In a longitudinal cross-section, the area of the connective tissue and muscle becomes equal approximately 8 mm from the pubic bone. The PVM originates bilaterally from the pubic bone via fibrous entheses whose collagen fibers arise tangentially from the periosteum of the pubic bone. Copyright © 2010 Wiley-Liss, Inc.

  20. On the Anatomy and Histology of the Pubovisceral Muscle Enthesis in Women

    PubMed Central

    Kim, Jinyong; Ramanah, Rajeev; DeLancey, John O. L.; Ashton-Miller, James A.

    2012-01-01

    Aims The origin of the pubovisceral muscle (PVM) from the pubic bone is known to be at elevated risk for injury during difficult vaginal births. We examined the anatomy and histology of its enthesial origin to classify its type and see if it differs from appendicular entheses. Methods Parasagittal sections of the pubic bone, PVM enthesis, myotendinous junction and muscle proper were harvested from five female cadavers (51 - 98 years). Histological sections were prepared with hematoxylin and eosin, Masson’s trichrome, and Verhoeff-Van Gieson stains. The type of enthesis was identified according to a published enthesial classification scheme. Quantitative imaging analysis was performed in sampling bands 2 mm apart along the enthesis to determine its cross-sectional area and composition. Results The PVM enthesis can be classified as a fibrous enthesis. The PVM muscle fibers terminated in collagenous fibers that insert tangentially onto the periosteum of the pubic bone for the most part. Sharpey’s fibers were not observed. In a longitudinal cross-section, the area of the connective tissue and muscle becomes equal approximately 8 mm from the pubic bone. Conclusion The PVM originates bilaterally from the pubic bone via fibrous entheses whose collagen fibers arise tangentially from the periosteum of the pubic bone. PMID:21567449

  1. Portable Video Media Versus Standard Verbal Communication in Surgical Information Delivery to Nurses: A Prospective Multicenter, Randomized Controlled Crossover Trial.

    PubMed

    Kam, Jonathan; Ainsworth, Hannah; Handmer, Marcus; Louie-Johnsun, Mark; Winter, Matthew

    2016-10-01

    Continuing education of health professionals is important for delivery of quality health care. Surgical nurses are often required to understand surgical procedures. Nurses need to be aware of the expected outcomes and recognize potential complications of such procedures during their daily work. Traditional educational methods, such as conferences and tutorials or informal education at the bedside, have many drawbacks for delivery of this information in a universal, standardized, and timely manner. The rapid uptake of portable media devices makes portable video media (PVM) a potential alternative to current educational methods. To compare PVM to standard verbal communication (SVC) for surgical information delivery and educational training for nurses and evaluate its impact on knowledge acquisition and participant satisfaction. Prospective, multicenter, randomized controlled crossover trial. Two hospitals: Gosford District Hospital and Wyong Hospital. Seventy-two nursing staff (36 at each site). Information delivery via PVM--7-minute video compared to information delivered via SVC. Knowledge acquisition was measured by a 32-point questionnaire, and satisfaction with the method of education delivery was measured using the validated Client Satisfaction Questionnaire (CSQ-8). Knowledge acquisition was higher via PVM compared to SVC 25.9 (95% confidence interval [CI] 25.2-26.6) versus 24.3 (95% CI 23.5-25.1), p = .004. Participant satisfaction was higher with PVM 29.5 (95% CI 28.3-30.7) versus 26.5 (95% CI 25.1-27.9), p = .003. Following information delivery via SVC, participants had a 6% increase in knowledge scores, 24.3 (95% CI 23.5-25.1) versus 25.7 (95% CI 24.9-26.5) p = .001, and a 13% increase in satisfaction scores, 26.5 (95% CI 25.1-27.9) versus 29.9 (95% CI 28.8-31.0) p < .001, when they crossed-over to information delivery via PVM. PVM provides a novel method for providing education to nurses that improves knowledge retention and satisfaction with the educational process. © 2016 Sigma Theta Tau International.

  2. Model institutional infrastructures for recycling of photovoltaic modules

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Reaven, S.J.; Moskowitz, P.D.; Fthenakis, V.

    1996-01-01

    How will photovoltaic modules (PVMS) be recycled at the end of their service lives? This question has technological and institutional components (Reaven, 1994a). The technological aspect concerns the physical means of recycling: what advantages and disadvantages of the several existing and emerging mechanical, thermal, and chemical recycling processes and facilities merit consideration? The institutional dimension refers to the arrangements for recycling: what are the operational and financial roles of the parties with an interest in PVM recycling? These parties include PVM manufacturers, trade organizations; distributors, and retailers; residential, commercial, and utility PVM users; waste collectors, transporters, reclaimers, and reclaimers; andmore » governments.« less

  3. Intranasal treatment with a novel immunomodulator mediates innate immune protection against lethal pneumonia virus of mice.

    PubMed

    Martinez, Elisa C; Garg, Ravendra; Shrivastava, Pratima; Gomis, Susantha; van Drunen Littel-van den Hurk, Sylvia

    2016-11-01

    Respiratory syncytial virus (RSV) is the leading cause of acute lower respiratory tract infections in infants and young children. There are no licensed RSV vaccines available, and the few treatment options for high-risk individuals are either extremely costly or cause severe side effects and toxicity. Immunomodulation mediated by a novel formulation consisting of the toll-like receptor 3 agonist poly(I:C), an innate defense regulator peptide and a polyphosphazene (P-I-P) was evaluated in the context of lethal infection with pneumonia virus of mice (PVM). Intranasal delivery of a single dose of P-I-P protected adult mice against PVM when given 24 h prior to challenge. These animals experienced minimal weight loss, no clinical disease, 100% survival, and reduced lung pathology. Similar clinical outcomes were observed in mice treated up to 3 days prior to infection. P-I-P pre-treatment induced early mRNA and protein expression of key chemokine and cytokine genes, reduced the recruitment of neutrophils and eosinophils, decreased virus titers in the lungs, and modulated the delayed exacerbated nature of PVM disease without any short-term side effects. On day 14 post-infection, P-I-P-treated mice were confirmed to be PVM-free. These results demonstrate the capacity of this formulation to prevent PVM and possibly other viral respiratory infections. Copyright © 2016 Elsevier B.V. All rights reserved.

  4. RISC Processors and High Performance Computing

    NASA Technical Reports Server (NTRS)

    Saini, Subhash; Bailey, David H.; Lasinski, T. A. (Technical Monitor)

    1995-01-01

    In this tutorial, we will discuss top five current RISC microprocessors: The IBM Power2, which is used in the IBM RS6000/590 workstation and in the IBM SP2 parallel supercomputer, the DEC Alpha, which is in the DEC Alpha workstation and in the Cray T3D; the MIPS R8000, which is used in the SGI Power Challenge; the HP PA-RISC 7100, which is used in the HP 700 series workstations and in the Convex Exemplar; and the Cray proprietary processor, which is used in the new Cray J916. The architecture of these microprocessors will first be presented. The effective performance of these processors will then be compared, both by citing standard benchmarks and also in the context of implementing a real applications. In the process, different programming models such as data parallel (CM Fortran and HPF) and message passing (PVM and MPI) will be introduced and compared. The latest NAS Parallel Benchmark (NPB) absolute performance and performance per dollar figures will be presented. The next generation of the NP13 will also be described. The tutorial will conclude with a discussion of general trends in the field of high performance computing, including likely future developments in hardware and software technology, and the relative roles of vector supercomputers tightly coupled parallel computers, and clusters of workstations. This tutorial will provide a unique cross-machine comparison not available elsewhere.

  5. Shared elements of host-targeting pathways among apicomplexan parasites of differing lifestyles.

    PubMed

    Pellé, Karell G; Jiang, Rays H Y; Mantel, Pierre-Yves; Xiao, Yu-Ping; Hjelmqvist, Daisy; Gallego-Lopez, Gina M; O T Lau, Audrey; Kang, Byung-Ho; Allred, David R; Marti, Matthias

    2015-11-01

    Apicomplexans are a diverse group of obligate parasites occupying different intracellular niches that require modification to meet the needs of the parasite. To efficiently manipulate their environment, apicomplexans translocate numerous parasite proteins into the host cell. Whereas some parasites remain contained within a parasitophorous vacuole membrane (PVM) throughout their developmental cycle, others do not, a difference that affects the machinery needed for protein export. A signal-mediated pathway for protein export into the host cell has been characterized in Plasmodium parasites, which maintain the PVM. Here, we functionally demonstrate an analogous host-targeting pathway involving organellar staging prior to secretion in the related bovine parasite, Babesia bovis, a parasite that destroys the PVM shortly after invasion. Taking into account recent identification of a similar signal-mediated pathway in the coccidian parasite Toxoplasma gondii, we suggest a model in which this conserved pathway has evolved in multiple steps from signal-mediated trafficking to specific secretory organelles for controlled secretion to a complex protein translocation process across the PVM. © 2015 John Wiley & Sons Ltd.

  6. Generalization of soft phonon modes

    NASA Astrophysics Data System (ADS)

    Rudin, Sven P.

    2018-04-01

    Soft phonon modes describe a collective movement of atoms that transform a higher-symmetry crystal structure into a lower-symmetry crystal structure. Such structural transformations occur at finite temperatures, where the phonons (i.e., the low-temperature vibrational modes) and the static perfect crystal structures provide an incomplete picture of the dynamics. Here, principal vibrational modes (PVMs) are introduced as descriptors of the dynamics of a material system with N atoms. The PVMs represent the independent collective movements of the atoms at a given temperature. Molecular dynamics (MD) simulations, here in the form of quantum MD using density functional theory calculations, provide both the data describing the atomic motion and the data used to construct the PVMs. The leading mode, PVM0, represents the 3 N -dimensional direction in which the system moves with greatest amplitude. For structural phase transitions, PVM0 serves as a generalization of soft phonon modes. At low temperatures, PVM0 reproduces the soft phonon mode in systems where one phonon dominates the phase transformation. In general, multiple phonon modes combine to describe a transformation, in which case PVM0 culls these phonon modes. Moreover, while soft phonon modes arise in the higher-symmetry crystal structure, PVM0 can be equally well calculated on either side of the structural phase transition. Two applications demonstrate these properties: first, transitions into and out of bcc titanium, and, second, the two crystal structures proposed for the β phase of uranium, the higher-symmetry structure of which stabilizes with temperature.

  7. Preparation of Chitosan-Based Hemostatic Sponges by Supercritical Fluid Technology

    PubMed Central

    Song, Hu-Fan; Chen, Ai-Zheng; Wang, Shi-Bin; Kang, Yong-Qiang; Ye, Shi-Fu; Liu, Yuan-Gang; Wu, Wen-Guo

    2014-01-01

    Using ammonium bicarbonate (AB) particles as a porogen, chitosan (CS)-based hemostatic porous sponges were prepared in supercritical carbon dioxide due to its low viscosity, small surface tension, and good compatibility with organic solvent. Fourier transform infrared spectroscopy (FTIR) spectra demonstrated that the chemical compositions of CS and poly-(methyl vinyl ether-co-maleic anhydride) (PVM/MA) were not altered during the phase inversion process. The morphology and structure of the sponge after the supercritical fluid (SCF) process were observed by scanning electron microscopy (SEM). The resulting hemostatic sponges showed a relatively high porosity (about 80%) with a controllable pore size ranging from 0.1 to 200 μm. The concentration of PVM/MA had no significant influence on the porosity of the sponges. Comparative experiments on biological assessment and hemostatic effect between the resulting sponges and Avitene® were also carried out. With the incorporation of PVM/MA into the CS-based sponges, the water absorption rate of the sponges increased significantly, and the CS-PVM/MA sponges showed a similar water absorption rate (about 90%) to that of Avitene®. The results of the whole blood clotting experiment and animal experiment also demonstrated that the clotting ability of the CS-PVM/MA sponges was similar to that of Avitene®. All these results elementarily verified that the sponges prepared in this study were suitable for hemostasis and demonstrated the feasibility of using SCF-assisted phase inversion technology to produce hemostatic porous sponges. PMID:28788577

  8. Animal model of respiratory syncytial virus: CD8+ T cells cause a cytokine storm that is chemically tractable by sphingosine-1-phosphate 1 receptor agonist therapy.

    PubMed

    Walsh, Kevin B; Teijaro, John R; Brock, Linda G; Fremgen, Daniel M; Collins, Peter L; Rosen, Hugh; Oldstone, Michael B A

    2014-06-01

    The cytokine storm is an intensified, dysregulated, tissue-injurious inflammatory response driven by cytokine and immune cell components. The cytokine storm during influenza virus infection, whereby the amplified innate immune response is primarily responsible for pulmonary damage, has been well characterized. Now we describe a novel event where virus-specific T cells induce a cytokine storm. The paramyxovirus pneumonia virus of mice (PVM) is a model of human respiratory syncytial virus (hRSV). Unexpectedly, when C57BL/6 mice were infected with PVM, the innate inflammatory response was undetectable until day 5 postinfection, at which time CD8(+) T cells infiltrated into the lung, initiating a cytokine storm by their production of gamma interferon (IFN-γ) and tumor necrosis factor alpha (TNF-α). Administration of an immunomodulatory sphingosine-1-phosphate (S1P) receptor 1 (S1P1R) agonist significantly inhibited PVM-elicited cytokine storm by blunting the PVM-specific CD8(+) T cell response, resulting in diminished pulmonary disease and enhanced survival. A dysregulated overly exuberant immune response, termed a "cytokine storm," accompanies virus-induced acute respiratory diseases (VARV), is primarily responsible for the accompanying high morbidity and mortality, and can be controlled therapeutically in influenza virus infection of mice and ferrets by administration of sphingosine-1-phosphate 1 receptor (S1P1R) agonists. Here, two novel findings are recorded. First, in contrast to influenza infection, where the cytokine storm is initiated early by the innate immune system, for pneumonia virus of mice (PVM), a model of RSV, the cytokine storm is initiated late in infection by the adaptive immune response: specifically, by virus-specific CD8 T cells via their release of IFN-γ and TNF-α. Blockading these cytokines with neutralizing antibodies blunts the cytokine storm and protects the host. Second, PVM infection is controlled by administration of an S1P1R agonist.

  9. Application of a distributed network in computational fluid dynamic simulations

    NASA Technical Reports Server (NTRS)

    Deshpande, Manish; Feng, Jinzhang; Merkle, Charles L.; Deshpande, Ashish

    1994-01-01

    A general-purpose 3-D, incompressible Navier-Stokes algorithm is implemented on a network of concurrently operating workstations using parallel virtual machine (PVM) and compared with its performance on a CRAY Y-MP and on an Intel iPSC/860. The problem is relatively computationally intensive, and has a communication structure based primarily on nearest-neighbor communication, making it ideally suited to message passing. Such problems are frequently encountered in computational fluid dynamics (CDF), and their solution is increasingly in demand. The communication structure is explicitly coded in the implementation to fully exploit the regularity in message passing in order to produce a near-optimal solution. Results are presented for various grid sizes using up to eight processors.

  10. Generalization of soft phonon modes

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Rudin, Sven P.

    Soft phonon modes describe a collective movement of atoms that transform a higher-symmetry crystal structure into a lower-symmetry crystal structure. Such structural transformations occur at finite temperatures, where the phonons (i.e., the low-temperature vibrational modes) and the static perfect crystal structures provide an incomplete picture of the dynamics. In this paper, principal vibrational modes (PVMs) are introduced as descriptors of the dynamics of a material system withmore » $N$ atoms. The PVMs represent the independent collective movements of the atoms at a given temperature. Molecular dynamics (MD) simulations, here in the form of quantum MD using density functional theory calculations, provide both the data describing the atomic motion and the data used to construct the PVMs. The leading mode, $${\\mathrm{PVM}}_{0}$$, represents the $3N$-dimensional direction in which the system moves with greatest amplitude. For structural phase transitions, $${\\mathrm{PVM}}_{0}$$ serves as a generalization of soft phonon modes. At low temperatures, $${\\mathrm{PVM}}_{0}$$ reproduces the soft phonon mode in systems where one phonon dominates the phase transformation. In general, multiple phonon modes combine to describe a transformation, in which case $${\\mathrm{PVM}}_{0}$$ culls these phonon modes. Moreover, while soft phonon modes arise in the higher-symmetry crystal structure, $${\\mathrm{PVM}}_{0}$$ can be equally well calculated on either side of the structural phase transition. Finally, two applications demonstrate these properties: first, transitions into and out of bcc titanium, and, second, the two crystal structures proposed for the $${\\beta}$$ phase of uranium, the higher-symmetry structure of which stabilizes with temperature.« less

  11. Generalization of soft phonon modes

    DOE PAGES

    Rudin, Sven P.

    2018-04-27

    Soft phonon modes describe a collective movement of atoms that transform a higher-symmetry crystal structure into a lower-symmetry crystal structure. Such structural transformations occur at finite temperatures, where the phonons (i.e., the low-temperature vibrational modes) and the static perfect crystal structures provide an incomplete picture of the dynamics. In this paper, principal vibrational modes (PVMs) are introduced as descriptors of the dynamics of a material system withmore » $N$ atoms. The PVMs represent the independent collective movements of the atoms at a given temperature. Molecular dynamics (MD) simulations, here in the form of quantum MD using density functional theory calculations, provide both the data describing the atomic motion and the data used to construct the PVMs. The leading mode, $${\\mathrm{PVM}}_{0}$$, represents the $3N$-dimensional direction in which the system moves with greatest amplitude. For structural phase transitions, $${\\mathrm{PVM}}_{0}$$ serves as a generalization of soft phonon modes. At low temperatures, $${\\mathrm{PVM}}_{0}$$ reproduces the soft phonon mode in systems where one phonon dominates the phase transformation. In general, multiple phonon modes combine to describe a transformation, in which case $${\\mathrm{PVM}}_{0}$$ culls these phonon modes. Moreover, while soft phonon modes arise in the higher-symmetry crystal structure, $${\\mathrm{PVM}}_{0}$$ can be equally well calculated on either side of the structural phase transition. Finally, two applications demonstrate these properties: first, transitions into and out of bcc titanium, and, second, the two crystal structures proposed for the $${\\beta}$$ phase of uranium, the higher-symmetry structure of which stabilizes with temperature.« less

  12. Modeling and analysis of solar distributed generation

    NASA Astrophysics Data System (ADS)

    Ortiz Rivera, Eduardo Ivan

    Recent changes in the global economy are creating a big impact in our daily life. The price of oil is increasing and the number of reserves are less every day. Also, dramatic demographic changes are impacting the viability of the electric infrastructure and ultimately the economic future of the industry. These are some of the reasons that many countries are looking for alternative energy to produce electric energy. The most common form of green energy in our daily life is solar energy. To convert solar energy into electrical energy is required solar panels, dc-dc converters, power control, sensors, and inverters. In this work, a photovoltaic module, PVM, model using the electrical characteristics provided by the manufacturer data sheet is presented for power system applications. Experimental results from testing are showed, verifying the proposed PVM model. Also in this work, three maximum power point tracker, MPPT, algorithms would be presented to obtain the maximum power from a PVM. The first MPPT algorithm is a method based on the Rolle's and Lagrange's Theorems and can provide at least an approximate answer to a family of transcendental functions that cannot be solved using differential calculus. The second MPPT algorithm is based on the approximation of the proposed PVM model using fractional polynomials where the shape, boundary conditions and performance of the proposed PVM model are satisfied. The third MPPT algorithm is based in the determination of the optimal duty cycle for a dc-dc converter and the previous knowledge of the load or load matching conditions. Also, four algorithms to calculate the effective irradiance level and temperature over a photovoltaic module are presented in this work. The main reasons to develop these algorithms are for monitoring climate conditions, the elimination of temperature and solar irradiance sensors, reductions in cost for a photovoltaic inverter system, and development of new algorithms to be integrated with maximum power point tracking algorithms. Finally, several PV power applications will be presented like circuit analysis for a load connected to two different PV arrays, speed control for a do motor connected to a PVM, and a novel single phase photovoltaic inverter system using the Z-source converter.

  13. Kinetic and kinematic evaluation of compensatory movements of the head, pelvis and thoracolumbar spine associated with asymmetric weight bearing of the pelvic limbs in trotting dogs.

    PubMed

    Hicks, D A; Millis, D L

    2014-01-01

    To determine ground reaction forces, head and pelvis vertical motion (HVM and PVM, respectively), and thoraco-lumbar lateral angular motion (LAM) of the spine using kinematic gait analysis in dogs with mild asymmetric weight-bearing of the pelvic limbs while trotting. Twenty-seven hound-type dogs were fitted with reflective markers placed on the sagittal crest of the skull, the ischiatic tuberosity, and thoracolumbar spine of dogs to track motion while trotting. Kinetic and kinematic data were used to characterize asymmetry between the left and right pelvic limbs, and to describe HVM, PVM and thoraco-lumbar LAM. Maximum and minimum position and total motion values were determined for each measured variable. Dogs with asymmetric weight bearing of the pelvic limbs had greater PVM on the side with a greater peak vertical force (PVF), and greater thoraco-lumbar LAM toward the side with a lower PVF while trotting. No differences in mean HVM were detected, and there were no significant correlations between the magnitude of HVM, PVM and thoraco-lumbar LAM and the degree of asymmetric weight bearing. Dogs with subtle asymmetric weight bearing of a pelvic limb had patterns of body motion that may be useful in identifying subtle lameness in dogs; greater PVM on the side with greater weight bearing and greater thoraco-lumbar LAM toward the side with less weight bearing while trotting. Description of these compensatory movements is valuable when evaluating dogs with subtle weight bearing asymmetry in the pelvic limbs and may improve the sensitivity of lameness detection during subjective clinical lameness examination.

  14. Higher operation temperature quadrant photon detectors of 2-11 μm wavelength radiation with large photosensitive areas

    NASA Astrophysics Data System (ADS)

    Pawluczyk, J.; Sosna, A.; Wojnowski, D.; Koźniewski, A.; Romanis, M.; Gawron, W.; Piotrowski, J.

    2017-10-01

    We report on the quadrant photon HgCdTe detectors optimized for 2-11 μm wavelength spectral range and Peltier or no cooling, and photosensitive area of a quad-cell of 1×1 to 4×4 mm. The devices are fabricated as photoconductors or multiple photovoltaic cells connected in series (PVM). The former are characterized by a relatively uniform photosensitive area. The PVM photovoltaic cells are distributed along the wafer surface, comprising a periodical stripe structure with a period of 20 μm. Within each period, there is an insensitive gap/trench < 9 μm wide between stripe mesas. The resulting spatial quantization error prevents positioning of the beam spot of size close to the period, but becomes negligible for the optimal spot size comparable to a quadrant-cell area. The photoconductors produce 1/f noise with about 10 kHz knee frequency, due to bias necessary for their operation. The PVM photodiodes are typically operated at 0 V bias, so they generate no 1/f noise and operation from DC is enabled. At 230 K, upper corner frequency of 16 to 100 MHz is obtained for photoconductor and 60 to 80 MHz for PVM, normalized detectivity D* 6×107 cm×Hz1/2/W to >1.4×108 cm×Hz1/2/W for photoconductor and >1.7×108 cm·Hz1/2/W for PVM, allowing for position control of the radiation beam with submicron accuracy at 16 MHz, 10.6 μm wavelength of pulsed radiation spot of 0.8 mm dia at the close-to-maximal input radiation power density in a range of detector linear operation.

  15. Characterization of the passive and active material parameters of the pubovisceralis muscle using an inverse numerical method.

    PubMed

    Silva, M E T; Parente, M P L; Brandão, S; Mascarenhas, T; Natal Jorge, R M

    2018-04-11

    The mechanical characteristics of the female pelvic floor are relevant to understand pelvic floor dysfunctions (PFD), and how they are related with changes in their biomechanical behavior. Urinary incontinence (UI) and pelvic organ prolapse (POP) are the most common pathologies, which can be associated with changes in the mechanical properties of the supportive structures in the female pelvic cavity. PFD have been studied through different methods, from experimental tensile tests using tissues from fresh female cadavers or tissues collected at the time of a transvaginal hysterectomy procedure, or by applying imaging techniques. In this work, an inverse finite element analysis (FEA) was applied to understand the passive and active behavior of the pubovisceralis muscle (PVM) during Valsalva maneuver and muscle active contraction, respectively. Individual numerical models of women without pathology, with stress UI (SUI) and POP were built based on magnetic resonance images, including the PVM and surrounding structures. The passive and active material parameters obtained for a transversely isotropic hyperelastic constitutive model were estimated for the three groups. The values for the material constants were significantly higher for the women with POP when compared with the other two groups. The PVM of women with POP showed the highest stiffness. Additionally, the influence of these parameters was analyzed by evaluating their stress-strain, and force-displacements responses. The force produced by the PVM in women with POP was 47% and 82% higher when compared to women without pathology and with SUI, respectively. The inverse FEA allowed estimating the material parameters of the PVM using input information acquired non-invasively. Copyright © 2018 Elsevier Ltd. All rights reserved.

  16. Implementation of Helioseismic Data Reduction and Diagnostic Techniques on Massively Parallel Architectures

    NASA Technical Reports Server (NTRS)

    Korzennik, Sylvain

    1997-01-01

    Under the direction of Dr. Rhodes, and the technical supervision of Dr. Korzennik, the data assimilation of high spatial resolution solar dopplergrams has been carried out throughout the program on the Intel Delta Touchstone supercomputer. With the help of a research assistant, partially supported by this grant, and under the supervision of Dr. Korzennik, code development was carried out at SAO, using various available resources. To ensure cross-platform portability, PVM was selected as the message passing library. A parallel implementation of power spectra computation for helioseismology data reduction, using PVM was successfully completed. It was successfully ported to SMP architectures (i.e. SUN), and to some MPP architectures (i.e. the CM5). Due to limitation of the implementation of PVM on the Cray T3D, the port to that architecture was not completed at the time.

  17. Self-assembled software and method of overriding software execution

    DOEpatents

    Bouchard, Ann M.; Osbourn, Gordon C.

    2013-01-08

    A computer-implemented software self-assembled system and method for providing an external override and monitoring capability to dynamically self-assembling software containing machines that self-assemble execution sequences and data structures. The method provides an external override machine that can be introduced into a system of self-assembling machines while the machines are executing such that the functionality of the executing software can be changed or paused without stopping the code execution and modifying the existing code. Additionally, a monitoring machine can be introduced without stopping code execution that can monitor specified code execution functions by designated machines and communicate the status to an output device.

  18. Software platform virtualization in chemistry research and university teaching

    PubMed Central

    2009-01-01

    Background Modern chemistry laboratories operate with a wide range of software applications under different operating systems, such as Windows, LINUX or Mac OS X. Instead of installing software on different computers it is possible to install those applications on a single computer using Virtual Machine software. Software platform virtualization allows a single guest operating system to execute multiple other operating systems on the same computer. We apply and discuss the use of virtual machines in chemistry research and teaching laboratories. Results Virtual machines are commonly used for cheminformatics software development and testing. Benchmarking multiple chemistry software packages we have confirmed that the computational speed penalty for using virtual machines is low and around 5% to 10%. Software virtualization in a teaching environment allows faster deployment and easy use of commercial and open source software in hands-on computer teaching labs. Conclusion Software virtualization in chemistry, mass spectrometry and cheminformatics is needed for software testing and development of software for different operating systems. In order to obtain maximum performance the virtualization software should be multi-core enabled and allow the use of multiprocessor configurations in the virtual machine environment. Server consolidation, by running multiple tasks and operating systems on a single physical machine, can lead to lower maintenance and hardware costs especially in small research labs. The use of virtual machines can prevent software virus infections and security breaches when used as a sandbox system for internet access and software testing. Complex software setups can be created with virtual machines and are easily deployed later to multiple computers for hands-on teaching classes. We discuss the popularity of bioinformatics compared to cheminformatics as well as the missing cheminformatics education at universities worldwide. PMID:20150997

  19. Software platform virtualization in chemistry research and university teaching.

    PubMed

    Kind, Tobias; Leamy, Tim; Leary, Julie A; Fiehn, Oliver

    2009-11-16

    Modern chemistry laboratories operate with a wide range of software applications under different operating systems, such as Windows, LINUX or Mac OS X. Instead of installing software on different computers it is possible to install those applications on a single computer using Virtual Machine software. Software platform virtualization allows a single guest operating system to execute multiple other operating systems on the same computer. We apply and discuss the use of virtual machines in chemistry research and teaching laboratories. Virtual machines are commonly used for cheminformatics software development and testing. Benchmarking multiple chemistry software packages we have confirmed that the computational speed penalty for using virtual machines is low and around 5% to 10%. Software virtualization in a teaching environment allows faster deployment and easy use of commercial and open source software in hands-on computer teaching labs. Software virtualization in chemistry, mass spectrometry and cheminformatics is needed for software testing and development of software for different operating systems. In order to obtain maximum performance the virtualization software should be multi-core enabled and allow the use of multiprocessor configurations in the virtual machine environment. Server consolidation, by running multiple tasks and operating systems on a single physical machine, can lead to lower maintenance and hardware costs especially in small research labs. The use of virtual machines can prevent software virus infections and security breaches when used as a sandbox system for internet access and software testing. Complex software setups can be created with virtual machines and are easily deployed later to multiple computers for hands-on teaching classes. We discuss the popularity of bioinformatics compared to cheminformatics as well as the missing cheminformatics education at universities worldwide.

  20. Improved approach to quantitative cardiac volumetrics using automatic thresholding and manual trimming: a cardiovascular MRI study.

    PubMed

    Rayarao, Geetha; Biederman, Robert W W; Williams, Ronald B; Yamrozik, June A; Lombardi, Richard; Doyle, Mark

    2018-01-01

    To establish the clinical validity and accuracy of automatic thresholding and manual trimming (ATMT) by comparing the method with the conventional contouring method for in vivo cardiac volume measurements. CMR was performed on 40 subjects (30 patients and 10 controls) using steady-state free precession cine sequences with slices oriented in the short-axis and acquired contiguously from base to apex. Left ventricular (LV) volumes, end-diastolic volume, end-systolic volume, and stroke volume (SV) were obtained with ATMT and with the conventional contouring method. Additionally, SV was measured independently using CMR phase velocity mapping (PVM) of the aorta for validation. Three methods of calculating SV were compared by applying Bland-Altman analysis. The Bland-Altman standard deviation of variation (SD) and offset bias for LV SV for the three sets of data were: ATMT-PVM (7.65, [Formula: see text]), ATMT-contours (7.85, [Formula: see text]), and contour-PVM (11.01, 4.97), respectively. Equating the observed range to the error contribution of each approach, the error magnitude of ATMT:PVM:contours was in the ratio 1:2.4:2.5. Use of ATMT for measuring ventricular volumes accommodates trabeculae and papillary structures more intuitively than contemporary contouring methods. This results in lower variation when analyzing cardiac structure and function and consequently improved accuracy in assessing chamber volumes.

  1. Start-up capabilities of photovoltaic module for the International Space Station

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Hajela, G.; Hague, L.

    1997-12-31

    The International Space Station (ISS) uses four photovoltaic modules (PVMs) to supply electric power for the US On-Orbit Segment (USOS). The ISS is assembled on orbit over a period of about 5 years and over 40 stages. PVMs are launched and integrated with the ISS at different times during the ISS assembly. During early stages, the electric power is provided by the integrated truss segment (ITS) P6; subsequently, ITS P4, S4, and S6 are launched. PVMs are launched into space in the National Space Transportation System (NSTS) cargo bay. Each PVM consists of two independent power channels. The NSTS docksmore » with the ISS, the PVM is removed from the cargo bay and installed on the ISS. At this stage the PVM is in stowed configuration and its batteries are in fully discharged state. The start-up consists of initialization and checkout of all hardware, deployment of SAW and photovoltaic radiator (PVR), thermal conditioning batteries, and charging batteries; not necessarily in the same order for all PVMs. PVMs are designed to be capable of on-orbit start-up, within a specified time period, when external power is applied to a specified electrical interface. This paper describes the essential steps required for PVM start-up and how these operations are performed for various PVMs. The integrated operations scenarios (IOS) prepared by the NASA, Johnson Space Center, details specific procedures and timelines for start-up of each PVM. The paper describes how dormant batteries are brought to their normal operating temperature range and then charged to 100% state of charge (SOC). Total time required to complete start-up is computed and compared to the IOS timelines. External power required during start-up is computed and compared to the requirements and/or available power on ISS. Also described is how these start-up procedures can be adopted for restart of PVMs when required.« less

  2. A heterogeneous computing environment for simulating astrophysical fluid flows

    NASA Technical Reports Server (NTRS)

    Cazes, J.

    1994-01-01

    In the Concurrent Computing Laboratory in the Department of Physics and Astronomy at Louisiana State University we have constructed a heterogeneous computing environment that permits us to routinely simulate complicated three-dimensional fluid flows and to readily visualize the results of each simulation via three-dimensional animation sequences. An 8192-node MasPar MP-1 computer with 0.5 GBytes of RAM provides 250 MFlops of execution speed for our fluid flow simulations. Utilizing the parallel virtual machine (PVM) language, at periodic intervals data is automatically transferred from the MP-1 to a cluster of workstations where individual three-dimensional images are rendered for inclusion in a single animation sequence. Work is underway to replace executions on the MP-1 with simulations performed on the 512-node CM-5 at NCSA and to simultaneously gain access to more potent volume rendering workstations.

  3. THE MODIFYING EFFECTS OF CERTAIN SUBSTANCES OF BACTERIAL ORIGIN ON THE COURSE OF INFECTION WITH PNEUMONIA VIRUS OF MICE (PVM)

    PubMed Central

    Horsfall, Frank L.; McCarty, Maclyn

    1947-01-01

    Evidence is presented which indicates that certain polysaccharide preparations derived from various bacterial species, as well as similar materials not of bacterial origin, are capable of lessening the severity of infection with pneumonia virus of mice (PVM) and inhibiting multiplication of the virus in mouse lungs infected with this agent. It seems probable that modification with respect to the virus is mediated by a substance which may be polysaccharide in nature. PMID:19871640

  4. Reflections on the ethics of participatory visual methods to engage communities in global health research

    PubMed Central

    Black, Gillian F.; Davies, Alun; Iskander, Dalia; Chambers, Mary

    2018-01-01

    ABSTRACT There is a growing body of literature describing conceptual frameworks for working with participatory visual methods (PVM). Through a global health lens, this paper examines some key themes within these frameworks. We reflect on our experiences of working with with an array of PVM to engage community members in Vietnam, Kenya, the Philippines and South Africa in biomedical research and public health. The participants that we have engaged in these processes live in under-resourced areas with high prevalence of communicable and non-communicable diseases. Our paper describes some of the challenges that we have encountered while using PVM to foster knowledge exchange, build relationships and facilitate change among individuals and families, community members, health workers, biomedical scientists and researchers. We consider multiple ethical situations that have arisen through our work and discuss the ways in which we have navigated and negotiated them. We offer our reflections and learning from facilitating these processes and in doing so we add novel contributions to ethical framework concepts. PMID:29434532

  5. A distributed monitoring system for photovoltaic arrays based on a two-level wireless sensor network

    NASA Astrophysics Data System (ADS)

    Su, F. P.; Chen, Z. C.; Zhou, H. F.; Wu, L. J.; Lin, P. J.; Cheng, S. Y.; Li, Y. F.

    2017-11-01

    In this paper, a distributed on-line monitoring system based on a two-level wireless sensor network (WSN) is proposed for real time status monitoring of photovoltaic (PV) arrays to support the fine management and maintenance of PV power plants. The system includes the sensing nodes installed on PV modules (PVM), sensing and routing nodes installed on combiner boxes of PV sub-arrays (PVA), a sink node and a data management centre (DMC) running on a host computer. The first level WSN is implemented by the low-cost wireless transceiver nRF24L01, and it is used to achieve single hop communication between the PVM nodes and their corresponding PVA nodes. The second level WSN is realized by the CC2530 based ZigBee network for multi-hop communication among PVA nodes and the sink node. The PVM nodes are used to monitor the PVM working voltage and backplane temperature, and they send the acquired data to their PVA node via the nRF24L01 based first level WSN. The PVA nodes are used to monitor the array voltage, PV string current and environment irradiance, and they send the acquired and received data to the DMC via the ZigBee based second level WSN. The DMC is designed using the MATLAB GUIDE and MySQL database. Laboratory experiment results show that the system can effectively acquire, display, store and manage the operating and environment parameters of PVA in real time.

  6. Empirical Analysis and Refinement of Expert System Knowledge Bases

    DTIC Science & Technology

    1990-03-31

    the number of hidden units and the error rates is listed in Figure 6. 3.3. Cancer Data A data qet for eva!ukting th.- Frognosis of breast cancer ...Alternative Rule Induction Methods A data set for evaluating the prognosis of breast cancer recurrence was analyzed by Michalski’s AQI5 rule induction program...AQ15 7 2 32% PVM 2 1 23% Figure 6-3: Comparative Summa-y for AQI5 and PVM on Breast Cancer Data 6.2.2. Alternative Decision Tree Induction Methods

  7. Methods For Self-Organizing Software

    DOEpatents

    Bouchard, Ann M.; Osbourn, Gordon C.

    2005-10-18

    A method for dynamically self-assembling and executing software is provided, containing machines that self-assemble execution sequences and data structures. In addition to ordered functions calls (found commonly in other software methods), mutual selective bonding between bonding sites of machines actuates one or more of the bonding machines. Two or more machines can be virtually isolated by a construct, called an encapsulant, containing a population of machines and potentially other encapsulants that can only bond with each other. A hierarchical software structure can be created using nested encapsulants. Multi-threading is implemented by populations of machines in different encapsulants that are interacting concurrently. Machines and encapsulants can move in and out of other encapsulants, thereby changing the functionality. Bonding between machines' sites can be deterministic or stochastic with bonding triggering a sequence of actions that can be implemented by each machine. A self-assembled execution sequence occurs as a sequence of stochastic binding between machines followed by their deterministic actuation. It is the sequence of bonding of machines that determines the execution sequence, so that the sequence of instructions need not be contiguous in memory.

  8. 76 FR 54800 - International Business Machines (IBM), Software Group Business Unit, Quality Assurance Group, San...

    Federal Register 2010, 2011, 2012, 2013, 2014

    2011-09-02

    ... Machines (IBM), Software Group Business Unit, Quality Assurance Group, San Jose, California; Notice of... workers of International Business Machines (IBM), Software Group Business Unit, Optim Data Studio Tools QA... February 2, 2011 (76 FR 5832). The subject worker group supplies acceptance testing services, design...

  9. Immunobiotic Lactobacillus administered post-exposure averts the lethal sequelae of respiratory virus infection.

    PubMed

    Percopo, Caroline M; Rice, Tyler A; Brenner, Todd A; Dyer, Kimberly D; Luo, Janice L; Kanakabandi, Kishore; Sturdevant, Daniel E; Porcella, Stephen F; Domachowske, Joseph B; Keicher, Jesse D; Rosenberg, Helene F

    2015-09-01

    We reported previously that priming of the respiratory tract with immunobiotic Lactobacillus prior to virus challenge protects mice against subsequent lethal infection with pneumonia virus of mice (PVM). We present here the results of gene microarray which document differential expression of proinflammatory mediators in response to PVM infection alone and those suppressed in response to Lactobacillus plantarum. We also demonstrate for the first time that intranasal inoculation with live or heat-inactivated L. plantarum or Lactobacillus reuteri promotes full survival from PVM infection when administered within 24h after virus challenge. Survival in response to L. plantarum administered after virus challenge is associated with suppression of proinflammatory cytokines, limited virus recovery, and diminished neutrophil recruitment to lung tissue and airways. Utilizing this post-virus challenge protocol, we found that protective responses elicited by L. plantarum at the respiratory tract were distinct from those at the gastrointestinal mucosa, as mice devoid of the anti-inflammatory cytokine, interleukin (IL)-10, exhibit survival and inflammatory responses that are indistinguishable from those of their wild-type counterparts. Finally, although L. plantarum interacts specifically with pattern recognition receptors TLR2 and NOD2, the respective gene-deleted mice were fully protected against lethal PVM infection by L. plantarum, as are mice devoid of type I interferon receptors. Taken together, L. plantarum is a versatile and flexible agent that is capable of averting the lethal sequelae of severe respiratory infection both prior to and post-virus challenge via complex and potentially redundant mechanisms. Published by Elsevier B.V.

  10. Crocodilian perivitelline membrane-bound sperm detection.

    PubMed

    Augustine, Lauren

    2017-05-01

    Advanced reproductive technologies (ART's) are often employed with various taxa to enhance captive breeding programs and maintain genetic diversity. Perivitelline membrane-bound (PVM-bound) sperm detection has previously been demonstrated in avian and chelonian species as a useful technique for breeding management. In the absence of embryotic development within an egg, this technique can detect the presence of sperm trapped on the oocyte membrane confirming breeding, male reproductive status, and pair compatibility. PVM-bound sperm were successfully detected in three clutches of Cuban crocodile (Crocodylus rhombifer) eggs at the Smithsonian's National Zoological Park (NZP) for the first time in any crocodilian species. PVM-bound sperm were detected in fresh and incubated C. rhombifer eggs, as well as eggs that were developing (banded) and those that were not (not banded). The results of this study showed significant differences in average sperm densities per egg between clutches (p = 0.001). Additionally, there was not a significant difference within clutches between eggs that banded and those that did not band (Clutch A, p = 0.505; Clutch B, p = 0.665; Clutch C, p = 0.266). The results of this study demonstrate the necessity to microscopically examine eggs that do not develop (do not band), to determine if sperm is present, which can help animal managers problem solve reproductive shortcomings. PVM-bound sperm detection could be a useful technique in assessing crocodilian breeding programs, as well as have potential uses in studies assessing sperm storage, artificial insemination, and artificial incubation. This article is a U.S. Government work and is in the public domain in the USA.

  11. Torins are potent antimalarials that block replenishment of Plasmodium liver stage parasitophorous vacuole membrane proteins

    PubMed Central

    Hanson, Kirsten K.; Ressurreição, Ana S.; Buchholz, Kathrin; Prudêncio, Miguel; Herman-Ornelas, Jonathan D.; Rebelo, Maria; Beatty, Wandy L.; Wirth, Dyann F.; Hänscheid, Thomas; Moreira, Rui; Marti, Matthias; Mota, Maria M.

    2013-01-01

    Residence within a customized vacuole is a highly successful strategy used by diverse intracellular microorganisms. The parasitophorous vacuole membrane (PVM) is the critical interface between Plasmodium parasites and their possibly hostile, yet ultimately sustaining, host cell environment. We show that torins, developed as ATP-competitive mammalian target of rapamycin (mTOR) kinase inhibitors, are fast-acting antiplasmodial compounds that unexpectedly target the parasite directly, blocking the dynamic trafficking of the Plasmodium proteins exported protein 1 (EXP1) and upregulated in sporozoites 4 (UIS4) to the liver stage PVM and leading to efficient parasite elimination by the hepatocyte. Torin2 has single-digit, or lower, nanomolar potency in both liver and blood stages of infection in vitro and is likewise effective against both stages in vivo, with a single oral dose sufficient to clear liver stage infection. Parasite elimination and perturbed trafficking of liver stage PVM-resident proteins are both specific aspects of torin-mediated Plasmodium liver stage inhibition, indicating that torins have a distinct mode of action compared with currently used antimalarials. PMID:23836641

  12. 76 FR 5832 - International Business Machines (IBM), Software Group Business Unit, Optim Data Studio Tools QA...

    Federal Register 2010, 2011, 2012, 2013, 2014

    2011-02-02

    ... DEPARTMENT OF LABOR Employment and Training Administration [TA-W-74,554] International Business Machines (IBM), Software Group Business Unit, Optim Data Studio Tools QA, San Jose, CA; Notice of... determination of the TAA petition filed on behalf of workers at International Business Machines (IBM), Software...

  13. Lactobacillus priming of the respiratory tract: Heterologous immunity and protection against lethal pneumovirus infection.

    PubMed

    Garcia-Crespo, Katia E; Chan, Calvin C; Gabryszewski, Stanislaw J; Percopo, Caroline M; Rigaux, Peter; Dyer, Kimberly D; Domachowske, Joseph B; Rosenberg, Helene F

    2013-03-01

    We showed previously that wild-type mice primed via intranasal inoculation with live or heat-inactivated Lactobacillus species were fully (100%) protected against the lethal sequelae of infection with the virulent pathogen, pneumonia virus of mice (PVM), a response that is associated with diminished expression of proinflammatory cytokines and diminished virus recovery. We show here that 40% of the mice primed with live Lactobacillus survived when PVM challenge was delayed for 5months. This robust and sustained resistance to PVM infection resulting from prior interaction with an otherwise unrelated microbe is a profound example of heterologous immunity. We undertook the present study in order to understand the nature and unique features of this response. We found that intranasal inoculation with L. reuteri elicited rapid, transient neutrophil recruitment in association with proinflammatory mediators (CXCL1, CCL3, CCL2, CXCL10, TNF-alpha and IL-17A) but not Th1 cytokines. IFNγ does not contribute to survival promoted by Lactobacillus-priming. Live L. reuteri detected in lung tissue underwent rapid clearance, and was undetectable at 24h after inoculation. In contrast, L. reuteri peptidoglycan (PGN) and L. reuteri genomic DNA (gDNA) were detected at 24 and 48h after inoculation, respectively. In contrast to live bacteria, intranasal inoculation with isolated L. reuteri gDNA elicited no neutrophil recruitment, had minimal impact on virus recovery and virus-associated production of CCL3, and provided no protection against the negative sequelae of virus infection. Isolated PGN elicited neutrophil recruitment and proinflammatory cytokines but did not promote sustained survival in response to subsequent PVM infection. Overall, further evaluation of the responses leading to Lactobacillus-mediated heterologous immunity may provide insight into novel antiviral preventive modalities. Published by Elsevier B.V.

  14. Lactobacillus priming of the respiratory tract: heterologous immunity and protection against lethal pneumovirus infection

    PubMed Central

    Garcia-Crespo, Katia E.; Chan, Calvin C.; Gabryszewski, Stanislaw J.; Percopo, Caroline M.; Rigaux, Peter; Dyer, Kimberly D.; Domachowske, Joseph B.; Rosenberg, Helene F.

    2013-01-01

    We showed previously that wild-type mice primed via intranasal inoculation with live or heat-inactivated Lactobacillus species were fully (100%) protected against the lethal sequelae of infection with the virulent pathogen, pneumonia virus of mice (PVM), a response that is associated with diminished expression of proinflammatory cytokines and diminished virus recovery. We show here that 40% of the mice primed with live Lactobacillus survived when PVM challenge was delayed for 5 months. This robust and sustained resistance to PVM infection resulting from prior interaction with an otherwise unrelated microbe is a profound example of heterologous immunity. We undertook the present study in order to understand the nature and unique features of this response. We found that intranasal inoculation with L. reuteri elicited rapid, transient neutrophil recruitment in association with proinflammatory mediators (CXCL1, CCL3, CCL2, CXCL10, TNF-alpha and IL-17A) but not Th1 cytokines. IFNγ does not contribute to survival promoted by Lactobacillus-priming. Live L. reuteri detected in lung tissue underwent rapid clearance, and was undetectable at 24 hrs after inoculation. In contrast, L. reuteri peptidoglycan (PGN) and L. reuteri genomic DNA (gDNA) were detected at 24 and 48 hours after inoculation, respectively. In contrast to live bacteria, intranasal inoculation with isolated L. reuteri gDNA elicited no neutrophil recruitment, had minimal impact on virus recovery and virus-associated production of CCL3, and provided no protection against the negative sequelae of virus infection. Isolated PGN elicited neutrophil recruitment and proinflammatory cytokines but did not promote sustained survival in response to subsequent PVM infection. Overall, further evaluation of the responses leading to Lactobacillus-mediated heterologous immunity may provide insight into novel antiviral preventive modalities. PMID:23274789

  15. Effect of a triclosan/PVM/MA copolymer/fluoride dentifrice on volatile sulfur compounds in vitro.

    PubMed

    Pilch, S; Williams, M I; Cummins, D

    2005-01-01

    The objective of the investigation was to document the in vitro efficacy of a triclosan/PVM/MA copolymer/fluoride (TCF) dentifrice against the formation of volatile sulfur compounds (VSC) as well as the growth of H2S-producing bacteria. Clinical studies using organoleptic judges, gas chromatography, or a portable sulfide monitor have generally been employed in the assessment of treatments for the control of oral malodor. However, these studies are not appropriate for screening purposes because of the expense and time required. An in vitro method was developed for the purpose of screening new compounds, agents or formulations for their ability to control VSC formation and for determining bio-equivalence of efficacy when implementing changes in existing formulations. The method combines basic microbiological methods, dynamic flow cell techniques and head space analysis. The in vitro VSC method was validated by comparing the efficacy of two dentifrices containing TCF with a control fluoride dentifrice as the TCF products have been clinically proven to control oral malodor. In the validation studies, the TCF-containing dentifrices were significantly better (P < 0.05) than the control dentifrice in inhibiting VSC formation and reducing H(2)S-producing bacteria. For example, when compared with baseline, the TCF dentifrices reduced VSC formation between 42 and 49% compared with the control dentifrice which reduced VSC formation 3%. There was no significant difference (P > 0.05) between the two TCF dentifrice formulations. Using an in vitro breath VSC model, it has been demonstrated that two variants of a dentifrice containing triclosan, PVM/MA copolymer and fluoride have efficacy that is significantly better than a control fluoridated dentifrice and that there is no significant difference between the triclosan/PVM/MA copolymer/fluoride dentifrice variants.

  16. Analysis of machining accuracy during free form surface milling simulation for different milling strategies

    NASA Astrophysics Data System (ADS)

    Matras, A.; Kowalczyk, R.

    2014-11-01

    The analysis results of machining accuracy after the free form surface milling simulations (based on machining EN AW- 7075 alloys) for different machining strategies (Level Z, Radial, Square, Circular) are presented in the work. Particular milling simulations were performed using CAD/CAM Esprit software. The accuracy of obtained allowance is defined as a difference between the theoretical surface of work piece element (the surface designed in CAD software) and the machined surface after a milling simulation. The difference between two surfaces describes a value of roughness, which is as the result of tool shape mapping on the machined surface. Accuracy of the left allowance notifies in direct way a surface quality after the finish machining. Described methodology of usage CAD/CAM software can to let improve a time design of machining process for a free form surface milling by a 5-axis CNC milling machine with omitting to perform the item on a milling machine in order to measure the machining accuracy for the selected strategies and cutting data.

  17. Heterogeneous voter models

    NASA Astrophysics Data System (ADS)

    Masuda, Naoki; Gibert, N.; Redner, S.

    2010-07-01

    We introduce the heterogeneous voter model (HVM), in which each agent has its own intrinsic rate to change state, reflective of the heterogeneity of real people, and the partisan voter model (PVM), in which each agent has an innate and fixed preference for one of two possible opinion states. For the HVM, the time until consensus is reached is much longer than in the classic voter model. For the PVM in the mean-field limit, a population evolves to a preference-based state, where each agent tends to be aligned with its internal preference. For finite populations, discrete fluctuations ultimately lead to consensus being reached in a time that scales exponentially with population size.

  18. Monitoring of antisolvent crystallization of sodium scutellarein by combined FBRM-PVM-NIR.

    PubMed

    Liu, Xuesong; Sun, Di; Wang, Feng; Wu, Yongjiang; Chen, Yong; Wang, Longhu

    2011-06-01

    Antisolvent crystallization can be used as an alternative to cooling or evaporation for the separation and purification of solid product in the pharmaceutical industry. To improve the process understanding of antisolvent crystallization, the use of in-line tools is vital. In this study, the process analytical technology (PAT) tools including focused beam reflectance measurement (FBRM), particle video microscope (PVM), and near-infrared spectroscopy (NIRS) were utilized to monitor antisolvent crystallization of sodium scutellarein. FBRM was used to monitor chord count and chord length distribution of sodium scutellarein particles in the crystallizer, and PVM, as an in-line video camera, provided pictures imaging particle shape and dimension. In addition, a quantitative model of PLS was established by in-line NIRS to detect the concentration of sodium scutellarein in the solvent and good calibration statistics were obtained (r(2) = 0.976) with the residual predictive deviation value of 11.3. The discussion over sensitivities, strengths, and weaknesses of the PAT tools may be helpful in selection of suitable PAT techniques. These in-line techniques eliminate the need for sample preparation and offer a time-saving approach to understand and monitor antisolvent crystallization process. Copyright © 2011 Wiley-Liss, Inc.

  19. System software for the finite element machine

    NASA Technical Reports Server (NTRS)

    Crockett, T. W.; Knott, J. D.

    1985-01-01

    The Finite Element Machine is an experimental parallel computer developed at Langley Research Center to investigate the application of concurrent processing to structural engineering analysis. This report describes system-level software which has been developed to facilitate use of the machine by applications researchers. The overall software design is outlined, and several important parallel processing issues are discussed in detail, including processor management, communication, synchronization, and input/output. Based on experience using the system, the hardware architecture and software design are critiqued, and areas for further work are suggested.

  20. Machine Tool Software

    NASA Technical Reports Server (NTRS)

    1988-01-01

    A NASA-developed software package has played a part in technical education of students who major in Mechanical Engineering Technology at William Rainey Harper College. Professor Hack has been using (APT) Automatically Programmed Tool Software since 1969 in his CAD/CAM Computer Aided Design and Manufacturing curriculum. Professor Hack teaches the use of APT programming languages for control of metal cutting machines. Machine tool instructions are geometry definitions written in APT Language to constitute a "part program." The part program is processed by the machine tool. CAD/CAM students go from writing a program to cutting steel in the course of a semester.

  1. Computation of Coupled Thermal-Fluid Problems in Distributed Memory Environment

    NASA Technical Reports Server (NTRS)

    Wei, H.; Shang, H. M.; Chen, Y. S.

    2001-01-01

    The thermal-fluid coupling problems are very important to aerospace and engineering applications. Instead of analyzing heat transfer and fluid flow separately, this study merged two well-accepted engineering solution methods, SINDA for thermal analysis and FDNS for fluid flow simulation, into a unified multi-disciplinary thermal fluid prediction method. A fully conservative patched grid interface algorithm for arbitrary two-dimensional and three-dimensional geometry has been developed. The state-of-the-art parallel computing concept was used to couple SINDA and FDNS for the communication of boundary conditions through PVM (Parallel Virtual Machine) libraries. Therefore, the thermal analysis performed by SINDA and the fluid flow calculated by FDNS are fully coupled to obtain steady state or transient solutions. The natural convection between two thick-walled eccentric tubes was calculated and the predicted results match the experiment data perfectly. A 3-D rocket engine model and a real 3-D SSME geometry were used to test the current model, and the reasonable temperature field was obtained.

  2. Searching for patterns in remote sensing image databases using neural networks

    NASA Technical Reports Server (NTRS)

    Paola, Justin D.; Schowengerdt, Robert A.

    1995-01-01

    We have investigated a method, based on a successful neural network multispectral image classification system, of searching for single patterns in remote sensing databases. While defining the pattern to search for and the feature to be used for that search (spectral, spatial, temporal, etc.) is challenging, a more difficult task is selecting competing patterns to train against the desired pattern. Schemes for competing pattern selection, including random selection and human interpreted selection, are discussed in the context of an example detection of dense urban areas in Landsat Thematic Mapper imagery. When applying the search to multiple images, a simple normalization method can alleviate the problem of inconsistent image calibration. Another potential problem, that of highly compressed data, was found to have a minimal effect on the ability to detect the desired pattern. The neural network algorithm has been implemented using the PVM (Parallel Virtual Machine) library and nearly-optimal speedups have been obtained that help alleviate the long process of searching through imagery.

  3. Quantum machine learning.

    PubMed

    Biamonte, Jacob; Wittek, Peter; Pancotti, Nicola; Rebentrost, Patrick; Wiebe, Nathan; Lloyd, Seth

    2017-09-13

    Fuelled by increasing computer power and algorithmic advances, machine learning techniques have become powerful tools for finding patterns in data. Quantum systems produce atypical patterns that classical systems are thought not to produce efficiently, so it is reasonable to postulate that quantum computers may outperform classical computers on machine learning tasks. The field of quantum machine learning explores how to devise and implement quantum software that could enable machine learning that is faster than that of classical computers. Recent work has produced quantum algorithms that could act as the building blocks of machine learning programs, but the hardware and software challenges are still considerable.

  4. Quantum machine learning

    NASA Astrophysics Data System (ADS)

    Biamonte, Jacob; Wittek, Peter; Pancotti, Nicola; Rebentrost, Patrick; Wiebe, Nathan; Lloyd, Seth

    2017-09-01

    Fuelled by increasing computer power and algorithmic advances, machine learning techniques have become powerful tools for finding patterns in data. Quantum systems produce atypical patterns that classical systems are thought not to produce efficiently, so it is reasonable to postulate that quantum computers may outperform classical computers on machine learning tasks. The field of quantum machine learning explores how to devise and implement quantum software that could enable machine learning that is faster than that of classical computers. Recent work has produced quantum algorithms that could act as the building blocks of machine learning programs, but the hardware and software challenges are still considerable.

  5. Software architecture standard for simulation virtual machine, version 2.0

    NASA Technical Reports Server (NTRS)

    Sturtevant, Robert; Wessale, William

    1994-01-01

    The Simulation Virtual Machine (SBM) is an Ada architecture which eases the effort involved in the real-time software maintenance and sustaining engineering. The Software Architecture Standard defines the infrastructure which all the simulation models are built from. SVM was developed for and used in the Space Station Verification and Training Facility.

  6. Research on axisymmetric aspheric surface numerical design and manufacturing technology

    NASA Astrophysics Data System (ADS)

    Wang, Zhen-zhong; Guo, Yin-biao; Lin, Zheng

    2006-02-01

    The key technology for aspheric machining offers exact machining path and machining aspheric lens with high accuracy and efficiency, in spite of the development of traditional manual manufacturing into nowadays numerical control (NC) machining. This paper presents a mathematical model between virtual cone and aspheric surface equations, and discusses the technology of uniform wear of grinding wheel and error compensation in aspheric machining. Finally, a software system for high precision aspheric surface manufacturing is designed and realized, based on the mentioned above. This software system can work out grinding wheel path according to input parameters and generate machining NC programs of aspheric surfaces.

  7. Intelligible machine learning with malibu.

    PubMed

    Langlois, Robert E; Lu, Hui

    2008-01-01

    malibu is an open-source machine learning work-bench developed in C/C++ for high-performance real-world applications, namely bioinformatics and medical informatics. It leverages third-party machine learning implementations for more robust bug-free software. This workbench handles several well-studied supervised machine learning problems including classification, regression, importance-weighted classification and multiple-instance learning. The malibu interface was designed to create reproducible experiments ideally run in a remote and/or command line environment. The software can be found at: http://proteomics.bioengr. uic.edu/malibu/index.html.

  8. Fundamental Roles of the Golgi-Associated Toxoplasma Aspartyl Protease, ASP5, at the Host-Parasite Interface

    PubMed Central

    Hammoudi, Pierre-Mehdi; Jacot, Damien; Mueller, Christina; Di Cristina, Manlio; Dogga, Sunil Kumar; Marq, Jean-Baptiste; Romano, Julia; Tosetti, Nicolò; Dubrot, Juan; Emre, Yalin; Lunghi, Matteo; Coppens, Isabelle; Yamamoto, Masahiro; Sojka, Daniel; Pino, Paco; Soldati-Favre, Dominique

    2015-01-01

    Toxoplasma gondii possesses sets of dense granule proteins (GRAs) that either assemble at, or cross the parasitophorous vacuole membrane (PVM) and exhibit motifs resembling the HT/PEXEL previously identified in a repertoire of exported Plasmodium proteins. Within Plasmodium spp., cleavage of the HT/PEXEL motif by the endoplasmic reticulum-resident protease Plasmepsin V precedes trafficking to and export across the PVM of proteins involved in pathogenicity and host cell remodelling. Here, we have functionally characterized the T. gondii aspartyl protease 5 (ASP5), a Golgi-resident protease that is phylogenetically related to Plasmepsin V. We show that deletion of ASP5 causes a significant loss in parasite fitness in vitro and an altered virulence in vivo. Furthermore, we reveal that ASP5 is necessary for the cleavage of GRA16, GRA19 and GRA20 at the PEXEL-like motif. In the absence of ASP5, the intravacuolar nanotubular network disappears and several GRAs fail to localize to the PVM, while GRA16 and GRA24, both known to be targeted to the host cell nucleus, are retained within the vacuolar space. Additionally, hypermigration of dendritic cells and bradyzoite cyst wall formation are impaired, critically impacting on parasite dissemination and persistence. Overall, the absence of ASP5 dramatically compromises the parasite’s ability to modulate host signalling pathways and immune responses. PMID:26473595

  9. An intelligent CNC machine control system architecture

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Miller, D.J.; Loucks, C.S.

    1996-10-01

    Intelligent, agile manufacturing relies on automated programming of digitally controlled processes. Currently, processes such as Computer Numerically Controlled (CNC) machining are difficult to automate because of highly restrictive controllers and poor software environments. It is also difficult to utilize sensors and process models for adaptive control, or to integrate machining processes with other tasks within a factory floor setting. As part of a Laboratory Directed Research and Development (LDRD) program, a CNC machine control system architecture based on object-oriented design and graphical programming has been developed to address some of these problems and to demonstrate automated agile machining applications usingmore » platform-independent software.« less

  10. VML 3.0 Reactive Sequencing Objects and Matrix Math Operations for Attitude Profiling

    NASA Technical Reports Server (NTRS)

    Grasso, Christopher A.; Riedel, Joseph E.

    2012-01-01

    VML (Virtual Machine Language) has been used as the sequencing flight software on over a dozen JPL deep-space missions, most recently flying on GRAIL and JUNO. In conjunction with the NASA SBIR entitled "Reactive Rendezvous and Docking Sequencer", VML version 3.0 has been enhanced to include object-oriented element organization, built-in queuing operations, and sophisticated matrix / vector operations. These improvements allow VML scripts to easily perform much of the work that formerly would have required a great deal of expensive flight software development to realize. Autonomous turning and tracking makes considerable use of new VML features. Profiles generated by flight software are managed using object-oriented VML data constructs executed in discrete time by the VML flight software. VML vector and matrix operations provide the ability to calculate and supply quaternions to the attitude controller flight software which produces torque requests. Using VML-based attitude planning components eliminates flight software development effort, and reduces corresponding costs. In addition, the direct management of the quaternions allows turning and tracking to be tied in with sophisticated high-level VML state machines. These state machines provide autonomous management of spacecraft operations during critical tasks like a hypothetic Mars sample return rendezvous and docking. State machines created for autonomous science observations can also use this sort of attitude planning system, allowing heightened autonomy levels to reduce operations costs. VML state machines cannot be considered merely sequences - they are reactive logic constructs capable of autonomous decision making within a well-defined domain. The state machine approach enabled by VML 3.0 is progressing toward flight capability with a wide array of applicable mission activities.

  11. A clinical investigation of the efficacy of three commercially available dentifrices for controlling established gingivitis and supragingival plaque.

    PubMed

    Singh, Surrendra; Chaknis, Patricia; DeVizio, William; Petrone, Margaret; Panagakos, Fotinos S; Proskin, Howard M

    2010-01-01

    To assess the efficacy ofa dentifrice containing 0.3% triclosan/2.0% PVM/MA copolymer/0.243% sodium fluoride for controlling established gingivitis and supragingival plaque relative to that of a dentifrice containing 0.454% stannous fluoride, sodium hexametaphosphate, and zinc lactate, and a dentifrice containing 0.243% sodium fluoride as a negative control. Following a baseline examination for gingivitis and supragingival plaque, qualifying adult male and female subjects from the Piscataway, NJ, USA area were randomized into three dentifrice groups. Subjects were instructed to brush their teeth twice daily (morning and evening) for one minute with their assigned dentifrice and a soft-bristled toothbrush. Examinations for gingivitis and supragingival plaque were repeated after six weeks of product use. One-hundred and seventy-one (171) subjects complied with the protocol and completed the study. Relative to the group using the dentifrice with 0.243% sodium fluoride alone, the 0.3% triclosan/2.0% PVM/MA copolymer/0.243% sodium fluoride group exhibited statistically significant reductions in gingival index and supragingival plaque index scores of 25.3% and 33.0%, respectively, after six weeks of product use. Similarly, relative to the group using the 0.243% sodium fluoride dentifrice, the 0.454% stannous fluoride, sodium hexametaphosphate, and zinc lactate dentifrice group exhibited statistically significant reductions in gingival index and plaque index scores of 8.1% and 14.1% after six weeks of product use. Further, relative to the 0.454% stannous fluoride, sodium hexametaphosphate, and zinc lactate dentifrice group, the 0.3% triclosan/2.0% PVM/MA copolymer/0.243% sodium fluoride dentifrice group exhibited statistically significant reductions in gingival index and plaque index scores of 18.7% and 22%, respectively. The overall results of this double-blind clinical study support the conclusion that a dentifrice containing 0.3% triclosan/2.0% PVM/MA copolymer/0.243% sodium fluoride is efficacious for the control of established gingivitis and supragingival plaque as compared to a regular fluoride dentifrice, and that it provides a greater level of efficacy for the control of gingivitis and supragingival plaque than does a dentifrice containing 0.454% stannous fluoride, sodium hexametaphosphate, and zinc lactate.

  12. Muscarinic excitation of parvalbumin-positive interneurons contributes to the severity of pilocarpine-induced seizures

    PubMed Central

    Yi, Feng; DeCan, Evan; Stoll, Kurt; Marceau, Eric; Deisseroth, Karl; Lawrence, J. Josh

    2014-01-01

    SUMMARY Objective A common rodent model in epilepsy research employs the muscarinic acetylcholine receptor (mAChR) agonist pilocarpine, yet the mechanisms underlying the induction of pilocarpine-induced seizures (PISs) remain unclear. Global M1 mAChR (M1R) knockout mice are resistant to PISs, implying that M1R activation disrupts excitation/inhibition balance. Parvalbumin-positive (PV) inhibitory neurons express M1 mAChRs, participate in cholinergically-induced oscillations, and can enter a state of depolarization block (DB) during epileptiform activity. Here, we test the hypothesis that pilocarpine activation of M1Rs expressed on PV cells contributes to PISs. Methods CA1 PV cells in PV-CRE mice were visualized with a floxed YFP or hM3Dq-mCherry adeno-associated virus, or by crossing PV-CRE mice with the RosaYFP reporter line. To eliminate M1Rs from PV cells, we generated PV-M1KO mice by crossing PV-CRE and floxed M1 mice. Action potential (AP) frequency was monitored during application of pilocarpine (200 µM). In behavioral experiments, locomotion and seizure symptoms were recorded in WT or PV-M1KO mice during PISs. Results Pilocarpine significantly increased AP frequency in CA1 PV cells into the gamma range. In the continued presence of pilocarpine, a subset (5/7) of PV cells progressed to DB, which was mimicked by hM3Dq activation of Gq-receptor signaling. Pilocarpine-induced depolarization, AP firing at gamma frequency, and progression to DB were prevented in CA1 PV cells of PV-M1KO mice. Finally, compared to WT mice, PV-M1KO mice were associated with reduced severity of PISs. Significance Pilocarpine can directly depolarize PV+ cells via M1R activation but a subset of these cells progress to DB. Our electrophysiological and behavioral results suggest that this mechanism is active during PISs, contributing to a collapse of PV-mediated GABAergic inhibition, dysregulation of excitation/inhibition balance, and increased susceptibility to PISs. PMID:25495999

  13. An open CAM system for dentistry on the basis of China-made 5-axis simultaneous contouring CNC machine tool and industrial CAM software.

    PubMed

    Lu, Li; Liu, Shusheng; Shi, Shenggen; Yang, Jianzhong

    2011-10-01

    China-made 5-axis simultaneous contouring CNC machine tool and domestically developed industrial computer-aided manufacture (CAM) technology were used for full crown fabrication and measurement of crown accuracy, with an attempt to establish an open CAM system for dental processing and to promote the introduction of domestic dental computer-aided design (CAD)/CAM system. Commercially available scanning equipment was used to make a basic digital tooth model after preparation of crown, and CAD software that comes with the scanning device was employed to design the crown by using domestic industrial CAM software to process the crown data in order to generate a solid model for machining purpose, and then China-made 5-axis simultaneous contouring CNC machine tool was used to complete machining of the whole crown and the internal accuracy of the crown internal was measured by using 3D-MicroCT. The results showed that China-made 5-axis simultaneous contouring CNC machine tool in combination with domestic industrial CAM technology can be used for crown making and the crown was well positioned in die. The internal accuracy was successfully measured by using 3D-MicroCT. It is concluded that an open CAM system for dentistry on the basis of China-made 5-axis simultaneous contouring CNC machine tool and domestic industrial CAM software has been established, and development of the system will promote the introduction of domestically-produced dental CAD/CAM system.

  14. Heterogeneous concurrent computing with exportable services

    NASA Technical Reports Server (NTRS)

    Sunderam, Vaidy

    1995-01-01

    Heterogeneous concurrent computing, based on the traditional process-oriented model, is approaching its functionality and performance limits. An alternative paradigm, based on the concept of services, supporting data driven computation, and built on a lightweight process infrastructure, is proposed to enhance the functional capabilities and the operational efficiency of heterogeneous network-based concurrent computing. TPVM is an experimental prototype system supporting exportable services, thread-based computation, and remote memory operations that is built as an extension of and an enhancement to the PVM concurrent computing system. TPVM offers a significantly different computing paradigm for network-based computing, while maintaining a close resemblance to the conventional PVM model in the interest of compatibility and ease of transition Preliminary experiences have demonstrated that the TPVM framework presents a natural yet powerful concurrent programming interface, while being capable of delivering performance improvements of upto thirty percent.

  15. Preparation and characterization of mucoadhesive nanoparticles of poly (methyl vinyl ether-co-maleic anhydride) containing glycyrrhizic acid intended for vaginal administration.

    PubMed

    Aguilar-Rosas, Irene; Alcalá-Alcalá, Sergio; Llera-Rojas, Viridiana; Ganem-Rondero, Adriana

    2015-01-01

    Traditional vaginal preparations reside in the vaginal cavity for relatively a short period of time, requiring multiple doses in order to attain the desired therapeutic effect. Therefore, mucoadhesive systems appear to be appropriate to prolong the residence time in the vaginal cavity. In the current study, mucoadhesive nanoparticles based on poly(methyl vinyl ether-co-maleic anhydride) (PVM/MA) intended for vaginal delivery of glycyrrhizic acid (GA) (a drug with well-known antiviral properties) were prepared and characterized. Nanoparticles were generated by a solvent displacement method. Incorporation of GA was performed during nanoprecipitation, followed by adsorption of drug once nanoparticles were formed. The prepared nanoparticles were characterized in terms of size, Z-potential, morphology, drug loading, interaction of GA with PVM/MA (by differential scanning calorimetry) and the in vitro interaction of nanoparticles with pig mucin (at two pH values, 3.6 and 5; with and without GA adsorbed). The preparation method led to nanoparticles of a mean diameter of 198.5 ± 24.3 nm, zeta potential of -44.8 ± 2.8 mV and drug loading of 15.07 ± 0.86 µg/mg polymer. The highest mucin interaction resulted at pH 3.6 for nanoparticles without GA adsorbed. The data obtained suggest the promise of using mucoadhesive nanoparticles of PVM/MA for intravaginal delivery of GA.

  16. Non-equilibrium plasma kinetics of reacting CO: an improved state to state approach

    NASA Astrophysics Data System (ADS)

    Pietanza, L. D.; Colonna, G.; Capitelli, M.

    2017-12-01

    Non-equilibrium plasma kinetics of reacting CO for conditions typically met in microwave discharges have been developed based on the coupling of excited state kinetics and the Boltzmann equation for the electron energy distribution function (EEDF). Particular attention is given to the insertion in the vibrational kinetics of a complete set of electron molecule resonant processes linking the whole vibrational ladder of the CO molecule, as well as to the role of Boudouard reaction, i.e. the process of forming CO2 by two vibrationally excited CO molecules, in shaping the vibrational distribution of CO and promoting reaction channels assisted by vibrational excitation (pure vibrational mechanisms, PVM). PVM mechanisms can become competitive with electron impact dissociation processes (DEM) in the activation of CO. A case study reproducing the conditions of a microwave discharge has been considered following the coupled kinetics also in the post discharge conditions. Results include the evolution of EEDF in discharge and post discharge conditions highlighting the role of superelastic vibrational and electronic collisions in shaping the EEDF. Moreover, PVM rate coefficients and DEM ones are studied as a function of gas temperature, showing a non-Arrhenius behavior, i.e. the rate coefficients increase with decreasing gas temperature as a result of a vibrational-vibrational (V-V) pumping up mechanism able to form plateaux in the vibrational distribution function. The accuracy of the results is discussed in particular in connection to the present knowledge of the activation energy of the Boudouard process.

  17. Effects of Ramadan Gasting on Postural Balance and Attentional Capacities in Elderly People.

    PubMed

    Laatar, R; Borji, R; Baccouch, R; Zahaf, F; Rebai, H; Sahli, S

    2016-01-01

    To evaluate the effects of Ramadan fasting on postural balance and attentional capacities in older adults. the Neurophysiology department of a University Hospital. Fifteen males aged between 65 and 80 years were asked to perform a postural balance protocol and a simple reaction time (SRT) test in four testing phases: one week before Ramadan (BR), during the second (SWR) and the fourth week of Ramadan (FWR) and 3 weeks after Ramadan (AR). Postural balance measurements were recorded in the bipedal stance in four different conditions: firm surface/eyes open (EO), firm surface/eyes closed (EC), foam surface/EO and foam surface/EC using a force platform. Results of the present study demonstrated that center of pressure (CoP) mean velocity (CoPVm), medio-lateral length (CoPLX) and antero-posterior length (CoPLY) were significantly higher during the SWR than BR. Likewise, values of CoPVm and CoPLX increased significantly during the FWR compared to BR. The CoPLX decreased significantly in the FWR compared to the SWR. Values of CoPVm and CoPLX were significantly higher AR in comparison with BR. In addition, SRT values increased significantly during the SWR and the FWR than BR. Ramadan fasting affects postural balance and attentional capacities in the elderly mainly in the SWR and it may, therefore, increase the risk of fall and fall-related injuries. More than three weeks are needed for older adults to recover postural balance impairment due to Ramadan fasting.

  18. Power Electronics and Electric Machines Publications | Transportation

    Science.gov Websites

    electric machines. For more information about the following publications, contact Sreekant Narumanchi. A , NREL Software Spray System Evaluation (Software 1.1 MB) Papers 2017 Electric Motor Thermal Management Source: Douglas DeVoto. 2017. 14 pp. NREL/MP-5400-67117. Power Electronics Thermal Management Research

  19. Real Time Network Monitoring and Reporting System

    ERIC Educational Resources Information Center

    Massengale, Ricky L., Sr.

    2009-01-01

    With the ability of modern system developers to develop intelligent programs that allows machines to learn, modify and evolve themselves, current trends of reactionary methods to detect and eradicate malicious software code from infected machines is proving to be too costly. Addressing malicious software after an attack is the current methodology…

  20. Open architecture CNC system

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Tal, J.; Lopez, A.; Edwards, J.M.

    1995-04-01

    In this paper, an alternative solution to the traditional CNC machine tool controller has been introduced. Software and hardware modules have been described and their incorporation in a CNC control system has been outlined. This type of CNC machine tool controller demonstrates that technology is accessible and can be readily implemented into an open architecture machine tool controller. Benefit to the user is greater controller flexibility, while being economically achievable. PC based, motion as well as non-motion features will provide flexibility through a Windows environment. Up-grading this type of controller system through software revisions will keep the machine tool inmore » a competitive state with minimal effort. Software and hardware modules are mass produced permitting competitive procurement and incorporation. Open architecture CNC systems provide diagnostics thus enhancing maintainability, and machine tool up-time. A major concern of traditional CNC systems has been operator training time. Training time can be greatly minimized by making use of Windows environment features.« less

  1. Parallelization of Rocket Engine Simulator Software (PRESS)

    NASA Technical Reports Server (NTRS)

    Cezzar, Ruknet

    1998-01-01

    We have outlined our work in the last half of the funding period. We have shown how a demo package for RESSAP using MPI can be done. However, we also mentioned the difficulties with the UNIX platform. We have reiterated some of the suggestions made during the presentation of the progress of the at Fourth Annual HBCU Conference. Although we have discussed, in some detail, how TURBDES/PUMPDES software can be run in parallel using MPI, at present, we are unable to experiment any further with either MPI or PVM. Due to X windows not being implemented, we are also not able to experiment further with XPVM, which it will be recalled, has a nice GUI interface. There are also some concerns, on our part, about MPI being an appropriate tool. The best thing about MPr is that it is public domain. Although and plenty of documentation exists for the intricacies of using MPI, little information is available on its actual implementations. Other than very typical, somewhat contrived examples, such as Jacobi algorithm for solving Laplace's equation, there are few examples which can readily be applied to real situations, such as in our case. In effect, the review of literature on both MPI and PVM, and there is a lot, indicate something similar to the enormous effort which was spent on LISP and LISP-like languages as tools for artificial intelligence research. During the development of a book on programming languages [12], when we searched the literature for very simple examples like taking averages, reading and writing records, multiplying matrices, etc., we could hardly find a any! Yet, so much was said and done on that topic in academic circles. It appears that we faced the same problem with MPI, where despite significant documentation, we could not find even a simple example which supports course-grain parallelism involving only a few processes. From the foregoing, it appears that a new direction may be required for more productive research during the extension period (10/19/98 - 10/18/99). At the least, the research would need to be done on Windows 95/Windows NT based platforms. Moreover, with the acquisition of Lahey Fortran package for PC platform, and the existing Borland C + + 5. 0, we can do work on C + + wrapper issues. We have carefully studied the blueprint for Space Transportation Propulsion Integrated Design Environment for the next 25 years [13] and found the inclusion of HBCUs in that effort encouraging. Especially in the long period for which a map is provided, there is no doubt that HBCUs will grow and become better equipped to do meaningful research. In the shorter period, as was suggested in our presentation at the HBCU conference, some key decisions regarding the aging Fortran based software for rocket propellants will need to be made. One important issue is whether or not object oriented languages such as C + + or Java should be used for distributed computing. Whether or not "distributed computing" is necessary for the existing software is yet another, larger, question to be tackled with.

  2. Development of the FITS tools package for multiple software environments

    NASA Technical Reports Server (NTRS)

    Pence, W. D.; Blackburn, J. K.

    1992-01-01

    The HEASARC is developing a package of general purpose software for analyzing data files in FITS format. This paper describes the design philosophy which makes the software both machine-independent (it runs on VAXs, Suns, and DEC-stations) and software environment-independent. Currently the software can be compiled and linked to produce IRAF tasks, or alternatively, the same source code can be used to generate stand-alone tasks using one of two implementations of a user-parameter interface library. The machine independence of the software is achieved by writing the source code in ANSI standard Fortran or C, using the machine-independent FITSIO subroutine interface for all data file I/O, and using a standard user-parameter subroutine interface for all user I/O. The latter interface is based on the Fortran IRAF Parameter File interface developed at STScI. The IRAF tasks are built by linking to the IRAF implementation of this parameter interface library. Two other implementations of this parameter interface library, which have no IRAF dependencies, are now available which can be used to generate stand-alone executable tasks. These stand-alone tasks can simply be executed from the machine operating system prompt either by supplying all the task parameters on the command line or by entering the task name after which the user will be prompted for any required parameters. A first release of this FTOOLS package is now publicly available. The currently available tasks are described, along with instructions on how to obtain a copy of the software.

  3. Acquiring Software Design Schemas: A Machine Learning Perspective

    NASA Technical Reports Server (NTRS)

    Harandi, Mehdi T.; Lee, Hing-Yan

    1991-01-01

    In this paper, we describe an approach based on machine learning that acquires software design schemas from design cases of existing applications. An overview of the technique, design representation, and acquisition system are presented. the paper also addresses issues associated with generalizing common features such as biases. The generalization process is illustrated using an example.

  4. Toward a Progress Indicator for Machine Learning Model Building and Data Mining Algorithm Execution: A Position Paper.

    PubMed

    Luo, Gang

    2017-12-01

    For user-friendliness, many software systems offer progress indicators for long-duration tasks. A typical progress indicator continuously estimates the remaining task execution time as well as the portion of the task that has been finished. Building a machine learning model often takes a long time, but no existing machine learning software supplies a non-trivial progress indicator. Similarly, running a data mining algorithm often takes a long time, but no existing data mining software provides a nontrivial progress indicator. In this article, we consider the problem of offering progress indicators for machine learning model building and data mining algorithm execution. We discuss the goals and challenges intrinsic to this problem. Then we describe an initial framework for implementing such progress indicators and two advanced, potential uses of them, with the goal of inspiring future research on this topic.

  5. Toward a Progress Indicator for Machine Learning Model Building and Data Mining Algorithm Execution: A Position Paper

    PubMed Central

    Luo, Gang

    2017-01-01

    For user-friendliness, many software systems offer progress indicators for long-duration tasks. A typical progress indicator continuously estimates the remaining task execution time as well as the portion of the task that has been finished. Building a machine learning model often takes a long time, but no existing machine learning software supplies a non-trivial progress indicator. Similarly, running a data mining algorithm often takes a long time, but no existing data mining software provides a nontrivial progress indicator. In this article, we consider the problem of offering progress indicators for machine learning model building and data mining algorithm execution. We discuss the goals and challenges intrinsic to this problem. Then we describe an initial framework for implementing such progress indicators and two advanced, potential uses of them, with the goal of inspiring future research on this topic. PMID:29177022

  6. Simulation Platform: a cloud-based online simulation environment.

    PubMed

    Yamazaki, Tadashi; Ikeno, Hidetoshi; Okumura, Yoshihiro; Satoh, Shunji; Kamiyama, Yoshimi; Hirata, Yutaka; Inagaki, Keiichiro; Ishihara, Akito; Kannon, Takayuki; Usui, Shiro

    2011-09-01

    For multi-scale and multi-modal neural modeling, it is needed to handle multiple neural models described at different levels seamlessly. Database technology will become more important for these studies, specifically for downloading and handling the neural models seamlessly and effortlessly. To date, conventional neuroinformatics databases have solely been designed to archive model files, but the databases should provide a chance for users to validate the models before downloading them. In this paper, we report our on-going project to develop a cloud-based web service for online simulation called "Simulation Platform". Simulation Platform is a cloud of virtual machines running GNU/Linux. On a virtual machine, various software including developer tools such as compilers and libraries, popular neural simulators such as GENESIS, NEURON and NEST, and scientific software such as Gnuplot, R and Octave, are pre-installed. When a user posts a request, a virtual machine is assigned to the user, and the simulation starts on that machine. The user remotely accesses to the machine through a web browser and carries out the simulation, without the need to install any software but a web browser on the user's own computer. Therefore, Simulation Platform is expected to eliminate impediments to handle multiple neural models that require multiple software. Copyright © 2011 Elsevier Ltd. All rights reserved.

  7. Reprint of: Simulation Platform: a cloud-based online simulation environment.

    PubMed

    Yamazaki, Tadashi; Ikeno, Hidetoshi; Okumura, Yoshihiro; Satoh, Shunji; Kamiyama, Yoshimi; Hirata, Yutaka; Inagaki, Keiichiro; Ishihara, Akito; Kannon, Takayuki; Usui, Shiro

    2011-11-01

    For multi-scale and multi-modal neural modeling, it is needed to handle multiple neural models described at different levels seamlessly. Database technology will become more important for these studies, specifically for downloading and handling the neural models seamlessly and effortlessly. To date, conventional neuroinformatics databases have solely been designed to archive model files, but the databases should provide a chance for users to validate the models before downloading them. In this paper, we report our on-going project to develop a cloud-based web service for online simulation called "Simulation Platform". Simulation Platform is a cloud of virtual machines running GNU/Linux. On a virtual machine, various software including developer tools such as compilers and libraries, popular neural simulators such as GENESIS, NEURON and NEST, and scientific software such as Gnuplot, R and Octave, are pre-installed. When a user posts a request, a virtual machine is assigned to the user, and the simulation starts on that machine. The user remotely accesses to the machine through a web browser and carries out the simulation, without the need to install any software but a web browser on the user's own computer. Therefore, Simulation Platform is expected to eliminate impediments to handle multiple neural models that require multiple software. Copyright © 2011 Elsevier Ltd. All rights reserved.

  8. High-throughput state-machine replication using software transactional memory.

    PubMed

    Zhao, Wenbing; Yang, William; Zhang, Honglei; Yang, Jack; Luo, Xiong; Zhu, Yueqin; Yang, Mary; Luo, Chaomin

    2016-11-01

    State-machine replication is a common way of constructing general purpose fault tolerance systems. To ensure replica consistency, requests must be executed sequentially according to some total order at all non-faulty replicas. Unfortunately, this could severely limit the system throughput. This issue has been partially addressed by identifying non-conflicting requests based on application semantics and executing these requests concurrently. However, identifying and tracking non-conflicting requests require intimate knowledge of application design and implementation, and a custom fault tolerance solution developed for one application cannot be easily adopted by other applications. Software transactional memory offers a new way of constructing concurrent programs. In this article, we present the mechanisms needed to retrofit existing concurrency control algorithms designed for software transactional memory for state-machine replication. The main benefit for using software transactional memory in state-machine replication is that general purpose concurrency control mechanisms can be designed without deep knowledge of application semantics. As such, new fault tolerance systems based on state-machine replications with excellent throughput can be easily designed and maintained. In this article, we introduce three different concurrency control mechanisms for state-machine replication using software transactional memory, namely, ordered strong strict two-phase locking, conventional timestamp-based multiversion concurrency control, and speculative timestamp-based multiversion concurrency control. Our experiments show that speculative timestamp-based multiversion concurrency control mechanism has the best performance in all types of workload, the conventional timestamp-based multiversion concurrency control offers the worst performance due to high abort rate in the presence of even moderate contention between transactions. The ordered strong strict two-phase locking mechanism offers the simplest solution with excellent performance in low contention workload, and fairly good performance in high contention workload.

  9. High-throughput state-machine replication using software transactional memory

    PubMed Central

    Yang, William; Zhang, Honglei; Yang, Jack; Luo, Xiong; Zhu, Yueqin; Yang, Mary; Luo, Chaomin

    2017-01-01

    State-machine replication is a common way of constructing general purpose fault tolerance systems. To ensure replica consistency, requests must be executed sequentially according to some total order at all non-faulty replicas. Unfortunately, this could severely limit the system throughput. This issue has been partially addressed by identifying non-conflicting requests based on application semantics and executing these requests concurrently. However, identifying and tracking non-conflicting requests require intimate knowledge of application design and implementation, and a custom fault tolerance solution developed for one application cannot be easily adopted by other applications. Software transactional memory offers a new way of constructing concurrent programs. In this article, we present the mechanisms needed to retrofit existing concurrency control algorithms designed for software transactional memory for state-machine replication. The main benefit for using software transactional memory in state-machine replication is that general purpose concurrency control mechanisms can be designed without deep knowledge of application semantics. As such, new fault tolerance systems based on state-machine replications with excellent throughput can be easily designed and maintained. In this article, we introduce three different concurrency control mechanisms for state-machine replication using software transactional memory, namely, ordered strong strict two-phase locking, conventional timestamp-based multiversion concurrency control, and speculative timestamp-based multiversion concurrency control. Our experiments show that speculative timestamp-based multiversion concurrency control mechanism has the best performance in all types of workload, the conventional timestamp-based multiversion concurrency control offers the worst performance due to high abort rate in the presence of even moderate contention between transactions. The ordered strong strict two-phase locking mechanism offers the simplest solution with excellent performance in low contention workload, and fairly good performance in high contention workload. PMID:29075049

  10. Modeling of spectral signatures of littoral waters

    NASA Astrophysics Data System (ADS)

    Haltrin, Vladimir I.

    1997-12-01

    The spectral values of remotely obtained radiance reflectance coefficient (RRC) are compared with the values of RRC computed from inherent optical properties measured during the shipborne experiment near the West Florida coast. The model calculations are based on the algorithm developed at the Naval Research Laboratory at Stennis Space Center and presented here. The algorithm is based on the radiation transfer theory and uses regression relationships derived from experimental data. Overall comparison of derived and measured RRCs shows that this algorithm is suitable for processing ground truth data for the purposes of remote data calibration. The second part of this work consists of the evaluation of the predictive visibility model (PVM). The simulated three-dimensional values of optical properties are compared with the measured ones. Preliminary results of comparison are encouraging and show that the PVM can qualitatively predict the evolution of inherent optical properties in littoral waters.

  11. Optimized and parallelized implementation of the electronegativity equalization method and the atom-bond electronegativity equalization method.

    PubMed

    Vareková, R Svobodová; Koca, J

    2006-02-01

    The most common way to calculate charge distribution in a molecule is ab initio quantum mechanics (QM). Some faster alternatives to QM have also been developed, the so-called "equalization methods" EEM and ABEEM, which are based on DFT. We have implemented and optimized the EEM and ABEEM methods and created the EEM SOLVER and ABEEM SOLVER programs. It has been found that the most time-consuming part of equalization methods is the reduction of the matrix belonging to the equation system generated by the method. Therefore, for both methods this part was replaced by the parallel algorithm WIRS and implemented within the PVM environment. The parallelized versions of the programs EEM SOLVER and ABEEM SOLVER showed promising results, especially on a single computer with several processors (compact PVM). The implemented programs are available through the Web page http://ncbr.chemi.muni.cz/~n19n/eem_abeem.

  12. DOE Office of Scientific and Technical Information (OSTI.GOV)

    Belley, M; Schmidt, M; Knutson, N

    Purpose: Physics second-checks for external beam radiation therapy are performed, in-part, to verify that the machine parameters in the Record-and-Verify (R&V) system that will ultimately be sent to the LINAC exactly match the values initially calculated by the Treatment Planning System (TPS). While performing the second-check, a large portion of the physicists’ time is spent navigating and arranging display windows to locate and compare the relevant numerical values (MLC position, collimator rotation, field size, MU, etc.). Here, we describe the development of a software tool that guides the physicist by aggregating and succinctly displaying machine parameter data relevant to themore » physics second-check process. Methods: A data retrieval software tool was developed using Python to aggregate data and generate a list of machine parameters that are commonly verified during the physics second-check process. This software tool imported values from (i) the TPS RT Plan DICOM file and (ii) the MOSAIQ (R&V) Structured Query Language (SQL) database. The machine parameters aggregated for this study included: MLC positions, X&Y jaw positions, collimator rotation, gantry rotation, MU, dose rate, wedges and accessories, cumulative dose, energy, machine name, couch angle, and more. Results: A GUI interface was developed to generate a side-by-side display of the aggregated machine parameter values for each field, and presented to the physicist for direct visual comparison. This software tool was tested for 3D conformal, static IMRT, sliding window IMRT, and VMAT treatment plans. Conclusion: This software tool facilitated the data collection process needed in order for the physicist to conduct a second-check, thus yielding an optimized second-check workflow that was both more user friendly and time-efficient. Utilizing this software tool, the physicist was able to spend less time searching through the TPS PDF plan document and the R&V system and focus the second-check efforts on assessing the patient-specific plan-quality.« less

  13. Data storage technology: Hardware and software, Appendix B

    NASA Technical Reports Server (NTRS)

    Sable, J. D.

    1972-01-01

    This project involves the development of more economical ways of integrating and interfacing new storage devices and data processing programs into a computer system. It involves developing interface standards and a software/hardware architecture which will make it possible to develop machine independent devices and programs. These will interface with the machine dependent operating systems of particular computers. The development project will not be to develop the software which would ordinarily be the responsibility of the manufacturer to supply, but to develop the standards with which that software is expected to confirm in providing an interface with the user or storage system.

  14. Army-NASA aircrew/aircraft integration program: Phase 4 A(3)I Man-Machine Integration Design and Analysis System (MIDAS) software detailed design document

    NASA Technical Reports Server (NTRS)

    Banda, Carolyn; Bushnell, David; Chen, Scott; Chiu, Alex; Constantine, Betsy; Murray, Jerry; Neukom, Christian; Prevost, Michael; Shankar, Renuka; Staveland, Lowell

    1991-01-01

    The Man-Machine Integration Design and Analysis System (MIDAS) is an integrated suite of software components that constitutes a prototype workstation to aid designers in applying human factors principles to the design of complex human-machine systems. MIDAS is intended to be used at the very early stages of conceptual design to provide an environment wherein designers can use computational representations of the crew station and operator, instead of hardware simulators and man-in-the-loop studies, to discover problems and ask 'what if' questions regarding the projected mission, equipment, and environment. This document is the Software Product Specification for MIDAS. Introductory descriptions of the processing requirements, hardware/software environment, structure, I/O, and control are given in the main body of the document for the overall MIDAS system, with detailed discussion of the individual modules included in Annexes A-J.

  15. Build and Execute Environment

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Guan, Qiang

    At exascale, the challenge becomes to develop applications that run at scale and use exascale platforms reliably, efficiently, and flexibly. Workflows become much more complex because they must seamlessly integrate simulation and data analytics. They must include down-sampling, post-processing, feature extraction, and visualization. Power and data transfer limitations require these analysis tasks to be run in-situ or in-transit. We expect successful workflows will comprise multiple linked simulations along with tens of analysis routines. Users will have limited development time at scale and, therefore, must have rich tools to develop, debug, test, and deploy applications. At this scale, successful workflows willmore » compose linked computations from an assortment of reliable, well-defined computation elements, ones that can come and go as required, based on the needs of the workflow over time. We propose a novel framework that utilizes both virtual machines (VMs) and software containers to create a workflow system that establishes a uniform build and execution environment (BEE) beyond the capabilities of current systems. In this environment, applications will run reliably and repeatably across heterogeneous hardware and software. Containers, both commercial (Docker and Rocket) and open-source (LXC and LXD), define a runtime that isolates all software dependencies from the machine operating system. Workflows may contain multiple containers that run different operating systems, different software, and even different versions of the same software. We will run containers in open-source virtual machines (KVM) and emulators (QEMU) so that workflows run on any machine entirely in user-space. On this platform of containers and virtual machines, we will deliver workflow software that provides services, including repeatable execution, provenance, checkpointing, and future proofing. We will capture provenance about how containers were launched and how they interact to annotate workflows for repeatable and partial re-execution. We will coordinate the physical snapshots of virtual machines with parallel programming constructs, such as barriers, to automate checkpoint and restart. We will also integrate with HPC-specific container runtimes to gain access to accelerators and other specialized hardware to preserve native performance. Containers will link development to continuous integration. When application developers check code in, it will automatically be tested on a suite of different software and hardware architectures.« less

  16. Modeling the 1958 Lituya Bay mega-tsunami with a PVM-IFCP GPU-based model

    NASA Astrophysics Data System (ADS)

    González-Vida, José M.; Arcas, Diego; de la Asunción, Marc; Castro, Manuel J.; Macías, Jorge; Ortega, Sergio; Sánchez-Linares, Carlos; Titov, Vasily

    2013-04-01

    In this work we present a numerical study, performed in collaboration with the NOAA Center for Tsunami Research (USA), that uses a GPU version of the PVM-IFCP landslide model for the simulation of the 1958 landslide generated tsunami of Lituya Bay. In this model, a layer composed of fluidized granular material is assumed to flow within an upper layer of an inviscid fluid (e. g. water). The model is discretized using a two dimensional PVM-IFCP [Fernández - Castro - Parés. On an Intermediate Field Capturing Riemann Solver Based on a Parabolic Viscosity Matrix for the Two-Layer Shallow Water System, J. Sci. Comput., 48 (2011):117-140] finite volume scheme implemented on GPU cards for increasing the speed-up. This model has been previously validated by using the two-dimensional physical laboratory experiments data from H. Fritz [Lituya Bay Landslide Impact Generated Mega-Tsunami 50th Anniversary. Pure Appl. Geophys., 166 (2009) pp. 153-175]. In the present work, the first step was to reconstruct the topobathymetry of the Lituya Bay before this event ocurred, this is based on USGS geological surveys data. Then, a sensitivity analysis of some model parameters has been performed in order to determine the parameters that better fit to reality, when model results are compared against available event data, as run-up areas. In this presentation, the reconstruction of the pre-tsunami scenario will be shown, a detailed simulation of the tsunami presented and several comparisons with real data (runup, wave height, etc.) shown.

  17. Long-term live imaging reveals cytosolic immune responses of host hepatocytes against Plasmodium infection and parasite escape mechanisms

    PubMed Central

    Prado, Monica; Eickel, Nina; De Niz, Mariana; Heitmann, Anna; Agop-Nersesian, Carolina; Wacker, Rahel; Schmuckli-Maurer, Jacqueline; Caldelari, Reto; Janse, Chris J; Khan, Shahid M; May, Jürgen; Meyer, Christian G; Heussler, Volker T

    2015-01-01

    Plasmodium parasites are transmitted by Anopheles mosquitoes to the mammalian host and actively infect hepatocytes after passive transport in the bloodstream to the liver. In their target host hepatocyte, parasites reside within a parasitophorous vacuole (PV). In the present study it was shown that the parasitophorous vacuole membrane (PVM) can be targeted by autophagy marker proteins LC3, ubiquitin, and SQSTM1/p62 as well as by lysosomes in a process resembling selective autophagy. The dynamics of autophagy marker proteins in individual Plasmodium berghei-infected hepatocytes were followed by live imaging throughout the entire development of the parasite in the liver. Although the host cell very efficiently recognized the invading parasite in its vacuole, the majority of parasites survived this initial attack. Successful parasite development correlated with the gradual loss of all analyzed autophagy marker proteins and associated lysosomes from the PVM. However, other autophagic events like nonselective canonical autophagy in the host cell continued. This was indicated as LC3, although not labeling the PVM anymore, still localized to autophagosomes in the infected host cell. It appears that growing parasites even benefit from this form of nonselective host cell autophagy as an additional source of nutrients, as in host cells deficient for autophagy, parasite growth was retarded and could partly be rescued by the supply of additional amino acid in the medium. Importantly, mouse infections with P. berghei sporozoites confirmed LC3 dynamics, the positive effect of autophagy activation on parasite growth, and negative effects upon autophagy inhibition. PMID:26208778

  18. The Automated Instrumentation and Monitoring System (AIMS) reference manual

    NASA Technical Reports Server (NTRS)

    Yan, Jerry; Hontalas, Philip; Listgarten, Sherry

    1993-01-01

    Whether a researcher is designing the 'next parallel programming paradigm,' another 'scalable multiprocessor' or investigating resource allocation algorithms for multiprocessors, a facility that enables parallel program execution to be captured and displayed is invaluable. Careful analysis of execution traces can help computer designers and software architects to uncover system behavior and to take advantage of specific application characteristics and hardware features. A software tool kit that facilitates performance evaluation of parallel applications on multiprocessors is described. The Automated Instrumentation and Monitoring System (AIMS) has four major software components: a source code instrumentor which automatically inserts active event recorders into the program's source code before compilation; a run time performance-monitoring library, which collects performance data; a trace file animation and analysis tool kit which reconstructs program execution from the trace file; and a trace post-processor which compensate for data collection overhead. Besides being used as prototype for developing new techniques for instrumenting, monitoring, and visualizing parallel program execution, AIMS is also being incorporated into the run-time environments of various hardware test beds to evaluate their impact on user productivity. Currently, AIMS instrumentors accept FORTRAN and C parallel programs written for Intel's NX operating system on the iPSC family of multi computers. A run-time performance-monitoring library for the iPSC/860 is included in this release. We plan to release monitors for other platforms (such as PVM and TMC's CM-5) in the near future. Performance data collected can be graphically displayed on workstations (e.g. Sun Sparc and SGI) supporting X-Windows (in particular, Xl IR5, Motif 1.1.3).

  19. Low Latency Messages on Distributed Memory Multiprocessors

    DOE PAGES

    Rosing, Matt; Saltz, Joel

    1995-01-01

    This article describes many of the issues in developing an efficient interface for communication on distributed memory machines. Although the hardware component of message latency is less than 1 ws on many distributed memory machines, the software latency associated with sending and receiving typed messages is on the order of 50 μs. The reason for this imbalance is that the software interface does not match the hardware. By changing the interface to match the hardware more closely, applications with fine grained communication can be put on these machines. This article describes several tests performed and many of the issues involvedmore » in supporting low latency messages on distributed memory machines.« less

  20. Channel Efficiency with Security Enhancement for Remote Condition Monitoring of Multi Machine System Using Hybrid Huffman Coding

    NASA Astrophysics Data System (ADS)

    Datta, Jinia; Chowdhuri, Sumana; Bera, Jitendranath

    2016-12-01

    This paper presents a novel scheme of remote condition monitoring of multi machine system where a secured and coded data of induction machine with different parameters is communicated between a state-of-the-art dedicated hardware Units (DHU) installed at the machine terminal and a centralized PC based machine data management (MDM) software. The DHUs are built for acquisition of different parameters from the respective machines, and hence are placed at their nearby panels in order to acquire different parameters cost effectively during their running condition. The MDM software collects these data through a communication channel where all the DHUs are networked using RS485 protocol. Before transmitting, the parameter's related data is modified with the adoption of differential pulse coded modulation (DPCM) and Huffman coding technique. It is further encrypted with a private key where different keys are used for different DHUs. In this way a data security scheme is adopted during its passage through the communication channel in order to avoid any third party attack into the channel. The hybrid mode of DPCM and Huffman coding is chosen to reduce the data packet length. A MATLAB based simulation and its practical implementation using DHUs at three machine terminals (one healthy three phase, one healthy single phase and one faulty three phase machine) proves its efficacy and usefulness for condition based maintenance of multi machine system. The data at the central control room are decrypted and decoded using MDM software. In this work it is observed that Chanel efficiency with respect to different parameter measurements has been increased very much.

  1. Parallel computing on Unix workstation arrays

    NASA Astrophysics Data System (ADS)

    Reale, F.; Bocchino, F.; Sciortino, S.

    1994-12-01

    We have tested arrays of general-purpose Unix workstations used as MIMD systems for massive parallel computations. In particular we have solved numerically a demanding test problem with a 2D hydrodynamic code, generally developed to study astrophysical flows, by exucuting it on arrays either of DECstations 5000/200 on Ethernet LAN, or of DECstations 3000/400, equipped with powerful Alpha processors, on FDDI LAN. The code is appropriate for data-domain decomposition, and we have used a library for parallelization previously developed in our Institute, and easily extended to work on Unix workstation arrays by using the PVM software toolset. We have compared the parallel efficiencies obtained on arrays of several processors to those obtained on a dedicated MIMD parallel system, namely a Meiko Computing Surface (CS-1), equipped with Intel i860 processors. We discuss the feasibility of using non-dedicated parallel systems and conclude that the convenience depends essentially on the size of the computational domain as compared to the relative processor power and network bandwidth. We point out that for future perspectives a parallel development of processor and network technology is important, and that the software still offers great opportunities of improvement, especially in terms of latency times in the message-passing protocols. In conditions of significant gain in terms of speedup, such workstation arrays represent a cost-effective approach to massive parallel computations.

  2. Flexible software architecture for user-interface and machine control in laboratory automation.

    PubMed

    Arutunian, E B; Meldrum, D R; Friedman, N A; Moody, S E

    1998-10-01

    We describe a modular, layered software architecture for automated laboratory instruments. The design consists of a sophisticated user interface, a machine controller and multiple individual hardware subsystems, each interacting through a client-server architecture built entirely on top of open Internet standards. In our implementation, the user-interface components are built as Java applets that are downloaded from a server integrated into the machine controller. The user-interface client can thereby provide laboratory personnel with a familiar environment for experiment design through a standard World Wide Web browser. Data management and security are seamlessly integrated at the machine-controller layer using QNX, a real-time operating system. This layer also controls hardware subsystems through a second client-server interface. This architecture has proven flexible and relatively easy to implement and allows users to operate laboratory automation instruments remotely through an Internet connection. The software architecture was implemented and demonstrated on the Acapella, an automated fluid-sample-processing system that is under development at the University of Washington.

  3. Advanced Cell Classifier: User-Friendly Machine-Learning-Based Software for Discovering Phenotypes in High-Content Imaging Data.

    PubMed

    Piccinini, Filippo; Balassa, Tamas; Szkalisity, Abel; Molnar, Csaba; Paavolainen, Lassi; Kujala, Kaisa; Buzas, Krisztina; Sarazova, Marie; Pietiainen, Vilja; Kutay, Ulrike; Smith, Kevin; Horvath, Peter

    2017-06-28

    High-content, imaging-based screens now routinely generate data on a scale that precludes manual verification and interrogation. Software applying machine learning has become an essential tool to automate analysis, but these methods require annotated examples to learn from. Efficiently exploring large datasets to find relevant examples remains a challenging bottleneck. Here, we present Advanced Cell Classifier (ACC), a graphical software package for phenotypic analysis that addresses these difficulties. ACC applies machine-learning and image-analysis methods to high-content data generated by large-scale, cell-based experiments. It features methods to mine microscopic image data, discover new phenotypes, and improve recognition performance. We demonstrate that these features substantially expedite the training process, successfully uncover rare phenotypes, and improve the accuracy of the analysis. ACC is extensively documented, designed to be user-friendly for researchers without machine-learning expertise, and distributed as a free open-source tool at www.cellclassifier.org. Copyright © 2017 Elsevier Inc. All rights reserved.

  4. Computer-Aided-Design of the Hydraulic System of Three-Dimensional Cartridge Valve Blocks (Selected Articles)

    DTIC Science & Technology

    1991-03-21

    sectional representation of the spatial figure can be correctly determined. 6 The AutoLisp language system in the AutoCAD software provides the most...softwares are developed on the 32-bit machines and little progress has been reported for the 16-bit machines. Even the AutoCAD is a two-ard-a-half... AutoCAD software as the basis, developed the design package of 3-D cartridge valve blocks on IM PC/AT. To realize the 3-D displaying of cartridge valves

  5. HeNCE: A Heterogeneous Network Computing Environment

    DOE PAGES

    Beguelin, Adam; Dongarra, Jack J.; Geist, George Al; ...

    1994-01-01

    Network computing seeks to utilize the aggregate resources of many networked computers to solve a single problem. In so doing it is often possible to obtain supercomputer performance from an inexpensive local area network. The drawback is that network computing is complicated and error prone when done by hand, especially if the computers have different operating systems and data formats and are thus heterogeneous. The heterogeneous network computing environment (HeNCE) is an integrated graphical environment for creating and running parallel programs over a heterogeneous collection of computers. It is built on a lower level package called parallel virtual machine (PVM).more » The HeNCE philosophy of parallel programming is to have the programmer graphically specify the parallelism of a computation and to automate, as much as possible, the tasks of writing, compiling, executing, debugging, and tracing the network computation. Key to HeNCE is a graphical language based on directed graphs that describe the parallelism and data dependencies of an application. Nodes in the graphs represent conventional Fortran or C subroutines and the arcs represent data and control flow. This article describes the present state of HeNCE, its capabilities, limitations, and areas of future research.« less

  6. Automated Concurrent Blackboard System Generation in C++

    NASA Technical Reports Server (NTRS)

    Kaplan, J. A.; McManus, J. W.; Bynum, W. L.

    1999-01-01

    In his 1992 Ph.D. thesis, "Design and Analysis Techniques for Concurrent Blackboard Systems", John McManus defined several performance metrics for concurrent blackboard systems and developed a suite of tools for creating and analyzing such systems. These tools allow a user to analyze a concurrent blackboard system design and predict the performance of the system before any code is written. The design can be modified until simulated performance is satisfactory. Then, the code generator can be invoked to generate automatically all of the code required for the concurrent blackboard system except for the code implementing the functionality of each knowledge source. We have completed the port of the source code generator and a simulator for a concurrent blackboard system. The source code generator generates the necessary C++ source code to implement the concurrent blackboard system using Parallel Virtual Machine (PVM) running on a heterogeneous network of UNIX(trademark) workstations. The concurrent blackboard simulator uses the blackboard specification file to predict the performance of the concurrent blackboard design. The only part of the source code for the concurrent blackboard system that the user must supply is the code implementing the functionality of the knowledge sources.

  7. Clinical data miner: an electronic case report form system with integrated data preprocessing and machine-learning libraries supporting clinical diagnostic model research.

    PubMed

    Installé, Arnaud Jf; Van den Bosch, Thierry; De Moor, Bart; Timmerman, Dirk

    2014-10-20

    Using machine-learning techniques, clinical diagnostic model research extracts diagnostic models from patient data. Traditionally, patient data are often collected using electronic Case Report Form (eCRF) systems, while mathematical software is used for analyzing these data using machine-learning techniques. Due to the lack of integration between eCRF systems and mathematical software, extracting diagnostic models is a complex, error-prone process. Moreover, due to the complexity of this process, it is usually only performed once, after a predetermined number of data points have been collected, without insight into the predictive performance of the resulting models. The objective of the study of Clinical Data Miner (CDM) software framework is to offer an eCRF system with integrated data preprocessing and machine-learning libraries, improving efficiency of the clinical diagnostic model research workflow, and to enable optimization of patient inclusion numbers through study performance monitoring. The CDM software framework was developed using a test-driven development (TDD) approach, to ensure high software quality. Architecturally, CDM's design is split over a number of modules, to ensure future extendability. The TDD approach has enabled us to deliver high software quality. CDM's eCRF Web interface is in active use by the studies of the International Endometrial Tumor Analysis consortium, with over 4000 enrolled patients, and more studies planned. Additionally, a derived user interface has been used in six separate interrater agreement studies. CDM's integrated data preprocessing and machine-learning libraries simplify some otherwise manual and error-prone steps in the clinical diagnostic model research workflow. Furthermore, CDM's libraries provide study coordinators with a method to monitor a study's predictive performance as patient inclusions increase. To our knowledge, CDM is the only eCRF system integrating data preprocessing and machine-learning libraries. This integration improves the efficiency of the clinical diagnostic model research workflow. Moreover, by simplifying the generation of learning curves, CDM enables study coordinators to assess more accurately when data collection can be terminated, resulting in better models or lower patient recruitment costs.

  8. Exploring machine-learning-based control plane intrusion detection techniques in software defined optical networks

    NASA Astrophysics Data System (ADS)

    Zhang, Huibin; Wang, Yuqiao; Chen, Haoran; Zhao, Yongli; Zhang, Jie

    2017-12-01

    In software defined optical networks (SDON), the centralized control plane may encounter numerous intrusion threatens which compromise the security level of provisioned services. In this paper, the issue of control plane security is studied and two machine-learning-based control plane intrusion detection techniques are proposed for SDON with properly selected features such as bandwidth, route length, etc. We validate the feasibility and efficiency of the proposed techniques by simulations. Results show an accuracy of 83% for intrusion detection can be achieved with the proposed machine-learning-based control plane intrusion detection techniques.

  9. Plasmacytoid Dendritic Cells Promote Host Defense Against Acute Pneumovirus Infection via the TLR7-MyD88-Dependent Signaling Pathway

    PubMed Central

    Davidson, Sophia; Kaiko, Gerard; Loh, Zhixuan; Lalwani, Amit; Zhang, Vivian; Spann, Kirsten; Foo, Shen Yun; Hansbro, Nicole; Uematsu, Satoshi; Akira, Shizuo; Matthaei, Klaus I.; Rosenberg, Helene F.; Foster, Paul S.; Phipps, Simon

    2012-01-01

    Human respiratory syncytial virus (RSV) is the leading cause of lower respiratory tract infection in infants. In human infants, plasmacytoid dendritic cells (pDC) are recruited to the nasal compartment during infection and initiate host defense through the secretion of type I IFN, IL-12 and IL-6. However, RSV-infected pDCs are refractory to TLR7-mediated activation. Here, we used the rodent-specific pathogen, pneumonia virus of mice (PVM), to determine the contribution of pDC and TLR7-signaling to the development of the innate inflammatory and early adaptive immune response. In wild-type (WT) but not TLR7- or myeloid differentiation protein 88 (MyD88)-deficient mice, PVM inoculation led to a marked infiltration of pDCs and increased expression of type I, II and III IFNs. The delayed induction of IFNs in the absence of TLR7 or MyD88 was associated with a diminished innate inflammatory response and augmented virus recovery from lung tissue. In the absence of TLR7, PVM-specific CD8+ T cell cytokine production was abrogated. The adoptive transfer of TLR7-sufficient but not TLR7-deficient pDC to TLR7-gene-deleted mice recapitulated the antiviral responses observed in WT mice and promoted virus clearance. In summary, TLR7-mediated signaling by pDC is required for appropriate innate responses to acute pneumovirus infection. It is conceivable that as-yet-unidentified defects in the TLR7 signaling pathway may be associated with elevated levels of RSV-associated morbidity and mortality among otherwise healthy human infants. PMID:21482736

  10. Culture media-based selection of endothelial cells, pericytes, and perivascular-resident macrophage-like melanocytes from the young mouse vestibular system.

    PubMed

    Zhang, Jinhui; Chen, Songlin; Cai, Jing; Hou, Zhiqiang; Wang, Xiaohan; Kachelmeier, Allan; Shi, Xiaorui

    2017-03-01

    The vestibular blood-labyrinth barrier (BLB) is comprised of perivascular-resident macrophage-like melanocytes (PVM/Ms) and pericytes (PCs), in addition to endothelial cells (ECs) and basement membrane (BM), and bears strong resemblance to the cochlear BLB in the stria vascularis. Over the past few decades, in vitro cell-based models have been widely used in blood-brain barrier (BBB) and blood-retina barrier (BRB) research, and have proved to be powerful tools for studying cell-cell interactions in their respective organs. Study of both the vestibular and strial BLB has been limited by the unavailability of primary culture cells from these barriers. To better understand how barrier component cells interact in the vestibular system to control BLB function, we developed a novel culture medium-based method for obtaining EC, PC, and PVM/M primary cells from tiny explants of the semicircular canal, sacculus, utriculus, and ampullae tissue of young mouse ears at post-natal age 8-12 d. Each phenotype is grown in a specific culture medium which selectively supports the phenotype in a mixed population of vestibular cell types. The unwanted phenotypes do not survive passaging. The protocol does not require additional equipment or special enzyme treatment. The harvesting process takes less than 2 h. Primary cell types are generated within 7-10 d. The primary culture ECs, PCs, and PVM/M shave consistent phenotypes more than 90% pure after two passages (∼ 3 weeks). The highly purified primary cell lines can be used for studying cell-cell interactions, barrier permeability, and angiogenesis. Copyright © 2017 Elsevier B.V. All rights reserved.

  11. General-Purpose Front End for Real-Time Data Processing

    NASA Technical Reports Server (NTRS)

    James, Mark

    2007-01-01

    FRONTIER is a computer program that functions as a front end for any of a variety of other software of both the artificial intelligence (AI) and conventional data-processing types. As used here, front end signifies interface software needed for acquiring and preprocessing data and making the data available for analysis by the other software. FRONTIER is reusable in that it can be rapidly tailored to any such other software with minimum effort. Each component of FRONTIER is programmable and is executed in an embedded virtual machine. Each component can be reconfigured during execution. The virtual-machine implementation making FRONTIER independent of the type of computing hardware on which it is executed.

  12. Multidisciplinary Concurrent Design Optimization via the Internet

    NASA Technical Reports Server (NTRS)

    Woodard, Stanley E.; Kelkar, Atul G.; Koganti, Gopichand

    2001-01-01

    A methodology is presented which uses commercial design and analysis software and the Internet to perform concurrent multidisciplinary optimization. The methodology provides a means to develop multidisciplinary designs without requiring that all software be accessible from the same local network. The procedures are amenable to design and development teams whose members, expertise and respective software are not geographically located together. This methodology facilitates multidisciplinary teams working concurrently on a design problem of common interest. Partition of design software to different machines allows each constituent software to be used on the machine that provides the most economy and efficiency. The methodology is demonstrated on the concurrent design of a spacecraft structure and attitude control system. Results are compared to those derived from performing the design with an autonomous FORTRAN program.

  13. RAGE deficiency predisposes mice to virus-induced paucigranulocytic asthma

    PubMed Central

    Arikkatt, Jaisy; Ullah, Md Ashik; Short, Kirsty Renfree; Zhang, Vivan; Gan, Wan Jun; Loh, Zhixuan; Werder, Rhiannon B; Simpson, Jennifer; Sly, Peter D; Mazzone, Stuart B; Spann, Kirsten M; Ferreira, Manuel AR; Upham, John W; Sukkar, Maria B; Phipps, Simon

    2017-01-01

    Asthma is a chronic inflammatory disease. Although many patients with asthma develop type-2 dominated eosinophilic inflammation, a number of individuals develop paucigranulocytic asthma, which occurs in the absence of eosinophilia or neutrophilia. The aetiology of paucigranulocytic asthma is unknown. However, both respiratory syncytial virus (RSV) infection and mutations in the receptor for advanced glycation endproducts (RAGE) are risk factors for asthma development. Here, we show that RAGE deficiency impairs anti-viral immunity during an early-life infection with pneumonia virus of mice (PVM; a murine analogue of RSV). The elevated viral load was associated with the release of high mobility group box-1 (HMGB1) which triggered airway smooth muscle remodelling in early-life. Re-infection with PVM in later-life induced many of the cardinal features of asthma in the absence of eosinophilic or neutrophilic inflammation. Anti-HMGB1 mitigated both early-life viral disease and asthma-like features, highlighting HMGB1 as a possible novel therapeutic target. DOI: http://dx.doi.org/10.7554/eLife.21199.001 PMID:28099113

  14. Sensory Neuron Fates Are Distinguished by a Transcriptional Switch that Regulates Dendrite Branch Stabilization

    PubMed Central

    Smith, Cody J.; O’Brien, Timothy; Chatzigeorgiou, Marios; Spencer, W. Clay; Feingold-Link, Elana; Husson, Steven J.; Hori, Sayaka; Mitani, Shohei; Gottschalk, Alexander; Schafer, William R.; Miller, David M.

    2013-01-01

    SUMMARY Sensory neurons adopt distinct morphologies and functional modalities to mediate responses to specific stimuli. Transcription factors and their downstream effectors orchestrate this outcome but are incompletely defined. Here, we show that different classes of mechanosensory neurons in C. elegans are distinguished by the combined action of the transcription factors MEC-3, AHR-1, and ZAG-1. Low levels of MEC-3 specify the elaborate branching pattern of PVD nociceptors, whereas high MEC-3 is correlated with the simple morphology of AVM and PVM touch neurons. AHR-1 specifies AVM touch neuron fate by elevating MEC-3 while simultaneously blocking expression of nociceptive genes such as the MEC-3 target, the claudin-like membrane protein HPO-30, that promotes the complex dendritic branching pattern of PVD. ZAG-1 exercises a parallel role to prevent PVM from adopting the PVD fate. The conserved dendritic branching function of the Drosophila AHR-1 homolog, Spineless, argues for similar pathways in mammals. PMID:23889932

  15. ICCE Policy Statement on Network and Multiple Machine Software.

    ERIC Educational Resources Information Center

    International Council for Computers in Education, Eugene, OR.

    Designed to provide educators with guidance for the lawful reproduction of computer software, this document contains suggested guidelines, sample forms, and several short articles concerning software copyright and license agreements. The initial policy statement calls for educators to provide software developers (or their agents) with a…

  16. ICCE Policy Statement on Network and Multiple Machine Software.

    ERIC Educational Resources Information Center

    Computing Teacher, 1983

    1983-01-01

    Issued to provide guidance for the resolution of problems inherent in providing and securing good educational software, this statement outlines responsibilities of educators, hardware vendors, and software developers/vendors. Sample policy statements for school districts and community colleges, suggested format for software licenses, and technical…

  17. A Low-Cost Audio Prescription Labeling System Using RFID for Thai Visually-Impaired People.

    PubMed

    Lertwiriyaprapa, Titipong; Fakkheow, Pirapong

    2015-01-01

    This research aims to develop a low-cost audio prescription labeling (APL) system for visually-impaired people by using the RFID system. The developed APL system includes the APL machine and APL software. The APL machine is for visually-impaired people while APL software allows caregivers to record all important information into the APL machine. The main objective of the development of the APL machine is to reduce costs and size by designing all of the electronic devices to fit into one print circuit board. Also, it is designed so that it is easy to use and can become an electronic aid for daily living. The developed APL software is based on Java and MySQL, both of which can operate on various operating platforms and are easy to develop as commercial software. The developed APL system was first evaluated by 5 experts. The APL system was also evaluated by 50 actual visually-impaired people (30 elders and 20 blind individuals) and 20 caregivers, pharmacists and nurses. After using the APL system, evaluations were carried out, and it can be concluded from the evaluation results that this proposed APL system can be effectively used for helping visually-impaired people in terms of self-medication.

  18. A real-time MPEG software decoder using a portable message-passing library

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Kwong, Man Kam; Tang, P.T. Peter; Lin, Biquan

    1995-12-31

    We present a real-time MPEG software decoder that uses message-passing libraries such as MPL, p4 and MPI. The parallel MPEG decoder currently runs on the IBM SP system but can be easil ported to other parallel machines. This paper discusses our parallel MPEG decoding algorithm as well as the parallel programming environment under which it uses. Several technical issues are discussed, including balancing of decoding speed, memory limitation, 1/0 capacities, and optimization of MPEG decoding components. This project shows that a real-time portable software MPEG decoder is feasible in a general-purpose parallel machine.

  19. Software design and documentation language, revision 1

    NASA Technical Reports Server (NTRS)

    Kleine, H.

    1979-01-01

    The Software Design and Documentation Language (SDDL) developed to provide an effective communications medium to support the design and documentation of complex software applications is described. Features of the system include: (1) a processor which can convert design specifications into an intelligible, informative machine-reproducible document; (2) a design and documentation language with forms and syntax that are simple, unrestrictive, and communicative; and (3) methodology for effective use of the language and processor. The SDDL processor is written in the SIMSCRIPT II programming language and is implemented on the UNIVAC 1108, the IBM 360/370, and Control Data machines.

  20. The 1988 Directory of Educational Software Publishing Companies.

    ERIC Educational Resources Information Center

    Electronic Learning, 1988

    1988-01-01

    Based on questionnaires sent to educational software companies in January 1988, this directory lists 78 companies. Information given includes company address, curriculum subject areas for which the company publishes software, types of machines and operating systems on which the software operates, and grade level for which it is targeted. (LRW)

  1. Using Microcomputers in Vocational Education to Teach Needed Skills in Machine Shop and Related Occupations. Final Report.

    ERIC Educational Resources Information Center

    Mercer County Schools, Princeton, WV.

    A project was undertaken to identify machine shop occupations requiring workers to use computers, identify the computer skills needed to perform machine shop tasks, and determine which software products are currently being used in machine shop programs. A search of the Dictionary of Occupational Titles revealed that computer skills will become…

  2. Survey of Mass Storage Systems

    DTIC Science & Technology

    1975-09-01

    software that Pre- cision Instruments can provide. System Name: IBM 3850 Mass Storage System Manufacturer and Location: International Business Machines...34 Datamation, pp. 52-58, October 1973. 15 17. International Business Machines, IBM 3850 Mass Storage System Facts Folder, White Plains, NY, n.d. 18... International Business Machines, Introduction to the IBM 3850 Mass Storage System (MSS), White Plains, NY, n.d. 19. International Business Machines

  3. The ASSERT Virtual Machine Kernel: Support for Preservation of Temporal Properties

    NASA Astrophysics Data System (ADS)

    Zamorano, J.; de la Puente, J. A.; Pulido, J. A.; Urueña

    2008-08-01

    A new approach to building embedded real-time software has been developed in the ASSERT project. One of its key elements is the concept of a virtual machine preserving the non-functional properties of the system, and especially real-time properties, all the way down from high- level design models down to executable code. The paper describes one instance of the virtual machine concept that provides support for the preservation of temporal properties both at the source code level —by accept- ing only "legal" entities, i.e. software components with statically analysable real-tim behaviour— and at run-time —by monitoring the temporal behaviour of the system. The virtual machine has been validated on several pilot projects carried out by aerospace companies in the framework of the ASSERT project.

  4. An application of machine learning to the organization of institutional software repositories

    NASA Technical Reports Server (NTRS)

    Bailin, Sidney; Henderson, Scott; Truszkowski, Walt

    1993-01-01

    Software reuse has become a major goal in the development of space systems, as a recent NASA-wide workshop on the subject made clear. The Data Systems Technology Division of Goddard Space Flight Center has been working on tools and techniques for promoting reuse, in particular in the development of satellite ground support software. One of these tools is the Experiment in Libraries via Incremental Schemata and Cobweb (ElvisC). ElvisC applies machine learning to the problem of organizing a reusable software component library for efficient and reliable retrieval. In this paper we describe the background factors that have motivated this work, present the design of the system, and evaluate the results of its application.

  5. The Historical Evolution of Educational Software.

    ERIC Educational Resources Information Center

    Troutner, Joanne

    This paper establishes the roots of computers and automated teaching in the field of psychology and describes Dr. S. L. Pressey's presentation of the teaching machine; B. F. Skinner's teaching machine; Meyer's steps in composing a program for the automated teaching machine; IBM's beginning research on automated courses and the development of the…

  6. Lifelong personal health data and application software via virtual machines in the cloud.

    PubMed

    Van Gorp, Pieter; Comuzzi, Marco

    2014-01-01

    Personal Health Records (PHRs) should remain the lifelong property of patients, who should be able to show them conveniently and securely to selected caregivers and institutions. In this paper, we present MyPHRMachines, a cloud-based PHR system taking a radically new architectural solution to health record portability. In MyPHRMachines, health-related data and the application software to view and/or analyze it are separately deployed in the PHR system. After uploading their medical data to MyPHRMachines, patients can access them again from remote virtual machines that contain the right software to visualize and analyze them without any need for conversion. Patients can share their remote virtual machine session with selected caregivers, who will need only a Web browser to access the pre-loaded fragments of their lifelong PHR. We discuss a prototype of MyPHRMachines applied to two use cases, i.e., radiology image sharing and personalized medicine.

  7. A Workstation Farm Optimized for Monte Carlo Shell Model Calculations : Alphleet

    NASA Astrophysics Data System (ADS)

    Watanabe, Y.; Shimizu, N.; Haruyama, S.; Honma, M.; Mizusaki, T.; Taketani, A.; Utsuno, Y.; Otsuka, T.

    We have built a workstation farm named ``Alphleet" which consists of 140 COMPAQ's Alpha 21264 CPUs, for Monte Carlo Shell Model (MCSM) calculations. It has achieved more than 90 % scalable performance with 140 CPUs when the MCSM calculation with PVM and 61.2 Gflops of LINPACK.

  8. Cognitive Foundry v. 3.0 (OSS)

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Basilico, Justin; Dixon, Kevin; McClain, Jonathan

    2009-11-18

    The Cognitive Foundry is a unified collection of tools designed for research and applications that use cognitive modeling, machine learning, or pattern recognition. The software library contains design patterns, interface definitions, and default implementations of reusable software components and algorithms designed to support a wide variety of research and development needs. The library contains three main software packages: the Common package that contains basic utilities and linear algebraic methods, the Cognitive Framework package that contains tools to assist in implementing and analyzing theories of cognition, and the Machine Learning package that provides general algorithms and methods for populating Cognitive Frameworkmore » components from domain-relevant data.« less

  9. Software tool for data mining and its applications

    NASA Astrophysics Data System (ADS)

    Yang, Jie; Ye, Chenzhou; Chen, Nianyi

    2002-03-01

    A software tool for data mining is introduced, which integrates pattern recognition (PCA, Fisher, clustering, hyperenvelop, regression), artificial intelligence (knowledge representation, decision trees), statistical learning (rough set, support vector machine), computational intelligence (neural network, genetic algorithm, fuzzy systems). It consists of nine function models: pattern recognition, decision trees, association rule, fuzzy rule, neural network, genetic algorithm, Hyper Envelop, support vector machine, visualization. The principle and knowledge representation of some function models of data mining are described. The software tool of data mining is realized by Visual C++ under Windows 2000. Nonmonotony in data mining is dealt with by concept hierarchy and layered mining. The software tool of data mining has satisfactorily applied in the prediction of regularities of the formation of ternary intermetallic compounds in alloy systems, and diagnosis of brain glioma.

  10. Visual Computing Environment

    NASA Technical Reports Server (NTRS)

    Lawrence, Charles; Putt, Charles W.

    1997-01-01

    The Visual Computing Environment (VCE) is a NASA Lewis Research Center project to develop a framework for intercomponent and multidisciplinary computational simulations. Many current engineering analysis codes simulate various aspects of aircraft engine operation. For example, existing computational fluid dynamics (CFD) codes can model the airflow through individual engine components such as the inlet, compressor, combustor, turbine, or nozzle. Currently, these codes are run in isolation, making intercomponent and complete system simulations very difficult to perform. In addition, management and utilization of these engineering codes for coupled component simulations is a complex, laborious task, requiring substantial experience and effort. To facilitate multicomponent aircraft engine analysis, the CFD Research Corporation (CFDRC) is developing the VCE system. This system, which is part of NASA's Numerical Propulsion Simulation System (NPSS) program, can couple various engineering disciplines, such as CFD, structural analysis, and thermal analysis. The objectives of VCE are to (1) develop a visual computing environment for controlling the execution of individual simulation codes that are running in parallel and are distributed on heterogeneous host machines in a networked environment, (2) develop numerical coupling algorithms for interchanging boundary conditions between codes with arbitrary grid matching and different levels of dimensionality, (3) provide a graphical interface for simulation setup and control, and (4) provide tools for online visualization and plotting. VCE was designed to provide a distributed, object-oriented environment. Mechanisms are provided for creating and manipulating objects, such as grids, boundary conditions, and solution data. This environment includes parallel virtual machine (PVM) for distributed processing. Users can interactively select and couple any set of codes that have been modified to run in a parallel distributed fashion on a cluster of heterogeneous workstations. A scripting facility allows users to dictate the sequence of events that make up the particular simulation.

  11. POLYSHIFT Communications Software for the Connection Machine System CM-200

    DOE PAGES

    George, William; Brickner, Ralph G.; Johnsson, S. Lennart

    1994-01-01

    We describe the use and implementation of a polyshift function PSHIFT for circular shifts and end-offs shifts. Polyshift is useful in many scientific codes using regular grids, such as finite difference codes in several dimensions, and multigrid codes, molecular dynamics computations, and in lattice gauge physics computations, such as quantum chromodynamics (QCD) calculations. Our implementation of the PSHIFT function on the Connection Machine systems CM-2 and CM-200 offers a speedup of up to a factor of 3–4 compared with CSHIFT when the local data motion within a node is small. The PSHIFT routine is included in the Connection Machine Scientificmore » Software Library (CMSSL).« less

  12. Vane Pump Casing Machining of Dumpling Machine Based on CAD/CAM

    NASA Astrophysics Data System (ADS)

    Huang, Yusen; Li, Shilong; Li, Chengcheng; Yang, Zhen

    Automatic dumpling forming machine is also called dumpling machine, which makes dumplings through mechanical motions. This paper adopts the stuffing delivery mechanism featuring the improved and specially-designed vane pump casing, which can contribute to the formation of dumplings. Its 3D modeling in Pro/E software, machining process planning, milling path optimization, simulation based on UG and compiling post program were introduced and verified. The results indicated that adoption of CAD/CAM offers firms the potential to pursue new innovative strategies.

  13. Research on software behavior trust based on hierarchy evaluation

    NASA Astrophysics Data System (ADS)

    Long, Ke; Xu, Haishui

    2017-08-01

    In view of the correlation software behavior, we evaluate software behavior credibility from two levels of control flow and data flow. In control flow level, method of the software behavior of trace based on support vector machine (SVM) is proposed. In data flow level, behavioral evidence evaluation based on fuzzy decision analysis method is put forward.

  14. TEACHING PHYSICS: Atwood's machine: experiments in an accelerating frame

    NASA Astrophysics Data System (ADS)

    Teck Chee, Chia; Hong, Chia Yee

    1999-03-01

    Experiments in an accelerating frame are often difficult to perform, but simple computer software allows sufficiently rapid and accurate measurements to be made on an arrangement of weights and pulleys known as Atwood's machine.

  15. Atwood's Machine: Experiments in an Accelerating Frame.

    ERIC Educational Resources Information Center

    Chee, Chia Teck; Hong, Chia Yee

    1999-01-01

    Experiments in an accelerating frame are hard to perform. Illustrates how simple computer software allows sufficiently rapid and accurate measurements to be made on an arrangement of weights and pulleys known as Atwood's machine. (Author/CCM)

  16. Software component quality evaluation

    NASA Technical Reports Server (NTRS)

    Clough, A. J.

    1991-01-01

    The paper describes a software inspection process that can be used to evaluate the quality of software components. Quality criteria, process application, independent testing of the process and proposed associated tool support are covered. Early results indicate that this technique is well suited for assessing software component quality in a standardized fashion. With automated machine assistance to facilitate both the evaluation and selection of software components, such a technique should promote effective reuse of software components.

  17. Orchid: a novel management, annotation and machine learning framework for analyzing cancer mutations.

    PubMed

    Cario, Clinton L; Witte, John S

    2018-03-15

    As whole-genome tumor sequence and biological annotation datasets grow in size, number and content, there is an increasing basic science and clinical need for efficient and accurate data management and analysis software. With the emergence of increasingly sophisticated data stores, execution environments and machine learning algorithms, there is also a need for the integration of functionality across frameworks. We present orchid, a python based software package for the management, annotation and machine learning of cancer mutations. Building on technologies of parallel workflow execution, in-memory database storage and machine learning analytics, orchid efficiently handles millions of mutations and hundreds of features in an easy-to-use manner. We describe the implementation of orchid and demonstrate its ability to distinguish tissue of origin in 12 tumor types based on 339 features using a random forest classifier. Orchid and our annotated tumor mutation database are freely available at https://github.com/wittelab/orchid. Software is implemented in python 2.7, and makes use of MySQL or MemSQL databases. Groovy 2.4.5 is optionally required for parallel workflow execution. JWitte@ucsf.edu. Supplementary data are available at Bioinformatics online.

  18. State-of-the-Art in Improved Parts Programming for Numerically Controlled Machines

    DTIC Science & Technology

    1976-10-01

    than expected let sizes for IIC. Cincinnati lilbcron, Inc., has built a $1.25 million Computer Ilumerical Control ( CNC ) 1,4nufacturing Center to "rw’ t...point-to- point user. Lathe and other turning operations are essentially two-axis opera- tions, and there has been some dissatisfaction over APT’s...a.particular machi-ne (50)." "Software is the key to CNC , the costs of which are easily overlooked. The cost of software development is growing in relation to

  19. Chronic Neuropsychological Sequelae of Cholinesterase Inhibitors in the Absence of Structural Brain Damage: Two Cases of Acute Poisoning

    PubMed Central

    Roldán-Tapia, Lola; Leyva, Antonia; Laynez, Francisco; Santed, Fernando Sánchez

    2005-01-01

    Here we describe two cases of carbamate poisoning. Patients AMF and PVM were accidentally poisoned by cholinesterase inhibitors. The medical diagnosis in both cases was overcholinergic syndrome, as demonstrated by exposure to cholinesterase inhibitors. The widespread use of cholinesterase inhibitors, especially as pesticides, produces a great number of human poisoning events annually. The main known neurotoxic effect of these substances is cholinesterase inhibition, which causes cholinergic overstimulation. Once AMF and PVM had recovered from acute intoxication, they were subjected to extensive neuropsychological evaluation 3 and 12 months after the poisoning event. These assessments point to a cognitive deficit in attention, memory, perceptual, and motor domains 3 months after intoxication. One year later these sequelae remained, even though the brain magnetic resonance imaging (MRI) and computed tomography (CT) scans were interpreted as being within normal limits. We present these cases as examples of neuropsychological profiles of long-term sequelae related to acute poisoning by cholinesterase inhibitor pesticides and show the usefulness of neuropsychological assessment in detecting central nervous system dysfunction in the absence of biochemical or structural markers. PMID:15929901

  20. The Machine Scoring of Writing

    ERIC Educational Resources Information Center

    McCurry, Doug

    2010-01-01

    This article provides an introduction to the kind of computer software that is used to score student writing in some high stakes testing programs, and that is being promoted as a teaching and learning tool to schools. It sketches the state of play with machines for the scoring of writing, and describes how these machines work and what they do.…

  1. 75 FR 71146 - In the Matter of Certain Machine Vision Software, Machine Vision Systems, and Products Containing...

    Federal Register 2010, 2011, 2012, 2013, 2014

    2010-11-22

    ...''); Amistar Automation, Inc. (``Amistar'') of San Marcos, California; Techno Soft Systemnics, Inc. (``Techno..., the ALJ's construction of the claim terms ``test,'' ``match score surface,'' and ``gradient direction...

  2. CNC Machining Of The Complex Copper Electrodes

    NASA Astrophysics Data System (ADS)

    Popan, Ioan Alexandru; Balc, Nicolae; Popan, Alina

    2015-07-01

    This paper presents the machining process of the complex copper electrodes. Machining of the complex shapes in copper is difficult because this material is soft and sticky. This research presents the main steps for processing those copper electrodes at a high dimensional accuracy and a good surface quality. Special tooling solutions are required for this machining process and optimal process parameters have been found for the accurate CNC equipment, using smart CAD/CAM software.

  3. Project SYNERGY: Software Support for Underprepared Students. Software Implementation Report.

    ERIC Educational Resources Information Center

    Anandam, Kamala; And Others

    Miami-Dade Community College's (MDCC's) implementation and assessment of computer software as a part of Project SYNERGY, a multi-institutional project funded by the International Business Machines (IBM) Corporation designed to seek technological solutions for helping students underprepared in reading, writing and mathematics, is described in this…

  4. Leveraging Code Comments to Improve Software Reliability

    ERIC Educational Resources Information Center

    Tan, Lin

    2009-01-01

    Commenting source code has long been a common practice in software development. This thesis, consisting of three pieces of work, made novel use of the code comments written in natural language to improve software reliability. Our solution combines Natural Language Processing (NLP), Machine Learning, Statistics, and Program Analysis techniques to…

  5. Objective detection of apoptosis in rat renal tissue sections using light microscopy and free image analysis software with subsequent machine learning: Detection of apoptosis in renal tissue.

    PubMed

    Macedo, Nayana Damiani; Buzin, Aline Rodrigues; de Araujo, Isabela Bastos Binotti Abreu; Nogueira, Breno Valentim; de Andrade, Tadeu Uggere; Endringer, Denise Coutinho; Lenz, Dominik

    2017-02-01

    The current study proposes an automated machine learning approach for the quantification of cells in cell death pathways according to DNA fragmentation. A total of 17 images of kidney histological slide samples from male Wistar rats were used. The slides were photographed using an Axio Zeiss Vert.A1 microscope with a 40x objective lens coupled with an Axio Cam MRC Zeiss camera and Zen 2012 software. The images were analyzed using CellProfiler (version 2.1.1) and CellProfiler Analyst open-source software. Out of the 10,378 objects, 4970 (47,9%) were identified as TUNEL positive, and 5408 (52,1%) were identified as TUNEL negative. On average, the sensitivity and specificity values of the machine learning approach were 0.80 and 0.77, respectively. Image cytometry provides a quantitative analytical alternative to the more traditional qualitative methods more commonly used in studies. Copyright © 2016 Elsevier Ltd. All rights reserved.

  6. Software Reviews.

    ERIC Educational Resources Information Center

    Science and Children, 1990

    1990-01-01

    Reviewed are seven computer software packages for IBM and/or Apple Computers. Included are "Windows on Science: Volume 1--Physical Science"; "Science Probe--Physical Science"; "Wildlife Adventures--Grizzly Bears"; "Science Skills--Development Programs"; "The Clean Machine"; "Rock Doctor";…

  7. A Spectral Element Ocean Model on the Cray T3D: the interannual variability of the Mediterranean Sea general circulation

    NASA Astrophysics Data System (ADS)

    Molcard, A. J.; Pinardi, N.; Ansaloni, R.

    A new numerical model, SEOM (Spectral Element Ocean Model, (Iskandarani et al, 1994)), has been implemented in the Mediterranean Sea. Spectral element methods combine the geometric flexibility of finite element techniques with the rapid convergence rate of spectral schemes. The current version solves the shallow water equations with a fifth (or sixth) order accuracy spectral scheme and about 50.000 nodes. The domain decomposition philosophy makes it possible to exploit the power of parallel machines. The original MIMD master/slave version of SEOM, written in F90 and PVM, has been ported to the Cray T3D. When critical for performance, Cray specific high-performance one-sided communication routines (SHMEM) have been adopted to fully exploit the Cray T3D interprocessor network. Tests performed with highly unstructured and irregular grid, on up to 128 processors, show an almost linear scalability even with unoptimized domain decomposition techniques. Results from various case studies on the Mediterranean Sea are shown, involving realistic coastline geometry, and monthly mean 1000mb winds from the ECMWF's atmospheric model operational analysis from the period January 1987 to December 1994. The simulation results show that variability in the wind forcing considerably affect the circulation dynamics of the Mediterranean Sea.

  8. The image-guided surgery toolkit IGSTK: an open source C++ software toolkit.

    PubMed

    Enquobahrie, Andinet; Cheng, Patrick; Gary, Kevin; Ibanez, Luis; Gobbi, David; Lindseth, Frank; Yaniv, Ziv; Aylward, Stephen; Jomier, Julien; Cleary, Kevin

    2007-11-01

    This paper presents an overview of the image-guided surgery toolkit (IGSTK). IGSTK is an open source C++ software library that provides the basic components needed to develop image-guided surgery applications. It is intended for fast prototyping and development of image-guided surgery applications. The toolkit was developed through a collaboration between academic and industry partners. Because IGSTK was designed for safety-critical applications, the development team has adopted lightweight software processes that emphasizes safety and robustness while, at the same time, supporting geographically separated developers. A software process that is philosophically similar to agile software methods was adopted emphasizing iterative, incremental, and test-driven development principles. The guiding principle in the architecture design of IGSTK is patient safety. The IGSTK team implemented a component-based architecture and used state machine software design methodologies to improve the reliability and safety of the components. Every IGSTK component has a well-defined set of features that are governed by state machines. The state machine ensures that the component is always in a valid state and that all state transitions are valid and meaningful. Realizing that the continued success and viability of an open source toolkit depends on a strong user community, the IGSTK team is following several key strategies to build an active user community. These include maintaining a users and developers' mailing list, providing documentation (application programming interface reference document and book), presenting demonstration applications, and delivering tutorial sessions at relevant scientific conferences.

  9. Configurable software for satellite graphics

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Hartzman, P D

    An important goal in interactive computer graphics is to provide users with both quick system responses for basic graphics functions and enough computing power for complex calculations. One solution is to have a distributed graphics system in which a minicomputer and a powerful large computer share the work. The most versatile type of distributed system is an intelligent satellite system in which the minicomputer is programmable by the application user and can do most of the work while the large remote machine is used for difficult computations. At New York University, the hardware was configured from available equipment. The levelmore » of system intelligence resulted almost completely from software development. Unlike previous work with intelligent satellites, the resulting system had system control centered in the satellite. It also had the ability to reconfigure software during realtime operation. The design of the system was done at a very high level using set theoretic language. The specification clearly illustrated processor boundaries and interfaces. The high-level specification also produced a compact, machine-independent virtual graphics data structure for picture representation. The software was written in a systems implementation language; thus, only one set of programs was needed for both machines. A user can program both machines in a single language. Tests of the system with an application program indicate that is has very high potential. A major result of this work is the demonstration that a gigantic investment in new hardware is not necessary for computing facilities interested in graphics.« less

  10. An Evaluation of Online Machine Translation of Arabic into English News Headlines: Implications on Students' Learning Purposes

    ERIC Educational Resources Information Center

    Kadhim, Kais A.; Habeeb, Luwaytha S.; Sapar, Ahmad Arifin; Hussin, Zaharah; Abdullah, Muhammad Ridhuan Tony Lim

    2013-01-01

    Nowadays, online Machine Translation (MT) is used widely with translation software, such as Google and Babylon, being easily available and downloadable. This study aims to test the translation quality of these two machine systems in translating Arabic news headlines into English. 40 Arabic news headlines were selected from three online sources,…

  11. Using Multiple FPGA Architectures for Real-time Processing of Low-level Machine Vision Functions

    Treesearch

    Thomas H. Drayer; William E. King; Philip A. Araman; Joseph G. Tront; Richard W. Conners

    1995-01-01

    In this paper, we investigate the use of multiple Field Programmable Gate Array (FPGA) architectures for real-time machine vision processing. The use of FPGAs for low-level processing represents an excellent tradeoff between software and special purpose hardware implementations. A library of modules that implement common low-level machine vision operations is presented...

  12. Rosen's (M,R) system as an X-machine.

    PubMed

    Palmer, Michael L; Williams, Richard A; Gatherer, Derek

    2016-11-07

    Robert Rosen's (M,R) system is an abstract biological network architecture that is allegedly both irreducible to sub-models of its component states and non-computable on a Turing machine. (M,R) stands as an obstacle to both reductionist and mechanistic presentations of systems biology, principally due to its self-referential structure. If (M,R) has the properties claimed for it, computational systems biology will not be possible, or at best will be a science of approximate simulations rather than accurate models. Several attempts have been made, at both empirical and theoretical levels, to disprove this assertion by instantiating (M,R) in software architectures. So far, these efforts have been inconclusive. In this paper, we attempt to demonstrate why - by showing how both finite state machine and stream X-machine formal architectures fail to capture the self-referential requirements of (M,R). We then show that a solution may be found in communicating X-machines, which remove self-reference using parallel computation, and then synthesise such machine architectures with object-orientation to create a formal basis for future software instantiations of (M,R) systems. Copyright © 2016 Elsevier Ltd. All rights reserved.

  13. Filament winding technique, experiment and simulation analysis on tubular structure

    NASA Astrophysics Data System (ADS)

    Quanjin, Ma; Rejab, M. R. M.; Kaige, Jiang; Idris, M. S.; Harith, M. N.

    2018-04-01

    Filament winding process has emerged as one of the potential composite fabrication processes with lower costs. Filament wound products involve classic axisymmetric parts (pipes, rings, driveshafts, high-pressure vessels and storage tanks), non-axisymmetric parts (prismatic nonround sections and pipe fittings). Based on the 3-axis filament winding machine has been designed with the inexpensive control system, it is completely necessary to make a relative comparison between experiment and simulation on tubular structure. In this technical paper, the aim of this paper is to perform a dry winding experiment using the 3-axis filament winding machine and simulate winding process on the tubular structure using CADWIND software with 30°, 45°, 60° winding angle. The main result indicates that the 3-axis filament winding machine can produce tubular structure with high winding pattern performance with different winding angle. This developed 3-axis winding machine still has weakness compared to CAWIND software simulation results with high axes winding machine about winding pattern, turnaround impact, process error, thickness, friction impact etc. In conclusion, the 3-axis filament winding machine improvements and recommendations come up with its comparison results, which can intuitively understand its limitations and characteristics.

  14. Image-Based Single Cell Profiling: High-Throughput Processing of Mother Machine Experiments

    PubMed Central

    Sachs, Christian Carsten; Grünberger, Alexander; Helfrich, Stefan; Probst, Christopher; Wiechert, Wolfgang; Kohlheyer, Dietrich; Nöh, Katharina

    2016-01-01

    Background Microfluidic lab-on-chip technology combined with live-cell imaging has enabled the observation of single cells in their spatio-temporal context. The mother machine (MM) cultivation system is particularly attractive for the long-term investigation of rod-shaped bacteria since it facilitates continuous cultivation and observation of individual cells over many generations in a highly parallelized manner. To date, the lack of fully automated image analysis software limits the practical applicability of the MM as a phenotypic screening tool. Results We present an image analysis pipeline for the automated processing of MM time lapse image stacks. The pipeline supports all analysis steps, i.e., image registration, orientation correction, channel/cell detection, cell tracking, and result visualization. Tailored algorithms account for the specialized MM layout to enable a robust automated analysis. Image data generated in a two-day growth study (≈ 90 GB) is analyzed in ≈ 30 min with negligible differences in growth rate between automated and manual evaluation quality. The proposed methods are implemented in the software molyso (MOther machine AnaLYsis SOftware) that provides a new profiling tool to analyze unbiasedly hitherto inaccessible large-scale MM image stacks. Conclusion Presented is the software molyso, a ready-to-use open source software (BSD-licensed) for the unsupervised analysis of MM time-lapse image stacks. molyso source code and user manual are available at https://github.com/modsim/molyso. PMID:27661996

  15. Support Vector Machine algorithm for regression and classification

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Yu, Chenggang; Zavaljevski, Nela

    2001-08-01

    The software is an implementation of the Support Vector Machine (SVM) algorithm that was invented and developed by Vladimir Vapnik and his co-workers at AT&T Bell Laboratories. The specific implementation reported here is an Active Set method for solving a quadratic optimization problem that forms the major part of any SVM program. The implementation is tuned to specific constraints generated in the SVM learning. Thus, it is more efficient than general-purpose quadratic optimization programs. A decomposition method has been implemented in the software that enables processing large data sets. The size of the learning data is virtually unlimited by themore » capacity of the computer physical memory. The software is flexible and extensible. Two upper bounds are implemented to regulate the SVM learning for classification, which allow users to adjust the false positive and false negative rates. The software can be used either as a standalone, general-purpose SVM regression or classification program, or be embedded into a larger software system.« less

  16. Craniux: A LabVIEW-Based Modular Software Framework for Brain-Machine Interface Research

    PubMed Central

    Degenhart, Alan D.; Kelly, John W.; Ashmore, Robin C.; Collinger, Jennifer L.; Tyler-Kabara, Elizabeth C.; Weber, Douglas J.; Wang, Wei

    2011-01-01

    This paper presents “Craniux,” an open-access, open-source software framework for brain-machine interface (BMI) research. Developed in LabVIEW, a high-level graphical programming environment, Craniux offers both out-of-the-box functionality and a modular BMI software framework that is easily extendable. Specifically, it allows researchers to take advantage of multiple features inherent to the LabVIEW environment for on-the-fly data visualization, parallel processing, multithreading, and data saving. This paper introduces the basic features and system architecture of Craniux and describes the validation of the system under real-time BMI operation using simulated and real electrocorticographic (ECoG) signals. Our results indicate that Craniux is able to operate consistently in real time, enabling a seamless work flow to achieve brain control of cursor movement. The Craniux software framework is made available to the scientific research community to provide a LabVIEW-based BMI software platform for future BMI research and development. PMID:21687575

  17. Craniux: a LabVIEW-based modular software framework for brain-machine interface research.

    PubMed

    Degenhart, Alan D; Kelly, John W; Ashmore, Robin C; Collinger, Jennifer L; Tyler-Kabara, Elizabeth C; Weber, Douglas J; Wang, Wei

    2011-01-01

    This paper presents "Craniux," an open-access, open-source software framework for brain-machine interface (BMI) research. Developed in LabVIEW, a high-level graphical programming environment, Craniux offers both out-of-the-box functionality and a modular BMI software framework that is easily extendable. Specifically, it allows researchers to take advantage of multiple features inherent to the LabVIEW environment for on-the-fly data visualization, parallel processing, multithreading, and data saving. This paper introduces the basic features and system architecture of Craniux and describes the validation of the system under real-time BMI operation using simulated and real electrocorticographic (ECoG) signals. Our results indicate that Craniux is able to operate consistently in real time, enabling a seamless work flow to achieve brain control of cursor movement. The Craniux software framework is made available to the scientific research community to provide a LabVIEW-based BMI software platform for future BMI research and development.

  18. How do I resolve problems reading the binary data?

    Atmospheric Science Data Center

    2014-12-08

    ... affecting compilation would be differing versions of the operating system and compilers the read software are being run on. Big ... Unix machines are Big Endian architecture while Linux systems are Little Endian architecture. Data generated on a Unix machine are ...

  19. Position Paper: Applying Machine Learning to Software Analysis to Achieve Trusted, Repeatable Scientific Computing

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Prowell, Stacy J; Symons, Christopher T

    2015-01-01

    Producing trusted results from high-performance codes is essential for policy and has significant economic impact. We propose combining rigorous analytical methods with machine learning techniques to achieve the goal of repeatable, trustworthy scientific computing.

  20. Software Tools for Emittance Measurement and Matching for 12 GeV CEBAF

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Turner, Dennis L.

    2016-05-01

    This paper discusses model-driven setup of the Continuous Electron Beam Accelerator Facility (CEBAF) for the 12GeV era, focusing on qsUtility. qsUtility is a set of software tools created to perform emittance measurements, analyze those measurements, and compute optics corrections based upon the measurements.qsUtility was developed as a toolset to facilitate reducing machine configuration time and reproducibility by way of an accurate accelerator model, and to provide Operations staff with tools to measure and correct machine optics with little or no assistance from optics experts.

  1. Imperfection and Thickness Measurement of Panels Using a Coordinate Measurement Machine

    NASA Technical Reports Server (NTRS)

    Thornburgh, Robert P.

    2006-01-01

    This paper summarizes the methodology used to measure imperfection and thickness variation for flat and curved panels using a Coordinate Measurement Machine (CMM) and the software program MeasPanel. The objective is to provide a reference document so that someone with a basic understanding of CMM operation can measure a panel with minimal training. Detailed information about both the measurement system setup and computer software is provided. Information is also provided about the format of the raw data, as well as how it is post-processed for use in finite-element analysis.

  2. Research study demonstrates computer simulation can predict warpage and assist in its elimination

    NASA Astrophysics Data System (ADS)

    Glozer, G.; Post, S.; Ishii, K.

    1994-10-01

    Programs for predicting warpage in injection molded parts are relatively new. Commercial software for simulating the flow and cooling stages of injection molding have steadily gained acceptance; however, warpage software is not yet as readily accepted. This study focused on gaining an understanding of the predictive capabilities of the warpage software. The following aspects of this study were unique. (1) Quantitative results were found using a statistically designed set of experiments. (2) Comparisons between experimental and simulation results were made with parts produced in a well-instrumented and controlled injection molding machine. (3) The experimental parts were accurately measured on a coordinate measuring machine with a non-contact laser probe. (4) The effect of part geometry on warpage was investigated.

  3. Hardware/software codesign for embedded RISC core

    NASA Astrophysics Data System (ADS)

    Liu, Peng

    2001-12-01

    This paper describes hardware/software codesign method of the extendible embedded RISC core VIRGO, which based on MIPS-I instruction set architecture. VIRGO is described by Verilog hardware description language that has five-stage pipeline with shared 32-bit cache/memory interface, and it is controlled by distributed control scheme. Every pipeline stage has one small controller, which controls the pipeline stage status and cooperation among the pipeline phase. Since description use high level language and structure is distributed, VIRGO core has highly extension that can meet the requirements of application. We take look at the high-definition television MPEG2 MPHL decoder chip, constructed the hardware/software codesign virtual prototyping machine that can research on VIRGO core instruction set architecture, and system on chip memory size requirements, and system on chip software, etc. We also can evaluate the system on chip design and RISC instruction set based on the virtual prototyping machine platform.

  4. Embedded control system for computerized franking machine

    NASA Astrophysics Data System (ADS)

    Shi, W. M.; Zhang, L. B.; Xu, F.; Zhan, H. W.

    2007-12-01

    This paper presents a novel control system for franking machine. A methodology for operating a franking machine using the functional controls consisting of connection, configuration and franking electromechanical drive is studied. A set of enabling technologies to synthesize postage management software architectures driven microprocessor-based embedded systems is proposed. The cryptographic algorithm that calculates mail items is analyzed to enhance the postal indicia accountability and security. The study indicated that the franking machine is reliability, performance and flexibility in printing mail items.

  5. Operating System For Numerically Controlled Milling Machine

    NASA Technical Reports Server (NTRS)

    Ray, R. B.

    1992-01-01

    OPMILL program is operating system for Kearney and Trecker milling machine providing fast easy way to program manufacture of machine parts with IBM-compatible personal computer. Gives machinist "equation plotter" feature, which plots equations that define movements and converts equations to milling-machine-controlling program moving cutter along defined path. System includes tool-manager software handling up to 25 tools and automatically adjusts to account for each tool. Developed on IBM PS/2 computer running DOS 3.3 with 1 MB of random-access memory.

  6. Modeling and simulation of five-axis virtual machine based on NX

    NASA Astrophysics Data System (ADS)

    Li, Xiaoda; Zhan, Xianghui

    2018-04-01

    Virtual technology in the machinery manufacturing industry has shown the role of growing. In this paper, the Siemens NX software is used to model the virtual CNC machine tool, and the parameters of the virtual machine are defined according to the actual parameters of the machine tool so that the virtual simulation can be carried out without loss of the accuracy of the simulation. How to use the machine builder of the CAM module to define the kinematic chain and machine components of the machine is described. The simulation of virtual machine can provide alarm information of tool collision and over cutting during the process to users, and can evaluate and forecast the rationality of the technological process.

  7. Nozzles for Focusing Aerosol Particles

    DTIC Science & Technology

    2009-10-01

    Fabrication of the nozzle with the desired shape was accomplished using EDM technology. First, a copper tungsten electrode was turned on a CNC lathe . The...small (0.9-mm diameter). The external portions of the nozzles were machined in a more conventional manner using computer numerical control ( CNC ... lathes and milling machines running programs written by computer aided machining (CAM) software. The close tolerance of concentricity of the two

  8. Characteristics study of the gears by the CAD/CAE

    NASA Astrophysics Data System (ADS)

    Wang, P. Y.; Chang, S. L.; Lee, B. Y.; Nguyen, D. H.; Cao, C. W.

    2017-09-01

    Gears are the most important transmission component in machines. The rapid development of the machines in industry requires a shorter time of the analysis process. In traditional, the gears are analyzed by setting up the complete mathematical model firstly, considering the profile of cutter and coordinate systems relationship between the machine and the cutter. It is a really complex and time-consuming process. Recently, the CAD/CAE software is well developed and useful in the mechanical design. In this paper, the Autodesk Inventor® software is introduced to model the spherical gears firstly, and then the models can also be transferred into ANSYS Workbench for the finite element analysis. The proposed process in this paper is helpful to the engineers to speed up the analyzing process of gears in the design stage.

  9. Re-designing a mechanism for higher speed: A case history from textile machinery

    NASA Astrophysics Data System (ADS)

    Douglas, S. S.; Rooney, G. T.

    The generation of general mechanism design software which is the formulation of suitable objective functions is discussed. There is a consistent drive towards higher speeds in the development of industrial sewing machines. This led to experimental analyses of dynamic performance and to a search for improved design methods. The experimental work highlighted the need for smoothness of motion at high speed, component inertias, and frame structural stiffness. Smoothness is associated with transmission properties and harmonic analysis. These are added to other design requirements of synchronization, mechanism size, and function. Some of the mechanism trains in overedte sewing machines are shown. All these trains are designed by digital optimization. The design software combines analysis of the sewing machine mechanisms, formulation of objectives innumerical terms, and suitable mathematical optimization ttechniques.

  10. Applying Computerized-Scoring Models of Written Biological Explanations across Courses and Colleges: Prospects and Limitations

    PubMed Central

    Ha, Minsu; Nehm, Ross H.; Urban-Lurain, Mark; Merrill, John E.

    2011-01-01

    Our study explored the prospects and limitations of using machine-learning software to score introductory biology students’ written explanations of evolutionary change. We investigated three research questions: 1) Do scoring models built using student responses at one university function effectively at another university? 2) How many human-scored student responses are needed to build scoring models suitable for cross-institutional application? 3) What factors limit computer-scoring efficacy, and how can these factors be mitigated? To answer these questions, two biology experts scored a corpus of 2556 short-answer explanations (from biology majors and nonmajors) at two universities for the presence or absence of five key concepts of evolution. Human- and computer-generated scores were compared using kappa agreement statistics. We found that machine-learning software was capable in most cases of accurately evaluating the degree of scientific sophistication in undergraduate majors’ and nonmajors’ written explanations of evolutionary change. In cases in which the software did not perform at the benchmark of “near-perfect” agreement (kappa > 0.80), we located the causes of poor performance and identified a series of strategies for their mitigation. Machine-learning software holds promise as an assessment tool for use in undergraduate biology education, but like most assessment tools, it is also characterized by limitations. PMID:22135372

  11. Applying computerized-scoring models of written biological explanations across courses and colleges: prospects and limitations.

    PubMed

    Ha, Minsu; Nehm, Ross H; Urban-Lurain, Mark; Merrill, John E

    2011-01-01

    Our study explored the prospects and limitations of using machine-learning software to score introductory biology students' written explanations of evolutionary change. We investigated three research questions: 1) Do scoring models built using student responses at one university function effectively at another university? 2) How many human-scored student responses are needed to build scoring models suitable for cross-institutional application? 3) What factors limit computer-scoring efficacy, and how can these factors be mitigated? To answer these questions, two biology experts scored a corpus of 2556 short-answer explanations (from biology majors and nonmajors) at two universities for the presence or absence of five key concepts of evolution. Human- and computer-generated scores were compared using kappa agreement statistics. We found that machine-learning software was capable in most cases of accurately evaluating the degree of scientific sophistication in undergraduate majors' and nonmajors' written explanations of evolutionary change. In cases in which the software did not perform at the benchmark of "near-perfect" agreement (kappa > 0.80), we located the causes of poor performance and identified a series of strategies for their mitigation. Machine-learning software holds promise as an assessment tool for use in undergraduate biology education, but like most assessment tools, it is also characterized by limitations.

  12. National Collegiate Software Clearinghouse Software for the Humanities and Social Sciences. Summer 1989 Catalog.

    ERIC Educational Resources Information Center

    National Collegiate Software Clearinghouse, Durham, NC.

    Over 250 microcomputer software packages, intended for use on MS-DOS machines by scholars and teachers in the humanities and social sciences, are included in this catalog. The clearinghouse's first Macintosh listing is included, with many more Macintosh programs and data sets being planned and tested for future inclusion. Most programs were…

  13. [The testing system for OCP of the digital X-ray machine].

    PubMed

    Wang, Yan; Mo, Guoming; Wang, Juru; Zhou, Tao; Yu, Jianguo

    2011-09-01

    In this paper, we designed a testing system for operator control panel of a high-voltage and high-frequency X-ray machine, and an online testing software for functional components, in order to help the testing engineers to improve their work efficiency.

  14. One of My Favorite Assignments: Automated Teller Machine Simulation.

    ERIC Educational Resources Information Center

    Oberman, Paul S.

    2001-01-01

    Describes an assignment for an introductory computer science class that requires the student to write a software program that simulates an automated teller machine. Highlights include an algorithm for the assignment; sample file contents; language features used; assignment variations; and discussion points. (LRW)

  15. Influence of cutting data on surface quality when machining 17-4 PH stainless steel

    NASA Astrophysics Data System (ADS)

    Popovici, T. D.; Dijmărescu, M. R.

    2017-08-01

    The aim of the research presented in this paper is to analyse the cutting data influence upon surface quality for 17-4 PH stainless steel milling machining. The cutting regime parameters considered for the experiments were established using cutting regimes from experimental researches or from industrial conditions as basis, within the recommended ranges. The experimental program structure was determined by taking into account compatibility and orthogonality conditions, minimal use of material and labour. The machined surface roughness was determined by measuring the Ra roughness parameter, followed by surface profile registration in the form of graphics which were saved on a computer with MarSurf PS1Explorer software. Based on Ra roughness parameter, maximum values were extracted from these graphics and the influence charts of the cutting regime parameters upon surface roughness were traced using Microsoft Excel software. After a thorough analysis of the resulting data, relevant conclusions were drawn, presenting the interdependence between the surface roughness of the machined 17-4 PH samples and the cutting data variation.

  16. A hardware/software environment to support R D in intelligent machines and mobile robotic systems

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Mann, R.C.

    1990-01-01

    The Center for Engineering Systems Advanced Research (CESAR) serves as a focal point at the Oak Ridge National Laboratory (ORNL) for basic and applied research in intelligent machines. R D at CESAR addresses issues related to autonomous systems, unstructured (i.e. incompletely known) operational environments, and multiple performing agents. Two mobile robot prototypes (HERMIES-IIB and HERMIES-III) are being used to test new developments in several robot component technologies. This paper briefly introduces the computing environment at CESAR which includes three hypercube concurrent computers (two on-board the mobile robots), a graphics workstation, VAX, and multiple VME-based systems (several on-board the mobile robots).more » The current software environment at CESAR is intended to satisfy several goals, e.g.: code portability, re-usability in different experimental scenarios, modularity, concurrent computer hardware transparent to applications programmer, future support for multiple mobile robots, support human-machine interface modules, and support for integration of software from other, geographically disparate laboratories with different hardware set-ups. 6 refs., 1 fig.« less

  17. Understanding dental CAD/CAM for restorations--dental milling machines from a mechanical engineering viewpoint. Part B: labside milling machines.

    PubMed

    Lebon, Nicolas; Tapie, Laurent; Duret, Francois; Attal, Jean-Pierre

    2016-01-01

    Nowadays, dental numerical controlled (NC) milling machines are available for dental laboratories (labside solution) and dental production centers. This article provides a mechanical engineering approach to NC milling machines to help dental technicians understand the involvement of technology in digital dentistry practice. The technical and economic criteria are described for four labside and two production center dental NC milling machines available on the market. The technical criteria are focused on the capacities of the embedded technologies of milling machines to mill prosthetic materials and various restoration shapes. The economic criteria are focused on investment cost and interoperability with third-party software. The clinical relevance of the technology is discussed through the accuracy and integrity of the restoration. It can be asserted that dental production center milling machines offer a wider range of materials and types of restoration shapes than labside solutions, while labside solutions offer a wider range than chairside solutions. The accuracy and integrity of restorations may be improved as a function of the embedded technologies provided. However, the more complex the technical solutions available, the more skilled the user must be. Investment cost and interoperability with third-party software increase according to the quality of the embedded technologies implemented. Each private dental practice may decide which fabrication option to use depending on the scope of the practice.

  18. Software Patents.

    ERIC Educational Resources Information Center

    Burke, Edmund B.

    1994-01-01

    Outlines basic patent law information that pertains to computer software programs. Topics addressed include protection in other countries; how to obtain patents; kinds of patents; duration; classes of patentable subject matter, including machines and processes; patentability searches; experimental use prior to obtaining a patent; and patent…

  19. Development of system decision support tools for behavioral trends monitoring of machinery maintenance in a competitive environment

    NASA Astrophysics Data System (ADS)

    Adeyeri, Michael Kanisuru; Mpofu, Khumbulani

    2017-06-01

    The article is centred on software system development for manufacturing company that produces polyethylene bags using mostly conventional machines in a competitive world where each business enterprise desires to stand tall. This is meant to assist in gaining market shares, taking maintenance and production decisions by the dynamism and flexibilities embedded in the package as customers' demand varies under the duress of meeting the set goals. The production and machine condition monitoring software (PMCMS) is programmed in C# and designed in such a way to support hardware integration, real-time machine conditions monitoring, which is based on condition maintenance approach, maintenance decision suggestions and suitable production strategies as the demand for products keeps changing in a highly competitive environment. PMCMS works with an embedded device which feeds it with data from the various machines being monitored at the workstation, and the data are read at the base station through transmission via a wireless transceiver and stored in a database. A case study was used in the implementation of the developed system, and the results show that it can monitor the machine's health condition effectively by displaying machines' health status, gives repair suggestions to probable faults, decides strategy for both production methods and maintenance, and, thus, can enhance maintenance performance obviously.

  20. Development and Application of a Parallel LCAO Cluster Method

    NASA Astrophysics Data System (ADS)

    Patton, David C.

    1997-08-01

    CPU intensive steps in the SCF electronic structure calculations of clusters and molecules with a first-principles LCAO method have been fully parallelized via a message passing paradigm. Identification of the parts of the code that are composed of many independent compute-intensive steps is discussed in detail as they are the most readily parallelized. Most of the parallelization involves spatially decomposing numerical operations on a mesh. One exception is the solution of Poisson's equation which relies on distribution of the charge density and multipole methods. The method we use to parallelize this part of the calculation is quite novel and is covered in detail. We present a general method for dynamically load-balancing a parallel calculation and discuss how we use this method in our code. The results of benchmark calculations of the IR and Raman spectra of PAH molecules such as anthracene (C_14H_10) and tetracene (C_18H_12) are presented. These benchmark calculations were performed on an IBM SP2 and a SUN Ultra HPC server with both MPI and PVM. Scalability and speedup for these calculations is analyzed to determine the efficiency of the code. In addition, performance and usage issues for MPI and PVM are presented.

  1. In Silico Identification Software (ISIS): A Machine Learning Approach to Tandem Mass Spectral Identification of Lipids

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Kangas, Lars J.; Metz, Thomas O.; Isaac, Georgis

    2012-05-15

    Liquid chromatography-mass spectrometry-based metabolomics has gained importance in the life sciences, yet it is not supported by software tools for high throughput identification of metabolites based on their fragmentation spectra. An algorithm (ISIS: in silico identification software) and its implementation are presented and show great promise in generating in silico spectra of lipids for the purpose of structural identification. Instead of using chemical reaction rate equations or rules-based fragmentation libraries, the algorithm uses machine learning to find accurate bond cleavage rates in a mass spectrometer employing collision-induced dissocia-tion tandem mass spectrometry. A preliminary test of the algorithm with 45 lipidsmore » from a subset of lipid classes shows both high sensitivity and specificity.« less

  2. Method and system for rendering and interacting with an adaptable computing environment

    DOEpatents

    Osbourn, Gordon Cecil [Albuquerque, NM; Bouchard, Ann Marie [Albuquerque, NM

    2012-06-12

    An adaptable computing environment is implemented with software entities termed "s-machines", which self-assemble into hierarchical data structures capable of rendering and interacting with the computing environment. A hierarchical data structure includes a first hierarchical s-machine bound to a second hierarchical s-machine. The first hierarchical s-machine is associated with a first layer of a rendering region on a display screen and the second hierarchical s-machine is associated with a second layer of the rendering region overlaying at least a portion of the first layer. A screen element s-machine is linked to the first hierarchical s-machine. The screen element s-machine manages data associated with a screen element rendered to the display screen within the rendering region at the first layer.

  3. Improvement of Computer Software Quality through Software Automated Tools.

    DTIC Science & Technology

    1986-08-30

    information that are returned from the tools to the human user, and the forms in which these outputs are presented. Page 2 of 4 STAGE OF DEVELOPMENT: What... AUTOMIATED SOFTWARE TOOL MONITORING SYSTEM APPENDIX 2 2-1 INTRODUCTION This document and Automated Software Tool Monitoring Program (Appendix 1) are...t Output Output features provide links from the tool to both the human user and the target machine (where applicable). They describe the types

  4. Claims-Based Authentication for a Web-Based Enterprise

    DTIC Science & Technology

    2013-07-01

    authority must use known and registered (or in specific cases defined ) certificate revocation and currency-checking software . B. Translation of...Machines and services are issued software certificates that contain the public key with the private key generated and remaining in hardware...publicly available) information. A hardware token that contains the certificate is preferred to software -only certificates. For enterprise users

  5. Verification and Validation of the Malicious Activity Simulation Tool (MAST) for Network Administrator Training and Evaluation

    DTIC Science & Technology

    2012-03-01

    to sell fake antivirus software ; Gammima, which was used to steal gaming login information; and Zeus, which was used to steal banking information...13 3. Viruses ......................................14 C. PROOF OF CONCEPT OF SOFTWARE TRAINING USING MALWARE MIMICS...33 2. Software .....................................34 3. COMPOSE CG-71 Virtual Machines ...............37 a. Integrated Shipboard Network System

  6. Control system software, simulation, and robotic applications

    NASA Technical Reports Server (NTRS)

    Frisch, Harold P.

    1991-01-01

    All essential existing capabilities needed to create a man-machine interaction dynamics and performance (MMIDAP) capability are reviewed. The multibody system dynamics software program Order N DISCOS will be used for machine and musculo-skeletal dynamics modeling. The program JACK will be used for estimating and animating whole body human response to given loading situations and motion constraints. The basic elements of performance (BEP) task decomposition methodologies associated with the Human Performance Institute database will be used for performance assessment. Techniques for resolving the statically indeterminant muscular load sharing problem will be used for a detailed understanding of potential musculotendon or ligamentous fatigue, pain, discomfort, and trauma. The envisioned capacity is to be used for mechanical system design, human performance assessment, extrapolation of man/machine interaction test data, biomedical engineering, and soft prototyping within a concurrent engineering (CE) system.

  7. Modeling and Composing Scenario-Based Requirements with Aspects

    NASA Technical Reports Server (NTRS)

    Araujo, Joao; Whittle, Jon; Ki, Dae-Kyoo

    2004-01-01

    There has been significant recent interest, within the Aspect-Oriented Software Development (AOSD) community, in representing crosscutting concerns at various stages of the software lifecycle. However, most of these efforts have concentrated on the design and implementation phases. We focus in this paper on representing aspects during use case modeling. In particular, we focus on scenario-based requirements and show how to compose aspectual and non-aspectual scenarios so that they can be simulated as a whole. Non-aspectual scenarios are modeled as UML sequence diagram. Aspectual scenarios are modeled as Interaction Pattern Specifications (IPS). In order to simulate them, the scenarios are transformed into a set of executable state machines using an existing state machine synthesis algorithm. Previous work composed aspectual and non-aspectual scenarios at the sequence diagram level. In this paper, the composition is done at the state machine level.

  8. Conquering Language Babel in the Classroom

    ERIC Educational Resources Information Center

    Minichino, Mario; Berson, Michael J.

    2012-01-01

    This article is an exploration of the available applications for speech to speech real-time translation software for use in the classroom. Three different types of machine language translation (MLT) software and devices are reviewed for their features and practical application in secondary education classrooms.

  9. Research on ARM Numerical Control System

    NASA Astrophysics Data System (ADS)

    Wei, Xu; JiHong, Chen

    Computerized Numerical Control (CNC) machine tools is the foundation of modern manufacturing systems, whose advanced digital technology is the key to solve the problem of sustainable development of machine tool manufacturing industry. The paper is to design CNC system embedded on ARM and indicates the hardware design and the software systems supported. On the hardware side: the driving chip of the motor control unit, as the core of components, is MCX314AL of DSP motion control which is developed by NOVA Electronics Co., Ltd. of Japan. It make convenient to control machine because of its excellent performance, simple interface, easy programming. On the Software side, the uC/OS-2 is selected as the embedded operating system of the open source, which makes a detailed breakdown of the modules of the CNC system. Those priorities are designed according to their actual requirements. The ways of communication between the module and the interrupt response are so different that it guarantees real-time property and reliability of the numerical control system. Therefore, it not only meets the requirements of the current social precision machining, but has good man-machine interface and network support to facilitate a variety of craftsmen use.

  10. RHE: A JVM Courseware

    ERIC Educational Resources Information Center

    Liu, S.; Tang, J.; Deng, C.; Li, X.-F.; Gaudiot, J.-L.

    2011-01-01

    Java Virtual Machine (JVM) education has become essential in training embedded software engineers as well as virtual machine researchers and practitioners. However, due to the lack of suitable instructional tools, it is difficult for students to obtain any kind of hands-on experience and to attain any deep understanding of JVM design. To address…

  11. 76 FR 32231 - International Business Machines (IBM), Sales and Distribution Business Unit, Global Sales...

    Federal Register 2010, 2011, 2012, 2013, 2014

    2011-06-03

    ... for the workers and former workers of International Business Machines (IBM), Sales and Distribution... reconsideration alleges that IBM outsourced to India and China. During the reconsideration investigation, it was..., Armonk, New York. The subject worker group supply computer software development and maintenance services...

  12. 75 FR 60478 - In the Matter of Certain Machine Vision Software, Machine Vision Systems, and Products Containing...

    Federal Register 2010, 2011, 2012, 2013, 2014

    2010-09-30

    ... Automation, Inc. (``Amistar'') of San Marcos, California; Techno Soft Systemnics, Inc. (``Techno Soft'') of... the claim terms ``test,'' ``match score surface,'' and ``gradient direction,'' all of his infringement... complainants' proposed construction for the claim terms ``test,'' ``match score surface,'' and ``gradient...

  13. Scientific bases of human-machine communication by voice.

    PubMed Central

    Schafer, R W

    1995-01-01

    The scientific bases for human-machine communication by voice are in the fields of psychology, linguistics, acoustics, signal processing, computer science, and integrated circuit technology. The purpose of this paper is to highlight the basic scientific and technological issues in human-machine communication by voice and to point out areas of future research opportunity. The discussion is organized around the following major issues in implementing human-machine voice communication systems: (i) hardware/software implementation of the system, (ii) speech synthesis for voice output, (iii) speech recognition and understanding for voice input, and (iv) usability factors related to how humans interact with machines. PMID:7479802

  14. Craniux: A LabVIEW-Based Modular Software Framework for Brain-Machine Interface Research

    DTIC Science & Technology

    2011-01-01

    open-source BMI software solu- tions are currently available, we feel that the Craniux software package fills a specific need in the realm of BMI...data, such as cortical source imaging using EEG or MEG recordings. It is with these characteristics in mind that we feel the Craniux software package...S. Adee, “Dean Kamen’s ‘luke arm’ prosthesis readies for clinical trials,” IEEE Spectrum, February 2008, http://spectrum .ieee.org/biomedical

  15. Concurrent extensions to the FORTRAN language for parallel programming of computational fluid dynamics algorithms

    NASA Technical Reports Server (NTRS)

    Weeks, Cindy Lou

    1986-01-01

    Experiments were conducted at NASA Ames Research Center to define multi-tasking software requirements for multiple-instruction, multiple-data stream (MIMD) computer architectures. The focus was on specifying solutions for algorithms in the field of computational fluid dynamics (CFD). The program objectives were to allow researchers to produce usable parallel application software as soon as possible after acquiring MIMD computer equipment, to provide researchers with an easy-to-learn and easy-to-use parallel software language which could be implemented on several different MIMD machines, and to enable researchers to list preferred design specifications for future MIMD computer architectures. Analysis of CFD algorithms indicated that extensions of an existing programming language, adaptable to new computer architectures, provided the best solution to meeting program objectives. The CoFORTRAN Language was written in response to these objectives and to provide researchers a means to experiment with parallel software solutions to CFD algorithms on machines with parallel architectures.

  16. Design and fabrication of complete dentures using CAD/CAM technology

    PubMed Central

    Han, Weili; Li, Yanfeng; Zhang, Yue; lv, Yuan; Zhang, Ying; Hu, Ping; Liu, Huanyue; Ma, Zheng; Shen, Yi

    2017-01-01

    Abstract The aim of the study was to test the feasibility of using commercially available computer-aided design and computer-aided manufacturing (CAD/CAM) technology including 3Shape Dental System 2013 trial version, WIELAND V2.0.049 and WIELAND ZENOTEC T1 milling machine to design and fabricate complete dentures. The modeling process of full denture available in the trial version of 3Shape Dental System 2013 was used to design virtual complete dentures on the basis of 3-dimensional (3D) digital edentulous models generated from the physical models. The virtual complete dentures designed were exported to CAM software of WIELAND V2.0.049. A WIELAND ZENOTEC T1 milling machine controlled by the CAM software was used to fabricate physical dentitions and baseplates by milling acrylic resin composite plates. The physical dentitions were bonded to the corresponding baseplates to form the maxillary and mandibular complete dentures. Virtual complete dentures were successfully designed using the software through several steps including generation of 3D digital edentulous models, model analysis, arrangement of artificial teeth, trimming relief area, and occlusal adjustment. Physical dentitions and baseplates were successfully fabricated according to the designed virtual complete dentures using milling machine controlled by a CAM software. Bonding physical dentitions to the corresponding baseplates generated the final physical complete dentures. Our study demonstrated that complete dentures could be successfully designed and fabricated by using CAD/CAM. PMID:28072686

  17. Ambiguity and variability of database and software names in bioinformatics.

    PubMed

    Duck, Geraint; Kovacevic, Aleksandar; Robertson, David L; Stevens, Robert; Nenadic, Goran

    2015-01-01

    There are numerous options available to achieve various tasks in bioinformatics, but until recently, there were no tools that could systematically identify mentions of databases and tools within the literature. In this paper we explore the variability and ambiguity of database and software name mentions and compare dictionary and machine learning approaches to their identification. Through the development and analysis of a corpus of 60 full-text documents manually annotated at the mention level, we report high variability and ambiguity in database and software mentions. On a test set of 25 full-text documents, a baseline dictionary look-up achieved an F-score of 46 %, highlighting not only variability and ambiguity but also the extensive number of new resources introduced. A machine learning approach achieved an F-score of 63 % (with precision of 74 %) and 70 % (with precision of 83 %) for strict and lenient matching respectively. We characterise the issues with various mention types and propose potential ways of capturing additional database and software mentions in the literature. Our analyses show that identification of mentions of databases and tools is a challenging task that cannot be achieved by relying on current manually-curated resource repositories. Although machine learning shows improvement and promise (primarily in precision), more contextual information needs to be taken into account to achieve a good degree of accuracy.

  18. Metacomputing on Commodity Computers

    DTIC Science & Technology

    1999-05-01

    on NOWs, and this has contributed to the popularity of systems such as PVM [59], MPI [67], Linda [33], and TreadMarks [2]. 26 Challenges Given that...presents the performance of Calypso and Persistent Linda (PLinda) [77] programs and compares how they can tolerate failures. A biological pattern...adds fault tolerance to Linda programs by using light-weight transac- tions, whereas Calypso uses the combination of eager scheduling and two-phase

  19. 78 FR 934 - 37 Wilton Road, Milford LLC, and 282 Route 101 LLC, PVM Commercial Center, LLC; Notice of...

    Federal Register 2010, 2011, 2012, 2013, 2014

    2013-01-07

    ... printed on the eLibrary link of Commission's Web site at http://www.ferc.gov/docs-filing/elibrary.asp... Commission's Web site under http://www.ferc.gov/docs-filing/efiling.asp . Commenters can submit brief comments up to 6,000 characters, without prior registration, using the eComment system at http://www.ferc...

  20. Machine learning research 1989-90

    NASA Technical Reports Server (NTRS)

    Porter, Bruce W.; Souther, Arthur

    1990-01-01

    Multifunctional knowledge bases offer a significant advance in artificial intelligence because they can support numerous expert tasks within a domain. As a result they amortize the costs of building a knowledge base over multiple expert systems and they reduce the brittleness of each system. Due to the inevitable size and complexity of multifunctional knowledge bases, their construction and maintenance require knowledge engineering and acquisition tools that can automatically identify interactions between new and existing knowledge. Furthermore, their use requires software for accessing those portions of the knowledge base that coherently answer questions. Considerable progress was made in developing software for building and accessing multifunctional knowledge bases. A language was developed for representing knowledge, along with software tools for editing and displaying knowledge, a machine learning program for integrating new information into existing knowledge, and a question answering system for accessing the knowledge base.

  1. Hybrid Optimization Parallel Search PACKage

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    2009-11-10

    HOPSPACK is open source software for solving optimization problems without derivatives. Application problems may have a fully nonlinear objective function, bound constraints, and linear and nonlinear constraints. Problem variables may be continuous, integer-valued, or a mixture of both. The software provides a framework that supports any derivative-free type of solver algorithm. Through the framework, solvers request parallel function evaluation, which may use MPI (multiple machines) or multithreading (multiple processors/cores on one machine). The framework provides a Cache and Pending Cache of saved evaluations that reduces execution time and facilitates restarts. Solvers can dynamically create other algorithms to solve subproblems, amore » useful technique for handling multiple start points and integer-valued variables. HOPSPACK ships with the Generating Set Search (GSS) algorithm, developed at Sandia as part of the APPSPACK open source software project.« less

  2. Model-based machine learning.

    PubMed

    Bishop, Christopher M

    2013-02-13

    Several decades of research in the field of machine learning have resulted in a multitude of different algorithms for solving a broad range of problems. To tackle a new application, a researcher typically tries to map their problem onto one of these existing methods, often influenced by their familiarity with specific algorithms and by the availability of corresponding software implementations. In this study, we describe an alternative methodology for applying machine learning, in which a bespoke solution is formulated for each new application. The solution is expressed through a compact modelling language, and the corresponding custom machine learning code is then generated automatically. This model-based approach offers several major advantages, including the opportunity to create highly tailored models for specific scenarios, as well as rapid prototyping and comparison of a range of alternative models. Furthermore, newcomers to the field of machine learning do not have to learn about the huge range of traditional methods, but instead can focus their attention on understanding a single modelling environment. In this study, we show how probabilistic graphical models, coupled with efficient inference algorithms, provide a very flexible foundation for model-based machine learning, and we outline a large-scale commercial application of this framework involving tens of millions of users. We also describe the concept of probabilistic programming as a powerful software environment for model-based machine learning, and we discuss a specific probabilistic programming language called Infer.NET, which has been widely used in practical applications.

  3. Model-based machine learning

    PubMed Central

    Bishop, Christopher M.

    2013-01-01

    Several decades of research in the field of machine learning have resulted in a multitude of different algorithms for solving a broad range of problems. To tackle a new application, a researcher typically tries to map their problem onto one of these existing methods, often influenced by their familiarity with specific algorithms and by the availability of corresponding software implementations. In this study, we describe an alternative methodology for applying machine learning, in which a bespoke solution is formulated for each new application. The solution is expressed through a compact modelling language, and the corresponding custom machine learning code is then generated automatically. This model-based approach offers several major advantages, including the opportunity to create highly tailored models for specific scenarios, as well as rapid prototyping and comparison of a range of alternative models. Furthermore, newcomers to the field of machine learning do not have to learn about the huge range of traditional methods, but instead can focus their attention on understanding a single modelling environment. In this study, we show how probabilistic graphical models, coupled with efficient inference algorithms, provide a very flexible foundation for model-based machine learning, and we outline a large-scale commercial application of this framework involving tens of millions of users. We also describe the concept of probabilistic programming as a powerful software environment for model-based machine learning, and we discuss a specific probabilistic programming language called Infer.NET, which has been widely used in practical applications. PMID:23277612

  4. CMMI (registered trademark) for Services, Version 1.2

    DTIC Science & Technology

    2009-02-01

    background in information technology, especially those familiar with disciplines like service - oriented architecture (SOA) or software as a service ( SaaS ). In... services , the Software Engineering Institute (SEI) has found several dimensions that an organization can focus on to improve its business. Figure...International Business Machines) and the SEI [Humphrey 1989]. Humphrey’s book, Managing the Software Process, provides a CMMI for Services Version 1.2

  5. Development of a support software system for real-time HAL/S applications

    NASA Technical Reports Server (NTRS)

    Smith, R. S.

    1984-01-01

    Methodologies employed in defining and implementing a software support system for the HAL/S computer language for real-time operations on the Shuttle are detailed. Attention is also given to the management and validation techniques used during software development and software maintenance. Utilities developed to support the real-time operating conditions are described. With the support system being produced on Cyber computers and executable code then processed through Cyber or PDP machines, the support system has a production level status and can serve as a model for other software development projects.

  6. Secure Video Surveillance System Acquisition Software

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    2009-12-04

    The SVSS Acquisition Software collects and displays video images from two cameras through a VPN, and store the images onto a collection controller. The software is configured to allow a user to enter a time window to display up to 2 1/2, hours of video review. The software collects images from the cameras at a rate of 1 image per second and automatically deletes images older than 3 hours. The software code operates in a linux environment and can be run in a virtual machine on Windows XP. The Sandia software integrates the different COTS software together to build themore » video review system.« less

  7. Airlift Operation Modeling Using Discrete Event Simulation (DES)

    DTIC Science & Technology

    2009-12-01

    Java ......................................................................................................20 2. Simkit...JRE Java Runtime Environment JVM Java Virtual Machine lbs Pounds LAM Load Allocation Mode LRM Landing Spot Reassignment Mode LEGO Listener Event...SOFTWARE DEVELOPMENT ENVIRONMENT The following are the software tools and development environment used for constructing the models. 1. Java Java

  8. Computerizing Your Program.

    ERIC Educational Resources Information Center

    Curtis, Rick

    This paper summarizes information about using computer hardware and software to aid in making purchase decisions that are based on user needs. The two major options in hardware are IBM-compatible machines and the Apple Macintosh line. The three basic software applications include word processing, database management, and spreadsheet applications.…

  9. The Bartlesville System; TGISS Software Documentation.

    ERIC Educational Resources Information Center

    Roberts, Tommy L.; And Others

    TGISS (Total Guidance Information Support System) is an information storage and retrieval system specifically designed to meet the needs and requirements of a counselor in the Bartlesville Public School environment. The system, which is a combination of man/machine capabilities, includes the hardware and software necessary to extend the…

  10. A Recommender System in the Cyber Defense Domain

    DTIC Science & Technology

    2014-03-27

    monitoring software is a java based program sending updates to the database on the sensor machine. The host monitoring program gathers information about...3.2.2 Database. A MySQL database located on the sensor machine acts as the storage for the sensors on the network. Snort, Nmap, vulnerability scores, and...machine with the IDS and the recommender is labeled “sensor”. The recommender system code is written in java and compiled using java version 1.6.024

  11. CAT-PUMA: CME Arrival Time Prediction Using Machine learning Algorithms

    NASA Astrophysics Data System (ADS)

    Liu, Jiajia; Ye, Yudong; Shen, Chenglong; Wang, Yuming; Erdélyi, Robert

    2018-04-01

    CAT-PUMA (CME Arrival Time Prediction Using Machine learning Algorithms) quickly and accurately predicts the arrival of Coronal Mass Ejections (CMEs) of CME arrival time. The software was trained via detailed analysis of CME features and solar wind parameters using 182 previously observed geo-effective partial-/full-halo CMEs and uses algorithms of the Support Vector Machine (SVM) to make its predictions, which can be made within minutes of providing the necessary input parameters of a CME.

  12. The Next Era: Deep Learning in Pharmaceutical Research.

    PubMed

    Ekins, Sean

    2016-11-01

    Over the past decade we have witnessed the increasing sophistication of machine learning algorithms applied in daily use from internet searches, voice recognition, social network software to machine vision software in cameras, phones, robots and self-driving cars. Pharmaceutical research has also seen its fair share of machine learning developments. For example, applying such methods to mine the growing datasets that are created in drug discovery not only enables us to learn from the past but to predict a molecule's properties and behavior in future. The latest machine learning algorithm garnering significant attention is deep learning, which is an artificial neural network with multiple hidden layers. Publications over the last 3 years suggest that this algorithm may have advantages over previous machine learning methods and offer a slight but discernable edge in predictive performance. The time has come for a balanced review of this technique but also to apply machine learning methods such as deep learning across a wider array of endpoints relevant to pharmaceutical research for which the datasets are growing such as physicochemical property prediction, formulation prediction, absorption, distribution, metabolism, excretion and toxicity (ADME/Tox), target prediction and skin permeation, etc. We also show that there are many potential applications of deep learning beyond cheminformatics. It will be important to perform prospective testing (which has been carried out rarely to date) in order to convince skeptics that there will be benefits from investing in this technique.

  13. Implementing finite state machines in a computer-based teaching system

    NASA Astrophysics Data System (ADS)

    Hacker, Charles H.; Sitte, Renate

    1999-09-01

    Finite State Machines (FSM) are models for functions commonly implemented in digital circuits such as timers, remote controls, and vending machines. Teaching FSM is core in the curriculum of many university digital electronic or discrete mathematics subjects. Students often have difficulties grasping the theoretical concepts in the design and analysis of FSM. This has prompted the author to develop an MS-WindowsTM compatible software, WinState, that provides a tutorial style teaching aid for understanding the mechanisms of FSM. The animated computer screen is ideal for visually conveying the required design and analysis procedures. WinState complements other software for combinatorial logic previously developed by the author, and enhances the existing teaching package by adding sequential logic circuits. WinState enables the construction of a students own FSM, which can be simulated, to test the design for functionality and possible errors.

  14. Provenance-Based Approaches to Semantic Web Service Discovery and Usage

    ERIC Educational Resources Information Center

    Narock, Thomas William

    2012-01-01

    The World Wide Web Consortium defines a Web Service as "a software system designed to support interoperable machine-to-machine interaction over a network." Web Services have become increasingly important both within and across organizational boundaries. With the recent advent of the Semantic Web, web services have evolved into semantic…

  15. Certification of highly complex safety-related systems.

    PubMed

    Reinert, D; Schaefer, M

    1999-01-01

    The BIA has now 15 years of experience with the certification of complex electronic systems for safety-related applications in the machinery sector. Using the example of machining centres this presentation will show the systematic procedure for verifying and validating control systems using Application Specific Integrated Circuits (ASICs) and microcomputers for safety functions. One section will describe the control structure of machining centres with control systems using "integrated safety." A diverse redundant architecture combined with crossmonitoring and forced dynamization is explained. In the main section the steps of the systematic certification procedure are explained showing some results of the certification of drilling machines. Specification reviews, design reviews with test case specification, statistical analysis, and walk-throughs are the analytical measures in the testing process. Systematic tests based on the test case specification, Electro Magnetic Interference (EMI), and environmental testing, and site acceptance tests on the machines are the testing measures for validation. A complex software driven system is always undergoing modification. Most of the changes are not safety-relevant but this has to be proven. A systematic procedure for certifying software modifications is presented in the last section of the paper.

  16. AstroCloud, a Cyber-Infrastructure for Astronomy Research: Data Access and Interoperability

    NASA Astrophysics Data System (ADS)

    Fan, D.; He, B.; Xiao, J.; Li, S.; Li, C.; Cui, C.; Yu, C.; Hong, Z.; Yin, S.; Wang, C.; Cao, Z.; Fan, Y.; Mi, L.; Wan, W.; Wang, J.

    2015-09-01

    Data access and interoperability module connects the observation proposals, data, virtual machines and software. According to the unique identifier of PI (principal investigator), an email address or an internal ID, data can be collected by PI's proposals, or by the search interfaces, e.g. conesearch. Files associated with the searched results could be easily transported to cloud storages, including the storage with virtual machines, or several commercial platforms like Dropbox. Benefitted from the standards of IVOA (International Observatories Alliance), VOTable formatted searching result could be sent to kinds of VO software. Latter endeavor will try to integrate more data and connect archives and some other astronomical resources.

  17. Ethoscopes: An open platform for high-throughput ethomics.

    PubMed

    Geissmann, Quentin; Garcia Rodriguez, Luis; Beckwith, Esteban J; French, Alice S; Jamasb, Arian R; Gilestro, Giorgio F

    2017-10-01

    Here, we present the use of ethoscopes, which are machines for high-throughput analysis of behavior in Drosophila and other animals. Ethoscopes provide a software and hardware solution that is reproducible and easily scalable. They perform, in real-time, tracking and profiling of behavior by using a supervised machine learning algorithm, are able to deliver behaviorally triggered stimuli to flies in a feedback-loop mode, and are highly customizable and open source. Ethoscopes can be built easily by using 3D printing technology and rely on Raspberry Pi microcomputers and Arduino boards to provide affordable and flexible hardware. All software and construction specifications are available at http://lab.gilest.ro/ethoscope.

  18. Process Approach for Modeling of Machine and Tractor Fleet Structure

    NASA Astrophysics Data System (ADS)

    Dokin, B. D.; Aletdinova, A. A.; Kravchenko, M. S.; Tsybina, Y. S.

    2018-05-01

    The existing software complexes on modelling of the machine and tractor fleet structure are mostly aimed at solving the task of optimization. However, the creators, choosing only one optimization criterion and incorporating it in their software, provide grounds on why it is the best without giving a decision maker the opportunity to choose it for their enterprise. To analyze “bottlenecks” of machine and tractor fleet modelling, the authors of this article created a process model, in which they included adjustment to the plan of using machinery based on searching through alternative technologies. As a result, the following recommendations for software complex development have been worked out: the introduction of a database of alternative technologies; the possibility for a user to change the timing of the operations even beyond the allowable limits and in that case the calculation of the incurred loss; the possibility to rule out the solution of an optimization task, and if there is a necessity in it - the possibility to choose an optimization criterion; introducing graphical display of an annual complex of works, which could be enough for the development and adjustment of a business strategy.

  19. Training Knowledge Bots for Physics-Based Simulations Using Artificial Neural Networks

    NASA Technical Reports Server (NTRS)

    Samareh, Jamshid A.; Wong, Jay Ming

    2014-01-01

    Millions of complex physics-based simulations are required for design of an aerospace vehicle. These simulations are usually performed by highly trained and skilled analysts, who execute, monitor, and steer each simulation. Analysts rely heavily on their broad experience that may have taken 20-30 years to accumulate. In addition, the simulation software is complex in nature, requiring significant computational resources. Simulations of system of systems become even more complex and are beyond human capacity to effectively learn their behavior. IBM has developed machines that can learn and compete successfully with a chess grandmaster and most successful jeopardy contestants. These machines are capable of learning some complex problems much faster than humans can learn. In this paper, we propose using artificial neural network to train knowledge bots to identify the idiosyncrasies of simulation software and recognize patterns that can lead to successful simulations. We examine the use of knowledge bots for applications of computational fluid dynamics (CFD), trajectory analysis, commercial finite-element analysis software, and slosh propellant dynamics. We will show that machine learning algorithms can be used to learn the idiosyncrasies of computational simulations and identify regions of instability without including any additional information about their mathematical form or applied discretization approaches.

  20. Live interactive computer music performance practice

    NASA Astrophysics Data System (ADS)

    Wessel, David

    2002-05-01

    A live-performance musical instrument can be assembled around current lap-top computer technology. One adds a controller such as a keyboard or other gestural input device, a sound diffusion system, some form of connectivity processor(s) providing for audio I/O and gestural controller input, and reactive real-time native signal processing software. A system consisting of a hand gesture controller; software for gesture analysis and mapping, machine listening, composition, and sound synthesis; and a controllable radiation pattern loudspeaker are described. Interactivity begins in the set up wherein the speaker-room combination is tuned with an LMS procedure. This system was designed for improvisation. It is argued that software suitable for carrying out an improvised musical dialog with another performer poses special challenges. The processes underlying the generation of musical material must be very adaptable, capable of rapid changes in musical direction. Machine listening techniques are used to help the performer adapt to new contexts. Machine learning can play an important role in the development of such systems. In the end, as with any musical instrument, human skill is essential. Practice is required not only for the development of musically appropriate human motor programs but for the adaptation of the computer-based instrument as well.

  1. Organization of functional interaction of corporate information systems

    NASA Astrophysics Data System (ADS)

    Safronov, V. V.; Barabanov, V. F.; Podvalniy, S. L.; Nuzhnyy, A. M.

    2018-03-01

    In this article the methods of specialized software systems integration are analyzed and the concept of seamless integration of production decisions is offered. In view of this concept developed structural and functional schemes of the specialized software are shown. The proposed schemes and models are improved for a machine-building enterprise.

  2. Chips: A Tool for Developing Software Interfaces Interactively.

    ERIC Educational Resources Information Center

    Cunningham, Robert E.; And Others

    This report provides a detailed description of Chips, an interactive tool for developing software employing graphical/computer interfaces on Xerox Lisp machines. It is noted that Chips, which is implemented as a collection of customizable classes, provides the programmer with a rich graphical interface for the creation of rich graphical…

  3. Advanced technologies for Mission Control Centers

    NASA Technical Reports Server (NTRS)

    Dalton, John T.; Hughes, Peter M.

    1991-01-01

    Advance technologies for Mission Control Centers are presented in the form of the viewgraphs. The following subject areas are covered: technology needs; current technology efforts at GSFC (human-machine interface development, object oriented software development, expert systems, knowledge-based software engineering environments, and high performance VLSI telemetry systems); and test beds.

  4. Instructional Support Software System. Final Report.

    ERIC Educational Resources Information Center

    McDonnell Douglas Astronautics Co. - East, St. Louis, MO.

    This report describes the development of the Instructional Support System (ISS), a large-scale, computer-based training system that supports both computer-assisted instruction and computer-managed instruction. Written in the Ada programming language, the ISS software package is designed to be machine independent. It is also grouped into functional…

  5. Diamond turning machine controller implementation

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Garrard, K.P.; Taylor, L.W.; Knight, B.F.

    The standard controller for a Pnuemo ASG 2500 Diamond Turning Machine, an Allen Bradley 8200, has been replaced with a custom high-performance design. This controller consists of four major components. Axis position feedback information is provided by a Zygo Axiom 2/20 laser interferometer with 0.1 micro-inch resolution. Hardware interface logic couples the computers digital and analog I/O channels to the diamond turning machine`s analog motor controllers, the laser interferometer, and other machine status and control information. It also provides front panel switches for operator override of the computer controller and implement the emergency stop sequence. The remaining two components, themore » control computer hardware and software, are discussed in detail below.« less

  6. Understanding dental CAD/CAM for restorations - dental milling machines from a mechanical engineering viewpoint. Part A: chairside milling machines.

    PubMed

    Lebon, Nicolas; Tapie, Laurent; Duret, Francois; Attal, Jean-Pierre

    2016-01-01

    The dental milling machine is an important device in the dental CAD/CAM chain. Nowadays, dental numerical controlled (NC) milling machines are available for dental surgeries (chairside solution). This article provides a mechanical engineering approach to NC milling machines to help dentists understand the involvement of technology in digital dentistry practice. First, some technical concepts and definitions associated with NC milling machines are described from a mechanical engineering viewpoint. The technical and economic criteria of four chairside dental NC milling machines that are available on the market are then described. The technical criteria are focused on the capacities of the embedded technologies of these milling machines to mill both prosthetic materials and types of shape restorations. The economic criteria are focused on investment costs and interoperability with third-party software. The clinical relevance of the technology is assessed in terms of the accuracy and integrity of the restoration.

  7. Study on Gap Flow Field Simulation in Small Hole Machining of Ultrasonic Assisted EDM

    NASA Astrophysics Data System (ADS)

    Liu, Yu; Chang, Hao; Zhang, Wenchao; Ma, Fujian; Sha, Zhihua; Zhang, Shengfang

    2017-12-01

    When machining a small hole with high aspect ratio in EDM, it is hard for the flushing liquid entering the bottom gap and the debris could hardly be removed, which results in the accumulation of debris and affects the machining efficiency and machining accuracy. The assisted ultrasonic vibration can improve the removal of debris in the gap. Based on dynamics simulation software Fluent, a 3D model of debris movement in the gap flow field of EDM small hole machining assisted with side flushing and ultrasonic vibration is established in this paper. When depth to ratio is 3, the laws of different amplitudes and frequencies on debris distribution and removal are quantitatively analysed. The research results show that periodic ultrasonic vibration can promote the movement of debris, which is beneficial to the removal of debris in the machining gap. Compared to traditional small hole machining in EDM, the debris in the machining gap is greatly reduced, which ensures the stability of machining process and improves the machining efficiency.

  8. RELAP5-3D Resolution of Known Restart/Backup Issues

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Mesina, George L.; Anderson, Nolan A.

    2014-12-01

    The state-of-the-art nuclear reactor system safety analysis computer program developed at the Idaho National Laboratory (INL), RELAP5-3D, continues to adapt to changes in computer hardware and software and to develop to meet the ever-expanding needs of the nuclear industry. To continue at the forefront, code testing must evolve with both code and industry developments, and it must work correctly. To best ensure this, the processes of Software Verification and Validation (V&V) are applied. Verification compares coding against its documented algorithms and equations and compares its calculations against analytical solutions and the method of manufactured solutions. A form of this, sequentialmore » verification, checks code specifications against coding only when originally written then applies regression testing which compares code calculations between consecutive updates or versions on a set of test cases to check that the performance does not change. A sequential verification testing system was specially constructed for RELAP5-3D to both detect errors with extreme accuracy and cover all nuclear-plant-relevant code features. Detection is provided through a “verification file” that records double precision sums of key variables. Coverage is provided by a test suite of input decks that exercise code features and capabilities necessary to model a nuclear power plant. A matrix of test features and short-running cases that exercise them is presented. This testing system is used to test base cases (called null testing) as well as restart and backup cases. It can test RELAP5-3D performance in both standalone and coupled (through PVM to other codes) runs. Application of verification testing revealed numerous restart and backup issues in both standalone and couple modes. This document reports the resolution of these issues.« less

  9. Development of a Pediatric Visual Field Test

    PubMed Central

    Miranda, Marco A.; Henson, David B.; Fenerty, Cecilia; Biswas, Susmito; Aslam, Tariq

    2016-01-01

    Purpose We describe a pediatric visual field (VF) test based on a computer game where software and hardware combine to provide an enjoyable test experience. Methods The test software consists of a platform-based computer game presented to the central VF. A storyline was created around the game as was a structure surrounding the computer monitor to enhance patients' experience. The patient is asked to help the central character collect magic coins (stimuli). To collect these coins a series of obstacles need to be overcome. The test was presented on a Sony PVM-2541A monitor calibrated from a central midpoint with a Minolta CS-100 photometer placed at 50 cm. Measurements were performed at 15 locations on the screen and the contrast calculated. Retinal sensitivity was determined by modulating stimulus in size. To test the feasibility of the novel approach 20 patients (4–16 years old) with no history of VF defects were recruited. Results For the 14 subjects completing the study, 31 ± 15 data points were collected on 1 eye of each patient. Mean background luminance and stimulus contrast were 9.9 ± 0.3 cd/m2 and 27.9 ± 0.1 dB, respectively. Sensitivity values obtained were similar to an adult population but variability was considerably higher – 8.3 ± 9.0 dB. Conclusions Preliminary data show the feasibility of a game-based VF test for pediatric use. Although the test was well accepted by the target population, test variability remained very high. Translational Relevance Traditional VF tests are not well tolerated by children. This study describes a child-friendly approach to test visual fields in the targeted population. PMID:27980876

  10. Parallel machine architecture and compiler design facilities

    NASA Technical Reports Server (NTRS)

    Kuck, David J.; Yew, Pen-Chung; Padua, David; Sameh, Ahmed; Veidenbaum, Alex

    1990-01-01

    The objective is to provide an integrated simulation environment for studying and evaluating various issues in designing parallel systems, including machine architectures, parallelizing compiler techniques, and parallel algorithms. The status of Delta project (which objective is to provide a facility to allow rapid prototyping of parallelized compilers that can target toward different machine architectures) is summarized. Included are the surveys of the program manipulation tools developed, the environmental software supporting Delta, and the compiler research projects in which Delta has played a role.

  11. Proceedings of the 1986 IEEE international conference on systems, man and cybernetics

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Not Available

    1986-01-01

    This book presents the papers given at a conference on man-machine systems. Topics considered at the conference included neural model-based cognitive theory and engineering, user interfaces, adaptive and learning systems, human interaction with robotics, decision making, the testing and evaluation of expert systems, software development, international conflict resolution, intelligent interfaces, automation in man-machine system design aiding, knowledge acquisition in expert systems, advanced architectures for artificial intelligence, pattern recognition, knowledge bases, and machine vision.

  12. Machine translation project alternatives analysis

    NASA Technical Reports Server (NTRS)

    Bajis, Catherine J.; Bedford, Denise A. D.

    1993-01-01

    The Machine Translation Project consists of several components, two of which, the Project Plan and the Requirements Analysis, have already been delivered. The Project Plan details the overall rationale, objectives and time-table for the project as a whole. The Requirements Analysis compares a number of available machine translation systems, their capabilities, possible configurations, and costs. The Alternatives Analysis has resulted in a number of conclusions and recommendations to the NASA STI program concerning the acquisition of specific MT systems and related hardware and software.

  13. DOD Weapon Systems Software Management Study. Appendix E. Bibliography

    DTIC Science & Technology

    1975-06-01

    PROGRAMMING. -U- FakerF.T. IBM Systems Journal. V-11. N-1. P.56-73. 1972. REPORT. International Business Machines Corp. UNCLAS- .* SI’TIED lqpp. 1972. 42...MODULAR PROGRAMS. -U- Myers,G.J. MEMO. International Business Machines Corp., Poughkee- psier N. Y. Pouqhkeepsie Lab. UNCLASSIFIED 81pp. 29JAN73 U7... International Business Machines Corp., Whitp Plains, N. Y. UNCLASSIFIED 20pp. JUN74. 81. DATA SECURTTY AND DATA PROCESSING. VOLUME 2. STUDY SUMMARY. C-320- 1371

  14. An Overview of High Performance Computing and Challenges for the Future

    ScienceCinema

    Google Tech Talks

    2017-12-09

    In this talk we examine how high performance computing has changed over the last 10-year and look toward the future in terms of trends. These changes have had and will continue to have a major impact on our software. A new generation of software libraries and lgorithms are needed for the effective and reliable use of (wide area) dynamic, distributed and parallel environments. Some of the software and algorithm challenges have already been encountered, such as management of communication and memory hierarchies through a combination of compile--time and run--time techniques, but the increased scale of computation, depth of memory hierarchies, range of latencies, and increased run--time environment variability will make these problems much harder. We will focus on the redesign of software to fit multicore architectures. Speaker: Jack Dongarra University of Tennessee Oak Ridge National Laboratory University of Manchester Jack Dongarra received a Bachelor of Science in Mathematics from Chicago State University in 1972 and a Master of Science in Computer Science from the Illinois Institute of Technology in 1973. He received his Ph.D. in Applied Mathematics from the University of New Mexico in 1980. He worked at the Argonne National Laboratory until 1989, becoming a senior scientist. He now holds an appointment as University Distinguished Professor of Computer Science in the Electrical Engineering and Computer Science Department at the University of Tennessee, has the position of a Distinguished Research Staff member in the Computer Science and Mathematics Division at Oak Ridge National Laboratory (ORNL), Turing Fellow in the Computer Science and Mathematics Schools at the University of Manchester, and an Adjunct Professor in the Computer Science Department at Rice University. He specializes in numerical algorithms in linear algebra, parallel computing, the use of advanced-computer architectures, programming methodology, and tools for parallel computers. His research includes the development, testing and documentation of high quality mathematical software. He has contributed to the design and implementation of the following open source software packages and systems: EISPACK, LINPACK, the BLAS, LAPACK, ScaLAPACK, Netlib, PVM, MPI, NetSolve, Top500, ATLAS, and PAPI. He has published approximately 200 articles, papers, reports and technical memoranda and he is coauthor of several books. He was awarded the IEEE Sid Fernbach Award in 2004 for his contributions in the application of high performance computers using innovative approaches. He is a Fellow of the AAAS, ACM, and the IEEE and a member of the National Academy of Engineering.

  15. An Overview of High Performance Computing and Challenges for the Future

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Google Tech Talks

    In this talk we examine how high performance computing has changed over the last 10-year and look toward the future in terms of trends. These changes have had and will continue to have a major impact on our software. A new generation of software libraries and lgorithms are needed for the effective and reliable use of (wide area) dynamic, distributed and parallel environments. Some of the software and algorithm challenges have already been encountered, such as management of communication and memory hierarchies through a combination of compile--time and run--time techniques, but the increased scale of computation, depth of memory hierarchies,more » range of latencies, and increased run--time environment variability will make these problems much harder. We will focus on the redesign of software to fit multicore architectures. Speaker: Jack Dongarra University of Tennessee Oak Ridge National Laboratory University of Manchester Jack Dongarra received a Bachelor of Science in Mathematics from Chicago State University in 1972 and a Master of Science in Computer Science from the Illinois Institute of Technology in 1973. He received his Ph.D. in Applied Mathematics from the University of New Mexico in 1980. He worked at the Argonne National Laboratory until 1989, becoming a senior scientist. He now holds an appointment as University Distinguished Professor of Computer Science in the Electrical Engineering and Computer Science Department at the University of Tennessee, has the position of a Distinguished Research Staff member in the Computer Science and Mathematics Division at Oak Ridge National Laboratory (ORNL), Turing Fellow in the Computer Science and Mathematics Schools at the University of Manchester, and an Adjunct Professor in the Computer Science Department at Rice University. He specializes in numerical algorithms in linear algebra, parallel computing, the use of advanced-computer architectures, programming methodology, and tools for parallel computers. His research includes the development, testing and documentation of high quality mathematical software. He has contributed to the design and implementation of the following open source software packages and systems: EISPACK, LINPACK, the BLAS, LAPACK, ScaLAPACK, Netlib, PVM, MPI, NetSolve, Top500, ATLAS, and PAPI. He has published approximately 200 articles, papers, reports and technical memoranda and he is coauthor of several books. He was awarded the IEEE Sid Fernbach Award in 2004 for his contributions in the application of high performance computers using innovative approaches. He is a Fellow of the AAAS, ACM, and the IEEE and a member of the National Academy of Engineering.« less

  16. Design and optimization of production parameters for boric acid crystals with the crystallization process in an MSMPR crystallizer using FBRM® and PVM® technologies

    NASA Astrophysics Data System (ADS)

    Kutluay, Sinan; Şahin, Ömer; Ceyhan, A. Abdullah; İzgi, M. Sait

    2017-06-01

    In crystallization studies, newly developed technologies, such as Focused Beam Reflectance Measurement (FBRM) and Particle Vision and Measurement (PVM) are applied for determining on-line monitoring of a representation of the Chord Length Distribution (CLD) and observe the photographs of crystals respectively; moreover recently they are widely used. Properly installed, the FBRM ensures on-line determination of the CLD, which is statistically associated to the Crystal Size Distribution (CSD). In industrial crystallization, CSD and mean crystal size as well as external habit and internal structure are important characteristics for further use of the crystals. In this paper, the effect of residence time, stirring speed, feeding rate, supersaturation level and the polyelectrolytes such as anionic polyacrylamide (APAM) and non-ionic polyacrylamide (NPAM) on the CLD as well as the shape of boric acid crystals were investigated by using the FBRM G600 and the PVM V819 probes respectively in an MSMPR (Mixed Suspension Mixed Product Removal) crystallizer. The CSD and kinetic data were determined experimentally using continuous MSMPR crystallizer running at steady state. The population density of nuclei, the nucleation rate and the growth rate were determined from the experimental population balance distribution when the steady state was reached.

  17. Cytoplasmic remodeling of erythrocyte raft lipids during infection by the human malaria parasite Plasmodium falciparum

    PubMed Central

    Murphy, Sean C.; Fernandez-Pol, Sebastian; Chung, Paul H.; Prasanna Murthy, S. N.; Milne, Stephen B.; Salomao, Marcela; Brown, H. Alex; Lomasney, Jon W.; Mohandas, Narla

    2007-01-01

    Studies of detergent-resistant membrane (DRM) rafts in mature erythrocytes have facilitated identification of proteins that regulate formation of endovacuolar structures such as the parasitophorous vacuolar membrane (PVM) induced by the malaria parasite Plasmodium falciparum. However, analyses of raft lipids have remained elusive because detergents interfere with lipid detection. Here, we use primaquine to perturb the erythrocyte membrane and induce detergent-free buoyant vesicles, which are enriched in cholesterol and major raft proteins flotillin and stomatin and contain low levels of cytoskeleton, all characteristics of raft microdomains. Lipid mass spectrometry revealed that phosphatidylethanolamine and phosphatidylglycerol are depleted in endovesicles while phosphoinositides are highly enriched, suggesting raft-based endovesiculation can be achieved by simple (non–receptor-mediated) mechanical perturbation of the erythrocyte plasma membrane and results in sorting of inner leaflet phospholipids. Live-cell imaging of lipid-specific protein probes showed that phosphatidylinositol (4,5) bisphosphate (PIP2) is highly concentrated in primaquine-induced vesicles, confirming that it is an erythrocyte raft lipid. However, the malarial PVM lacks PIP2, although another raft lipid, phosphatidylserine, is readily detected. Thus, different remodeling/sorting of cytoplasmic raft phospholipids may occur in distinct endovacuoles. Importantly, erythrocyte raft lipids recruited to the invasion junction by mechanical stimulation may be remodeled by the malaria parasite to establish blood-stage infection. PMID:17526861

  18. Local Epidermal Growth Factor Receptor Signaling Mediates the Systemic Pathogenic Effects of Staphylococcus aureus Toxic Shock Syndrome.

    PubMed

    Breshears, Laura M; Gillman, Aaron N; Stach, Christopher S; Schlievert, Patrick M; Peterson, Marnie L

    2016-01-01

    Secreted factors of Staphylococcus aureus can activate host signaling from the epidermal growth factor receptor (EGFR). The superantigen toxic shock syndrome toxin-1 (TSST-1) contributes to mucosal cytokine production through a disintegrin and metalloproteinase (ADAM)-mediated shedding of EGFR ligands and subsequent EGFR activation. The secreted hemolysin, α-toxin, can also induce EGFR signaling and directly interacts with ADAM10, a sheddase of EGFR ligands. The current work explores the role of EGFR signaling in menstrual toxic shock syndrome (mTSS), a disease mediated by TSST-1. The data presented show that TSST-1 and α-toxin induce ADAM- and EGFR-dependent cytokine production from human vaginal epithelial cells. TSST-1 and α-toxin also induce cytokine production from an ex vivo porcine vaginal mucosa (PVM) model. EGFR signaling is responsible for the majority of IL-8 production from PVM in response to secreted toxins and live S. aureus. Finally, data are presented demonstrating that inhibition of EGFR signaling with the EGFR-specific tyrosine kinase inhibitor AG1478 significantly increases survival in a rabbit model of mTSS. These data indicate that EGFR signaling is critical for progression of an S. aureus exotoxin-mediated disease and may represent an attractive host target for therapeutics.

  19. The nonstructural proteins of Pneumoviruses are remarkably distinct in substrate diversity and specificity.

    PubMed

    Ribaudo, Michael; Barik, Sailen

    2017-11-06

    Interferon (IFN) inhibits viruses by inducing several hundred cellular genes, aptly named 'interferon (IFN)-stimulated genes' (ISGs). The only two RNA viruses of the Pneumovirus genus of the Paramyxoviridae family, namely Respiratory Syncytial Virus (RSV) and Pneumonia Virus of Mice (PVM), each encode two nonstructural (NS) proteins that share no sequence similarity but yet suppress IFN. Since suppression of IFN underlies the ability of these viruses to replicate in the host cells, the mechanism of such suppression has become an important area of research. This Short Report is an important extension of our previous efforts in defining this mechanism. We show that, like their PVM counterparts, the RSV NS proteins also target multiple members of the ISG family. While significantly extending the substrate repertoire of the RSV NS proteins, these results, unexpectedly, also reveal that the target preferences of the NS proteins of the two viruses are entirely different. This is surprising since the two Pneumoviruses are phylogenetically close with similar genome organization and gene function, and the NS proteins of both also serve as suppressors of host IFN response. The finding that the NS proteins of the two highly similar viruses suppress entirely different members of the ISG family raises intriguing questions of pneumoviral NS evolution and mechanism of action.

  20. New views of the Toxoplasma gondii parasitophorous vacuole as revealed by Helium Ion Microscopy (HIM).

    PubMed

    de Souza, Wanderley; Attias, Marcia

    2015-07-01

    The Helium Ion Microscope (HIM) is a new technology that uses a highly focused helium ion beam to scan and interact with the sample, which is not coated. The images have resolution and depth of field superior to field emission scanning electron microscopes. In this paper, we used HIM to study LLC-MK2 cells infected with Toxoplasma gondii. These samples were chemically fixed and, after critical point drying, were scraped with adhesive tape to expose the inner structure of the cell and parasitophorous vacuoles. We confirmed some of the previous findings made by field emission-scanning electron microscopy and showed that the surface of the parasite is rich in structures suggestive of secretion, that the nanotubules of the intravacuolar network (IVN) are not always straight, and that bifurcations are less frequent than previously thought. Fusion of the tubules with the parasite membrane or the parasitophorous vacuole membrane (PVM) was also infrequent. Tiny adhesive links were observed for the first time connecting the IVN tubules. The PVM showed openings of various sizes that even allowed the observation of endoplasmic reticulum membranes in the cytoplasm of the host cell. These findings are discussed in relation to current knowledge on the cell biology of T. gondii. Copyright © 2015 Elsevier Inc. All rights reserved.

  1. In Vitro Ability of a Novel Nanohydroxyapatite Oral Rinse to Occlude Dentine Tubules

    PubMed Central

    Hill, Robert G.; Chen, Xiaohui; Gillam, David G.

    2015-01-01

    Objectives. The aim of the study was to investigate the ability of a novel nanohydroxyapatite (nHA) desensitizing oral rinse to occlude dentine tubules compared to selected commercially available desensitizing oral rinses. Methods. 25 caries-free extracted molars were sectioned into 1 mm thick dentine discs. The dentine discs (n = 25) were etched with 6% citric acid for 2 minutes and rinsed with distilled water, prior to a 30-second application of test and control oral rinses. Evaluation was by (1) Scanning Electron Microscopy (SEM) of the dentine surface and (2) fluid flow measurements through a dentine disc. Results. Most of the oral rinses failed to adequately cover the dentine surface apart from the nHa oral rinse. However the hydroxyapatite, 1.4% potassium oxalate, and arginine/PVM/MA copolymer oral rinses, appeared to be relatively more effective than the nHA test and negative control rinses (potassium nitrate) in relation to a reduction in fluid flow measurements. Conclusions. Although the novel nHA oral rinse demonstrated the ability to occlude the dentine tubules and reduce the fluid flow measurements, some of the other oral rinses appeared to demonstrate a statistically significant reduction in fluid flow through the dentine disc, in particular the arginine/PVM/MA copolymer oral rinse. PMID:26161093

  2. PSGMiner: A modular software for polysomnographic analysis.

    PubMed

    Umut, İlhan

    2016-06-01

    Sleep disorders affect a great percentage of the population. The diagnosis of these disorders is usually made by polysomnography. This paper details the development of new software to carry out feature extraction in order to perform robust analysis and classification of sleep events using polysomnographic data. The software, called PSGMiner, is a tool, which visualizes, processes and classifies bioelectrical data. The purpose of this program is to provide researchers with a platform with which to test new hypotheses by creating tests to check for correlations that are not available in commercially available software. The software is freely available under the GPL3 License. PSGMiner is composed of a number of diverse modules such as feature extraction, annotation, and machine learning modules, all of which are accessible from the main module. Using the software, it is possible to extract features of polysomnography using digital signal processing and statistical methods and to perform different analyses. The features can be classified through the use of five classification algorithms. PSGMiner offers an architecture designed for integrating new methods. Automatic scoring, which is available in almost all commercial PSG software, is not inherently available in this program, though it can be implemented by two different methodologies (machine learning and algorithms). While similar software focuses on a certain signal or event composed of a small number of modules with no expansion possibility, the software introduced here can handle all polysomnographic signals and events. The software simplifies the processing of polysomnographic signals for researchers and physicians that are not experts in computer programming. It can find correlations between different events which could help predict an oncoming event such as sleep apnea. The software could also be used for educational purposes. Copyright © 2016 Elsevier Ltd. All rights reserved.

  3. Evaluation of rotor axial vibrations in a turbo pump unit equipped with an automatic unloading machine

    NASA Astrophysics Data System (ADS)

    Martsynkovskyy, V. A.; Deineka, A.; Kovalenko, V.

    2017-08-01

    The article presents forced axial vibrations of the rotor with an automatic unloading machine in an oxidizer pump. A feature of the design is the use in the autoloading system of slotted throttles with mutually inverse throttling. Their conductivity is determined by a numerical experiment in the ANSYS CFX software package.

  4. An Evaluation of Output Quality of Machine Translation (Padideh Software vs. Google Translate)

    ERIC Educational Resources Information Center

    Azer, Haniyeh Sadeghi; Aghayi, Mohammad Bagher

    2015-01-01

    This study aims to evaluate the translation quality of two machine translation systems in translating six different text-types, from English to Persian. The evaluation was based on criteria proposed by Van Slype (1979). The proposed model for evaluation is a black-box type, comparative and adequacy-oriented evaluation. To conduct the evaluation, a…

  5. The use of point-of-sale machines in school cafeterias as a method of parental influence over child lunch food choices

    USDA-ARS?s Scientific Manuscript database

    Computerized point-of-sale (POS) machine software that allows parents to place restrictions on their child’s school meal accounts is available. Parents could restrict specific foods (eg, chips), identify specific days the child can purchase extra foods, or set monetary limits. This descriptive study...

  6. A Multiple Sensor Machine Vision System for Automatic Hardwood Feature Detection

    Treesearch

    D. Earl Kline; Richard W. Conners; Daniel L. Schmoldt; Philip A. Araman; Robert L. Brisbin

    1993-01-01

    A multiple sensor machine vision prototype is being developed to scan full size hardwood lumber at industrial speeds for automatically detecting features such as knots holes, wane, stain, splits, checks, and color. The prototype integrates a multiple sensor imaging system, a materials handling system, a computer system, and application software. The prototype provides...

  7. Elevating Virtual Machine Introspection for Fine-Grained Process Monitoring: Techniques and Applications

    ERIC Educational Resources Information Center

    Srinivasan, Deepa

    2013-01-01

    Recent rapid malware growth has exposed the limitations of traditional in-host malware-defense systems and motivated the development of secure virtualization-based solutions. By running vulnerable systems as virtual machines (VMs) and moving security software from inside VMs to the outside, the out-of-VM solutions securely isolate the anti-malware…

  8. Applicability of SREM to the Verification of Management Information System Software Requirements. Volume I.

    DTIC Science & Technology

    1981-04-30

    However, SREM was not designed to harmonize these kinds of problems. Rather, it is a tool to investigate the logic of the processing specified in the... design . Supoorting programs were also conducted to perform basic research into such areas as software reliability, static and dynamic validation techniques...development. 0 Maintain requirements development independent of the target machine and the eventual software design . 0. Allow for easy response to

  9. Commonality and Variability Analysis for Xenon Family of Separation Virtual Machine Monitors (CVAX)

    DTIC Science & Technology

    2017-07-18

    technical approach is a systematic application of Software Product Line Engineering (SPLE). A systematic application requires describing the family and... engineering Software family September 2016 – October 2016 OSD/OUSD/ATL/ASD(R&E)/RDOffice of Information Systems & Cyber Security RD / ASD(R&E) / AT&L...by the evolving open-source Xen hypervisor. The technical approach is a systematic application of Software Product Line Engineering (SPLE). A

  10. Functional and Software Considerations for Bibliographic Data Base Utilization.

    ERIC Educational Resources Information Center

    Cadwallader, Gouverneur

    This is the fourth in a series of eight reports of a research study for the National Agricultural Library (NAL) on the effective utilization of bibliographic data bases in machine-readable form. It describes the general functional and software requirements of an NAL system using external sources of bibliographic data. Various system design…

  11. A new software-based architecture for quantum computer

    NASA Astrophysics Data System (ADS)

    Wu, Nan; Song, FangMin; Li, Xiangdong

    2010-04-01

    In this paper, we study a reliable architecture of a quantum computer and a new instruction set and machine language for the architecture, which can improve the performance and reduce the cost of the quantum computing. We also try to address some key issues in detail in the software-driven universal quantum computers.

  12. Dimensional Error in Rapid Prototyping with Open Source Software and Low-cost 3D-printer

    PubMed Central

    Andrade-Delgado, Laura; Telich-Tarriba, Jose E.; Fuente-del-Campo, Antonio; Altamirano-Arcos, Carlos A.

    2018-01-01

    Summary: Rapid prototyping models (RPMs) had been extensively used in craniofacial and maxillofacial surgery, especially in areas such as orthognathic surgery, posttraumatic or oncological reconstructions, and implantology. Economic limitations are higher in developing countries such as Mexico, where resources dedicated to health care are limited, therefore limiting the use of RPM to few selected centers. This article aims to determine the dimensional error of a low-cost fused deposition modeling 3D printer (Tronxy P802MA, Shenzhen, Tronxy Technology Co), with Open source software. An ordinary dry human mandible was scanned with a computed tomography device. The data were processed with open software to build a rapid prototype with a fused deposition machine. Linear measurements were performed to find the mean absolute and relative difference. The mean absolute and relative difference was 0.65 mm and 1.96%, respectively (P = 0.96). Low-cost FDM machines and Open Source Software are excellent options to manufacture RPM, with the benefit of low cost and a similar relative error than other more expensive technologies. PMID:29464171

  13. Dimensional Error in Rapid Prototyping with Open Source Software and Low-cost 3D-printer.

    PubMed

    Rendón-Medina, Marco A; Andrade-Delgado, Laura; Telich-Tarriba, Jose E; Fuente-Del-Campo, Antonio; Altamirano-Arcos, Carlos A

    2018-01-01

    Rapid prototyping models (RPMs) had been extensively used in craniofacial and maxillofacial surgery, especially in areas such as orthognathic surgery, posttraumatic or oncological reconstructions, and implantology. Economic limitations are higher in developing countries such as Mexico, where resources dedicated to health care are limited, therefore limiting the use of RPM to few selected centers. This article aims to determine the dimensional error of a low-cost fused deposition modeling 3D printer (Tronxy P802MA, Shenzhen, Tronxy Technology Co), with Open source software. An ordinary dry human mandible was scanned with a computed tomography device. The data were processed with open software to build a rapid prototype with a fused deposition machine. Linear measurements were performed to find the mean absolute and relative difference. The mean absolute and relative difference was 0.65 mm and 1.96%, respectively ( P = 0.96). Low-cost FDM machines and Open Source Software are excellent options to manufacture RPM, with the benefit of low cost and a similar relative error than other more expensive technologies.

  14. Pandora Operation and Analysis Software

    NASA Technical Reports Server (NTRS)

    Herman, Jay; Cede, Alexander; Abuhassan, Nader

    2012-01-01

    Pandora Operation and Analysis Software controls the Pandora Sun- and sky-pointing optical head and built-in filter wheels (neutral density, UV bandpass, polarization filters, and opaque). The software also controls the attached spectrometer exposure time and thermoelectric cooler to maintain the spectrometer temperature to within 1 C. All functions are available through a GUI so as to be easily accessible by the user. The data are automatically stored on a miniature computer (netbook) for automatic download to a designated server at user defined intervals (once per day, once per week, etc.), or to a USB external device. An additional software component reduces the raw data (spectrometer counts) to preliminary scientific products for quick-view purposes. The Pandora systems are built from off-the-shelf commercial parts and from mechanical parts machined using electronic machine shop drawings. The Pandora spectrometer system is designed to look at the Sun (tracking to within 0.1 ), or to look at the sky at any zenith or azimuth angle, to gather information about the amount of trace gases or aerosols that are present.

  15. Final Report: Enabling Exascale Hardware and Software Design through Scalable System Virtualization

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Bridges, Patrick G.

    2015-02-01

    In this grant, we enhanced the Palacios virtual machine monitor to increase its scalability and suitability for addressing exascale system software design issues. This included a wide range of research on core Palacios features, large-scale system emulation, fault injection, perfomrance monitoring, and VMM extensibility. This research resulted in large number of high-impact publications in well-known venues, the support of a number of students, and the graduation of two Ph.D. students and one M.S. student. In addition, our enhanced version of the Palacios virtual machine monitor has been adopted as a core element of the Hobbes operating system under active DOE-fundedmore » research and development.« less

  16. Use of containerisation as an alternative to full virtualisation in grid environments.

    NASA Astrophysics Data System (ADS)

    Long, Robin

    2015-12-01

    Virtualisation is a key tool on the grid. It can be used to provide varying work environments or as part of a cloud infrastructure. Virtualisation itself carries certain overheads that decrease the performance of the system through requiring extra resources to virtualise the software and hardware stack, and CPU-cycles wasted instantiating or destroying virtual machines for each job. With the rise and improvements in containerisation, where only the software stack is kept separate and no hardware or kernel virtualisation is used, there is scope for speed improvements and efficiency increases over standard virtualisation. We compare containerisation and virtualisation, including a comparison against bare-metal machines as a benchmark.

  17. Ethoscopes: An open platform for high-throughput ethomics

    PubMed Central

    Geissmann, Quentin; Garcia Rodriguez, Luis; Beckwith, Esteban J.; French, Alice S.; Jamasb, Arian R.

    2017-01-01

    Here, we present the use of ethoscopes, which are machines for high-throughput analysis of behavior in Drosophila and other animals. Ethoscopes provide a software and hardware solution that is reproducible and easily scalable. They perform, in real-time, tracking and profiling of behavior by using a supervised machine learning algorithm, are able to deliver behaviorally triggered stimuli to flies in a feedback-loop mode, and are highly customizable and open source. Ethoscopes can be built easily by using 3D printing technology and rely on Raspberry Pi microcomputers and Arduino boards to provide affordable and flexible hardware. All software and construction specifications are available at http://lab.gilest.ro/ethoscope. PMID:29049280

  18. The Next Era: Deep Learning in Pharmaceutical Research

    PubMed Central

    Ekins, Sean

    2016-01-01

    Over the past decade we have witnessed the increasing sophistication of machine learning algorithms applied in daily use from internet searches, voice recognition, social network software to machine vision software in cameras, phones, robots and self-driving cars. Pharmaceutical research has also seen its fair share of machine learning developments. For example, applying such methods to mine the growing datasets that are created in drug discovery not only enables us to learn from the past but to predict a molecule’s properties and behavior in future. The latest machine learning algorithm garnering significant attention is deep learning, which is an artificial neural network with multiple hidden layers. Publications over the last 3 years suggest that this algorithm may have advantages over previous machine learning methods and offer a slight but discernable edge in predictive performance. The time has come for a balanced review of this technique but also to apply machine learning methods such as deep learning across a wider array of endpoints relevant to pharmaceutical research for which the datasets are growing such as physicochemical property prediction, formulation prediction, absorption, distribution, metabolism, excretion and toxicity (ADME/Tox), target prediction and skin permeation, etc. We also show that there are many potential applications of deep learning beyond cheminformatics. It will be important to perform prospective testing (which has been carried out rarely to date) in order to convince skeptics that there will be benefits from investing in this technique. PMID:27599991

  19. Software Design Improvements. Part 1; Software Benefits and Limitations

    NASA Technical Reports Server (NTRS)

    Lalli, Vincent R.; Packard, Michael H.; Ziemianski, Tom

    1997-01-01

    Computer hardware and associated software have been used for many years to process accounting information, to analyze test data and to perform engineering analysis. Now computers and software also control everything from automobiles to washing machines and the number and type of applications are growing at an exponential rate. The size of individual program has shown similar growth. Furthermore, software and hardware are used to monitor and/or control potentially dangerous products and safety-critical systems. These uses include everything from airplanes and braking systems to medical devices and nuclear plants. The question is: how can this hardware and software be made more reliable? Also, how can software quality be improved? What methodology needs to be provided on large and small software products to improve the design and how can software be verified?

  20. Clock Agreement Among Parallel Supercomputer Nodes

    DOE Data Explorer

    Jones, Terry R.; Koenig, Gregory A.

    2014-04-30

    This dataset presents measurements that quantify the clock synchronization time-agreement characteristics among several high performance computers including the current world's most powerful machine for open science, the U.S. Department of Energy's Titan machine sited at Oak Ridge National Laboratory. These ultra-fast machines derive much of their computational capability from extreme node counts (over 18000 nodes in the case of the Titan machine). Time-agreement is commonly utilized by parallel programming applications and tools, distributed programming application and tools, and system software. Our time-agreement measurements detail the degree of time variance between nodes and how that variance changes over time. The dataset includes empirical measurements and the accompanying spreadsheets.

  1. Multi-objective optimization model of CNC machining to minimize processing time and environmental impact

    NASA Astrophysics Data System (ADS)

    Hamada, Aulia; Rosyidi, Cucuk Nur; Jauhari, Wakhid Ahmad

    2017-11-01

    Minimizing processing time in a production system can increase the efficiency of a manufacturing company. Processing time are influenced by application of modern technology and machining parameter. Application of modern technology can be apply by use of CNC machining, one of the machining process can be done with a CNC machining is turning. However, the machining parameters not only affect the processing time but also affect the environmental impact. Hence, optimization model is needed to optimize the machining parameters to minimize the processing time and environmental impact. This research developed a multi-objective optimization to minimize the processing time and environmental impact in CNC turning process which will result in optimal decision variables of cutting speed and feed rate. Environmental impact is converted from environmental burden through the use of eco-indicator 99. The model were solved by using OptQuest optimization software from Oracle Crystal Ball.

  2. The JPL telerobot operator control station. Part 2: Software

    NASA Technical Reports Server (NTRS)

    Kan, Edwin P.; Landell, B. Patrick; Oxenberg, Sheldon; Morimoto, Carl

    1989-01-01

    The Operator Control Station of the Jet Propulsion Laboratory (JPL)/NASA Telerobot Demonstrator System provides the man-machine interface between the operator and the system. It provides all the hardware and software for accepting human input for the direct and indirect (supervised) manipulation of the robot arms and tools for task execution. Hardware and software are also provided for the display and feedback of information and control data for the operator's consumption and interaction with the task being executed. The software design of the operator control system is discussed.

  3. Software development for teleroentgenogram analysis

    NASA Astrophysics Data System (ADS)

    Goshkoderov, A. A.; Khlebnikov, N. A.; Obabkov, I. N.; Serkov, K. V.; Gajniyarov, I. M.; Aliev, A. A.

    2017-09-01

    A framework for the analysis and calculation of teleroentgenograms was developed. Software development was carried out in the Department of Children's Dentistry and Orthodontics in Ural State Medical University. The software calculates the teleroentgenogram by the original method which was developed in this medical department. Program allows designing its own methods for calculating the teleroentgenograms by new methods. It is planned to use the technology of machine learning (Neural networks) in the software. This will help to make the process of calculating the teleroentgenograms easier because methodological points will be placed automatically.

  4. Applying CASE Tools for On-Board Software Development

    NASA Astrophysics Data System (ADS)

    Brammer, U.; Hönle, A.

    For many space projects the software development is facing great pressure with respect to quality, costs and schedule. One way to cope with these challenges is the application of CASE tools for automatic generation of code and documentation. This paper describes two CASE tools: Rhapsody (I-Logix) featuring UML and ISG (BSSE) that provides modeling of finite state machines. Both tools have been used at Kayser-Threde in different space projects for the development of on-board software. The tools are discussed with regard to the full software development cycle.

  5. Advanced fingerprint verification software

    NASA Astrophysics Data System (ADS)

    Baradarani, A.; Taylor, J. R. B.; Severin, F.; Maev, R. Gr.

    2016-05-01

    We have developed a fingerprint software package that can be used in a wide range of applications from law enforcement to public and private security systems, and to personal devices such as laptops, vehicles, and door- locks. The software and processing units are a unique implementation of new and sophisticated algorithms that compete with the current best systems in the world. Development of the software package has been in line with the third generation of our ultrasonic fingerprinting machine1. Solid and robust performance is achieved in the presence of misplaced and low quality fingerprints.

  6. Software packager user's guide

    NASA Technical Reports Server (NTRS)

    Callahan, John R.

    1995-01-01

    Software integration is a growing area of concern for many programmers and software managers because the need to build new programs quickly from existing components is greater than ever. This includes building versions of software products for multiple hardware platforms and operating systems, building programs from components written in different languages, and building systems from components that must execute on different machines in a distributed network. The goal of software integration is to make building new programs from existing components more seamless -- programmers should pay minimal attention to the underlying configuration issues involved. Libraries of reusable components and classes are important tools but only partial solutions to software development problems. Even though software components may have compatible interfaces, there may be other reasons, such as differences between execution environments, why they cannot be integrated. Often, components must be adapted or reimplemented to fit into another application because of implementation differences -- they are implemented in different programming languages, dependent on different operating system resources, or must execute on different physical machines. The software packager is a tool that allows programmers to deal with interfaces between software components and ignore complex integration details. The packager takes modular descriptions of the structure of a software system written in the package specification language and produces an integration program in the form of a makefile. If complex integration tools are needed to integrate a set of components, such as remote procedure call stubs, their use is implied by the packager automatically and stub generation tools are invoked in the corresponding makefile. The programmer deals only with the components themselves and not the details of how to build the system on any given platform.

  7. Program Converts VAX Floating-Point Data To UNIX

    NASA Technical Reports Server (NTRS)

    Alves, Marcos; Chapman, Bruce; Chu, Eugene

    1996-01-01

    VAX Floating Point to Host Floating Point Conversion (VAXFC) software converts non-ASCII files to unformatted floating-point representation of UNIX machine. This is done by reading bytes bit by bit, converting them to floating-point numbers, then writing results to another file. Useful when data files created by VAX computer must be used on other machines. Written in C language.

  8. The IHMC CmapTools software in research and education: a multi-level use case in Space Meteorology

    NASA Astrophysics Data System (ADS)

    Messerotti, Mauro

    2010-05-01

    The IHMC (Institute for Human and Machine Cognition, Florida University System, USA) CmapTools software is a powerful multi-platform tool for knowledge modelling in graphical form based on concept maps. In this work we present its application for the high-level development of a set of multi-level concept maps in the framework of Space Meteorology to act as the kernel of a space meteorology domain ontology. This is an example of a research use case, as a domain ontology coded in machine-readable form via e.g. OWL (Web Ontology Language) is suitable to be an active layer of any knowledge management system embedded in a Virtual Observatory (VO). Apart from being manageable at machine level, concept maps developed via CmapTools are intrinsically human-readable and can embed hyperlinks and objects of many kinds. Therefore they are suitable to be published on the web: the coded knowledge can be exploited for educational purposes by the students and the public, as the level of information can be naturally organized among linked concept maps in progressively increasing complexity levels. Hence CmapTools and its advanced version COE (Concept-map Ontology Editor) represent effective and user-friendly software tools for high-level knowledge represention in research and education.

  9. Concurrent Image Processing Executive (CIPE). Volume 1: Design overview

    NASA Technical Reports Server (NTRS)

    Lee, Meemong; Groom, Steven L.; Mazer, Alan S.; Williams, Winifred I.

    1990-01-01

    The design and implementation of a Concurrent Image Processing Executive (CIPE), which is intended to become the support system software for a prototype high performance science analysis workstation are described. The target machine for this software is a JPL/Caltech Mark 3fp Hypercube hosted by either a MASSCOMP 5600 or a Sun-3, Sun-4 workstation; however, the design will accommodate other concurrent machines of similar architecture, i.e., local memory, multiple-instruction-multiple-data (MIMD) machines. The CIPE system provides both a multimode user interface and an applications programmer interface, and has been designed around four loosely coupled modules: user interface, host-resident executive, hypercube-resident executive, and application functions. The loose coupling between modules allows modification of a particular module without significantly affecting the other modules in the system. In order to enhance hypercube memory utilization and to allow expansion of image processing capabilities, a specialized program management method, incremental loading, was devised. To minimize data transfer between host and hypercube, a data management method which distributes, redistributes, and tracks data set information was implemented. The data management also allows data sharing among application programs. The CIPE software architecture provides a flexible environment for scientific analysis of complex remote sensing image data, such as planetary data and imaging spectrometry, utilizing state-of-the-art concurrent computation capabilities.

  10. A Virtual Astronomical Research Machine in No Time (VARMiNT)

    NASA Astrophysics Data System (ADS)

    Beaver, John

    2012-05-01

    We present early results of using virtual machine software to help make astronomical research computing accessible to a wider range of individuals. Our Virtual Astronomical Research Machine in No Time (VARMiNT) is an Ubuntu Linux virtual machine with free, open-source software already installed and configured (and in many cases documented). The purpose of VARMiNT is to provide a ready-to-go astronomical research computing environment that can be freely shared between researchers, or between amateur and professional, teacher and student, etc., and to circumvent the often-difficult task of configuring a suitable computing environment from scratch. Thus we hope that VARMiNT will make it easier for individuals to engage in research computing even if they have no ready access to the facilities of a research institution. We describe our current version of VARMiNT and some of the ways it is being used at the University of Wisconsin - Fox Valley, a two-year teaching campus of the University of Wisconsin System, as a means to enhance student independent study research projects and to facilitate collaborations with researchers at other locations. We also outline some future plans and prospects.

  11. Application of Metamorphic Testing to Supervised Classifiers

    PubMed Central

    Xie, Xiaoyuan; Ho, Joshua; Kaiser, Gail; Xu, Baowen; Chen, Tsong Yueh

    2010-01-01

    Many applications in the field of scientific computing - such as computational biology, computational linguistics, and others - depend on Machine Learning algorithms to provide important core functionality to support solutions in the particular problem domains. However, it is difficult to test such applications because often there is no “test oracle” to indicate what the correct output should be for arbitrary input. To help address the quality of such software, in this paper we present a technique for testing the implementations of supervised machine learning classification algorithms on which such scientific computing software depends. Our technique is based on an approach called “metamorphic testing”, which has been shown to be effective in such cases. More importantly, we demonstrate that our technique not only serves the purpose of verification, but also can be applied in validation. In addition to presenting our technique, we describe a case study we performed on a real-world machine learning application framework, and discuss how programmers implementing machine learning algorithms can avoid the common pitfalls discovered in our study. We also discuss how our findings can be of use to other areas outside scientific computing, as well. PMID:21243103

  12. Windows .NET Network Distributed Basic Local Alignment Search Toolkit (W.ND-BLAST)

    PubMed Central

    Dowd, Scot E; Zaragoza, Joaquin; Rodriguez, Javier R; Oliver, Melvin J; Payton, Paxton R

    2005-01-01

    Background BLAST is one of the most common and useful tools for Genetic Research. This paper describes a software application we have termed Windows .NET Distributed Basic Local Alignment Search Toolkit (W.ND-BLAST), which enhances the BLAST utility by improving usability, fault recovery, and scalability in a Windows desktop environment. Our goal was to develop an easy to use, fault tolerant, high-throughput BLAST solution that incorporates a comprehensive BLAST result viewer with curation and annotation functionality. Results W.ND-BLAST is a comprehensive Windows-based software toolkit that targets researchers, including those with minimal computer skills, and provides the ability increase the performance of BLAST by distributing BLAST queries to any number of Windows based machines across local area networks (LAN). W.ND-BLAST provides intuitive Graphic User Interfaces (GUI) for BLAST database creation, BLAST execution, BLAST output evaluation and BLAST result exportation. This software also provides several layers of fault tolerance and fault recovery to prevent loss of data if nodes or master machines fail. This paper lays out the functionality of W.ND-BLAST. W.ND-BLAST displays close to 100% performance efficiency when distributing tasks to 12 remote computers of the same performance class. A high throughput BLAST job which took 662.68 minutes (11 hours) on one average machine was completed in 44.97 minutes when distributed to 17 nodes, which included lower performance class machines. Finally, there is a comprehensive high-throughput BLAST Output Viewer (BOV) and Annotation Engine components, which provides comprehensive exportation of BLAST hits to text files, annotated fasta files, tables, or association files. Conclusion W.ND-BLAST provides an interactive tool that allows scientists to easily utilizing their available computing resources for high throughput and comprehensive sequence analyses. The install package for W.ND-BLAST is freely downloadable from . With registration the software is free, installation, networking, and usage instructions are provided as well as a support forum. PMID:15819992

  13. UFMulti: A new parallel processing software system for HEP

    NASA Astrophysics Data System (ADS)

    Avery, Paul; White, Andrew

    1989-12-01

    UFMulti is a multiprocessing software package designed for general purpose high energy physics applications, including physics and detector simulation, data reduction and DST physics analysis. The system is particularly well suited for installations where several workstation or computers are connected through a local area network (LAN). The initial configuration of the software is currently running on VAX/VMS machines with a planned extension to ULTRIX, using the new RISC CPUs from Digital, in the near future.

  14. Laboratory process control using natural language commands from a personal computer

    NASA Technical Reports Server (NTRS)

    Will, Herbert A.; Mackin, Michael A.

    1989-01-01

    PC software is described which provides flexible natural language process control capability with an IBM PC or compatible machine. Hardware requirements include the PC, and suitable hardware interfaces to all controlled devices. Software required includes the Microsoft Disk Operating System (MS-DOS) operating system, a PC-based FORTRAN-77 compiler, and user-written device drivers. Instructions for use of the software are given as well as a description of an application of the system.

  15. Human-machine interface (HMI) report for 241-SY-101 data acquisition [and control] system (DACS) upgrade study

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Truitt, R.W.

    1997-10-22

    This report provides an independent evaluation of information for a Windows based Human Machine Interface (HMI) to replace the existing DOS based Iconics HMI currently used in the Data Acquisition and Control System (DACS) used at Tank 241-SY-101. A fundamental reason for this evaluation is because of the difficulty of maintaining the system with obsolete, unsupported software. The DACS uses a software operator interface (Genesis for DOS HMI) that is no longer supported by its manufacturer, Iconics. In addition to its obsolescence, it is complex and difficult to train additional personnel on. The FY 1997 budget allocated $40K for phasemore » 1 of a software/hardware upgrade that would have allowed the old DOS based system to be replaced by a current Windows based system. Unfortunately, budget constraints during FY 1997 has prompted deferral of the upgrade. The upgrade needs to be performed at the earliest possible time, before other failures render the system useless. Once completed, the upgrade could alleviate other concerns: spare pump software may be able to be incorporated into the same software as the existing pump, thereby eliminating the parallel path dilemma; and the newer, less complex software should expedite training of future personnel, and in the process, require that less technical time be required to maintain the system.« less

  16. [Extension of cardiac monitoring function by used of ordinary ECG machine].

    PubMed

    Chen, Zhencheng; Jiang, Yong; Ni, Lili; Wang, Hongyan

    2002-06-01

    This paper deals with a portable monitor system on liquid crystal display (LCD) based on this available ordinary ECG machine, which is low power and suitable for China's specific condition. Apart from developing the overall scheme of the system, this paper also has completed the design of the hardware and the software. The 80c196 single chip microcomputer is taken as the central microprocessor and real time electrocardiac single is data treated and analyzed in the system. With the performance of ordinary monitor, this machine also possesses the following functions: five types of arrhythmia analysis, alarm, freeze, and record of automatic pappering, convenient in carrying, with alternate-current (AC) or direct-current (DC) powered. The hardware circuit is simplified and the software structure is optimized in this paper. Multiple low power designs and LCD unit design are adopted and completed in it. Popular in usage, low in cost price, the portable monitor system will have a valuable influence on China's monitor system field.

  17. A low-cost machine vision system for the recognition and sorting of small parts

    NASA Astrophysics Data System (ADS)

    Barea, Gustavo; Surgenor, Brian W.; Chauhan, Vedang; Joshi, Keyur D.

    2018-04-01

    An automated machine vision-based system for the recognition and sorting of small parts was designed, assembled and tested. The system was developed to address a need to expose engineering students to the issues of machine vision and assembly automation technology, with readily available and relatively low-cost hardware and software. This paper outlines the design of the system and presents experimental performance results. Three different styles of plastic gears, together with three different styles of defective gears, were used to test the system. A pattern matching tool was used for part classification. Nine experiments were conducted to demonstrate the effects of changing various hardware and software parameters, including: conveyor speed, gear feed rate, classification, and identification score thresholds. It was found that the system could achieve a maximum system accuracy of 95% at a feed rate of 60 parts/min, for a given set of parameter settings. Future work will be looking at the effect of lighting.

  18. The potential of latent semantic analysis for machine grading of clinical case summaries.

    PubMed

    Kintsch, Walter

    2002-02-01

    This paper introduces latent semantic analysis (LSA), a machine learning method for representing the meaning of words, sentences, and texts. LSA induces a high-dimensional semantic space from reading a very large amount of texts. The meaning of words and texts can be represented as vectors in this space and hence can be compared automatically and objectively. A generative theory of the mental lexicon based on LSA is described. The word vectors LSA constructs are context free, and each word, irrespective of how many meanings or senses it has, is represented by a single vector. However, when a word is used in different contexts, context appropriate word senses emerge. Several applications of LSA to educational software are described, involving the ability of LSA to quickly compare the content of texts, such as an essay written by a student and a target essay. An LSA-based software tool is sketched for machine grading of clinical case summaries written by medical students.

  19. Table-driven software architecture for a stitching system

    NASA Technical Reports Server (NTRS)

    Thrash, Patrick J. (Inventor); Miller, Jeffrey L. (Inventor); Pallas, Ken (Inventor); Trank, Robert C. (Inventor); Fox, Rhoda (Inventor); Korte, Mike (Inventor); Codos, Richard (Inventor); Korolev, Alexandre (Inventor); Collan, William (Inventor)

    2001-01-01

    Native code for a CNC stitching machine is generated by generating a geometry model of a preform; generating tool paths from the geometry model, the tool paths including stitching instructions for making stitches; and generating additional instructions indicating thickness values. The thickness values are obtained from a lookup table. When the stitching machine runs the native code, it accesses a lookup table to determine a thread tension value corresponding to the thickness value. The stitching machine accesses another lookup table to determine a thread path geometry value corresponding to the thickness value.

  20. Interlock system for machine protection of the KOMAC 100-MeV proton linac

    NASA Astrophysics Data System (ADS)

    Song, Young-Gi

    2015-02-01

    The 100-MeV proton linear accelerator of the Korea Multi-purpose Accelerator Complex (KOMAC) has been developed. The beam service started this year after performing the beam commissioning. If the very sensitive and essential equipment is to be protected during machine operation, a machine interlock system is required, and the interlock system has been implemented. The purpose of the interlock system is to shut off the beam when the radio-frequency (RF) and ion source are unstable or a beam loss occurs. The interlock signal of the KOMAC linac includes a variety of sources, such as the beam loss, RF and high-voltage converter modulator faults, and fast closing valves of the vacuum window at the beam lines and so on. This system consists of a hardware-based interlock system using analog circuits and a software-based interlock system using an industrial programmable logic controller (PLC). The hardware-based interlock system has been fabricated, and the requirement has been satisfied with the results being within 10 µs. The software logic interlock system using the PLC has been connected to the framework of with the experimental physics and industrial control system (EPICS) to integrate a variety of interlock signals and to control the machine components when an interlock occurs. This paper will describe the design and the construction of the machine interlock system for the KOMAC 100-MeV linac.

  1. Predicting Software Suitability Using a Bayesian Belief Network

    NASA Technical Reports Server (NTRS)

    Beaver, Justin M.; Schiavone, Guy A.; Berrios, Joseph S.

    2005-01-01

    The ability to reliably predict the end quality of software under development presents a significant advantage for a development team. It provides an opportunity to address high risk components earlier in the development life cycle, when their impact is minimized. This research proposes a model that captures the evolution of the quality of a software product, and provides reliable forecasts of the end quality of the software being developed in terms of product suitability. Development team skill, software process maturity, and software problem complexity are hypothesized as driving factors of software product quality. The cause-effect relationships between these factors and the elements of software suitability are modeled using Bayesian Belief Networks, a machine learning method. This research presents a Bayesian Network for software quality, and the techniques used to quantify the factors that influence and represent software quality. The developed model is found to be effective in predicting the end product quality of small-scale software development efforts.

  2. A Talking Computers System for Persons with Vision and Speech Handicaps. Final Report.

    ERIC Educational Resources Information Center

    Visek & Maggs, Urbana, IL.

    This final report contains a detailed description of six software systems designed to assist individuals with blindness and/or speech disorders in using inexpensive, off-the-shelf computers rather than expensive custom-made devices. The developed software is not written in the native machine language of any particular brand of computer, but in the…

  3. Relationship between Effective Application of Machine Learning and Malware Detection: A Quantitative Study

    ERIC Educational Resources Information Center

    Enfinger, Kerry Wayne

    2016-01-01

    The number of malicious files present in the public domain continues to rise at a substantial rate. Current anti-malware software utilizes a signature-based method to detect the presence of malicious software. Generating these pattern signatures is time consuming due to malicious code complexity and the need for expert analysis, however, by making…

  4. Software Engineering for User Interfaces. Technical Report.

    ERIC Educational Resources Information Center

    Draper, Stephen W.; Norman, Donald A.

    The discipline of software engineering can be extended in a natural way to deal with the issues raised by a systematic approach to the design of human-machine interfaces. The user should be treated as part of the system being designed and projects should be organized to take into account the current lack of a priori knowledge of user interface…

  5. Use of Student Experiments for Teaching Embedded Software Development Including HW/SW Co-Design

    ERIC Educational Resources Information Center

    Mitsui, H.; Kambe, H.; Koizumi, H.

    2009-01-01

    Embedded systems have been applied widely, not only to consumer products and industrial machines, but also to new applications such as ubiquitous or sensor networking. The increasing role of software (SW) in embedded system development has caused a great demand for embedded SW engineers, and university education for embedded SW engineering has…

  6. Domain Engineering

    NASA Astrophysics Data System (ADS)

    Bjørner, Dines

    Before software can be designed we must know its requirements. Before requirements can be expressed we must understand the domain. So it follows, from our dogma, that we must first establish precise descriptions of domains; then, from such descriptions, “derive” at least domain and interface requirements; and from those and machine requirements design the software, or, more generally, the computing systems.

  7. Industrial Inspection with Open Eyes: Advance with Machine Vision Technology

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Liu, Zheng; Ukida, H.; Niel, Kurt

    Machine vision systems have evolved significantly with the technology advances to tackle the challenges from modern manufacturing industry. A wide range of industrial inspection applications for quality control are benefiting from visual information captured by different types of cameras variously configured in a machine vision system. This chapter screens the state of the art in machine vision technologies in the light of hardware, software tools, and major algorithm advances for industrial inspection. The inspection beyond visual spectrum offers a significant complementary to the visual inspection. The combination with multiple technologies makes it possible for the inspection to achieve a bettermore » performance and efficiency in varied applications. The diversity of the applications demonstrates the great potential of machine vision systems for industry.« less

  8. Software model of a machine vision system based on the common house fly.

    PubMed

    Madsen, Robert; Barrett, Steven; Wilcox, Michael

    2005-01-01

    The vision system of the common house fly has many properties, such as hyperacuity and parallel structure, which would be advantageous in a machine vision system. A software model has been developed which is ultimately intended to be a tool to guide the design of an analog real time vision system. The model starts by laying out cartridges over an image. The cartridges are analogous to the ommatidium of the fly's eye and contain seven photoreceptors each with a Gaussian profile. The spacing between photoreceptors is variable providing for more or less detail as needed. The cartridges provide information on what type of features they see and neighboring cartridges share information to construct a feature map.

  9. Development Of Knowledge Systems For Trouble Shooting Complex Production Machinery

    NASA Astrophysics Data System (ADS)

    Sanford, Richard L.; Novak, Thomas; Meigs, James R.

    1987-05-01

    This paper discusses the use of knowledge base system software for microcomputers to aid repairmen in diagnosing electrical failures in complex mining machinery. The knowledge base is constructed to allow the user to input initial symptoms of the failed machine, and the most probable cause of failure is traced through the knowledge base, with the software requesting additional information such as voltage or resistance measurements as needed. Although the case study presented is for an underground mining machine, results have application to any industry using complex machinery. Two commercial expert-system development tools (M1 TM and Insight 2+TM) and an Al language (Turbo PrologTM) are discussed with emphasis on ease of application and suitability for this study.

  10. GRIDVIEW: Recent Improvements in Research and Education Software for Exploring Mars Topography

    NASA Technical Reports Server (NTRS)

    Roark, J. H.; Masuoka, C. M.; Frey, H. V.

    2004-01-01

    GRIDVIEW is being developed by the GEODYNAMICS Branch at NASA's Goddard Space Flight Center and can be downloaded on the web at http://geodynamics.gsfc.nasa.gov/gridview/. The program is very mature and has been successfully used for more than four years, but is still under development as we add new features for data analysis and visualization. The software can run on any computer supported by the IDL virtual machine application supplied by RSI. The virtual machine application is currently available for recent versions of MS Windows, MacOS X, Red Hat Linux and UNIX. Minimum system memory requirement is 32 MB, however loading large data sets may require larger amounts of RAM to function adequately.

  11. A Prolog Emulator

    NASA Technical Reports Server (NTRS)

    Tick, Evan

    1987-01-01

    This note describes an efficient software emulator for the Warren Abstract Machine (WAM) Prolog architecture. The version of the WAM implemented is called Lcode. The Lcode emulator, written in C, executes the 'naive reverse' benchmark at 3900 LIPS. The emulator is one of a set of tools used to measure the memory-referencing characteristics and performance of Prolog programs. These tools include a compiler, assembler, and memory simulators. An overview of the Lcode architecture is given here, followed by a description and listing of the emulator code implementing each Lcode instruction. This note will be of special interest to those studying the WAM and its performance characteristics. In general, this note will be of interest to those creating efficient software emulators for abstract machine architectures.

  12. Mold Heating and Cooling Pump Package Operator Interface Controls Upgrade

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Josh A. Salmond

    2009-08-07

    The modernization of the Mold Heating and Cooling Pump Package Operator Interface (MHC PP OI) consisted of upgrading the antiquated single board computer with a proprietary operating system to off-the-shelf hardware and off-the-shelf software with customizable software options. The pump package is the machine interface between a central heating and cooling system that pumps heat transfer fluid through an injection or compression mold base on a local plastic molding machine. The operator interface provides the intelligent means of controlling this pumping process. Strict temperature control of a mold allows the production of high quality parts with tight tolerances and lowmore » residual stresses. The products fabricated are used on multiple programs.« less

  13. The Use of Point-of-Sale Machines in School Cafeterias as a Method of Parental Influence over Child Lunch Food Choices

    ERIC Educational Resources Information Center

    Andrepont, Emmy; Cullen, Karen W.; Taylor, Wendell C.

    2011-01-01

    Background: Computerized point-of-sale (POS) machine software that allows parents to place restrictions on their child's school meal accounts is available. Parents could restrict specific foods (e.g., chips), identify specific days the child can purchase extra foods, or set monetary limits. This descriptive study examines the use of parental…

  14. Running VisIt Software on the Peregrine System | High-Performance Computing

    Science.gov Websites

    kilobyte range. VisIt features a robust remote visualization capability. VisIt can be started on a local machine and used to visualize data on a remote compute cluster.The remote machine must be able to send VisIt module must be loaded as part of this process. To enable remote visualization the 'module load

  15. Research interface on a programmable ultrasound scanner.

    PubMed

    Shamdasani, Vijay; Bae, Unmin; Sikdar, Siddhartha; Yoo, Yang Mo; Karadayi, Kerem; Managuli, Ravi; Kim, Yongmin

    2008-07-01

    Commercial ultrasound machines in the past did not provide the ultrasound researchers access to raw ultrasound data. Lack of this ability has impeded evaluation and clinical testing of novel ultrasound algorithms and applications. Recently, we developed a flexible ultrasound back-end where all the processing for the conventional ultrasound modes, such as B, M, color flow and spectral Doppler, was performed in software. The back-end has been incorporated into a commercial ultrasound machine, the Hitachi HiVision 5500. The goal of this work is to develop an ultrasound research interface on the back-end for acquiring raw ultrasound data from the machine. The research interface has been designed as a software module on the ultrasound back-end. To increase the amount of raw ultrasound data that can be spooled in the limited memory available on the back-end, we have developed a method that can losslessly compress the ultrasound data in real time. The raw ultrasound data could be obtained in any conventional ultrasound mode, including duplex and triplex modes. Furthermore, use of the research interface does not decrease the frame rate or otherwise affect the clinical usability of the machine. The lossless compression of the ultrasound data in real time can increase the amount of data spooled by approximately 2.3 times, thus allowing more than 6s of raw ultrasound data to be acquired in all the modes. The interface has been used not only for early testing of new ideas with in vitro data from phantoms, but also for acquiring in vivo data for fine-tuning ultrasound applications and conducting clinical studies. We present several examples of how newer ultrasound applications, such as elastography, vibration imaging and 3D imaging, have benefited from this research interface. Since the research interface is entirely implemented in software, it can be deployed on existing HiVision 5500 ultrasound machines and may be easily upgraded in the future. The developed research interface can aid researchers in the rapid testing and clinical evaluation of new ultrasound algorithms and applications. Additionally, we believe that our approach would be applicable to designing research interfaces on other ultrasound machines.

  16. Optimization of large matrix calculations for execution on the Cray X-MP vector supercomputer

    NASA Technical Reports Server (NTRS)

    Hornfeck, William A.

    1988-01-01

    A considerable volume of large computational computer codes were developed for NASA over the past twenty-five years. This code represents algorithms developed for machines of earlier generation. With the emergence of the vector supercomputer as a viable, commercially available machine, an opportunity exists to evaluate optimization strategies to improve the efficiency of existing software. This result is primarily due to architectural differences in the latest generation of large-scale machines and the earlier, mostly uniprocessor, machines. A sofware package being used by NASA to perform computations on large matrices is described, and a strategy for conversion to the Cray X-MP vector supercomputer is also described.

  17. Concept Map Value Propagation for Tactical Intelligence

    DTIC Science & Technology

    2007-06-01

    meaningful diagrams: KMap, SmartDraw , MindGenius, and so on. However, CmapTools is the package we are using for this project. The software , produced by the...Cmap, driven by expected variability in the value of a datum and cost to get a new value. We use the CmapTools software developed with DoD support at... software developed with DoD support at the Institute of Human and Machine Cognition as a structural basis for creating and assessing tactical Cmaps. The

  18. A general observatory control software framework design for existing small and mid-size telescopes

    NASA Astrophysics Data System (ADS)

    Ge, Liang; Lu, Xiao-Meng; Jiang, Xiao-Jun

    2015-07-01

    A general framework for observatory control software would help to improve the efficiency of observation and operation of telescopes, and would also be advantageous for remote and joint observations. We describe a general framework for observatory control software, which considers principles of flexibility and inheritance to meet the expectations from observers and technical personnel. This framework includes observation scheduling, device control and data storage. The design is based on a finite state machine that controls the whole process.

  19. Physical Therapy Machine

    NASA Technical Reports Server (NTRS)

    1989-01-01

    Loredan Biomedical, Inc.'s LIDO, a computerized physical therapy system, was purchased by NASA in 1985 for evaluation as a Space Station Freedom exercise program. In 1986, while involved in an ARC muscle conditioning project, Malcom Bond, Loredan's chairman, designed an advanced software package for NASA which became the basis for LIDOSOFT software used in the commercially available system. The system employs a "proprioceptive" software program which perceives internal body conditions, induces perturbations to muscular effort and evaluates the response. Biofeedback on a screen allows a patient to observe his own performance.

  20. Open-source software for collision detection in external beam radiation therapy

    NASA Astrophysics Data System (ADS)

    Suriyakumar, Vinith M.; Xu, Renee; Pinter, Csaba; Fichtinger, Gabor

    2017-03-01

    PURPOSE: Collision detection for external beam radiation therapy (RT) is important for eliminating the need for dryruns that aim to ensure patient safety. Commercial treatment planning systems (TPS) offer this feature but they are expensive and proprietary. Cobalt-60 RT machines are a viable solution to RT practice in low-budget scenarios. However, such clinics are hesitant to invest in these machines due to a lack of affordable treatment planning software. We propose the creation of an open-source room's eye view visualization module with automated collision detection as part of the development of an open-source TPS. METHODS: An openly accessible linac 3D geometry model is sliced into the different components of the treatment machine. The model's movements are based on the International Electrotechnical Commission standard. Automated collision detection is implemented between the treatment machine's components. RESULTS: The room's eye view module was built in C++ as part of SlicerRT, an RT research toolkit built on 3D Slicer. The module was tested using head and neck and prostate RT plans. These tests verified that the module accurately modeled the movements of the treatment machine and radiation beam. Automated collision detection was verified using tests where geometric parameters of the machine's components were changed, demonstrating accurate collision detection. CONCLUSION: Room's eye view visualization and automated collision detection are essential in a Cobalt-60 treatment planning system. Development of these features will advance the creation of an open-source TPS that will potentially help increase the feasibility of adopting Cobalt-60 RT.

  1. Distributed communications and control network for robotic mining

    NASA Technical Reports Server (NTRS)

    Schiffbauer, William H.

    1989-01-01

    The application of robotics to coal mining machines is one approach pursued to increase productivity while providing enhanced safety for the coal miner. Toward that end, a network composed of microcontrollers, computers, expert systems, real time operating systems, and a variety of program languages are being integrated that will act as the backbone for intelligent machine operation. Actual mining machines, including a few customized ones, have been given telerobotic semiautonomous capabilities by applying the described network. Control devices, intelligent sensors and computers onboard these machines are showing promise of achieving improved mining productivity and safety benefits. Current research using these machines involves navigation, multiple machine interaction, machine diagnostics, mineral detection, and graphical machine representation. Guidance sensors and systems employed include: sonar, laser rangers, gyroscopes, magnetometers, clinometers, and accelerometers. Information on the network of hardware/software and its implementation on mining machines are presented. Anticipated coal production operations using the network are discussed. A parallelism is also drawn between the direction of present day underground coal mining research to how the lunar soil (regolith) may be mined. A conceptual lunar mining operation that employs a distributed communication and control network is detailed.

  2. Beamline Insertions Manager at Jefferson Lab

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Johnson, Michael C.

    2015-09-01

    The beam viewer system at Jefferson Lab provides operators and beam physicists with qualitative and quantitative information on the transverse electron beam properties. There are over 140 beam viewers installed on the 12 GeV CEBAF accelerator. This paper describes an upgrade consisting of replacing the EPICS-based system tasked with managing all viewers with a mixed system utilizing EPICS and high-level software. Most devices, particularly the beam viewers, cannot be safely inserted into the beam line during high-current beam operations. Software is partly responsible for protecting the machine from untimely insertions. The multiplicity of beam-blocking and beam-vulnerable devices motivates us tomore » try a data-driven approach. The beamline insertions application components are centrally managed and configured through an object-oriented software framework created for this purpose. A rules-based engine tracks the configuration and status of every device, along with the beam status of the machine segment containing the device. The application uses this information to decide on which device actions are allowed at any given time.« less

  3. Jacobian-free approximate solvers for hyperbolic systems: Application to relativistic magnetohydrodynamics

    NASA Astrophysics Data System (ADS)

    Castro, Manuel J.; Gallardo, José M.; Marquina, Antonio

    2017-10-01

    We present recent advances in PVM (Polynomial Viscosity Matrix) methods based on internal approximations to the absolute value function, and compare them with Chebyshev-based PVM solvers. These solvers only require a bound on the maximum wave speed, so no spectral decomposition is needed. Another important feature of the proposed methods is that they are suitable to be written in Jacobian-free form, in which only evaluations of the physical flux are used. This is particularly interesting when considering systems for which the Jacobians involve complex expressions, e.g., the relativistic magnetohydrodynamics (RMHD) equations. On the other hand, the proposed Jacobian-free solvers have also been extended to the case of approximate DOT (Dumbser-Osher-Toro) methods, which can be regarded as simple and efficient approximations to the classical Osher-Solomon method, sharing most of it interesting features and being applicable to general hyperbolic systems. To test the properties of our schemes a number of numerical experiments involving the RMHD equations are presented, both in one and two dimensions. The obtained results are in good agreement with those found in the literature and show that our schemes are robust and accurate, running stable under a satisfactory time step restriction. It is worth emphasizing that, although this work focuses on RMHD, the proposed schemes are suitable to be applied to general hyperbolic systems.

  4. [Differences in momentum development when standing up from a chair between elderly with and without frequent falls history].

    PubMed

    Guzmán, Rodrigo Antonio; Prado, Hugo Enrique; Porcel Melián, Helvio; Cordier, Benoit

    2009-01-01

    The momentum of the upper body (UB) during transfer sit-to-stand (STS) could be sensitive to the deterioration of dynamic postural control, and also the risk of falls. The aim of this study is to quantify the differences in the momentum development on UB during the STS in a sample of fall and no-fall elderly subjects. MATERIAL AND MEHODS: The sample consisted of twenty three voluntary elderly subjects (n=23), six elderly adults with antecedents of frequent falls (more than two within a year period) and seventeen without histories of frequent falls. Through a motion analysis system we registered the kinematics of UB during STS, from which we calculated the momentum of UB. The determined analysis variables were: the maximum values of the vertical (P(V)M) and horizontal (P(H)M) lineal momenta, the minimum (L(Max)) and maximum (L(Min)) values of the angular momentum and maximum trunk flexion (thetaM(UB)). No difference was observed in P(H)M, L(Max) and L(Min) (P>0.05) between both groups. However, a significant difference was found for the variable P(V)M (P=0.03) and thetaM(UB) (P=0.03) between both groups. We can conclude that, for the sample studied, the frequent fall condition relates to a smaller capacity to develop vertical momentum and increase flexion of the upper body.

  5. Toward Intelligent Software Defect Detection

    NASA Technical Reports Server (NTRS)

    Benson, Markland J.

    2011-01-01

    Source code level software defect detection has gone from state of the art to a software engineering best practice. Automated code analysis tools streamline many of the aspects of formal code inspections but have the drawback of being difficult to construct and either prone to false positives or severely limited in the set of defects that can be detected. Machine learning technology provides the promise of learning software defects by example, easing construction of detectors and broadening the range of defects that can be found. Pinpointing software defects with the same level of granularity as prominent source code analysis tools distinguishes this research from past efforts, which focused on analyzing software engineering metrics data with granularity limited to that of a particular function rather than a line of code.

  6. Intent Specifications: An Approach to Building Human-Centered Specifications

    NASA Technical Reports Server (NTRS)

    Leveson, Nancy G.

    1999-01-01

    This paper examines and proposes an approach to writing software specifications, based on research in systems theory, cognitive psychology, and human-machine interaction. The goal is to provide specifications that support human problem solving and the tasks that humans must perform in software development and evolution. A type of specification, called intent specifications, is constructed upon this underlying foundation.

  7. QUICK - An interactive software environment for engineering design

    NASA Technical Reports Server (NTRS)

    Skinner, David L.

    1989-01-01

    QUICK, an interactive software environment for engineering design, provides a programmable FORTRAN-like calculator interface to a wide range of data structures as well as both built-in and user created functions. QUICK also provides direct access to the operating systems of eight different machine architectures. The evolution of QUICK and a brief overview of the current version are presented.

  8. Software design and documentation language

    NASA Technical Reports Server (NTRS)

    Kleine, H.

    1977-01-01

    A communications medium to support the design and documentation of complex software applications is studied. The medium also provides the following: (1) a processor which can convert design specifications into an intelligible, informative machine reproducible document; (2) a design and documentation language with forms and syntax that are simple, unrestrictive, and communicative; and (3) methodology for effective use of the language and processor.

  9. A new optimization tool path planning for 3-axis end milling of free-form surfaces based on efficient machining intervals

    NASA Astrophysics Data System (ADS)

    Vu, Duy-Duc; Monies, Frédéric; Rubio, Walter

    2018-05-01

    A large number of studies, based on 3-axis end milling of free-form surfaces, seek to optimize tool path planning. Approaches try to optimize the machining time by reducing the total tool path length while respecting the criterion of the maximum scallop height. Theoretically, the tool path trajectories that remove the most material follow the directions in which the machined width is the largest. The free-form surface is often considered as a single machining area. Therefore, the optimization on the entire surface is limited. Indeed, it is difficult to define tool trajectories with optimal feed directions which generate largest machined widths. Another limiting point of previous approaches for effectively reduce machining time is the inadequate choice of the tool. Researchers use generally a spherical tool on the entire surface. However, the gains proposed by these different methods developed with these tools lead to relatively small time savings. Therefore, this study proposes a new method, using toroidal milling tools, for generating toolpaths in different regions on the machining surface. The surface is divided into several regions based on machining intervals. These intervals ensure that the effective radius of the tool, at each cutter-contact points on the surface, is always greater than the radius of the tool in an optimized feed direction. A parallel plane strategy is then used on the sub-surfaces with an optimal specific feed direction for each sub-surface. This method allows one to mill the entire surface with efficiency greater than with the use of a spherical tool. The proposed method is calculated and modeled using Maple software to find optimal regions and feed directions in each region. This new method is tested on a free-form surface. A comparison is made with a spherical cutter to show the significant gains obtained with a toroidal milling cutter. Comparisons with CAM software and experimental validations are also done. The results show the efficiency of the method.

  10. A gesture-controlled projection display for CT-guided interventions.

    PubMed

    Mewes, A; Saalfeld, P; Riabikin, O; Skalej, M; Hansen, C

    2016-01-01

    The interaction with interventional imaging systems within a sterile environment is a challenging task for physicians. Direct physician-machine interaction during an intervention is rather limited because of sterility and workspace restrictions. We present a gesture-controlled projection display that enables a direct and natural physician-machine interaction during computed tomography (CT)-based interventions. Therefore, a graphical user interface is projected on a radiation shield located in front of the physician. Hand gestures in front of this display are captured and classified using a leap motion controller. We propose a gesture set to control basic functions of intervention software such as gestures for 2D image exploration, 3D object manipulation and selection. Our methods were evaluated in a clinically oriented user study with 12 participants. The results of the performed user study confirm that the display and the underlying interaction concept are accepted by clinical users. The recognition of the gestures is robust, although there is potential for improvements. The gesture training times are less than 10 min, but vary heavily between the participants of the study. The developed gestures are connected logically to the intervention software and intuitive to use. The proposed gesture-controlled projection display counters current thinking, namely it gives the radiologist complete control of the intervention software. It opens new possibilities for direct physician-machine interaction during CT-based interventions and is well suited to become an integral part of future interventional suites.

  11. Dynamically allocated virtual clustering management system

    NASA Astrophysics Data System (ADS)

    Marcus, Kelvin; Cannata, Jess

    2013-05-01

    The U.S Army Research Laboratory (ARL) has built a "Wireless Emulation Lab" to support research in wireless mobile networks. In our current experimentation environment, our researchers need the capability to run clusters of heterogeneous nodes to model emulated wireless tactical networks where each node could contain a different operating system, application set, and physical hardware. To complicate matters, most experiments require the researcher to have root privileges. Our previous solution of using a single shared cluster of statically deployed virtual machines did not sufficiently separate each user's experiment due to undesirable network crosstalk, thus only one experiment could be run at a time. In addition, the cluster did not make efficient use of our servers and physical networks. To address these concerns, we created the Dynamically Allocated Virtual Clustering management system (DAVC). This system leverages existing open-source software to create private clusters of nodes that are either virtual or physical machines. These clusters can be utilized for software development, experimentation, and integration with existing hardware and software. The system uses the Grid Engine job scheduler to efficiently allocate virtual machines to idle systems and networks. The system deploys stateless nodes via network booting. The system uses 802.1Q Virtual LANs (VLANs) to prevent experimentation crosstalk and to allow for complex, private networks eliminating the need to map each virtual machine to a specific switch port. The system monitors the health of the clusters and the underlying physical servers and it maintains cluster usage statistics for historical trends. Users can start private clusters of heterogeneous nodes with root privileges for the duration of the experiment. Users also control when to shutdown their clusters.

  12. To select the best tool for generating 3D maintenance data and to set the detailed process for obtaining the 3D maintenance data

    NASA Astrophysics Data System (ADS)

    Prashanth, B. N.; Roy, Kingshuk

    2017-07-01

    Three Dimensional (3D) maintenance data provides a link between design and technical documentation creating interactive 3D graphical training and maintenance material. It becomes difficult for an operator to always go through huge paper manuals or come running to the computer for doing maintenance of a machine which makes the maintenance work fatigue. Above being the case, a 3D animation makes maintenance work very simple since, there is no language barrier. The research deals with the generation of 3D maintenance data of any given machine. The best tool for obtaining the 3D maintenance is selected and the tool is analyzed. Using the same tool, a detailed process for extracting the 3D maintenance data for any machine is set. This project aims at selecting the best tool for obtaining 3D maintenance data and to select the detailed process for obtaining 3D maintenance data. 3D maintenance reduces use of big volumes of manuals which creates human errors and makes the work of an operator fatiguing. Hence 3-D maintenance would help in training and maintenance and would increase productivity. 3Dvia when compared with Cortona 3D and Deep Exploration proves to be better than them. 3Dvia is good in data translation and it has the best renderings compared to the other two 3D maintenance software. 3Dvia is very user friendly and it has various options for creating 3D animations. Its Interactive Electronic Technical Publication (IETP) integration is also better than the other two software. Hence 3Dvia proves to be the best software for obtaining 3D maintenance data of any machine.

  13. The Design of the Automatic Control System of the Gripping-Belt Speed in Long-Rootstalk Traditional Chinese Herbal Harvester

    NASA Astrophysics Data System (ADS)

    Huang, Jinxia; Wang, Junfa; Yu, Yonghong

    This article aims to design a kind of gripping-belt speed automatic tracking system of traditional Chinese herbal harvester by AT89C52 single-chip micro computer as a core combined with fuzzy PID control algorithm. The system can adjust the gripping-belt speed in accordance with the variation of the machine's operation, so there is a perfect matching between the machine operation speed and the gripping-belt speed. The harvesting performance of the machine can be improved greatly. System design includes hardware and software.

  14. Application of dynamic milling in stainless steel processing

    NASA Astrophysics Data System (ADS)

    Shan, Wenju

    2017-09-01

    This paper mainly introduces the method of parameter setting for NC programming of stainless steel parts by dynamic milling. Stainless steel is of high plasticity and toughness, serious hard working, large cutting force, high temperature in cutting area and easy wear of tool. It is difficult to process material. Dynamic motion technology is the newest NC programming technology of Mastercam software. It is an advanced machining idea. The tool path generated by the dynamic motion technology is more smooth, more efficient and more stable in the machining process. Dynamic motion technology is very suitable for cutting hard machining materials.

  15. Pointright: a system to redirect mouse and keyboard control among multiple machines

    DOEpatents

    Johanson, Bradley E [Palo Alto, CA; Winograd, Terry A [Stanford, CA; Hutchins, Gregory M [Mountain View, CA

    2008-09-30

    The present invention provides a software system, PointRight, that allows for smooth and effortless control of pointing and input devices among multiple displays. With PointRight, a single free-floating mouse and keyboard can be used to control multiple screens. When the cursor reaches the edge of a screen it seamlessly moves to the adjacent screen and keyboard control is simultaneously redirected to the appropriate machine. Laptops may also redirect their keyboard and pointing device, and multiple pointers are supported simultaneously. The system automatically reconfigures itself as displays go on, go off, or change the machine they display.

  16. An Operating Environment for the Jellybean Machine

    DTIC Science & Technology

    1988-05-01

    MODEL 48 5.4.4 Restarting a Context The operating system provides one primitive message (RESTART-CONTEXT) and two system calls (XFERID and XFER.ADDR) to...efficient, powerful services is reqired to support this "stem. To provide this supportive operating environment, I developed an operating system kernel that...serves many of the initial needs of our machine. This Jellybean Operating System Software provides an object- based storage model, where typed

  17. DOE Office of Scientific and Technical Information (OSTI.GOV)

    Trujillo, Angelina Michelle

    Strategy, Planning, Acquiring- very large scale computing platforms come and go and planning for immensely scalable machines often precedes actual procurement by 3 years. Procurement can be another year or more. Integration- After Acquisition, machines must be integrated into the computing environments at LANL. Connection to scalable storage via large scale storage networking, assuring correct and secure operations. Management and Utilization – Ongoing operations, maintenance, and trouble shooting of the hardware and systems software at massive scale is required.

  18. CDC to CRAY FORTRAN conversion manual

    NASA Technical Reports Server (NTRS)

    Mcgary, C.; Diebert, D.

    1983-01-01

    Documentation describing software differences between two general purpose computers for scientific applications is presented. Descriptions of the use of the FORTRAN and FORTRAN 77 high level programming language on a CDC 7600 under SCOPE and a CRAY XMP under COS are offered. Itemized differences of the FORTRAN language sets of the two machines are also included. The material is accompanied by numerous examples of preferred programming techniques for the two machines.

  19. Low latency messages on distributed memory multiprocessors

    NASA Technical Reports Server (NTRS)

    Rosing, Matthew; Saltz, Joel

    1993-01-01

    Many of the issues in developing an efficient interface for communication on distributed memory machines are described and a portable interface is proposed. Although the hardware component of message latency is less than one microsecond on many distributed memory machines, the software latency associated with sending and receiving typed messages is on the order of 50 microseconds. The reason for this imbalance is that the software interface does not match the hardware. By changing the interface to match the hardware more closely, applications with fine grained communication can be put on these machines. Based on several tests that were run on the iPSC/860, an interface that will better match current distributed memory machines is proposed. The model used in the proposed interface consists of a computation processor and a communication processor on each node. Communication between these processors and other nodes in the system is done through a buffered network. Information that is transmitted is either data or procedures to be executed on the remote processor. The dual processor system is better suited for efficiently handling asynchronous communications compared to a single processor system. The ability to send data or procedure is very flexible for minimizing message latency, based on the type of communication being performed. The test performed and the proposed interface are described.

  20. Detection of Periodic Leg Movements by Machine Learning Methods Using Polysomnographic Parameters Other Than Leg Electromyography

    PubMed Central

    Umut, İlhan; Çentik, Güven

    2016-01-01

    The number of channels used for polysomnographic recording frequently causes difficulties for patients because of the many cables connected. Also, it increases the risk of having troubles during recording process and increases the storage volume. In this study, it is intended to detect periodic leg movement (PLM) in sleep with the use of the channels except leg electromyography (EMG) by analysing polysomnography (PSG) data with digital signal processing (DSP) and machine learning methods. PSG records of 153 patients of different ages and genders with PLM disorder diagnosis were examined retrospectively. A novel software was developed for the analysis of PSG records. The software utilizes the machine learning algorithms, statistical methods, and DSP methods. In order to classify PLM, popular machine learning methods (multilayer perceptron, K-nearest neighbour, and random forests) and logistic regression were used. Comparison of classified results showed that while K-nearest neighbour classification algorithm had higher average classification rate (91.87%) and lower average classification error value (RMSE = 0.2850), multilayer perceptron algorithm had the lowest average classification rate (83.29%) and the highest average classification error value (RMSE = 0.3705). Results showed that PLM can be classified with high accuracy (91.87%) without leg EMG record being present. PMID:27213008

  1. Detection of Periodic Leg Movements by Machine Learning Methods Using Polysomnographic Parameters Other Than Leg Electromyography.

    PubMed

    Umut, İlhan; Çentik, Güven

    2016-01-01

    The number of channels used for polysomnographic recording frequently causes difficulties for patients because of the many cables connected. Also, it increases the risk of having troubles during recording process and increases the storage volume. In this study, it is intended to detect periodic leg movement (PLM) in sleep with the use of the channels except leg electromyography (EMG) by analysing polysomnography (PSG) data with digital signal processing (DSP) and machine learning methods. PSG records of 153 patients of different ages and genders with PLM disorder diagnosis were examined retrospectively. A novel software was developed for the analysis of PSG records. The software utilizes the machine learning algorithms, statistical methods, and DSP methods. In order to classify PLM, popular machine learning methods (multilayer perceptron, K-nearest neighbour, and random forests) and logistic regression were used. Comparison of classified results showed that while K-nearest neighbour classification algorithm had higher average classification rate (91.87%) and lower average classification error value (RMSE = 0.2850), multilayer perceptron algorithm had the lowest average classification rate (83.29%) and the highest average classification error value (RMSE = 0.3705). Results showed that PLM can be classified with high accuracy (91.87%) without leg EMG record being present.

  2. Comparison of fMRI data analysis by SPM99 on different operating systems.

    PubMed

    Shinagawa, Hideo; Honda, Ei-ichi; Ono, Takashi; Kurabayashi, Tohru; Ohyama, Kimie

    2004-09-01

    The hardware chosen for fMRI data analysis may depend on the platform already present in the laboratory or the supporting software. In this study, we ran SPM99 software on multiple platforms to examine whether we could analyze fMRI data by SPM99, and to compare their differences and limitations in processing fMRI data, which can be attributed to hardware capabilities. Six normal right-handed volunteers participated in a study of hand-grasping to obtain fMRI data. Each subject performed a run that consisted of 98 images. The run was measured using a gradient echo-type echo planar imaging sequence on a 1.5T apparatus with a head coil. We used several personal computer (PC), Unix and Linux machines to analyze the fMRI data. There were no differences in the results obtained on several PC, Unix and Linux machines. The only limitations in processing large amounts of the fMRI data were found using PC machines. This suggests that the results obtained with different machines were not affected by differences in hardware components, such as the CPU, memory and hard drive. Rather, it is likely that the limitations in analyzing a huge amount of the fMRI data were due to differences in the operating system (OS).

  3. DOE Office of Scientific and Technical Information (OSTI.GOV)

    Kowalkowski, Jim; Lyon, Adam; Paterno, Marc

    Over the past few years, container technology has become increasingly promising as a means to seamlessly make our software available across a wider range of platforms. In December 2015, we decided to put together a set of docker images that serve as a demonstration of this container technology for managing a run-time environment for art-related software projects, and also serve as a set of test cases for evaluation of performance. Docker[1] containers provide a way to “wrap up a piece of software in a complete filesystem that contains everything it needs to run”. In combination with Shifter[2], such containers providemore » a way to run software developed and deployed on “typical” HEP platforms (such as SLF 6, in common use at Fermilab and on OSG platforms) on HPC facilities at NERSC. Docker containers provide a means of delivering software that can be run on a variety of hosts without needing to be compiled specially for each OS to be supported. This could substantially reduce the effort required to create and validate a new release, since one build could be suitable for use on both grid machines (both FermiGrid and OSG) as well as any machine capable of running the Docker container. In addition, docker containers may provide for a quick and easy way for users to install and use a software release in a standardized environment. This report contains the results and status of this demonstration and evaluation.« less

  4. iTOUGH2 V6.5

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Finsterle, Stefan A.

    2010-11-01

    iTOUGH2 (inverse TOUGH2) provides inverse modeling capabilities for TOUGH2, a simulator for multi-dimensional , multi-phase, multi-component, non-isothermal flow and transport in fractured porous media. It performs sensitivity analysis, parameter estimation, and uncertainty propagation, analysis in geosciences and reservoir engineering and other application areas. It supports a number of different combination of fluids and components [equation-of-state (EOS) modules]. In addition, the optimization routines implemented in iTOUGH2 can also be used or sensitivity analysis, automatic model calibration, and uncertainty quantification of any external code that uses text-based input and output files. This link is achieved by means of the PEST application programmingmore » interface. iTOUGH2 solves the inverse problem by minimizing a non-linear objective function of the weighted differences between model output and the corresponding observations. Multiple minimization algorithms (derivative fee, gradient-based and second-order; local and global) are available. iTOUGH2 also performs Latin Hypercube Monte Carlos simulation for uncertainty propagation analysis. A detailed residual and error analysis is provided. This upgrade includes new EOS modules (specifically EOS7c, ECO2N and TMVOC), hysteretic relative permeability and capillary pressure functions and the PEST API. More details can be found at http://esd.lbl.gov/iTOUGH2 and the publications cited there. Hardware Req.: Multi-platform; Related/auxiliary software PVM (if running in parallel).« less

  5. MLBCD: a machine learning tool for big clinical data.

    PubMed

    Luo, Gang

    2015-01-01

    Predictive modeling is fundamental for extracting value from large clinical data sets, or "big clinical data," advancing clinical research, and improving healthcare. Machine learning is a powerful approach to predictive modeling. Two factors make machine learning challenging for healthcare researchers. First, before training a machine learning model, the values of one or more model parameters called hyper-parameters must typically be specified. Due to their inexperience with machine learning, it is hard for healthcare researchers to choose an appropriate algorithm and hyper-parameter values. Second, many clinical data are stored in a special format. These data must be iteratively transformed into the relational table format before conducting predictive modeling. This transformation is time-consuming and requires computing expertise. This paper presents our vision for and design of MLBCD (Machine Learning for Big Clinical Data), a new software system aiming to address these challenges and facilitate building machine learning predictive models using big clinical data. The paper describes MLBCD's design in detail. By making machine learning accessible to healthcare researchers, MLBCD will open the use of big clinical data and increase the ability to foster biomedical discovery and improve care.

  6. A computer architecture for intelligent machines

    NASA Technical Reports Server (NTRS)

    Lefebvre, D. R.; Saridis, G. N.

    1992-01-01

    The theory of intelligent machines proposes a hierarchical organization for the functions of an autonomous robot based on the principle of increasing precision with decreasing intelligence. An analytic formulation of this theory using information-theoretic measures of uncertainty for each level of the intelligent machine has been developed. The authors present a computer architecture that implements the lower two levels of the intelligent machine. The architecture supports an event-driven programming paradigm that is independent of the underlying computer architecture and operating system. Execution-level controllers for motion and vision systems are briefly addressed, as well as the Petri net transducer software used to implement coordination-level functions. A case study illustrates how this computer architecture integrates real-time and higher-level control of manipulator and vision systems.

  7. Proton Therapy

    MedlinePlus

    ... effects of the treatment. top of page What equipment is used? Proton beam therapy uses special machines, ... tumor cells. top of page Who operates the equipment? With backgrounds in mechanical, electrical, software, hardware and ...

  8. ControlShell: A real-time software framework

    NASA Technical Reports Server (NTRS)

    Schneider, Stanley A.; Chen, Vincent W.; Pardo-Castellote, Gerardo

    1994-01-01

    The ControlShell system is a programming environment that enables the development and implementation of complex real-time software. It includes many building tools for complex systems, such as a graphical finite state machine (FSM) tool to provide strategic control. ControlShell has a component-based design, providing interface definitions and mechanisms for building real-time code modules along with providing basic data management. Some of the system-building tools incorporated in ControlShell are a graphical data flow editor, a component data requirement editor, and a state-machine editor. It also includes a distributed data flow package, an execution configuration manager, a matrix package, and an object database and dynamic binding facility. This paper presents an overview of ControlShell's architecture and examines the functions of several of its tools.

  9. Mathematical support for automated geometry analysis of lathe machining of oblique peakless round-nose tools

    NASA Astrophysics Data System (ADS)

    Filippov, A. V.; Tarasov, S. Yu; Podgornyh, O. A.; Shamarin, N. N.; Filippova, E. O.

    2017-01-01

    Automatization of engineering processes requires developing relevant mathematical support and a computer software. Analysis of metal cutting kinematics and tool geometry is a necessary key task at the preproduction stage. This paper is focused on developing a procedure for determining the geometry of oblique peakless round-nose tool lathe machining with the use of vector/matrix transformations. Such an approach allows integration into modern mathematical software packages in distinction to the traditional analytic description. Such an advantage is very promising for developing automated control of the preproduction process. A kinematic criterion for the applicable tool geometry has been developed from the results of this study. The effect of tool blade inclination and curvature on the geometry-dependent process parameters was evaluated.

  10. Credit WCT. Original 21" x 2A" color negative is housed ...

    Library of Congress Historic Buildings Survey, Historic Engineering Record, Historic Landscapes Survey

    Credit WCT. Original 2-1" x 2-A" color negative is housed in the JPL Photography Laboratory, Pasadena, California. The mixing pot of the 150-gallon (Size 16-PVM) Baker-Perkins vertical mixer appears in its lowered position, exposing the mixer paddles. JPL employees Harold "Andy" Anderson and Ron Wright in protective clothing demonstrate how to scrape mixed propellant from mixer blades (JPL negative JPL10284BC, 27 January 1989) - Jet Propulsion Laboratory Edwards Facility, Mixer, Edwards Air Force Base, Boron, Kern County, CA

  11. Automated Engineering Design (AED); An approach to automated documentation

    NASA Technical Reports Server (NTRS)

    Mcclure, C. W.

    1970-01-01

    The automated engineering design (AED) is reviewed, consisting of a high level systems programming language, a series of modular precoded subroutines, and a set of powerful software machine tools that effectively automate the production and design of new languages. AED is used primarily for development of problem and user-oriented languages. Software production phases are diagramed, and factors which inhibit effective documentation are evaluated.

  12. Guidelines for Processing and Cataloging Computer Software for Schools and Area Education Agencies. Suggestions to Aid Schools and AEAs.

    ERIC Educational Resources Information Center

    Martin, Elizabeth; And Others

    Based on definitions of a machine-readable data file (MRDF) taken from the Anglo-American Cataloging Rules, second edition (AACR2) and Standards for Cataloging Nonprint Materials, the following recommendations for processing items of computer software are provided: (1) base main and added entry determination on AACR2; (2) place designation of form…

  13. A Compatible Hardware/Software Reliability Prediction Model.

    DTIC Science & Technology

    1981-07-22

    machines. In particular, he was interested in the following problem: assu me that one has a collection of connected elements computing and transmitting...software reliability prediction model is desirable, the findings about the Weibull distribution are intriguing. After collecting failure data from several...capacitor, some of the added charge carriers are collected by the capacitor. If the added charge is sufficiently large, the information stored is changed

  14. Crux: Rapid Open Source Protein Tandem Mass Spectrometry Analysis

    PubMed Central

    2015-01-01

    Efficiently and accurately analyzing big protein tandem mass spectrometry data sets requires robust software that incorporates state-of-the-art computational, machine learning, and statistical methods. The Crux mass spectrometry analysis software toolkit (http://cruxtoolkit.sourceforge.net) is an open source project that aims to provide users with a cross-platform suite of analysis tools for interpreting protein mass spectrometry data. PMID:25182276

  15. Software Process Automation: Interviews, Survey, and Workshop Results.

    DTIC Science & Technology

    1997-10-01

    International Business Machines Coproration Foundation is a pending trademark of Foundation Software , Inc. FrameMaker is a registered trademark of Adobe, Inc...amount of technology Integration of technologies, con- flicting points of view between adopting org. and consultants E CM FrameMaker Labor/resource...Weaver FrameMaker , CM System Integration of CM tool L InConcert Cadre, AutoPlan, DBStar Ineffective process integration, poor training, time

  16. Independent backup mode transfer and mechanism for digital control computers

    NASA Technical Reports Server (NTRS)

    Tulpule, Bhalchandra R. (Inventor); Oscarson, Edward M. (Inventor)

    1992-01-01

    An interrupt is provided to a signal processor having a non-maskable interrupt input, in response to the detection of a request for transfer to backup software. The signal processor provides a transfer signal to a transfer mechanism only after completion of the present machine cycle. Transfer to the backup software is initiated by the transfer mechanism only upon reception of the transfer signal.

  17. Concurrent Image Processing Executive (CIPE)

    NASA Technical Reports Server (NTRS)

    Lee, Meemong; Cooper, Gregory T.; Groom, Steven L.; Mazer, Alan S.; Williams, Winifred I.

    1988-01-01

    The design and implementation of a Concurrent Image Processing Executive (CIPE), which is intended to become the support system software for a prototype high performance science analysis workstation are discussed. The target machine for this software is a JPL/Caltech Mark IIIfp Hypercube hosted by either a MASSCOMP 5600 or a Sun-3, Sun-4 workstation; however, the design will accommodate other concurrent machines of similar architecture, i.e., local memory, multiple-instruction-multiple-data (MIMD) machines. The CIPE system provides both a multimode user interface and an applications programmer interface, and has been designed around four loosely coupled modules; (1) user interface, (2) host-resident executive, (3) hypercube-resident executive, and (4) application functions. The loose coupling between modules allows modification of a particular module without significantly affecting the other modules in the system. In order to enhance hypercube memory utilization and to allow expansion of image processing capabilities, a specialized program management method, incremental loading, was devised. To minimize data transfer between host and hypercube a data management method which distributes, redistributes, and tracks data set information was implemented.

  18. Proceedings of the Second NASA Formal Methods Symposium

    NASA Technical Reports Server (NTRS)

    Munoz, Cesar (Editor)

    2010-01-01

    This publication contains the proceedings of the Second NASA Formal Methods Symposium sponsored by the National Aeronautics and Space Administration and held in Washington D.C. April 13-15, 2010. Topics covered include: Decision Engines for Software Analysis using Satisfiability Modulo Theories Solvers; Verification and Validation of Flight-Critical Systems; Formal Methods at Intel -- An Overview; Automatic Review of Abstract State Machines by Meta Property Verification; Hardware-independent Proofs of Numerical Programs; Slice-based Formal Specification Measures -- Mapping Coupling and Cohesion Measures to Formal Z; How Formal Methods Impels Discovery: A Short History of an Air Traffic Management Project; A Machine-Checked Proof of A State-Space Construction Algorithm; Automated Assume-Guarantee Reasoning for Omega-Regular Systems and Specifications; Modeling Regular Replacement for String Constraint Solving; Using Integer Clocks to Verify the Timing-Sync Sensor Network Protocol; Can Regulatory Bodies Expect Efficient Help from Formal Methods?; Synthesis of Greedy Algorithms Using Dominance Relations; A New Method for Incremental Testing of Finite State Machines; Verification of Faulty Message Passing Systems with Continuous State Space in PVS; Phase Two Feasibility Study for Software Safety Requirements Analysis Using Model Checking; A Prototype Embedding of Bluespec System Verilog in the PVS Theorem Prover; SimCheck: An Expressive Type System for Simulink; Coverage Metrics for Requirements-Based Testing: Evaluation of Effectiveness; Software Model Checking of ARINC-653 Flight Code with MCP; Evaluation of a Guideline by Formal Modelling of Cruise Control System in Event-B; Formal Verification of Large Software Systems; Symbolic Computation of Strongly Connected Components Using Saturation; Towards the Formal Verification of a Distributed Real-Time Automotive System; Slicing AADL Specifications for Model Checking; Model Checking with Edge-valued Decision Diagrams; and Data-flow based Model Analysis.

  19. Evaluation and selection of open-source EMR software packages based on integrated AHP and TOPSIS.

    PubMed

    Zaidan, A A; Zaidan, B B; Al-Haiqi, Ahmed; Kiah, M L M; Hussain, Muzammil; Abdulnabi, Mohamed

    2015-02-01

    Evaluating and selecting software packages that meet the requirements of an organization are difficult aspects of software engineering process. Selecting the wrong open-source EMR software package can be costly and may adversely affect business processes and functioning of the organization. This study aims to evaluate and select open-source EMR software packages based on multi-criteria decision-making. A hands-on study was performed and a set of open-source EMR software packages were implemented locally on separate virtual machines to examine the systems more closely. Several measures as evaluation basis were specified, and the systems were selected based a set of metric outcomes using Integrated Analytic Hierarchy Process (AHP) and TOPSIS. The experimental results showed that GNUmed and OpenEMR software can provide better basis on ranking score records than other open-source EMR software packages. Copyright © 2014 Elsevier Inc. All rights reserved.

  20. Celeris: A GPU-accelerated open source software with a Boussinesq-type wave solver for real-time interactive simulation and visualization

    NASA Astrophysics Data System (ADS)

    Tavakkol, Sasan; Lynett, Patrick

    2017-08-01

    In this paper, we introduce an interactive coastal wave simulation and visualization software, called Celeris. Celeris is an open source software which needs minimum preparation to run on a Windows machine. The software solves the extended Boussinesq equations using a hybrid finite volume-finite difference method and supports moving shoreline boundaries. The simulation and visualization are performed on the GPU using Direct3D libraries, which enables the software to run faster than real-time. Celeris provides a first-of-its-kind interactive modeling platform for coastal wave applications and it supports simultaneous visualization with both photorealistic and colormapped rendering capabilities. We validate our software through comparison with three standard benchmarks for non-breaking and breaking waves.

  1. User-driven integrated software lives: ``Paleomag'' paleomagnetics analysis on the Macintosh

    NASA Astrophysics Data System (ADS)

    Jones, Craig H.

    2002-12-01

    "PaleoMag," a paleomagnetics analysis package originally developed for the Macintosh operating system in 1988, allows examination of demagnetization of individual samples and analysis of directional data from collections of samples. Prior to recent reinvigorated development of the software for both Macintosh and Windows, it was widely used despite not running properly on machines and operating systems sold after 1995. This somewhat surprising situation demonstrates that there is a continued need for integrated analysis software within the earth sciences, in addition to well-developed scripting and batch-mode software. One distinct advantage of software like PaleoMag is in the ability to combine quality control with analysis within a unique graphical environment. Because such demands are frequent within the earth sciences, means of nurturing the development of similar software should be found.

  2. Software Support Measurement and Estimating for Oracle Database Applications Using Mark II Function Points

    DTIC Science & Technology

    1992-12-01

    36 V.33. Coe ncint of De minstioi ........................ 37 V3A. F-Raio .................................... 37 V3.5... de ations. Instructions ae defined as lines of code or card images. Thus, a line containin two or mome souce statements counts as one instruction; a...understand the productivity paradox, recall de concept of virtual machines. When a higher level machine groups ogether many instructm of a lower level

  3. Direct Digital Manufacturing of Integrated Naval Systems Using Ultrasonic Consolidation, Support Material Deposition and Direct Write Technologies

    DTIC Science & Technology

    2012-02-17

    tool should be combined with a user-friendly Windows-based software interface that utilizes the best practices for process planning developed by us and...best practices developed through this project, resulting in the commercial availability of machines for the Navy and others. These machines will...research 2011 Outstanding Paper Award, VRAP 2011, for paper "Some Studies on Dislocation Density based Finite Element Modeling of Ultrasonic

  4. The Burn Medical Assistant: Developing Machine Learning Algorithms to Aid in the Estimation of Burn Wound Size

    DTIC Science & Technology

    2017-10-01

    hypothesis that a computer machine learning algorithm can analyze and classify burn injures using multispectral imaging within 5% of an expert clinician...morbidity. In response to these challenges, the USAISR developed and obtained FDA 510(k) clearance of the Burn Navigator™, a computer decision support... computer decision support software (CDSS), can significantly change the CDSS algorithm’s recommendations and thus the total fluid administered to a

  5. Partners | Energy Systems Integration Facility | NREL

    Science.gov Websites

    LiquidCool Solutions Mercedes-Benz MidAmerican Energy Company N-Dimension Nanosonic Nebland Software OMNETRIC Group Pacific Gas and Electric Company Panasonic Parker Hannifin PDC Machines Pecan Street

  6. Optimization of a novel large field of view distortion phantom for MR-only treatment planning.

    PubMed

    Price, Ryan G; Knight, Robert A; Hwang, Ken-Pin; Bayram, Ersin; Nejad-Davarani, Siamak P; Glide-Hurst, Carri K

    2017-07-01

    MR-only treatment planning requires images of high geometric fidelity, particularly for large fields of view (FOV). However, the availability of large FOV distortion phantoms with analysis software is currently limited. This work sought to optimize a modular distortion phantom to accommodate multiple bore configurations and implement distortion characterization in a widely implementable solution. To determine candidate materials, 1.0 T MR and CT images were acquired of twelve urethane foam samples of various densities and strengths. Samples were precision-machined to accommodate 6 mm diameter paintballs used as landmarks. Final material candidates were selected by balancing strength, machinability, weight, and cost. Bore sizes and minimum aperture width resulting from couch position were tabulated from the literature (14 systems, 5 vendors). Bore geometry and couch position were simulated using MATLAB to generate machine-specific models to optimize the phantom build. Previously developed software for distortion characterization was modified for several magnet geometries (1.0 T, 1.5 T, 3.0 T), compared against previously published 1.0 T results, and integrated into the 3D Slicer application platform. All foam samples provided sufficient MR image contrast with paintball landmarks. Urethane foam (compressive strength ∼1000 psi, density ~20 lb/ft 3 ) was selected for its accurate machinability and weight characteristics. For smaller bores, a phantom version with the following parameters was used: 15 foam plates, 55 × 55 × 37.5 cm 3 (L×W×H), 5,082 landmarks, and weight ~30 kg. To accommodate > 70 cm wide bores, an extended build used 20 plates spanning 55 × 55 × 50 cm 3 with 7,497 landmarks and weight ~44 kg. Distortion characterization software was implemented as an external module into 3D Slicer's plugin framework and results agreed with the literature. The design and implementation of a modular, extendable distortion phantom was optimized for several bore configurations. The phantom and analysis software will be available for multi-institutional collaborations and cross-validation trials to support MR-only planning. © 2017 The Authors. Journal of Applied Clinical Medical Physics published by Wiley Periodicals, Inc. on behalf of American Association of Physicists in Medicine.

  7. Effects of virtualization on a scientific application - Running a hyperspectral radiative transfer code on virtual machines.

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Tikotekar, Anand A; Vallee, Geoffroy R; Naughton III, Thomas J

    2008-01-01

    The topic of system-level virtualization has recently begun to receive interest for high performance computing (HPC). This is in part due to the isolation and encapsulation offered by the virtual machine. These traits enable applications to customize their environments and maintain consistent software configurations in their virtual domains. Additionally, there are mechanisms that can be used for fault tolerance like live virtual machine migration. Given these attractive benefits to virtualization, a fundamental question arises, how does this effect my scientific application? We use this as the premise for our paper and observe a real-world scientific code running on a Xenmore » virtual machine. We studied the effects of running a radiative transfer simulation, Hydrolight, on a virtual machine. We discuss our methodology and report observations regarding the usage of virtualization with this application.« less

  8. Hardware support for software controlled fast multiplexing of performance counters

    DOEpatents

    Salapura, Valentina; Wisniewski, Robert W

    2013-10-01

    Performance counters may be operable to collect one or more counts of one or more selected activities, and registers may be operable to store a set of performance counter configurations. A state machine may be operable to automatically select a register from the registers for reconfiguring the one or more performance counters in response to receiving a first signal. The state machine may be further operable to reconfigure the one or more performance counters based on a configuration specified in the selected register. The state machine yet further may be operable to copy data in selected one or more of the performance counters to a memory location, or to copy data from the memory location to the counters, in response to receiving a second signal. The state machine may be operable to store or restore the counter values and state machine configuration in response to a context switch event.

  9. Hardware support for software controlled fast multiplexing of performance counters

    DOEpatents

    Salapura, Valentina; Wisniewski, Robert W.

    2013-01-01

    Performance counters may be operable to collect one or more counts of one or more selected activities, and registers may be operable to store a set of performance counter configurations. A state machine may be operable to automatically select a register from the registers for reconfiguring the one or more performance counters in response to receiving a first signal. The state machine may be further operable to reconfigure the one or more performance counters based on a configuration specified in the selected register. The state machine yet further may be operable to copy data in selected one or more of the performance counters to a memory location, or to copy data from the memory location to the counters, in response to receiving a second signal. The state machine may be operable to store or restore the counter values and state machine configuration in response to a context switch event.

  10. A comparative study of electrochemical machining process parameters by using GA and Taguchi method

    NASA Astrophysics Data System (ADS)

    Soni, S. K.; Thomas, B.

    2017-11-01

    In electrochemical machining quality of machined surface strongly depend on the selection of optimal parameter settings. This work deals with the application of Taguchi method and genetic algorithm using MATLAB to maximize the metal removal rate and minimize the surface roughness and overcut. In this paper a comparative study is presented for drilling of LM6 AL/B4C composites by comparing the significant impact of numerous machining process parameters such as, electrolyte concentration (g/l),machining voltage (v),frequency (hz) on the response parameters (surface roughness, material removal rate and over cut). Taguchi L27 orthogonal array was chosen in Minitab 17 software, for the investigation of experimental results and also multiobjective optimization done by genetic algorithm is employed by using MATLAB. After obtaining optimized results from Taguchi method and genetic algorithm, a comparative results are presented.

  11. Design and Fabrication of Automatic Glass Cutting Machine

    NASA Astrophysics Data System (ADS)

    Veena, T. R.; Kadadevaramath, R. S.; Nagaraj, P. M.; Madhusudhan, S. V.

    2016-09-01

    This paper deals with the design and fabrication of the automatic glass or mirror cutting machine. In order to increase the accuracy of cut and production rate; and decrease the production time and accidents caused due to manual cutting of mirror or glass, this project aims at development of an automatic machine which uses a programmable logic controller (PLC) for controlling the movement of the conveyer and also to control the pneumatic circuit. In this machine, the work of the operator is to load and unload the mirror. The cutter used in this machine is carbide wheel with its cutting edge ground to a V-shaped profile. The PLC controls the pneumatic cylinder and intern actuates the cutter along the glass, a fracture layer is formed causing a mark to be formed below the fracture layer and a crack to be formed below the rib mark. The machine elements are designed using CATIA V5R20 and pneumatic circuit are designed using FESTO FLUID SIM software.

  12. Machine learning and data science in soft materials engineering

    NASA Astrophysics Data System (ADS)

    Ferguson, Andrew L.

    2018-01-01

    In many branches of materials science it is now routine to generate data sets of such large size and dimensionality that conventional methods of analysis fail. Paradigms and tools from data science and machine learning can provide scalable approaches to identify and extract trends and patterns within voluminous data sets, perform guided traversals of high-dimensional phase spaces, and furnish data-driven strategies for inverse materials design. This topical review provides an accessible introduction to machine learning tools in the context of soft and biological materials by ‘de-jargonizing’ data science terminology, presenting a taxonomy of machine learning techniques, and surveying the mathematical underpinnings and software implementations of popular tools, including principal component analysis, independent component analysis, diffusion maps, support vector machines, and relative entropy. We present illustrative examples of machine learning applications in soft matter, including inverse design of self-assembling materials, nonlinear learning of protein folding landscapes, high-throughput antimicrobial peptide design, and data-driven materials design engines. We close with an outlook on the challenges and opportunities for the field.

  13. Machine learning and data science in soft materials engineering.

    PubMed

    Ferguson, Andrew L

    2018-01-31

    In many branches of materials science it is now routine to generate data sets of such large size and dimensionality that conventional methods of analysis fail. Paradigms and tools from data science and machine learning can provide scalable approaches to identify and extract trends and patterns within voluminous data sets, perform guided traversals of high-dimensional phase spaces, and furnish data-driven strategies for inverse materials design. This topical review provides an accessible introduction to machine learning tools in the context of soft and biological materials by 'de-jargonizing' data science terminology, presenting a taxonomy of machine learning techniques, and surveying the mathematical underpinnings and software implementations of popular tools, including principal component analysis, independent component analysis, diffusion maps, support vector machines, and relative entropy. We present illustrative examples of machine learning applications in soft matter, including inverse design of self-assembling materials, nonlinear learning of protein folding landscapes, high-throughput antimicrobial peptide design, and data-driven materials design engines. We close with an outlook on the challenges and opportunities for the field.

  14. Parallel processors and nonlinear structural dynamics algorithms and software

    NASA Technical Reports Server (NTRS)

    Belytschko, Ted; Gilbertsen, Noreen D.; Neal, Mark O.; Plaskacz, Edward J.

    1989-01-01

    The adaptation of a finite element program with explicit time integration to a massively parallel SIMD (single instruction multiple data) computer, the CONNECTION Machine is described. The adaptation required the development of a new algorithm, called the exchange algorithm, in which all nodal variables are allocated to the element with an exchange of nodal forces at each time step. The architectural and C* programming language features of the CONNECTION Machine are also summarized. Various alternate data structures and associated algorithms for nonlinear finite element analysis are discussed and compared. Results are presented which demonstrate that the CONNECTION Machine is capable of outperforming the CRAY XMP/14.

  15. Personal manufacturing systems

    NASA Astrophysics Data System (ADS)

    Bailey, P.

    1992-04-01

    Personal Manufacturing Systems are the missing link in the automation of the design-to- manufacture process. A PMS will act as a CAD peripheral, closing the loop around the designer enabling him to directly produce models, short production runs or soft tooling with as little fuss as he might otherwise plot a drawing. Whereas conventional 5-axis CNC machines are based on orthogonal axes and simple incremental movements, the PMS is based on a geodetic structure and complex co-ordinated 'spline' movements. The software employs a novel 3D pixel technique for give itself 'spatial awareness' and an expert system to determine the optimum machining conditions. A completely automatic machining strategy can then be determined.

  16. The evolution and practical application of machine translation system (1)

    NASA Astrophysics Data System (ADS)

    Tominaga, Isao; Sato, Masayuki

    This paper describes a development, practical applicatioin, problem of a system, evaluation of practical system, and development trend of machine translation. Most recent system contains next four problems. 1) the vagueness of a text, 2) a difference of the definition of the terminology between different language, 3) the preparing of a large-scale translation dictionary, 4) the development of a software for the logical inference. Machine translation system is already used practically in many industry fields. However, many problems are not solved. The implementation of an ideal system will be after 15 years. Also, this paper described seven evaluation items detailedly. This English abstract was made by Mu system.

  17. Ring Image Analyzer

    NASA Technical Reports Server (NTRS)

    Strekalov, Dmitry V.

    2012-01-01

    Ring Image Analyzer software analyzes images to recognize elliptical patterns. It determines the ellipse parameters (axes ratio, centroid coordinate, tilt angle). The program attempts to recognize elliptical fringes (e.g., Newton Rings) on a photograph and determine their centroid position, the short-to-long-axis ratio, and the angle of rotation of the long axis relative to the horizontal direction on the photograph. These capabilities are important in interferometric imaging and control of surfaces. In particular, this program has been developed and applied for determining the rim shape of precision-machined optical whispering gallery mode resonators. The program relies on a unique image recognition algorithm aimed at recognizing elliptical shapes, but can be easily adapted to other geometric shapes. It is robust against non-elliptical details of the image and against noise. Interferometric analysis of precision-machined surfaces remains an important technological instrument in hardware development and quality analysis. This software automates and increases the accuracy of this technique. The software has been developed for the needs of an R&TD-funded project and has become an important asset for the future research proposal to NASA as well as other agencies.

  18. Man-machine Integration Design and Analysis System (MIDAS) Task Loading Model (TLM) experimental and software detailed design report

    NASA Technical Reports Server (NTRS)

    Staveland, Lowell

    1994-01-01

    This is the experimental and software detailed design report for the prototype task loading model (TLM) developed as part of the man-machine integration design and analysis system (MIDAS), as implemented and tested in phase 6 of the Army-NASA Aircrew/Aircraft Integration (A3I) Program. The A3I program is an exploratory development effort to advance the capabilities and use of computational representations of human performance and behavior in the design, synthesis, and analysis of manned systems. The MIDAS TLM computationally models the demands designs impose on operators to aide engineers in the conceptual design of aircraft crewstations. This report describes TLM and the results of a series of experiments which were run this phase to test its capabilities as a predictive task demand modeling tool. Specifically, it includes discussions of: the inputs and outputs of TLM, the theories underlying it, the results of the test experiments, the use of the TLM as both stand alone tool and part of a complete human operator simulation, and a brief introduction to the TLM software design.

  19. CernVM WebAPI - Controlling Virtual Machines from the Web

    NASA Astrophysics Data System (ADS)

    Charalampidis, I.; Berzano, D.; Blomer, J.; Buncic, P.; Ganis, G.; Meusel, R.; Segal, B.

    2015-12-01

    Lately, there is a trend in scientific projects to look for computing resources in the volunteering community. In addition, to reduce the development effort required to port the scientific software stack to all the known platforms, the use of Virtual Machines (VMs)u is becoming increasingly popular. Unfortunately their use further complicates the software installation and operation, restricting the volunteer audience to sufficiently expert people. CernVM WebAPI is a software solution addressing this specific case in a way that opens wide new application opportunities. It offers a very simple API for setting-up, controlling and interfacing with a VM instance in the users computer, while in the same time offloading the user from all the burden of downloading, installing and configuring the hypervisor. WebAPI comes with a lightweight javascript library that guides the user through the application installation process. Malicious usage is prohibited by offering a per-domain PKI validation mechanism. In this contribution we will overview this new technology, discuss its security features and examine some test cases where it is already in use.

  20. Proteus: a reconfigurable computational network for computer vision

    NASA Astrophysics Data System (ADS)

    Haralick, Robert M.; Somani, Arun K.; Wittenbrink, Craig M.; Johnson, Robert; Cooper, Kenneth; Shapiro, Linda G.; Phillips, Ihsin T.; Hwang, Jenq N.; Cheung, William; Yao, Yung H.; Chen, Chung-Ho; Yang, Larry; Daugherty, Brian; Lorbeski, Bob; Loving, Kent; Miller, Tom; Parkins, Larye; Soos, Steven L.

    1992-04-01

    The Proteus architecture is a highly parallel MIMD, multiple instruction, multiple-data machine, optimized for large granularity tasks such as machine vision and image processing The system can achieve 20 Giga-flops (80 Giga-flops peak). It accepts data via multiple serial links at a rate of up to 640 megabytes/second. The system employs a hierarchical reconfigurable interconnection network with the highest level being a circuit switched Enhanced Hypercube serial interconnection network for internal data transfers. The system is designed to use 256 to 1,024 RISC processors. The processors use one megabyte external Read/Write Allocating Caches for reduced multiprocessor contention. The system detects, locates, and replaces faulty subsystems using redundant hardware to facilitate fault tolerance. The parallelism is directly controllable through an advanced software system for partitioning, scheduling, and development. System software includes a translator for the INSIGHT language, a parallel debugger, low and high level simulators, and a message passing system for all control needs. Image processing application software includes a variety of point operators neighborhood, operators, convolution, and the mathematical morphology operations of binary and gray scale dilation, erosion, opening, and closing.

  1. Software and Human-Machine Interface Development for Environmental Controls Subsystem Support

    NASA Technical Reports Server (NTRS)

    Dobson, Matthew

    2018-01-01

    The Space Launch System (SLS) is the next premier launch vehicle for NASA. It is the next stage of manned space exploration from American soil, and will be the platform in which we push further beyond Earth orbit. In preparation of the SLS maiden voyage on Exploration Mission 1 (EM-1), the existing ground support architecture at Kennedy Space Center required significant overhaul and updating. A comprehensive upgrade of controls systems was necessary, including programmable logic controller software, as well as Launch Control Center (LCC) firing room and local launch pad displays for technician use. Environmental control acts as an integral component in these systems, being the foremost system for conditioning the pad and extremely sensitive launch vehicle until T-0. The Environmental Controls Subsystem (ECS) required testing and modification to meet the requirements of the designed system, as well as the human factors requirements of NASA software for Validation and Verification (V&V). This term saw significant strides in the progress and functionality of the human-machine interfaces used at the launch pad, and improved integration with the controller code.

  2. Integrated computer-aided design using minicomputers

    NASA Technical Reports Server (NTRS)

    Storaasli, O. O.

    1980-01-01

    Computer-Aided Design/Computer-Aided Manufacturing (CAD/CAM), a highly interactive software, has been implemented on minicomputers at the NASA Langley Research Center. CAD/CAM software integrates many formerly fragmented programs and procedures into one cohesive system; it also includes finite element modeling and analysis, and has been interfaced via a computer network to a relational data base management system and offline plotting devices on mainframe computers. The CAD/CAM software system requires interactive graphics terminals operating at a minimum of 4800 bits/sec transfer rate to a computer. The system is portable and introduces 'interactive graphics', which permits the creation and modification of models interactively. The CAD/CAM system has already produced designs for a large area space platform, a national transonic facility fan blade, and a laminar flow control wind tunnel model. Besides the design/drafting element analysis capability, CAD/CAM provides options to produce an automatic program tooling code to drive a numerically controlled (N/C) machine. Reductions in time for design, engineering, drawing, finite element modeling, and N/C machining will benefit productivity through reduced costs, fewer errors, and a wider range of configuration.

  3. How Does One Manage ’Information? Making Sense of the Information Being Received

    DTIC Science & Technology

    2012-12-01

    to manage. (Photo by PFC Franklin E. Mercado .) Report Documentation Page Form ApprovedOMB No. 0704-0188 Public reporting burden for the collection of...in choosing the right application. Appli- cation software is written to perform a specific task or function, and it becomes increasingly difficult...common data, virtu- alizing machines for all software (using one computer/server, but dividing it into logical segments), and standardizing

  4. Source Code Analysis Laboratory (SCALe)

    DTIC Science & Technology

    2012-04-01

    Versus Flagged Nonconformities (FNC) Software System TP/FNC Ratio Mozilla Firefox version 2.0 6/12 50% Linux kernel version 2.6.15 10/126 8...is inappropriately tuned for analysis of the Linux kernel, which has anomalous results. Customizing SCALe to work with software for a particular...servers support a collection of virtual machines (VMs) that can be configured to support analysis in various environments, such as Windows XP and Linux . A

  5. Knowledge-Based Software Development Tools

    DTIC Science & Technology

    1993-09-01

    GREEN, C., AND WESTFOLD, S. Knowledge-based programming self-applied. In Machine Intelligence 10, J. E. Hayes, D. Mitchie, and Y. Pao, Eds., Wiley...Technical Report KES.U.84.2, Kestrel Institute, April 1984. [181 KORF, R. E. Toward a model of representation changes. Artificial Intelligence 14, 1...Artificial Intelligence 27, 1 (February 1985), 43-96. Replinted in Readings in Artificial Intelligence and Software Engineering, C. Rich •ad R. Waters

  6. Verification of operation of the actuator control system using the integration the B&R Automation Studio software with a virtual model of the actuator system

    NASA Astrophysics Data System (ADS)

    Herbuś, K.; Ociepka, P.

    2017-08-01

    In the work is analysed a sequential control system of a machine for separating and grouping work pieces for processing. Whereas, the area of the considered problem is related with verification of operation of an actuator system of an electro-pneumatic control system equipped with a PLC controller. Wherein to verification is subjected the way of operation of actuators in view of logic relationships assumed in the control system. The actuators of the considered control system were three drives of linear motion (pneumatic cylinders). And the logical structure of the system of operation of the control system is based on the signals flow graph. The tested logical structure of operation of the electro-pneumatic control system was implemented in the Automation Studio software of B&R company. This software is used to create programs for the PLC controllers. Next, in the FluidSIM software was created the model of the actuator system of the control system of a machine. To verify the created program for the PLC controller, simulating the operation of the created model, it was utilized the approach of integration these two programs using the tool for data exchange in the form of the OPC server.

  7. Automated identification of retained surgical items in radiological images

    NASA Astrophysics Data System (ADS)

    Agam, Gady; Gan, Lin; Moric, Mario; Gluncic, Vicko

    2015-03-01

    Retained surgical items (RSIs) in patients is a major operating room (OR) patient safety concern. An RSI is any surgical tool, sponge, needle or other item inadvertently left in a patients body during the course of surgery. If left undetected, RSIs may lead to serious negative health consequences such as sepsis, internal bleeding, and even death. To help physicians efficiently and effectively detect RSIs, we are developing computer-aided detection (CADe) software for X-ray (XR) image analysis, utilizing large amounts of currently available image data to produce a clinically effective RSI detection system. Physician analysis of XRs for the purpose of RSI detection is a relatively lengthy process that may take up to 45 minutes to complete. It is also error prone due to the relatively low acuity of the human eye for RSIs in XR images. The system we are developing is based on computer vision and machine learning algorithms. We address the problem of low incidence by proposing synthesis algorithms. The CADe software we are developing may be integrated into a picture archiving and communication system (PACS), be implemented as a stand-alone software application, or be integrated into portable XR machine software through application programming interfaces. Preliminary experimental results on actual XR images demonstrate the effectiveness of the proposed approach.

  8. 3D OCT imaging in clinical settings: toward quantitative measurements of retinal structures

    NASA Astrophysics Data System (ADS)

    Zawadzki, Robert J.; Fuller, Alfred R.; Zhao, Mingtao; Wiley, David F.; Choi, Stacey S.; Bower, Bradley A.; Hamann, Bernd; Izatt, Joseph A.; Werner, John S.

    2006-02-01

    The acquisition speed of current FD-OCT (Fourier Domain - Optical Coherence Tomography) instruments allows rapid screening of three-dimensional (3D) volumes of human retinas in clinical settings. To take advantage of this ability requires software used by physicians to be capable of displaying and accessing volumetric data as well as supporting post processing in order to access important quantitative information such as thickness maps and segmented volumes. We describe our clinical FD-OCT system used to acquire 3D data from the human retina over the macula and optic nerve head. B-scans are registered to remove motion artifacts and post-processed with customized 3D visualization and analysis software. Our analysis software includes standard 3D visualization techniques along with a machine learning support vector machine (SVM) algorithm that allows a user to semi-automatically segment different retinal structures and layers. Our program makes possible measurements of the retinal layer thickness as well as volumes of structures of interest, despite the presence of noise and structural deformations associated with retinal pathology. Our software has been tested successfully in clinical settings for its efficacy in assessing 3D retinal structures in healthy as well as diseased cases. Our tool facilitates diagnosis and treatment monitoring of retinal diseases.

  9. LHCb experience with running jobs in virtual machines

    NASA Astrophysics Data System (ADS)

    McNab, A.; Stagni, F.; Luzzi, C.

    2015-12-01

    The LHCb experiment has been running production jobs in virtual machines since 2013 as part of its DIRAC-based infrastructure. We describe the architecture of these virtual machines and the steps taken to replicate the WLCG worker node environment expected by user and production jobs. This relies on the uCernVM system for providing root images for virtual machines. We use the CernVM-FS distributed filesystem to supply the root partition files, the LHCb software stack, and the bootstrapping scripts necessary to configure the virtual machines for us. Using this approach, we have been able to minimise the amount of contextualisation which must be provided by the virtual machine managers. We explain the process by which the virtual machine is able to receive payload jobs submitted to DIRAC by users and production managers, and how this differs from payloads executed within conventional DIRAC pilot jobs on batch queue based sites. We describe our operational experiences in running production on VM based sites managed using Vcycle/OpenStack, Vac, and HTCondor Vacuum. Finally we show how our use of these resources is monitored using Ganglia and DIRAC.

  10. Gram staining with an automatic machine.

    PubMed

    Felek, S; Arslan, A

    1999-01-01

    This study was undertaken to develop a new Gram-staining machine controlled by a micro-controller and to investigate the quality of slides that were stained in the machine. The machine was designed and produced by the authors. It uses standard 220 V AC. Staining, washing, and drying periods are controlled by a timer built in the micro-controller. A software was made that contains a certain algorithm and time intervals for the staining mode. One-hundred and forty smears were prepared from Escherichia coli, Staphylococcus aureus, Neisseria sp., blood culture, trypticase soy broth, direct pus and sputum smears for comparison studies. Half of the slides in each group were stained with the machine, the other half by hand and then examined by four different microbiologists. Machine-stained slides had a higher clarity and less debris than the hand-stained slides (p < 0.05). In hand-stained slides, some Gram-positive organisms showed poor Gram-positive staining features (p < 0.05). In conclusion, we suggest that Gram staining with the automatic machine increases the staining quality and helps to decrease the work load in a busy diagnostic laboratory.

  11. Intelligent Systems and Advanced User Interfaces for Design, Operation, and Maintenance of Command Management Systems

    NASA Technical Reports Server (NTRS)

    Mitchell, Christine M.

    1998-01-01

    Historically Command Management Systems (CMS) have been large, expensive, spacecraft-specific software systems that were costly to build, operate, and maintain. Current and emerging hardware, software, and user interface technologies may offer an opportunity to facilitate the initial formulation and design of a spacecraft-specific CMS as well as a to develop a more generic or a set of core components for CMS systems. Current MOC (mission operations center) hardware and software include Unix workstations, the C/C++ and Java programming languages, and X and Java window interfaces representations. This configuration provides the power and flexibility to support sophisticated systems and intelligent user interfaces that exploit state-of-the-art technologies in human-machine systems engineering, decision making, artificial intelligence, and software engineering. One of the goals of this research is to explore the extent to which technologies developed in the research laboratory can be productively applied in a complex system such as spacecraft command management. Initial examination of some of the issues in CMS design and operation suggests that application of technologies such as intelligent planning, case-based reasoning, design and analysis tools from a human-machine systems engineering point of view (e.g., operator and designer models) and human-computer interaction tools, (e.g., graphics, visualization, and animation), may provide significant savings in the design, operation, and maintenance of a spacecraft-specific CMS as well as continuity for CMS design and development across spacecraft with varying needs. The savings in this case is in software reuse at all stages of the software engineering process.

  12. Present and Future of M2M

    NASA Astrophysics Data System (ADS)

    Ono, Satoru; Watanabe, Takashi

    In recent years, the rapid progress in the development of hardware and software technologies enables tiny and low cost information devices hereinafter referred to as Machine to be widely available. M2M (Machine to Machine) has been of much attention where many tiny machines are connected to each other through networks with minimal human intervention to provide smooth and intelligent management. M2M is a promising core technology providing timely, flexible, efficient and comprehensive service at low cost. M2M has wide variety of applications including energy management system, environmental monitoring system, intelligent transport system, industrial automation system and other applications. M2M consists of terminals and networks that connect them. In this paper, we mainly focus on M2M networking and mention the future direction of the technology.

  13. The Machine / Job Features Mechanism

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Alef, M.; Cass, T.; Keijser, J. J.

    Within the HEPiX virtualization group and the Worldwide LHC Computing Grid’s Machine/Job Features Task Force, a mechanism has been developed which provides access to detailed information about the current host and the current job to the job itself. This allows user payloads to access meta information, independent of the current batch system or virtual machine model. The information can be accessed either locally via the filesystem on a worker node, or remotely via HTTP(S) from a webserver. This paper describes the final version of the specification from 2016 which was published as an HEP Software Foundation technical note, and themore » design of the implementations of this version for batch and virtual machine platforms. We discuss early experiences with these implementations and how they can be exploited by experiment frameworks.« less

  14. The machine/job features mechanism

    NASA Astrophysics Data System (ADS)

    Alef, M.; Cass, T.; Keijser, J. J.; McNab, A.; Roiser, S.; Schwickerath, U.; Sfiligoi, I.

    2017-10-01

    Within the HEPiX virtualization group and the Worldwide LHC Computing Grid’s Machine/Job Features Task Force, a mechanism has been developed which provides access to detailed information about the current host and the current job to the job itself. This allows user payloads to access meta information, independent of the current batch system or virtual machine model. The information can be accessed either locally via the filesystem on a worker node, or remotely via HTTP(S) from a webserver. This paper describes the final version of the specification from 2016 which was published as an HEP Software Foundation technical note, and the design of the implementations of this version for batch and virtual machine platforms. We discuss early experiences with these implementations and how they can be exploited by experiment frameworks.

  15. Quality-by-Design (QbD): An integrated process analytical technology (PAT) approach for a dynamic pharmaceutical co-precipitation process characterization and process design space development.

    PubMed

    Wu, Huiquan; White, Maury; Khan, Mansoor A

    2011-02-28

    The aim of this work was to develop an integrated process analytical technology (PAT) approach for a dynamic pharmaceutical co-precipitation process characterization and design space development. A dynamic co-precipitation process by gradually introducing water to the ternary system of naproxen-Eudragit L100-alcohol was monitored at real-time in situ via Lasentec FBRM and PVM. 3D map of count-time-chord length revealed three distinguishable process stages: incubation, transition, and steady-state. The effects of high risk process variables (slurry temperature, stirring rate, and water addition rate) on both derived co-precipitation process rates and final chord-length-distribution were evaluated systematically using a 3(3) full factorial design. Critical process variables were identified via ANOVA for both transition and steady state. General linear models (GLM) were then used for parameter estimation for each critical variable. Clear trends about effects of each critical variable during transition and steady state were found by GLM and were interpreted using fundamental process principles and Nyvlt's transfer model. Neural network models were able to link process variables with response variables at transition and steady state with R(2) of 0.88-0.98. PVM images evidenced nucleation and crystal growth. Contour plots illustrated design space via critical process variables' ranges. It demonstrated the utility of integrated PAT approach for QbD development. Published by Elsevier B.V.

  16. Selfconsistent vibrational and free electron kinetics for CO2 dissociation in cold plasmas

    NASA Astrophysics Data System (ADS)

    Capitelli, Mario

    2016-09-01

    The activation of CO2 by cold plasmas is receiving new theoretical interest thanks to two European groups. The Bogaerts group developed a global model for the activation of CO2 trying to reproduce the experimental values for DBD and microwave discharges. The approach of Pietanza et al was devoted to understand the dependence of electron energy distribution function (eedf) of pure CO2 on the presence of concentrations of electronically and vibrationally excited states taken as parameter. To understand the importance of the vibrational excitation in the dissociation process Pietanza et al compared an upper limit to the dissociation process from a pure vibrational mechanism (PVM) with the corresponding electron impact dissociation rate, the prevalence of the two models depending on the reduced electric field and on the choice of the electron molecule cross section database. Improvement of the Pietanza et al model is being considered by coupling the time dependent Boltzmann solver with the non equilibrium vibrational kinetics of asymmetric mode and with simplified plasma chemistry kinetics describing the ionization/recombination process and the excitation-deexcitation of a metastable level at 10.5eV. A new PVM mechanism is also considered. Preliminary results, for both discharge and post discharge conditions, emphasize the action of superelastic collisions involving both vibrationally and electronically excited states in affecting the eedf. The new results can be used to plan a road map for future developments of numerical codes for rationalizing existing experimental values, as well as, for indicating new experimental situations.

  17. Service learning: Priority 4 Paws mobile surgical service for shelter animals.

    PubMed

    Freeman, Lynetta J; Ferguson, Nancy; Litster, Annette; Arighi, Mimi

    2013-01-01

    The increasing attention given to competencies needed to enter the workforce has revealed a need for veterinary students to gain more experience in performing small-animal elective surgery before graduation. In addition, guidelines for standards of care for shelter animals recommend that all dogs and cats should be spayed or neutered before adoption. Teaching surgical skills while serving the needs of local animal shelters represents an ideal service-learning opportunity. Following a pilot study and the benchmarking of other programs, an elective course in shelter medicine and surgery was created at Purdue University College of Veterinary Medicine (PVM) to allow senior DVM students an opportunity to spend 2 weeks on a mobile surgery unit (Priority 4 Paws) and 1 week at an animal shelter. With financial assistance from sponsors and donors, PVM purchased and equipped a mobile surgery unit, hired a full-time veterinarian and a registered veterinary technician, and established relationships with 12 animal shelters. From July 30, 2012, to March 22, 2013, 1,941 spays and neuters were performed with excellent postsurgical outcomes while training 33 veterinary students on rotation and 26 veterinary technician students. The program was well accepted by both students and the shelters being served. The Priority 4 Paws program is an example of an integrated, community-based service-learning opportunity that not only helps to improve the surgical skills of veterinary students but also helps to meet an identified community need.

  18. Plasmid mediated antimicrobial resistance in Ontario isolates of Actinobacillus (Haemophilus) pleuropneumoniae.

    PubMed Central

    Gilbride, K A; Rosendal, S; Brunton, J L

    1989-01-01

    The genetic basis of antimicrobial resistance in Ontario isolates of Actinobacillus (Haemophilus) pleuropneumoniae was studied. Two Ontario isolates of A. pleuropneumoniae were found to be resistant to sulfonamides (Su), streptomycin (Sm) and ampicillin (Amp). Resistance to Su and Sm was specified by a 2.3 megadalton (Mdal) plasmid which appeared to be identical to pVM104, which has been described in isolates of A. pleuropneumoniae from South Dakota. Southern hybridization showed that the 2.3 Mdal Su Sm plasmid was highly related to those Hinc II fragments of RSF1010 known to carry the Su Sm genes, but was unrelated to the remainder of this Salmonella resistance plasmid. Resistance to Su and Amp was specified by a 3.5 Mdal plasmid and appeared identical to pVM105 previously reported. The beta-lactamase enzyme had an isoelectric point of approximately 9.0. Southern hybridization showed no relationship to the TEM beta-lactamase. A third isolate of A. pleuropneumoniae was found to be resistant to chloramphenicol (Cm), Su and Sm by virtue of a 3.0 Mdal plasmid which specified a chloramphenicol acetyl transferase. We conclude that resistance to Su, Sm, Amp and Cm is mediated by small plasmids in A. pleuropneumoniae. Although the Su and Sm resistance determinants are highly related to those found in Enterobacteriaceae, the plasmids themselves and the beta-lactamase determinant are different. Images Fig. 1. Fig. 2. Fig. 3. Fig. 4. PMID:2914226

  19. Natural Language Processing.

    ERIC Educational Resources Information Center

    Chowdhury, Gobinda G.

    2003-01-01

    Discusses issues related to natural language processing, including theoretical developments; natural language understanding; tools and techniques; natural language text processing systems; abstracting; information extraction; information retrieval; interfaces; software; Internet, Web, and digital library applications; machine translation for…

  20. Model development, testing and experimentation in a CyberWorkstation for Brain-Machine Interface research.

    PubMed

    Rattanatamrong, Prapaporn; Matsunaga, Andrea; Raiturkar, Pooja; Mesa, Diego; Zhao, Ming; Mahmoudi, Babak; Digiovanna, Jack; Principe, Jose; Figueiredo, Renato; Sanchez, Justin; Fortes, Jose

    2010-01-01

    The CyberWorkstation (CW) is an advanced cyber-infrastructure for Brain-Machine Interface (BMI) research. It allows the development, configuration and execution of BMI computational models using high-performance computing resources. The CW's concept is implemented using a software structure in which an "experiment engine" is used to coordinate all software modules needed to capture, communicate and process brain signals and motor-control commands. A generic BMI-model template, which specifies a common interface to the CW's experiment engine, and a common communication protocol enable easy addition, removal or replacement of models without disrupting system operation. This paper reviews the essential components of the CW and shows how templates can facilitate the processes of BMI model development, testing and incorporation into the CW. It also discusses the ongoing work towards making this process infrastructure independent.

  1. 3D Printed, Customized Cranial Implant for Surgical Planning

    NASA Astrophysics Data System (ADS)

    Bogu, Venkata Phanindra; Ravi Kumar, Yennam; Asit Kumar, Khanra

    2018-06-01

    The main objective of the present work is to model cranial implant and printed in FDM machine (printer model used: mojo). Actually this is peculiar case and the skull has been damaged in frontal, parietal and temporal regions and a small portion of frontal region damaged away from saggital plane, complexity is to fill this frontal region with proper curvature. The Patient CT-data (Number of slices was 381 and thickness of each slice is 0.488 mm) was processed in mimics14.1 software, mimics file was sent to 3-matic software and calculated thickness of skull at different sections where cranial implant is needed then corrected the edges of cranial implant to overcome CSF (cerebrospinal fluid) leakage and proper fitting. Finally the implant average thickness is decided as 2.5 mm and printed in FDM machine with ABS plastic.

  2. 3D Printed, Customized Cranial Implant for Surgical Planning

    NASA Astrophysics Data System (ADS)

    Bogu, Venkata Phanindra; Ravi Kumar, Yennam; Asit Kumar, Khanra

    2016-06-01

    The main objective of the present work is to model cranial implant and printed in FDM machine (printer model used: mojo). Actually this is peculiar case and the skull has been damaged in frontal, parietal and temporal regions and a small portion of frontal region damaged away from saggital plane, complexity is to fill this frontal region with proper curvature. The Patient CT-data (Number of slices was 381 and thickness of each slice is 0.488 mm) was processed in mimics14.1 software, mimics file was sent to 3-matic software and calculated thickness of skull at different sections where cranial implant is needed then corrected the edges of cranial implant to overcome CSF (cerebrospinal fluid) leakage and proper fitting. Finally the implant average thickness is decided as 2.5 mm and printed in FDM machine with ABS plastic.

  3. Development of a State Machine Sequencer for the Keck Interferometer: Evolution, Development and Lessons Learned using a CASE Tool Approach

    NASA Technical Reports Server (NTRS)

    Rede, Leonard J.; Booth, Andrew; Hsieh, Jonathon; Summer, Kellee

    2004-01-01

    This paper presents a discussion of the evolution of a sequencer from a simple EPICS (Experimental Physics and Industrial Control System) based sequencer into a complex implementation designed utilizing UML (Unified Modeling Language) methodologies and a CASE (Computer Aided Software Engineering) tool approach. The main purpose of the sequencer (called the IF Sequencer) is to provide overall control of the Keck Interferometer to enable science operations be carried out by a single operator (and/or observer). The interferometer links the two 10m telescopes of the W. M. Keck Observatory at Mauna Kea, Hawaii. The IF Sequencer is a high-level, multi-threaded, Hare1 finite state machine, software program designed to orchestrate several lower-level hardware and software hard real time subsystems that must perform their work in a specific and sequential order. The sequencing need not be done in hard real-time. Each state machine thread commands either a high-speed real-time multiple mode embedded controller via CORB A, or slower controllers via EPICS Channel Access interfaces. The overall operation of the system is simplified by the automation. The UML is discussed and our use of it to implement the sequencer is presented. The decision to use the Rhapsody product as our CASE tool is explained and reflected upon. Most importantly, a section on lessons learned is presented and the difficulty of integrating CASE tool automatically generated C++ code into a large control system consisting of multiple infrastructures is presented.

  4. SU-G-BRB-02: An Open-Source Software Analysis Library for Linear Accelerator Quality Assurance

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Kerns, J; Yaldo, D

    Purpose: Routine linac quality assurance (QA) tests have become complex enough to require automation of most test analyses. A new data analysis software library was built that allows physicists to automate routine linear accelerator quality assurance tests. The package is open source, code tested, and benchmarked. Methods: Images and data were generated on a TrueBeam linac for the following routine QA tests: VMAT, starshot, CBCT, machine logs, Winston Lutz, and picket fence. The analysis library was built using the general programming language Python. Each test was analyzed with the library algorithms and compared to manual measurements taken at the timemore » of acquisition. Results: VMAT QA results agreed within 0.1% between the library and manual measurements. Machine logs (dynalogs & trajectory logs) were successfully parsed; mechanical axis positions were verified for accuracy and MLC fluence agreed well with EPID measurements. CBCT QA measurements were within 10 HU and 0.2mm where applicable. Winston Lutz isocenter size measurements were within 0.2mm of TrueBeam’s Machine Performance Check. Starshot analysis was within 0.2mm of the Winston Lutz results for the same conditions. Picket fence images with and without a known error showed that the library was capable of detecting MLC offsets within 0.02mm. Conclusion: A new routine QA software library has been benchmarked and is available for use by the community. The library is open-source and extensible for use in larger systems.« less

  5. Development of a state machine sequencer for the Keck Interferometer: evolution, development, and lessons learned using a CASE tool approach

    NASA Astrophysics Data System (ADS)

    Reder, Leonard J.; Booth, Andrew; Hsieh, Jonathan; Summers, Kellee R.

    2004-09-01

    This paper presents a discussion of the evolution of a sequencer from a simple Experimental Physics and Industrial Control System (EPICS) based sequencer into a complex implementation designed utilizing UML (Unified Modeling Language) methodologies and a Computer Aided Software Engineering (CASE) tool approach. The main purpose of the Interferometer Sequencer (called the IF Sequencer) is to provide overall control of the Keck Interferometer to enable science operations to be carried out by a single operator (and/or observer). The interferometer links the two 10m telescopes of the W. M. Keck Observatory at Mauna Kea, Hawaii. The IF Sequencer is a high-level, multi-threaded, Harel finite state machine software program designed to orchestrate several lower-level hardware and software hard real-time subsystems that must perform their work in a specific and sequential order. The sequencing need not be done in hard real-time. Each state machine thread commands either a high-speed real-time multiple mode embedded controller via CORBA, or slower controllers via EPICS Channel Access interfaces. The overall operation of the system is simplified by the automation. The UML is discussed and our use of it to implement the sequencer is presented. The decision to use the Rhapsody product as our CASE tool is explained and reflected upon. Most importantly, a section on lessons learned is presented and the difficulty of integrating CASE tool automatically generated C++ code into a large control system consisting of multiple infrastructures is presented.

  6. Programmable and autonomous computing machine made of biomolecules

    PubMed Central

    Benenson, Yaakov; Paz-Elizur, Tamar; Adar, Rivka; Keinan, Ehud; Livneh, Zvi; Shapiro, Ehud

    2013-01-01

    Devices that convert information from one form into another according to a definite procedure are known as automata. One such hypothetical device is the universal Turing machine1, which stimulated work leading to the development of modern computers. The Turing machine and its special cases2, including finite automata3, operate by scanning a data tape, whose striking analogy to information-encoding biopolymers inspired several designs for molecular DNA computers4–8. Laboratory-scale computing using DNA and human-assisted protocols has been demonstrated9–15, but the realization of computing devices operating autonomously on the molecular scale remains rare16–20. Here we describe a programmable finite automaton comprising DNA and DNA-manipulating enzymes that solves computational problems autonomously. The automaton’s hardware consists of a restriction nuclease and ligase, the software and input are encoded by double-stranded DNA, and programming amounts to choosing appropriate software molecules. Upon mixing solutions containing these components, the automaton processes the input molecule via a cascade of restriction, hybridization and ligation cycles, producing a detectable output molecule that encodes the automaton’s final state, and thus the computational result. In our implementation 1012 automata sharing the same software run independently and in parallel on inputs (which could, in principle, be distinct) in 120 μl solution at room temperature at a combined rate of 109 transitions per second with a transition fidelity greater than 99.8%, consuming less than 10−10 W. PMID:11719800

  7. Virtual pools for interactive analysis and software development through an integrated Cloud environment

    NASA Astrophysics Data System (ADS)

    Grandi, C.; Italiano, A.; Salomoni, D.; Calabrese Melcarne, A. K.

    2011-12-01

    WNoDeS, an acronym for Worker Nodes on Demand Service, is software developed at CNAF-Tier1, the National Computing Centre of the Italian Institute for Nuclear Physics (INFN) located in Bologna. WNoDeS provides on demand, integrated access to both Grid and Cloud resources through virtualization technologies. Besides the traditional use of computing resources in batch mode, users need to have interactive and local access to a number of systems. WNoDeS can dynamically select these computers instantiating Virtual Machines, according to the requirements (computing, storage and network resources) of users through either the Open Cloud Computing Interface API, or through a web console. An interactive use is usually limited to activities in user space, i.e. where the machine configuration is not modified. In some other instances the activity concerns development and testing of services and thus implies the modification of the system configuration (and, therefore, root-access to the resource). The former use case is a simple extension of the WNoDeS approach, where the resource is provided in interactive mode. The latter implies saving the virtual image at the end of each user session so that it can be presented to the user at subsequent requests. This work describes how the LHC experiments at INFN-Bologna are testing and making use of these dynamically created ad-hoc machines via WNoDeS to support flexible, interactive analysis and software development at the INFN Tier-1 Computing Centre.

  8. Comparison of the MPP with other supercomputers for LANDSAT data processing

    NASA Technical Reports Server (NTRS)

    Ozga, Martin

    1987-01-01

    The massively parallel processor is compared to the CRAY X-MP and the CYBER-205 for LANDSAT data processing. The maximum likelihood classification algorithm is the basis for comparison since this algorithm is simple to implement and vectorizes very well. The algorithm was implemented on all three machines and tested by classifying the same full scene of LANDSAT multispectral scan data. Timings are compared as well as features of the machines and available software.

  9. Feasibility of Virtual Machine and Cloud Computing Technologies for High Performance Computing

    DTIC Science & Technology

    2014-05-01

    Hat Enterprise Linux SaaS software as a service VM virtual machine vNUMA virtual non-uniform memory access WRF weather research and forecasting...previously mentioned in Chapter I Section B1 of this paper, which is used to run the weather research and forecasting ( WRF ) model in their experiments...against a VMware virtualization solution of WRF . The experiment consisted of running WRF in a standard configuration between the D-VTM and VMware while

  10. Prospects for Evidence -Based Software Assurance: Models and Analysis

    DTIC Science & Technology

    2015-09-01

    virtual machine is much lighter than the workstation. The virtual machine doesn’t need to run anti- virus , firewalls, intrusion preven- tion systems...34] Maiorca, D., Corona , I., and Giacinto, G. Looking at the bag is not enough to find the bomb: An evasion of structural methods for malicious PDF...CCS ’13, ACM, pp. 119–130. [35] Maiorca, D., Giacinto, G., and Corona , I. A pattern recognition system for malicious PDF files detection. In

  11. Solving the Software Legacy Problem with RISA

    NASA Astrophysics Data System (ADS)

    Ibarra, A.; Gabriel, C.

    2012-09-01

    Nowadays hardware and system infrastructure evolve on time scales much shorter than the typical duration of space astronomy missions. Data processing software capabilities have to evolve to preserve the scientific return during the entire experiment life time. Software preservation is a key issue that has to be tackled before the end of the project to keep the data usable over many years. We present RISA (Remote Interface to Science Analysis) as a solution to decouple data processing software and infrastructure life-cycles, using JAVA applications and web-services wrappers to existing software. This architecture employs embedded SAS in virtual machines assuring a homogeneous job execution environment. We will also present the first studies to reactivate the data processing software of the EXOSAT mission, the first ESA X-ray astronomy mission launched in 1983, using the generic RISA approach.

  12. Productivity improvement using discrete events simulation

    NASA Astrophysics Data System (ADS)

    Hazza, M. H. F. Al; Elbishari, E. M. Y.; Ismail, M. Y. Bin; Adesta, E. Y. T.; Rahman, Nur Salihah Binti Abdul

    2018-01-01

    The increasing in complexity of the manufacturing systems has increased the cost of investment in many industries. Furthermore, the theoretical feasibility studies are not enough to take the decision in investing for that particular area. Therefore, the development of the new advanced software is protecting the manufacturer from investing money in production lines that may not be sufficient and effective with their requirement in terms of machine utilization and productivity issue. By conducting a simulation, using accurate model will reduce and eliminate the risk associated with their new investment. The aim of this research is to prove and highlight the importance of simulation in decision-making process. Delmia quest software was used as a simulation program to run a simulation for the production line. A simulation was first done for the existing production line and show that the estimated production rate is 261 units/day. The results have been analysed based on utilization percentage and idle time. Two different scenarios have been proposed based on different objectives. The first scenario is by focusing on low utilization machines and their idle time, this was resulted in minimizing the number of machines used by three with the addition of the works who maintain them without having an effect on the production rate. The second scenario is to increase the production rate by upgrading the curing machine which lead to the increase in the daily productivity by 7% from 261 units to 281 units.

  13. Staghorn: An Automated Large-Scale Distributed System Analysis Platform

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Gabert, Kasimir; Burns, Ian; Elliott, Steven

    2016-09-01

    Conducting experiments on large-scale distributed computing systems is becoming significantly easier with the assistance of emulation. Researchers can now create a model of a distributed computing environment and then generate a virtual, laboratory copy of the entire system composed of potentially thousands of virtual machines, switches, and software. The use of real software, running at clock rate in full virtual machines, allows experiments to produce meaningful results without necessitating a full understanding of all model components. However, the ability to inspect and modify elements within these models is bound by the limitation that such modifications must compete with the model,more » either running in or alongside it. This inhibits entire classes of analyses from being conducted upon these models. We developed a mechanism to snapshot an entire emulation-based model as it is running. This allows us to \\freeze time" and subsequently fork execution, replay execution, modify arbitrary parts of the model, or deeply explore the model. This snapshot includes capturing packets in transit and other input/output state along with the running virtual machines. We were able to build this system in Linux using Open vSwitch and Kernel Virtual Machines on top of Sandia's emulation platform Firewheel. This primitive opens the door to numerous subsequent analyses on models, including state space exploration, debugging distributed systems, performance optimizations, improved training environments, and improved experiment repeatability.« less

  14. Tool vibration detection with eddy current sensors in machining process and computation of stability lobes using fuzzy classifiers

    NASA Astrophysics Data System (ADS)

    Devillez, Arnaud; Dudzinski, Daniel

    2007-01-01

    Today the knowledge of a process is very important for engineers to find optimal combination of control parameters warranting productivity, quality and functioning without defects and failures. In our laboratory, we carry out research in the field of high speed machining with modelling, simulation and experimental approaches. The aim of our investigation is to develop a software allowing the cutting conditions optimisation to limit the number of predictive tests, and the process monitoring to prevent any trouble during machining operations. This software is based on models and experimental data sets which constitute the knowledge of the process. In this paper, we deal with the problem of vibrations occurring during a machining operation. These vibrations may cause some failures and defects to the process, like workpiece surface alteration and rapid tool wear. To measure on line the tool micro-movements, we equipped a lathe with a specific instrumentation using eddy current sensors. Obtained signals were correlated with surface finish and a signal processing algorithm was used to determine if a test is stable or unstable. Then, a fuzzy classification method was proposed to classify the tests in a space defined by the width of cut and the cutting speed. Finally, it was shown that the fuzzy classification takes into account of the measurements incertitude to compute the stability limit or stability lobes of the process.

  15. Software architecture for time-constrained machine vision applications

    NASA Astrophysics Data System (ADS)

    Usamentiaga, Rubén; Molleda, Julio; García, Daniel F.; Bulnes, Francisco G.

    2013-01-01

    Real-time image and video processing applications require skilled architects, and recent trends in the hardware platform make the design and implementation of these applications increasingly complex. Many frameworks and libraries have been proposed or commercialized to simplify the design and tuning of real-time image processing applications. However, they tend to lack flexibility, because they are normally oriented toward particular types of applications, or they impose specific data processing models such as the pipeline. Other issues include large memory footprints, difficulty for reuse, and inefficient execution on multicore processors. We present a novel software architecture for time-constrained machine vision applications that addresses these issues. The architecture is divided into three layers. The platform abstraction layer provides a high-level application programming interface for the rest of the architecture. The messaging layer provides a message-passing interface based on a dynamic publish/subscribe pattern. A topic-based filtering in which messages are published to topics is used to route the messages from the publishers to the subscribers interested in a particular type of message. The application layer provides a repository for reusable application modules designed for machine vision applications. These modules, which include acquisition, visualization, communication, user interface, and data processing, take advantage of the power of well-known libraries such as OpenCV, Intel IPP, or CUDA. Finally, the proposed architecture is applied to a real machine vision application: a jam detector for steel pickling lines.

  16. Using CAD/CAM to improve productivity - The IPAD approach

    NASA Technical Reports Server (NTRS)

    Fulton, R. E.

    1981-01-01

    Progress in designing and implementing CAD/CAM systems as a result of the NASA Integrated Programs for Aerospace-Vehicle Design is discussed. Essential software packages have been identified as executive, data management, general user, and geometry and graphics software. Data communication, as a means to integrate data over a network of computers of different vendors, provides data management with the capability of meeting design and manufacturing requirements of the vendors. Geometry software is dependent on developmental success with solid geometry software, which is necessary for continual measurements of, for example, a block of metal while it is being machined. Applications in the aerospace industry, such as for design, analysis, tooling, testing, quality control, etc., are outlined.

  17. Software implementation of the SKIPSM paradigm under PIP

    NASA Astrophysics Data System (ADS)

    Hack, Ralf; Waltz, Frederick M.; Batchelor, Bruce G.

    1997-09-01

    SKIPSM (separated-kernel image processing using finite state machines) is a technique for implementing large-kernel binary- morphology operators and many other operations. While earlier papers on SKIPSM concentrated mainly on implementations using pipelined hardware, there is considerable scope for achieving major speed improvements in software systems. Using identical control software, one-pass binary erosion and dilation structuring elements (SEs) ranging from the trivial (3 by 3) to the gigantic (51 by 51, or even larger), are readily available. Processing speed is independent of the size of the SE, making the SKIPSM approach practical for work with very large SEs on ordinary desktop computers. PIP (prolog image processing) is an interactive machine vision prototyping environment developed at the University of Wales Cardiff. It consists of a large number of image processing operators embedded within the standard AI language Prolog. This paper describes the SKIPSM implementation of binary morphology operators within PIP. A large set of binary erosion and dilation operations (circles, squares, diamonds, octagons, etc.) is available to the user through a command-line driven dialogue, via pull-down menus, or incorporated into standard (Prolog) programs. Little has been done thus far to optimize speed on this first software implementation of SKIPSM. Nevertheless, the results are impressive. The paper describes sample applications and presents timing figures. Readers have the opportunity to try out these operations on demonstration software written by the University of Wales, or via their WWW home page at http://bruce.cs.cf.ac.uk/bruce/index.html .

  18. Impact of Spot Size and Spacing on the Quality of Robustly Optimized Intensity Modulated Proton Therapy Plans for Lung Cancer.

    PubMed

    Liu, Chenbin; Schild, Steven E; Chang, Joe Y; Liao, Zhongxing; Korte, Shawn; Shen, Jiajian; Ding, Xiaoning; Hu, Yanle; Kang, Yixiu; Keole, Sameer R; Sio, Terence T; Wong, William W; Sahoo, Narayan; Bues, Martin; Liu, Wei

    2018-06-01

    To investigate how spot size and spacing affect plan quality, robustness, and interplay effects of robustly optimized intensity modulated proton therapy (IMPT) for lung cancer. Two robustly optimized IMPT plans were created for 10 lung cancer patients: first by a large-spot machine with in-air energy-dependent large spot size at isocenter (σ: 6-15 mm) and spacing (1.3 σ), and second by a small-spot machine with in-air energy-dependent small spot size (σ: 2-6 mm) and spacing (5 mm). Both plans were generated by optimizing radiation dose to internal target volume on averaged 4-dimensional computed tomography scans using an in-house-developed IMPT planning system. The dose-volume histograms band method was used to evaluate plan robustness. Dose evaluation software was developed to model time-dependent spot delivery to incorporate interplay effects with randomized starting phases for each field per fraction. Patient anatomy voxels were mapped phase-to-phase via deformable image registration, and doses were scored using in-house-developed software. Dose-volume histogram indices, including internal target volume dose coverage, homogeneity, and organs at risk (OARs) sparing, were compared using the Wilcoxon signed-rank test. Compared with the large-spot machine, the small-spot machine resulted in significantly lower heart and esophagus mean doses, with comparable target dose coverage, homogeneity, and protection of other OARs. Plan robustness was comparable for targets and most OARs. With interplay effects considered, significantly lower heart and esophagus mean doses with comparable target dose coverage and homogeneity were observed using smaller spots. Robust optimization with a small spot-machine significantly improves heart and esophagus sparing, with comparable plan robustness and interplay effects compared with robust optimization with a large-spot machine. A small-spot machine uses a larger number of spots to cover the same tumors compared with a large-spot machine, which gives the planning system more freedom to compensate for the higher sensitivity to uncertainties and interplay effects for lung cancer treatments. Copyright © 2018 Elsevier Inc. All rights reserved.

  19. Application of Artificial Intelligence technology to the analysis and synthesis of reliable software systems

    NASA Technical Reports Server (NTRS)

    Wild, Christian; Eckhardt, Dave

    1987-01-01

    The development of a methodology for the production of highly reliable software is one of the greatest challenges facing the computer industry. Meeting this challenge will undoubtably involve the integration of many technologies. This paper describes the use of Artificial Intelligence technologies in the automated analysis of the formal algebraic specifications of abstract data types. These technologies include symbolic execution of specifications using techniques of automated deduction and machine learning through the use of examples. On-going research into the role of knowledge representation and problem solving in the process of developing software is also discussed.

  20. Modeling reliability measurement of interface on information system: Towards the forensic of rules

    NASA Astrophysics Data System (ADS)

    Nasution, M. K. M.; Sitompul, Darwin; Harahap, Marwan

    2018-02-01

    Today almost all machines depend on the software. As a software and hardware system depends also on the rules that are the procedures for its use. If the procedure or program can be reliably characterized by involving the concept of graph, logic, and probability, then regulatory strength can also be measured accordingly. Therefore, this paper initiates an enumeration model to measure the reliability of interfaces based on the case of information systems supported by the rules of use by the relevant agencies. An enumeration model is obtained based on software reliability calculation.

  1. A New Look at NASA: Strategic Research In Information Technology

    NASA Technical Reports Server (NTRS)

    Alfano, David; Tu, Eugene (Technical Monitor)

    2002-01-01

    This viewgraph presentation provides information on research undertaken by NASA to facilitate the development of information technologies. Specific ideas covered here include: 1) Bio/nano technologies: biomolecular and nanoscale systems and tools for assembly and computing; 2) Evolvable hardware: autonomous self-improving, self-repairing hardware and software for survivable space systems in extreme environments; 3) High Confidence Software Technologies: formal methods, high-assurance software design, and program synthesis; 4) Intelligent Controls and Diagnostics: Next generation machine learning, adaptive control, and health management technologies; 5) Revolutionary computing: New computational models to increase capability and robustness to enable future NASA space missions.

  2. Formal specification and verification of Ada software

    NASA Technical Reports Server (NTRS)

    Hird, Geoffrey R.

    1991-01-01

    The use of formal methods in software development achieves levels of quality assurance unobtainable by other means. The Larch approach to specification is described, and the specification of avionics software designed to implement the logic of a flight control system is given as an example. Penelope is described which is an Ada-verification environment. The Penelope user inputs mathematical definitions, Larch-style specifications and Ada code and performs machine-assisted proofs that the code obeys its specifications. As an example, the verification of a binary search function is considered. Emphasis is given to techniques assisting the reuse of a verification effort on modified code.

  3. Instrument control software requirement specification for Extremely Large Telescopes

    NASA Astrophysics Data System (ADS)

    Young, Peter J.; Kiekebusch, Mario J.; Chiozzi, Gianluca

    2010-07-01

    Engineers in several observatories are now designing the next generation of optical telescopes, the Extremely Large Telescopes (ELT). These are very complex machines that will host sophisticated astronomical instruments to be used for a wide range of scientific studies. In order to carry out scientific observations, a software infrastructure is required to orchestrate the control of the multiple subsystems and functions. This paper will focus on describing the considerations, strategies and main issues related to the definition and analysis of the software requirements for the ELT's Instrument Control System using modern development processes and modelling tools like SysML.

  4. Europe Report, Science and Technology.

    DTIC Science & Technology

    1986-05-29

    developed the institutional software training of today, primarily in institutions of higher learning or in study courses. TV BASIC-- it really would be a...only to develop software but also to offer large catalogs to users of its machines. Bull is studying the product-marketing strategies of companies such...by DEC, an XCON + XSEL package). A prototype is under study ; it was developed on an SPS 9 with Kool and a relational data base. The first

  5. Survivable Software

    DTIC Science & Technology

    2013-04-01

    machine transitions. 2. We developed the TraceContract API for trace analysis in the Scala programming language. TraceContract combines a high-level...DATE (DD-MM-YYYY) 2. REPORT TYPE 3. DATES COVERED (From - To) 4. TITLE AND SUBTITLE 5a. CONTRACT NUMBER 5b. GRANT NUMBER 5c. PROGRAM ELEMENT NUMBER...awarded within the Software and Systems program . The original Pro- gram Manager was David Luginbuhl. Bob Bonneau took over as PM in March 2011. The award

  6. State transition storyboards: A tool for designing the Goldstone solar system radar data acquisition system user interface software

    NASA Technical Reports Server (NTRS)

    Howard, S. D.

    1987-01-01

    Effective user interface design in software systems is a complex task that takes place without adequate modeling tools. By combining state transition diagrams and the storyboard technique of filmmakers, State Transition Storyboards were developed to provide a detailed modeling technique for the Goldstone Solar System Radar Data Acquisition System human-machine interface. Illustrations are included with a description of the modeling technique.

  7. Advanced Online Survival Analysis Tool for Predictive Modelling in Clinical Data Science.

    PubMed

    Montes-Torres, Julio; Subirats, José Luis; Ribelles, Nuria; Urda, Daniel; Franco, Leonardo; Alba, Emilio; Jerez, José Manuel

    2016-01-01

    One of the prevailing applications of machine learning is the use of predictive modelling in clinical survival analysis. In this work, we present our view of the current situation of computer tools for survival analysis, stressing the need of transferring the latest results in the field of machine learning to biomedical researchers. We propose a web based software for survival analysis called OSA (Online Survival Analysis), which has been developed as an open access and user friendly option to obtain discrete time, predictive survival models at individual level using machine learning techniques, and to perform standard survival analysis. OSA employs an Artificial Neural Network (ANN) based method to produce the predictive survival models. Additionally, the software can easily generate survival and hazard curves with multiple options to personalise the plots, obtain contingency tables from the uploaded data to perform different tests, and fit a Cox regression model from a number of predictor variables. In the Materials and Methods section, we depict the general architecture of the application and introduce the mathematical background of each of the implemented methods. The study concludes with examples of use showing the results obtained with public datasets.

  8. Advanced Online Survival Analysis Tool for Predictive Modelling in Clinical Data Science

    PubMed Central

    Montes-Torres, Julio; Subirats, José Luis; Ribelles, Nuria; Urda, Daniel; Franco, Leonardo; Alba, Emilio; Jerez, José Manuel

    2016-01-01

    One of the prevailing applications of machine learning is the use of predictive modelling in clinical survival analysis. In this work, we present our view of the current situation of computer tools for survival analysis, stressing the need of transferring the latest results in the field of machine learning to biomedical researchers. We propose a web based software for survival analysis called OSA (Online Survival Analysis), which has been developed as an open access and user friendly option to obtain discrete time, predictive survival models at individual level using machine learning techniques, and to perform standard survival analysis. OSA employs an Artificial Neural Network (ANN) based method to produce the predictive survival models. Additionally, the software can easily generate survival and hazard curves with multiple options to personalise the plots, obtain contingency tables from the uploaded data to perform different tests, and fit a Cox regression model from a number of predictor variables. In the Materials and Methods section, we depict the general architecture of the application and introduce the mathematical background of each of the implemented methods. The study concludes with examples of use showing the results obtained with public datasets. PMID:27532883

  9. Summary of vulnerability related technologies based on machine learning

    NASA Astrophysics Data System (ADS)

    Zhao, Lei; Chen, Zhihao; Jia, Qiong

    2018-04-01

    As the scale of information system increases by an order of magnitude, the complexity of system software is getting higher. The vulnerability interaction from design, development and deployment to implementation stages greatly increases the risk of the entire information system being attacked successfully. Considering the limitations and lags of the existing mainstream security vulnerability detection techniques, this paper summarizes the development and current status of related technologies based on the machine learning methods applied to deal with massive and irregular data, and handling security vulnerabilities.

  10. A neurite quality index and machine vision software for improved quantification of neurodegeneration.

    PubMed

    Romero, Peggy; Miller, Ted; Garakani, Arman

    2009-12-01

    Current methods to assess neurodegradation in dorsal root ganglion cultures as a model for neurodegenerative diseases are imprecise and time-consuming. Here we describe two new methods to quantify neuroprotection in these cultures. The neurite quality index (NQI) builds upon earlier manual methods, incorporating additional morphological events to increase detection sensitivity for the detection of early degeneration events. Neurosight is a machine vision-based method that recapitulates many of the strengths of NQI while enabling high-throughput screening applications with decreased costs.

  11. Experimental Investigation of Mechanical Behavior of an Oxide/Oxide Ceramic Composite in Interlaminar Shear and under Combined Tension-Torsion Loading

    DTIC Science & Technology

    2014-03-27

    testing machine was warmed up for at least 30 min using a cyclic command with a sine waveform in displacement control . Gripping sections of each test...the test specimen was inserted into the susceptor. Then the testing machine is placed in displacement control and the top portion of the specimen...the MTS software also triggered the operation of the high speed cameras. 31 The testing system was placed in displacement /rotation control and the

  12. Reducing maintenance costs in agreement with CNC machine tools reliability

    NASA Astrophysics Data System (ADS)

    Ungureanu, A. L.; Stan, G.; Butunoi, P. A.

    2016-08-01

    Aligning maintenance strategy with reliability is a challenge due to the need to find an optimal balance between them. Because the various methods described in the relevant literature involve laborious calculations or use of software that can be costly, this paper proposes a method that is easier to implement on CNC machine tools. The new method, called the Consequence of Failure Analysis (CFA) is based on technical and economic optimization, aimed at obtaining a level of required performance with minimum investment and maintenance costs.

  13. The Data Acquisition System for the AAO 2-Degree Field Project

    NASA Astrophysics Data System (ADS)

    Shortridge, K.; Farrell, T. J.; Bailey, J. A.

    1993-01-01

    The software system being produced by AAO to control the new 2-degree field fibre positioner and spectrographs is described. The system has to mesh cleanly with the ADAM systems used at AAO for CCD data acquisition, and has to run on a network of disparate machines including VMS Vaxes, UNIX workstations, and VME systems running VxWorks. The basis of the new system is a task control layer that operates by sending self-defining hierarchically-structured and machine-independent messages.

  14. Image understanding and the man-machine interface II; Proceedings of the Meeting, Los Angeles, CA, Jan. 17, 18, 1989

    NASA Technical Reports Server (NTRS)

    Barrett, Eamon B. (Editor); Pearson, James J. (Editor)

    1989-01-01

    Image understanding concepts and models, image understanding systems and applications, advanced digital processors and software tools, and advanced man-machine interfaces are among the topics discussed. Particular papers are presented on such topics as neural networks for computer vision, object-based segmentation and color recognition in multispectral images, the application of image algebra to image measurement and feature extraction, and the integration of modeling and graphics to create an infrared signal processing test bed.

  15. Algorithm for detection the QRS complexes based on support vector machine

    NASA Astrophysics Data System (ADS)

    Van, G. V.; Podmasteryev, K. V.

    2017-11-01

    The efficiency of computer ECG analysis depends on the accurate detection of QRS-complexes. This paper presents an algorithm for QRS complex detection based of support vector machine (SVM). The proposed algorithm is evaluated on annotated standard databases such as MIT-BIH Arrhythmia database. The QRS detector obtained a sensitivity Se = 98.32% and specificity Sp = 95.46% for MIT-BIH Arrhythmia database. This algorithm can be used as the basis for the software to diagnose electrical activity of the heart.

  16. A self-configuring control system for storage and computing departments at INFN-CNAF Tierl

    NASA Astrophysics Data System (ADS)

    Gregori, Daniele; Dal Pra, Stefano; Ricci, Pier Paolo; Pezzi, Michele; Prosperini, Andrea; Sapunenko, Vladimir

    2015-05-01

    The storage and farming departments at the INFN-CNAF Tier1[1] manage approximately thousands of computing nodes and several hundreds of servers that provides access to the disk and tape storage. In particular, the storage server machines should provide the following services: an efficient access to about 15 petabytes of disk space with different cluster of GPFS file system, the data transfers between LHC Tiers sites (Tier0, Tier1 and Tier2) via GridFTP cluster and Xrootd protocol and finally the writing and reading data operations on magnetic tape backend. One of the most important and essential point in order to get a reliable service is a control system that can warn if problems arise and which is able to perform automatic recovery operations in case of service interruptions or major failures. Moreover, during daily operations the configurations can change, i.e. if the GPFS cluster nodes roles can be modified and therefore the obsolete nodes must be removed from the control system production, and the new servers should be added to the ones that are already present. The manual management of all these changes is an operation that can be somewhat difficult in case of several changes, it can also take a long time and is easily subject to human error or misconfiguration. For these reasons we have developed a control system with the feature of self-configure itself if any change occurs. Currently, this system has been in production for about a year at the INFN-CNAF Tier1 with good results and hardly any major drawback. There are three major key points in this system. The first is a software configurator service (e.g. Quattor or Puppet) for the servers machines that we want to monitor with the control system; this service must ensure the presence of appropriate sensors and custom scripts on the nodes to check and should be able to install and update software packages on them. The second key element is a database containing information, according to a suitable format, on all the machines in production and able to provide for each of them the principal information such as the type of hardware, the network switch to which the machine is connected, if the machine is real (physical) or virtual, the possible hypervisor to which it belongs and so on. The last key point is a control system software (in our implementation we choose the Nagios software), capable of assessing the status of the servers and services, and that can attempt to restore the working state, restart or inhibit software services and send suitable alarm messages to the site administrators. The integration of these three elements was made by appropriate scripts and custom implementation that allow the self-configuration of the system according to a decisional logic and the whole combination of all the above-mentioned components will be deeply discussed in this paper.

  17. Machine learning phases of matter

    NASA Astrophysics Data System (ADS)

    Carrasquilla, Juan; Melko, Roger G.

    2017-02-01

    Condensed-matter physics is the study of the collective behaviour of infinitely complex assemblies of electrons, nuclei, magnetic moments, atoms or qubits. This complexity is reflected in the size of the state space, which grows exponentially with the number of particles, reminiscent of the `curse of dimensionality' commonly encountered in machine learning. Despite this curse, the machine learning community has developed techniques with remarkable abilities to recognize, classify, and characterize complex sets of data. Here, we show that modern machine learning architectures, such as fully connected and convolutional neural networks, can identify phases and phase transitions in a variety of condensed-matter Hamiltonians. Readily programmable through modern software libraries, neural networks can be trained to detect multiple types of order parameter, as well as highly non-trivial states with no conventional order, directly from raw state configurations sampled with Monte Carlo.

  18. Impact of Machine Virtualization on Timing Precision for Performance-critical Tasks

    NASA Astrophysics Data System (ADS)

    Karpov, Kirill; Fedotova, Irina; Siemens, Eduard

    2017-07-01

    In this paper we present a measurement study to characterize the impact of hardware virtualization on basic software timing, as well as on precise sleep operations of an operating system. We investigated how timer hardware is shared among heavily CPU-, I/O- and Network-bound tasks on a virtual machine as well as on the host machine. VMware ESXi and QEMU/KVM have been chosen as commonly used examples of hypervisor- and host-based models. Based on statistical parameters of retrieved distributions, our results provide a very good estimation of timing behavior. It is essential for real-time and performance-critical applications such as image processing or real-time control.

  19. The Mathematical Modeling and Computer Simulation of Electrochemical Micromachining Using Ultrashort Pulses

    NASA Astrophysics Data System (ADS)

    Kozak, J.; Gulbinowicz, D.; Gulbinowicz, Z.

    2009-05-01

    The need for complex and accurate three dimensional (3-D) microcomponents is increasing rapidly for many industrial and consumer products. Electrochemical machining process (ECM) has the potential of generating desired crack-free and stress-free surfaces of microcomponents. This paper reports a study of pulse electrochemical micromachining (PECMM) using ultrashort (nanoseconds) pulses for generating complex 3-D microstructures of high accuracy. A mathematical model of the microshaping process with taking into consideration unsteady phenomena in electrical double layer has been developed. The software for computer simulation of PECM has been developed and the effects of machining parameters on anodic localization and final shape of machined surface are presented.

  20. An alternative model to distribute VO software to WLCG sites based on CernVM-FS: a prototype at PIC Tier1

    NASA Astrophysics Data System (ADS)

    Lanciotti, E.; Merino, G.; Bria, A.; Blomer, J.

    2011-12-01

    In a distributed computing model as WLCG the software of experiment specific application software has to be efficiently distributed to any site of the Grid. Application software is currently installed in a shared area of the site visible for all Worker Nodes (WNs) of the site through some protocol (NFS, AFS or other). The software is installed at the site by jobs which run on a privileged node of the computing farm where the shared area is mounted in write mode. This model presents several drawbacks which cause a non-negligible rate of job failure. An alternative model for software distribution based on the CERN Virtual Machine File System (CernVM-FS) has been tried at PIC, the Spanish Tierl site of WLCG. The test bed used and the results are presented in this paper.

  1. Unresolved Galaxy Classifier for ESA/Gaia mission: Support Vector Machines approach

    NASA Astrophysics Data System (ADS)

    Bellas-Velidis, Ioannis; Kontizas, Mary; Dapergolas, Anastasios; Livanou, Evdokia; Kontizas, Evangelos; Karampelas, Antonios

    A software package Unresolved Galaxy Classifier (UGC) is being developed for the ground-based pipeline of ESA's Gaia mission. It aims to provide an automated taxonomic classification and specific parameters estimation analyzing Gaia BP/RP instrument low-dispersion spectra of unresolved galaxies. The UGC algorithm is based on a supervised learning technique, the Support Vector Machines (SVM). The software is implemented in Java as two separate modules. An offline learning module provides functions for SVM-models training. Once trained, the set of models can be repeatedly applied to unknown galaxy spectra by the pipeline's application module. A library of galaxy models synthetic spectra, simulated for the BP/RP instrument, is used to train and test the modules. Science tests show a very good classification performance of UGC and relatively good regression performance, except for some of the parameters. Possible approaches to improve the performance are discussed.

  2. Man-machine interfaces in LACIE/ERIPS

    NASA Technical Reports Server (NTRS)

    Duprey, B. B. (Principal Investigator)

    1979-01-01

    One of the most important aspects of the interactive portion of the LACIE/ERIPS software system is the way in which the analysis and decision-making capabilities of a human being are integrated with the speed and accuracy of a computer to produce a powerful analysis system. The three major man-machine interfaces in the system are (1) the use of menus for communications between the software and the interactive user; (2) the checkpoint/restart facility to recreate in one job the internal environment achieved in an earlier one; and (3) the error recovery capability which would normally cause job termination. This interactive system, which executes on an IBM 360/75 mainframe, was adapted for use in noninteractive (batch) mode. A case study is presented to show how the interfaces work in practice by defining some fields based on an image screen display, noting the field definitions, and obtaining a film product of the classification map.

  3. Final Report, University Research Program in Robotics (URPR), Nuclear Facilities Clean-up

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Tesar, Delbert; Kapoor, Chetan; Pryor, Mitch

    This final report describes the research activity at the University of Texas at Austin with application to EM needs at DOE. This research activity is divided in to two major thrusts and contributes to the overall University Research Program in Robotics (URPR) thrust by providing mechanically oriented robotic solutions based on modularity and generalized software. These thrusts are also the core strengths of the UTA program that has a 40-year history in machine development, 30 years specifically devoted to robotics. Since 1975, much of this effort has been to establish the general analytical and design infrastructure for an open (modular)more » architecture of systems with many degrees of freedom that are able to satisfy a broad range of applications for future production machines. This work has coalesced from two principal areas: standardized actuators and generalized software.« less

  4. Magnetic properties measurement of soft magnetic composite material (SOMALOY 700) by using 3-D tester

    NASA Astrophysics Data System (ADS)

    Asari, Ashraf; Guo, Youguang; Zhu, Jianguo

    2017-08-01

    Core losses of rotating electrical machine can be predicted by identifying the magnetic properties of the magnetic material. The magnetic properties should be properly measured since there are some variations of vector flux density in the rotating machine. In this paper, the SOMALOY 700 material has been measured under x, y and z- axes flux density penetration by using the 3-D tester. The calibrated sensing coils are used in detecting the flux densities which have been generated by the Labview software. The measured sensing voltages are used in obtaining the magnetic properties of the sample such as magnetic flux density B, magnetic field strength H, hysteresis loop which can be used to calculate the total core loss of the sample. The results of the measurement are analyzed by using the Mathcad software before being compared to another material.

  5. DOE Office of Scientific and Technical Information (OSTI.GOV)

    Jones, J.P.; Bangs, A.L.; Butler, P.L.

    Hetero Helix is a programming environment which simulates shared memory on a heterogeneous network of distributed-memory computers. The machines in the network may vary with respect to their native operating systems and internal representation of numbers. Hetero Helix presents a simple programming model to developers, and also considers the needs of designers, system integrators, and maintainers. The key software technology underlying Hetero Helix is the use of a compiler'' which analyzes the data structures in shared memory and automatically generates code which translates data representations from the format native to each machine into a common format, and vice versa. Themore » design of Hetero Helix was motivated in particular by the requirements of robotics applications. Hetero Helix has been used successfully in an integration effort involving 27 CPUs in a heterogeneous network and a body of software totaling roughly 100,00 lines of code. 25 refs., 6 figs.« less

  6. Perl Extension to the Bproc Library

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Grunau, Daryl W.

    2004-06-07

    The Beowulf Distributed process Space (Bproc) software stack is comprised of UNIX/Linux kernel modifications and a support library by which a cluster of machines, each running their own private kernel, can present itself as a unified process space to the user. A Bproc cluster contains a single front-end machine and many back-end nodes which receive and run processes given to them by the front-end. Any process which is migrated to a back-end node is also visible as a ghost process on the fron-end, and may be controlled there using traditional UNIX semantics (e.g. ps(1), kill(1), etc). This software is amore » Perl extension to the Bproc library which enables the Perl programmer to make direct calls to functions within the Bproc library. See http://www.clustermatic.org, http://bproc.sourceforge.net, and http://www.perl.org« less

  7. UltraNet Target Parameters. Chapter 1

    NASA Technical Reports Server (NTRS)

    Kislitzin, Katherine T.; Blaylock, Bruce T. (Technical Monitor)

    1992-01-01

    The UltraNet is a high speed network capable of rates up to one gigabit per second. It is a hub based network with four optical fiber links connecting each hub. Each link can carry up to 256 megabits of data, and the hub backplane is capable of one gigabit aggregate throughput. Host connections to the hub may be fiber, coax, or channel based. Bus based machines have adapter boards that connect to transceivers in the hub, while channel based machines use a personality module in the hub. One way that the UltraNet achieves its high transfer rates is by off-loading the protocol processing from the hosts to special purpose protocol engines in the UltraNet hubs. In addition, every hub has a PC connected to it by StarLAN for network management purposes. Although there is hub resident and PC resident UltraNet software, this document treats only the host resident UltraNet software.

  8. History of visual systems in the Systems Engineering Simulator

    NASA Technical Reports Server (NTRS)

    Christianson, David C.

    1989-01-01

    The Systems Engineering Simulator (SES) houses a variety of real-time computer generated visual systems. The earliest machine dates from the mid-1960's and is one of the first real-time graphics systems in the world. The latest acquisition is the state-of-the-art Evans and Sutherland CT6. Between the span of time from the mid-1960's to the late 1980's, tremendous strides have been made in the real-time graphics world. These strides include advances in both software and hardware engineering. The purpose is to explore the history of the development of these real-time computer generated image systems from the first machine to the present. Hardware advances as well as software algorithm changes are presented. This history is not only quite interesting but also provides us with a perspective with which we can look backward and forward.

  9. Diskless supercomputers: Scalable, reliable I/O for the Tera-Op technology base

    NASA Technical Reports Server (NTRS)

    Katz, Randy H.; Ousterhout, John K.; Patterson, David A.

    1993-01-01

    Computing is seeing an unprecedented improvement in performance; over the last five years there has been an order-of-magnitude improvement in the speeds of workstation CPU's. At least another order of magnitude seems likely in the next five years, to machines with 500 MIPS or more. The goal of the ARPA Teraop program is to realize even larger, more powerful machines, executing as many as a trillion operations per second. Unfortunately, we have seen no comparable breakthroughs in I/O performance; the speeds of I/O devices and the hardware and software architectures for managing them have not changed substantially in many years. We have completed a program of research to demonstrate hardware and software I/O architectures capable of supporting the kinds of internetworked 'visualization' workstations and supercomputers that will appear in the mid 1990s. The project had three overall goals: high performance, high reliability, and scalable, multipurpose system.

  10. Math Machines: Using Actuators in Physics Classes

    NASA Astrophysics Data System (ADS)

    Thomas, Frederick J.; Chaney, Robert A.; Gruesbeck, Marta

    2018-01-01

    Probeware (sensors combined with data-analysis software) is a well-established part of physics education. In engineering and technology, sensors are frequently paired with actuators—motors, heaters, buzzers, valves, color displays, medical dosing systems, and other devices that are activated by electrical signals to produce intentional physical change. This article describes how a 20-year project aimed at better integration of the STEM disciplines (science, technology, engineering and mathematics) uses brief actuator activities in physics instruction. Math Machines "actionware" includes software and hardware that convert virtually any free-form, time-dependent algebraic function into the dynamic actions of a stepper motor, servo motor, or RGB (red, green, blue) color mixer. With wheels and a platform, the stepper motor becomes LACI, a programmable vehicle. Adding a low-power laser module turns the servo motor into a programmable Pointer. Adding a gear and platform can transform the Pointer into an earthquake simulator.

  11. WE-G-BRA-02: SafetyNet: Automating Radiotherapy QA with An Event Driven Framework

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Hadley, S; Kessler, M; Litzenberg, D

    2015-06-15

    Purpose: Quality assurance is an essential task in radiotherapy that often requires many manual tasks. We investigate the use of an event driven framework in conjunction with software agents to automate QA and eliminate wait times. Methods: An in house developed subscription-publication service, EventNet, was added to the Aria OIS to be a message broker for critical events occurring in the OIS and software agents. Software agents operate without user intervention and perform critical QA steps. The results of the QA are documented and the resulting event is generated and passed back to EventNet. Users can subscribe to those eventsmore » and receive messages based on custom filters designed to send passing or failing results to physicists or dosimetrists. Agents were developed to expedite the following QA tasks: Plan Revision, Plan 2nd Check, SRS Winston-Lutz isocenter, Treatment History Audit, Treatment Machine Configuration. Results: Plan approval in the Aria OIS was used as the event trigger for plan revision QA and Plan 2nd check agents. The agents pulled the plan data, executed the prescribed QA, stored the results and updated EventNet for publication. The Winston Lutz agent reduced QA time from 20 minutes to 4 minutes and provided a more accurate quantitative estimate of radiation isocenter. The Treatment Machine Configuration agent automatically reports any changes to the Treatment machine or HDR unit configuration. The agents are reliable, act immediately, and execute each task identically every time. Conclusion: An event driven framework has inverted the data chase in our radiotherapy QA process. Rather than have dosimetrists and physicists push data to QA software and pull results back into the OIS, the software agents perform these steps immediately upon receiving the sentinel events from EventNet. Mr Keranen is an employee of Varian Medical Systems. Dr. Moran’s institution receives research support for her effort for a linear accelerator QA project from Varian Medical Systems. Other quality projects involving her effort are funded by Blue Cross Blue Shield of Michigan, Breast Cancer Research Foundation, and the NIH.« less

  12. Needs of ergonomic design at control units in production industries.

    PubMed

    Levchuk, I; Schäfer, A; Lang, K-H; Gebhardt, Hj; Klussmann, A

    2012-01-01

    During the last decades, an increasing use of innovative technologies in manufacturing areas was monitored. A huge amount of physical workload was replaced by the change from conventional machine tools to computer-controlled units. CNC systems spread in current production processes. Because of this, machine operators today mostly have an observational function. This caused increasing of static work (e.g., standing, sitting) and cognitive demands (e.g., process observation). Machine operators have a high responsibility, because mistakes may lead to human injuries as well as to product losses - and in consequence may lead to high monetary losses (for the company) as well. Being usable often means for a CNC machine being efficient. An intuitive usability and an ergonomic organization of CNC workplaces can be an essential basis to reduce the risk of failures in operation as well as physical complaints (e.g. pain or diseases because of bad body posture during work). In contrast to conventional machines, CNC machines are equipped both with hardware and software. An intuitive and clear-sighted operating of CNC systems is a requirement for quick learning of new systems. Within this study, a survey was carried out among trainees learning the operation of CNC machines.

  13. Function library programming to support B89 evaluation of Sheffield Apollo RS50 DCC (Direct Computer Control) CMM (Coordinate Measuring Machine)

    DOE Office of Scientific and Technical Information (OSTI.GOV)

    Frank, R.N.

    1990-02-28

    The Inspection Shop at Lawrence Livermore Lab recently purchased a Sheffield Apollo RS50 Direct Computer Control Coordinate Measuring Machine. The performance of the machine was specified to conform to B89 standard which relies heavily upon using the measuring machine in its intended manner to verify its accuracy (rather than parametric tests). Although it would be possible to use the interactive measurement system to perform these tasks, a more thorough and efficient job can be done by creating Function Library programs for certain tasks which integrate Hewlett-Packard Basic 5.0 language and calls to proprietary analysis and machine control routines. This combinationmore » provides efficient use of the measuring machine with a minimum of keyboard input plus an analysis of the data with respect to the B89 Standard rather than a CMM analysis which would require subsequent interpretation. This paper discusses some characteristics of the Sheffield machine control and analysis software and my use of H-P Basic language to create automated measurement programs to support the B89 performance evaluation of the CMM. 1 ref.« less

  14. Protyping machine vision software on the World Wide Web

    NASA Astrophysics Data System (ADS)

    Karantalis, George; Batchelor, Bruce G.

    1998-10-01

    Interactive image processing is a proven technique for analyzing industrial vision applications and building prototype systems. Several of the previous implementations have used dedicated hardware to perform the image processing, with a top layer of software providing a convenient user interface. More recently, self-contained software packages have been devised and these run on a standard computer. The advent of the Java programming language has made it possible to write platform-independent software, operating over the Internet, or a company-wide Intranet. Thus, there arises the possibility of designing at least some shop-floor inspection/control systems, without the vision engineer ever entering the factories where they will be used. It successful, this project will have a major impact on the productivity of vision systems designers.

  15. [Microcomputer control of a LED stimulus display device].

    PubMed

    Ohmoto, S; Kikuchi, T; Kumada, T

    1987-02-01

    A visual stimulus display system controlled by a microcomputer was constructed at low cost. The system consists of a LED stimulus display device, a microcomputer, two interface boards, a pointing device (a "mouse") and two kinds of software. The first software package is written in BASIC. Its functions are: to construct stimulus patterns using the mouse, to construct letter patterns (alphabet, digit, symbols and Japanese letters--kanji, hiragana, katakana), to modify the patterns, to store the patterns on a floppy disc, to translate the patterns into integer data which are used to display the patterns in the second software. The second software package, written in BASIC and machine language, controls display of a sequence of stimulus patterns in predetermined time schedules in visual experiments.

  16. Neural network technologies

    NASA Technical Reports Server (NTRS)

    Villarreal, James A.

    1991-01-01

    A whole new arena of computer technologies is now beginning to form. Still in its infancy, neural network technology is a biologically inspired methodology which draws on nature's own cognitive processes. The Software Technology Branch has provided a software tool, Neural Execution and Training System (NETS), to industry, government, and academia to facilitate and expedite the use of this technology. NETS is written in the C programming language and can be executed on a variety of machines. Once a network has been debugged, NETS can produce a C source code which implements the network. This code can then be incorporated into other software systems. Described here are various software projects currently under development with NETS and the anticipated future enhancements to NETS and the technology.

  17. Heuristic for Critical Machine Based a Lot Streaming for Two-Stage Hybrid Production Environment

    NASA Astrophysics Data System (ADS)

    Vivek, P.; Saravanan, R.; Chandrasekaran, M.; Pugazhenthi, R.

    2017-03-01

    Lot streaming in Hybrid flowshop [HFS] is encountered in many real world problems. This paper deals with a heuristic approach for Lot streaming based on critical machine consideration for a two stage Hybrid Flowshop. The first stage has two identical parallel machines and the second stage has only one machine. In the second stage machine is considered as a critical by valid reasons these kind of problems is known as NP hard. A mathematical model developed for the selected problem. The simulation modelling and analysis were carried out in Extend V6 software. The heuristic developed for obtaining optimal lot streaming schedule. The eleven cases of lot streaming were considered. The proposed heuristic was verified and validated by real time simulation experiments. All possible lot streaming strategies and possible sequence under each lot streaming strategy were simulated and examined. The heuristic consistently yielded optimal schedule consistently in all eleven cases. The identification procedure for select best lot streaming strategy was suggested.

  18. Space Station man-machine automation trade-off analysis

    NASA Technical Reports Server (NTRS)

    Zimmerman, W. F.; Bard, J.; Feinberg, A.

    1985-01-01

    The man machine automation tradeoff methodology presented is of four research tasks comprising the autonomous spacecraft system technology (ASST) project. ASST was established to identify and study system level design problems for autonomous spacecraft. Using the Space Station as an example spacecraft system requiring a certain level of autonomous control, a system level, man machine automation tradeoff methodology is presented that: (1) optimizes man machine mixes for different ground and on orbit crew functions subject to cost, safety, weight, power, and reliability constraints, and (2) plots the best incorporation plan for new, emerging technologies by weighing cost, relative availability, reliability, safety, importance to out year missions, and ease of retrofit. A fairly straightforward approach is taken by the methodology to valuing human productivity, it is still sensitive to the important subtleties associated with designing a well integrated, man machine system. These subtleties include considerations such as crew preference to retain certain spacecraft control functions; or valuing human integration/decision capabilities over equivalent hardware/software where appropriate.

  19. Study on Parallel 2-DOF Rotation Machanism in Radar

    NASA Astrophysics Data System (ADS)

    Jiang, Ming; Hu, Xuelong; Liu, Lei; Yu, Yunfei

    The spherical parallel machine has become the world's academic and industrial focus of the field in recent years due to its simple and economical manufacture as well as its structural compactness especially suitable for areas where space gesture changes. This paper dwells upon its present research and development home and abroad. The newer machine (RGRR-II) can rotate around the axis z within 360° and the axis y1 from -90° to +90°. It has the advantages such as less moving parts (only 3 parts), larger ratio of work space to machine size, zero mechanic coupling, no singularity. Constructing rotation machine with spherical parallel 2-DOF rotation join (RGRR-II) may realize semispherical movement with zero dead point and extent the range. Control card (PA8000NT Series CNC) is installed in the computer. The card can run the corresponding software which realizes radar movement control. The machine meets the need of radars in plane and satellite which require larger detection range, lighter weight and compacter structure.

  20. Automated Analysis of ARM Binaries using the Low-Level Virtual Machine Compiler Framework

    DTIC Science & Technology

    2011-03-01

    president to insist on keeping his smartphone [CNN09]. A self-proclaimed BlackBerry addict , President Obama fought hard to keep his mobile device after his... smartphone but renders a device non-functional on installation [FSe09][Hof07]. Complex interactions between hardware and software components both within... smartphone (which is a big assumption), the phone may still be vulnerable if the hardware or software does not correctly implement the design

  1. Protecting Externally Supplied Software in Small Computers.

    DTIC Science & Technology

    1980-09-01

    smiall computer sysiemrS incor-poratinig these SecCurity fetReNS reCquireCs careful analysis of a number of options in making tradeoffs amnong perfbrmance...alarming statistic is not representative of the market as a whole or that it is not indicative of the fate of sales of such software in the future. In ... market as well. Although the size of this market ( in numbers of machines) may not approach that of personal computers, small business computers may

  2. Drone Control System

    NASA Technical Reports Server (NTRS)

    1983-01-01

    Drones, subscale vehicles like the Firebees, and full scale retired military aircraft are used to test air defense missile systems. The DFCS (Drone Formation Control System) computer, developed by IBM (International Business Machines) Federal Systems Division, can track ten drones at once. A program called ORACLS is used to generate software to track and control Drones. It was originally developed by Langley and supplied by COSMIC (Computer Software Management and Information Center). The program saved the company both time and money.

  3. Technique for Early Reliability Prediction of Software Components Using Behaviour Models

    PubMed Central

    Ali, Awad; N. A. Jawawi, Dayang; Adham Isa, Mohd; Imran Babar, Muhammad

    2016-01-01

    Behaviour models are the most commonly used input for predicting the reliability of a software system at the early design stage. A component behaviour model reveals the structure and behaviour of the component during the execution of system-level functionalities. There are various challenges related to component reliability prediction at the early design stage based on behaviour models. For example, most of the current reliability techniques do not provide fine-grained sequential behaviour models of individual components and fail to consider the loop entry and exit points in the reliability computation. Moreover, some of the current techniques do not tackle the problem of operational data unavailability and the lack of analysis results that can be valuable for software architects at the early design stage. This paper proposes a reliability prediction technique that, pragmatically, synthesizes system behaviour in the form of a state machine, given a set of scenarios and corresponding constraints as input. The state machine is utilized as a base for generating the component-relevant operational data. The state machine is also used as a source for identifying the nodes and edges of a component probabilistic dependency graph (CPDG). Based on the CPDG, a stack-based algorithm is used to compute the reliability. The proposed technique is evaluated by a comparison with existing techniques and the application of sensitivity analysis to a robotic wheelchair system as a case study. The results indicate that the proposed technique is more relevant at the early design stage compared to existing works, and can provide a more realistic and meaningful prediction. PMID:27668748

  4. An evaluation of software tools for the design and development of cockpit displays

    NASA Technical Reports Server (NTRS)

    Ellis, Thomas D., Jr.

    1993-01-01

    The use of all-glass cockpits at the NASA Langley Research Center (LaRC) simulation facility has changed the means of design, development, and maintenance of instrument displays. The human-machine interface has evolved from a physical hardware device to a software-generated electronic display system. This has subsequently caused an increased workload at the facility. As computer processing power increases and the glass cockpit becomes predominant in facilities, software tools used in the design and development of cockpit displays are becoming both feasible and necessary for a more productive simulation environment. This paper defines LaRC requirements of a display software development tool and compares two available applications against these requirements. As a part of the software engineering process, these tools reduce development time, provide a common platform for display development, and produce exceptional real-time results.

  5. [Porting Radiotherapy Software of Varian to Cloud Platform].

    PubMed

    Zou, Lian; Zhang, Weisha; Liu, Xiangxiang; Xie, Zhao; Xie, Yaoqin

    2017-09-30

    To develop a low-cost private cloud platform of radiotherapy software. First, a private cloud platform which was based on OpenStack and the virtual GPU hardware was builded. Then on the private cloud platform, all the Varian radiotherapy software modules were installed to the virtual machine, and the corresponding function configuration was completed. Finally the software on the cloud was able to be accessed by virtual desktop client. The function test results of the cloud workstation show that a cloud workstation is equivalent to an isolated physical workstation, and any clients on the LAN can use the cloud workstation smoothly. The cloud platform transplantation in this study is economical and practical. The project not only improves the utilization rates of radiotherapy software, but also makes it possible that the cloud computing technology can expand its applications to the field of radiation oncology.

  6. Long-term Preservation of Data Analysis Capabilities

    NASA Astrophysics Data System (ADS)

    Gabriel, C.; Arviset, C.; Ibarra, A.; Pollock, A.

    2015-09-01

    While the long-term preservation of scientific data obtained by large astrophysics missions is ensured through science archives, the issue of data analysis software preservation has hardly been addressed. Efforts by large data centres have contributed so far to maintain some instrument or mission-specific data reduction packages on top of high-level general purpose data analysis software. However, it is always difficult to keep software alive without support and maintenance once the active phase of a mission is over. This is especially difficult in the budgetary model followed by space agencies. We discuss the importance of extending the lifetime of dedicated data analysis packages and review diverse strategies under development at ESA using new paradigms such as Virtual Machines, Cloud Computing, and Software as a Service for making possible full availability of data analysis and calibration software for decades at minimal cost.

  7. Performance analysis of single stage libr-water absorption machine operated by waste thermal energy of internal combustion engine: Case study

    NASA Astrophysics Data System (ADS)

    Sharif, Hafiz Zafar; Leman, A. M.; Muthuraman, S.; Salleh, Mohd Najib Mohd; Zakaria, Supaat

    2017-09-01

    Combined heating, cooling, and power is also known as Tri-generation. Tri-generation system can provide power, hot water, space heating and air -conditioning from single source of energy. The objective of this study is to propose a method to evaluate the characteristic and performance of a single stage lithium bromide-water (LiBr-H2O) absorption machine operated with waste thermal energy of internal combustion engine which is integral part of trigeneration system. Correlations for computer sensitivity analysis are developed in data fit software for (P-T-X), (H-T-X), saturated liquid (water), saturated vapor, saturation pressure and crystallization temperature curve of LiBr-H2O Solution. Number of equations were developed with data fit software and exported into excel work sheet for the evaluation of number of parameter concerned with the performance of vapor absorption machine such as co-efficient of performance, concentration of solution, mass flow rate, size of heat exchangers of the unit in relation to the generator, condenser, absorber and evaporator temperatures. Size of vapor absorption machine within its crystallization limits for cooling and heating by waste energy recovered from exhaust gas, and jacket water of internal combustion engine also presented in this study to save the time and cost for the facilities managers who are interested to utilize the waste thermal energy of their buildings or premises for heating and air conditioning applications.

  8. A semi-automated process for the production of custom-made shoes

    NASA Technical Reports Server (NTRS)

    Farmer, Franklin H.

    1991-01-01

    A more efficient, cost-effective and timely way of designing and manufacturing custom footware is needed. A potential solution to this problem lies in the use of computer-aided design and manufacturing (CAD/CAM) techniques in the production of custom shoes. A prototype computer-based system was developed, and the system is primarily a software entity which directs and controls a 3-D scanner, a lathe or milling machine, and a pattern-cutting machine to produce the shoe last and the components to be assembled into a shoe. The steps in this process are: (1) scan the surface of the foot to obtain a 3-D image; (2) thin the foot surface data and create a tiled wire model of the foot; (3) interactively modify the wire model of the foot to produce a model of the shoe last; (4) machine the last; (5) scan the surface of the last and verify that it correctly represents the last model; (6) design cutting patterns for shoe uppers; (7) cut uppers; (8) machine an inverse mold for the shoe innersole/sole combination; (9) mold the innersole/sole; and (10) assemble the shoe. For all its capabilities, this system still requires the direction and assistance of skilled operators, and shoemakers to assemble the shoes. Currently, the system is running on a SUN3/260 workstation with TAAC application accelerator. The software elements of the system are written in either Fortran or C and run under a UNIX operator system.

  9. Parameter Estimation of the Thermal Network Model of a Machine Tool Spindle by Self-made Bluetooth Temperature Sensor Module

    PubMed Central

    Lo, Yuan-Chieh; Hu, Yuh-Chung; Chang, Pei-Zen

    2018-01-01

    Thermal characteristic analysis is essential for machine tool spindles because sudden failures may occur due to unexpected thermal issue. This article presents a lumped-parameter Thermal Network Model (TNM) and its parameter estimation scheme, including hardware and software, in order to characterize both the steady-state and transient thermal behavior of machine tool spindles. For the hardware, the authors develop a Bluetooth Temperature Sensor Module (BTSM) which accompanying with three types of temperature-sensing probes (magnetic, screw, and probe). Its specification, through experimental test, achieves to the precision ±(0.1 + 0.0029|t|) °C, resolution 0.00489 °C, power consumption 7 mW, and size Ø40 mm × 27 mm. For the software, the heat transfer characteristics of the machine tool spindle correlative to rotating speed are derived based on the theory of heat transfer and empirical formula. The predictive TNM of spindles was developed by grey-box estimation and experimental results. Even under such complicated operating conditions as various speeds and different initial conditions, the experiments validate that the present modeling methodology provides a robust and reliable tool for the temperature prediction with normalized mean square error of 99.5% agreement, and the present approach is transferable to the other spindles with a similar structure. For realizing the edge computing in smart manufacturing, a reduced-order TNM is constructed by Model Order Reduction (MOR) technique and implemented into the real-time embedded system. PMID:29473877

  10. Parameter Estimation of the Thermal Network Model of a Machine Tool Spindle by Self-made Bluetooth Temperature Sensor Module.

    PubMed

    Lo, Yuan-Chieh; Hu, Yuh-Chung; Chang, Pei-Zen

    2018-02-23

    Thermal characteristic analysis is essential for machine tool spindles because sudden failures may occur due to unexpected thermal issue. This article presents a lumped-parameter Thermal Network Model (TNM) and its parameter estimation scheme, including hardware and software, in order to characterize both the steady-state and transient thermal behavior of machine tool spindles. For the hardware, the authors develop a Bluetooth Temperature Sensor Module (BTSM) which accompanying with three types of temperature-sensing probes (magnetic, screw, and probe). Its specification, through experimental test, achieves to the precision ±(0.1 + 0.0029|t|) °C, resolution 0.00489 °C, power consumption 7 mW, and size Ø40 mm × 27 mm. For the software, the heat transfer characteristics of the machine tool spindle correlative to rotating speed are derived based on the theory of heat transfer and empirical formula. The predictive TNM of spindles was developed by grey-box estimation and experimental results. Even under such complicated operating conditions as various speeds and different initial conditions, the experiments validate that the present modeling methodology provides a robust and reliable tool for the temperature prediction with normalized mean square error of 99.5% agreement, and the present approach is transferable to the other spindles with a similar structure. For realizing the edge computing in smart manufacturing, a reduced-order TNM is constructed by Model Order Reduction (MOR) technique and implemented into the real-time embedded system.

  11. The Linux operating system: An introduction

    NASA Technical Reports Server (NTRS)

    Bokhari, Shahid H.

    1995-01-01

    Linux is a Unix-like operating system for Intel 386/486/Pentium based IBM-PCs and compatibles. The kernel of this operating system was written from scratch by Linus Torvalds and, although copyrighted by the author, may be freely distributed. A world-wide group has collaborated in developing Linux on the Internet. Linux can run the powerful set of compilers and programming tools of the Free Software Foundation, and XFree86, a port of the X Window System from MIT. Most capabilities associated with high performance workstations, such as networking, shared file systems, electronic mail, TeX, LaTeX, etc. are freely available for Linux. It can thus transform cheap IBM-PC compatible machines into Unix workstations with considerable capabilities. The author explains how Linux may be obtained, installed and networked. He also describes some interesting applications for Linux that are freely available. The enormous consumer market for IBM-PC compatible machines continually drives down prices of CPU chips, memory, hard disks, CDROMs, etc. Linux can convert such machines into powerful workstations that can be used for teaching, research and software development. For professionals who use Unix based workstations at work, Linux permits virtually identical working environments on their personal home machines. For cost conscious educational institutions Linux can create world-class computing environments from cheap, easily maintained, PC clones. Finally, for university students, it provides an essentially cost-free path away from DOS into the world of Unix and X Windows.

  12. DOE Office of Scientific and Technical Information (OSTI.GOV)

    Simpson, Wayne; Borders, Tammie

    INL successfully developed a proof of concept for "Software Defined Anything" by emulating the laboratory's business applications that run on Virtual Machines. The work INL conducted demonstrates to industry on how this methodology can be used to improve security, automate and repeat processes, and improve consistency.

  13. Mission and data operations IBM 360 user's guide

    NASA Technical Reports Server (NTRS)

    Balakirsky, J.

    1973-01-01

    The M and DO computer systems are introduced and supplemented. The hardware and software status is discussed, along with standard processors and user libraries. Data management techniques are presented, as well as machine independence, debugging facilities, and overlay considerations.

  14. 75 FR 65632 - Granting of Request for Early Termination of the Waiting Period Under the Premerger Notification...

    Federal Register 2010, 2011, 2012, 2013, 2014

    2010-10-26

    ... Fortify Software, Inc. 20101047 G International Business Machines Corporation. G Unica Corporation. G... Potash Corporation of Saskatchewan Inc. 20101124 G Taleo Corporation. G James Barrett Riley. G Learn.com...

  15. Data Publication and Interoperability for Long Tail Researchers via the Open Data Repository's (ODR) Data Publisher.

    NASA Astrophysics Data System (ADS)

    Stone, N.; Lafuente, B.; Bristow, T.; Keller, R.; Downs, R. T.; Blake, D. F.; Fonda, M.; Pires, A.

    2016-12-01

    Working primarily with astrobiology researchers at NASA Ames, the Open Data Repository (ODR) has been conducting a software pilot to meet the varying needs of this multidisciplinary community. Astrobiology researchers often have small communities or operate individually with unique data sets that don't easily fit into existing database structures. The ODR constructed its Data Publisher software to allow researchers to create databases with common metadata structures and subsequently extend them to meet their individual needs and data requirements. The software accomplishes these tasks through a web-based interface that allows collaborative creation and revision of common metadata templates and individual extensions to these templates for custom data sets. This allows researchers to search disparate datasets based on common metadata established through the metadata tools, but still facilitates distinct analyses and data that may be stored alongside the required common metadata. The software produces web pages that can be made publicly available at the researcher's discretion so that users may search and browse the data in an effort to make interoperability and data discovery a human-friendly task while also providing semantic data for machine-based discovery. Once relevant data has been identified, researchers can utilize the built-in application programming interface (API) that exposes the data for machine-based consumption and integration with existing data analysis tools (e.g. R, MATLAB, Project Jupyter - http://jupyter.org). The current evolution of the project has created the Astrobiology Habitable Environments Database (AHED)[1] which provides an interface to databases connected through a common metadata core. In the next project phase, the goal is for small research teams and groups to be self-sufficient in publishing their research data to meet funding mandates and academic requirements as well as fostering increased data discovery and interoperability through human-readable and machine-readable interfaces. This project is supported by the Science-Enabling Research Activity (SERA) and NASA NNX11AP82A, MSL. [1] B. Lafuente et al. (2016) AGU, submitted.

  16. Open access for ALICE analysis based on virtualization technology

    NASA Astrophysics Data System (ADS)

    Buncic, P.; Gheata, M.; Schutz, Y.

    2015-12-01

    Open access is one of the important leverages for long-term data preservation for a HEP experiment. To guarantee the usability of data analysis tools beyond the experiment lifetime it is crucial that third party users from the scientific community have access to the data and associated software. The ALICE Collaboration has developed a layer of lightweight components built on top of virtualization technology to hide the complexity and details of the experiment-specific software. Users can perform basic analysis tasks within CernVM, a lightweight generic virtual machine, paired with an ALICE specific contextualization. Once the virtual machine is launched, a graphical user interface is automatically started without any additional configuration. This interface allows downloading the base ALICE analysis software and running a set of ALICE analysis modules. Currently the available tools include fully documented tutorials for ALICE analysis, such as the measurement of strange particle production or the nuclear modification factor in Pb-Pb collisions. The interface can be easily extended to include an arbitrary number of additional analysis modules. We present the current status of the tools used by ALICE through the CERN open access portal, and the plans for future extensions of this system.

  17. [Prenatal risk calculation: comparison between Fast Screen pre I plus software and ViewPoint software. Evaluation of the risk calculation algorithms].

    PubMed

    Morin, Jean-François; Botton, Eléonore; Jacquemard, François; Richard-Gireme, Anouk

    2013-01-01

    The Fetal medicine foundation (FMF) has developed a new algorithm called Prenatal Risk Calculation (PRC) to evaluate Down syndrome screening based on free hCGβ, PAPP-A and nuchal translucency. The peculiarity of this algorithm is to use the degree of extremeness (DoE) instead of the multiple of the median (MoM). The biologists measuring maternal seric markers on Kryptor™ machines (Thermo Fisher Scientific) use Fast Screen pre I plus software for the prenatal risk calculation. This software integrates the PRC algorithm. Our study evaluates the data of 2.092 patient files of which 19 show a fœtal abnormality. These files have been first evaluated with the ViewPoint software based on MoM. The link between DoE and MoM has been analyzed and the different calculated risks compared. The study shows that Fast Screen pre I plus software gives the same risk results as ViewPoint software, but yields significantly fewer false positive results.

  18. B cells are not essential for Lactobacillus-mediated protection against lethal pneumovirus infection*

    PubMed Central

    Percopo, Caroline M.; Dyer, Kimberly D.; Garcia-Crespo, Katia E.; Gabryszewski, Stanislaw J.; Shaffer, Arthur L.; Domachowske, Joseph B.; Rosenberg, Helene F.

    2014-01-01

    We have shown previously that priming of respiratory mucosa with live Lactobacillus species promotes robust and prolonged survival from an otherwise lethal infection with pneumonia virus of mice (PVM), a property known as heterologous immunity. Lactobacillus-priming results in a moderate reduction in virus recovery and a dramatic reduction in virus-induced proinflammatory cytokine production; the precise mechanisms underlying these findings remain to be elucidated. As B cells have been shown to promote heterologous immunity against respiratory virus pathogens under similar conditions, here we explore the role of B cells in Lactobacillus-mediated protection against acute pneumovirus infection. We found that Lactobacillus-primed mice feature elevated levels of airway immunoglobulins IgG, IgA and IgM and lung tissues with dense, B cell (B220+) enriched peribronchial and perivascular infiltrates with germinal centers consistent with descriptions of bronchus-associated lymphoid tissue. No B cells were detected in lung tissue of Lactobacillus-primed B-cell deficient μMT mice or Jh mice, and Lactobacillus-primed μMT mice had no characteristic infiltrates or airway immunoglobulins. Nonetheless, we observed diminished virus recovery and profound suppression of virus-induced proinflammatory cytokines CCL2, IFN-gamma, and CXCL10 in both wild-type and Lactobacillus-primed μMT mice. Furthermore, L. plantarum-primed, B-cell deficient μMT and Jh mice were fully protected from an otherwise lethal PVM infection, as were their respective wild-types. We conclude that B cells are dispensable for Lactobacillus-mediated heterologous immunity and were not crucial for promoting survival in response to an otherwise lethal pneumovirus infection. PMID:24748495

  19. Study of the influence of the cutting temperature on the magnitude of the contact forces in the machining fixtures

    NASA Astrophysics Data System (ADS)

    Cioată, V. G.; Kiss, I.; Alexa, V.; Raţiu, S. A.; Racov, M.

    2018-01-01

    In the machining process, the workpieces are installed in machining fixtures in order to establish a strictly determined position with the cutting tool or its trajectory. During the cutting process, the weight of the workpiece, the forces and moments of inertia, cutting forces and moments, clamping forces, the heat released during the cutting process determine the contact forces between the locators and the workpiece. The magnitude of these forces is important because too large value can destroy the surface of the workpiece, and a too small value can cause the workpiece to slip on the locators or even the loss of the contact with the workpiece. Both situations must be avoided. The paper presents a study, realized with CAE software, regarding the influence of the cutting temperature on the magnitude of the contact forces in a machining fixture for the milling a rectangular workpiece.

  20. Machine learning of network metrics in ATLAS Distributed Data Management

    NASA Astrophysics Data System (ADS)

    Lassnig, Mario; Toler, Wesley; Vamosi, Ralf; Bogado, Joaquin; ATLAS Collaboration

    2017-10-01

    The increasing volume of physics data poses a critical challenge to the ATLAS experiment. In anticipation of high luminosity physics, automation of everyday data management tasks has become necessary. Previously many of these tasks required human decision-making and operation. Recent advances in hardware and software have made it possible to entrust more complicated duties to automated systems using models trained by machine learning algorithms. In this contribution we show results from one of our ongoing automation efforts that focuses on network metrics. First, we describe our machine learning framework built atop the ATLAS Analytics Platform. This framework can automatically extract and aggregate data, train models with various machine learning algorithms, and eventually score the resulting models and parameters. Second, we use these models to forecast metrics relevant for networkaware job scheduling and data brokering. We show the characteristics of the data and evaluate the forecasting accuracy of our models.

Top