Science.gov

Sample records for imaging workspace software

  1. Image Processing Software

    NASA Astrophysics Data System (ADS)

    Bosio, M. A.

    1990-11-01

    ABSTRACT: A brief description of astronomical image software is presented. This software was developed in a Digital Micro Vax II Computer System. : St presenta una somera descripci6n del software para procesamiento de imagenes. Este software fue desarrollado en un equipo Digital Micro Vax II. : DATA ANALYSIS - IMAGE PROCESSING

  2. Biological imaging software tools.

    PubMed

    Eliceiri, Kevin W; Berthold, Michael R; Goldberg, Ilya G; Ibáñez, Luis; Manjunath, B S; Martone, Maryann E; Murphy, Robert F; Peng, Hanchuan; Plant, Anne L; Roysam, Badrinath; Stuurman, Nico; Stuurmann, Nico; Swedlow, Jason R; Tomancak, Pavel; Carpenter, Anne E

    2012-06-28

    Few technologies are more widespread in modern biological laboratories than imaging. Recent advances in optical technologies and instrumentation are providing hitherto unimagined capabilities. Almost all these advances have required the development of software to enable the acquisition, management, analysis and visualization of the imaging data. We review each computational step that biologists encounter when dealing with digital images, the inherent challenges and the overall status of available software for bioimage informatics, focusing on open-source options.

  3. Biological Imaging Software Tools

    PubMed Central

    Eliceiri, Kevin W.; Berthold, Michael R.; Goldberg, Ilya G.; Ibáñez, Luis; Manjunath, B.S.; Martone, Maryann E.; Murphy, Robert F.; Peng, Hanchuan; Plant, Anne L.; Roysam, Badrinath; Stuurman, Nico; Swedlow, Jason R.; Tomancak, Pavel; Carpenter, Anne E.

    2013-01-01

    Few technologies are more widespread in modern biological laboratories than imaging. Recent advances in optical technologies and instrumentation are providing hitherto unimagined capabilities. Almost all these advances have required the development of software to enable the acquisition, management, analysis, and visualization of the imaging data. We review each computational step that biologists encounter when dealing with digital images, the challenges in that domain, and the overall status of available software for bioimage informatics, focusing on open source options. PMID:22743775

  4. SAPHIRE 8 Volume 5 - Workspaces

    SciTech Connect

    C. L. Smith; J. K. Knudsen; D. O'Neal

    2011-03-01

    The Systems Analysis Programs for Hands-on Integrated Reliability Evaluations (SAPHIRE) Version 8 is a software application developed for performing a complete probabilistic risk assessment using a personal computer running the Microsoft Windows™ operating system. SAPHIRE 8 is funded by the U.S. Nuclear Regulatory Commission (NRC). The role of the Idaho National Laboratory (INL) in this project is that of software developer and tester. In older versions of SAPHIRE, the model creation and analysis functions were intermingled. However, in SAPHIRE 8, the act of creating a model has been separated from the analysis of that model in order to improve the quality of both the model (e.g., by avoiding inadvertent changes) and the analysis. Consequently, in SAPHIRE 8, the analysis of models is performed by using what are called Workspaces. Currently, there are Workspaces for three types of analyses: (1) the NRC’s Accident Sequence Precursor program, where the workspace is called “Events and Condition Assessment (ECA);” (2) the NRC’s Significance Determination Process (SDP); and (3) the General Analysis (GA) workspace. Workspaces for each type are created and saved separately from the base model which keeps the original database intact. Workspaces are independent of each other and modifications or calculations made within one workspace will not affect another. In addition, each workspace has a user interface and reports tailored for their intended uses.

  5. Image Processing Software

    NASA Technical Reports Server (NTRS)

    1992-01-01

    To convert raw data into environmental products, the National Weather Service and other organizations use the Global 9000 image processing system marketed by Global Imaging, Inc. The company's GAE software package is an enhanced version of the TAE, developed by Goddard Space Flight Center to support remote sensing and image processing applications. The system can be operated in three modes and is combined with HP Apollo workstation hardware.

  6. Confined Space Imager (CSI) Software

    SciTech Connect

    Karelilz, David

    2013-07-03

    The software provides real-time image capture, enhancement, and display, and sensor control for the Confined Space Imager (CSI) sensor system The software captures images over a Cameralink connection and provides the following image enhancements: camera pixel to pixel non-uniformity correction, optical distortion correction, image registration and averaging, and illumination non-uniformity correction. The software communicates with the custom CSI hardware over USB to control sensor parameters and is capable of saving enhanced sensor images to an external USB drive. The software provides sensor control, image capture, enhancement, and display for the CSI sensor system. It is designed to work with the custom hardware.

  7. Flyover Animation of Phoenix Workspace

    NASA Technical Reports Server (NTRS)

    2008-01-01

    [figure removed for brevity, see original site] Click on image for animation

    This animated 'flyover' of the workspace of NASA's Phoenix Mars Lander's was created from images taken by the Surface Stereo Imager on Sol 14 (June 8, 2008), or the 14th Martian day after landing.

    The visualization uses both of the camera's 'eyes' to provide depth perception and ranging. The camera is looking north over the workspace.

    The Phoenix Mission is led by the University of Arizona, Tucson, on behalf of NASA. Project management of the mission is by NASA's Jet Propulsion Laboratory, Pasadena, Calif. Spacecraft development is by Lockheed Martin Space Systems, Denver.

  8. SOFT-1: Imaging Processing Software

    NASA Technical Reports Server (NTRS)

    1984-01-01

    Five levels of image processing software are enumerated and discussed: (1) logging and formatting; (2) radiometric correction; (3) correction for geometric camera distortion; (4) geometric/navigational corrections; and (5) general software tools. Specific concerns about access to and analysis of digital imaging data within the Planetary Data System are listed.

  9. Image processing software for imaging spectrometry

    NASA Technical Reports Server (NTRS)

    Mazer, Alan S.; Martin, Miki; Lee, Meemong; Solomon, Jerry E.

    1988-01-01

    The paper presents a software system, Spectral Analysis Manager (SPAM), which has been specifically designed and implemented to provide the exploratory analysis tools necessary for imaging spectrometer data, using only modest computational resources. The basic design objectives are described as well as the major algorithms designed or adapted for high-dimensional images. Included in a discussion of system implementation are interactive data display, statistical analysis, image segmentation and spectral matching, and mixture analysis.

  10. Spotlight-8 Image Analysis Software

    NASA Technical Reports Server (NTRS)

    Klimek, Robert; Wright, Ted

    2006-01-01

    Spotlight is a cross-platform GUI-based software package designed to perform image analysis on sequences of images generated by combustion and fluid physics experiments run in a microgravity environment. Spotlight can perform analysis on a single image in an interactive mode or perform analysis on a sequence of images in an automated fashion. Image processing operations can be employed to enhance the image before various statistics and measurement operations are performed. An arbitrarily large number of objects can be analyzed simultaneously with independent areas of interest. Spotlight saves results in a text file that can be imported into other programs for graphing or further analysis. Spotlight can be run on Microsoft Windows, Linux, and Apple OS X platforms.

  11. FITS Liberator: Image processing software

    NASA Astrophysics Data System (ADS)

    Lindberg Christensen, Lars; Nielsen, Lars Holm; Nielsen, Kaspar K.; Johansen, Teis; Hurt, Robert; de Martin, David

    2012-06-01

    The ESA/ESO/NASA FITS Liberator makes it possible to process and edit astronomical science data in the FITS format to produce stunning images of the universe. Formerly a plugin for Adobe Photoshop, the current version of FITS Liberator is a stand-alone application and no longer requires Photoshop. This image processing software makes it possible to create color images using raw observations from a range of telescopes; the FITS Liberator continues to support the FITS and PDS formats, preferred by astronomers and planetary scientists respectively, which enables data to be processed from a wide range of telescopes and planetary probes, including ESO's Very Large Telescope, the NASA/ESA Hubble Space Telescope, NASA's Spitzer Space Telescope, ESA's XMM-Newton Telescope and Cassini-Huygens or Mars Reconnaissance Orbiter.

  12. Automatic AVHRR image navigation software

    NASA Technical Reports Server (NTRS)

    Baldwin, Dan; Emery, William

    1992-01-01

    This is the final report describing the work done on the project entitled Automatic AVHRR Image Navigation Software funded through NASA-Washington, award NAGW-3224, Account 153-7529. At the onset of this project, we had developed image navigation software capable of producing geo-registered images from AVHRR data. The registrations were highly accurate but required a priori knowledge of the spacecraft's axes alignment deviations, commonly known as attitude. The three angles needed to describe the attitude are called roll, pitch, and yaw, and are the components of the deviations in the along scan, along track and about center directions. The inclusion of the attitude corrections in the navigation software results in highly accurate georegistrations, however, the computation of the angles is very tedious and involves human interpretation for several steps. The technique also requires easily identifiable ground features which may not be available due to cloud cover or for ocean data. The current project was motivated by the need for a navigation system which was automatic and did not require human intervention or ground control points. The first step in creating such a system must be the ability to parameterize the spacecraft's attitude. The immediate goal of this project was to study the attitude fluctuations and determine if they displayed any systematic behavior which could be modeled or parameterized. We chose a period in 1991-1992 to study the attitude of the NOAA 11 spacecraft using data from the Tiros receiving station at the Colorado Center for Astrodynamic Research (CCAR) at the University of Colorado.

  13. Phoenix Robotic Arm's Workspace After 90 Sols

    NASA Technical Reports Server (NTRS)

    2008-01-01

    During the first 90 Martian days, or sols, after its May 25, 2008, landing on an arctic plain of Mars, NASA's Phoenix Mars Lander dug several trenches in the workspace reachable with the lander's robotic arm.

    The lander's Surface Stereo Imager camera recorded this view of the workspace on Sol 90, early afternoon local Mars time (overnight Aug. 25 to Aug. 26, 2008). The shadow of the the camera itself, atop its mast, is just left of the center of the image and roughly a third of a meter (one foot) wide.

    The workspace is on the north side of the lander. The trench just to the right of center is called 'Neverland.'

    The Phoenix Mission is led by the University of Arizona, Tucson, on behalf of NASA. Project management of the mission is by NASA's Jet Propulsion Laboratory, Pasadena, Calif. Spacecraft development is by Lockheed Martin Space Systems, Denver.

  14. 3-D Visualization on Workspace of Parallel Manipulators

    NASA Astrophysics Data System (ADS)

    Tanaka, Yoshito; Yokomichi, Isao; Ishii, Junko; Makino, Toshiaki

    In parallel mechanisms, the form and volume of workspace also change variously with the attitude of a platform. This paper presents a method to search for the workspace of parallel mechanisms with 6-DOF and 3D visualization of the workspace. Workspace is a search for the movable range of the central point of a platform when it moves with a given orientation. In order to search workspace, geometric analysis based on inverse kinematics is considered. Plots of 2D of calculations are compared with those measured by position sensors. The test results are shown to have good agreement with simulation results. The workspace variations are demonstrated in terms of 3D and 2D plots for prototype mechanisms. The workspace plots are created with OpenGL and Visual C++ by implementation of the algorithm. An application module is developed, which displays workspace of the mechanism in 3D images. The effectiveness and practicability of 3D visualization on workspace are successfully demonstrated by 6-DOF parallel mechanisms.

  15. Easy and Accessible Imaging Software

    NASA Technical Reports Server (NTRS)

    2003-01-01

    DATASTAR, Inc., of Picayune, Mississippi, has taken NASA s award-winning Earth Resources Laboratory Applications Software (ELAS) program and evolved it into a user-friendly desktop application and Internet service to perform processing, analysis, and manipulation of remotely sensed imagery data. NASA s Stennis Space Center developed ELAS in the early 1980s to process satellite and airborne sensor imagery data of the Earth s surface into readable and accessible information. Since then, ELAS information has been applied worldwide to determine soil content, rainfall levels, and numerous other variances of topographical information. However, end-users customarily had to depend on scientific or computer experts to provide the results, because the imaging processing system was intricate and labor intensive.

  16. Software for Automated Image-to-Image Co-registration

    NASA Technical Reports Server (NTRS)

    Benkelman, Cody A.; Hughes, Heidi

    2007-01-01

    The project objectives are: a) Develop software to fine-tune image-to-image co-registration, presuming images are orthorectified prior to input; b) Create a reusable software development kit (SDK) to enable incorporation of these tools into other software; d) provide automated testing for quantitative analysis; and e) Develop software that applies multiple techniques to achieve subpixel precision in the co-registration of image pairs.

  17. Software For Computing Image Ratios

    NASA Technical Reports Server (NTRS)

    Yates, Gigi L.

    1993-01-01

    RATIO_TOOL is interactive computer program for viewing and analyzing large sets of multispectral image data created by imaging spectrometer. Uses ratios between intensities in different spectral bands in order to spot significant areas of interest within multispectral image. Each image band viewed iteratively, or selected image band of set of data requested and displayed. When image ratios computed, result displayed as grayscale image. Written in C Language.

  18. OSIRIX: open source multimodality image navigation software

    NASA Astrophysics Data System (ADS)

    Rosset, Antoine; Pysher, Lance; Spadola, Luca; Ratib, Osman

    2005-04-01

    The goal of our project is to develop a completely new software platform that will allow users to efficiently and conveniently navigate through large sets of multidimensional data without the need of high-end expensive hardware or software. We also elected to develop our system on new open source software libraries allowing other institutions and developers to contribute to this project. OsiriX is a free and open-source imaging software designed manipulate and visualize large sets of medical images: http://homepage.mac.com/rossetantoine/osirix/

  19. Identifying Image Manipulation Software from Image Features

    DTIC Science & Technology

    2015-03-26

    6 2.1 Mathematical Image Definition , Image Formats, and Interpolation Algorithms...6 2.1.1 Mathematical Image Definition ...Farid [9]. The following section defines the mathematical definition of an image, discusses two image formats, and three interpolation algo- rithms. 2.1

  20. Image analysis library software development

    NASA Technical Reports Server (NTRS)

    Guseman, L. F., Jr.; Bryant, J.

    1977-01-01

    The Image Analysis Library consists of a collection of general purpose mathematical/statistical routines and special purpose data analysis/pattern recognition routines basic to the development of image analysis techniques for support of current and future Earth Resources Programs. Work was done to provide a collection of computer routines and associated documentation which form a part of the Image Analysis Library.

  1. Infrared Imaging Data Reduction Software and Techniques

    NASA Astrophysics Data System (ADS)

    Sabbey, C. N.; McMahon, R. G.; Lewis, J. R.; Irwin, M. J.

    Developed to satisfy certain design requirements not met in existing packages (e.g., full weight map handling) and to optimize the software for large data sets (non-interactive tasks that are CPU and disk efficient), the InfraRed Data Reduction software package is a small ANSI C library of fast image processing routines for automated pipeline reduction of infrared (dithered) observations. The software includes stand-alone C programs for tasks such as running sky frame subtraction with object masking, image registration and co-addition with weight maps, dither offset measurement using cross-correlation, and object mask dilation. Although currently used for near-IR mosaic images, the modular software is concise and readily adaptable for reuse in other work. IRDR, available via anonymous ftp at ftp.ast.cam.ac.uk in pub/sabbey

  2. Software for Managing an Archive of Images

    NASA Technical Reports Server (NTRS)

    Hallai, Charles; Jones, Helene; Callac, Chris

    2003-01-01

    This is a revised draft by Innovators concerning the report on Software for Managing and Archive of Images.The SSC Multimedia Archive is an automated electronic system to manage images, acquired both by film and digital cameras, for the Public Affairs Office (PAO) at Stennis Space Center (SSC). Previously, the image archive was based on film photography and utilized a manual system that, by todays standards, had become inefficient and expensive. Now, the SSC Multimedia Archive, based on a server at SSC, contains both catalogs and images for pictures taken both digitally and with a traditional film-based camera, along with metadata about each image.

  3. Video Image Stabilization and Registration (VISAR) Software

    NASA Technical Reports Server (NTRS)

    1999-01-01

    Two scientists at NASA's Marshall Space Flight Center, atmospheric scientist Paul Meyer and solar physicist Dr. David Hathaway, developed promising new software, called Video Image Stabilization and Registration (VISAR), which is illustrated in this Quick Time movie. VISAR is a computer algorithm that stabilizes camera motion in the horizontal and vertical as well as rotation and zoom effects producing clearer images of moving objects, smoothes jagged edges, enhances still images, and reduces video noise or snow. It could steady images of ultrasounds, which are infamous for their grainy, blurred quality. VISAR could also have applications in law enforcement, medical, and meteorological imaging. The software can be used for defense application by improving reconnaissance video imagery made by military vehicles, aircraft, and ships traveling in harsh, rugged environments.

  4. Video Image Stabilization and Registration (VISAR) Software

    NASA Technical Reports Server (NTRS)

    1999-01-01

    Two scientists at NASA's Marshall Space Flight Center,atmospheric scientist Paul Meyer and solar physicist Dr. David Hathaway, developed promising new software, called Video Image stabilization and Registration (VISAR), which is illustrated in this Quick Time movie. VISAR is a computer algorithm that stabilizes camera motion in the horizontal and vertical as well as rotation and zoom effects producing clearer images of moving objects, smoothes jagged edges, enhances still images, and reduces video noise or snow. It could steady images of ultrasounds, which are infamous for their grainy, blurred quality. VISAR could also have applications in law enforcement, medical, and meteorological imaging. The software can be used for defense application by improving reconnaissance video imagery made by military vehicles, aircraft, and ships traveling in harsh, rugged environments.

  5. Video Image Stabilization and Registration (VISAR) Software

    NASA Technical Reports Server (NTRS)

    1999-01-01

    Two scientists at NASA's Marshall Space Flight Center,atmospheric scientist Paul Meyer and solar physicist Dr. David Hathaway, developed promising new software, called Video Image Stabilization and Registration (VISAR). VISAR may help law enforcement agencies catch criminals by improving the quality of video recorded at crime scenes. In this photograph, the single frame at left, taken at night, was brightened in order to enhance details and reduce noise or snow. To further overcome the video defects in one frame, Law enforcement officials can use VISAR software to add information from multiple frames to reveal a person. Images from less than a second of videotape were added together to create the clarified image at right. VISAR stabilizes camera motion in the horizontal and vertical as well as rotation and zoom effects producing clearer images of moving objects, smoothes jagged edges, enhances still images, and reduces video noise or snow. VISAR could also have applications in medical and meteorological imaging. It could steady images of ultrasounds, which are infamous for their grainy, blurred quality. The software can be used for defense application by improving recornaissance video imagery made by military vehicles, aircraft, and ships traveling in harsh, rugged environments.

  6. Video Image Stabilization and Registration (VISAR) Software

    NASA Technical Reports Server (NTRS)

    1999-01-01

    Two scientists at NASA Marshall Space Flight Center, atmospheric scientist Paul Meyer (left) and solar physicist Dr. David Hathaway, have developed promising new software, called Video Image Stabilization and Registration (VISAR), that may help law enforcement agencies to catch criminals by improving the quality of video recorded at crime scenes, VISAR stabilizes camera motion in the horizontal and vertical as well as rotation and zoom effects; produces clearer images of moving objects; smoothes jagged edges; enhances still images; and reduces video noise of snow. VISAR could also have applications in medical and meteorological imaging. It could steady images of Ultrasounds which are infamous for their grainy, blurred quality. It would be especially useful for tornadoes, tracking whirling objects and helping to determine the tornado's wind speed. This image shows two scientists reviewing an enhanced video image of a license plate taken from a moving automobile.

  7. Workspaces that move people.

    PubMed

    Waber, Ben; Magnolfi, Jennifer; Lindsay, Greg

    2014-10-01

    Few companies measure whether the design of their workspaces helps or hurts performance, but they should. The authors have collected data that capture individuals' interactions, communications, and location information. They've learned that face-to-face interactions are by far the most important activity in an office; creating chance encounters between knowledge workers, both inside and outside the organization, improves performance. The Norwegian telecom company Telenor was ahead of its time in 2003, when it incorporated "hot desking" (no assigned seats) and spaces that could easily be reconfigured for different tasks and evolving teams. The CEO credits the design of the offices with helping Telenor shift from a state-run monopoly to a competitive multinational carrier with 150 million subscribers. In another example, data collected at one pharmaceuticals company showed that when a salesperson increased interactions with coworkers on other teams by 10%, his or her sales increased by 10%. To get the sales staff running into colleagues from other departments, management shifted from one coffee machine for every six employees to one for every 120 and created a new large cafeteria for everyone. Sales rose by 20%, or $200 million, afterjust one quarter, quickly justifying the capital investment in the redesign.

  8. False Color Terrain Model of Phoenix Workspace

    NASA Technical Reports Server (NTRS)

    2008-01-01

    This is a terrain model of Phoenix's Robotic Arm workspace. It has been color coded by depth with a lander model for context. The model has been derived using images from the depth perception feature from Phoenix's Surface Stereo Imager (SSI). Red indicates low-lying areas that appear to be troughs. Blue indicates higher areas that appear to be polygons.

    The Phoenix Mission is led by the University of Arizona, Tucson, on behalf of NASA. Project management of the mission is by NASA's Jet Propulsion Laboratory, Pasadena, Calif. Spacecraft development is by Lockheed Martin Space Systems, Denver.

  9. Salvo: Seismic imaging software for complex geologies

    SciTech Connect

    OBER,CURTIS C.; GJERTSEN,ROB; WOMBLE,DAVID E.

    2000-03-01

    This report describes Salvo, a three-dimensional seismic-imaging software for complex geologies. Regions of complex geology, such as overthrusts and salt structures, can cause difficulties for many seismic-imaging algorithms used in production today. The paraxial wave equation and finite-difference methods used within Salvo can produce high-quality seismic images in these difficult regions. However this approach comes with higher computational costs which have been too expensive for standard production. Salvo uses improved numerical algorithms and methods, along with parallel computing, to produce high-quality images and to reduce the computational and the data input/output (I/O) costs. This report documents the numerical algorithms implemented for the paraxial wave equation, including absorbing boundary conditions, phase corrections, imaging conditions, phase encoding, and reduced-source migration. This report also describes I/O algorithms for large seismic data sets and images and parallelization methods used to obtain high efficiencies for both the computations and the I/O of seismic data sets. Finally, this report describes the required steps to compile, port and optimize the Salvo software, and describes the validation data sets used to help verify a working copy of Salvo.

  10. Objective facial photograph analysis using imaging software.

    PubMed

    Pham, Annette M; Tollefson, Travis T

    2010-05-01

    Facial analysis is an integral part of the surgical planning process. Clinical photography has long been an invaluable tool in the surgeon's practice not only for accurate facial analysis but also for enhancing communication between the patient and surgeon, for evaluating postoperative results, for medicolegal documentation, and for educational and teaching opportunities. From 35-mm slide film to the digital technology of today, clinical photography has benefited greatly from technological advances. With the development of computer imaging software, objective facial analysis becomes easier to perform and less time consuming. Thus, while the original purpose of facial analysis remains the same, the process becomes much more efficient and allows for some objectivity. Although clinical judgment and artistry of technique is never compromised, the ability to perform objective facial photograph analysis using imaging software may become the standard in facial plastic surgery practices in the future.

  11. Awake Animal Imaging Motion Tracking Software

    SciTech Connect

    Goddard, James

    2010-03-15

    The Awake Animal Motion Tracking Software code calculates the 3D movement of the head motion of a live, awake animal during a medical imaging scan. In conjunction with markers attached to the head, images acquired from multiple cameras are processed and marker locations precisely determined. Using off-line camera calibration data, the 3D positions of the markers are calculated along with a 6 degree of freedom position and orientation (pose) relative to a fixed initial position. This calculation is performed in real time at frame rates up to 30 frames per second. A time stamp with microsecond accuracy from a time base source is attached to each pose measurement.

  12. Imaging Sensor Flight and Test Equipment Software

    NASA Technical Reports Server (NTRS)

    Freestone, Kathleen; Simeone, Louis; Robertson, Byran; Frankford, Maytha; Trice, David; Wallace, Kevin; Wilkerson, DeLisa

    2007-01-01

    The Lightning Imaging Sensor (LIS) is one of the components onboard the Tropical Rainfall Measuring Mission (TRMM) satellite, and was designed to detect and locate lightning over the tropics. The LIS flight code was developed to run on a single onboard digital signal processor, and has operated the LIS instrument since 1997 when the TRMM satellite was launched. The software provides controller functions to the LIS Real-Time Event Processor (RTEP) and onboard heaters, collects the lightning event data from the RTEP, compresses and formats the data for downlink to the satellite, collects housekeeping data and formats the data for downlink to the satellite, provides command processing and interface to the spacecraft communications and data bus, and provides watchdog functions for error detection. The Special Test Equipment (STE) software was designed to operate specific test equipment used to support the LIS hardware through development, calibration, qualification, and integration with the TRMM spacecraft. The STE software provides the capability to control instrument activation, commanding (including both data formatting and user interfacing), data collection, decompression, and display and image simulation. The LIS STE code was developed for the DOS operating system in the C programming language. Because of the many unique data formats implemented by the flight instrument, the STE software was required to comprehend the same formats, and translate them for the test operator. The hardware interfaces to the LIS instrument using both commercial and custom computer boards, requiring that the STE code integrate this variety into a working system. In addition, the requirement to provide RTEP test capability dictated the need to provide simulations of background image data with short-duration lightning transients superimposed. This led to the development of unique code used to control the location, intensity, and variation above background for simulated lightning strikes

  13. Terahertz/mm wave imaging simulation software

    NASA Astrophysics Data System (ADS)

    Fetterman, M. R.; Dougherty, J.; Kiser, W. L., Jr.

    2006-10-01

    We have developed a mm wave/terahertz imaging simulation package from COTS graphic software and custom MATLAB code. In this scheme, a commercial ray-tracing package was used to simulate the emission and reflections of radiation from scenes incorporating highly realistic imagery. Accurate material properties were assigned to objects in the scenes, with values obtained from the literature, and from our own terahertz spectroscopy measurements. The images were then post-processed with custom Matlab code to include the blur introduced by the imaging system and noise levels arising from system electronics and detector noise. The Matlab code was also used to simulate the effect of fog, an important aspect for mm wave imaging systems. Several types of image scenes were evaluated, including bar targets, contrast detail targets, a person in a portal screening situation, and a sailboat on the open ocean. The images produced by this simulation are currently being used as guidance for a 94 GHz passive mm wave imaging system, but have broad applicability for frequencies extending into the terahertz region.

  14. ICER-3D Hyperspectral Image Compression Software

    NASA Technical Reports Server (NTRS)

    Xie, Hua; Kiely, Aaron; Klimesh, matthew; Aranki, Nazeeh

    2010-01-01

    Software has been developed to implement the ICER-3D algorithm. ICER-3D effects progressive, three-dimensional (3D), wavelet-based compression of hyperspectral images. If a compressed data stream is truncated, the progressive nature of the algorithm enables reconstruction of hyperspectral data at fidelity commensurate with the given data volume. The ICER-3D software is capable of providing either lossless or lossy compression, and incorporates an error-containment scheme to limit the effects of data loss during transmission. The compression algorithm, which was derived from the ICER image compression algorithm, includes wavelet-transform, context-modeling, and entropy coding subalgorithms. The 3D wavelet decomposition structure used by ICER-3D exploits correlations in all three dimensions of sets of hyperspectral image data, while facilitating elimination of spectral ringing artifacts, using a technique summarized in "Improving 3D Wavelet-Based Compression of Spectral Images" (NPO-41381), NASA Tech Briefs, Vol. 33, No. 3 (March 2009), page 7a. Correlation is further exploited by a context-modeling subalgorithm, which exploits spectral dependencies in the wavelet-transformed hyperspectral data, using an algorithm that is summarized in "Context Modeler for Wavelet Compression of Hyperspectral Images" (NPO-43239), which follows this article. An important feature of ICER-3D is a scheme for limiting the adverse effects of loss of data during transmission. In this scheme, as in the similar scheme used by ICER, the spatial-frequency domain is partitioned into rectangular error-containment regions. In ICER-3D, the partitions extend through all the wavelength bands. The data in each partition are compressed independently of those in the other partitions, so that loss or corruption of data from any partition does not affect the other partitions. Furthermore, because compression is progressive within each partition, when data are lost, any data from that partition received

  15. Image processing software for imaging spectrometry data analysis

    NASA Technical Reports Server (NTRS)

    Mazer, Alan; Martin, Miki; Lee, Meemong; Solomon, Jerry E.

    1988-01-01

    Imaging spectrometers simultaneously collect image data in hundreds of spectral channels, from the near-UV to the IR, and can thereby provide direct surface materials identification by means resembling laboratory reflectance spectroscopy. Attention is presently given to a software system, the Spectral Analysis Manager (SPAM) for the analysis of imaging spectrometer data. SPAM requires only modest computational resources and is composed of one main routine and a set of subroutine libraries. Additions and modifications are relatively easy, and special-purpose algorithms have been incorporated that are tailored to geological applications.

  16. Image processing software for imaging spectrometry data analysis

    NASA Astrophysics Data System (ADS)

    Mazer, Alan; Martin, Miki; Lee, Meemong; Solomon, Jerry E.

    1988-02-01

    Imaging spectrometers simultaneously collect image data in hundreds of spectral channels, from the near-UV to the IR, and can thereby provide direct surface materials identification by means resembling laboratory reflectance spectroscopy. Attention is presently given to a software system, the Spectral Analysis Manager (SPAM) for the analysis of imaging spectrometer data. SPAM requires only modest computational resources and is composed of one main routine and a set of subroutine libraries. Additions and modifications are relatively easy, and special-purpose algorithms have been incorporated that are tailored to geological applications.

  17. User questionnaire to evaluate the radiological workspace.

    PubMed

    van Ooijen, Peter M A; Koesoema, Allya P; Oudkerk, Matthijs

    2006-01-01

    Over the past few years, an increase in digitalization of radiology departments can be seen, which has a large impact on the work of the radiologists. This impact is not only demonstrated by the increased use of digital images but also by changing demands on the whole reading environment. In this study, we evaluated the satisfaction of our radiologists with our digital Picture Archival and Communication System environment and their workspace. This evaluation was performed by distribution of a questionnaire consisting of a score sheet and some open questions to all radiologists and residents. Out of 25 questionnaires, 12 were adequately answered and returned. Results clearly showed that most problems were present in the area of reading room design and layout and comfort and ergonomics. Based on the results from this study, adaptations were made and the results were also used in the planning of the redesign of the entire department of radiology.

  18. Workspaces in the Semantic Web

    NASA Technical Reports Server (NTRS)

    Wolfe, Shawn R.; Keller, RIchard M.

    2005-01-01

    Due to the recency and relatively limited adoption of Semantic Web technologies. practical issues related to technology scaling have received less attention than foundational issues. Nonetheless, these issues must be addressed if the Semantic Web is to realize its full potential. In particular, we concentrate on the lack of scoping methods that reduce the size of semantic information spaces so they are more efficient to work with and more relevant to an agent's needs. We provide some intuition to motivate the need for such reduced information spaces, called workspaces, give a formal definition, and suggest possible methods of deriving them.

  19. Digital image processing software system using an array processor

    SciTech Connect

    Sherwood, R.J.; Portnoff, M.R.; Journeay, C.H.; Twogood, R.E.

    1981-03-10

    A versatile array processor-based system for general-purpose image processing was developed. At the heart of this system is an extensive, flexible software package that incorporates the array processor for effective interactive image processing. The software system is described in detail, and its application to a diverse set of applications at LLNL is briefly discussed. 4 figures, 1 table.

  20. Perspective automated inkless fingerprinting imaging software for fingerprint research.

    PubMed

    Nanakorn, Somsong; Poosankam, Pongsakorn; Mongconthawornchai, Paiboon

    2008-01-01

    Fingerprint collection using ink-and-paper image is a conventional method i.e. an ink-print, transparent-adhesive tape techniques which are slower and cumbersome. This is a pilot research for software development aimed at imaging an automated, inkless fingerprint using a fingerprint sensor, a development kit of the IT WORKS Company Limited, PC camera, and printer The development of software was performed to connect with the fingerprint sensor for collection of fingerprint images and recorded into a hard disk. It was also developed to connect with the PC camera for recording a face image of persons' fingerprints or identification card images. These images had been appropriately arranged in a PDF file prior to printing. This software is able to scan ten fingerprints and store high-quality electronics fingertip images with rapid, large, and clear images without dirt of ink or carbon. This fingerprint technology is helpful in a potential application in public health and clinical medicine research.

  1. Survey of geographical information system and image processing software

    USGS Publications Warehouse

    Vanderzee, D.; Singh, A.

    1995-01-01

    The Global Resource Information Database—a part of the United Nations Environment Programme—conducts a bi-annual survey of geographical information system (GIS) and image processing (IP) software. This survey makes information about software products available in developing countries. The 1993 survey showed that the number of installations of GIS, IP, and related software products increased dramatically from 1991 to 1993, mostly in North America and Europe.

  2. Software Helps Extract Information From Astronomical Images

    NASA Technical Reports Server (NTRS)

    Hartley, Booth; Ebert, Rick; Laughlin, Gaylin

    1995-01-01

    PAC Skyview 2.0 is interactive program for display and analysis of astronomical images. Includes large set of functions for display, analysis and manipulation of images. "Man" pages with descriptions of functions and examples of usage included. Skyview used interactively or in "server" mode, in which another program calls Skyview and executes commands itself. Skyview capable of reading image data files of four types, including those in FITS, S, IRAF, and Z formats. Written in C.

  3. Earth Observation Services (Image Processing Software)

    NASA Technical Reports Server (NTRS)

    1992-01-01

    San Diego State University and Environmental Systems Research Institute, with other agencies, have applied satellite imaging and image processing techniques to geographic information systems (GIS) updating. The resulting images display land use and are used by a regional planning agency for applications like mapping vegetation distribution and preserving wildlife habitats. The EOCAP program provides government co-funding to encourage private investment in, and to broaden the use of NASA-developed technology for analyzing information about Earth and ocean resources.

  4. A Software Package For Biomedical Image Processing And Analysis

    NASA Astrophysics Data System (ADS)

    Goncalves, Joao G. M.; Mealha, Oscar

    1988-06-01

    The decreasing cost of computing power and the introduction of low cost imaging boards justifies the increasing number of applications of digital image processing techniques in the area of biomedicine. There is however a large software gap to be fulfilled, between the application and the equipment. The requirements to bridge this gap are twofold: good knowledge of the hardware provided and its interface to the host computer, and expertise in digital image processing and analysis techniques. A software package incorporating these two requirements was developped using the C programming language, in order to create a user friendly image processing programming environment. The software package can be considered in two different ways: as a data structure adapted to image processing and analysis, which acts as the backbone and the standard of communication for all the software; and as a set of routines implementing the basic algorithms used in image processing and analysis. Hardware dependency is restricted to a single module upon which all hardware calls are based. The data structure that was built has four main features: hierchical, open, object oriented, and object dependent dimensions. Considering the vast amount of memory needed by imaging applications and the memory available in small imaging systems, an effective image memory management scheme was implemented. This software package is being used for more than one and a half years by users with different applications. It proved to be an efficient tool for helping people to get adapted into the system, and for standardizing and exchanging software, yet preserving flexibility allowing for users' specific implementations. The philosophy of the software package is discussed and the data structure that was built is described in detail.

  5. Workspace Program for Complex-Number Arithmetic

    NASA Technical Reports Server (NTRS)

    Patrick, M. C.; Howell, Leonard W., Jr.

    1986-01-01

    COMPLEX is workspace program designed to empower APL with complexnumber capabilities. Complex-variable methods provide analytical tools invaluable for applications in mathematics, science, and engineering. COMPLEX written in APL.

  6. A Workspace-oriented Needle Guiding Robot for 3T MRI-guided Transperineal Prostate Intervention: Evaluation of In-bore Workspace and MRI Compatibility

    PubMed Central

    Song, Sang-Eun; Hata, Nobuhiko; Iordachita, Iulian; Fichtinger, Gabor; Tempany, Clare; Tokuda, Junichi

    2013-01-01

    Background Magnetic Resonance Imaging (MRI) guided prostate interventions have been introduced to enhance the cancer detection. For accurate needle positioning, in-bore operated robotic systems have been developed and optimal use of the confined in-bore space become a critical engineering challenge. Methods As preliminary evaluation of our prostate intervention robot, we conducted a workspace design analysis using a new evaluation method that we developed for in-bore operated robots for transperineal prostate interventions, and an MRI compatibility study. Results The workspace analysis resulted in the effective workspace (VW) of 0.32, which is greater than that of our early prototype despite that the current robot is approximately 50% larger than the early prototype in sectional space. The MRI compatibility study resulted in less than 15% signal-to-noise ratio (SNR) reduction. Conclusions The new workspace evaluation method quantifies the workspace utilization of the in-bore operated robots for MRI-guided transperineal prostate interventions, providing a useful tool for evaluation and new robot design. The robot creates insignificant electromagnetic noise during typical prostate imaging sequences. PMID:22492680

  7. Software for Acquiring Image Data for PIV

    NASA Technical Reports Server (NTRS)

    Wernet, Mark P.; Cheung, H. M.; Kressler, Brian

    2003-01-01

    PIV Acquisition (PIVACQ) is a computer program for acquisition of data for particle-image velocimetry (PIV). In the PIV system for which PIVACQ was developed, small particles entrained in a flow are illuminated with a sheet of light from a pulsed laser. The illuminated region is monitored by a charge-coupled-device camera that operates in conjunction with a data-acquisition system that includes a frame grabber and a counter-timer board, both installed in a single computer. The camera operates in "frame-straddle" mode where a pair of images can be obtained closely spaced in time (on the order of microseconds). The frame grabber acquires image data from the camera and stores the data in the computer memory. The counter/timer board triggers the camera and synchronizes the pulsing of the laser with acquisition of data from the camera. PIVPROC coordinates all of these functions and provides a graphical user interface, through which the user can control the PIV data-acquisition system. PIVACQ enables the user to acquire a sequence of single-exposure images, display the images, process the images, and then save the images to the computer hard drive. PIVACQ works in conjunction with the PIVPROC program which processes the images of particles into the velocity field in the illuminated plane.

  8. Development and implementation of software systems for imaging spectroscopy

    USGS Publications Warehouse

    Boardman, J.W.; Clark, R.N.; Mazer, A.S.; Biehl, L.L.; Kruse, F.A.; Torson, J.; Staenz, K.

    2006-01-01

    Specialized software systems have played a crucial role throughout the twenty-five year course of the development of the new technology of imaging spectroscopy, or hyperspectral remote sensing. By their very nature, hyperspectral data place unique and demanding requirements on the computer software used to visualize, analyze, process and interpret them. Often described as a marriage of the two technologies of reflectance spectroscopy and airborne/spaceborne remote sensing, imaging spectroscopy, in fact, produces data sets with unique qualities, unlike previous remote sensing or spectrometer data. Because of these unique spatial and spectral properties hyperspectral data are not readily processed or exploited with legacy software systems inherited from either of the two parent fields of study. This paper provides brief reviews of seven important software systems developed specifically for imaging spectroscopy.

  9. Software for Viewing Landsat Mosaic Images

    NASA Technical Reports Server (NTRS)

    Watts, Zack; Farve, Catharine L.; Harvey, Craig

    2003-01-01

    A Windows-based computer program has been written to enable novice users (especially educators and students) to view images of large areas of the Earth (e.g., the continental United States) generated from image data acquired in the Landsat observations performed circa the year 1990. The large-area images are constructed as mosaics from the original Landsat images, which were acquired in several wavelength bands and each of which spans an area (in effect, one tile of a mosaic) of .5 in latitude by .6 in longitude. Whereas the original Landsat data are registered on a universal transverse Mercator (UTM) grid, the program converts the UTM coordinates of a mouse pointer in the image to latitude and longitude, which are continuously updated and displayed as the pointer is moved. The mosaic image currently on display can be exported as a Windows bitmap file. Other images (e.g., of state boundaries or interstate highways) can be overlaid on Landsat mosaics. The program interacts with the user via standard toolbar, keyboard, and mouse user interfaces. The program is supplied on a compact disk along with tutorial and educational information.

  10. gr-MRI: A software package for magnetic resonance imaging using software defined radios

    NASA Astrophysics Data System (ADS)

    Hasselwander, Christopher J.; Cao, Zhipeng; Grissom, William A.

    2016-09-01

    The goal of this work is to develop software that enables the rapid implementation of custom MRI spectrometers using commercially-available software defined radios (SDRs). The developed gr-MRI software package comprises a set of Python scripts, flowgraphs, and signal generation and recording blocks for GNU Radio, an open-source SDR software package that is widely used in communications research. gr-MRI implements basic event sequencing functionality, and tools for system calibrations, multi-radio synchronization, and MR signal processing and image reconstruction. It includes four pulse sequences: a single-pulse sequence to record free induction signals, a gradient-recalled echo imaging sequence, a spin echo imaging sequence, and an inversion recovery spin echo imaging sequence. The sequences were used to perform phantom imaging scans with a 0.5 Tesla tabletop MRI scanner and two commercially-available SDRs. One SDR was used for RF excitation and reception, and the other for gradient pulse generation. The total SDR hardware cost was approximately 2000. The frequency of radio desynchronization events and the frequency with which the software recovered from those events was also measured, and the SDR's ability to generate frequency-swept RF waveforms was validated and compared to the scanner's commercial spectrometer. The spin echo images geometrically matched those acquired using the commercial spectrometer, with no unexpected distortions. Desynchronization events were more likely to occur at the very beginning of an imaging scan, but were nearly eliminated if the user invoked the sequence for a short period before beginning data recording. The SDR produced a 500 kHz bandwidth frequency-swept pulse with high fidelity, while the commercial spectrometer produced a waveform with large frequency spike errors. In conclusion, the developed gr-MRI software can be used to develop high-fidelity, low-cost custom MRI spectrometers using commercially-available SDRs.

  11. gr-MRI: A software package for magnetic resonance imaging using software defined radios.

    PubMed

    Hasselwander, Christopher J; Cao, Zhipeng; Grissom, William A

    2016-09-01

    The goal of this work is to develop software that enables the rapid implementation of custom MRI spectrometers using commercially-available software defined radios (SDRs). The developed gr-MRI software package comprises a set of Python scripts, flowgraphs, and signal generation and recording blocks for GNU Radio, an open-source SDR software package that is widely used in communications research. gr-MRI implements basic event sequencing functionality, and tools for system calibrations, multi-radio synchronization, and MR signal processing and image reconstruction. It includes four pulse sequences: a single-pulse sequence to record free induction signals, a gradient-recalled echo imaging sequence, a spin echo imaging sequence, and an inversion recovery spin echo imaging sequence. The sequences were used to perform phantom imaging scans with a 0.5Tesla tabletop MRI scanner and two commercially-available SDRs. One SDR was used for RF excitation and reception, and the other for gradient pulse generation. The total SDR hardware cost was approximately $2000. The frequency of radio desynchronization events and the frequency with which the software recovered from those events was also measured, and the SDR's ability to generate frequency-swept RF waveforms was validated and compared to the scanner's commercial spectrometer. The spin echo images geometrically matched those acquired using the commercial spectrometer, with no unexpected distortions. Desynchronization events were more likely to occur at the very beginning of an imaging scan, but were nearly eliminated if the user invoked the sequence for a short period before beginning data recording. The SDR produced a 500kHz bandwidth frequency-swept pulse with high fidelity, while the commercial spectrometer produced a waveform with large frequency spike errors. In conclusion, the developed gr-MRI software can be used to develop high-fidelity, low-cost custom MRI spectrometers using commercially-available SDRs.

  12. MOSAIC: Software for creating mosaics from collections of images

    NASA Technical Reports Server (NTRS)

    Varosi, F.; Gezari, D. Y.

    1992-01-01

    We have developed a powerful, versatile image processing and analysis software package called MOSAIC, designed specifically for the manipulation of digital astronomical image data obtained with (but not limited to) two-dimensional array detectors. The software package is implemented using the Interactive Data Language (IDL), and incorporates new methods for processing, calibration, analysis, and visualization of astronomical image data, stressing effective methods for the creation of mosaic images from collections of individual exposures, while at the same time preserving the photometric integrity of the original data. Since IDL is available on many computers, the MOSAIC software runs on most UNIX and VAX workstations with the X-Windows or Sun View graphics interface.

  13. A Global Workspace perspective on mental disorders

    PubMed Central

    Wallace, Rodrick

    2005-01-01

    Background Recent developments in Global Workspace theory suggest that human consciousness can suffer interpenetrating dysfunctions of mutual and reciprocal interaction with embedding environments which will have early onset and often insidious staged developmental progression, possibly according to a cancer model, in which a set of long-evolved control strategies progressively fails. Methods and results A rate distortion argument implies that, if an external information source carries a damaging 'message', then sufficient exposure to it, particularly during critical developmental periods, is sure to write a sufficiently accurate image of it on mind and body in a punctuated manner so as to initiate or promote similarly progressively punctuated developmental disorder, in essence either a staged failure affecting large-scale brain connectivity, which is the sine qua non of human consciousness, or else damaging the ability of embedding goal contexts to contain conscious dynamics. Conclusion The key intervention, at the population level, is clearly to limit exposure to factors triggering developmental disorders, a question of proper environmental sanitation, in a large sense, primarily a matter of social justice which has long been known to be determined almost entirely by the interactions of cultural trajectory, group power relations, and economic structure, with public policy. Intervention at the individual level appears limited to triggering or extending periods of remission, representing reestablishment of an extensive, but largely unexplored, spectrum of evolved control strategies, in contrast with the far better-understood case of cancer. PMID:16371149

  14. The influence of software filtering in digital mammography image quality

    NASA Astrophysics Data System (ADS)

    Michail, C.; Spyropoulou, V.; Kalyvas, N.; Valais, I.; Dimitropoulos, N.; Fountos, G.; Kandarakis, I.; Panayiotakis, G.

    2009-05-01

    Breast cancer is one of the most frequently diagnosed cancers among women. Several techniques have been developed to help in the early detection of breast cancer such as conventional and digital x-ray mammography, positron and single-photon emission mammography, etc. A key advantage in digital mammography is that images can be manipulated as simple computer image files. Thus non-dedicated commercially available image manipulation software can be employed to process and store the images. The image processing tools of the Photoshop (CS 2) software usually incorporate digital filters which may be used to reduce image noise, enhance contrast and increase spatial resolution. However, improving an image quality parameter may result in degradation of another. The aim of this work was to investigate the influence of three sharpening filters, named hereafter sharpen, sharpen more and sharpen edges on image resolution and noise. Image resolution was assessed by means of the Modulation Transfer Function (MTF).In conclusion it was found that the correct use of commercial non-dedicated software on digital mammograms may improve some aspects of image quality.

  15. Colonoscopy tutorial software made with a cadaver's sectioned images.

    PubMed

    Chung, Beom Sun; Chung, Min Suk; Park, Hyung Seon; Shin, Byeong-Seok; Kwon, Koojoo

    2016-11-01

    Novice doctors may watch tutorial videos in training for actual or computed tomographic (CT) colonoscopy. The conventional learning videos can be complemented by virtual colonoscopy software made with a cadaver's sectioned images (SIs). The objective of this study was to assist colonoscopy trainees with the new interactive software. Submucosal segmentation on the SIs was carried out through the whole length of the large intestine. With the SIs and segmented images, a three dimensional model was reconstructed. Six-hundred seventy-one proximal colonoscopic views (conventional views) and corresponding distal colonoscopic views (simulating the retroflexion of a colonoscope) were produced. Not only navigation views showing the current location of the colonoscope tip and its course, but also, supplementary description views were elaborated. The four corresponding views were put into convenient browsing software to be downloaded free from the homepage (anatomy.co.kr). The SI colonoscopy software with the realistic images and supportive tools was available to anybody. Users could readily notice the position and direction of the virtual colonoscope tip and recognize meaningful structures in colonoscopic views. The software is expected to be an auxiliary learning tool to improve technique and related knowledge in actual and CT colonoscopies. Hopefully, the software will be updated using raw images from the Visible Korean project.

  16. Uses of software in digital image analysis: a forensic report

    NASA Astrophysics Data System (ADS)

    Sharma, Mukesh; Jha, Shailendra

    2010-02-01

    Forensic image analysis is required an expertise to interpret the content of an image or the image itself in legal matters. Major sub-disciplines of forensic image analysis with law enforcement applications include photo-grammetry, photographic comparison, content analysis and image authentication. It has wide applications in forensic science range from documenting crime scenes to enhancing faint or indistinct patterns such as partial fingerprints. The process of forensic image analysis can involve several different tasks, regardless of the type of image analysis performed. Through this paper authors have tried to explain these tasks, which are described in to three categories: Image Compression, Image Enhancement & Restoration and Measurement Extraction. With the help of examples like signature comparison, counterfeit currency comparison and foot-wear sole impression using the software Canvas and Corel Draw.

  17. Accuracy of 3D Imaging Software in Cephalometric Analysis

    DTIC Science & Technology

    2013-06-21

    Imaging and Communication in Medicine ( DICOM ) files into personal computer-based software to enable 3D reconstruction of the craniofacial skeleton. These...tissue profile. CBCT data can be imported as DICOM files into personal computer–based software to provide 3D reconstruction of the craniofacial...been acquired for the three pig models. The CBCT data were exported into DICOM multi-file format. They will be imported into a proprietary

  18. Single-molecule localization software applied to photon counting imaging.

    PubMed

    Hirvonen, Liisa M; Kilfeather, Tiffany; Suhling, Klaus

    2015-06-01

    Centroiding in photon counting imaging has traditionally been accomplished by a single-step, noniterative algorithm, often implemented in hardware. Single-molecule localization techniques in superresolution fluorescence microscopy are conceptually similar, but use more sophisticated iterative software-based fitting algorithms to localize the fluorophore. Here, we discuss common features and differences between single-molecule localization and photon counting imaging and investigate the suitability of single-molecule localization software for photon event localization. We find that single-molecule localization software packages designed for superresolution microscopy-QuickPALM, rapidSTORM, and ThunderSTORM-can work well when applied to photon counting imaging with a microchannel-plate-based intensified camera system: photon event recognition can be excellent, fixed pattern noise can be low, and the microchannel plate pores can easily be resolved.

  19. Modified control software for imaging ultracold atomic clouds

    SciTech Connect

    Whitaker, D. L.; Sharma, A.; Brown, J. M.

    2006-12-15

    A charge-coupled device (CCD) camera capable of taking high-quality images of ultracold atomic samples can often represent a significant portion of the equipment costs in atom trapping experiment. We have modified the commercial control software of a CCD camera designed for astronomical imaging to take absorption images of ultracold rubidium clouds. This camera is sensitive at 780 nm and has been modified to take three successive 16-bit images at full resolution. The control software can be integrated into a Matlab graphical user interface with fitting routines written as Matlab functions. This camera is capable of recording high-quality images at a fraction of the cost of similar cameras typically used in atom trapping experiments.

  20. Computer Software Configuration Item-Specific Flight Software Image Transfer Script Generator

    NASA Technical Reports Server (NTRS)

    Bolen, Kenny; Greenlaw, Ronald

    2010-01-01

    A K-shell UNIX script enables the International Space Station (ISS) Flight Control Team (FCT) operators in NASA s Mission Control Center (MCC) in Houston to transfer an entire or partial computer software configuration item (CSCI) from a flight software compact disk (CD) to the onboard Portable Computer System (PCS). The tool is designed to read the content stored on a flight software CD and generate individual CSCI transfer scripts that are capable of transferring the flight software content in a given subdirectory on the CD to the scratch directory on the PCS. The flight control team can then transfer the flight software from the PCS scratch directory to the Electronically Erasable Programmable Read Only Memory (EEPROM) of an ISS Multiplexer/ Demultiplexer (MDM) via the Indirect File Transfer capability. The individual CSCI scripts and the CSCI Specific Flight Software Image Transfer Script Generator (CFITSG), when executed a second time, will remove all components from their original execution. The tool will identify errors in the transfer process and create logs of the transferred software for the purposes of configuration management.

  1. Software Development for Ring Imaging Detector

    NASA Astrophysics Data System (ADS)

    Torisky, Benjamin

    2016-03-01

    Jefferson Lab (Jlab) is performing a large-scale upgrade to their Continuous Electron Beam Accelerator Facility (CEBAF) up to 12GeV beam. The Large Acceptance Spectrometer (CLAS12) in Hall B is being upgraded and a new Ring Imaging Cherenkov (RICH) detector is being developed to provide better kaon - pion separation throughout the 3 to 12 GeV range. With this addition, when the electron beam hits the target, the resulting pions, kaons, and other particles will pass through a wall of translucent aerogel tiles and create Cherenkov radiation. This light can then be accurately detected by a large array of Multi-Anode PhotoMultiplier Tubes (MA-PMT). I am presenting an update on my work on the implementation of Java based reconstruction programs for the RICH in the CLAS12 main analysis package.

  2. Software to model AXAF image quality

    NASA Technical Reports Server (NTRS)

    Ahmad, Anees

    1993-01-01

    This draft final report describes the work performed under this delivery order from May 1992 through June 1993. The purpose of this contract was to enhance and develop an integrated optical performance modeling software for complex x-ray optical systems such as AXAF. The GRAZTRACE program developed by the MSFC Optical Systems Branch for modeling VETA-I was used as the starting baseline program. The original program was a large single file program and, therefore, could not be modified very efficiently. The original source code has been reorganized, and a 'Make Utility' has been written to update the original program. The new version of the source code consists of 36 small source files to make it easier for the code developer to manage and modify the program. A user library has also been built and a 'Makelib' utility has been furnished to update the library. With the user library, the users can easily access the GRAZTRACE source files and build a custom library. A user manual for the new version of GRAZTRACE has been compiled. The plotting capability for the 3-D point spread functions and contour plots has been provided in the GRAZTRACE using the graphics package DISPLAY. The Graphics emulator over the network has been set up for programming the graphics routine. The point spread function and the contour plot routines have also been modified to display the plot centroid, and to allow the user to specify the plot range, and the viewing angle options. A Command Mode version of GRAZTRACE has also been developed. More than 60 commands have been implemented in a Code-V like format. The functions covered in this version include data manipulation, performance evaluation, and inquiry and setting of internal parameters. The user manual for these commands has been formatted as in Code-V, showing the command syntax, synopsis, and options. An interactive on-line help system for the command mode has also been accomplished to allow the user to find valid commands, command syntax

  3. Metamodels for Planar 3R Workspace Optimization.

    SciTech Connect

    Turner, C. J.

    2002-01-01

    Robotic workspace optimization is a central element of robot system design. To formulate the optimization problem, the complex relationships between design variables, tuning parameters, and performance indices need to be accurately and efficiently represented. The nature of the relationships suggests that metamodels, or models of the models, should be used to derive suitable objective functions. A comparison of two metamodeling techniques for robotic workspace optimization problems for several trial cases suggests that non-uniform rational B-spline models, derived from computer graphics and computer-aided design techniques, are as suitable as response surface models to solve planar 3R workspace optimization problems. Promising nonlinear modeling results with B-spline models suggest future work is justified and performance gains can be realized.

  4. Non-Imaging Software/Data Analysis Requirements

    NASA Technical Reports Server (NTRS)

    1984-01-01

    The analysis software needs of the non-imaging planetary data user are discussed. Assumptions as to the nature of the planetary science data centers where the data are physically stored are advanced, the scope of the non-imaging data is outlined, and facilities that users are likely to need to define and access data are identified. Data manipulation and analysis needs and display graphics are discussed.

  5. Software to model AXAF-I image quality

    NASA Technical Reports Server (NTRS)

    Ahmad, Anees; Feng, Chen

    1995-01-01

    A modular user-friendly computer program for the modeling of grazing-incidence type x-ray optical systems has been developed. This comprehensive computer software GRAZTRACE covers the manipulation of input data, ray tracing with reflectivity and surface deformation effects, convolution with x-ray source shape, and x-ray scattering. The program also includes the capabilities for image analysis, detector scan modeling, and graphical presentation of the results. A number of utilities have been developed to interface the predicted Advanced X-ray Astrophysics Facility-Imaging (AXAF-I) mirror structural and thermal distortions with the ray-trace. This software is written in FORTRAN 77 and runs on a SUN/SPARC station. An interactive command mode version and a batch mode version of the software have been developed.

  6. Stromatoporoid biometrics using image analysis software: A first order approach

    NASA Astrophysics Data System (ADS)

    Wolniewicz, Pawel

    2010-04-01

    Strommetric is a new image analysis computer program that performs morphometric measurements of stromatoporoid sponges. The program measures 15 features of skeletal elements (pillars and laminae) visible in both longitudinal and transverse thin sections. The software is implemented in C++, using the Open Computer Vision (OpenCV) library. The image analysis system distinguishes skeletal elements from sparry calcite using Otsu's method for image thresholding. More than 150 photos of thin sections were used as a test set, from which 36,159 measurements were obtained. The software provided about one hundred times more data than the current method applied until now. The data obtained are reproducible, even if the work is repeated by different workers. Thus the method makes the biometric studies of stromatoporoids objective.

  7. Data to Pictures to Data: Outreach Imaging Software and Metadata

    NASA Astrophysics Data System (ADS)

    Levay, Z.

    2011-07-01

    A convergence between astronomy science and digital photography has enabled a steady stream of visually rich imagery from state-of-the-art data. The accessibility of hardware and software has facilitated an explosion of astronomical images for outreach, from space-based observatories, ground-based professional facilities and among the vibrant amateur astrophotography community. Producing imagery from science data involves a combination of custom software to understand FITS data (FITS Liberator), off-the-shelf, industry-standard software to composite multi-wavelength data and edit digital photographs (Adobe Photoshop), and application of photo/image-processing techniques. Some additional effort is needed to close the loop and enable this imagery to be conveniently available for various purposes beyond web and print publication. The metadata paradigms in digital photography are now complying with FITS and science software to carry information such as keyword tags and world coordinates, enabling these images to be usable in more sophisticated, imaginative ways exemplified by Sky in Google Earth and World Wide Telescope.

  8. Pattern Recognition Software and Techniques for Biological Image Analysis

    PubMed Central

    Shamir, Lior; Delaney, John D.; Orlov, Nikita; Eckley, D. Mark; Goldberg, Ilya G.

    2010-01-01

    The increasing prevalence of automated image acquisition systems is enabling new types of microscopy experiments that generate large image datasets. However, there is a perceived lack of robust image analysis systems required to process these diverse datasets. Most automated image analysis systems are tailored for specific types of microscopy, contrast methods, probes, and even cell types. This imposes significant constraints on experimental design, limiting their application to the narrow set of imaging methods for which they were designed. One of the approaches to address these limitations is pattern recognition, which was originally developed for remote sensing, and is increasingly being applied to the biology domain. This approach relies on training a computer to recognize patterns in images rather than developing algorithms or tuning parameters for specific image processing tasks. The generality of this approach promises to enable data mining in extensive image repositories, and provide objective and quantitative imaging assays for routine use. Here, we provide a brief overview of the technologies behind pattern recognition and its use in computer vision for biological and biomedical imaging. We list available software tools that can be used by biologists and suggest practical experimental considerations to make the best use of pattern recognition techniques for imaging assays. PMID:21124870

  9. Pattern recognition software and techniques for biological image analysis.

    PubMed

    Shamir, Lior; Delaney, John D; Orlov, Nikita; Eckley, D Mark; Goldberg, Ilya G

    2010-11-24

    The increasing prevalence of automated image acquisition systems is enabling new types of microscopy experiments that generate large image datasets. However, there is a perceived lack of robust image analysis systems required to process these diverse datasets. Most automated image analysis systems are tailored for specific types of microscopy, contrast methods, probes, and even cell types. This imposes significant constraints on experimental design, limiting their application to the narrow set of imaging methods for which they were designed. One of the approaches to address these limitations is pattern recognition, which was originally developed for remote sensing, and is increasingly being applied to the biology domain. This approach relies on training a computer to recognize patterns in images rather than developing algorithms or tuning parameters for specific image processing tasks. The generality of this approach promises to enable data mining in extensive image repositories, and provide objective and quantitative imaging assays for routine use. Here, we provide a brief overview of the technologies behind pattern recognition and its use in computer vision for biological and biomedical imaging. We list available software tools that can be used by biologists and suggest practical experimental considerations to make the best use of pattern recognition techniques for imaging assays.

  10. Image Fusion Software in the Clearpem-Sonic Project

    NASA Astrophysics Data System (ADS)

    Pizzichemi, M.; di Vara, N.; Cucciati, G.; Ghezzi, A.; Paganoni, M.; Farina, F.; Frisch, B.; Bugalho, R.

    2012-08-01

    ClearPEM-Sonic is a mammography scanner that combines Positron Emission Tomography with 3D ultrasound echographic and elastographic imaging. It has been developed to improve early stage detection of breast cancer by combining metabolic and anatomical information. The PET system has been developed by the Crystal Clear Collaboration, while the 3D ultrasound probe has been provided by SuperSonic Imagine. In this framework, the visualization and fusion software is an essential tool for the radiologists in the diagnostic process. This contribution discusses the design choices, the issues faced during the implementation, and the commissioning of the software tools developed for ClearPEM-Sonic.

  11. Current and future trends in marine image annotation software

    NASA Astrophysics Data System (ADS)

    Gomes-Pereira, Jose Nuno; Auger, Vincent; Beisiegel, Kolja; Benjamin, Robert; Bergmann, Melanie; Bowden, David; Buhl-Mortensen, Pal; De Leo, Fabio C.; Dionísio, Gisela; Durden, Jennifer M.; Edwards, Luke; Friedman, Ariell; Greinert, Jens; Jacobsen-Stout, Nancy; Lerner, Steve; Leslie, Murray; Nattkemper, Tim W.; Sameoto, Jessica A.; Schoening, Timm; Schouten, Ronald; Seager, James; Singh, Hanumant; Soubigou, Olivier; Tojeira, Inês; van den Beld, Inge; Dias, Frederico; Tempera, Fernando; Santos, Ricardo S.

    2016-12-01

    Given the need to describe, analyze and index large quantities of marine imagery data for exploration and monitoring activities, a range of specialized image annotation tools have been developed worldwide. Image annotation - the process of transposing objects or events represented in a video or still image to the semantic level, may involve human interactions and computer-assisted solutions. Marine image annotation software (MIAS) have enabled over 500 publications to date. We review the functioning, application trends and developments, by comparing general and advanced features of 23 different tools utilized in underwater image analysis. MIAS requiring human input are basically a graphical user interface, with a video player or image browser that recognizes a specific time code or image code, allowing to log events in a time-stamped (and/or geo-referenced) manner. MIAS differ from similar software by the capability of integrating data associated to video collection, the most simple being the position coordinates of the video recording platform. MIAS have three main characteristics: annotating events in real time, posteriorly to annotation and interact with a database. These range from simple annotation interfaces, to full onboard data management systems, with a variety of toolboxes. Advanced packages allow to input and display data from multiple sensors or multiple annotators via intranet or internet. Posterior human-mediated annotation often include tools for data display and image analysis, e.g. length, area, image segmentation, point count; and in a few cases the possibility of browsing and editing previous dive logs or to analyze the annotations. The interaction with a database allows the automatic integration of annotations from different surveys, repeated annotation and collaborative annotation of shared datasets, browsing and querying of data. Progress in the field of automated annotation is mostly in post processing, for stable platforms or still images

  12. The application of image processing software: Photoshop in environmental design

    NASA Astrophysics Data System (ADS)

    Dong, Baohua; Zhang, Chunmi; Zhuo, Chen

    2011-02-01

    In the process of environmental design and creation, the design sketch holds a very important position in that it not only illuminates the design's idea and concept but also shows the design's visual effects to the client. In the field of environmental design, computer aided design has made significant improvement. Many types of specialized design software for environmental performance of the drawings and post artistic processing have been implemented. Additionally, with the use of this software, working efficiency has greatly increased and drawings have become more specific and more specialized. By analyzing the application of photoshop image processing software in environmental design and comparing and contrasting traditional hand drawing and drawing with modern technology, this essay will further explore the way for computer technology to play a bigger role in environmental design.

  13. Determining Angle of Humeral Torsion Using Image Software Technique

    PubMed Central

    Sethi, Madhu; Vasudeva, Neelam

    2016-01-01

    Introduction Several researches have been done on the measurement of angles of humeral torsion in different parts of the world. Previously described methods were more complicated, not much accurate, cumbersome or required sophisticated instruments. Aim The present study was conducted with the aim to determine the angles of humeral torsion with a newer simple technique using digital images and image tool software. Materials and Methods A total of 250 dry normal adult human humeri were obtained from the bone bank of Department of Anatomy. The length and mid-shaft circumference of each bone was measured with the help of measuring tape. The angle of humeral torsion was measured directly from the digital images by the image analysis using Image Tool 3.0 software program. The data was analysed statistically with SPSS version 17 using unpaired t-test and Spearman’s rank order correlation coefficient. Results The mean angle of torsion was 64.57°±7.56°. On the right side it was 66.84°±9.69°, whereas, on the left side it was found to be 63.31±9.50°. The mean humeral length was 31.6 cm on right side and 30.33 cm on left side. Mid shaft circumference was 5.79 on right side and 5.63 cm on left side. No statistical differences were seen in angles between right and left humeri (p>0.001). Conclusion From our study, it was concluded that circumference of shaft is inversely proportional to angle of humeral torsion. The length and side of humerus has no relation with the humeral torsion. With advancement of digital technology, it is better to use new image softwares for anatomical studies. PMID:27891326

  14. Software for visualization, analysis, and manipulation of laser scan images

    NASA Astrophysics Data System (ADS)

    Burnsides, Dennis B.

    1997-03-01

    The recent introduction of laser surface scanning to scientific applications presents a challenge to computer scientists and engineers. Full utilization of this two- dimensional (2-D) and three-dimensional (3-D) data requires advances in techniques and methods for data processing and visualization. This paper explores the development of software to support the visualization, analysis and manipulation of laser scan images. Specific examples presented are from on-going efforts at the Air Force Computerized Anthropometric Research and Design (CARD) Laboratory.

  15. Software phantom for the synthesis of equilibrium radionuclide ventriculography images.

    PubMed

    Ruiz-de-Jesus, Oscar; Yanez-Suarez, Oscar; Jimenez-Angeles, Luis; Vallejo-Venegas, Enrique

    2006-01-01

    This paper presents the novel design of a software phantom for the evaluation of equilibrium radionuclide ventriculography systems. Through singular value decomposition, the data matrix corresponding to an equilibrium image series is decomposed into both spatial and temporal fundamental components that can be parametrized. This parametric model allows for the application of user-controlled conditions related to a desired dynamic behavior. Being invertible, the decomposition is used to regenerate the radionuclide image series, which is then translated into a DICOM ventriculography file that can be read by commercial equipment.

  16. Sungrabber - Software for Measurements on Solar Synoptic Images

    NASA Astrophysics Data System (ADS)

    Hržina, D.; Roša, D.; Hanslmeier, A.; Ruždjak, V.; Brajša, R.

    Measurement of positions of the tracers on synoptic solar images and conversion to heliographic coordinates is a time-consuming procedure with different sources of errors. To make measurements faster and easier, the application "Sungrabber" was developed. The data of the measured heliographic coordinates are stored in text files which are linked to the related solar images, which allows also a fast and simple comparison of the measurements from different sources. Extension of the software is possible and therefore Sungrabber can be used for different purposes (e.g. determining the solar rotation rate, proper motions of the tracers on the Sun, etc.).

  17. Parallel-Processing Software for Creating Mosaic Images

    NASA Technical Reports Server (NTRS)

    Klimeck, Gerhard; Deen, Robert; McCauley, Michael; DeJong, Eric

    2008-01-01

    A computer program implements parallel processing for nearly real-time creation of panoramic mosaics of images of terrain acquired by video cameras on an exploratory robotic vehicle (e.g., a Mars rover). Because the original images are typically acquired at various camera positions and orientations, it is necessary to warp the images into the reference frame of the mosaic before stitching them together to create the mosaic. [Also see "Parallel-Processing Software for Correlating Stereo Images," Software Supplement to NASA Tech Briefs, Vol. 31, No. 9 (September 2007) page 26.] The warping algorithm in this computer program reflects the considerations that (1) for every pixel in the desired final mosaic, a good corresponding point must be found in one or more of the original images and (2) for this purpose, one needs a good mathematical model of the cameras and a good correlation of individual pixels with respect to their positions in three dimensions. The desired mosaic is divided into slices, each of which is assigned to one of a number of central processing units (CPUs) operating simultaneously. The results from the CPUs are gathered and placed into the final mosaic. The time taken to create the mosaic depends upon the number of CPUs, the speed of each CPU, and whether a local or a remote data-staging mechanism is used.

  18. NASA's MERBoard: An Interactive Collaborative Workspace Platform. Chapter 4

    NASA Technical Reports Server (NTRS)

    Trimble, Jay; Wales, Roxana; Gossweiler, Rich

    2003-01-01

    This chapter describes the ongoing process by which a multidisciplinary group at NASA's Ames Research Center is designing and implementing a large interactive work surface called the MERBoard Collaborative Workspace. A MERBoard system involves several distributed, large, touch-enabled, plasma display systems with custom MERBoard software. A centralized server and database back the system. We are continually tuning MERBoard to support over two hundred scientists and engineers during the surface operations of the Mars Exploration Rover Missions. These scientists and engineers come from various disciplines and are working both in small and large groups over a span of space and time. We describe the multidisciplinary, human-centered process by which this h4ERBoard system is being designed, the usage patterns and social interactions that we have observed, and issues we are currently facing.

  19. Software components for medical image visualization and surgical planning

    NASA Astrophysics Data System (ADS)

    Starreveld, Yves P.; Gobbi, David G.; Finnis, Kirk; Peters, Terence M.

    2001-05-01

    Purpose: The development of new applications in medical image visualization and surgical planning requires the completion of many common tasks such as image reading and re-sampling, segmentation, volume rendering, and surface display. Intra-operative use requires an interface to a tracking system and image registration, and the application requires basic, easy to understand user interface components. Rapid changes in computer and end-application hardware, as well as in operating systems and network environments make it desirable to have a hardware and operating system as an independent collection of reusable software components that can be assembled rapidly to prototype new applications. Methods: Using the OpenGL based Visualization Toolkit as a base, we have developed a set of components that implement the above mentioned tasks. The components are written in both C++ and Python, but all are accessible from Python, a byte compiled scripting language. The components have been used on the Red Hat Linux, Silicon Graphics Iris, Microsoft Windows, and Apple OS X platforms. Rigorous object-oriented software design methods have been applied to ensure hardware independence and a standard application programming interface (API). There are components to acquire, display, and register images from MRI, MRA, CT, Computed Rotational Angiography (CRA), Digital Subtraction Angiography (DSA), 2D and 3D ultrasound, video and physiological recordings. Interfaces to various tracking systems for intra-operative use have also been implemented. Results: The described components have been implemented and tested. To date they have been used to create image manipulation and viewing tools, a deep brain functional atlas, a 3D ultrasound acquisition and display platform, a prototype minimally invasive robotic coronary artery bypass graft planning system, a tracked neuro-endoscope guidance system and a frame-based stereotaxy neurosurgery planning tool. The frame-based stereotaxy module has been

  20. Woods Hole Image Processing System Software implementation; using NetCDF as a software interface for image processing

    USGS Publications Warehouse

    Paskevich, Valerie F.

    1992-01-01

    The Branch of Atlantic Marine Geology has been involved in the collection, processing and digital mosaicking of high, medium and low-resolution side-scan sonar data during the past 6 years. In the past, processing and digital mosaicking has been accomplished with a dedicated, shore-based computer system. With the need to process sidescan data in the field with increased power and reduced cost of major workstations, a need to have an image processing package on a UNIX based computer system which could be utilized in the field as well as be more generally available to Branch personnel was identified. This report describes the initial development of that package referred to as the Woods Hole Image Processing System (WHIPS). The software was developed using the Unidata NetCDF software interface to allow data to be more readily portable between different computer operating systems.

  1. Development of Software to Model AXAF-I Image Quality

    NASA Technical Reports Server (NTRS)

    Ahmad, Anees; Hawkins, Lamar

    1996-01-01

    This draft final report describes the work performed under the delivery order number 145 from May 1995 through August 1996. The scope of work included a number of software development tasks for the performance modeling of AXAF-I. A number of new capabilities and functions have been added to the GT software, which is the command mode version of the GRAZTRACE software, originally developed by MSFC. A structural data interface has been developed for the EAL (old SPAR) finite element analysis FEA program, which is being used by MSFC Structural Analysis group for the analysis of AXAF-I. This interface utility can read the structural deformation file from the EAL and other finite element analysis programs such as NASTRAN and COSMOS/M, and convert the data to a suitable format that can be used for the deformation ray-tracing to predict the image quality for a distorted mirror. There is a provision in this utility to expand the data from finite element models assuming 180 degrees symmetry. This utility has been used to predict image characteristics for the AXAF-I HRMA, when subjected to gravity effects in the horizontal x-ray ground test configuration. The development of the metrology data processing interface software has also been completed. It can read the HDOS FITS format surface map files, manipulate and filter the metrology data, and produce a deformation file, which can be used by GT for ray tracing for the mirror surface figure errors. This utility has been used to determine the optimum alignment (axial spacing and clocking) for the four pairs of AXAF-I mirrors. Based on this optimized alignment, the geometric images and effective focal lengths for the as built mirrors were predicted to cross check the results obtained by Kodak.

  2. Demineralization Depth Using QLF and a Novel Image Processing Software

    PubMed Central

    Wu, Jun; Donly, Zachary R.; Donly, Kevin J.; Hackmyer, Steven

    2010-01-01

    Quantitative Light-Induced fluorescence (QLF) has been widely used to detect tooth demineralization indicated by fluorescence loss with respect to surrounding sound enamel. The correlation between fluorescence loss and demineralization depth is not fully understood. The purpose of this project was to study this correlation to estimate demineralization depth. Extracted teeth were collected. Artificial caries-like lesions were created and imaged with QLF. Novel image processing software was developed to measure the largest percent of fluorescence loss in the region of interest. All teeth were then sectioned and imaged by polarized light microscopy. The largest depth of demineralization was measured by NIH ImageJ software. The statistical linear regression method was applied to analyze these data. The linear regression model was Y = 0.32X + 0.17, where X was the percent loss of fluorescence and Y was the depth of demineralization. The correlation coefficient was 0.9696. The two-tailed t-test for coefficient was 7.93, indicating the P-value = .0014. The F test for the entire model was 62.86, which shows the P-value = .0013. The results indicated statistically significant linear correlation between the percent loss of fluorescence and depth of the enamel demineralization. PMID:20445755

  3. Performance of Personal Workspace Controls Final Report

    SciTech Connect

    Rubinstein, Francis; Kiliccote, Sila; Loffeld, John; Pettler,Pete; Snook, Joel

    2004-12-01

    One of the key deliverables for the DOE-funded controls research at LBNL for FY04 was the development of a prototype Personal Workspace Control system. The successful development of this system is a critical milestone for the LBNL Lighting Controls Research effort because this system demonstrates how IBECS can add value to today's Task Ambient lighting systems. LBNL has argued that by providing both the occupant and the facilities manager with the ability to precisely control the operation of overhead lighting and all task lighting in a coordinated manner, that task ambient lighting can optimize energy performance and occupant comfort simultaneously [Reference Task Ambient Foundation Document]. The Personal Workspace Control system is the application of IBECS to this important lighting problem. This report discusses the development of the Personal Workspace Control to date including descriptions of the different fixture types that have been converted to IBECS operation and a detailed description of the operation of PWC Scene Controller, which provides the end user with precise control of his task ambient lighting system. The objective, from the Annual Plan, is to demonstrate improvements in efficiency, lighting quality and occupant comfort realized using Personal Workspace Controls (PWC) designed to optimize the delivery of lighting to the individual's workstation regardless of which task-ambient lighting solution is chosen. The PWC will be capable of controlling floor-mounted, desk lamps, furniture-mounted and overhead lighting fixtures from a personal computer and handheld remote. The PWC will use an environmental sensor to automatically monitor illuminance, temperature and occupancy and to appropriately modulate ambient lighting according to daylight availability and to switch off task lighting according to local occupancy. [Adding occupancy control to the system would blunt the historical criticism of occupant-controlled lighting - the tendency of the occupant

  4. Software and Algorithms for Biomedical Image Data Processing and Visualization

    NASA Technical Reports Server (NTRS)

    Talukder, Ashit; Lambert, James; Lam, Raymond

    2004-01-01

    A new software equipped with novel image processing algorithms and graphical-user-interface (GUI) tools has been designed for automated analysis and processing of large amounts of biomedical image data. The software, called PlaqTrak, has been specifically used for analysis of plaque on teeth of patients. New algorithms have been developed and implemented to segment teeth of interest from surrounding gum, and a real-time image-based morphing procedure is used to automatically overlay a grid onto each segmented tooth. Pattern recognition methods are used to classify plaque from surrounding gum and enamel, while ignoring glare effects due to the reflection of camera light and ambient light from enamel regions. The PlaqTrak system integrates these components into a single software suite with an easy-to-use GUI (see Figure 1) that allows users to do an end-to-end run of a patient s record, including tooth segmentation of all teeth, grid morphing of each segmented tooth, and plaque classification of each tooth image. The automated and accurate processing of the captured images to segment each tooth [see Figure 2(a)] and then detect plaque on a tooth-by-tooth basis is a critical component of the PlaqTrak system to do clinical trials and analysis with minimal human intervention. These features offer distinct advantages over other competing systems that analyze groups of teeth or synthetic teeth. PlaqTrak divides each segmented tooth into eight regions using an advanced graphics morphing procedure [see results on a chipped tooth in Figure 2(b)], and a pattern recognition classifier is then used to locate plaque [red regions in Figure 2(d)] and enamel regions. The morphing allows analysis within regions of teeth, thereby facilitating detailed statistical analysis such as the amount of plaque present on the biting surfaces on teeth. This software system is applicable to a host of biomedical applications, such as cell analysis and life detection, or robotic applications, such

  5. IDL Object Oriented Software for Hinode/XRT Image Analysis

    NASA Astrophysics Data System (ADS)

    Higgins, P. A.; Gallagher, P. T.

    2008-09-01

    We have developed a set of object oriented IDL routines that enable users to search, download and analyse images from the X-Ray Telescope (XRT) on-board Hinode. In this paper, we give specific examples of how the object can be used and how multi-instrument data analysis can be performed. The XRT object is a highly versatile and powerful IDL object, which will prove to be a useful tool for solar researchers. This software utilizes the generic Framework object available within the GEN branch of SolarSoft.

  6. Special Software for Planetary Image Processing and Research

    NASA Astrophysics Data System (ADS)

    Zubarev, A. E.; Nadezhdina, I. E.; Kozlova, N. A.; Brusnikin, E. S.; Karachevtseva, I. P.

    2016-06-01

    The special modules of photogrammetric processing of remote sensing data that provide the opportunity to effectively organize and optimize the planetary studies were developed. As basic application the commercial software package PHOTOMOD™ is used. Special modules were created to perform various types of data processing: calculation of preliminary navigation parameters, calculation of shape parameters of celestial body, global view image orthorectification, estimation of Sun illumination and Earth visibilities from planetary surface. For photogrammetric processing the different types of data have been used, including images of the Moon, Mars, Mercury, Phobos, Galilean satellites and Enceladus obtained by frame or push-broom cameras. We used modern planetary data and images that were taken over the years, shooting from orbit flight path with various illumination and resolution as well as obtained by planetary rovers from surface. Planetary data image processing is a complex task, and as usual it can take from few months to years. We present our efficient pipeline procedure that provides the possibilities to obtain different data products and supports a long way from planetary images to celestial body maps. The obtained data - new three-dimensional control point networks, elevation models, orthomosaics - provided accurate maps production: a new Phobos atlas (Karachevtseva et al., 2015) and various thematic maps that derived from studies of planetary surface (Karachevtseva et al., 2016a).

  7. Software architecture for intelligent image processing using Prolog

    NASA Astrophysics Data System (ADS)

    Jones, Andrew C.; Batchelor, Bruce G.

    1994-10-01

    We describe a prototype system for interactive image processing using Prolog, implemented by the first author on an Apple Macintosh computer. This system is inspired by Prolog+, but differs from it in two particularly important respects. The first is that whereas Prolog+ assumes the availability of dedicated image processing hardware, with which the Prolog system communicates, our present system implements image processing functions in software using the C programming language. The second difference is that although our present system supports Prolog+ commands, these are implemented in terms of lower-level Prolog predicates which provide a more flexible approach to image manipulation. We discuss the impact of the Apple Macintosh operating system upon the implementation of the image-processing functions, and the interface between these functions and the Prolog system. We also explain how the Prolog+ commands have been implemented. The system described in this paper is a fairly early prototype, and we outline how we intend to develop the system, a task which is expedited by the extensible architecture we have implemented.

  8. 'Face value': new medical imaging software in commercial view.

    PubMed

    Coopmans, Catelijne

    2011-04-01

    Based on three ethnographic vignettes describing the engagements of a small start-up company with prospective competitors, partners and customers, this paper shows how commercial considerations are folded into the ways visual images become 'seeable'. When company members mount demonstrations of prototype mammography software, they seek to generate interest but also to protect their intellectual property. Pivotal to these efforts to manage revelation and concealment is the visual interface, which is variously performed as obstacle and ally in the development of a profitable product. Using the concept of 'face value', the paper seeks to develop further insight into contemporary dynamics of seeing and showing by tracing the way techno-visual presentations and commercial considerations become entangled in practice. It also draws attention to the salience and significance of enactments of surface and depth in image-based practices.

  9. Electrophoretic gel image analysis software for the molecular biology laboratory.

    PubMed

    Redman, T; Jacobs, T

    1991-06-01

    We present GelReader 1.0, a microcomputer program designed to make precision, digital analysis of one-dimensional electrophoretic gels accessible to the molecular biology laboratory of modest means. Images of electrophoretic gels are digitized via a desktop flatbed scanner from instant photographs, autoradiograms or chromogenically stained blotting media. GelReader is then invoked to locate lanes and bands and generate a report of molecular weights of unknowns, based on specified sets of standards. Frequently used standards can be stored in the program. Lanes and bands can be added or removed, based upon users' subjective preferences. A unique lane histogram feature facilitates precise manual addition of bands missed by the software. Image enhancement features include palette manipulation, histogram equalization, shadowing and magnification. The user interface strikes a balance between program autonomy and user intervention, in recognition of the variability in electrophoretic gel quality and users' analytical needs.

  10. Development of Automated Image Analysis Software for Suspended Marine Particle Classification

    DTIC Science & Technology

    2003-09-30

    Development of Automated Image Analysis Software for Suspended Marine Particle Classification Scott Samson Center for Ocean Technology...REPORT TYPE 3. DATES COVERED 00-00-2003 to 00-00-2003 4. TITLE AND SUBTITLE Development of Automated Image Analysis Software for Suspended...objective is to develop automated image analysis software to reduce the effort and time required for manual identification of plankton images. Automated

  11. Predictive images of postoperative levator resection outcome using image processing software

    PubMed Central

    Mawatari, Yuki; Fukushima, Mikiko

    2016-01-01

    Purpose This study aims to evaluate the efficacy of processed images to predict postoperative appearance following levator resection. Methods Analysis involved 109 eyes from 65 patients with blepharoptosis who underwent advancement of levator aponeurosis and Müller’s muscle complex (levator resection). Predictive images were prepared from preoperative photographs using the image processing software (Adobe Photoshop®). Images of selected eyes were digitally enlarged in an appropriate manner and shown to patients prior to surgery. Results Approximately 1 month postoperatively, we surveyed our patients using questionnaires. Fifty-six patients (89.2%) were satisfied with their postoperative appearances, and 55 patients (84.8%) positively responded to the usefulness of processed images to predict postoperative appearance. Conclusion Showing processed images that predict postoperative appearance to patients prior to blepharoptosis surgery can be useful for those patients concerned with their postoperative appearance. This approach may serve as a useful tool to simulate blepharoptosis surgery. PMID:27757008

  12. Software defined multi-spectral imaging for Arctic sensor networks

    NASA Astrophysics Data System (ADS)

    Siewert, Sam; Angoth, Vivek; Krishnamurthy, Ramnarayan; Mani, Karthikeyan; Mock, Kenrick; Singh, Surjith B.; Srivistava, Saurav; Wagner, Chris; Claus, Ryan; Vis, Matthew Demi

    2016-05-01

    Availability of off-the-shelf infrared sensors combined with high definition visible cameras has made possible the construction of a Software Defined Multi-Spectral Imager (SDMSI) combining long-wave, near-infrared and visible imaging. The SDMSI requires a real-time embedded processor to fuse images and to create real-time depth maps for opportunistic uplink in sensor networks. Researchers at Embry Riddle Aeronautical University working with University of Alaska Anchorage at the Arctic Domain Awareness Center and the University of Colorado Boulder have built several versions of a low-cost drop-in-place SDMSI to test alternatives for power efficient image fusion. The SDMSI is intended for use in field applications including marine security, search and rescue operations and environmental surveys in the Arctic region. Based on Arctic marine sensor network mission goals, the team has designed the SDMSI to include features to rank images based on saliency and to provide on camera fusion and depth mapping. A major challenge has been the design of the camera computing system to operate within a 10 to 20 Watt power budget. This paper presents a power analysis of three options: 1) multi-core, 2) field programmable gate array with multi-core, and 3) graphics processing units with multi-core. For each test, power consumed for common fusion workloads has been measured at a range of frame rates and resolutions. Detailed analyses from our power efficiency comparison for workloads specific to stereo depth mapping and sensor fusion are summarized. Preliminary mission feasibility results from testing with off-the-shelf long-wave infrared and visible cameras in Alaska and Arizona are also summarized to demonstrate the value of the SDMSI for applications such as ice tracking, ocean color, soil moisture, animal and marine vessel detection and tracking. The goal is to select the most power efficient solution for the SDMSI for use on UAVs (Unoccupied Aerial Vehicles) and other drop

  13. Reducing depth uncertainty in large surgical workspaces, with applications to veterinary medicine

    NASA Astrophysics Data System (ADS)

    Audette, Michel A.; Kolahi, Ahmad; Enquobahrie, Andinet; Gatti, Claudio; Cleary, Kevin

    2010-02-01

    This paper presents on-going research that addresses uncertainty along the Z-axis in image-guided surgery, for applications to large surgical workspaces, including those found in veterinary medicine. Veterinary medicine lags human medicine in using image guidance, despite MR and CT data scanning of animals. The positional uncertainty of a surgical tracking device can be modeled as an octahedron with one long axis coinciding with the depth axis of the sensor, where the short axes are determined by pixel resolution and workspace dimensions. The further a 3D point is from this device, the more elongated is this long axis, and the greater the uncertainty along Z of this point's position, in relation to its components along X and Y. Moreover, for a triangulation-based tracker, its position error degrades with the square of distance. Our approach is to use two or more Micron Trackers to communicate with each other, and combine this feature with flexible positioning. Prior knowledge of the type of surgical procedure, and if applicable, the species of animal that determines the scale of the workspace, would allow the surgeon to pre-operatively configure the trackers in the OR for optimal accuracy. Our research also leverages the open-source Image-guided Surgery Toolkit (IGSTK).

  14. WorkstationJ: workstation emulation software for medical image perception and technology evaluation research

    NASA Astrophysics Data System (ADS)

    Schartz, Kevin M.; Berbaum, Kevin S.; Caldwell, Robert T.; Madsen, Mark T.

    2007-03-01

    We developed image presentation software that mimics the functionality available in the clinic, but also records time-stamped, observer-display interactions and is readily deployable on diverse workstations making it possible to collect comparable observer data at multiple sites. Commercial image presentation software for clinical use has limited application for research on image perception, ergonomics, computer-aids and informatics because it does not collect observer responses, or other information on observer-display interactions, in real time. It is also very difficult to collect observer data from multiple institutions unless the same commercial software is available at different sites. Our software not only records observer reports of abnormalities and their locations, but also inspection time until report, inspection time for each computed radiograph and for each slice of tomographic studies, window/level, and magnification settings used by the observer. The software is a modified version of the open source ImageJ software available from the National Institutes of Health. Our software involves changes to the base code and extensive new plugin code. Our free software is currently capable of displaying computed tomography and computed radiography images. The software is packaged as Java class files and can be used on Windows, Linux, or Mac systems. By deploying our software together with experiment-specific script files that administer experimental procedures and image file handling, multi-institutional studies can be conducted that increase reader and/or case sample sizes or add experimental conditions.

  15. Quantification of fungal infection of leaves with digital images and Scion Image software.

    PubMed

    Goodwin, Paul H; Hsiang, Tom

    2010-01-01

    Digital image analysis has been used to distinguish and quantify leaf color changes arising from a variety of factors. Its use to assess the percentage of leaf area with color differences caused by plant disease symptoms, such as necrosis, chlorosis, or sporulation, can provide a rigorous and quantitative means of assessing disease severity. A method is described for measuring symptoms of different fungal foliar infections that involves capturing the image with a standard flatbed scanner or digital camera followed by quantifying the area, where the color has been affected because of fungal infection. The method uses the freely available program, Scion Image for Windows or MAC, which is derived from the public domain software, NIH Image. The method has thus far been used to quantify the percentage of tissue with necrosis, chlorosis, or sporulation on leaves of variety of plants with several different diseases (anthracnose, apple scab, powdery mildew or rust).

  16. Development of Automated Image Analysis Software for Suspended Marine Particle Classification

    DTIC Science & Technology

    2002-09-30

    Development of Automated Image Analysis Software for Suspended Marine Particle Classification Scott Samson Center for Ocean Technology...and global water column. 1 OBJECTIVES The project’s objective is to develop automated image analysis software to reduce the effort and time

  17. Quantifying fungal infection of plant leaves by digital image analysis using Scion Image software.

    PubMed

    Wijekoon, C P; Goodwin, P H; Hsiang, T

    2008-08-01

    A digital image analysis method previously used to evaluate leaf color changes due to nutritional changes was modified to measure the severity of several foliar fungal diseases. Images captured with a flatbed scanner or digital camera were analyzed with a freely available software package, Scion Image, to measure changes in leaf color caused by fungal sporulation or tissue damage. High correlations were observed between the percent diseased leaf area estimated by Scion Image analysis and the percent diseased leaf area from leaf drawings. These drawings of various foliar diseases came from a disease key previously developed to aid in visual estimation of disease severity. For leaves of Nicotiana benthamiana inoculated with different spore concentrations of the anthracnose fungus Colletotrichum destructivum, a high correlation was found between the percent diseased tissue measured by Scion Image analysis and the number of leaf spots. The method was adapted to quantify percent diseased leaf area ranging from 0 to 90% for anthracnose of lily-of-the-valley, apple scab, powdery mildew of phlox and rust of golden rod. In some cases, the brightness and contrast of the images were adjusted and other modifications were made, but these were standardized for each disease. Detached leaves were used with the flatbed scanner, but a method using attached leaves with a digital camera was also developed to make serial measurements of individual leaves to quantify symptom progression. This was successfully applied to monitor anthracnose on N. benthamiana leaves. Digital image analysis using Scion Image software is a useful tool for quantifying a wide variety of fungal interactions with plant leaves.

  18. Design and validation of Segment - freely available software for cardiovascular image analysis

    PubMed Central

    2010-01-01

    Background Commercially available software for cardiovascular image analysis often has limited functionality and frequently lacks the careful validation that is required for clinical studies. We have already implemented a cardiovascular image analysis software package and released it as freeware for the research community. However, it was distributed as a stand-alone application and other researchers could not extend it by writing their own custom image analysis algorithms. We believe that the work required to make a clinically applicable prototype can be reduced by making the software extensible, so that researchers can develop their own modules or improvements. Such an initiative might then serve as a bridge between image analysis research and cardiovascular research. The aim of this article is therefore to present the design and validation of a cardiovascular image analysis software package (Segment) and to announce its release in a source code format. Results Segment can be used for image analysis in magnetic resonance imaging (MRI), computed tomography (CT), single photon emission computed tomography (SPECT) and positron emission tomography (PET). Some of its main features include loading of DICOM images from all major scanner vendors, simultaneous display of multiple image stacks and plane intersections, automated segmentation of the left ventricle, quantification of MRI flow, tools for manual and general object segmentation, quantitative regional wall motion analysis, myocardial viability analysis and image fusion tools. Here we present an overview of the validation results and validation procedures for the functionality of the software. We describe a technique to ensure continued accuracy and validity of the software by implementing and using a test script that tests the functionality of the software and validates the output. The software has been made freely available for research purposes in a source code format on the project home page http

  19. Software.

    ERIC Educational Resources Information Center

    Journal of Chemical Education, 1989

    1989-01-01

    Presented are reviews of two computer software packages for Apple II computers; "Organic Spectroscopy," and "Videodisc Display Program" for use with "The Periodic Table Videodisc." A sample spectrograph from "Organic Spectroscopy" is included. (CW)

  20. Emerging standards for still image compression: A software implementation and simulation study

    NASA Technical Reports Server (NTRS)

    Pollara, F.; Arnold, S.

    1991-01-01

    The software implementation is described of an emerging standard for the lossy compression of continuous tone still images. This software program can be used to compress planetary images and other 2-D instrument data. It provides a high compression image coding capability that preserves image fidelity at compression rates competitive or superior to most known techniques. This software implementation confirms the usefulness of such data compression and allows its performance to be compared with other schemes used in deep space missions and for data based storage.

  1. NEIGHBOUR-IN: Image processing software for spatial analysis of animal grouping

    PubMed Central

    Caubet, Yves; Richard, Freddie-Jeanne

    2015-01-01

    Abstract Animal grouping is a very complex process that occurs in many species, involving many individuals under the influence of different mechanisms. To investigate this process, we have created an image processing software, called NEIGHBOUR-IN, designed to analyse individuals’ coordinates belonging to up to three different groups. The software also includes statistical analysis and indexes to discriminate aggregates based on spatial localisation of individuals and their neighbours. After the description of the software, the indexes computed by the software are illustrated using both artificial patterns and case studies using the spatial distribution of woodlice. The added strengths of this software and methods are also discussed. PMID:26261448

  2. NEIGHBOUR-IN: Image processing software for spatial analysis of animal grouping.

    PubMed

    Caubet, Yves; Richard, Freddie-Jeanne

    2015-01-01

    Animal grouping is a very complex process that occurs in many species, involving many individuals under the influence of different mechanisms. To investigate this process, we have created an image processing software, called NEIGHBOUR-IN, designed to analyse individuals' coordinates belonging to up to three different groups. The software also includes statistical analysis and indexes to discriminate aggregates based on spatial localisation of individuals and their neighbours. After the description of the software, the indexes computed by the software are illustrated using both artificial patterns and case studies using the spatial distribution of woodlice. The added strengths of this software and methods are also discussed.

  3. Platform-independent software for medical image processing on the Internet

    NASA Astrophysics Data System (ADS)

    Mancuso, Michael E.; Pathak, Sayan D.; Kim, Yongmin

    1997-05-01

    We have developed a software tool for image processing over the Internet. The tool is a general purpose, easy to use, flexible, platform independent image processing software package with functions most commonly used in medical image processing.It provides for processing of medical images located wither remotely on the Internet or locally. The software was written in Java - the new programming language developed by Sun Microsystems. It was compiled and tested using Microsoft's Visual Java 1.0 and Microsoft's Just in Time Compiler 1.00.6211. The software is simple and easy to use. In order to use the tool, the user needs to download the software from our site before he/she runs it using any Java interpreter, such as those supplied by Sun, Symantec, Borland or Microsoft. Future versions of the operating systems supplied by Sun, Microsoft, Apple, IBM, and others will include Java interpreters. The software is then able to access and process any image on the iNternet or on the local computer. Using a 512 X 512 X 8-bit image, a 3 X 3 convolution took 0.88 seconds on an Intel Pentium Pro PC running at 200 MHz with 64 Mbytes of memory. A window/level operation took 0.38 seconds while a 3 X 3 median filter took 0.71 seconds. These performance numbers demonstrate the feasibility of using this software interactively on desktop computes. Our software tool supports various image processing techniques commonly used in medical image processing and can run without the need of any specialized hardware. It can become an easily accessible resource over the Internet to promote the learning and of understanding image processing algorithms. Also, it could facilitate sharing of medical image databases and collaboration amongst researchers and clinicians, regardless of location.

  4. Application of open source image guided therapy software in MR-guided therapies.

    PubMed

    Hata, Nobuhiko; Piper, Steve; Jolesz, Ferenc A; Tempany, Clare M C; Black, Peter McL; Morikawa, Shigehiro; Iseki, Horoshi; Hashizume, Makoto; Kikinis, Ron

    2007-01-01

    We present software engineering methods to provide free open-source software for MR-guided therapy. We report that graphical representation of the surgical tools, interconnectively with the tracking device, patient-to-image registration, and MRI-based thermal mapping are crucial components of MR-guided therapy in sharing such software. Software process includes a network-based distribution mechanism by multi-platform compiling tool CMake, CVS, quality assurance software DART. We developed six procedures in four separate clinical sites using proposed software engineering and process, and found the proposed method is feasible to facilitate multicenter clinical trial of MR-guided therapies. Our future studies include use of the software in non-MR-guided therapies.

  5. Planning maximally smooth hand movements constrained to nonplanar workspaces.

    PubMed

    Liebermann, Dario G; Krasovsky, Tal; Berman, Sigal

    2008-11-01

    The article characterizes hand paths and speed profiles for movements performed in a nonplanar, 2-dimensional workspace (a hemisphere of constant curvature). The authors assessed endpoint kinematics (i.e., paths and speeds) under the minimum-jerk model assumptions and calculated minimal amplitude paths (geodesics) and the corresponding speed profiles. The authors also calculated hand speeds using the 2/3 power law. They then compared modeled results with the empirical observations. In all, 10 participants moved their hands forward and backward from a common starting position toward 3 targets located within a hemispheric workspace of small or large curvature. Comparisons of modeled observed differences using 2-way RM-ANOVAs showed that movement direction had no clear influence on hand kinetics (p < .05). Workspace curvature affected the hand paths, which seldom followed geodesic lines. Constraining the paths to different curvatures did not affect the hand speed profiles. Minimum-jerk speed profiles closely matched the observations and were superior to those predicted by 2/3 power law (p < .001). The authors conclude that speed and path cannot be unambiguously linked under the minimum-jerk assumption when individuals move the hand in a nonplanar 2-dimensional workspace. In such a case, the hands do not follow geodesic paths, but they preserve the speed profile, regardless of the geometric features of the workspace.

  6. Software Compression for Partially Parallel Imaging with Multi-channels.

    PubMed

    Huang, Feng; Vijayakumar, Sathya; Akao, James

    2005-01-01

    In magnetic resonance imaging, multi-channel phased array coils enjoy a high signal to noise ratio (SNR) and better parallel imaging performance. But with the increase in number of channels, the reconstruction time and requirement for computer memory become inevitable problems. In this work, principle component analysis is applied to reduce the size of data and protect the performance of parallel imaging. Clinical data collected using a 32-channel cardiac coil are used in the experiments. Experimental results show that the proposed method dramatically reduces the processing time without much damage to the reconstructed image.

  7. Fluorescence Image Analyzer - FLIMA: software for quantitative analysis of fluorescence in situ hybridization.

    PubMed

    Silva, H C M; Martins-Júnior, M M C; Ribeiro, L B; Matoso, D A

    2017-03-30

    The Fluorescence Image Analyzer (FLIMA) software was developed for the quantitative analysis of images generated by fluorescence in situ hybridization (FISH). Currently, the images of FISH are examined without a coefficient that enables a comparison between them. Through GD Graphics Library, the FLIMA software calculates the amount of pixels on image and recognizes each present color. The coefficient generated by the algorithm shows the percentage of marks (probes) hybridized on the chromosomes. This software can be used for any type of image generated by a fluorescence microscope and is able to quantify digoxigenin probes exhibiting a red color, biotin probes exhibiting a green color, and double-FISH probes (digoxigenin and biotin used together), where the white color is displayed.

  8. ESO C Library for an Image Processing Software Environment (eclipse)

    NASA Astrophysics Data System (ADS)

    Devillard, N.

    Written in ANSI C, eclipse is a library offering numerous services related to astronomical image processing: FITS data access, various image and cube loading methods, binary image handling and filtering (including convolution and morphological filters), 2-D cross-correlation, connected components, cube and image arithmetic, dead pixel detection and correction, object detection, data extraction, flat-fielding with robust fit, image generation, statistics, photometry, image-space resampling, image combination, and cube stacking. It also contains support for mathematical tools like random number generation, FFT, curve fitting, matrices, fast median computation, and point-pattern matching. The main feature of this library is its ability to handle large amounts of input data (up to 2 GB in the current version) regardless of the amount of memory and swap available on the local machine. Another feature is the very high speed allowed by optimized C, making it an ideal base tool for programming efficient number-crunching applications, e.g., on parallel (Beowulf) systems. Running on all Unix-like platforms, eclipse is portable. A high-level interface to Python is foreseen that would allow programmers to prototype their applications much faster than through C programs.

  9. Eclipse: ESO C Library for an Image Processing Software Environment

    NASA Astrophysics Data System (ADS)

    Devillard, Nicolas

    2011-12-01

    Written in ANSI C, eclipse is a library offering numerous services related to astronomical image processing: FITS data access, various image and cube loading methods, binary image handling and filtering (including convolution and morphological filters), 2-D cross-correlation, connected components, cube and image arithmetic, dead pixel detection and correction, object detection, data extraction, flat-fielding with robust fit, image generation, statistics, photometry, image-space resampling, image combination, and cube stacking. It also contains support for mathematical tools like random number generation, FFT, curve fitting, matrices, fast median computation, and point-pattern matching. The main feature of this library is its ability to handle large amounts of input data (up to 2GB in the current version) regardless of the amount of memory and swap available on the local machine. Another feature is the very high speed allowed by optimized C, making it an ideal base tool for programming efficient number-crunching applications, e.g., on parallel (Beowulf) systems.

  10. Polarization information processing and software system design for simultaneously imaging polarimetry

    NASA Astrophysics Data System (ADS)

    Wang, Yahui; Liu, Jing; Jin, Weiqi; Wen, Renjie

    2015-08-01

    Simultaneous imaging polarimetry can realize real-time polarization imaging of the dynamic scene, which has wide application prospect. This paper first briefly illustrates the design of the double separate Wollaston Prism simultaneous imaging polarimetry, and then emphases are put on the polarization information processing methods and software system design for the designed polarimetry. Polarization information processing methods consist of adaptive image segmentation, high-accuracy image registration, instrument matrix calibration. Morphological image processing was used for image segmentation by taking dilation of an image; The accuracy of image registration can reach 0.1 pixel based on the spatial and frequency domain cross-correlation; Instrument matrix calibration adopted four-point calibration method. The software system was implemented under Windows environment based on C++ programming language, which realized synchronous polarization images acquisition and preservation, image processing and polarization information extraction and display. Polarization data obtained with the designed polarimetry shows that: the polarization information processing methods and its software system effectively performs live realize polarization measurement of the four Stokes parameters of a scene. The polarization information processing methods effectively improved the polarization detection accuracy.

  11. BIRP: Software for interactive search and retrieval of image engineering data

    NASA Technical Reports Server (NTRS)

    Arvidson, R. E.; Bolef, L. K.; Guinness, E. A.; Norberg, P.

    1980-01-01

    Better Image Retrieval Programs (BIRP), a set of programs to interactively sort through and to display a database, such as engineering data for images acquired by spacecraft is described. An overview of the philosophy of BIRP design, the structure of BIRP data files, and examples that illustrate the capabilities of the software are provided.

  12. A Review of Diffusion Tensor Magnetic Resonance Imaging Computational Methods and Software Tools

    PubMed Central

    Hasan, Khader M.; Walimuni, Indika S.; Abid, Humaira; Hahn, Klaus R.

    2010-01-01

    In this work we provide an up-to-date short review of computational magnetic resonance imaging (MRI) and software tools that are widely used to process and analyze diffusion-weighted MRI data. A review of different methods used to acquire, model and analyze diffusion-weighted imaging data (DWI) is first provided with focus on diffusion tensor imaging (DTI). The major preprocessing, processing and post-processing procedures applied to DTI data are discussed. A list of freely available software packages to analyze diffusion MRI data is also provided. PMID:21087766

  13. Spatial data software integration - Merging CAD/CAM/mapping with GIS and image processing

    NASA Technical Reports Server (NTRS)

    Logan, Thomas L.; Bryant, Nevin A.

    1987-01-01

    The integration of CAD/CAM/mapping with image processing using geographic information systems (GISs) as the interface is examined. Particular emphasis is given to the development of software interfaces between JPL's Video Image Communication and Retrieval (VICAR)/Imaged Based Information System (IBIS) raster-based GIS and the CAD/CAM/mapping system. The design and functions of the VICAR and IBIS are described. Vector data capture and editing are studied. Various software programs for interfacing between the VICAR/IBIS and CAD/CAM/mapping are presented and analyzed.

  14. 75 FR 28058 - In the Matter of Certain Digital Imaging Devices and Related Software; Notice of Investigation

    Federal Register 2010, 2011, 2012, 2013, 2014

    2010-05-19

    ... COMMISSION In the Matter of Certain Digital Imaging Devices and Related Software; Notice of Investigation... software by reason of infringement of certain claims of U.S. Patent Nos. 6,031,964 and RE38,911. The... importation of certain digital imaging devices and related software that infringe one or more of claim 1-3...

  15. JUPOS : Amateur analysis of Jupiter images with specialized measurement software

    NASA Astrophysics Data System (ADS)

    Jacquesson, M.; Mettig, H.-J.

    2008-09-01

    Introduction - Beginning of JUPOS in 1989, H.J. Mettig and Grischa Hahn in Dresden - 350.000 positional measures on electronic images of almost 180 observers have been gathered since 1998 - What do we mean by "electronic images": o digitized hi-res chemical photographs o traditional CCD technique o webcam images - 2002 started the implementation of WinJUPOS by Grischa, now finished - Cooperation with the Jupiter Section of the BAA for many years - At present, JUPOS has four active measurers: o Gianluigi Adamoli and Marco Vedovato (Italy) o H.J. Mettig (Germany) o Michel Jacquesson (France) How we work together - Each member of the team measures images of several observers (useful to detect and avoid systematic errors) - When necessary, exchange of problems and ideas by e-mail - During the period of visibility of Jupiter, team leader H.J. Mettig: o collects all recent measurements about once a month o produces drift charts - Once a year or two, the JUPOS team helds an informal meeting Criteria for image selection 1) Validity of time and date: Origins of time errors: 1. local zonal times are wrongly (or, not at all) converted to UT. This is still easy to find out: Either "only" the full hour is erroneous, or/and the date (problem of observers with UTC+10) 2. the computer clock has not been synchronised over a longer period 3. exposure of the final image exceeds the recommended two minutes or observers communicate begin or end of the total period of image recording instead of its middle How to test the validity of the time? - Position of a galilean satellite visible on or near Jupiter - Position of a satellite shadow - Measuring the longitude of permanent or long lived objects whose positions are known from former measurements 2) Avoid measuring the same object on different images taken at about the same time. One image every 1 ½ hours is sufficient 3) Duration of exposure: not more than 2-3 minutes because of the rapid rotation of Jupiter 4) Spectral range: Visual

  16. Software optimization for electrical conductivity imaging in polycrystalline diamond cutters

    SciTech Connect

    Bogdanov, G.; Ludwig, R.; Wiggins, J.; Bertagnolli, K.

    2014-02-18

    We previously reported on an electrical conductivity imaging instrument developed for measurements on polycrystalline diamond cutters. These cylindrical cutters for oil and gas drilling feature a thick polycrystalline diamond layer on a tungsten carbide substrate. The instrument uses electrical impedance tomography to profile the conductivity in the diamond table. Conductivity images must be acquired quickly, on the order of 5 sec per cutter, to be useful in the manufacturing process. This paper reports on successful efforts to optimize the conductivity reconstruction routine, porting major portions of it to NVIDIA GPUs, including a custom CUDA kernel for Jacobian computation.

  17. IDP: Image and data processing (software) in C++

    SciTech Connect

    Lehman, S.

    1994-11-15

    IDP++(Image and Data Processing in C++) is a complied, multidimensional, multi-data type, signal processing environment written in C++. It is being developed within the Radar Ocean Imaging group and is intended as a partial replacement for View. IDP++ takes advantage of the latest object-oriented compiler technology to provide `information hiding.` Users need only know C, not C++. Signals are treated like any other variable with a defined set of operators and functions in an intuitive manner. IDP++ is being designed for real-time environment where interpreted signal processing packages are less efficient.

  18. Image compression software for the SOHO LASCO and EIT experiments

    NASA Technical Reports Server (NTRS)

    Grunes, Mitchell R.; Howard, Russell A.; Hoppel, Karl; Mango, Stephen A.; Wang, Dennis

    1994-01-01

    This paper describes the lossless and lossy image compression algorithms to be used on board the Solar Heliospheric Observatory (SOHO) in conjunction with the Large Angle Spectrometric Coronograph and Extreme Ultraviolet Imaging Telescope experiments. It also shows preliminary results obtained using similar prior imagery and discusses the lossy compression artifacts which will result. This paper is in part intended for the use of SOHO investigators who need to understand the results of SOHO compression in order to better allocate the transmission bits which they have been allocated.

  19. Development of Software to Model AXAF-I Image Quality

    NASA Technical Reports Server (NTRS)

    Geary, Joseph; Hawkins, Lamar; Ahmad, Anees; Gong, Qian

    1997-01-01

    This report describes work conducted on Delivery Order 181 between October 1996 through June 1997. During this period software was written to: compute axial PSD's from RDOS AXAF-I mirror surface maps; plot axial surface errors and compute PSD's from HDOS "Big 8" axial scans; plot PSD's from FITS format PSD files; plot band-limited RMS vs axial and azimuthal position for multiple PSD files; combine and organize PSD's from multiple mirror surface measurements formatted as input to GRAZTRACE; modify GRAZTRACE to read FITS formatted PSD files; evaluate AXAF-I test results; improve and expand the capabilities of the GT x-ray mirror analysis package. During this period work began on a more user-friendly manual for the GT program, and improvements were made to the on-line help manual.

  20. Image analysis software for following progression of peripheral neuropathy

    NASA Astrophysics Data System (ADS)

    Epplin-Zapf, Thomas; Miller, Clayton; Larkin, Sean; Hermesmeyer, Eduardo; Macy, Jenny; Pellegrini, Marco; Luccarelli, Saverio; Staurenghi, Giovanni; Holmes, Timothy

    2009-02-01

    A relationship has been reported by several research groups [1 - 4] between the density and shapes of nerve fibers in the cornea and the existence and severity of peripheral neuropathy. Peripheral neuropathy is a complication of several prevalent diseases or conditions, which include diabetes, HIV, prolonged alcohol overconsumption and aging. A common clinical technique for confirming the condition is intramuscular electromyography (EMG), which is invasive, so a noninvasive technique like the one proposed here carries important potential advantages for the physician and patient. A software program that automatically detects the nerve fibers, counts them and measures their shapes is being developed and tested. Tests were carried out with a database of subjects with levels of severity of diabetic neuropathy as determined by EMG testing. Results from this testing, that include a linear regression analysis are shown.

  1. Validated novel software to measure the conspicuity index of lesions in DICOM images

    NASA Astrophysics Data System (ADS)

    Szczepura, K. R.; Manning, D. J.

    2016-03-01

    A novel software programme and associated Excel spreadsheet has been developed to provide an objective measure of the expected visual detectability of focal abnormalities within DICOM images. ROIs are drawn around the abnormality, the software then fits the lesion using a least squares method to recognize the edges of the lesion based on the full width half maximum. 180 line profiles are then plotted around the lesion, giving 360 edge profiles.

  2. Parallel-Processing Software for Correlating Stereo Images

    NASA Technical Reports Server (NTRS)

    Klimeck, Gerhard; Deen, Robert; Mcauley, Michael; DeJong, Eric

    2007-01-01

    A computer program implements parallel- processing algorithms for cor relating images of terrain acquired by stereoscopic pairs of digital stereo cameras on an exploratory robotic vehicle (e.g., a Mars rove r). Such correlations are used to create three-dimensional computatio nal models of the terrain for navigation. In this program, the scene viewed by the cameras is segmented into subimages. Each subimage is assigned to one of a number of central processing units (CPUs) opera ting simultaneously.

  3. A role for the anterior insular cortex in the global neuronal workspace model of consciousness.

    PubMed

    Michel, Matthias

    2017-03-01

    According to the global neuronal workspace model of consciousness, consciousness results from the global broadcast of information throughout the brain. The global neuronal workspace is mainly constituted by a fronto-parietal network. The anterior insular cortex is part of this global neuronal workspace, but the function of this region has not yet been defined within the global neuronal workspace model of consciousness. In this review, I hypothesize that the anterior insular cortex implements a cross-modal priority map, the function of which is to determine priorities for the processing of information and subsequent entrance in the global neuronal workspace.

  4. New StatPhantom software for assessment of digital image quality

    NASA Astrophysics Data System (ADS)

    Gurvich, Victor A.; Davydenko, George I.

    2002-04-01

    The rapid development of digital imaging and computers networks, using Picture Archiving and Communication Systems (PACS) and DICOM compatible devices increase requirements to the quality control process in medical imaging departments, but provide new opportunities for evaluation of image quality. New StatPhantom software simplifies statistical techniques based on modern detection theory and ROC analysis improving the accuracy and reliability of known methods and allowing to implement statistical analysis with phantoms of any design. In contrast to manual statistical methods, all calculation, analysis of results, and test elements positions changes in the image of phantom are implemented by computer. This paper describes the user interface and functionality of StatPhantom software, its opportunities and advantages in the assessment of various imaging modalities, and the diagnostic preference of an observer. The results obtained by the conventional ROC analysis, manual, and computerized statistical methods are analyzed. Different designs of phantoms are considered.

  5. Oxygen octahedra picker: A software tool to extract quantitative information from STEM images.

    PubMed

    Wang, Yi; Salzberger, Ute; Sigle, Wilfried; Eren Suyolcu, Y; van Aken, Peter A

    2016-09-01

    In perovskite oxide based materials and hetero-structures there are often strong correlations between oxygen octahedral distortions and functionality. Thus, atomistic understanding of the octahedral distortion, which requires accurate measurements of atomic column positions, will greatly help to engineer their properties. Here, we report the development of a software tool to extract quantitative information of the lattice and of BO6 octahedral distortions from STEM images. Center-of-mass and 2D Gaussian fitting methods are implemented to locate positions of individual atom columns. The precision of atomic column distance measurements is evaluated on both simulated and experimental images. The application of the software tool is demonstrated using practical examples.

  6. Vertical bone measurements from cone beam computed tomography images using different software packages.

    PubMed

    Vasconcelos, Taruska Ventorini; Neves, Frederico Sampaio; Moraes, Lívia Almeida Bueno; Freitas, Deborah Queiroz

    2015-01-01

    This article aimed at comparing the accuracy of linear measurement tools of different commercial software packages. Eight fully edentulous dry mandibles were selected for this study. Incisor, canine, premolar, first molar and second molar regions were selected. Cone beam computed tomography (CBCT) images were obtained with i-CAT Next Generation. Linear bone measurements were performed by one observer on the cross-sectional images using three different software packages: XoranCat®, OnDemand3D® and KDIS3D®, all able to assess DICOM images. In addition, 25% of the sample was reevaluated for the purpose of reproducibility. The mandibles were sectioned to obtain the gold standard for each region. Intraclass coefficients (ICC) were calculated to examine the agreement between the two periods of evaluation; the one-way analysis of variance performed with the post-hoc Dunnett test was used to compare each of the software-derived measurements with the gold standard. The ICC values were excellent for all software packages. The least difference between the software-derived measurements and the gold standard was obtained with the OnDemand3D and KDIS3D (-0.11 and -0.14 mm, respectively), and the greatest, with the XoranCAT (+0.25 mm). However, there was no statistical significant difference between the measurements obtained with the different software packages and the gold standard (p> 0.05). In conclusion, linear bone measurements were not influenced by the software package used to reconstruct the image from CBCT DICOM data.

  7. A software framework for diagnostic medical image perception with feedback, and a novel perception visualization technique

    NASA Astrophysics Data System (ADS)

    Phillips, Peter W.; Manning, David J.; Donovan, Tim; Crawford, Trevor; Higham, Stephen

    2005-04-01

    This paper describes a software framework and analysis tool to support the collection and analysis of eye movement and perceptual feedback data for a variety of diagnostic imaging modalities. The framework allows the rapid creation of experiment software that can display a collection of medical images of a particular modality, capture eye trace data, and record marks added to an image by the observer, together with their final decision. There are also a number of visualisation techniques for the display of eye trace information. The analysis tool supports the comparison of individual eye traces for a particular observer or traces from multiple observers for a particular image. Saccade and fixation data can be visualised, with user control of fixation identification functions and properties. Observer markings are displayed, and predefined regions of interest are supported. The software also supports some interactive and multi-image modalities. The analysis tool includes a novel visualisation of scan paths across multi-image modalities. Using an exploded 3D view of a stack of MRI scan sections, an observer's scan path can be shown traversing between images, in addition to inspecting them.

  8. Plume Ascent Tracker: Interactive Matlab software for analysis of ascending plumes in image data

    NASA Astrophysics Data System (ADS)

    Valade, S. A.; Harris, A. J. L.; Cerminara, M.

    2014-05-01

    This paper presents Matlab-based software designed to track and analyze an ascending plume as it rises above its source, in image data. It reads data recorded in various formats (video files, image files, or web-camera image streams), and at various wavelengths (infrared, visible, or ultra-violet). Using a set of filters which can be set interactively, the plume is first isolated from its background. A user-friendly interface then allows tracking of plume ascent and various parameters that characterize plume evolution during emission and ascent. These include records of plume height, velocity, acceleration, shape, volume, ash (fine-particle) loading, spreading rate, entrainment coefficient and inclination angle, as well as axial and radial profiles for radius and temperature (if data are radiometric). Image transformations (dilatation, rotation, resampling) can be performed to create new images with a vent-centered metric coordinate system. Applications may interest both plume observers (monitoring agencies) and modelers. For the first group, the software is capable of providing quantitative assessments of plume characteristics from image data, for post-event analysis or in near real-time analysis. For the second group, extracted data can serve as benchmarks for plume ascent models, and as inputs for cloud dispersal models. We here describe the software's tracking methodology and main graphical interfaces, using thermal infrared image data of an ascending volcanic ash plume at Santiaguito volcano.

  9. Towards a multi-site international public dataset for the validation of retinal image analysis software.

    PubMed

    Trucco, Emanuele; Ruggeri, Alfredo

    2013-01-01

    This paper discusses concisely the main issues and challenges posed by the validation of retinal image analysis algorithms. It is designed to set the discussion for the IEEE EBMC 2013 invited session "From laboratory to clinic: the validation of retinal image processing tools ". The session carries forward an international initiative started at EMBC 2011, Boston, which resulted in the first large-consensus paper (14 international sites) on the validation of retinal image processing software, appearing in IOVS. This paper is meant as a focus for the session discussion, but the ubiquity and importance of validation makes its contents, arguably, of interest for the wider medical image processing community.

  10. Web-based interactive 2D/3D medical image processing and visualization software.

    PubMed

    Mahmoudi, Seyyed Ehsan; Akhondi-Asl, Alireza; Rahmani, Roohollah; Faghih-Roohi, Shahrooz; Taimouri, Vahid; Sabouri, Ahmad; Soltanian-Zadeh, Hamid

    2010-05-01

    There are many medical image processing software tools available for research and diagnosis purposes. However, most of these tools are available only as local applications. This limits the accessibility of the software to a specific machine, and thus the data and processing power of that application are not available to other workstations. Further, there are operating system and processing power limitations which prevent such applications from running on every type of workstation. By developing web-based tools, it is possible for users to access the medical image processing functionalities wherever the internet is available. In this paper, we introduce a pure web-based, interactive, extendable, 2D and 3D medical image processing and visualization application that requires no client installation. Our software uses a four-layered design consisting of an algorithm layer, web-user-interface layer, server communication layer, and wrapper layer. To compete with extendibility of the current local medical image processing software, each layer is highly independent of other layers. A wide range of medical image preprocessing, registration, and segmentation methods are implemented using open source libraries. Desktop-like user interaction is provided by using AJAX technology in the web-user-interface. For the visualization functionality of the software, the VRML standard is used to provide 3D features over the web. Integration of these technologies has allowed implementation of our purely web-based software with high functionality without requiring powerful computational resources in the client side. The user-interface is designed such that the users can select appropriate parameters for practical research and clinical studies.

  11. Despeckle filtering software toolbox for ultrasound imaging of the common carotid artery.

    PubMed

    Loizou, Christos P; Theofanous, Charoula; Pantziaris, Marios; Kasparis, Takis

    2014-04-01

    Ultrasound imaging of the common carotid artery (CCA) is a non-invasive tool used in medicine to assess the severity of atherosclerosis and monitor its progression through time. It is also used in border detection and texture characterization of the atherosclerotic carotid plaque in the CCA, the identification and measurement of the intima-media thickness (IMT) and the lumen diameter that all are very important in the assessment of cardiovascular disease (CVD). Visual perception, however, is hindered by speckle, a multiplicative noise, that degrades the quality of ultrasound B-mode imaging. Noise reduction is therefore essential for improving the visual observation quality or as a pre-processing step for further automated analysis, such as image segmentation of the IMT and the atherosclerotic carotid plaque in ultrasound images. In order to facilitate this preprocessing step, we have developed in MATLAB(®) a unified toolbox that integrates image despeckle filtering (IDF), texture analysis and image quality evaluation techniques to automate the pre-processing and complement the disease evaluation in ultrasound CCA images. The proposed software, is based on a graphical user interface (GUI) and incorporates image normalization, 10 different despeckle filtering techniques (DsFlsmv, DsFwiener, DsFlsminsc, DsFkuwahara, DsFgf, DsFmedian, DsFhmedian, DsFad, DsFnldif, DsFsrad), image intensity normalization, 65 texture features, 15 quantitative image quality metrics and objective image quality evaluation. The software is publicly available in an executable form, which can be downloaded from http://www.cs.ucy.ac.cy/medinfo/. It was validated on 100 ultrasound images of the CCA, by comparing its results with quantitative visual analysis performed by a medical expert. It was observed that the despeckle filters DsFlsmv, and DsFhmedian improved image quality perception (based on the expert's assessment and the image texture and quality metrics). It is anticipated that the

  12. MMX-I: data-processing software for multimodal X-ray imaging and tomography

    PubMed Central

    Bergamaschi, Antoine; Medjoubi, Kadda; Messaoudi, Cédric; Marco, Sergio; Somogyi, Andrea

    2016-01-01

    A new multi-platform freeware has been developed for the processing and reconstruction of scanning multi-technique X-ray imaging and tomography datasets. The software platform aims to treat different scanning imaging techniques: X-ray fluorescence, phase, absorption and dark field and any of their combinations, thus providing an easy-to-use data processing tool for the X-ray imaging user community. A dedicated data input stream copes with the input and management of large datasets (several hundred GB) collected during a typical multi-technique fast scan at the Nanoscopium beamline and even on a standard PC. To the authors’ knowledge, this is the first software tool that aims at treating all of the modalities of scanning multi-technique imaging and tomography experiments. PMID:27140159

  13. Development of HydroImage, A User Friendly Hydrogeophysical Characterization Software

    SciTech Connect

    Mok, Chin Man; Hubbard, Susan; Chen, Jinsong; Suribhatla, Raghu; Kaback, Dawn Samara

    2014-01-29

    HydroImage, user friendly software that utilizes high-resolution geophysical data for estimating hydrogeological parameters in subsurface strate, was developed under this grant. HydroImage runs on a personal computer platform to promote broad use by hydrogeologists to further understanding of subsurface processes that govern contaminant fate, transport, and remediation. The unique software provides estimates of hydrogeological properties over continuous volumes of the subsurface, whereas previous approaches only allow estimation of point locations. thus, this unique tool can be used to significantly enhance site conceptual models and improve design and operation of remediation systems. The HydroImage technical approach uses statistical models to integrate geophysical data with borehole geological data and hydrological measurements to produce hydrogeological parameter estimates as 2-D or 3-D images.

  14. Enhancement of the Shared Graphics Workspace.

    DTIC Science & Technology

    1986-05-01

    RESPONSIBLE INDIVIDUAL 22b. TELEPHONE (Include Area Code) 22c OFFICE SYMBOL Ginger Winn (202) 694-5919 DARPA DO FORM 1473, 84 MAR 83 APR edition may...photograph and the image is less sharp. Text images appear on the SGWS red -green-blue (RGB) monitor in black on amber, with a resolution of 1,024 x 512 pixels...boards no longer send extra characters when an operator presses a button for longer than usual. The system confused the operator by showing all options

  15. Digital processing of side-scan sonar data with the Woods Hole image processing system software

    USGS Publications Warehouse

    Paskevich, Valerie F.

    1992-01-01

    Since 1985, the Branch of Atlantic Marine Geology has been involved in collecting, processing and digitally mosaicking high and low-resolution side-scan sonar data. Recent development of a UNIX-based image-processing software system includes a series of task specific programs for processing side-scan sonar data. This report describes the steps required to process the collected data and to produce an image that has equal along- and across-track resol

  16. Digital image measurement of specimen deformation based on CCD cameras and Image J software: an application to human pelvic biomechanics

    NASA Astrophysics Data System (ADS)

    Jia, Yongwei; Cheng, Liming; Yu, Guangrong; Lou, Yongjian; Yu, Yan; Chen, Bo; Ding, Zuquan

    2008-03-01

    A method of digital image measurement of specimen deformation based on CCD cameras and Image J software was developed. This method was used to measure the biomechanics behavior of human pelvis. Six cadaveric specimens from the third lumbar vertebra to the proximal 1/3 part of femur were tested. The specimens without any structural abnormalities were dissected of all soft tissue, sparing the hip joint capsules and the ligaments of the pelvic ring and floor. Markers with black dot on white background were affixed to the key regions of the pelvis. Axial loading from the proximal lumbar was applied by MTS in the gradient of 0N to 500N, which simulated the double feet standing stance. The anterior and lateral images of the specimen were obtained through two CCD cameras. Based on Image J software, digital image processing software, which can be freely downloaded from the National Institutes of Health, digital 8-bit images were processed. The procedure includes the recognition of digital marker, image invert, sub-pixel reconstruction, image segmentation, center of mass algorithm based on weighted average of pixel gray values. Vertical displacements of S1 (the first sacral vertebrae) in front view and micro-angular rotation of sacroiliac joint in lateral view were calculated according to the marker movement. The results of digital image measurement showed as following: marker image correlation before and after deformation was excellent. The average correlation coefficient was about 0.983. According to the 768 × 576 pixels image (pixel size 0.68mm × 0.68mm), the precision of the displacement detected in our experiment was about 0.018 pixels and the comparatively error could achieve 1.11\\perthou. The average vertical displacement of S1 of the pelvis was 0.8356+/-0.2830mm under vertical load of 500 Newtons and the average micro-angular rotation of sacroiliac joint in lateral view was 0.584+/-0.221°. The load-displacement curves obtained from our optical measure system

  17. Enhancing Image Characteristics of Retinal Images of Aggressive Posterior Retinopathy of Prematurity Using a Novel Software, (RetiView)

    PubMed Central

    Jayadev, Chaitra; Vinekar, Anand; Mohanachandra, Poornima; Desai, Samit; Suveer, Amit; Mangalesh, Shwetha; Bauer, Noel; Shetty, Bhujang

    2015-01-01

    Purpose. To report pilot data from a novel image analysis software “RetiView,” to highlight clinically relevant information in RetCam images of infants with aggressive posterior retinopathy of prematurity (APROP). Methods. Twenty-three imaging sessions of consecutive infants of Asian Indian origin with clinically diagnosed APROP underwent three protocols (Grey Enhanced (GE), Color Enhanced (CE), and “Vesselness Measure” (VNM)) of the software. The postprocessed images were compared to baseline data from the archived unprocessed images and clinical exam by the retinopathy of prematurity (ROP) specialist for anterior extent of the vessels, capillary nonperfusion zones (CNP), loops, hemorrhages, and flat neovascularization. Results. There was better visualization of tortuous loops in the GE protocol (56.5%); “bald” zones within the CNP zones (26.1%), hemorrhages (13%), and edge of the disease (34.8%) in the CE images; neovascularization on both GE and CE protocols (13% each); clinically relevant information in cases with poor pupillary dilatation (8.7%); anterior extent of vessels on the VNM protocol (13%) effecting a “reclassification” from zone 1 to zone 2 posterior. Conclusions. RetiView is a noninvasive and inexpensive method of customized image enhancement to detect clinically difficult characteristics in a subset of APROP images with a potential to influence treatment planning. PMID:26240830

  18. Enhancing Image Characteristics of Retinal Images of Aggressive Posterior Retinopathy of Prematurity Using a Novel Software, (RetiView).

    PubMed

    Jayadev, Chaitra; Vinekar, Anand; Mohanachandra, Poornima; Desai, Samit; Suveer, Amit; Mangalesh, Shwetha; Bauer, Noel; Shetty, Bhujang

    2015-01-01

    Purpose. To report pilot data from a novel image analysis software "RetiView," to highlight clinically relevant information in RetCam images of infants with aggressive posterior retinopathy of prematurity (APROP). Methods. Twenty-three imaging sessions of consecutive infants of Asian Indian origin with clinically diagnosed APROP underwent three protocols (Grey Enhanced (GE), Color Enhanced (CE), and "Vesselness Measure" (VNM)) of the software. The postprocessed images were compared to baseline data from the archived unprocessed images and clinical exam by the retinopathy of prematurity (ROP) specialist for anterior extent of the vessels, capillary nonperfusion zones (CNP), loops, hemorrhages, and flat neovascularization. Results. There was better visualization of tortuous loops in the GE protocol (56.5%); "bald" zones within the CNP zones (26.1%), hemorrhages (13%), and edge of the disease (34.8%) in the CE images; neovascularization on both GE and CE protocols (13% each); clinically relevant information in cases with poor pupillary dilatation (8.7%); anterior extent of vessels on the VNM protocol (13%) effecting a "reclassification" from zone 1 to zone 2 posterior. Conclusions. RetiView is a noninvasive and inexpensive method of customized image enhancement to detect clinically difficult characteristics in a subset of APROP images with a potential to influence treatment planning.

  19. Capturing a failure of an ASIC in-situ, using infrared radiometry and image processing software

    NASA Technical Reports Server (NTRS)

    Ruiz, Ronald P.

    2003-01-01

    Failures in electronic devices can sometimes be tricky to locate-especially if they are buried inside radiation-shielded containers designed to work in outer space. Such was the case with a malfunctioning ASIC (Application Specific Integrated Circuit) that was drawing excessive power at a specific temperature during temperature cycle testing. To analyze the failure, infrared radiometry (thermography) was used in combination with image processing software to locate precisely where the power was being dissipated at the moment the failure took place. The IR imaging software was used to make the image of the target and background, appear as unity. As testing proceeded and the failure mode was reached, temperature changes revealed the precise location of the fault. The results gave the design engineers the information they needed to fix the problem. This paper describes the techniques and equipment used to accomplish this failure analysis.

  20. IHE cross-enterprise document sharing for imaging: interoperability testing software

    PubMed Central

    2010-01-01

    Background With the deployments of Electronic Health Records (EHR), interoperability testing in healthcare is becoming crucial. EHR enables access to prior diagnostic information in order to assist in health decisions. It is a virtual system that results from the cooperation of several heterogeneous distributed systems. Interoperability between peers is therefore essential. Achieving interoperability requires various types of testing. Implementations need to be tested using software that simulates communication partners, and that provides test data and test plans. Results In this paper we describe a software that is used to test systems that are involved in sharing medical images within the EHR. Our software is used as part of the Integrating the Healthcare Enterprise (IHE) testing process to test the Cross Enterprise Document Sharing for imaging (XDS-I) integration profile. We describe its architecture and functionalities; we also expose the challenges encountered and discuss the elected design solutions. Conclusions EHR is being deployed in several countries. The EHR infrastructure will be continuously evolving to embrace advances in the information technology domain. Our software is built on a web framework to allow for an easy evolution with web technology. The testing software is publicly available; it can be used by system implementers to test their implementations. It can also be used by site integrators to verify and test the interoperability of systems, or by developers to understand specifications ambiguities, or to resolve implementations difficulties. PMID:20858241

  1. Software for MR image overlay guided needle insertions: the clinical translation process

    NASA Astrophysics Data System (ADS)

    Ungi, Tamas; U-Thainual, Paweena; Fritz, Jan; Iordachita, Iulian I.; Flammang, Aaron J.; Carrino, John A.; Fichtinger, Gabor

    2013-03-01

    PURPOSE: Needle guidance software using augmented reality image overlay was translated from the experimental phase to support preclinical and clinical studies. Major functional and structural changes were needed to meet clinical requirements. We present the process applied to fulfill these requirements, and selected features that may be applied in the translational phase of other image-guided surgical navigation systems. METHODS: We used an agile software development process for rapid adaptation to unforeseen clinical requests. The process is based on iterations of operating room test sessions, feedback discussions, and software development sprints. The open-source application framework of 3D Slicer and the NA-MIC kit provided sufficient flexibility and stable software foundations for this work. RESULTS: All requirements were addressed in a process with 19 operating room test iterations. Most features developed in this phase were related to workflow simplification and operator feedback. CONCLUSION: Efficient and affordable modifications were facilitated by an open source application framework and frequent clinical feedback sessions. Results of cadaver experiments show that software requirements were successfully solved after a limited number of operating room tests.

  2. A software to digital image processing to be used in the voxel phantom development.

    PubMed

    Vieira, J W; Lima, F R A

    2009-11-15

    Anthropomorphic models used in computational dosimetry, also denominated phantoms, are based on digital images recorded from scanning of real people by Computed Tomography (CT) or Magnetic Resonance Imaging (MRI). The voxel phantom construction requests computational processing for transformations of image formats, to compact two-dimensional (2-D) images forming of three-dimensional (3-D) matrices, image sampling and quantization, image enhancement, restoration and segmentation, among others. Hardly the researcher of computational dosimetry will find all these available abilities in single software, and almost always this difficulty presents as a result the decrease of the rhythm of his researches or the use, sometimes inadequate, of alternative tools. The need to integrate the several tasks mentioned above to obtain an image that can be used in an exposure computational model motivated the development of the Digital Image Processing (DIP) software, mainly to solve particular problems in Dissertations and Thesis developed by members of the Grupo de Pesquisa em Dosimetria Numérica (GDN/CNPq). Because of this particular objective, the software uses the Portuguese idiom in their implementations and interfaces. This paper presents the second version of the DIP, whose main changes are the more formal organization on menus and menu items, and menu for digital image segmentation. Currently, the DIP contains the menus Fundamentos, Visualizações, Domínio Espacial, Domínio de Frequências, Segmentações and Estudos. Each menu contains items and sub-items with functionalities that, usually, request an image as input and produce an image or an attribute in the output. The DIP reads edits and writes binary files containing the 3-D matrix corresponding to a stack of axial images from a given geometry that can be a human body or other volume of interest. It also can read any type of computational image and to make conversions. When the task involves only an output image

  3. Software tools of the Computis European project to process mass spectrometry images.

    PubMed

    Robbe, Marie-France; Both, Jean-Pierre; Prideaux, Brendan; Klinkert, Ivo; Picaud, Vincent; Schramm, Thorsten; Hester, Atfons; Guevara, Victor; Stoeckli, Markus; Roempp, Andreas; Heeren, Ron M A; Spengler, Bernhard; Gala, Olivier; Haan, Serge

    2014-01-01

    Among the needs usually expressed by teams using mass spectrometry imaging, one that often arises is that for user-friendly software able to manage huge data volumes quickly and to provide efficient assistance for the interpretation of data. To answer this need, the Computis European project developed several complementary software tools to process mass spectrometry imaging data. Data Cube Explorer provides a simple spatial and spectral exploration for matrix-assisted laser desorption/ionisation-time of flight (MALDI-ToF) and time of flight-secondary-ion mass spectrometry (ToF-SIMS) data. SpectViewer offers visualisation functions, assistance to the interpretation of data, classification functionalities, peak list extraction to interrogate biological database and image overlay, and it can process data issued from MALDI-ToF, ToF-SIMS and desorption electrospray ionisation (DESI) equipment. EasyReg2D is able to register two images, in American Standard Code for Information Interchange (ASCII) format, issued from different technologies. The collaboration between the teams was hampered by the multiplicity of equipment and data formats, so the project also developed a common data format (imzML) to facilitate the exchange of experimental data and their interpretation by the different software tools. The BioMap platform for visualisation and exploration of MALDI-ToF and DESI images was adapted to parse imzML files, enabling its access to all project partners and, more globally, to a larger community of users. Considering the huge advantages brought by the imzML standard format, a specific editor (vBrowser) for imzML files and converters from proprietary formats to imzML were developed to enable the use of the imzML format by a broad scientific community. This initiative paves the way toward the development of a large panel of software tools able to process mass spectrometry imaging datasets in the future.

  4. OsiriX: an open-source software for navigating in multidimensional DICOM images.

    PubMed

    Rosset, Antoine; Spadola, Luca; Ratib, Osman

    2004-09-01

    A multidimensional image navigation and display software was designed for display and interpretation of large sets of multidimensional and multimodality images such as combined PET-CT studies. The software is developed in Objective-C on a Macintosh platform under the MacOS X operating system using the GNUstep development environment. It also benefits from the extremely fast and optimized 3D graphic capabilities of the OpenGL graphic standard widely used for computer games optimized for taking advantage of any hardware graphic accelerator boards available. In the design of the software special attention was given to adapt the user interface to the specific and complex tasks of navigating through large sets of image data. An interactive jog-wheel device widely used in the video and movie industry was implemented to allow users to navigate in the different dimensions of an image set much faster than with a traditional mouse or on-screen cursors and sliders. The program can easily be adapted for very specific tasks that require a limited number of functions, by adding and removing tools from the program's toolbar and avoiding an overwhelming number of unnecessary tools and functions. The processing and image rendering tools of the software are based on the open-source libraries ITK and VTK. This ensures that all new developments in image processing that could emerge from other academic institutions using these libraries can be directly ported to the OsiriX program. OsiriX is provided free of charge under the GNU open-source licensing agreement at http://homepage.mac.com/rossetantoine/osirix.

  5. A New Effort for Atmospherical Forecast: Meteorological Image Processing Software (MIPS) for Astronomical Observations

    NASA Astrophysics Data System (ADS)

    Shameoni Niaei, M.; Kilic, Y.; Yildiran, B. E.; Yüzlükoglu, F.; Yesilyaprak, C.

    2016-12-01

    We have described a new software (MIPS) about the analysis and image processing of the meteorological satellite (Meteosat) data for an astronomical observatory. This software will be able to help to make some atmospherical forecast (cloud, humidity, rain) using meteosat data for robotic telescopes. MIPS uses a python library for Eumetsat data that aims to be completely open-source and licenced under GNU/General Public Licence (GPL). MIPS is a platform independent and uses h5py, numpy, and PIL with the general-purpose and high-level programming language Python and the QT framework.

  6. Comparison of human observers and CDCOM software reading for CDMAM images

    NASA Astrophysics Data System (ADS)

    Lanconelli, Nico; Rivetti, Stefano; Golinelli, Paola; Serafini, Marco; Bertolini, Marco; Borasi, Giovanni

    2007-03-01

    Contrast-detail analysis is one the most common way for the assessment of the performance of an imaging system. Usually, the reading of phantoms, such as CDMAM, is obtained by human observers. The main drawbacks of this practice is the presence of inter-observer variability and the great amount of time needed. However, software programs are available, for reading CDMAM images in an automatic way. In this paper we present a comparison of human and software reading of CDMAM images coming from three different FFDM clinical units. Images were acquired at different exposures in the same conditions for the three systems. Once software has completed the reading, the interpretation of the results is achieved on the same way used for the human case. CDCOM results are consistent with human analysis, if we consider figures such as COR and IQF. On the other hand, we find out some discrepancies along the CD curves obtained by human observers, with respect to those estimated by automated CDCOM analysis.

  7. MedXViewer: an extensible web-enabled software package for medical imaging

    NASA Astrophysics Data System (ADS)

    Looney, P. T.; Young, K. C.; Mackenzie, Alistair; Halling-Brown, Mark D.

    2014-03-01

    MedXViewer (Medical eXtensible Viewer) is an application designed to allow workstation-independent, PACS-less viewing and interaction with anonymised medical images (e.g. observer studies). The application was initially implemented for use in digital mammography and tomosynthesis but the flexible software design allows it to be easily extended to other imaging modalities. Regions of interest can be identified by a user and any associated information about a mark, an image or a study can be added. The questions and settings can be easily configured depending on the need of the research allowing both ROC and FROC studies to be performed. The extensible nature of the design allows for other functionality and hanging protocols to be available for each study. Panning, windowing, zooming and moving through slices are all available while modality-specific features can be easily enabled e.g. quadrant zooming in mammographic studies. MedXViewer can integrate with a web-based image database allowing results and images to be stored centrally. The software and images can be downloaded remotely from this centralised data-store. Alternatively, the software can run without a network connection where the images and results can be encrypted and stored locally on a machine or external drive. Due to the advanced workstation-style functionality, the simple deployment on heterogeneous systems over the internet without a requirement for administrative access and the ability to utilise a centralised database, MedXViewer has been used for running remote paper-less observer studies and is capable of providing a training infrastructure and co-ordinating remote collaborative viewing sessions (e.g. cancer reviews, interesting cases).

  8. Technical Note: DIRART- A software suite for deformable image registration and adaptive radiotherapy research

    SciTech Connect

    Yang Deshan; Brame, Scott; El Naqa, Issam; Aditya, Apte; Wu Yu; Murty Goddu, S.; Mutic, Sasa; Deasy, Joseph O.; Low, Daniel A.

    2011-01-15

    Purpose: Recent years have witnessed tremendous progress in image guide radiotherapy technology and a growing interest in the possibilities for adapting treatment planning and delivery over the course of treatment. One obstacle faced by the research community has been the lack of a comprehensive open-source software toolkit dedicated for adaptive radiotherapy (ART). To address this need, the authors have developed a software suite called the Deformable Image Registration and Adaptive Radiotherapy Toolkit (DIRART). Methods: DIRART is an open-source toolkit developed in MATLAB. It is designed in an object-oriented style with focus on user-friendliness, features, and flexibility. It contains four classes of DIR algorithms, including the newer inverse consistency algorithms to provide consistent displacement vector field in both directions. It also contains common ART functions, an integrated graphical user interface, a variety of visualization and image-processing features, dose metric analysis functions, and interface routines. These interface routines make DIRART a powerful complement to the Computational Environment for Radiotherapy Research (CERR) and popular image-processing toolkits such as ITK. Results: DIRART provides a set of image processing/registration algorithms and postprocessing functions to facilitate the development and testing of DIR algorithms. It also offers a good amount of options for DIR results visualization, evaluation, and validation. Conclusions: By exchanging data with treatment planning systems via DICOM-RT files and CERR, and by bringing image registration algorithms closer to radiotherapy applications, DIRART is potentially a convenient and flexible platform that may facilitate ART and DIR research.

  9. Apero, AN Open Source Bundle Adjusment Software for Automatic Calibration and Orientation of Set of Images

    NASA Astrophysics Data System (ADS)

    Pierrot Deseilligny, M.; Clery, I.

    2011-09-01

    IGN has developed a set of photogrammetric tools, APERO and MICMAC, for computing 3D models from set of images. This software, developed initially for its internal needs are now delivered as open source code. This paper focuses on the presentation of APERO the orientation software. Compared to some other free software initiatives, it is probably more complex but also more complete, its targeted user is rather professionals (architects, archaeologist, geomophologist) than people. APERO uses both computer vision approach for estimation of initial solution and photogrammetry for a rigorous compensation of the total error; it has a large library of parametric model of distortion allowing a precise modelization of all the kind of pinhole camera we know, including several model of fish-eye; there is also several tools for geo-referencing the result. The results are illustrated on various application, including the data-set of 3D-Arch workshop.

  10. Workspace Analysis and Optimization of 3-PUU Parallel Mechanism in Medicine Base on Genetic Algorithm

    PubMed Central

    Hou, Yongchao; Zhao, Yang

    2015-01-01

    A novel 3-PUU parallel robot was put forward, on which kinematic analysis was conducted to obtain its inverse kinematics solution, and on this basis, the limitations of the sliding pair and the Hooke joint on the workspace were analyzed. Moreover, the workspace was solved through the three dimensional limit search method, and then optimization analysis was performed on the workspace of this parallel robot, which laid the foundations for the configuration design and further analysis of the parallel mechanism, with the result indicated that this type of robot was equipped with promising application prospect. In addition that, the workspace after optimization can meet more requirements of patients. PMID:26628930

  11. Workspace Analysis and Optimization of 3-PUU Parallel Mechanism in Medicine Base on Genetic Algorithm.

    PubMed

    Hou, Yongchao; Zhao, Yang

    2015-01-01

    A novel 3-PUU parallel robot was put forward, on which kinematic analysis was conducted to obtain its inverse kinematics solution, and on this basis, the limitations of the sliding pair and the Hooke joint on the workspace were analyzed. Moreover, the workspace was solved through the three dimensional limit search method, and then optimization analysis was performed on the workspace of this parallel robot, which laid the foundations for the configuration design and further analysis of the parallel mechanism, with the result indicated that this type of robot was equipped with promising application prospect. In addition that, the workspace after optimization can meet more requirements of patients.

  12. The Performance Evaluation of Multi-Image 3d Reconstruction Software with Different Sensors

    NASA Astrophysics Data System (ADS)

    Mousavi, V.; Khosravi, M.; Ahmadi, M.; Noori, N.; Naveh, A. Hosseini; Varshosaz, M.

    2015-12-01

    Today, multi-image 3D reconstruction is an active research field and generating three dimensional model of the objects is one the most discussed issues in Photogrammetry and Computer Vision that can be accomplished using range-based or image-based methods. Very accurate and dense point clouds generated by range-based methods such as structured light systems and laser scanners has introduced them as reliable tools in the industry. Image-based 3D digitization methodologies offer the option of reconstructing an object by a set of unordered images that depict it from different viewpoints. As their hardware requirements are narrowed down to a digital camera and a computer system, they compose an attractive 3D digitization approach, consequently, although range-based methods are generally very accurate, image-based methods are low-cost and can be easily used by non-professional users. One of the factors affecting the accuracy of the obtained model in image-based methods is the software and algorithm used to generate three dimensional model. These algorithms are provided in the form of commercial software, open source and web-based services. Another important factor in the accuracy of the obtained model is the type of sensor used. Due to availability of mobile sensors to the public, popularity of professional sensors and the advent of stereo sensors, a comparison of these three sensors plays an effective role in evaluating and finding the optimized method to generate three-dimensional models. Lots of research has been accomplished to identify a suitable software and algorithm to achieve an accurate and complete model, however little attention is paid to the type of sensors used and its effects on the quality of the final model. The purpose of this paper is deliberation and the introduction of an appropriate combination of a sensor and software to provide a complete model with the highest accuracy. To do this, different software, used in previous studies, were compared and

  13. The Spectral Image Processing System (SIPS): Software for integrated analysis of AVIRIS data

    NASA Technical Reports Server (NTRS)

    Kruse, F. A.; Lefkoff, A. B.; Boardman, J. W.; Heidebrecht, K. B.; Shapiro, A. T.; Barloon, P. J.; Goetz, A. F. H.

    1992-01-01

    The Spectral Image Processing System (SIPS) is a software package developed by the Center for the Study of Earth from Space (CSES) at the University of Colorado, Boulder, in response to a perceived need to provide integrated tools for analysis of imaging spectrometer data both spectrally and spatially. SIPS was specifically designed to deal with data from the Airborne Visible/Infrared Imaging Spectrometer (AVIRIS) and the High Resolution Imaging Spectrometer (HIRIS), but was tested with other datasets including the Geophysical and Environmental Research Imaging Spectrometer (GERIS), GEOSCAN images, and Landsat TM. SIPS was developed using the 'Interactive Data Language' (IDL). It takes advantage of high speed disk access and fast processors running under the UNIX operating system to provide rapid analysis of entire imaging spectrometer datasets. SIPS allows analysis of single or multiple imaging spectrometer data segments at full spatial and spectral resolution. It also allows visualization and interactive analysis of image cubes derived from quantitative analysis procedures such as absorption band characterization and spectral unmixing. SIPS consists of three modules: SIPS Utilities, SIPS_View, and SIPS Analysis. SIPS version 1.1 is described below.

  14. Simulation of tomosynthesis images based on an anthropomorphic software breast tissue phantom

    NASA Astrophysics Data System (ADS)

    Ruiter, Nicole V.; Zhang, Cuiping; Bakic, Predrag R.; Carton, Ann-Katherine; Kuo, Johnny; Maidment, Andrew D. A.

    2008-03-01

    The aim of this work is to provide a simulation framework for generation of synthetic tomosynthesis images to be used for evaluation of future developments in the field of tomosynthesis. An anthropomorphic software tissue phantom was previously used in a number of applications for evaluation of acquisition modalities and image post-processing algorithms for mammograms. This software phantom has been extended for similar use with tomosynthesis. The new features of the simulation framework include a finite element deformation model to obtain realistic mammographic deformation and projection simulation for a variety of tomosynthesis geometries. The resulting projections are provided in DICOM format to be applicable for clinically applied reconstruction algorithms. Examples of simulations using parameters of a currently applied clinical setup are presented. The overall simulation model is generic, allowing multiple degrees of freedom to cover anatomical variety in the amount of glandular tissue, degrees of compression, material models for breast tissues, and tomosynthesis geometries.

  15. 2D-CELL: image processing software for extraction and analysis of 2-dimensional cellular structures

    NASA Astrophysics Data System (ADS)

    Righetti, F.; Telley, H.; Leibling, Th. M.; Mocellin, A.

    1992-01-01

    2D-CELL is a software package for the processing and analyzing of photographic images of cellular structures in a largely interactive way. Starting from a binary digitized image, the programs extract the line network (skeleton) of the structure and determine the graph representation that best models it. Provision is made for manually correcting defects such as incorrect node positions or dangling bonds. Then a suitable algorithm retrieves polygonal contours which define individual cells — local boundary curvatures are neglected for simplicity. Using elementary analytical geometry relations, a range of metric and topological parameters describing the population are then computed, organized into statistical distributions and graphically displayed.

  16. TiLIA: a software package for image analysis of firefly flash patterns.

    PubMed

    Konno, Junsuke; Hatta-Ohashi, Yoko; Akiyoshi, Ryutaro; Thancharoen, Anchana; Silalom, Somyot; Sakchoowong, Watana; Yiu, Vor; Ohba, Nobuyoshi; Suzuki, Hirobumi

    2016-05-01

    As flash signaling patterns of fireflies are species specific, signal-pattern analysis is important for understanding this system of communication. Here, we present time-lapse image analysis (TiLIA), a free open-source software package for signal and flight pattern analyses of fireflies that uses video-recorded image data. TiLIA enables flight path tracing of individual fireflies and provides frame-by-frame coordinates and light intensity data. As an example of TiLIA capabilities, we demonstrate flash pattern analysis of the fireflies Luciola cruciata and L. lateralis during courtship behavior.

  17. SOFI Simulation Tool: A Software Package for Simulating and Testing Super-Resolution Optical Fluctuation Imaging

    PubMed Central

    Sharipov, Azat; Geissbuehler, Stefan; Leutenegger, Marcel; Vandenberg, Wim; Dedecker, Peter; Hofkens, Johan; Lasser, Theo

    2016-01-01

    Super-resolution optical fluctuation imaging (SOFI) allows one to perform sub-diffraction fluorescence microscopy of living cells. By analyzing the acquired image sequence with an advanced correlation method, i.e. a high-order cross-cumulant analysis, super-resolution in all three spatial dimensions can be achieved. Here we introduce a software tool for a simple qualitative comparison of SOFI images under simulated conditions considering parameters of the microscope setup and essential properties of the biological sample. This tool incorporates SOFI and STORM algorithms, displays and describes the SOFI image processing steps in a tutorial-like fashion. Fast testing of various parameters simplifies the parameter optimization prior to experimental work. The performance of the simulation tool is demonstrated by comparing simulated results with experimentally acquired data. PMID:27583365

  18. SOFI Simulation Tool: A Software Package for Simulating and Testing Super-Resolution Optical Fluctuation Imaging.

    PubMed

    Girsault, Arik; Lukes, Tomas; Sharipov, Azat; Geissbuehler, Stefan; Leutenegger, Marcel; Vandenberg, Wim; Dedecker, Peter; Hofkens, Johan; Lasser, Theo

    2016-01-01

    Super-resolution optical fluctuation imaging (SOFI) allows one to perform sub-diffraction fluorescence microscopy of living cells. By analyzing the acquired image sequence with an advanced correlation method, i.e. a high-order cross-cumulant analysis, super-resolution in all three spatial dimensions can be achieved. Here we introduce a software tool for a simple qualitative comparison of SOFI images under simulated conditions considering parameters of the microscope setup and essential properties of the biological sample. This tool incorporates SOFI and STORM algorithms, displays and describes the SOFI image processing steps in a tutorial-like fashion. Fast testing of various parameters simplifies the parameter optimization prior to experimental work. The performance of the simulation tool is demonstrated by comparing simulated results with experimentally acquired data.

  19. Vobi One: a data processing software package for functional optical imaging

    PubMed Central

    Takerkart, Sylvain; Katz, Philippe; Garcia, Flavien; Roux, Sébastien; Reynaud, Alexandre; Chavane, Frédéric

    2014-01-01

    Optical imaging is the only technique that allows to record the activity of a neuronal population at the mesoscopic scale. A large region of the cortex (10–20 mm diameter) is directly imaged with a CCD camera while the animal performs a behavioral task, producing spatio-temporal data with an unprecedented combination of spatial and temporal resolutions (respectively, tens of micrometers and milliseconds). However, researchers who have developed and used this technique have relied on heterogeneous software and methods to analyze their data. In this paper, we introduce Vobi One, a software package entirely dedicated to the processing of functional optical imaging data. It has been designed to facilitate the processing of data and the comparison of different analysis methods. Moreover, it should help bring good analysis practices to the community because it relies on a database and a standard format for data handling and it provides tools that allow producing reproducible research. Vobi One is an extension of the BrainVISA software platform, entirely written with the Python programming language, open source and freely available for download at https://trac.int.univ-amu.fr/vobi_one. PMID:24478623

  20. SPLASSH: Open source software for camera-based high-speed, multispectral in-vivo optical image acquisition.

    PubMed

    Sun, Ryan; Bouchard, Matthew B; Hillman, Elizabeth M C

    2010-08-02

    Camera-based in-vivo optical imaging can provide detailed images of living tissue that reveal structure, function, and disease. High-speed, high resolution imaging can reveal dynamic events such as changes in blood flow and responses to stimulation. Despite these benefits, commercially available scientific cameras rarely include software that is suitable for in-vivo imaging applications, making this highly versatile form of optical imaging challenging and time-consuming to implement. To address this issue, we have developed a novel, open-source software package to control high-speed, multispectral optical imaging systems. The software integrates a number of modular functions through a custom graphical user interface (GUI) and provides extensive control over a wide range of inexpensive IEEE 1394 Firewire cameras. Multispectral illumination can be incorporated through the use of off-the-shelf light emitting diodes which the software synchronizes to image acquisition via a programmed microcontroller, allowing arbitrary high-speed illumination sequences. The complete software suite is available for free download. Here we describe the software's framework and provide details to guide users with development of this and similar software.

  1. Image 100 procedures manual development: Applications system library definition and Image 100 software definition

    NASA Technical Reports Server (NTRS)

    Guseman, L. F., Jr.; Decell, H. P., Jr.

    1975-01-01

    An outline for an Image 100 procedures manual for Earth Resources Program image analysis was developed which sets forth guidelines that provide a basis for the preparation and updating of an Image 100 Procedures Manual. The scope of the outline was limited to definition of general features of a procedures manual together with special features of an interactive system. Computer programs were identified which should be implemented as part of an applications oriented library for the system.

  2. A Survey of DICOM Viewer Software to Integrate Clinical Research and Medical Imaging.

    PubMed

    Haak, Daniel; Page, Charles-E; Deserno, Thomas M

    2016-04-01

    The digital imaging and communications in medicine (DICOM) protocol is the leading standard for image data management in healthcare. Imaging biomarkers and image-based surrogate endpoints in clinical trials and medical registries require DICOM viewer software with advanced functionality for visualization and interfaces for integration. In this paper, a comprehensive evaluation of 28 DICOM viewers is performed. The evaluation criteria are obtained from application scenarios in clinical research rather than patient care. They include (i) platform, (ii) interface, (iii) support, (iv) two-dimensional (2D), and (v) three-dimensional (3D) viewing. On the average, 4.48 and 1.43 of overall 8 2D and 5 3D image viewing criteria are satisfied, respectively. Suitable DICOM interfaces for central viewing in hospitals are provided by GingkoCADx, MIPAV, and OsiriX Lite. The viewers ImageJ, MicroView, MIPAV, and OsiriX Lite offer all included 3D-rendering features for advanced viewing. Interfaces needed for decentral viewing in web-based systems are offered by Oviyam, Weasis, and Xero. Focusing on open source components, MIPAV is the best candidate for 3D imaging as well as DICOM communication. Weasis is superior for workflow optimization in clinical trials. Our evaluation shows that advanced visualization and suitable interfaces can also be found in the open source field and not only in commercial products.

  3. Study on image processing of panoramic X-ray using deviation improvement software.

    PubMed

    Kim, Tae-Gon; Lee, Yang-Sun; Kim, Young-Pyo; Park, Yong-Pil; Cheon, Min-Woo

    2014-01-01

    Utilization of panoramic X-ray device is getting wider. Panoramic X-ray has low resolution than general X-ray device and it occurs to distortion by deviation of image synthesis. Due to structural problems, it has been used restrictively to identify of tooth structure, not for whole head. Therefore, it designed and produced panoramic X-ray device which is possible to diagnostic coverage can be extended and had to be adjusted interval control between X-ray generator and image processing for whole of Maxillofacia's diagnosis. Produced panoramic X-ray device is composed basically of short image synthesis. In addition, it was confirmed the results by used the device which was applied deviation of the brightness of the image, filter to improve the location of the deviation and interpolation method. In this study, it was used 13 images including the front. It occurs to brightness deviation, position deviation, and geometric correction when synthesis of image, but it had been solved by deviation improvement software and a change of CCD camera's scan line which is used for image acquisition. Therefore, it confirmed expansion possibility of utilization range to commonly used panoramic X-ray device.

  4. A User Assessment of Workspaces in Selected Music Education Computer Laboratories.

    ERIC Educational Resources Information Center

    Badolato, Michael Jeremy

    A study of 120 students selected from the user populations of four music education computer laboratories was conducted to determine the applicability of current ergonomic and environmental design guidelines in satisfying the needs of users of educational computing workspaces. Eleven categories of workspace factors were organized into a…

  5. Development of an Open Source Image-Based Flow Modeling Software - SimVascular

    NASA Astrophysics Data System (ADS)

    Updegrove, Adam; Merkow, Jameson; Schiavazzi, Daniele; Wilson, Nathan; Marsden, Alison; Shadden, Shawn

    2014-11-01

    SimVascular (www.simvascular.org) is currently the only comprehensive software package that provides a complete pipeline from medical image data segmentation to patient specific blood flow simulation. This software and its derivatives have been used in hundreds of conference abstracts and peer-reviewed journal articles, as well as the foundation of medical startups. SimVascular was initially released in August 2007, yet major challenges and deterrents for new adopters were the requirement of licensing three expensive commercial libraries utilized by the software, a complicated build process, and a lack of documentation, support and organized maintenance. In the past year, the SimVascular team has made significant progress to integrate open source alternatives for the linear solver, solid modeling, and mesh generation commercial libraries required by the original public release. In addition, the build system, available distributions, and graphical user interface have been significantly enhanced. Finally, the software has been updated to enable users to directly run simulations using models and boundary condition values, included in the Vascular Model Repository (vascularmodel.org). In this presentation we will briefly overview the capabilities of the new SimVascular 2.0 release. National Science Foundation.

  6. CONRAD—A software framework for cone-beam imaging in radiology

    PubMed Central

    Maier, Andreas; Hofmann, Hannes G.; Berger, Martin; Fischer, Peter; Schwemmer, Chris; Wu, Haibo; Müller, Kerstin; Hornegger, Joachim; Choi, Jang-Hwan; Riess, Christian; Keil, Andreas; Fahrig, Rebecca

    2013-01-01

    Purpose: In the community of x-ray imaging, there is a multitude of tools and applications that are used in scientific practice. Many of these tools are proprietary and can only be used within a certain lab. Often the same algorithm is implemented multiple times by different groups in order to enable comparison. In an effort to tackle this problem, the authors created CONRAD, a software framework that provides many of the tools that are required to simulate basic processes in x-ray imaging and perform image reconstruction with consideration of nonlinear physical effects. Methods: CONRAD is a Java-based state-of-the-art software platform with extensive documentation. It is based on platform-independent technologies. Special libraries offer access to hardware acceleration such as OpenCL. There is an easy-to-use interface for parallel processing. The software package includes different simulation tools that are able to generate up to 4D projection and volume data and respective vector motion fields. Well known reconstruction algorithms such as FBP, DBP, and ART are included. All algorithms in the package are referenced to a scientific source. Results: A total of 13 different phantoms and 30 processing steps have already been integrated into the platform at the time of writing. The platform comprises 74.000 nonblank lines of code out of which 19% are used for documentation. The software package is available for download at http://conrad.stanford.edu. To demonstrate the use of the package, the authors reconstructed images from two different scanners, a table top system and a clinical C-arm system. Runtimes were evaluated using the RabbitCT platform and demonstrate state-of-the-art runtimes with 2.5 s for the 256 problem size and 12.4 s for the 512 problem size. Conclusions: As a common software framework, CONRAD enables the medical physics community to share algorithms and develop new ideas. In particular this offers new opportunities for scientific collaboration and

  7. CONRAD—A software framework for cone-beam imaging in radiology

    SciTech Connect

    Maier, Andreas; Choi, Jang-Hwan; Riess, Christian; Keil, Andreas; Fahrig, Rebecca; Hofmann, Hannes G.; Berger, Martin; Fischer, Peter; Schwemmer, Chris; Wu, Haibo; Müller, Kerstin; Hornegger, Joachim

    2013-11-15

    Purpose: In the community of x-ray imaging, there is a multitude of tools and applications that are used in scientific practice. Many of these tools are proprietary and can only be used within a certain lab. Often the same algorithm is implemented multiple times by different groups in order to enable comparison. In an effort to tackle this problem, the authors created CONRAD, a software framework that provides many of the tools that are required to simulate basic processes in x-ray imaging and perform image reconstruction with consideration of nonlinear physical effects.Methods: CONRAD is a Java-based state-of-the-art software platform with extensive documentation. It is based on platform-independent technologies. Special libraries offer access to hardware acceleration such as OpenCL. There is an easy-to-use interface for parallel processing. The software package includes different simulation tools that are able to generate up to 4D projection and volume data and respective vector motion fields. Well known reconstruction algorithms such as FBP, DBP, and ART are included. All algorithms in the package are referenced to a scientific source.Results: A total of 13 different phantoms and 30 processing steps have already been integrated into the platform at the time of writing. The platform comprises 74.000 nonblank lines of code out of which 19% are used for documentation. The software package is available for download at http://conrad.stanford.edu. To demonstrate the use of the package, the authors reconstructed images from two different scanners, a table top system and a clinical C-arm system. Runtimes were evaluated using the RabbitCT platform and demonstrate state-of-the-art runtimes with 2.5 s for the 256 problem size and 12.4 s for the 512 problem size.Conclusions: As a common software framework, CONRAD enables the medical physics community to share algorithms and develop new ideas. In particular this offers new opportunities for scientific collaboration and

  8. CAVASS: a computer-assisted visualization and analysis software system - image processing aspects

    NASA Astrophysics Data System (ADS)

    Udupa, Jayaram K.; Grevera, George J.; Odhner, Dewey; Zhuge, Ying; Souza, Andre; Mishra, Shipra; Iwanaga, Tad

    2007-03-01

    The development of the concepts within 3DVIEWNIX and of the software system 3DVIEWNIX itself dates back to the 1970s. Since then, a series of software packages for Computer Assisted Visualization and Analysis (CAVA) of images came out from our group, 3DVIEWNIX released in 1993, being the most recent, and all were distributed with source code. CAVASS, an open source system, is the latest in this series, and represents the next major incarnation of 3DVIEWNIX. It incorporates four groups of operations: IMAGE PROCESSING (including ROI, interpolation, filtering, segmentation, registration, morphological, and algebraic operations), VISUALIZATION (including slice display, reslicing, MIP, surface rendering, and volume rendering), MANIPULATION (for modifying structures and surgery simulation), ANALYSIS (various ways of extracting quantitative information). CAVASS is designed to work on all platforms. Its key features are: (1) most major CAVA operations incorporated; (2) very efficient algorithms and their highly efficient implementations; (3) parallelized algorithms for computationally intensive operations; (4) parallel implementation via distributed computing on a cluster of PCs; (5) interface to other systems such as CAD/CAM software, ITK, and statistical packages; (6) easy to use GUI. In this paper, we focus on the image processing operations and compare the performance of CAVASS with that of ITK. Our conclusions based on assessing performance by utilizing a regular (6 MB), large (241 MB), and a super (873 MB) 3D image data set are as follows: CAVASS is considerably more efficient than ITK, especially in those operations which are computationally intensive. It can handle considerably larger data sets than ITK. It is easy and ready to use in applications since it provides an easy to use GUI. The users can easily build a cluster from ordinary inexpensive PCs and reap the full power of CAVASS inexpensively compared to expensive multiprocessing systems which are less

  9. Intracoronary optical coherence tomography: Clinical and research applications and intravascular imaging software overview.

    PubMed

    Tenekecioglu, Erhan; Albuquerque, Felipe N; Sotomi, Yohei; Zeng, Yaping; Suwannasom, Pannipa; Tateishi, Hiroki; Cavalcante, Rafael; Ishibashi, Yuki; Nakatani, Shimpei; Abdelghani, Mohammad; Dijkstra, Jouke; Bourantas, Christos; Collet, Carlos; Karanasos, Antonios; Radu, Maria; Wang, Ancong; Muramatsu, Takashi; Landmesser, Ulf; Okamura, Takayuki; Regar, Evelyn; Räber, Lorenz; Guagliumi, Giulio; Pyo, Robert T; Onuma, Yoshinobu; Serruys, Patrick W

    2017-01-21

    By providing valuable information about the coronary artery wall and lumen, intravascular imaging may aid in optimizing interventional procedure results and thereby could improve clinical outcomes following percutaneous coronary intervention (PCI). Intravascular optical coherence tomography (OCT) is a light-based technology with a tissue penetration of approximately 1 to 3 mm and provides near histological resolution. It has emerged as a technological breakthrough in intravascular imaging with multiple clinical and research applications. OCT provides detailed visualization of the vessel following PCI and provides accurate assessment of post-procedural stent performance including detection of edge dissection, stent struts apposition, tissue prolapse, and healing parameters. Additionally, it can provide accurate characterization of plaque morphology and provides key information to optimize post-procedural outcomes. This manuscript aims to review the current clinical and research applications of intracoronary OCT and summarize the analytic OCT imaging software packages currently available. © 2017 Wiley Periodicals, Inc.

  10. The role of camera-bundled image management software in the consumer digital imaging value chain

    NASA Astrophysics Data System (ADS)

    Mueller, Milton; Mundkur, Anuradha; Balasubramanian, Ashok; Chirania, Virat

    2005-02-01

    This research was undertaken by the Convergence Center at the Syracuse University School of Information Studies (www.digital-convergence.info). Project ICONICA, the name for the research, focuses on the strategic implications of digital Images and the CONvergence of Image management and image CApture. Consumer imaging - the activity that we once called "photography" - is now recognized as in the throes of a digital transformation. At the end of 2003, market researchers estimated that about 30% of the households in the U.S. and 40% of the households in Japan owned digital cameras. In 2004, of the 86 million new cameras sold (excluding one-time use cameras), a majority (56%) were estimated to be digital cameras. Sales of photographic film, while still profitable, are declining precipitously.

  11. 76 FR 43724 - In the Matter of Certain Digital Imaging Devices and Related Software; Notice of Commission...

    Federal Register 2010, 2011, 2012, 2013, 2014

    2011-07-21

    ... From the Federal Register Online via the Government Publishing Office INTERNATIONAL TRADE COMMISSION In the Matter of Certain Digital Imaging Devices and Related Software; Notice of Commission... related software by reason of infringement of various claims of United States Patent Nos. 6,031,964 and...

  12. A comprehensive software system for image processing and programming. Final report

    SciTech Connect

    Rasure, J.; Hallett, S.; Jordan, R.

    1994-12-31

    XVision is an example of a comprehensive software system dedicated to the processing of multidimensional scientific data. Because it is comprehensive it is necessarily complex. This design complexity is dealt with by considering XVision as nine overlapping software systems, their components and the required standards. The complexity seen by a user of XVision is minimized by the different interfaces providing access to the image processing routines as well as an interface to ease the incorporation of new routines. The XVision project has stressed the importance of having: (1) interfaces to accommodate users with differing preferences and backgrounds and (2) tools to support the programmer and the scientist. The result is a system that provides a framework for building a powerful research, education and development tool.

  13. Mississippi Company Using NASA Software Program to Provide Unique Imaging Service: DATASTAR Success Story

    NASA Technical Reports Server (NTRS)

    2001-01-01

    DATASTAR, Inc., of Picayune, Miss., has taken NASA's award-winning Earth Resources Laboratory Applications (ELAS) software program and evolved it to the point that the company is now providing a unique, spatial imagery service over the Internet. ELAS was developed in the early 80's to process satellite and airborne sensor imagery data of the Earth's surface into readable and useable information. While there are several software packages on the market that allow the manipulation of spatial data into useable products, this is usually a laborious task. The new program, called the DATASTAR Image Processing Exploitation, or DIPX, Delivery Service, is a subscription service available over the Internet that takes the work out of the equation and provides normalized geo-spatial data in the form of decision products.

  14. Features of the Upgraded Imaging for Hypersonic Experimental Aeroheating Testing (IHEAT) Software

    NASA Technical Reports Server (NTRS)

    Mason, Michelle L.; Rufer, Shann J.

    2016-01-01

    The Imaging for Hypersonic Experimental Aeroheating Testing (IHEAT) software is used at the NASA Langley Research Center to analyze global aeroheating data on wind tunnel models tested in the Langley Aerothermodynamics Laboratory. One-dimensional, semi-infinite heating data derived from IHEAT are used in the design of thermal protection systems for hypersonic vehicles that are exposed to severe aeroheating loads, such as reentry vehicles during descent and landing procedures. This software program originally was written in the PV-WAVE(Registered Trademark) programming language to analyze phosphor thermography data from the two-color, relative-intensity system developed at Langley. To increase the efficiency, functionality, and reliability of IHEAT, the program was migrated to MATLAB(Registered Trademark) syntax and compiled as a stand-alone executable file labeled version 4.0. New features of IHEAT 4.0 include the options to perform diagnostic checks of the accuracy of the acquired data during a wind tunnel test, to extract data along a specified multi-segment line following a feature such as a leading edge or a streamline, and to batch process all of the temporal frame data from a wind tunnel run. Results from IHEAT 4.0 were compared on a pixel level to the output images from the legacy software to validate the program. The absolute differences between the heat transfer data output from the two programs were on the order of 10(exp -5) to 10(exp -7). IHEAT 4.0 replaces the PV-WAVE(Registered Trademark) version as the production software for aeroheating experiments conducted in the hypersonic facilities at NASA Langley.

  15. Scalable, high-performance 3D imaging software platform: system architecture and application to virtual colonoscopy.

    PubMed

    Yoshida, Hiroyuki; Wu, Yin; Cai, Wenli; Brett, Bevin

    2012-01-01

    One of the key challenges in three-dimensional (3D) medical imaging is to enable the fast turn-around time, which is often required for interactive or real-time response. This inevitably requires not only high computational power but also high memory bandwidth due to the massive amount of data that need to be processed. In this work, we have developed a software platform that is designed to support high-performance 3D medical image processing for a wide range of applications using increasingly available and affordable commodity computing systems: multi-core, clusters, and cloud computing systems. To achieve scalable, high-performance computing, our platform (1) employs size-adaptive, distributable block volumes as a core data structure for efficient parallelization of a wide range of 3D image processing algorithms; (2) supports task scheduling for efficient load distribution and balancing; and (3) consists of a layered parallel software libraries that allow a wide range of medical applications to share the same functionalities. We evaluated the performance of our platform by applying it to an electronic cleansing system in virtual colonoscopy, with initial experimental results showing a 10 times performance improvement on an 8-core workstation over the original sequential implementation of the system.

  16. A complete software application for automatic registration of x-ray mammography and magnetic resonance images

    SciTech Connect

    Solves-Llorens, J. A.; Rupérez, M. J. Monserrat, C.; Lloret, M.

    2014-08-15

    Purpose: This work presents a complete and automatic software application to aid radiologists in breast cancer diagnosis. The application is a fully automated method that performs a complete registration of magnetic resonance (MR) images and x-ray (XR) images in both directions (from MR to XR and from XR to MR) and for both x-ray mammograms, craniocaudal (CC), and mediolateral oblique (MLO). This new approximation allows radiologists to mark points in the MR images and, without any manual intervention, it provides their corresponding points in both types of XR mammograms and vice versa. Methods: The application automatically segments magnetic resonance images and x-ray images using the C-Means method and the Otsu method, respectively. It compresses the magnetic resonance images in both directions, CC and MLO, using a biomechanical model of the breast that distinguishes the specific biomechanical behavior of each one of its three tissues (skin, fat, and glandular tissue) separately. It makes a projection of both compressions and registers them with the original XR images using affine transformations and nonrigid registration methods. Results: The application has been validated by two expert radiologists. This was carried out through a quantitative validation on 14 data sets in which the Euclidean distance between points marked by the radiologists and the corresponding points obtained by the application were measured. The results showed a mean error of 4.2 ± 1.9 mm for the MRI to CC registration, 4.8 ± 1.3 mm for the MRI to MLO registration, and 4.1 ± 1.3 mm for the CC and MLO to MRI registration. Conclusions: A complete software application that automatically registers XR and MR images of the breast has been implemented. The application permits radiologists to estimate the position of a lesion that is suspected of being a tumor in an imaging modality based on its position in another different modality with a clinically acceptable error. The results show that the

  17. Efficient 3D rendering for web-based medical imaging software: a proof of concept

    NASA Astrophysics Data System (ADS)

    Cantor-Rivera, Diego; Bartha, Robert; Peters, Terry

    2011-03-01

    Medical Imaging Software (MIS) found in research and in clinical practice, such as in Picture and Archiving Communication Systems (PACS) and Radiology Information Systems (RIS), has not been able to take full advantage of the Internet as a deployment platform. MIS is usually tightly coupled to algorithms that have substantial hardware and software requirements. Consequently, MIS is deployed on thick clients which usually leads project managers to allocate more resources during the deployment phase of the application than the resources that would be allocated if the application were deployed through a web interface.To minimize the costs associated with this scenario, many software providers use or develop plug-ins to provide the delivery platform (internet browser) with the features to load, interact and analyze medical images. Nevertheless there has not been a successful standard means to achieve this goal so far. This paper presents a study of WebGL as an alternative to plug-in development for efficient rendering of 3D medical models and DICOM images. WebGL is a technology that enables the internet browser to have access to the local graphics hardware in a native fashion. Because it is based in OpenGL, a widely accepted graphic industry standard, WebGL is being implemented in most of the major commercial browsers. After a discussion on the details of the technology, a series of experiments are presented to determine the operational boundaries in which WebGL is adequate for MIS. A comparison with current alternatives is also addressed. Finally conclusions and future work are discussed.

  18. A comparison of strain calculation using digital image correlation and finite element software

    NASA Astrophysics Data System (ADS)

    Iadicola, M.; Banerjee, D.

    2016-08-01

    Digital image correlation (DIC) data are being extensively used for many forming applications and for comparisons with finite element analysis (FEA) simulated results. The most challenging comparisons are often in the area of strain localizations just prior to material failure. While qualitative comparisons can be misleading, quantitative comparisons are difficult because of insufficient information about the type of strain output. In this work, strains computed from DIC displacements from a forming limit test are compared to those from three commercial FEA software. Quantitative differences in calculated strains are assessed to determine if the scale of variations seen between FEA and DIC calculated strains constitute real behavior or just calculation differences.

  19. Two-Dimensional Gel Electrophoresis Image Analysis via Dedicated Software Packages.

    PubMed

    Maurer, Martin H

    2016-01-01

    Analyzing two-dimensional gel electrophoretic images is supported by a number of freely and commercially available software. Although the respective program is highly specific, all the programs follow certain standardized algorithms. General steps are: (1) detecting and separating individual spots, (2) subtracting background, (3) creating a reference gel and (4) matching the spots to the reference gel, (5) modifying the reference gel, (6) normalizing the gel measurements for comparison, (7) calibrating for isoelectric point and molecular weight markers, and moreover, (8) constructing a database containing the measurement results and (9) comparing data by statistical and bioinformatic methods.

  20. HYPOTrace: image analysis software for measuring hypocotyl growth and shape demonstrated on Arabidopsis seedlings undergoing photomorphogenesis.

    PubMed

    Wang, Liya; Uilecan, Ioan Vlad; Assadi, Amir H; Kozmik, Christine A; Spalding, Edgar P

    2009-04-01

    Analysis of time series of images can quantify plant growth and development, including the effects of genetic mutations (phenotypes) that give information about gene function. Here is demonstrated a software application named HYPOTrace that automatically extracts growth and shape information from electronic gray-scale images of Arabidopsis (Arabidopsis thaliana) seedlings. Key to the method is the iterative application of adaptive local principal components analysis to extract a set of ordered midline points (medial axis) from images of the seedling hypocotyl. Pixel intensity is weighted to avoid the medial axis being diverted by the cotyledons in areas where the two come in contact. An intensity feature useful for terminating the midline at the hypocotyl apex was isolated in each image by subtracting the baseline with a robust local regression algorithm. Applying the algorithm to time series of images of Arabidopsis seedlings responding to light resulted in automatic quantification of hypocotyl growth rate, apical hook opening, and phototropic bending with high spatiotemporal resolution. These functions are demonstrated here on wild-type, cryptochrome1, and phototropin1 seedlings for the purpose of showing that HYPOTrace generated expected results and to show how much richer the machine-vision description is compared to methods more typical in plant biology. HYPOTrace is expected to benefit seedling development research, particularly in the photomorphogenesis field, by replacing many tedious, error-prone manual measurements with a precise, largely automated computational tool.

  1. The 3D scanner prototype utilize object profile imaging using line laser and octave software

    NASA Astrophysics Data System (ADS)

    Nurdini, Mugi; Manunggal, Trikarsa Tirtadwipa; Samsi, Agus

    2016-11-01

    Three-dimensional scanner or 3D Scanner is a device to reconstruct the real object into digital form on a computer. 3D Scanner is a technology that is being developed, especially in developed countries, where the current 3D Scanner devices is the advanced version with a very expensive prices. This study is basically a simple prototype of 3D Scanner with a very low investment costs. 3D Scanner prototype device consists of a webcam, a rotating desk system controlled by a stepper motor and Arduino UNO, and a line laser. Objects that limit the research is the object with same radius from its center point (object pivot). Scanning is performed by using object profile imaging by line laser which is then captured by the camera and processed by a computer (image processing) using Octave software. On each image acquisition, the scanned object on a rotating desk rotated by a certain degree, so for one full turn multiple images of a number of existing side are finally obtained. Then, the profile of the entire images is extracted in order to obtain digital object dimension. Digital dimension is calibrated by length standard, called gage block. Overall dimensions are then digitally reconstructed into a three-dimensional object. Validation of the scanned object reconstruction of the original object dimensions expressed as a percentage error. Based on the results of data validation, horizontal dimension error is about 5% to 23% and vertical dimension error is about +/- 3%.

  2. Review of free software tools for image analysis of fluorescence cell micrographs.

    PubMed

    Wiesmann, V; Franz, D; Held, C; Münzenmayer, C; Palmisano, R; Wittenberg, T

    2015-01-01

    An increasing number of free software tools have been made available for the evaluation of fluorescence cell micrographs. The main users are biologists and related life scientists with no or little knowledge of image processing. In this review, we give an overview of available tools and guidelines about which tools the users should use to segment fluorescence micrographs. We selected 15 free tools and divided them into stand-alone, Matlab-based, ImageJ-based, free demo versions of commercial tools and data sharing tools. The review consists of two parts: First, we developed a criteria catalogue and rated the tools regarding structural requirements, functionality (flexibility, segmentation and image processing filters) and usability (documentation, data management, usability and visualization). Second, we performed an image processing case study with four representative fluorescence micrograph segmentation tasks with figure-ground and cell separation. The tools display a wide range of functionality and usability. In the image processing case study, we were able to perform figure-ground separation in all micrographs using mainly thresholding. Cell separation was not possible with most of the tools, because cell separation methods are provided only by a subset of the tools and are difficult to parametrize and to use. Most important is that the usability matches the functionality of a tool. To be usable, specialized tools with less functionality need to fulfill less usability criteria, whereas multipurpose tools need a well-structured menu and intuitive graphical user interface.

  3. The i5k Workspace@NAL--enabling genomic data access, visualization and curation of arthropod genomes.

    PubMed

    Poelchau, Monica; Childers, Christopher; Moore, Gary; Tsavatapalli, Vijaya; Evans, Jay; Lee, Chien-Yueh; Lin, Han; Lin, Jun-Wei; Hackett, Kevin

    2015-01-01

    The 5000 arthropod genomes initiative (i5k) has tasked itself with coordinating the sequencing of 5000 insect or related arthropod genomes. The resulting influx of data, mostly from small research groups or communities with little bioinformatics experience, will require visualization, dissemination and curation, preferably from a centralized platform. The National Agricultural Library (NAL) has implemented the i5k Workspace@NAL (http://i5k.nal.usda.gov/) to help meet the i5k initiative's genome hosting needs. Any i5k member is encouraged to contact the i5k Workspace with their genome project details. Once submitted, new content will be accessible via organism pages, genome browsers and BLAST search engines, which are implemented via the open-source Tripal framework, a web interface for the underlying Chado database schema. We also implement the Web Apollo software for groups that choose to curate gene models. New content will add to the existing body of 35 arthropod species, which include species relevant for many aspects of arthropod genomic research, including agriculture, invasion biology, systematics, ecology and evolution, and developmental research.

  4. The i5k Workspace@NAL—enabling genomic data access, visualization and curation of arthropod genomes

    PubMed Central

    Poelchau, Monica; Childers, Christopher; Moore, Gary; Tsavatapalli, Vijaya; Evans, Jay; Lee, Chien-Yueh; Lin, Han; Lin, Jun-Wei; Hackett, Kevin

    2015-01-01

    The 5000 arthropod genomes initiative (i5k) has tasked itself with coordinating the sequencing of 5000 insect or related arthropod genomes. The resulting influx of data, mostly from small research groups or communities with little bioinformatics experience, will require visualization, dissemination and curation, preferably from a centralized platform. The National Agricultural Library (NAL) has implemented the i5k Workspace@NAL (http://i5k.nal.usda.gov/) to help meet the i5k initiative's genome hosting needs. Any i5k member is encouraged to contact the i5k Workspace with their genome project details. Once submitted, new content will be accessible via organism pages, genome browsers and BLAST search engines, which are implemented via the open-source Tripal framework, a web interface for the underlying Chado database schema. We also implement the Web Apollo software for groups that choose to curate gene models. New content will add to the existing body of 35 arthropod species, which include species relevant for many aspects of arthropod genomic research, including agriculture, invasion biology, systematics, ecology and evolution, and developmental research. PMID:25332403

  5. Applying Workspace Limitations in a Velocity-Controlled Robotic Mechanism

    NASA Technical Reports Server (NTRS)

    Abdallah, Muhammad E. (Inventor); Hargrave, Brian (Inventor); Platt, Robert J., Jr. (Inventor)

    2014-01-01

    A robotic system includes a robotic mechanism responsive to velocity control signals, and a permissible workspace defined by a convex-polygon boundary. A host machine determines a position of a reference point on the mechanism with respect to the boundary, and includes an algorithm for enforcing the boundary by automatically shaping the velocity control signals as a function of the position, thereby providing smooth and unperturbed operation of the mechanism along the edges and corners of the boundary. The algorithm is suited for application with higher speeds and/or external forces. A host machine includes an algorithm for enforcing the boundary by shaping the velocity control signals as a function of the reference point position, and a hardware module for executing the algorithm. A method for enforcing the convex-polygon boundary is also provided that shapes a velocity control signal via a host machine as a function of the reference point position.

  6. Software development for ACR-approved phantom-based nuclear medicine tomographic image quality control with cross-platform compatibility

    NASA Astrophysics Data System (ADS)

    Oh, Jungsu S.; Choi, Jae Min; Nam, Ki Pyo; Chae, Sun Young; Ryu, Jin-Sook; Moon, Dae Hyuk; Kim, Jae Seung

    2015-07-01

    Quality control and quality assurance (QC/QA) have been two of the most important issues in modern nuclear medicine (NM) imaging for both clinical practices and academic research. Whereas quantitative QC analysis software is common to modern positron emission tomography (PET) scanners, the QC of gamma cameras and/or single-photon-emission computed tomography (SPECT) scanners has not been sufficiently addressed. Although a thorough standard operating process (SOP) for mechanical and software maintenance may help the QC/QA of a gamma camera and SPECT-computed tomography (CT), no previous study has addressed a unified platform or process to decipher or analyze SPECT phantom images acquired from various scanners thus far. In addition, a few approaches have established cross-platform software to enable the technologists and physicists to assess the variety of SPECT scanners from different manufacturers. To resolve these issues, we have developed Interactive Data Language (IDL)-based in-house software for crossplatform (in terms of not only operating systems (OS) but also manufacturers) analyses of the QC data on an ACR SPECT phantom, which is essential for assessing and assuring the tomographical image quality of SPECT. We applied our devised software to our routine quarterly QC of ACR SPECT phantom images acquired from a number of platforms (OS/manufacturers). Based on our experience, we suggest that our devised software can offer a unified platform that allows images acquired from various types of scanners to be analyzed with great precision and accuracy.

  7. User's Guide for the MapImage Reprojection Software Package, Version 1.01

    USGS Publications Warehouse

    Finn, Michael P.; Trent, Jason R.

    2004-01-01

    Scientists routinely accomplish small-scale geospatial modeling in the raster domain, using high-resolution datasets (such as 30-m data) for large parts of continents and low-resolution to high-resolution datasets for the entire globe. Recently, Usery and others (2003a) expanded on the previously limited empirical work with real geographic data by compiling and tabulating the accuracy of categorical areas in projected raster datasets of global extent. Geographers and applications programmers at the U.S. Geological Survey's (USGS) Mid-Continent Mapping Center (MCMC) undertook an effort to expand and evolve an internal USGS software package, MapImage, or mapimg, for raster map projection transformation (Usery and others, 2003a). Daniel R. Steinwand of Science Applications International Corporation, Earth Resources Observation Systems Data Center in Sioux Falls, S. Dak., originally developed mapimg for the USGS, basing it on the USGS's General Cartographic Transformation Package (GCTP). It operated as a command line program on the Unix operating system. Through efforts at MCMC, and in coordination with Mr. Steinwand, this program has been transformed from an application based on a command line into a software package based on a graphic user interface for Windows, Linux, and Unix machines. Usery and others (2003b) pointed out that many commercial software packages do not use exact projection equations and that even when exact projection equations are used, the software often results in error and sometimes does not complete the transformation for specific projections, at specific resampling resolutions, and for specific singularities. Direct implementation of point-to-point transformation with appropriate functions yields the variety of projections available in these software packages, but implementation with data other than points requires specific adaptation of the equations or prior preparation of the data to allow the transformation to succeed. Additional

  8. Hierarchical Image Segmentation of Remotely Sensed Data using Massively Parallel GNU-LINUX Software

    NASA Technical Reports Server (NTRS)

    Tilton, James C.

    2003-01-01

    A hierarchical set of image segmentations is a set of several image segmentations of the same image at different levels of detail in which the segmentations at coarser levels of detail can be produced from simple merges of regions at finer levels of detail. In [1], Tilton, et a1 describes an approach for producing hierarchical segmentations (called HSEG) and gave a progress report on exploiting these hierarchical segmentations for image information mining. The HSEG algorithm is a hybrid of region growing and constrained spectral clustering that produces a hierarchical set of image segmentations based on detected convergence points. In the main, HSEG employs the hierarchical stepwise optimization (HSWO) approach to region growing, which was described as early as 1989 by Beaulieu and Goldberg. The HSWO approach seeks to produce segmentations that are more optimized than those produced by more classic approaches to region growing (e.g. Horowitz and T. Pavlidis, [3]). In addition, HSEG optionally interjects between HSWO region growing iterations, merges between spatially non-adjacent regions (i.e., spectrally based merging or clustering) constrained by a threshold derived from the previous HSWO region growing iteration. While the addition of constrained spectral clustering improves the utility of the segmentation results, especially for larger images, it also significantly increases HSEG s computational requirements. To counteract this, a computationally efficient recursive, divide-and-conquer, implementation of HSEG (RHSEG) was devised, which includes special code to avoid processing artifacts caused by RHSEG s recursive subdivision of the image data. The recursive nature of RHSEG makes for a straightforward parallel implementation. This paper describes the HSEG algorithm, its recursive formulation (referred to as RHSEG), and the implementation of RHSEG using massively parallel GNU-LINUX software. Results with Landsat TM data are included comparing RHSEG with classic

  9. SNARK09 - a software package for reconstruction of 2D images from 1D projections.

    PubMed

    Klukowska, Joanna; Davidi, Ran; Herman, Gabor T

    2013-06-01

    The problem of reconstruction of slices and volumes from 1D and 2D projections has arisen in a large number of scientific fields (including computerized tomography, electron microscopy, X-ray microscopy, radiology, radio astronomy and holography). Many different methods (algorithms) have been suggested for its solution. In this paper we present a software package, SNARK09, for reconstruction of 2D images from their 1D projections. In the area of image reconstruction, researchers often desire to compare two or more reconstruction techniques and assess their relative merits. SNARK09 provides a uniform framework to implement algorithms and evaluate their performance. It has been designed to treat both parallel and divergent projection geometries and can either create test data (with or without noise) for use by reconstruction algorithms or use data collected by another software or a physical device. A number of frequently-used classical reconstruction algorithms are incorporated. The package provides a means for easy incorporation of new algorithms for their testing, comparison and evaluation. It comes with tools for statistical analysis of the results and ten worked examples.

  10. Developing a new software package for PSF estimation and fitting of adaptive optics images

    NASA Astrophysics Data System (ADS)

    Schreiber, Laura; Diolaiti, Emiliano; Sollima, Antonio; Arcidiacono, Carmelo; Bellazzini, Michele; Ciliegi, Paolo; Falomo, Renato; Foppiani, Italo; Greggio, Laura; Lanzoni, Barbara; Lombini, Matteo; Montegriffo, Paolo; Dalessandro, Emanuele; Massari, Davide

    2012-07-01

    Adaptive Optics (AO) images are characterized by structured Point Spread Function (PSF), with sharp core and extended halo, and by significant variations across the field of view. In order to enable the extraction of high-precision quantitative information and improve the scientific exploitation of AO data, efforts in the PSF modeling and in the integration of suitable models in a code for image analysis are needed. We present the current status of a study on the modeling of AO PSFs based on observational data taken with present telescopes (VLT and LBT). The methods under development include parametric models and hybrid (i.e. analytical / numerical) models adapted to various types of PSFs that can show up in AO images. The specific features of AO data, such as the mainly radial variation of the PSF with respect to the guide star position in single-reference AO, are taken into account as much as possible. The final objective of this project is the development of a flexible software package, based on the Starfinder code (Diolaiati et Al 2000), specifically dedicated to the PSF estimation and to the astrometric and photometric analysis of AO images with complex and spatially variable PSF.

  11. Higher-order continuation for the determination of robot workspace boundaries

    NASA Astrophysics Data System (ADS)

    Hentz, Gauthier; Charpentier, Isabelle; Renaud, Pierre

    2016-02-01

    In the medical and surgical fields, robotics may be of great interest for safer and more accurate procedures. Space constraints for a robotic assistant are however strict. Therefore, roboticists study non-conventional mechanisms with advantageous size/workspace ratios. The determination of mechanism workspace, and primarily its boundaries, is thus of major importance. This Note builds on boundary equation definition, continuation and automatic differentiation to propose a general, accurate, fast and automated method for the determination of mechanism workspace. The method is illustrated with a planar RRR mechanism and a three-dimensional Orthoglide parallel mechanism.

  12. An Effective On-line Polymer Characterization Technique by Using SALS Image Processing Software and Wavelet Analysis

    PubMed Central

    Xian, Guang-ming; Qu, Jin-ping; Zeng, Bi-qing

    2008-01-01

    This paper describes an effective on-line polymer characterization technique by using small-angle light-scattering (SALS) image processing software and wavelet analysis. The phenomenon of small-angle light scattering has been applied to give information about transparent structures on morphology. Real-time visualization of various scattered light image and light intensity matrices is performed by the optical image real-time processing software for SALS. The software can measure the signal intensity of light scattering images, draw the frequency-intensity curves and the amplitude-intensity curves to indicate the variation of the intensity of scattered light in different processing conditions, and estimate the parameters. The current study utilizes a one-dimensional wavelet to delete noise from the original SALS signal and estimate the variation trend of maximum intensity area of the scattered light. So, the system brought the qualitative analysis of the structural information of transparent film success. PMID:19229343

  13. Comprehensive, powerful, efficient, intuitive: a new software framework for clinical imaging applications

    NASA Astrophysics Data System (ADS)

    Augustine, Kurt E.; Holmes, David R., III; Hanson, Dennis P.; Robb, Richard A.

    2006-03-01

    One of the greatest challenges for a software engineer is to create a complex application that is comprehensive enough to be useful to a diverse set of users, yet focused enough for individual tasks to be carried out efficiently with minimal training. This "powerful yet simple" paradox is particularly prevalent in advanced medical imaging applications. Recent research in the Biomedical Imaging Resource (BIR) at Mayo Clinic has been directed toward development of an imaging application framework that provides powerful image visualization/analysis tools in an intuitive, easy-to-use interface. It is based on two concepts very familiar to physicians - Cases and Workflows. Each case is associated with a unique patient and a specific set of routine clinical tasks, or a workflow. Each workflow is comprised of an ordered set of general-purpose modules which can be re-used for each unique workflow. Clinicians help describe and design the workflows, and then are provided with an intuitive interface to both patient data and analysis tools. Since most of the individual steps are common to many different workflows, the use of general-purpose modules reduces development time and results in applications that are consistent, stable, and robust. While the development of individual modules may reflect years of research by imaging scientists, new customized workflows based on the new modules can be developed extremely fast. If a powerful, comprehensive application is difficult to learn and complicated to use, it will be unacceptable to most clinicians. Clinical image analysis tools must be intuitive and effective or they simply will not be used.

  14. Software-based high-level synthesis design of FPGA beamformers for synthetic aperture imaging.

    PubMed

    Amaro, Joao; Yiu, Billy Y S; Falcao, Gabriel; Gomes, Marco A C; Yu, Alfred C H

    2015-05-01

    Field-programmable gate arrays (FPGAs) can potentially be configured as beamforming platforms for ultrasound imaging, but a long design time and skilled expertise in hardware programming are typically required. In this article, we present a novel approach to the efficient design of FPGA beamformers for synthetic aperture (SA) imaging via the use of software-based high-level synthesis techniques. Software kernels (coded in OpenCL) were first developed to stage-wise handle SA beamforming operations, and their corresponding FPGA logic circuitry was emulated through a high-level synthesis framework. After design space analysis, the fine-tuned OpenCL kernels were compiled into register transfer level descriptions to configure an FPGA as a beamformer module. The processing performance of this beamformer was assessed through a series of offline emulation experiments that sought to derive beamformed images from SA channel-domain raw data (40-MHz sampling rate, 12 bit resolution). With 128 channels, our FPGA-based SA beamformer can achieve 41 frames per second (fps) processing throughput (3.44 × 10(8) pixels per second for frame size of 256 × 256 pixels) at 31.5 W power consumption (1.30 fps/W power efficiency). It utilized 86.9% of the FPGA fabric and operated at a 196.5 MHz clock frequency (after optimization). Based on these findings, we anticipate that FPGA and high-level synthesis can together foster rapid prototyping of real-time ultrasound processor modules at low power consumption budgets.

  15. An image-based software tool for screening retinal fundus images using vascular morphology and network transport analysis

    NASA Astrophysics Data System (ADS)

    Clark, Richard D.; Dickrell, Daniel J.; Meadows, David L.

    2014-03-01

    As the number of digital retinal fundus images taken each year grows at an increasing rate, there exists a similarly increasing need for automatic eye disease detection through image-based analysis. A new method has been developed for classifying standard color fundus photographs into both healthy and diseased categories. This classification was based on the calculated network fluid conductance, a function of the geometry and connectivity of the vascular segments. To evaluate the network resistance, the retinal vasculature was first manually separated from the background to ensure an accurate representation of the geometry and connectivity. The arterial and venous networks were then semi-automatically separated into two separate binary images. The connectivity of the arterial network was then determined through a series of morphological image operations. The network comprised of segments of vasculature and points of bifurcation, with each segment having a characteristic geometric and fluid properties. Based on the connectivity and fluid resistance of each vascular segment, an arterial network flow conductance was calculated, which described the ease with which blood can pass through a vascular system. In this work, 27 eyes (13 healthy and 14 diabetic) from patients roughly 65 years in age were evaluated using this methodology. Healthy arterial networks exhibited an average fluid conductance of 419 ± 89 μm3/mPa-s while the average network fluid conductance of the diabetic set was 165 ± 87 μm3/mPa-s (p < 0.001). The results of this new image-based software demonstrated an ability to automatically, quantitatively and efficiently screen diseased eyes from color fundus imagery.

  16. Current status of magnetic resonance imaging (MRI) and ultrasonography fusion software platforms for guidance of prostate biopsies.

    PubMed

    Logan, Jennifer K; Rais-Bahrami, Soroush; Turkbey, Baris; Gomella, Andrew; Amalou, Hayet; Choyke, Peter L; Wood, Bradford J; Pinto, Peter A

    2014-11-01

    Prostate MRI is currently the best diagnostic imaging method for detecting PCa. Magnetic resonance imaging (MRI)/ultrasonography (US) fusion allows the sensitivity and specificity of MRI to be combined with the real-time capabilities of transrectal ultrasonography (TRUS). Multiple approaches and techniques exist for MRI/US fusion and include direct 'in bore' MRI biopsies, cognitive fusion, and MRI/US fusion via software-based image coregistration platforms.

  17. Pre-Hardware Optimization of Spacecraft Image Processing Software Algorithms and Hardware Implementation

    NASA Technical Reports Server (NTRS)

    Kizhner, Semion; Flatley, Thomas P.; Hestnes, Phyllis; Jentoft-Nilsen, Marit; Petrick, David J.; Day, John H. (Technical Monitor)

    2001-01-01

    Spacecraft telemetry rates have steadily increased over the last decade presenting a problem for real-time processing by ground facilities. This paper proposes a solution to a related problem for the Geostationary Operational Environmental Spacecraft (GOES-8) image processing application. Although large super-computer facilities are the obvious heritage solution, they are very costly, making it imperative to seek a feasible alternative engineering solution at a fraction of the cost. The solution is based on a Personal Computer (PC) platform and synergy of optimized software algorithms and re-configurable computing hardware technologies, such as Field Programmable Gate Arrays (FPGA) and Digital Signal Processing (DSP). It has been shown in [1] and [2] that this configuration can provide superior inexpensive performance for a chosen application on the ground station or on-board a spacecraft. However, since this technology is still maturing, intensive pre-hardware steps are necessary to achieve the benefits of hardware implementation. This paper describes these steps for the GOES-8 application, a software project developed using Interactive Data Language (IDL) (Trademark of Research Systems, Inc.) on a Workstation/UNIX platform. The solution involves converting the application to a PC/Windows/RC platform, selected mainly by the availability of low cost, adaptable high-speed RC hardware. In order for the hybrid system to run, the IDL software was modified to account for platform differences. It was interesting to examine the gains and losses in performance on the new platform, as well as unexpected observations before implementing hardware. After substantial pre-hardware optimization steps, the necessity of hardware implementation for bottleneck code in the PC environment became evident and solvable beginning with the methodology described in [1], [2], and implementing a novel methodology for this specific application [6]. The PC-RC interface bandwidth problem for the

  18. Automated software for CCD-image processing and detection of small Solar System bodies

    NASA Astrophysics Data System (ADS)

    Savanevych, V.; Bryukhovetskiy, A.; Sokovikova, N.; Bezkrovniy, M.; Khlamov, S.; Elenin, L.; Movsesian, I.; Dihtyar, M.

    2014-07-01

    Efficiency is a crucial factor in the discovery of near-Earth asteroids (NEA) and potentially-hazardous asteroids. Current asteroid surveys yield many images per night. It is no longer possible for the observer to quickly view these images in the the blinking mode. This cause a serious difficulty for large-aperture wide-field telescopes, capturing up to several tens of asteroids in one image. To achieve better asteroid-survey efficiency it is necessary to design and develop automated software for the frame processing. Currently the CoLiTec software solves the problem of the frame processing for asteroid surveys in the real mode. The automatically detected asteroids are subject to follow-up visual confirmation. The CoLiTec software is in use for the automated detection of asteroids in Andrushivka Astronomical Observatory, in the Russian remote observatory ISON-NM (Mayhill, New Mexico, USA), as well as in the observatory ISON-Kislovodsk and in ISON-Ussuriysk starting from the fall 2013. The CoLiTec led to the first automated asteroid and comet discoveries in the CIS (Commonwealth of Independent States) and Baltic countries. In 2012 (2011) 80 (86) % of observations and 74 (75) % of discoveries of asteroids in these countries were made using the CoLiTec. The comet C/2010 X1 (Elenin), discovered using the CoLiTec on December 10, 2010, was the first comet discovered by a CIS astronomer over the past 20 years. In total, out of 7 recently discovered in the CIS and Baltic countries comets 4 comets were discovered due to the CoLiTec, namely C/2010 X1 (Elenin), P/2011 NO1 (Elenin), C/2012 S1 (ISON), and P/2013 V3 (Nevski). About 500,000 CoLiTec-used measurements were reported to MPC, including over 1,500 preliminary discovered objects. These objects include 21 Jupiter Trojan asteroids, 4 NEAs and 1 Centaur. Three other discovered asteroids were reported via dedicated electronic MPC circulars. In 2012 the CoLiTec users were ranked as No. 10, 13, and 22 in the list of the most

  19. JHelioviewer: Open-Source Software for Discovery and Image Access in the Petabyte Age (Invited)

    NASA Astrophysics Data System (ADS)

    Mueller, D.; Dimitoglou, G.; Langenberg, M.; Pagel, S.; Dau, A.; Nuhn, M.; Garcia Ortiz, J. P.; Dietert, H.; Schmidt, L.; Hughitt, V. K.; Ireland, J.; Fleck, B.

    2010-12-01

    The unprecedented torrent of data returned by the Solar Dynamics Observatory is both a blessing and a barrier: a blessing for making available data with significantly higher spatial and temporal resolution, but a barrier for scientists to access, browse and analyze them. With such staggering data volume, the data is bound to be accessible only from a few repositories and users will have to deal with data sets effectively immobile and practically difficult to download. From a scientist's perspective this poses three challenges: accessing, browsing and finding interesting data while avoiding the proverbial search for a needle in a haystack. To address these challenges, we have developed JHelioviewer, an open-source visualization software that lets users browse large data volumes both as still images and movies. We did so by deploying an efficient image encoding, storage, and dissemination solution using the JPEG 2000 standard. This solution enables users to access remote images at different resolution levels as a single data stream. Users can view, manipulate, pan, zoom, and overlay JPEG 2000 compressed data quickly, without severe network bandwidth penalties. Besides viewing data, the browser provides third-party metadata and event catalog integration to quickly locate data of interest, as well as an interface to the Virtual Solar Observatory to download science-quality data. As part of the Helioviewer Project, JHelioviewer offers intuitive ways to browse large amounts of heterogeneous data remotely and provides an extensible and customizable open-source platform for the scientific community.

  20. A collaborative resource management workspace and project management application for data collection, analysis and visualization: OpenNRM

    NASA Astrophysics Data System (ADS)

    Osti, A.

    2013-12-01

    During the process of research and design for OpenNRM, we imagined a place where diverse groups of people and communities could effectively and efficiently collaborate to manage large-scale environmental problems and projects. Our research revealed the need to combine a variety of software components. Users can explore and analyze a topic while simultaneously develop stories and solve problems in a way that the end result is consumable by their colleagues and the general public. To do this we brought together software modules that are typically separate: Document and Asset Management, GIS and Interactive Mapping, WIKI and Information Libraries, Data Catalogs and Services, Project Management Tools and Storytelling templates. These components, along with others are supported by extensive data catalogs (NWIS, Storet, CDEC, Cuahsi), data analysis tools and web services for a turn-key workspace that will allow you to quickly build project communities and data stories. In this presentation we will show you how our investigation into these collaborative efforts are implemented and working for some of our clients, including the State of California's Sacramento San Joaquin Bay-Delta and San Joaquin River Basin. The case study will display the use of the OpenNRM workspace for real time environmental conditions management, data visualization, project operations, environmental restoration, high frequency monitoring and data reporting. We will demonstrate how scientists and policy makers are working together to tell the story of this complicated and divisive system and how they are becoming better managers of that system. Using the genius of web services, we will show you how OpenNRM was designed to allow you to build your own community while easily sharing data stories, project data, monitoring results, document libraries, interactive maps and datasets with others. We will get into more technical detail by presenting how our data interpolation tools can show high frequency

  1. BAMS2 Workspace: a comprehensive and versatile neuroinformatic platform for collating and processing neuroanatomical connections

    PubMed Central

    Bota, Mihail; Talpalaru, Ştefan; Hintiryan, Houri; Dong, Hong-Wei; Swanson, Larry W.

    2014-01-01

    We present in this paper a novel neuroinformatic platform, the BAMS2 Workspace (http://brancusi1.usc.edu), designed for storing and processing information about gray matter region axonal connections. This de novo constructed module allows registered users to directly collate their data by using a simple and versatile visual interface. It also allows construction and analysis of sets of connections associated with gray matter region nomenclatures from any designated species. The Workspace includes a set of tools allowing the display of data in matrix and networks formats, and the uploading of processed information in visual, PDF, CSV, and Excel formats. Finally, the Workspace can be accessed anonymously by third party systems to create individualized connectivity networks. All features of the BAMS2 Workspace are described in detail, and are demonstrated with connectivity reports collated in BAMS and associated with the rat sensory-motor cortex, medial frontal cortex, and amygdalar regions. PMID:24668342

  2. A software tool for stitching two PET/CT body segments into a single whole-body image set.

    PubMed

    Chang, Tingting; Chang, Guoping; Clark, John W; Rohren, Eric M; Mawlawi, Osama R

    2012-05-10

    A whole-body PET/CT scan extending from the vertex of the head to the toes of the patient is not feasible on a number of commercially available PET/CT scanners due to a limitation in the extent of bed travel on these systems. In such cases, the PET scan has to be divided into two parts: one covering the upper body segment, while the other covering the lower body segment. The aim of this paper is to describe and evaluate, using phantom and patient studies, a software tool that was developed to stitch two body segments and output a single whole-body image set, thereby facilitating the interpretation of whole-body PET scans. A mathematical model was first developed to stitch images from two body segments using three landmarks. The model calculates the relative positions of the landmarks on the two segments and then generates a rigid transformation that aligns these landmarks on the two segments. A software tool was written to implement this model while correcting for radioactive decay between the two body segments, and output a single DICOM whole-body image set with all the necessary tags. One phantom, and six patient studies were conducted to evaluate the performance of the software. In these studies, six radio-opaque markers (BBs) were used as landmarks (three on each leg). All studies were acquired in two body segments with BBs placed in the overlap region of the two segments. The PET/CT images of each segment were then stitched using the software tool to create a single DICOM whole-body PET/CT image. Evaluation of the stitching tool was based on visual inspection, consistency of radiotracer uptake in the two segments, and ability to display the resultant DICOM image set on two independent workstations. The software tool successfully stitched the two segments of the phantom image, and generated a single whole-body DICOM PET/CT image set that had the correct alignment and activity concentration throughout the image. The stitched images were viewed by two independent

  3. Digital mapping of side-scan sonar data with the Woods Hole Image Processing System software

    USGS Publications Warehouse

    Paskevich, Valerie F.

    1992-01-01

    Since 1985, the Branch of Atlantic Marine Geology has been involved in collecting, processing and digitally mosaicking high and low resolution sidescan sonar data. In the past, processing and digital mosaicking has been accomplished with a dedicated, shore-based computer system. Recent development of a UNIX-based image-processing software system includes a series of task specific programs for pre-processing sidescan sonar data. To extend the capabilities of the UNIX-based programs, development of digital mapping techniques have been developed. This report describes the initial development of an automated digital mapping procedure. Included is a description of the programs and steps required to complete the digital mosaicking on a UNIXbased computer system, and a comparison of techniques that the user may wish to select.

  4. Integration of XNAT/PACS, DICOM, and research software for automated multi-modal image analysis

    NASA Astrophysics Data System (ADS)

    Gao, Yurui; Burns, Scott S.; Lauzon, Carolyn B.; Fong, Andrew E.; James, Terry A.; Lubar, Joel F.; Thatcher, Robert W.; Twillie, David A.; Wirt, Michael D.; Zola, Marc A.; Logan, Bret W.; Anderson, Adam W.; Landman, Bennett A.

    2013-03-01

    Traumatic brain injury (TBI) is an increasingly important public health concern. While there are several promising avenues of intervention, clinical assessments are relatively coarse and comparative quantitative analysis is an emerging field. Imaging data provide potentially useful information for evaluating TBI across functional, structural, and microstructural phenotypes. Integration and management of disparate data types are major obstacles. In a multi-institution collaboration, we are collecting electroencephalogy (EEG), structural MRI, diffusion tensor MRI (DTI), and single photon emission computed tomography (SPECT) from a large cohort of US Army service members exposed to mild or moderate TBI who are undergoing experimental treatment. We have constructed a robust informatics backbone for this project centered on the DICOM standard and eXtensible Neuroimaging Archive Toolkit (XNAT) server. Herein, we discuss (1) optimization of data transmission, validation and storage, (2) quality assurance and workflow management, and (3) integration of high performance computing with research software.

  5. Integration of XNAT/PACS, DICOM, and Research Software for Automated Multi-modal Image Analysis

    PubMed Central

    Gao, Yurui; Burns, Scott S.; Lauzon, Carolyn B.; Fong, Andrew E.; James, Terry A.; Lubar, Joel F.; Thatcher, Robert W.; Twillie, David A.; Wirt, Michael D.; Zola, Marc A.; Logan, Bret W.; Anderson, Adam W.; Landman, Bennett A.

    2013-01-01

    Traumatic brain injury (TBI) is an increasingly important public health concern. While there are several promising avenues of intervention, clinical assessments are relatively coarse and comparative quantitative analysis is an emerging field. Imaging data provide potentially useful information for evaluating TBI across functional, structural, and microstructural phenotypes. Integration and management of disparate data types are major obstacles. In a multi-institution collaboration, we are collecting electroencephalogy (EEG), structural MRI, diffusion tensor MRI (DTI), and single photon emission computed tomography (SPECT) from a large cohort of US Army service members exposed to mild or moderate TBI who are undergoing experimental treatment. We have constructed a robust informatics backbone for this project centered on the DICOM standard and eXtensible Neuroimaging Archive Toolkit (XNAT) server. Herein, we discuss (1) optimization of data transmission, validation and storage, (2) quality assurance and workflow management, and (3) integration of high performance computing with research software. PMID:24386548

  6. Comparison between three methods to value lower tear meniscus measured by image software

    NASA Astrophysics Data System (ADS)

    García-Resúa, Carlos; Pena-Verdeal, Hugo; Lira, Madalena; Oliveira, M. Elisabete Real; Giráldez, María. Jesús; Yebra-Pimentel, Eva

    2013-11-01

    To measure different parameters of lower tear meniscus height (TMH) by using photography with open software of measurement. TMH was addressed from lower eyelid to the top of the meniscus (absolute TMH) and to the brightest meniscus reflex (reflex TMH). 121 young healthy subjects were included in the study. The lower tear meniscus was videotaped by a digital camera attached to a slit lamp. Three videos were recorded in central meniscus portion on three different methods: slit lamp without fluorescein instillation, slit lamp with fluorescein instillation and TearscopeTM without fluorescein instillation. Then, a masked observed obtained an image from each video and measured TMH by using open source software of measurement based on Java (NIH ImageJ). Absolute central (TMH-CA), absolute with fluorescein (TMH-F) and absolute using the Tearscope (TMH-Tc) were compared each other as well as reflex central (TMH-CR) and reflex Tearscope (TMH-TcR). Mean +/- S.D. values of TMH-CA, TMH-CR, TMH-F, TMH-Tc and TMH-TcR of 0.209 +/- 0.049, 0.139 +/- 0.031, 0.222 +/- 0.058, 0.175 +/- 0.045 and 0.109 +/- 0.029 mm, respectively were found. Paired t-test was performed for the relationship between TMH-CA - TMH-CR, TMH-CA - TMH-F, TMH-CA - TMH-Tc, TMH-F - TMH-Tc, TMH-Tc - TMH-TcR and TMH-CR - TMH-TcR. In all cases, it was found a significant difference between both variables (all p < 0.008). This study showed a useful tool to objectively measure TMH by photography. Eye care professionals should maintain the same TMH parameter in the follow-up visits, due to the difference between them.

  7. SU-E-J-264: Comparison of Two Commercially Available Software Platforms for Deformable Image Registration

    SciTech Connect

    Tuohy, R; Stathakis, S; Mavroidis, P; Bosse, C; Papanikolaou, N

    2014-06-01

    Purpose: To evaluate and compare the deformable image registration algorithms available in the Velocity (Velocity Medical Solutions, Atlanta, GA) and RayStation (RaySearch Americas, Inc., Garden city NY). Methods: Ten consecutive patient cone beam CTs (CBCT) for each fraction were collected. The CBCTs along with the simulation CT were exported to the Velocity and the RayStation software. Each CBCT was registered using deformable image registration to the simulation CT and the resulting deformable vector matrix was generated. Each registration was visually inspected by a physicist and the prescribing physician. The volumes of the critical organs were calculated for each deformable CT and used for comparison. Results: The resulting deformable registrations revealed differences between the two algorithms. These differences were realized when the organs at risk were contoured on each deformed CBCT. Differences in the order of 10% ±30% in volume were observed for bladder, 17 ±21% for rectum and 16±10% for sigmoid. The prostate and PTV volume differences were in the order of 3±5%. The volumetric differences observed had a respective impact on the DVHs of all organs at risk. Differences of 8–10% in the mean dose were observed for all organs above. Conclusion: Deformable registration is a powerful tool that aids in the definition of critical structures and is often used for the evaluation of daily dose delivered to the patient. It should be noted that extended QA should be performed before clinical implementation of the software and the users should be aware of advantages and limitations of the methods.

  8. Reliability evaluation of I-123 ADAM SPECT imaging using SPM software and AAL ROI methods

    NASA Astrophysics Data System (ADS)

    Yang, Bang-Hung; Tsai, Sung-Yi; Wang, Shyh-Jen; Su, Tung-Ping; Chou, Yuan-Hwa; Chen, Chia-Chieh; Chen, Jyh-Cheng

    2011-08-01

    The level of serotonin was regulated by serotonin transporter (SERT), which is a decisive protein in regulation of serotonin neurotransmission system. Many psychiatric disorders and therapies were also related to concentration of cerebral serotonin. I-123 ADAM was the novel radiopharmaceutical to image SERT in brain. The aim of this study was to measure reliability of SERT densities of healthy volunteers by automated anatomical labeling (AAL) method. Furthermore, we also used statistic parametric mapping (SPM) on a voxel by voxel analysis to find difference of cortex between test and retest of I-123 ADAM single photon emission computed tomography (SPECT) images.Twenty-one healthy volunteers were scanned twice with SPECT at 4 h after intravenous administration of 185 MBq of 123I-ADAM. The image matrix size was 128×128 and pixel size was 3.9 mm. All images were obtained through filtered back-projection (FBP) reconstruction algorithm. Region of interest (ROI) definition was performed based on the AAL brain template in PMOD version 2.95 software package. ROI demarcations were placed on midbrain, pons, striatum, and cerebellum. All images were spatially normalized to the SPECT MNI (Montreal Neurological Institute) templates supplied with SPM2. And each image was transformed into standard stereotactic space, which was matched to the Talairach and Tournoux atlas. Then differences across scans were statistically estimated on a voxel by voxel analysis using paired t-test (population main effect: 2 cond's, 1 scan/cond.), which was applied to compare concentration of SERT between the test and retest cerebral scans.The average of specific uptake ratio (SUR: target/cerebellum-1) of 123I-ADAM binding to SERT in midbrain was 1.78±0.27, pons was 1.21±0.53, and striatum was 0.79±0.13. The cronbach's α of intra-class correlation coefficient (ICC) was 0.92. Besides, there was also no significant statistical finding in cerebral area using SPM2 analysis. This finding might help us

  9. Optimal design of a 3-leg 6-DOF parallel manipulator for a specific workspace

    NASA Astrophysics Data System (ADS)

    Fu, Jianxun; Gao, Feng

    2016-07-01

    Researchers seldom study optimum design of a six-degree-of-freedom(DOF) parallel manipulator with three legs based upon the given workspace. An optimal design method of a novel three-leg six-DOF parallel manipulator(TLPM) is presented. The mechanical structure of this robot is introduced, with this structure the kinematic constrain equations is decoupled. Analytical solutions of the forward kinematics are worked out, one configuration of this robot, including position and orientation of the end-effector are graphically displayed. Then, on the basis of several extreme positions of the kinematic performances, the task workspace is given. An algorithm of optimal designing is introduced to find the smallest dimensional parameters of the proposed robot. Examples illustrate the design results, and a design stability index is introduced, which ensures that the robot remains a safe distance from the boundary of sits actual workspace. Finally, one prototype of the robot is developed based on this method. This method can easily find appropriate kinematic parameters that can size a robot having the smallest workspace enclosing a predefined task workspace. It improves the design efficiency, ensures that the robot has a small mechanical size possesses a large given workspace volume, and meets the lightweight design requirements.

  10. Upper extremity 3D reachable workspace analysis in dystrophinopathy using Kinect

    PubMed Central

    Han, Jay J.; Kurillo, Gregorij; Abresch, Richard T.; de Bie, Evan; Nicorici, Alina; Bajcsy, Ruzena

    2015-01-01

    Introduction An innovative upper extremity 3D reachable workspace outcome measure acquired using Kinect sensor is applied towards Duchenne/Becker muscular dystrophy (DMD/BMD). The validity, sensitivity, and clinical meaningfulness of the novel outcome is examined. Methods Upper extremity function assessment (Brooke scale, NeuroQOL questionnaire) and Kinect-based reachable workspace analyses were conducted in 43 individuals with dystrophinopathy (30-DMD, 13-BMD; ages 7–60) and 46 controls (ages 6–68). Results The reachable workspace measure reliably captured a wide-range of upper extremity impairments encountered in both pediatric and adult, as well as ambulatory and non-ambulatory individuals with dystrophinopathy. Reduced reachable workspaces were noted for the dystrophinopathy cohort compared to controls, and they correlated with Brooke grades. Additionally, progressive reduction in reachable workspace directly correlated with worsening ability to perform activities of daily living, as self-reported on the NeuroQOL. Discussion This study demonstrates the utility and potential of the novel sensor-acquired reachable workspace outcome measure in dystrophinopathy. PMID:25597487

  11. A flexible software architecture for scalable real-time image and video processing applications

    NASA Astrophysics Data System (ADS)

    Usamentiaga, Rubén; Molleda, Julio; García, Daniel F.; Bulnes, Francisco G.

    2012-06-01

    Real-time image and video processing applications require skilled architects, and recent trends in the hardware platform make the design and implementation of these applications increasingly complex. Many frameworks and libraries have been proposed or commercialized to simplify the design and tuning of real-time image processing applications. However, they tend to lack flexibility because they are normally oriented towards particular types of applications, or they impose specific data processing models such as the pipeline. Other issues include large memory footprints, difficulty for reuse and inefficient execution on multicore processors. This paper presents a novel software architecture for real-time image and video processing applications which addresses these issues. The architecture is divided into three layers: the platform abstraction layer, the messaging layer, and the application layer. The platform abstraction layer provides a high level application programming interface for the rest of the architecture. The messaging layer provides a message passing interface based on a dynamic publish/subscribe pattern. A topic-based filtering in which messages are published to topics is used to route the messages from the publishers to the subscribers interested in a particular type of messages. The application layer provides a repository for reusable application modules designed for real-time image and video processing applications. These modules, which include acquisition, visualization, communication, user interface and data processing modules, take advantage of the power of other well-known libraries such as OpenCV, Intel IPP, or CUDA. Finally, we present different prototypes and applications to show the possibilities of the proposed architecture.

  12. Free digital image analysis software helps to resolve equivocal scores in HER2 immunohistochemistry.

    PubMed

    Helin, Henrik O; Tuominen, Vilppu J; Ylinen, Onni; Helin, Heikki J; Isola, Jorma

    2016-02-01

    Evaluation of human epidermal growth factor receptor 2 (HER2) immunohistochemistry (IHC) is subject to interobserver variation and lack of reproducibility. Digital image analysis (DIA) has been shown to improve the consistency and accuracy of the evaluation and its use is encouraged in current testing guidelines. We studied whether digital image analysis using a free software application (ImmunoMembrane) can assist in interpreting HER2 IHC in equivocal 2+ cases. We also compared digital photomicrographs with whole-slide images (WSI) as material for ImmunoMembrane DIA. We stained 750 surgical resection specimens of invasive breast cancers immunohistochemically for HER2 and analysed staining with ImmunoMembrane. The ImmunoMembrane DIA scores were compared with the originally responsible pathologists' visual scores, a researcher's visual scores and in situ hybridisation (ISH) results. The originally responsible pathologists reported 9.1 % positive 3+ IHC scores, for the researcher this was 8.4 % and for ImmunoMembrane 9.5 %. Equivocal 2+ scores were 34 % for the pathologists, 43.7 % for the researcher and 10.1 % for ImmunoMembrane. Negative 0/1+ scores were 57.6 % for the pathologists, 46.8 % for the researcher and 80.8 % for ImmunoMembrane. There were six false positive cases, which were classified as 3+ by ImmunoMembrane and negative by ISH. Six cases were false negative defined as 0/1+ by IHC and positive by ISH. ImmunoMembrane DIA using digital photomicrographs and WSI showed almost perfect agreement. In conclusion, digital image analysis by ImmunoMembrane can help to resolve a majority of equivocal 2+ cases in HER2 IHC, which reduces the need for ISH testing.

  13. SU-E-I-13: Evaluation of Metal Artifact Reduction (MAR) Software On Computed Tomography (CT) Images

    SciTech Connect

    Huang, V; Kohli, K

    2015-06-15

    Purpose: A new commercially available metal artifact reduction (MAR) software in computed tomography (CT) imaging was evaluated with phantoms in the presence of metals. The goal was to assess the ability of the software to restore the CT number in the vicinity of the metals without impacting the image quality. Methods: A Catphan 504 was scanned with a GE Optima RT 580 CT scanner (GE Healthcare, Milwaukee, WI) and the images were reconstructed with and without the MAR software. Both datasets were analyzed with Image Owl QA software (Image Owl Inc, Greenwich, NY). CT number sensitometry, MTF, low contrast, uniformity, noise and spatial accuracy were compared for scans with and without MAR software. In addition, an in-house made phantom was scanned with and without a stainless steel insert at three different locations. The accuracy of the CT number and metal insert dimension were investigated as well. Results: Comparisons between scans with and without MAR algorithm on the Catphan phantom demonstrate similar results for image quality. However, noise was slightly higher for the MAR algorithm. Evaluation of the CT number at various locations of the in-house made phantom was also performed. The baseline HU, obtained from the scan without metal insert, was compared to scans with the stainless steel insert at 3 different locations. The HU difference between the baseline scan versus metal scan was improved when the MAR algorithm was applied. In addition, the physical diameter of the stainless steel rod was over-estimated by the MAR algorithm by 0.9 mm. Conclusion: This work indicates with the presence of metal in CT scans, the MAR algorithm is capable of providing a more accurate CT number without compromising the overall image quality. Future work will include the dosimetric impact on the MAR algorithm.

  14. Real-time telemedicine using shared three-dimensional workspaces over ATM

    NASA Astrophysics Data System (ADS)

    Cahoon, Peter; Forsey, David R.; Hutchison, Susan

    1999-03-01

    During the past five years a high speed ATM network has been developed at UBC that provides a campus testbed, a local testbed to the hospitals, and a National testbed between here and the BADLAB in Ottawa. This testbed has been developed to combine a commercial shared audio/video/whiteboard environment coupled with a shared interactive 3-dimensional solid model. This solid model ranges from a skull reconstructed from a CT scan with muscles and an overlying skin, to a model of the ventricle system of the human brain. Typical interactions among surgeon, radiologist and modeler consist of having image slices of the original scan shared by all and the ability to adjust the surface of the model to conform to each individuals perception of what the final object should look like. The purpose of this interaction can range from forensic reconstruction from partial remains to pre-maxillofacial surgery. A joint project with the forensic unit of the R.C.M.P. in Ottawa using the BADLAB is now in the stages of testing this methodology on a real case beginning with a CT scan of partial remains. A second study underway with the department of Maxiofacial reconstruction at Dalhousie University in Halifax Nova Scotia and concerns a subject who is about to undergo orthognathic surgery, in particular a mandibular advancement. This subject has been MRI scanned, a solid model constructed of the mandible and the virtual surgery constructed on the model. This model and the procedure have been discussed and modified by the modeler and the maxillofacial specialist using these shared workspaces. The procedure will be repeated after the actual surgery to verify the modeled procedure. The advantage of this technique is that none of the specialists need be in the same room, or city. Given the scarcity of time and specialists this methodology shows great promise. In November of this last year a shared live demonstration of this facial modeler was done between Vancouver and Dalhousie University in

  15. Digital map and situation surface: a team-oriented multidisplay workspace for network enabled situation analysis

    NASA Astrophysics Data System (ADS)

    Peinsipp-Byma, E.; Geisler, Jürgen; Bader, Thomas

    2009-05-01

    System concepts for network enabled image-based ISR (intelligence, surveillance, reconnaissance) is the major mission of Fraunhofer IITB's applied research in the area of defence and security solutions. For the TechDemo08 as part of the NATO CNAD POW Defence against terrorism Fraunhofer IITB advanced a new multi display concept to handle the shear amount and high complexity of ISR data acquired by networked, distributed surveillance systems with the objective to support the generation of a common situation picture. Amount and Complexity of ISR data demands an innovative man-machine interface concept for humans to deal with it. The IITB's concept is the Digital Map & Situation Surface. This concept offers to the user a coherent multi display environment combining a horizontal surface for the situation overview from the bird's eye view, an attached vertical display for collateral information and so-called foveatablets as personalized magic lenses in order to obtain high resolved and role-specific information about a focused areaof- interest and to interact with it. In the context of TechDemo08 the Digital Map & Situation Surface served as workspace for team-based situation visualization and analysis. Multiple sea- and landside surveillance components were connected to the system.

  16. JHelioviewer: Open-Source Software for Discovery and Image Access in the Petabyte Age

    NASA Astrophysics Data System (ADS)

    Mueller, D.; Dimitoglou, G.; Garcia Ortiz, J.; Langenberg, M.; Nuhn, M.; Dau, A.; Pagel, S.; Schmidt, L.; Hughitt, V. K.; Ireland, J.; Fleck, B.

    2011-12-01

    The unprecedented torrent of data returned by the Solar Dynamics Observatory is both a blessing and a barrier: a blessing for making available data with significantly higher spatial and temporal resolution, but a barrier for scientists to access, browse and analyze them. With such staggering data volume, the data is accessible only from a few repositories and users have to deal with data sets effectively immobile and practically difficult to download. From a scientist's perspective this poses three challenges: accessing, browsing and finding interesting data while avoiding the proverbial search for a needle in a haystack. To address these challenges, we have developed JHelioviewer, an open-source visualization software that lets users browse large data volumes both as still images and movies. We did so by deploying an efficient image encoding, storage, and dissemination solution using the JPEG 2000 standard. This solution enables users to access remote images at different resolution levels as a single data stream. Users can view, manipulate, pan, zoom, and overlay JPEG 2000 compressed data quickly, without severe network bandwidth penalties. Besides viewing data, the browser provides third-party metadata and event catalog integration to quickly locate data of interest, as well as an interface to the Virtual Solar Observatory to download science-quality data. As part of the ESA/NASA Helioviewer Project, JHelioviewer offers intuitive ways to browse large amounts of heterogeneous data remotely and provides an extensible and customizable open-source platform for the scientific community. In addition, the easy-to-use graphical user interface enables the general public and educators to access, enjoy and reuse data from space missions without barriers.

  17. Caltech/JPL Conference on Image Processing Technology, Data Sources and Software for Commercial and Scientific Applications

    NASA Technical Reports Server (NTRS)

    Redmann, G. H.

    1976-01-01

    Recent advances in image processing and new applications are presented to the user community to stimulate the development and transfer of this technology to industrial and commercial applications. The Proceedings contains 37 papers and abstracts, including many illustrations (some in color) and provides a single reference source for the user community regarding the ordering and obtaining of NASA-developed image-processing software and science data.

  18. Development of a viability standard curve for microencapsulated probiotic bacteria using confocal microscopy and image analysis software.

    PubMed

    Moore, Sarah; Kailasapathy, Kasipathy; Phillips, Michael; Jones, Mark R

    2015-07-01

    Microencapsulation is proposed to protect probiotic strains from food processing procedures and to maintain probiotic viability. Little research has described the in situ viability of microencapsulated probiotics. This study successfully developed a real-time viability standard curve for microencapsulated bacteria using confocal microscopy, fluorescent dyes and image analysis software.

  19. Position, singularity and workspace analysis of 3-PSR-O spatial parallel manipulator

    NASA Astrophysics Data System (ADS)

    Shao, Jiejie; Chen, Wenyu; Fu, Xin

    2015-04-01

    Although the parallel mechanisms have the advantages of high accuracy, velocity, stiffness, and payload capacity, the shortcomings of the space utilization and workspace limit the applications in the confined space. A novel 3 degrees of freedom spatial parallel manipulator 3-PSR-O (prismatic-spherical-revolute) is proposed, which possesses a compact architecture and extended workspace while maintaining the inherent advantages of the parallel mechanisms. The direct-inverse position, singularity and workspace are investigated. The mapping method is adopted in the position analysis, and the closed form solution is derived in the form of a six order equation. The singularity analysis of the mechanism is also carried out based on the geometrical constraints, including six singularity boundaries. A feature boundary, which is independent of the prismatic joints' stroke limit, is obtained by integrating the six singularity boundaries. According to the formation of the reachable workspace, a concept of basic workspace is also introduced and presented in the analytical way. By demarcating the basic workspace along the central height with the feature boundary, the reachable workspace can be derived and analyzed more efficiently. Finally, a comparative study on the space utilization between the 3-PSP parallel mechanism and the new mechanism is also presented. The area of feature boundary of the new mechanism is about 140% of the 3-PSP parallel mechanism, while its installation radius is only 1/2 of the 3-PSP parallel mechanism. The proposed parallel mechanism shows great space utilization, and is ideally suited for applications in confined space occasions such as immersion lithography, nano-imprint etc.

  20. Holistic approach to design and implementation of a medical teleconsultation workspace.

    PubMed

    Czekierda, Łukasz; Malawski, Filip; Wyszkowski, Przemysław

    2015-10-01

    While there are many state-of-the-art approaches to introducing telemedical services in the area of medical imaging, it is hard to point to studies which would address all relevant aspects in a complete and comprehensive manner. In this paper we describe our approach to design and implementation of a universal platform for imaging medicine which is based on our longstanding experience in this area. We claim it is holistic, because, contrary to most of the available studies it addresses all aspects related to creation and utilization of a medical teleconsultation workspace. We present an extensive analysis of requirements, including possible usage scenarios, user needs, organizational and security issues and infrastructure components. We enumerate and analyze multiple usage scenarios related to medical imaging data in treatment, research and educational applications - with typical teleconsultations treated as just one of many possible options. Certain phases common to all these scenarios have been identified, with the resulting classification distinguishing several modes of operation (local vs. remote, collaborative vs. non-interactive etc.). On this basis we propose a system architecture which addresses all of the identified requirements, applying two key concepts: Service Oriented Architecture (SOA) and Virtual Organizations (VO). The SOA paradigm allows us to decompose the functionality of the system into several distinct building blocks, ensuring flexibility and reliability. The VO paradigm defines the cooperation model for all participating healthcare institutions. Our approach is validated by an ICT platform called TeleDICOM II which implements the proposed architecture. All of its main elements are described in detail and cross-checked against the listed requirements. A case study presents the role and usage of the platform in a specific scenario. Finally, our platform is compared with similar systems described into-date studies and available on the market.

  1. An image analysis pipeline for the semi-automated analysis of clinical fMRI images based on freely available software.

    PubMed

    Karmonik, Christof; York, Michele; Grossman, Robert; Kakkar, Ekta; Patel, Krutina; Haykal, Hani; King, David

    2010-03-01

    The technique of functional Magnetic Resonance Imaging (fMRI) has evolved in the last 15 years from a research concept into a clinically relevant medical procedure. In this study, an efficient, semi-automated and cost-effective solution for the analysis of fMRI images acquired in a clinical setting is presented relying heavily on open source software. The core of the pipeline is the software Analysis of Functional NeuroImages (AFNI, National Institute of Mental Health (NIMH)) combined with K-PACS and ImageJ. Its application is illustrated with clinical fMRI exams and with a research study involving comparing subjects diagnosed with Parkinson's disease and age-matched controls.

  2. Software for Collaborative Use of Large Interactive Displays

    NASA Technical Reports Server (NTRS)

    Trimble, Jay; Shab, Thodore; Wales, Roxana; Vera, Alonso; Tollinger, Irene; McCurdy, Michael; Lyubimov, Dmitriy

    2006-01-01

    The MERBoard Collaborative Workspace, which is currently being deployed to support the Mars Exploration Rover (MER) Missions, is the first instantiation of a new computing architecture designed to support collaborative and group computing using computing devices situated in NASA mission operations room. It is a software system for generation of large-screen interactive displays by multiple users

  3. A user-friendly LabVIEW software platform for grating based X-ray phase-contrast imaging.

    PubMed

    Wang, Shenghao; Han, Huajie; Gao, Kun; Wang, Zhili; Zhang, Can; Yang, Meng; Wu, Zhao; Wu, Ziyu

    2015-01-01

    X-ray phase-contrast imaging can provide greatly improved contrast over conventional absorption-based imaging for weakly absorbing samples, such as biological soft tissues and fibre composites. In this study, we introduced an easy and fast way to develop a user-friendly software platform dedicated to the new grating-based X-ray phase-contrast imaging setup at the National Synchrotron Radiation Laboratory of the University of Science and Technology of China. The control of 21 motorized stages, of a piezoelectric stage and of an X-ray tube are achieved with this software, it also covers image acquisition with a flat panel detector for automatic phase stepping scan. Moreover, a data post-processing module for signals retrieval and other custom features are in principle available. With a seamless integration of all the necessary functions in one software package, this platform greatly facilitate users' activities during experimental runs with this grating based X-ray phase contrast imaging setup.

  4. 3D reconstruction of SEM images by use of optical photogrammetry software.

    PubMed

    Eulitz, Mona; Reiss, Gebhard

    2015-08-01

    Reconstruction of the three-dimensional (3D) surface of an object to be examined is widely used for structure analysis in science and many biological questions require information about their true 3D structure. For Scanning Electron Microscopy (SEM) there has been no efficient non-destructive solution for reconstruction of the surface morphology to date. The well-known method of recording stereo pair images generates a 3D stereoscope reconstruction of a section, but not of the complete sample surface. We present a simple and non-destructive method of 3D surface reconstruction from SEM samples based on the principles of optical close range photogrammetry. In optical close range photogrammetry a series of overlapping photos is used to generate a 3D model of the surface of an object. We adapted this method to the special SEM requirements. Instead of moving a detector around the object, the object itself was rotated. A series of overlapping photos was stitched and converted into a 3D model using the software commonly used for optical photogrammetry. A rabbit kidney glomerulus was used to demonstrate the workflow of this adaption. The reconstruction produced a realistic and high-resolution 3D mesh model of the glomerular surface. The study showed that SEM micrographs are suitable for 3D reconstruction by optical photogrammetry. This new approach is a simple and useful method of 3D surface reconstruction and suitable for various applications in research and teaching.

  5. New image processing software for analyzing object size-frequency distributions, geometry, orientation, and spatial distribution

    NASA Astrophysics Data System (ADS)

    Beggan, Ciarán; Hamilton, Christopher W.

    2010-04-01

    Geological Image Analysis Software (GIAS) combines basic tools for calculating object area, abundance, radius, perimeter, eccentricity, orientation, and centroid location, with the first automated method for characterizing the aerial distribution of objects using sample-size-dependent nearest neighbor (NN) statistics. The NN analyses include tests for (1) Poisson, (2) Normalized Poisson, (3) Scavenged k=1, and (4) Scavenged k=2 NN distributions. GIAS is implemented in MATLAB with a Graphical User Interface (GUI) that is available as pre-parsed pseudocode for use with MATLAB, or as a stand-alone application that runs on Windows and Unix systems. GIAS can process raster data (e.g., satellite imagery, photomicrographs, etc.) and tables of object coordinates to characterize the size, geometry, orientation, and spatial organization of a wide range of geological features. This information expedites quantitative measurements of 2D object properties, provides criteria for validating the use of stereology to transform 2D object sections into 3D models, and establishes a standardized NN methodology that can be used to compare the results of different geospatial studies and identify objects using non-morphological parameters.

  6. ORBS: A data reduction software for the imaging Fourier transform spectrometers SpIOMM and SITELLE

    NASA Astrophysics Data System (ADS)

    Martin, T.; Drissen, L.; Joncas, G.

    2012-09-01

    SpIOMM (Spectromètre-Imageur de l'Observatoire du Mont Mégantic) is still the only operational astronomical Imaging Fourier Transform Spectrometer (IFTS) capable of obtaining the visible spectrum of every source of light in a field of view of 12 arc-minutes. Even if it has been designed to work with both outputs of the Michelson interferometer, up to now only one output has been used. Here we present ORBS (Outils de Réduction Binoculaire pour SpIOMM/SITELLE), the reduction software we designed in order to take advantage of the two output data. ORBS will also be used to reduce the data of SITELLE (Spectromètre-Imageur pour l' Étude en Long et en Large des raies d' Émissions) { the direct successor of SpIOMM, which will be in operation at the Canada-France- Hawaii Telescope (CFHT) in early 2013. SITELLE will deliver larger data cubes than SpIOMM (up to 2 cubes of 34 Go each). We thus have made a strong effort in optimizing its performance efficiency in terms of speed and memory usage in order to ensure the best compliance with the quality characteristics discussed with the CFHT team. As a result ORBS is now capable of reducing 68 Go of data in less than 20 hours using only 5 Go of random-access memory (RAM).

  7. I-SPINE: a software package for advances in image-guided and minimally invasive spine procedures

    NASA Astrophysics Data System (ADS)

    Choi, Jae Jeong; Cleary, Kevin R.; Zeng, Jianchao; Gary, Kevin A.; Freedman, Matthew T.; Watson, Vance; Lindisch, David; Mun, Seong K.

    2000-05-01

    While image guidance is now routinely used in the brain in the form of frameless stereotaxy, it is beginning to be more widely used in other clinical areas such as the spine. At Georgetown University Medical Center, we are developing a program to provide advanced visualization and image guidance for minimally invasive spine procedures. This is a collaboration between an engineering-based research group and physicians from the radiology, neurosurgery, and orthopaedics departments. A major component of this work is the ISIS Center Spine Procedures Imaging and Navigation Engine, which is a software package under development as the base platform for technical advances.

  8. Army technology development. IBIS query. Software to support the Image Based Information System (IBIS) expansion for mapping, charting and geodesy

    NASA Technical Reports Server (NTRS)

    Friedman, S. Z.; Walker, R. E.; Aitken, R. B.

    1986-01-01

    The Image Based Information System (IBIS) has been under development at the Jet Propulsion Laboratory (JPL) since 1975. It is a collection of more than 90 programs that enable processing of image, graphical, tabular data for spatial analysis. IBIS can be utilized to create comprehensive geographic data bases. From these data, an analyst can study various attributes describing characteristics of a given study area. Even complex combinations of disparate data types can be synthesized to obtain a new perspective on spatial phenomena. In 1984, new query software was developed enabling direct Boolean queries of IBIS data bases through the submission of easily understood expressions. An improved syntax methodology, a data dictionary, and display software simplified the analysts' tasks associated with building, executing, and subsequently displaying the results of a query. The primary purpose of this report is to describe the features and capabilities of the new query software. A secondary purpose of this report is to compare this new query software to the query software developed previously (Friedman, 1982). With respect to this topic, the relative merits and drawbacks of both approaches are covered.

  9. Analyses of requirements for computer control and data processing experiment subsystems: Image data processing system (IDAPS) software description (7094 version), volume 2

    NASA Technical Reports Server (NTRS)

    1973-01-01

    A description of each of the software modules of the Image Data Processing System (IDAPS) is presented. The changes in the software modules are the result of additions to the application software of the system and an upgrade of the IBM 7094 Mod(1) computer to a 1301 disk storage configuration. Necessary information about IDAPS sofware is supplied to the computer programmer who desires to make changes in the software system or who desires to use portions of the software outside of the IDAPS system. Each software module is documented with: module name, purpose, usage, common block(s) description, method (algorithm of subroutine) flow diagram (if needed), subroutines called, and storage requirements.

  10. The 'Densitometric Image Analysis Software' and its application to determine stepwise equilibrium constants from electrophoretic mobility shift assays.

    PubMed

    van Oeffelen, Liesbeth; Peeters, Eveline; Nguyen Le Minh, Phu; Charlier, Daniël

    2014-01-01

    Current software applications for densitometric analysis, such as ImageJ, QuantityOne (BioRad) and the Intelligent or Advanced Quantifier (Bio Image) do not allow to take the non-linearity of autoradiographic films into account during calibration. As a consequence, quantification of autoradiographs is often regarded as problematic, and phosphorimaging is the preferred alternative. However, the non-linear behaviour of autoradiographs can be described mathematically, so it can be accounted for. Therefore, the 'Densitometric Image Analysis Software' has been developed, which allows to quantify electrophoretic bands in autoradiographs, as well as in gels and phosphorimages, while providing optimized band selection support to the user. Moreover, the program can determine protein-DNA binding constants from Electrophoretic Mobility Shift Assays (EMSAs). For this purpose, the software calculates a chosen stepwise equilibrium constant for each migration lane within the EMSA, and estimates the errors due to non-uniformity of the background noise, smear caused by complex dissociation or denaturation of double-stranded DNA, and technical errors such as pipetting inaccuracies. Thereby, the program helps the user to optimize experimental parameters and to choose the best lanes for estimating an average equilibrium constant. This process can reduce the inaccuracy of equilibrium constants from the usual factor of 2 to about 20%, which is particularly useful when determining position weight matrices and cooperative binding constants to predict genomic binding sites. The MATLAB source code, platform-dependent software and installation instructions are available via the website http://micr.vub.ac.be.

  11. Parameter-based estimation of CT dose index and image quality using an in-house android™-based software

    NASA Astrophysics Data System (ADS)

    Mubarok, S.; Lubis, L. E.; Pawiro, S. A.

    2016-03-01

    Compromise between radiation dose and image quality is essential in the use of CT imaging. CT dose index (CTDI) is currently the primary dosimetric formalisms in CT scan, while the low and high contrast resolutions are aspects indicating the image quality. This study was aimed to estimate CTDIvol and image quality measures through a range of exposure parameters variation. CTDI measurements were performed using PMMA (polymethyl methacrylate) phantom of 16 cm diameter, while the image quality test was conducted by using catphan ® 600. CTDI measurements were carried out according to IAEA TRS 457 protocol using axial scan mode, under varied parameters of tube voltage, collimation or slice thickness, and tube current. Image quality test was conducted accordingly under the same exposure parameters with CTDI measurements. An Android™ based software was also result of this study. The software was designed to estimate the value of CTDIvol with maximum difference compared to actual CTDIvol measurement of 8.97%. Image quality can also be estimated through CNR parameter with maximum difference to actual CNR measurement of 21.65%.

  12. ImageMiner: a software system for comparative analysis of tissue microarrays using content-based image retrieval, high-performance computing, and grid technology

    PubMed Central

    Foran, David J; Yang, Lin; Hu, Jun; Goodell, Lauri A; Reiss, Michael; Wang, Fusheng; Kurc, Tahsin; Pan, Tony; Sharma, Ashish; Saltz, Joel H

    2011-01-01

    Objective and design The design and implementation of ImageMiner, a software platform for performing comparative analysis of expression patterns in imaged microscopy specimens such as tissue microarrays (TMAs), is described. ImageMiner is a federated system of services that provides a reliable set of analytical and data management capabilities for investigative research applications in pathology. It provides a library of image processing methods, including automated registration, segmentation, feature extraction, and classification, all of which have been tailored, in these studies, to support TMA analysis. The system is designed to leverage high-performance computing machines so that investigators can rapidly analyze large ensembles of imaged TMA specimens. To support deployment in collaborative, multi-institutional projects, ImageMiner features grid-enabled, service-based components so that multiple instances of ImageMiner can be accessed remotely and federated. Results The experimental evaluation shows that: (1) ImageMiner is able to support reliable detection and feature extraction of tumor regions within imaged tissues; (2) images and analysis results managed in ImageMiner can be searched for and retrieved on the basis of image-based features, classification information, and any correlated clinical data, including any metadata that have been generated to describe the specified tissue and TMA; and (3) the system is able to reduce computation time of analyses by exploiting computing clusters, which facilitates analysis of larger sets of tissue samples. PMID:21606133

  13. Stereoscopic adapter based system using HMD and image processing software for supporting inner ear operations performed using operating microscope

    NASA Astrophysics Data System (ADS)

    Leśniewski, Marcin; Kujawińska, Malgorzata; Kucharski, Tomasz; Niemczyk, Kazimierz

    2006-02-01

    Recently surgery requires extensive support from imaging technologies in order to increase effectiveness and safety of operations. One of important tasks is to enhance visualisation of quasi-phase (transparent) 3D structures. In this paper authors present a few of practical hardware solutions using of operational stereoscopic microscope with two image acquisition channels, stereoscopic adapter and Helmet Mounted Display (HMD) for stereoscopic visualization of operational field "in real time". Special attention is paid to the development of opto- mechanical unit. The authors focus on searching cheap, accurate and ergonomic solutions. A few proposals are analyzed: typical stereoscopic adapter with two image acquisition channels equipped with developed software for image low contrast enhancement for stereoscopic observation in stereoscopic HMD of operational field, visual - picture adapter (real operational view through microscope channels or processed operational field images observation in "real time").

  14. Workspace Safe Operation of a Force- or Impedance-Controlled Robot

    NASA Technical Reports Server (NTRS)

    Abdallah, Muhammad E. (Inventor); Hargrave, Brian (Inventor); Yamokoski, John D. (Inventor); Strawser, Philip A. (Inventor)

    2013-01-01

    A method of controlling a robotic manipulator of a force- or impedance-controlled robot within an unstructured workspace includes imposing a saturation limit on a static force applied by the manipulator to its surrounding environment, and may include determining a contact force between the manipulator and an object in the unstructured workspace, and executing a dynamic reflex when the contact force exceeds a threshold to thereby alleviate an inertial impulse not addressed by the saturation limited static force. The method may include calculating a required reflex torque to be imparted by a joint actuator to a robotic joint. A robotic system includes a robotic manipulator having an unstructured workspace and a controller that is electrically connected to the manipulator, and which controls the manipulator using force- or impedance-based commands. The controller, which is also disclosed herein, automatically imposes the saturation limit and may execute the dynamic reflex noted above.

  15. Creating the optimal workspace for hospital staff using human centred design.

    PubMed

    Cawood, T; Saunders, E; Drennan, C; Cross, N; Nicholl, D; Kenny, A; Meates, D; Laing, R

    2016-07-01

    We were tasked with creating best possible non-clinical workspace solutions for approximately 450 hospital staff across 11 departments encompassing medical, nursing, allied health, administrative and other support staff. We used a Human-Centred Design process, involving 'Hear, Create and Deliver' stages. We used observations, contextual enquiry and role-specific workshops to understand needs, key interactions and drivers of behaviour. Co-design workshops were then used to explore and prototype-test concepts for the final design. With extensive employee engagement and design process expertise, an innovative solution was created that focussed on meeting the functional workspace needs of a diverse group of staff requiring a range of different spaces, incorporating space constraints and equity. This project demonstrated the strength of engaging employees in an expert-led Human-Centred Design process. We believe this is a successful blueprint process for other institutions to embrace when facing similar workspace design challenges.

  16. A redundant, 6-DOF parallel manipulator structure with improved workspace and dexterity

    SciTech Connect

    Stoughton, R.S.; Salerno, R.; Canfield, S.; Reinholtz, C.

    1994-08-01

    This paper presents a novel manipulator structure which combines two known parallel manipulator structures--a Stewart Platform (SP), and a double octahedral Variable Geometry Truss (VGT). The combined VGT + SP structure is redundant, using nine actuators to realize six-DOF motion. Combining the two structures allows the translational and orientational workspaces of the two individual structures to sum together to a much larger workspace than is generally achievable with parallel manipulator structures. In addition, the VGT portion of the structure allows the configuration of the Stewart Platform to be changed ``on the fly`` from one with a large workspace to one with high dexterity. A useful application of this structure is at the distal end of a truss-based manipulator, where it can serve as a dexterous wrist while preserving an internal passageway for cabling and/or conveyance systems.

  17. Web-based spatial analysis with the ILWIS open source GIS software and satellite images from GEONETCast

    NASA Astrophysics Data System (ADS)

    Lemmens, R.; Maathuis, B.; Mannaerts, C.; Foerster, T.; Schaeffer, B.; Wytzisk, A.

    2009-12-01

    This paper involves easy accessible integrated web-based analysis of satellite images with a plug-in based open source software. The paper is targeted to both users and developers of geospatial software. Guided by a use case scenario, we describe the ILWIS software and its toolbox to access satellite images through the GEONETCast broadcasting system. The last two decades have shown a major shift from stand-alone software systems to networked ones, often client/server applications using distributed geo-(web-)services. This allows organisations to combine without much effort their own data with remotely available data and processing functionality. Key to this integrated spatial data analysis is a low-cost access to data from within a user-friendly and flexible software. Web-based open source software solutions are more often a powerful option for developing countries. The Integrated Land and Water Information System (ILWIS) is a PC-based GIS & Remote Sensing software, comprising a complete package of image processing, spatial analysis and digital mapping and was developed as commercial software from the early nineties onwards. Recent project efforts have migrated ILWIS into a modular, plug-in-based open source software, and provide web-service support for OGC-based web mapping and processing. The core objective of the ILWIS Open source project is to provide a maintainable framework for researchers and software developers to implement training components, scientific toolboxes and (web-) services. The latest plug-ins have been developed for multi-criteria decision making, water resources analysis and spatial statistics analysis. The development of this framework is done since 2007 in the context of 52°North, which is an open initiative that advances the development of cutting edge open source geospatial software, using the GPL license. GEONETCast, as part of the emerging Global Earth Observation System of Systems (GEOSS), puts essential environmental data at the

  18. New Instruments for Survey: on Line Softwares for 3d Recontruction from Images

    NASA Astrophysics Data System (ADS)

    Fratus de Balestrini, E.; Guerra, F.

    2011-09-01

    3d scanning technologies had a significant development and have been widely used in documentation of cultural, architectural and archeological heritages. Modern methods of three-dimensional acquiring and modeling allow to represent an object through a digital model that combines visual potentialities of images (normally used for documentation) to the accuracy of the survey, becoming at the same time support for the visualization that for metric evaluation of any artefact that have an historical or artistic interest, opening up new possibilities for cultural heritage's fruition, cataloging and study. Despite this development, because of the small catchment area and the 3D laser scanner's sophisticated technologies, the cost of these instruments is very high and beyond the reach of most operators in the field of cultural heritages. This is the reason why they have appeared low-cost technologies or even free, allowing anyone to approach the issues of acquisition and 3D modeling, providing tools that allow to create three-dimensional models in a simple and economical way. The research, conducted by the Laboratory of Photogrammetry of the University IUAV of Venice, of which we present here some results, is intended to figure out whether, with Arc3D, it is possible to obtain results that can be somehow comparable, in therms of overall quality, to those of the laser scanner, and/or whether it is possible to integrate them. They were carried out a series of tests on certain types of objects, models made with Arc3D, from raster images, were compared with those obtained using the point clouds from laser scanner. We have also analyzed the conditions for an optimal use of Arc3D: environmental conditions (lighting), acquisition tools (digital cameras) and type and size of objects. After performing the tests described above, we analyzed the patterns generated by Arc3D to check what other graphic representations can be obtained from them: orthophotos and drawings. The research

  19. Time Efficiency and Diagnostic Accuracy of New Automated Myocardial Perfusion Analysis Software in 320-Row CT Cardiac Imaging

    PubMed Central

    Rief, Matthias; Stenzel, Fabian; Kranz, Anisha; Schlattmann, Peter

    2013-01-01

    Objective We aimed to evaluate the time efficiency and diagnostic accuracy of automated myocardial computed tomography perfusion (CTP) image analysis software. Materials and Methods 320-row CTP was performed in 30 patients, and analyses were conducted independently by three different blinded readers by the use of two recent software releases (version 4.6 and novel version 4.71GR001, Toshiba, Tokyo, Japan). Analysis times were compared, and automated epi- and endocardial contour detection was subjectively rated in five categories (excellent, good, fair, poor and very poor). As semi-quantitative perfusion parameters, myocardial attenuation and transmural perfusion ratio (TPR) were calculated for each myocardial segment and agreement was tested by using the intraclass correlation coefficient (ICC). Conventional coronary angiography served as reference standard. Results The analysis time was significantly reduced with the novel automated software version as compared with the former release (Reader 1: 43:08 ± 11:39 min vs. 09:47 ± 04:51 min, Reader 2: 42:07 ± 06:44 min vs. 09:42 ± 02:50 min and Reader 3: 21:38 ± 3:44 min vs. 07:34 ± 02:12 min; p < 0.001 for all). Epi- and endocardial contour detection for the novel software was rated to be significantly better (p < 0.001) than with the former software. ICCs demonstrated strong agreement (≥ 0.75) for myocardial attenuation in 93% and for TPR in 82%. Diagnostic accuracy for the two software versions was not significantly different (p = 0.169) as compared with conventional coronary angiography. Conclusion The novel automated CTP analysis software offers enhanced time efficiency with an improvement by a factor of about four, while maintaining diagnostic accuracy. PMID:23323027

  20. [Design of the 2D-FFT image reconstruction software based on Matlab].

    PubMed

    Xu, Hong-yu; Wang, Hong-zhi

    2008-09-01

    This paper presents a Matlab's implementation for 2D-FFT image reconstruction algorithm of magnetic resonance imaging, with the universal COM component that Windows system can identify. This allows to segregate the 2D-FFT image reconstruction algorithm from the business magnetic resonance imaging closed system, providing the ability for initial data processing before reconstruction, which would be important for improving the image quality, diagnostic value and image post-processing.

  1. Comparison of grey scale median (GSM) measurement in ultrasound images of human carotid plaques using two different softwares.

    PubMed

    Östling, Gerd; Persson, Margaretha; Hedblad, Bo; Gonçalves, Isabel

    2013-11-01

    Grey scale median (GSM) measured on ultrasound images of carotid plaques has been used for several years now in research to find the vulnerable plaque. Centres have used different software and also different methods for GSM measurement. This has resulted in a wide range of GSM values and cut-off values for the detection of the vulnerable plaque. The aim of this study was to compare the values obtained with two different softwares, using different standardization methods, for the measurement of GSM on ultrasound images of carotid human plaques. GSM was measured with Adobe Photoshop(®) and with Artery Measurement System (AMS) on duplex ultrasound images of 100 consecutive medium- to large-sized carotid plaques of the Beta-blocker Cholesterol-lowering Asymptomatic Plaque Study (BCAPS). The mean values of GSM were 35·2 ± 19·3 and 55·8 ± 22·5 for Adobe Photoshop(®) and AMS, respectively. Mean difference was 20·45 (95% CI: 19·17-21·73). Although the absolute values of GSM differed, the agreement between the two measurements was good, correlation coefficient 0·95. A chi-square test revealed a kappa value of 0·68 when studying quartiles of GSM. The intra-observer variability was 1·9% for AMS and 2·5% for Adobe Photoshop. The difference between softwares and standardization methods must be taken into consideration when comparing studies. To avoid these problems, researcher should come to a consensus regarding software and standardization method for GSM measurement on ultrasound images of plaque in the arteries.

  2. Analysis of nuclear organization with TANGO, software for high-throughput quantitative analysis of 3D fluorescence microscopy images.

    PubMed

    Ollion, Jean; Cochennec, Julien; Loll, François; Escudé, Christophe; Boudier, Thomas

    2015-01-01

    The cell nucleus is a highly organized cellular organelle that contains the genome. An important step to understand the relationships between genome positioning and genome functions is to extract quantitative data from three-dimensional (3D) fluorescence imaging. However, such approaches are limited by the requirement for processing and analyzing large sets of images. Here we present a practical approach using TANGO (Tools for Analysis of Nuclear Genome Organization), an image analysis tool dedicated to the study of nuclear architecture. TANGO is a generic tool able to process large sets of images, allowing quantitative study of nuclear organization. In this chapter a practical description of the software is drawn in order to give an overview of its different concepts and functionalities. This description is illustrated with a precise example that can be performed step-by-step on experimental data provided on the website http://biophysique.mnhn.fr/tango/HomePage.

  3. Development and Evaluation of an Open-Source Software Package “CGITA” for Quantifying Tumor Heterogeneity with Molecular Images

    PubMed Central

    Fang, Yu-Hua Dean; Lin, Chien-Yu; Shih, Meng-Jung; Wang, Hung-Ming; Ho, Tsung-Ying; Liao, Chun-Ta; Yen, Tzu-Chen

    2014-01-01

    Background. The quantification of tumor heterogeneity with molecular images, by analyzing the local or global variation in the spatial arrangements of pixel intensity with texture analysis, possesses a great clinical potential for treatment planning and prognosis. To address the lack of available software for computing the tumor heterogeneity on the public domain, we develop a software package, namely, Chang-Gung Image Texture Analysis (CGITA) toolbox, and provide it to the research community as a free, open-source project. Methods. With a user-friendly graphical interface, CGITA provides users with an easy way to compute more than seventy heterogeneity indices. To test and demonstrate the usefulness of CGITA, we used a small cohort of eighteen locally advanced oral cavity (ORC) cancer patients treated with definitive radiotherapies. Results. In our case study of ORC data, we found that more than ten of the current implemented heterogeneity indices outperformed SUVmean for outcome prediction in the ROC analysis with a higher area under curve (AUC). Heterogeneity indices provide a better area under the curve up to 0.9 than the SUVmean and TLG (0.6 and 0.52, resp.). Conclusions. CGITA is a free and open-source software package to quantify tumor heterogeneity from molecular images. CGITA is available for free for academic use at http://code.google.com/p/cgita. PMID:24757667

  4. Biological Visualization, Imaging and Simulation(Bio-VIS) at NASA Ames Research Center: Developing New Software and Technology for Astronaut Training and Biology Research in Space

    NASA Technical Reports Server (NTRS)

    Smith, Jeffrey

    2003-01-01

    The Bio- Visualization, Imaging and Simulation (BioVIS) Technology Center at NASA's Ames Research Center is dedicated to developing and applying advanced visualization, computation and simulation technologies to support NASA Space Life Sciences research and the objectives of the Fundamental Biology Program. Research ranges from high resolution 3D cell imaging and structure analysis, virtual environment simulation of fine sensory-motor tasks, computational neuroscience and biophysics to biomedical/clinical applications. Computer simulation research focuses on the development of advanced computational tools for astronaut training and education. Virtual Reality (VR) and Virtual Environment (VE) simulation systems have become important training tools in many fields from flight simulation to, more recently, surgical simulation. The type and quality of training provided by these computer-based tools ranges widely, but the value of real-time VE computer simulation as a method of preparing individuals for real-world tasks is well established. Astronauts routinely use VE systems for various training tasks, including Space Shuttle landings, robot arm manipulations and extravehicular activities (space walks). Currently, there are no VE systems to train astronauts for basic and applied research experiments which are an important part of many missions. The Virtual Glovebox (VGX) is a prototype VE system for real-time physically-based simulation of the Life Sciences Glovebox where astronauts will perform many complex tasks supporting research experiments aboard the International Space Station. The VGX consists of a physical display system utilizing duel LCD projectors and circular polarization to produce a desktop-sized 3D virtual workspace. Physically-based modeling tools (Arachi Inc.) provide real-time collision detection, rigid body dynamics, physical properties and force-based controls for objects. The human-computer interface consists of two magnetic tracking devices

  5. Interference-free ultrasound imaging during HIFU therapy, using software tools

    NASA Technical Reports Server (NTRS)

    Vaezy, Shahram (Inventor); Held, Robert (Inventor); Sikdar, Siddhartha (Inventor); Managuli, Ravi (Inventor); Zderic, Vesna (Inventor)

    2010-01-01

    Disclosed herein is a method for obtaining a composite interference-free ultrasound image when non-imaging ultrasound waves would otherwise interfere with ultrasound imaging. A conventional ultrasound imaging system is used to collect frames of ultrasound image data in the presence of non-imaging ultrasound waves, such as high-intensity focused ultrasound (HIFU). The frames are directed to a processor that analyzes the frames to identify portions of the frame that are interference-free. Interference-free portions of a plurality of different ultrasound image frames are combined to generate a single composite interference-free ultrasound image that is displayed to a user. In this approach, a frequency of the non-imaging ultrasound waves is offset relative to a frequency of the ultrasound imaging waves, such that the interference introduced by the non-imaging ultrasound waves appears in a different portion of the frames.

  6. SU-E-J-42: Customized Deformable Image Registration Using Open-Source Software SlicerRT

    SciTech Connect

    Gaitan, J Cifuentes; Chin, L; Pignol, J; Kirby, N; Pouliot, J; Lasso, A; Pinter, C; Fichtinger, G

    2014-06-01

    Purpose: SlicerRT is a flexible platform that allows the user to incorporate the necessary images registration and processing tools to improve clinical workflow. This work validates the accuracy and the versatility of the deformable image registration algorithm of the free open-source software SlicerRT using a deformable physical pelvic phantom versus available commercial image fusion algorithms. Methods: Optical camera images of nonradiopaque markers implanted in an anatomical pelvic phantom were used to measure the ground-truth deformation and evaluate the theoretical deformations for several DIR algorithms. To perform the registration, full and empty bladder computed tomography (CT) images of the phantom were obtained and used as fixed and moving images, respectively. The DIR module, found in SlicerRT, used a B-spline deformable image registration with multiple optimization parameters that allowed customization of the registration including a regularization term that controlled the amount of local voxel displacement. The virtual deformation field at the center of the phantom was obtained and compared to the experimental ground-truth values. The parameters of SlicerRT were then varied to improve spatial accuracy. To quantify image similarity, the mean absolute difference (MAD) parameter using Hounsfield units was calculated. In addition, the Dice coefficient of the contoured rectum was evaluated to validate the strength of the algorithm to transfer anatomical contours. Results: Overall, SlicerRT achieved one of the lowest MAD values across the algorithm spectrum, but slightly smaller mean spatial errors in comparison to MIM software (MIM). On the other hand, SlicerRT created higher mean spatial errors than Velocity Medical Solutions (VEL), although obtaining an improvement on the DICE to 0.91. The large spatial errors were attributed to the poor contrast in the prostate bladder interface of the phantom. Conclusion: Based phantom validation, SlicerRT is capable of

  7. Nuquantus: Machine learning software for the characterization and quantification of cell nuclei in complex immunofluorescent tissue images

    PubMed Central

    Gross, Polina; Honnorat, Nicolas; Varol, Erdem; Wallner, Markus; Trappanese, Danielle M.; Sharp, Thomas E.; Starosta, Timothy; Duran, Jason M.; Koller, Sarah; Davatzikos, Christos; Houser, Steven R.

    2016-01-01

    Determination of fundamental mechanisms of disease often hinges on histopathology visualization and quantitative image analysis. Currently, the analysis of multi-channel fluorescence tissue images is primarily achieved by manual measurements of tissue cellular content and sub-cellular compartments. Since the current manual methodology for image analysis is a tedious and subjective approach, there is clearly a need for an automated analytical technique to process large-scale image datasets. Here, we introduce Nuquantus (Nuclei quantification utility software) - a novel machine learning-based analytical method, which identifies, quantifies and classifies nuclei based on cells of interest in composite fluorescent tissue images, in which cell borders are not visible. Nuquantus is an adaptive framework that learns the morphological attributes of intact tissue in the presence of anatomical variability and pathological processes. Nuquantus allowed us to robustly perform quantitative image analysis on remodeling cardiac tissue after myocardial infarction. Nuquantus reliably classifies cardiomyocyte versus non-cardiomyocyte nuclei and detects cell proliferation, as well as cell death in different cell classes. Broadly, Nuquantus provides innovative computerized methodology to analyze complex tissue images that significantly facilitates image analysis and minimizes human bias. PMID:27005843

  8. Nuquantus: Machine learning software for the characterization and quantification of cell nuclei in complex immunofluorescent tissue images.

    PubMed

    Gross, Polina; Honnorat, Nicolas; Varol, Erdem; Wallner, Markus; Trappanese, Danielle M; Sharp, Thomas E; Starosta, Timothy; Duran, Jason M; Koller, Sarah; Davatzikos, Christos; Houser, Steven R

    2016-03-23

    Determination of fundamental mechanisms of disease often hinges on histopathology visualization and quantitative image analysis. Currently, the analysis of multi-channel fluorescence tissue images is primarily achieved by manual measurements of tissue cellular content and sub-cellular compartments. Since the current manual methodology for image analysis is a tedious and subjective approach, there is clearly a need for an automated analytical technique to process large-scale image datasets. Here, we introduce Nuquantus (Nuclei quantification utility software) - a novel machine learning-based analytical method, which identifies, quantifies and classifies nuclei based on cells of interest in composite fluorescent tissue images, in which cell borders are not visible. Nuquantus is an adaptive framework that learns the morphological attributes of intact tissue in the presence of anatomical variability and pathological processes. Nuquantus allowed us to robustly perform quantitative image analysis on remodeling cardiac tissue after myocardial infarction. Nuquantus reliably classifies cardiomyocyte versus non-cardiomyocyte nuclei and detects cell proliferation, as well as cell death in different cell classes. Broadly, Nuquantus provides innovative computerized methodology to analyze complex tissue images that significantly facilitates image analysis and minimizes human bias.

  9. Nuquantus: Machine learning software for the characterization and quantification of cell nuclei in complex immunofluorescent tissue images

    NASA Astrophysics Data System (ADS)

    Gross, Polina; Honnorat, Nicolas; Varol, Erdem; Wallner, Markus; Trappanese, Danielle M.; Sharp, Thomas E.; Starosta, Timothy; Duran, Jason M.; Koller, Sarah; Davatzikos, Christos; Houser, Steven R.

    2016-03-01

    Determination of fundamental mechanisms of disease often hinges on histopathology visualization and quantitative image analysis. Currently, the analysis of multi-channel fluorescence tissue images is primarily achieved by manual measurements of tissue cellular content and sub-cellular compartments. Since the current manual methodology for image analysis is a tedious and subjective approach, there is clearly a need for an automated analytical technique to process large-scale image datasets. Here, we introduce Nuquantus (Nuclei quantification utility software) - a novel machine learning-based analytical method, which identifies, quantifies and classifies nuclei based on cells of interest in composite fluorescent tissue images, in which cell borders are not visible. Nuquantus is an adaptive framework that learns the morphological attributes of intact tissue in the presence of anatomical variability and pathological processes. Nuquantus allowed us to robustly perform quantitative image analysis on remodeling cardiac tissue after myocardial infarction. Nuquantus reliably classifies cardiomyocyte versus non-cardiomyocyte nuclei and detects cell proliferation, as well as cell death in different cell classes. Broadly, Nuquantus provides innovative computerized methodology to analyze complex tissue images that significantly facilitates image analysis and minimizes human bias.

  10. Software workflow for the automatic tagging of medieval manuscript images (SWATI)

    NASA Astrophysics Data System (ADS)

    Chandna, Swati; Tonne, Danah; Jejkal, Thomas; Stotzka, Rainer; Krause, Celia; Vanscheidt, Philipp; Busch, Hannah; Prabhune, Ajinkya

    2015-01-01

    Digital methods, tools and algorithms are gaining in importance for the analysis of digitized manuscript collections in the arts and humanities. One example is the BMBF-funded research project "eCodicology" which aims to design, evaluate and optimize algorithms for the automatic identification of macro- and micro-structural layout features of medieval manuscripts. The main goal of this research project is to provide better insights into high-dimensional datasets of medieval manuscripts for humanities scholars. The heterogeneous nature and size of the humanities data and the need to create a database of automatically extracted reproducible features for better statistical and visual analysis are the main challenges in designing a workflow for the arts and humanities. This paper presents a concept of a workflow for the automatic tagging of medieval manuscripts. As a starting point, the workflow uses medieval manuscripts digitized within the scope of the project Virtual Scriptorium St. Matthias". Firstly, these digitized manuscripts are ingested into a data repository. Secondly, specific algorithms are adapted or designed for the identification of macro- and micro-structural layout elements like page size, writing space, number of lines etc. And lastly, a statistical analysis and scientific evaluation of the manuscripts groups are performed. The workflow is designed generically to process large amounts of data automatically with any desired algorithm for feature extraction. As a result, a database of objectified and reproducible features is created which helps to analyze and visualize hidden relationships of around 170,000 pages. The workflow shows the potential of automatic image analysis by enabling the processing of a single page in less than a minute. Furthermore, the accuracy tests of the workflow on a small set of manuscripts with respect to features like page size and text areas show that automatic and manual analysis are comparable. The usage of a computer

  11. An adaptive workspace hypothesis about the neural correlates of consciousness: insights from neuroscience and meditation studies.

    PubMed

    Raffone, Antonino; Srinivasan, Narayanan

    2009-01-01

    While enormous progress has been made to identify neural correlates of consciousness (NCC), crucial NCC aspects are still very controversial. A major hurdle is the lack of an adequate definition and characterization of different aspects of conscious experience and also its relationship to attention and metacognitive processes like monitoring. In this paper, we therefore attempt to develop a unitary theoretical framework for NCC, with an interdependent characterization of endogenous attention, access consciousness, phenomenal awareness, metacognitive consciousness, and a non-referential form of unified consciousness. We advance an adaptive workspace hypothesis about the NCC based on the global workspace model emphasizing transient resonant neurodynamics and prefrontal cortex function, as well as meditation-related characterizations of conscious experiences. In this hypothesis, transient dynamic links within an adaptive coding net in prefrontal cortex, especially in anterior prefrontal cortex, and between it and the rest of the brain, in terms of ongoing intrinsic and long-range signal exchanges, flexibly regulate the interplay between endogenous attention, access consciousness, phenomenal awareness, and metacognitive consciousness processes. Such processes are established in terms of complementary aspects of an ongoing transition between context-sensitive global workspace assemblies, modulated moment-to-moment by body and environment states. Brain regions associated to momentary interoceptive and exteroceptive self-awareness, or first-person experiential perspective as emphasized in open monitoring meditation, play an important modulatory role in adaptive workspace transitions.

  12. Investigating Uses and Perceptions of an Online Collaborative Workspace for the Dissertation Process

    ERIC Educational Resources Information Center

    Rockinson-Szapkiw, Amanda J.

    2012-01-01

    The intent of this study was to investigate 93 doctoral candidates' perceptions and use of an online collaboration workspace and content management server, Microsoft Office SharePoint, for dissertation process. All candidates were enrolled in an Ed.D. programme in the United States. Descriptive statistics demonstrate that candidates frequently use…

  13. Reachable Workspace and Performance of Upper Limb (PUL) in Duchenne muscular dystrophy

    PubMed Central

    Han, Jay J.; de Bie, Evan; Nicorici, Alina; Abresch, Richard T.; Anthonisen, Colleen; Bajcsy, Ruzena; Kurillo, Gregorij; McDonald, Craig M.

    2015-01-01

    Introduction The Kinect-based reachable workspace relative surface area (RSA) is compared with the Performance of Upper Limb (PUL) assessment in Duchenne muscular dystrophy (DMD). Methods 29 individuals with DMD (ages: 7–23, Brooke: 1–5) underwent both Kinect-based reachable workspace RSA and PUL assessments. RSAs were also collected from 24 age-matched controls. Total and quadrant RSAs were compared with the PUL total, shoulder-, middle-, and distal-dimension scores. Results The total reachable workspace RSA correlated well with the total PUL score (Spearman ρ=−0.602, P<0.001), and with each of the PUL dimensional scores: shoulder (ρ=−0.624, P<0.001), middle (ρ=−0.564, P=0.001), and distal (ρ=−0.630, P<0.001). With quadrant RSA, reachability in a particular quadrant was closely associated with respective PUL dimensional-level function (lateral-upper quadrant for shoulder-, lateral-upper/lower quadrants for middle-, and lateral-lower quadrant for distal-level function). Discussion This study demonstrates concurrent validity of the reachable workspace outcome measure (RSA) with the DMD-specific upper extremity outcome measure (PUL). PMID:26342193

  14. Revolute manipulator workspace optimization using a modified bacteria foraging algorithm: A comparative study

    NASA Astrophysics Data System (ADS)

    Panda, S.; Mishra, D.; Biswal, B. B.; Tripathy, M.

    2014-02-01

    Robotic manipulators with three-revolute (3R) motions to attain desired positional configurations are very common in industrial robots. The capability of these robots depends largely on the workspace of the manipulator in addition to other parameters. In this study, an evolutionary optimization algorithm based on the foraging behaviour of the Escherichia coli bacteria present in the human intestine is utilized to optimize the workspace volume of a 3R manipulator. The new optimization method is modified from the original algorithm for faster convergence. This method is also useful for optimization problems in a highly constrained environment, such as robot workspace optimization. The new approach for workspace optimization of 3R manipulators is tested using three cases. The test results are compared with standard results available using other optimization algorithms, i.e. the differential evolution algorithm, the genetic algorithm and the particle swarm optimization algorithm. The present method is found to be superior to the other methods in terms of computational efficiency.

  15. Mapping and correction of the CMM workspace error with the use of an electronic gyroscope and neural networks--practical application.

    PubMed

    Swornowski, Pawel J

    2013-01-01

    The article presents the application of neural networks in determining and correction of the deformation of a coordinate measuring machine (CMM) workspace. The information about the CMM errors is acquired using an ADXRS401 electronic gyroscope. A test device (PS-20 module) was built and integrated with a commercial measurement system based on the SP25M passive scanning probe and with a PH10M module (Renishaw). The proposed solution was tested on a Kemco 600 CMM and on a DEA Global Clima CMM. In the former case, correction of the CMM errors was performed using the source code of WinIOS software owned by The Institute of Advanced Manufacturing Technology, Cracow, Poland and in the latter on an external PC. Optimum parameters of full and simplified mapping of a given layer of the CMM workspace were determined for practical applications. The proposed method can be employed for the interim check (ISO 10360-2 procedure) or to detect local CMM deformations, occurring when the CMM works at high scanning speeds (>20 mm/s).

  16. Automatic Detection of Mitosis and Nuclei From Cytogenetic Images by CellProfiler Software for Mitotic Index Estimation.

    PubMed

    González, Jorge Ernesto; Radl, Analía; Romero, Ivonne; Barquinero, Joan Francesc; García, Omar; Di Giorgio, Marina

    2016-12-01

    Mitotic Index (MI) estimation expressed as percentage of mitosis plays an important role as quality control endpoint. To this end, MI is applied to check the lot of media and reagents to be used throughout the assay and also to check cellular viability after blood sample shipping, indicating satisfactory/unsatisfactory conditions for the progression of cell culture. The objective of this paper was to apply the CellProfiler open-source software for automatic detection of mitotic and nuclei figures from digitized images of cultured human lymphocytes for MI assessment, and to compare its performance to that performed through semi-automatic and visual detection. Lymphocytes were irradiated and cultured for mitosis detection. Sets of images from cultures were analyzed visually and findings were compared with those using CellProfiler software. The CellProfiler pipeline includes the detection of nuclei and mitosis with 80% sensitivity and more than 99% specificity. We conclude that CellProfiler is a reliable tool for counting mitosis and nuclei from cytogenetic images, saves considerable time compared to manual operation and reduces the variability derived from the scoring criteria of different scorers. The CellProfiler automated pipeline achieves good agreement with visual counting workflow, i.e. it allows fully automated mitotic and nuclei scoring in cytogenetic images yielding reliable information with minimal user intervention.

  17. Phenotiki: an open software and hardware platform for affordable and easy image-based phenotyping of rosette-shaped plants.

    PubMed

    Minervini, Massimo; Giuffrida, Mario V; Perata, Pierdomenico; Tsaftaris, Sotirios A

    2017-04-01

    Phenotyping is important to understand plant biology, but current solutions are costly, not versatile or are difficult to deploy. To solve this problem, we present Phenotiki, an affordable system for plant phenotyping that, relying on off-the-shelf parts, provides an easy to install and maintain platform, offering an out-of-box experience for a well-established phenotyping need: imaging rosette-shaped plants. The accompanying software (with available source code) processes data originating from our device seamlessly and automatically. Our software relies on machine learning to devise robust algorithms, and includes an automated leaf count obtained from 2D images without the need of depth (3D). Our affordable device (~€200) can be deployed in growth chambers or greenhouse to acquire optical 2D images of approximately up to 60 adult Arabidopsis rosettes concurrently. Data from the device are processed remotely on a workstation or via a cloud application (based on CyVerse). In this paper, we present a proof-of-concept validation experiment on top-view images of 24 Arabidopsis plants in a combination of genotypes that has not been compared previously. Phenotypic analysis with respect to morphology, growth, color and leaf count has not been performed comprehensively before now. We confirm the findings of others on some of the extracted traits, showing that we can phenotype at reduced cost. We also perform extensive validations with external measurements and with higher fidelity equipment, and find no loss in statistical accuracy when we use the affordable setting that we propose. Device set-up instructions and analysis software are publicly available ( http://phenotiki.com).

  18. NeuroGam Software Analysis in Epilepsy Diagnosis Using 99mTc-ECD Brain Perfusion SPECT Imaging.

    PubMed

    Fu, Peng; Zhang, Fang; Gao, Jianqing; Jing, Jianmin; Pan, Liping; Li, Dongxue; Wei, Lingge

    2015-09-20

    BACKGROUND The aim of this study was to explore the value of NeuroGam software in diagnosis of epilepsy by 99Tcm-ethyl cysteinate dimer (ECD) brain imaging. MATERIAL AND METHODS NeuroGam was used to analyze 52 cases of clinically proven epilepsy by 99Tcm-ECD brain imaging. The results were compared with EEG and MRI, and the positive rates and localization to epileptic foci were analyzed. RESULTS NeuroGam analysis showed that 42 of 52 epilepsy cases were abnormal. 99Tcm-ECD brain imaging revealed a positive rate of 80.8% (42/52), with 36 out of 42 patients (85.7%) clearly showing an abnormal area. Both were higher than that of brain perfusion SPECT, with a consistency of 64.5% (34/52) using these 2 methods. Decreased regional cerebral blood flow (rCBF) was observed in frontal (18), temporal (20), and parietal lobes (2). Decreased rCBF was seen in frontal and temporal lobes in 4 out of 36 patients, and in temporal and parietal lobes of 2 out of 36 patients. NeuroGam further showed that the abnormal area was located in a different functional area of the brain. EEG abnormalities were detected in 29 out of 52 patients (55.8%) with 16 cases (55.2%) clearly showing an abnormal area. MRI abnormalities were detected in 17 out of 43 cases (39.5%), including 9 cases (52.9%) clearly showing an abnormal area. The consistency of NeuroGam software analysis, and EEG and MRI were 48.1% (25/52) and 34.9% (15/43), respectively. CONCLUSIONS NeuroGam software analysis offers a higher sensitivity in detecting epilepsy than EEG or MRI. It is a powerful tool in 99Tcm-ECD brain imaging.

  19. Filtering Chromatic Aberration for Wide Acceptance Angle Electrostatic Lenses II--Experimental Evaluation and Software-Based Imaging Energy Analyzer.

    PubMed

    Fazekas, Ádám; Daimon, Hiroshi; Matsuda, Hiroyuki; Tóth, László

    2016-03-01

    Here, the experimental results of the method of filtering the effect of chromatic aberration for wide acceptance angle electrostatic lens-based system are described. This method can eliminate the effect of chromatic aberration from the images of a measured spectral image sequence by determining and removing the effect of higher and lower kinetic energy electrons on each different energy image, which leads to significant improvement of image and spectral quality. The method is based on the numerical solution of a large system of linear equations and equivalent with a multivariate strongly nonlinear deconvolution method. A matrix whose elements describe the strongly nonlinear chromatic aberration-related transmission function of the lens system acts on the vector of the ordered pixels of the distortion free spectral image sequence, and produces the vector of the ordered pixels of the measured spectral image sequence. Since the method can be applied not only on 2D real- and $k$ -space diffraction images, but also along a third dimension of the image sequence that is along the optical or in the 3D parameter space, the energy axis, it functions as a software-based imaging energy analyzer (SBIEA). It can also be applied in cases of light or other type of optics for different optical aberrations and distortions. In case of electron optics, the SBIEA method makes possible the spectral imaging without the application of any other energy filter. It is notable that this method also eliminates the disturbing background significantly in the present investigated case of reflection electron energy loss spectra. It eliminates the instrumental effects and makes possible to measure the real physical processes better.

  20. SU-E-J-104: Evaluation of Accuracy for Various Deformable Image Registrations with Virtual Deformation QA Software

    SciTech Connect

    Han, S; Kim, K; Kim, M; Jung, H; Ji, Y; Choi, S; Park, S

    2015-06-15

    Purpose: The accuracy of deformable image registration (DIR) has a significant dosimetric impact in radiation treatment planning. We evaluated accuracy of various DIR algorithms using virtual deformation QA software (ImSimQA, Oncology System Limited, UK). Methods: The reference image (Iref) and volume (Vref) was first generated with IMSIMQA software. We deformed Iref with axial movement of deformation point and Vref depending on the type of deformation that are the deformation1 is to increase the Vref (relaxation) and the deformation 2 is to decrease the Vref (contraction) .The deformed image (Idef) and volume (Vdef) were inversely deformed to Iref and Vref using DIR algorithms. As a Result, we acquired deformed image (Iid) and volume (Vid). The DIR algorithms were optical flow (HS, IOF) and demons (MD, FD) of the DIRART. The image similarity evaluation between Iref and Iid was calculated by Normalized Mutual Information (NMI) and Normalized Cross Correlation (NCC). The value of Dice Similarity Coefficient (DSC) was used for evaluation of volume similarity. Results: When moving distance of deformation point was 4 mm, the value of NMI was above 1.81 and NCC was above 0.99 in all DIR algorithms. Since the degree of deformation was increased, the degree of image similarity was decreased. When the Vref increased or decreased about 12%, the difference between Vref and Vid was within ±5% regardless of the type of deformation. The value of DSC was above 0.95 in deformation1 except for the MD algorithm. In case of deformation 2, that of DSC was above 0.95 in all DIR algorithms. Conclusion: The Idef and Vdef have not been completely restored to Iref and Vref and the accuracy of DIR algorithms was different depending on the degree of deformation. Hence, the performance of DIR algorithms should be verified for the desired applications.

  1. Mission planning for Shuttle Imaging Radar-C (SIR-C) with a real-time interactive planning software

    NASA Technical Reports Server (NTRS)

    Potts, Su K.

    1993-01-01

    The Shuttle Imaging Radar-C (SIR-C) mission will operate from the payload bay of the space shuttle for 8 days, gathering Synthetic Aperture Radar (SAR) data over specific sites on the Earth. The short duration of the mission and the requirement for realtime planning offer challenges in mission planning and in the design of the Planning and Analysis Subsystem (PAS). The PAS generates shuttle ephemerides and mission planning data and provides an interactive real-time tool for quick mission replanning. It offers a multi-user and multiprocessing environment, and it is able to keep multiple versions of the mission timeline data while maintaining data integrity and security. Its flexible design allows one software to provide different menu options based on the user's operational function, and makes it easy to tailor the software for other Earth orbiting missions.

  2. SpotMetrics: An Open-Source Image-Analysis Software Plugin for Automatic Chromatophore Detection and Measurement

    PubMed Central

    Hadjisolomou, Stavros P.; El-Haddad, George

    2017-01-01

    Coleoid cephalopods (squid, octopus, and sepia) are renowned for their elaborate body patterning capabilities, which are employed for camouflage or communication. The specific chromatic appearance of a cephalopod, at any given moment, is a direct result of the combined action of their intradermal pigmented chromatophore organs and reflecting cells. Therefore, a lot can be learned about the cephalopod coloration system by video recording and analyzing the activation of individual chromatophores in time. The fact that adult cephalopods have small chromatophores, up to several hundred thousand in number, makes measurement and analysis over several seconds a difficult task. However, current advancements in videography enable high-resolution and high framerate recording, which can be used to record chromatophore activity in more detail and accuracy in both space and time domains. In turn, the additional pixel information and extra frames per video from such recordings result in large video files of several gigabytes, even when the recording spans only few minutes. We created a software plugin, “SpotMetrics,” that can automatically analyze high resolution, high framerate video of chromatophore organ activation in time. This image analysis software can track hundreds of individual chromatophores over several hundred frames to provide measurements of size and color. This software may also be used to measure differences in chromatophore activation during different behaviors which will contribute to our understanding of the cephalopod sensorimotor integration system. In addition, this software can potentially be utilized to detect numbers of round objects and size changes in time, such as eye pupil size or number of bacteria in a sample. Thus, we are making this software plugin freely available as open-source because we believe it will be of benefit to other colleagues both in the cephalopod biology field and also within other disciplines. PMID:28298896

  3. SpotMetrics: An Open-Source Image-Analysis Software Plugin for Automatic Chromatophore Detection and Measurement.

    PubMed

    Hadjisolomou, Stavros P; El-Haddad, George

    2017-01-01

    Coleoid cephalopods (squid, octopus, and sepia) are renowned for their elaborate body patterning capabilities, which are employed for camouflage or communication. The specific chromatic appearance of a cephalopod, at any given moment, is a direct result of the combined action of their intradermal pigmented chromatophore organs and reflecting cells. Therefore, a lot can be learned about the cephalopod coloration system by video recording and analyzing the activation of individual chromatophores in time. The fact that adult cephalopods have small chromatophores, up to several hundred thousand in number, makes measurement and analysis over several seconds a difficult task. However, current advancements in videography enable high-resolution and high framerate recording, which can be used to record chromatophore activity in more detail and accuracy in both space and time domains. In turn, the additional pixel information and extra frames per video from such recordings result in large video files of several gigabytes, even when the recording spans only few minutes. We created a software plugin, "SpotMetrics," that can automatically analyze high resolution, high framerate video of chromatophore organ activation in time. This image analysis software can track hundreds of individual chromatophores over several hundred frames to provide measurements of size and color. This software may also be used to measure differences in chromatophore activation during different behaviors which will contribute to our understanding of the cephalopod sensorimotor integration system. In addition, this software can potentially be utilized to detect numbers of round objects and size changes in time, such as eye pupil size or number of bacteria in a sample. Thus, we are making this software plugin freely available as open-source because we believe it will be of benefit to other colleagues both in the cephalopod biology field and also within other disciplines.

  4. Determining of a robot workspace using the integration of a CAD system with a virtual control system

    NASA Astrophysics Data System (ADS)

    Herbuś, K.; Ociepka, P.

    2016-08-01

    The paper presents a method for determining the workspace of an industrial robot using an approach consisting in integration a 3D model of an industrial robot with a virtual control system. The robot model with his work environment, prepared for motion simulation, was created in the “Motion Simulation” module of the Siemens PLM NX software. In the mentioned model components of the “link” type were created which map the geometrical form of particular elements of the robot and the components of “joint” type mapping way of cooperation of components of the “link” type. In the paper is proposed the solution in which the control process of a virtual robot is similar to the control process of a real robot using the manual control panel (teach pendant). For this purpose, the control application “JOINT” was created, which provides the manipulation of a virtual robot in accordance with its internal control system. The set of procedures stored in an .xlsx file is the element integrating the 3D robot model working in the CAD/CAE class system with the elaborated control application.

  5. ImaSim, a software tool for basic education of medical x-ray imaging in radiotherapy and radiology

    NASA Astrophysics Data System (ADS)

    Landry, Guillaume; deBlois, François; Verhaegen, Frank

    2013-11-01

    Introduction: X-ray imaging is an important part of medicine and plays a crucial role in radiotherapy. Education in this field is mostly limited to textbook teaching due to equipment restrictions. A novel simulation tool, ImaSim, for teaching the fundamentals of the x-ray imaging process based on ray-tracing is presented in this work. ImaSim is used interactively via a graphical user interface (GUI). Materials and methods: The software package covers the main x-ray based medical modalities: planar kilo voltage (kV), planar (portal) mega voltage (MV), fan beam computed tomography (CT) and cone beam CT (CBCT) imaging. The user can modify the photon source, object to be imaged and imaging setup with three-dimensional editors. Objects are currently obtained by combining blocks with variable shapes. The imaging of three-dimensional voxelized geometries is currently not implemented, but can be added in a later release. The program follows a ray-tracing approach, ignoring photon scatter in its current implementation. Simulations of a phantom CT scan were generated in ImaSim and were compared to measured data in terms of CT number accuracy. Spatial variations in the photon fluence and mean energy from an x-ray tube caused by the heel effect were estimated from ImaSim and Monte Carlo simulations and compared. Results: In this paper we describe ImaSim and provide two examples of its capabilities. CT numbers were found to agree within 36 Hounsfield Units (HU) for bone, which corresponds to a 2% attenuation coefficient difference. ImaSim reproduced the heel effect reasonably well when compared to Monte Carlo simulations. Discussion: An x-ray imaging simulation tool is made available for teaching and research purposes. ImaSim provides a means to facilitate the teaching of medical x-ray imaging.

  6. Panoramic Images Mapping Tools Integrated Within the ESRI ArcGIS Software

    NASA Astrophysics Data System (ADS)

    Guo, Jiao; Zhong, Ruofei; Zeng, Fanyang

    2014-03-01

    There is a general study on panoramic images which are presented along with appearance of the Google street map. Despite 360 degree viewing of street, we can realize more applications over panoramic images. This paper developed a toolkits plugged in ArcGIS, which can view panoramic photographs at street level directly from ArcMap and measure and capture all visible elements as frontages, trees and bridges. We use a series of panoramic images adjoined with absolute coordinate through GPS and IMU. There are two methods in this paper to measure object from these panoramic images: one is to intersect object position through a stereogram; the other one is multichip matching involved more than three images which all cover the object. While someone wants to measure objects from these panoramic images, each two panoramic images which both contain the object can be chosen to display on ArcMap. Then we calculate correlation coefficient of the two chosen panoramic images so as to calculate the coordinate of object. Our study test different patterns of panoramic pairs and compare the results of measurement to the real value of objects so as to offer the best choosing suggestion. The article has mainly elaborated the principles of calculating correlation coefficient and multichip matching.

  7. Software-based turbulence mitigation of short exposure image data with motion detection and background segmentation

    NASA Astrophysics Data System (ADS)

    Huebner, Claudia S.

    2011-11-01

    The degree of image degradation due to atmospheric turbulence is particularly severe when imaging over long horizontal paths since the turbulence is strongest close to the ground. The most pronounced effects include image blurring and image dancing and in case of strong turbulence image distortion as well. To mitigate these effects a number of methods from the field of image processing have been proposed most of which aim exclusively at the restoration of static scenes. But there is also an increasing interest in advancing turbulence mitigation to encompass moving objects as well. Therefore, in this paper a procedure is described that employs block-matching for the segmentation of static scene elements and moving objects such that image restoration can be carried out for both separately. This way motion blurring is taken into account in addition to atmospheric blurring, effectively reducing motion artefacts and improving the overall restoration result. Motion-compensated averaging with subsequent blind deconvolution is used for the actual image restoration.

  8. Simultaneous iterative reconstruction technique software for spectral-spatial EPR imaging

    NASA Astrophysics Data System (ADS)

    Spitzbarth, Martin; Drescher, Malte

    2015-08-01

    Continuous wave electron paramagnetic resonance imaging (EPRI) experiments often suffer from low signal to noise ratio. The increase in spectrometer time required to acquire data of sufficient quality to allow further analysis can be counteracted in part by more processing effort during the image reconstruction step. We suggest a simultaneous iterative reconstruction algorithm (SIRT) for reconstruction of continuous wave EPRI experimental data as an alternative to the widely applied filtered back projection algorithm (FBP). We show experimental and numerical test data of 2d spatial images and spectral-spatial images. We find that for low signal to noise ratio and spectral-spatial images that are limited by the maximum magnetic field gradient strength SIRT is more suitable than FBP.

  9. User's Guide for MapIMG 2: Map Image Re-projection Software Package

    USGS Publications Warehouse

    Finn, Michael P.; Trent, Jason R.; Buehler, Robert A.

    2006-01-01

    BACKGROUND Scientists routinely accomplish small-scale geospatial modeling in the raster domain, using high-resolution datasets for large parts of continents and low-resolution to high-resolution datasets for the entire globe. Direct implementation of point-to-point transformation with appropriate functions yields the variety of projections available in commercial software packages, but implementation with data other than points requires specific adaptation of the transformation equations or prior preparation of the data to allow the transformation to succeed. It seems that some of these packages use the U.S. Geological Survey's (USGS) General Cartographic Transformation Package (GCTP) or similar point transformations without adaptation to the specific characteristics of raster data (Usery and others, 2003a). Usery and others (2003b) compiled and tabulated the accuracy of categorical areas in projected raster datasets of global extent. Based on the shortcomings identified in these studies, geographers and applications programmers at the USGS expanded and evolved a USGS software package, MapIMG, for raster map projection transformation (Finn and Trent, 2004). Daniel R. Steinwand of Science Applications International Corporation, National Center for Earth Resources Observation and Science, originally developed MapIMG for the USGS, basing it on GCTP. Through previous and continuing efforts at the USGS' National Geospatial Technical Operations Center, this program has been transformed from an application based on command line input into a software package based on a graphical user interface for Windows, Linux, and other UNIX machines.

  10. Multiparametric Cell Cycle Analysis Using the Operetta High-Content Imager and Harmony Software with PhenoLOGIC.

    PubMed

    Massey, Andrew J

    2015-01-01

    High-content imaging is a powerful tool for determining cell phenotypes at the single cell level. Characterising the effect of small molecules on cell cycle distribution is important for understanding their mechanism of action especially in oncology drug discovery but also for understanding potential toxicology liabilities. Here, a high-throughput phenotypic assay utilising the PerkinElmer Operetta high-content imager and Harmony software to determine cell cycle distribution is described. PhenoLOGIC, a machine learning algorithm within Harmony software was employed to robustly separate single cells from cell clumps. DNA content, EdU incorporation and pHH3 (S10) expression levels were subsequently utilised to separate cells into the various phases of the cell cycle. The assay is amenable to multiplexing with an additional pharmacodynamic marker to assess cell cycle changes within a specific cellular sub-population. Using this approach, the cell cycle distribution of γH2AX positive nuclei was determined following treatment with DNA damaging agents. Likewise, the assay can be multiplexed with Ki67 to determine the fraction of quiescent cells and with BrdU dual labelling to determine S-phase duration. This methodology therefore provides a relatively cheap, quick and high-throughput phenotypic method for determining accurate cell cycle distribution for small molecule mechanism of action and drug toxicity studies.

  11. Novel mass spectrometry imaging software assisting labeled normalization and quantitation of drugs and neuropeptides directly in tissue sections.

    PubMed

    Källback, Patrik; Shariatgorji, Mohammadreza; Nilsson, Anna; Andrén, Per E

    2012-08-30

    MALDI MS imaging has been extensively used to produce qualitative distribution maps of proteins, peptides, lipids, small molecule pharmaceuticals and their metabolites directly in biological tissue sections. There is growing demand to quantify the amount of target compounds in the tissue sections of different organs. We present a novel MS imaging software including protocol for the quantitation of drugs, and for the first time, an endogenous neuropeptide directly in tissue sections. After selecting regions of interest on the tissue section, data is read and processed by the software using several available methods for baseline corrections, subtractions, denoising, smoothing, recalibration and normalization. The concentrations of in vivo administered drugs or endogenous compounds are then determined semi-automatically using either external standard curves, or by using labeled compounds, i.e., isotope labeled analogs as standards. As model systems, we have quantified the distribution of imipramine and tiotropium in the brain and lung of dosed rats. Substance P was quantified in different mouse brain structures, which correlated well with previously reported peptide levels. Our approach facilitates quantitative data processing and labeled standards provide better reproducibility and may be considered as an efficient tool to quantify drugs and endogenous compounds in tissue regions of interest.

  12. Online Workspace to Connect Scientists with NASA's Science E/PO Efforts and Practitioners

    NASA Astrophysics Data System (ADS)

    Shipp, Stephanie; Bartolone , Lindsay; Peticolas, Laura; Woroner, Morgan; Dalton, Heather; Schwerin, Theresa; Smith, Denise

    2014-11-01

    There is a growing awareness of the need for a scientifically literate public in light of challenges facing society today, and also a growing concern about the preparedness of our future workforce to meet those challenges. Federal priorities for science, technology, engineering, and math (STEM) education call for improvement of teacher training, increased youth and public engagement, greater involvement of underrepresented populations, and investment in undergraduate and graduate education. How can planetary scientists contribute to these priorities? How can they “make their work and findings comprehensible, appealing, and available to the public” as called for in the Planetary Decadal Survey?NASA’s Science Mission Directorate (SMD) Education and Public Outreach (E/PO) workspace provides the SMD E/PO community of practice - scientists and educators funded to conduct SMD E/PO or those using NASA’s science discoveries in E/PO endeavors - with an online environment in which to communicate, collaborate, and coordinate activities, thus helping to increase effectiveness of E/PO efforts. The workspace offers interested scientists avenues to partner with SMD E/PO practitioners and learn about E/PO projects and impacts, as well as to advertise their own efforts to reach a broader audience. Through the workspace, scientists can become aware of opportunities for involvement and explore resources to improve professional practice, including literature reviews of best practices for program impact, mechanisms for engaging diverse audiences, and large- and small-scale program evaluation. Scientists will find “how to” manuals for getting started and increasing impact with public presentations, classroom visits, and other audiences, as well as primers with activity ideas and resources that can augment E/PO interactions with different audiences. The poster will introduce the workspace to interested scientists and highlight pathways to resources of interest that can help

  13. The Human-Information Workspace (HI-Space): Ambient Table Top Entertainment

    SciTech Connect

    Cowell, Andrew J.; May, Richard A.; Cramer, Nick O.; Matthias Rauterberg

    2004-09-01

    This paper introduces the Human Information Workspace (HISpace) as a test-bed for evaluating new information exploration mechanisms. In moving from dated interaction devices and small computer monitors, we aim to utilize more natural surfaces such as tables and walls as our interaction space. In testing our theories, we have produced a number of gaming applications as test cases. Here, we report on our most popular application, Virtual Hockey.

  14. A Method to Determine an Organization’s Compatibility with Hybrid Workspaces

    DTIC Science & Technology

    2014-03-27

    individuals; instead, a variety of different work areas are designated to allow individuals to choose where they accomplish their work-related...set number of unassigned office spaces. Members choose their seating based on availability and preferred location. Studies of change management...towards hybrid workspaces. Interviews consisted of 22 questions, with 20 of these questions being open-ended and 2 designed to capture generational

  15. Measurement of ground displacement from optical satellite image correlation using the free open-source software MicMac

    NASA Astrophysics Data System (ADS)

    Rosu, Ana-Maria; Pierrot-Deseilligny, Marc; Delorme, Arthur; Binet, Renaud; Klinger, Yann

    2015-02-01

    Image correlation is one of the most efficient techniques to determine horizontal ground displacements due to earthquakes, landslides, ice flows or sand dune migrations. Analyzing these deformations allows a better understanding of the causes and mechanisms of the events. By using sub-pixel correlation on before- and after-event ortho-images obtained from high resolution satellite images it is possible to compute the displacement field with high planimetric resolution. In this paper, we focus on measuring the ground displacements due to seismotectonic events. The three sub-pixel correlators used are: COSI-Corr - developed by Caltech, a free, closed-source correlator, dependent on commercial software (ENVI) and widely used by the geoscience community for measuring ground displacement; Medicis - developed by CNES, also a closed-source correlator capable of measuring this type of deformation; and MicMac - developed by IGN, the free open-source correlator we study and tune for measuring fine ground displacements. We measured horizontal ground deformation using these three correlators on SPOT images in three study cases: the 2001 Kokoxili earthquake, the 2005 dyke intrusion in the Afar depression and the 2008 Yutian earthquake.

  16. Mathematically gifted adolescents mobilize enhanced workspace configuration of theta cortical network during deductive reasoning.

    PubMed

    Zhang, L; Gan, J Q; Wang, H

    2015-03-19

    Previous studies have established the importance of the fronto-parietal brain network in the information processing of reasoning. At the level of cortical source analysis, this eletroencepalogram (EEG) study investigates the functional reorganization of the theta-band (4-8Hz) neurocognitive network of mathematically gifted adolescents during deductive reasoning. Depending on the dense increase of long-range phase synchronizations in the reasoning process, math-gifted adolescents show more significant adaptive reorganization and enhanced "workspace" configuration in the theta network as compared with average-ability control subjects. The salient areas are mainly located in the anterior cortical vertices of the fronto-parietal network. Further correlation analyses have shown that the enhanced workspace configuration with respect to the global topological metrics of the theta network in math-gifted subjects is correlated with the intensive frontal midline theta (fm theta) response that is related to strong neural effort for cognitive events. These results suggest that by investing more cognitive resources math-gifted adolescents temporally mobilize an enhanced task-related global neuronal workspace, which is manifested as a highly integrated fronto-parietal information processing network during the reasoning process.

  17. Longitudinal evaluation of upper extremity reachable workspace in ALS by Kinect sensor.

    PubMed

    de Bie, Evan; Oskarsson, Bjorn; Joyce, Nanette C; Nicorici, Alina; Kurillo, Gregorij; Han, Jay J

    2017-02-01

    Our objective was to evaluate longitudinal changes in Microsoft Kinect measured upper extremity reachable workspace relative surface area (RSA) versus the revised Amyotrophic Lateral Sclerosis Functional Rating Scale (ALSFRS-R), ALSFRS-R upper extremity sub-scale and Forced Vital Capacity (FVC) in a cohort of patients diagnosed with amyotrophic lateral sclerosis (ALS). Ten patients diagnosed with ALS (ages 52-76 years, ALSFRS-R: 8-41 at entry) were tested using single 3D depth sensor, Microsoft Kinect, to measure reachable workspace RSA across five visits spanning one year. Changes in RSA, ALSFRS-R, ALSFRS-R upper extremity sub-scale, and FVC were assessed using a linear mixed model. Results showed that upper lateral quadrant RSA declined significantly in one year by approximately 19% (p <0.01) while all other quadrants and total RSA did not change significantly in this time-period. Simultaneously, ALSFRS-R upper extremity sub-scale worsened significantly by 25% (p <0.01). In conclusion, upper extremity reachable workspace RSA as a novel ALS outcome measure is capable of objectively quantifying declines in upper extremity ability over time in patients with ALS with more granularity than other common outcome measures. RSA may serve as a clinical endpoint for the evaluation of upper extremity targeted therapeutics.

  18. Dual-energy CT with virtual monochromatic images and metal artifact reduction software for reducing metallic dental artifacts.

    PubMed

    Cha, Jihoon; Kim, Hyung-Jin; Kim, Sung Tae; Kim, Yi Kyung; Kim, Ha Youn; Park, Gyeong Min

    2017-01-01

    Background Metallic dental prostheses may degrade image quality on head and neck computed tomography (CT). However, there is little information available on the use of dual-energy CT (DECT) and metal artifact reduction software (MARS) in the head and neck regions to reduce metallic dental artifacts. Purpose To assess the usefulness of DECT with virtual monochromatic imaging and MARS to reduce metallic dental artifacts. Material and Methods DECT was performed using fast kilovoltage (kV)-switching between 80-kV and 140-kV in 20 patients with metallic dental prostheses. CT data were reconstructed with and without MARS, and with synthesized monochromatic energy in the range of 40-140-kiloelectron volt (keV). For quantitative analysis, the artifact index of the tongue, buccal, and parotid areas was calculated for each scan. For qualitative analysis, two radiologists evaluated 70-keV and 100-keV images with and without MARS for tongue, buccal, parotid areas, and metallic denture. The locations and characteristics of the MARS-related artifacts, if any, were also recorded. Results DECT with MARS markedly reduced metallic dental artifacts and improved image quality in the buccal area ( P < 0.001) and the tongue ( P < 0.001), but not in the parotid area. The margin and internal architecture of the metallic dentures were more clearly delineated with MARS ( P < 0.001) and in the higher-energy images than in the lower-energy images ( P = 0.042). MARS-related artifacts most commonly occurred in the deep center of the neck. Conclusion DECT with MARS can reduce metallic dental artifacts and improve delineation of the metallic prosthesis and periprosthetic region.

  19. Analysis of image sharpness reproducibility on a novel engineered micro-CT scanner with variable geometry and embedded recalibration software.

    PubMed

    Panetta, D; Belcari, N; Del Guerra, A; Bartolomei, A; Salvadori, P A

    2012-04-01

    This study investigates the reproducibility of the reconstructed image sharpness, after modifications of the geometry setup, for a variable magnification micro-CT (μCT) scanner. All the measurements were performed on a novel engineered μCT scanner for in vivo imaging of small animals (Xalt), which has been recently built at the Institute of Clinical Physiology of the National Research Council (IFC-CNR, Pisa, Italy), in partnership with the University of Pisa. The Xalt scanner is equipped with an integrated software for on-line geometric recalibration, which will be used throughout the experiments. In order to evaluate the losses of image quality due to modifications of the geometry setup, we have made 22 consecutive acquisitions by changing alternatively the system geometry between two different setups (Large FoV - LF, and High Resolution - HR). For each acquisition, the tomographic images have been reconstructed before and after the on-line geometric recalibration. For each reconstruction, the image sharpness was evaluated using two different figures of merit: (i) the percentage contrast on a small bar pattern of fixed frequency (f = 5.5 lp/mm for the LF setup and f = 10 lp/mm for the HR setup) and (ii) the image entropy. We have found that, due to the small-scale mechanical uncertainty (in the order of the voxel size), a recalibration is necessary for each geometric setup after repositioning of the system's components; the resolution losses due to the lack of recalibration are worse for the HR setup (voxel size = 18.4 μm). The integrated on-line recalibration algorithm of the Xalt scanner allowed to perform the recalibration quickly, by restoring the spatial resolution of the system to the reference resolution obtained after the initial (off-line) calibration.

  20. Familiarity effects in the construction of facial-composite images using modern software systems.

    PubMed

    Frowd, Charlie D; Skelton, Faye C; Butt, Neelam; Hassan, Amal; Fields, Stephen; Hancock, Peter J B

    2011-12-01

    We investigate the effect of target familiarity on the construction of facial composites, as used by law enforcement to locate criminal suspects. Two popular software construction methods were investigated. Participants were shown a target face that was either familiar or unfamiliar to them and constructed a composite of it from memory using a typical 'feature' system, involving selection of individual facial features, or one of the newer 'holistic' types, involving repeated selection and breeding from arrays of whole faces. This study found that composites constructed of a familiar face were named more successfully than composites of an unfamiliar face; also, naming of composites of internal and external features was equivalent for construction of unfamiliar targets, but internal features were better named than the external features for familiar targets. These findings applied to both systems, although benefit emerged for the holistic type due to more accurate construction of internal features and evidence for a whole-face advantage. STATEMENT OF RELEVANCE: This work is of relevance to practitioners who construct facial composites with witnesses to and victims of crime, as well as for software designers to help them improve the effectiveness of their composite systems.

  1. Seismic reflection imaging of underground cavities using open-source software

    SciTech Connect

    Mellors, R J

    2011-12-20

    The Comprehensive Nuclear Test Ban Treaty (CTBT) includes provisions for an on-site inspection (OSI), which allows the use of specific techniques to detect underground anomalies including cavities and rubble zones. One permitted technique is active seismic surveys such as seismic refraction or reflection. The purpose of this report is to conduct some simple modeling to evaluate the potential use of seismic reflection in detecting cavities and to test the use of open-source software in modeling possible scenarios. It should be noted that OSI inspections are conducted under specific constraints regarding duration and logistics. These constraints are likely to significantly impact active seismic surveying, as a seismic survey typically requires considerable equipment, effort, and expertise. For the purposes of this study, which is a first-order feasibility study, these issues will not be considered. This report provides a brief description of the seismic reflection method along with some commonly used software packages. This is followed by an outline of a simple processing stream based on a synthetic model, along with results from a set of models representing underground cavities. A set of scripts used to generate the models are presented in an appendix. We do not consider detection of underground facilities in this work and the geologic setting used in these tests is an extremely simple one.

  2. Motion correction of PET brain images through deconvolution: I. Theoretical development and analysis in software simulations

    NASA Astrophysics Data System (ADS)

    Faber, T. L.; Raghunath, N.; Tudorascu, D.; Votaw, J. R.

    2009-02-01

    Image quality is significantly degraded even by small amounts of patient motion in very high-resolution PET scanners. Existing correction methods that use known patient motion obtained from tracking devices either require multi-frame acquisitions, detailed knowledge of the scanner, or specialized reconstruction algorithms. A deconvolution algorithm has been developed that alleviates these drawbacks by using the reconstructed image to estimate the original non-blurred image using maximum likelihood estimation maximization (MLEM) techniques. A high-resolution digital phantom was created by shape-based interpolation of the digital Hoffman brain phantom. Three different sets of 20 movements were applied to the phantom. For each frame of the motion, sinograms with attenuation and three levels of noise were simulated and then reconstructed using filtered backprojection. The average of the 20 frames was considered the motion blurred image, which was restored with the deconvolution algorithm. After correction, contrast increased from a mean of 2.0, 1.8 and 1.4 in the motion blurred images, for the three increasing amounts of movement, to a mean of 2.5, 2.4 and 2.2. Mean error was reduced by an average of 55% with motion correction. In conclusion, deconvolution can be used for correction of motion blur when subject motion is known.

  3. Integration of instrumentation and processing software of a laser speckle contrast imaging system

    NASA Astrophysics Data System (ADS)

    Carrick, Jacob J.

    Laser speckle contrast imaging (LSCI) has the potential to be a powerful tool in medicine, but more research in the field is required so it can be used properly. To help in the progression of Michigan Tech's research in the field, a graphical user interface (GUI) was designed in Matlab to control the instrumentation of the experiments as well as process the raw speckle images into contrast images while they are being acquired. The design of the system was successful and is currently being used by Michigan Tech's Biomedical Engineering department. This thesis describes the development of the LSCI GUI as well as offering a full introduction into the history, theory and applications of LSCI.

  4. Novel collaboration and situational awareness environment for leaders and their support staff via self assembling software.

    SciTech Connect

    Bouchard, Ann Marie; Osbourn, Gordon Cecil; Bartholomew, John Warren

    2008-02-01

    This is the final report on the Sandia Fellow LDRD, project 117865, 08-0281. This presents an investigation of self-assembling software intended to create shared workspace environment to allow online collaboration and situational awareness for use by high level managers and their teams.

  5. Automated image mosaics by non-automated light microscopes: the MicroMos software tool.

    PubMed

    Piccinini, F; Bevilacqua, A; Lucarelli, E

    2013-12-01

    Light widefield microscopes and digital imaging are the basis for most of the analyses performed in every biological laboratory. In particular, the microscope's user is typically interested in acquiring high-detailed images for analysing observed cells and tissues, meanwhile being representative of a wide area to have reliable statistics. The microscopist has to choose between higher magnification factor and extension of the observed area, due to the finite size of the camera's field of view. To overcome the need of arrangement, mosaicing techniques have been developed in the past decades for increasing the camera's field of view by stitching together more images. Nevertheless, these approaches typically work in batch mode and rely on motorized microscopes. Or alternatively, the methods are conceived just to provide visually pleasant mosaics not suitable for quantitative analyses. This work presents a tool for building mosaics of images acquired with nonautomated light microscopes. The method proposed is based on visual information only and the mosaics are built by incrementally stitching couples of images, making the approach available also for online applications. Seams in the stitching regions as well as tonal inhomogeneities are corrected by compensating the vignetting effect. In the experiments performed, we tested different registration approaches, confirming that the translation model is not always the best, despite the fact that the motion of the sample holder of the microscope is apparently translational and typically considered as such. The method's implementation is freely distributed as an open source tool called MicroMos. Its usability makes building mosaics of microscope images at subpixel accuracy easier. Furthermore, optional parameters for building mosaics according to different strategies make MicroMos an easy and reliable tool to compare different registration approaches, warping models and tonal corrections.

  6. Toolkits and Software for Developing Biomedical Image Processing and Analysis Applications

    NASA Astrophysics Data System (ADS)

    Wolf, Ivo

    Solutions in biomedical image processing and analysis usually consist of much more than a single method. Typically, a whole pipeline of algorithms is necessary, combined with visualization components to display and verify the results as well as possibilities to interact with the data. Therefore, successful research in biomedical image processing and analysis requires a solid base to start from. This is the case regardless whether the goal is the development of a new method (e.g., for segmentation) or to solve a specific task (e.g., computer-assisted planning of surgery).

  7. Quantification of Abdominal Fat in Obese and Healthy Adolescents Using 3 Tesla Magnetic Resonance Imaging and Free Software for Image Analysis

    PubMed Central

    Eloi, Juliana Cristina; Epifanio, Matias; de Gonçalves, Marília Maia; Pellicioli, Augusto; Vieira, Patricia Froelich Giora; Dias, Henrique Bregolin; Bruscato, Neide; Soder, Ricardo Bernardi; Santana, João Carlos Batista; Mouzaki, Marialena; Baldisserotto, Matteo

    2017-01-01

    Background and Aims Computed tomography, which uses ionizing radiation and expensive software packages for analysis of scans, can be used to quantify abdominal fat. The objective of this study is to measure abdominal fat with 3T MRI using free software for image analysis and to correlate these findings with anthropometric and laboratory parameters in adolescents. Methods This prospective observational study included 24 overweight/obese and 33 healthy adolescents (mean age 16.55 years). All participants underwent abdominal MRI exams. Visceral and subcutaneous fat area and percentage were correlated with anthropometric parameters, lipid profile, glucose metabolism, and insulin resistance. Student’s t test and Mann-Whitney’s test was applied. Pearson’s chi-square test was used to compare proportions. To determine associations Pearson’s linear correlation or Spearman’s correlation were used. Results In both groups, waist circumference (WC) was associated with visceral fat area (P = 0.001 and P = 0.01 respectively), and triglycerides were associated with fat percentage (P = 0.046 and P = 0.071 respectively). In obese individuals, total cholesterol/HDL ratio was associated with visceral fat area (P = 0.03) and percentage (P = 0.09), and insulin and HOMA-IR were associated with visceral fat area (P = 0.001) and percentage (P = 0.005). Conclusions 3T MRI can provide reliable and good quality images for quantification of visceral and subcutaneous fat by using a free software package. The results demonstrate that WC is a good predictor of visceral fat in obese adolescents and visceral fat area is associated with total cholesterol/HDL ratio, insulin and HOMA-IR. PMID:28129354

  8. eSlide suite: an open source software system for whole slide imaging.

    PubMed

    Della Mea, V; Bortolotti, N; Beltrami, C A

    2009-08-01

    This short report briefly describes the principles underlying the telepathology technique known as whole slide imaging, and the design and implementation of a system for acquisition and visualisation of digital slides. The developed system, including an acquisition module and a visualisation module, is available as an open source on the Internet, together with sample acquired slides.

  9. 3D Imaging for hand gesture recognition: Exploring the software-hardware interaction of current technologies

    NASA Astrophysics Data System (ADS)

    Periverzov, Frol; Ilieş, Horea T.

    2012-09-01

    Interaction with 3D information is one of the fundamental and most familiar tasks in virtually all areas of engineering and science. Several recent technological advances pave the way for developing hand gesture recognition capabilities available to all, which will lead to more intuitive and efficient 3D user interfaces (3DUI). These developments can unlock new levels of expression and productivity in all activities concerned with the creation and manipulation of virtual 3D shapes and, specifically, in engineering design. Building fully automated systems for tracking and interpreting hand gestures requires robust and efficient 3D imaging techniques as well as potent shape classifiers. We survey and explore current and emerging 3D imaging technologies, and focus, in particular, on those that can be used to build interfaces between the users' hands and the machine. The purpose of this paper is to categorize and highlight the relevant differences between these existing 3D imaging approaches in terms of the nature of the information provided, output data format, as well as the specific conditions under which these approaches yield reliable data. Furthermore we explore the impact of each of these approaches on the computational cost and reliability of the required image processing algorithms. Finally we highlight the main challenges and opportunities in developing natural user interfaces based on hand gestures, and conclude with some promising directions for future research. [Figure not available: see fulltext.

  10. DullRazor: a software approach to hair removal from images.

    PubMed

    Lee, T; Ng, V; Gallagher, R; Coldman, A; McLean, D

    1997-11-01

    Recently, there has been a growing number of studies applying image processing techniques to analyze melanocytic lesions for atypia and possible malignancy and for total-body mole mapping. However, such lesions can be partially obscured by body hairs. None of these studies has fully addressed the problem of human hairs occluding the imaged lesions. In our previous study we designed an automatic segmentation program to differentiate skin lesions from the normal healthy skin, and learned that the program performed well with most of the images, the exception being those with hairs, especially dark thick hairs, covering part of the lesions. These thick dark hairs confused the program, resulting in unsatisfactory segmentation results. In this paper, we present a method to remove hairs from an image using a pre-processing program we have called DullRazor. This pre-processing step enables the segmentation program to achieve satisfactory results. DullRazor can be downloaded as shareware from http:/(/)www.derm.ubc.ca.

  11. Advances in software for faster procedure and lower radiotracer dose myocardial perfusion imaging.

    PubMed

    Piccinelli, Marina; Garcia, Ernest V

    2015-01-01

    The American Society of Nuclear Cardiology has recently published documents that encourage laboratories to take all the appropriate steps to greatly decrease patient radiation dose and has set the goal of 50% of all myocardial perfusion studies performed with an associated radiation exposure of 9mSv by 2014. In the present work, a description of the major software techniques readily available to shorten procedure time and decrease injected activity is presented. Particularly new reconstruction methods and their ability to include means for resolution recovery and noise regularization are described. The use of these improved reconstruction algorithms results in a consistent reduction in acquisition time, injected activity and consequently in the radiation dose absorbed by the patient. The clinical implications to the use of these techniques are also described in terms of maintained and even improved study quality, accuracy and sensitivity for the detection of heart disease.

  12. Full-sun synchronic EUV and coronal hole mapping using multi-instrument images: Data and software made available

    NASA Astrophysics Data System (ADS)

    Caplan, R. M.; Downs, C.; Linker, J.

    2015-12-01

    A method for the automatic generation of EUV and coronal hole (CH) maps using simultaneous multi-instrument imaging data is described. Synchronized EUV images from STEREO/EUVI A&B 195Å and SDO/AIA 193Å undergo preprocessing steps that include PSF-deconvolution and the application of nonlinear data-derived intensity corrections that account for center-to-limb variations (limb-brightening) and inter-instrument intensity normalization. The latter two corrections are derived using a robust, systematic approach that takes advantage of unbiased long-term averages of data and serve to flatten the images by converting all pixel intensities to a unified disk center equivalent. While the number of applications are broad, we demonstrate how this technique is very useful for CH detection as it enables the use of a fast and simplified image segmentation algorithm to obtain consistent detection results. The multi-instrument nature of the technique also allows one to track evolving features consistently for longer periods than is possible with a single instrument, and preliminary results quantifying CH area and shape evolution are shown.Most importantly, several data and software products are made available to the community for use. For the ~4 year period of 6/10/2010 to 8/18/2014, we provide synchronic EUV and coronal hole maps at 6-hour cadence as well as the data-derived limb brightening and inter-instrument correction factors that we applied. We also make available a ready-to-use MATLAB script EUV2CHM used to generate the maps, which loads EUV images, applies our preprocessing steps, and then uses our GPU-accelerated/CPU-multithreaded segmentation algorithm EZSEG to detect coronal holes.

  13. Grid-less imaging with antiscatter correction software in 2D mammography: the effects on image quality and MGD under a partial virtual clinical validation study

    NASA Astrophysics Data System (ADS)

    Van Peteghem, Nelis; Bemelmans, Frédéric; Bramaje Adversalo, Xenia; Salvagnini, Elena; Marshall, Nicholas; Bosmans, Hilde; Van Ongeval, Chantal

    2016-03-01

    This work investigated the effect of the grid-less acquisition mode with scatter correction software developed by Siemens Healthcare (PRIME mode) on image quality and mean glandular dose (MGD) in a comparative study against a standard mammography system with grid. Image quality was technically quantified with contrast-detail (c-d) analysis and by calculating detectability indices (d') using a non-prewhitening with eye filter model observer (NPWE). MGD was estimated technically using slabs of PMMA and clinically on a set of 11439 patient images. The c-d analysis gave similar results for all mammographic systems examined, although the d' values were slightly lower for the system with PRIME mode when compared to the same system in standard mode (-2.8% to -5.7%, depending on the PMMA thickness). The MGD values corresponding to the PMMA measurements with automatic exposure control indicated a dose reduction from 11.0% to 20.8% for the system with PRIME mode compared to the same system without PRIME mode. The largest dose reductions corresponded to the thinnest PMMA thicknesses. The results from the clinical dosimetry study showed an overall population-averaged dose reduction of 11.6% (up to 27.7% for thinner breasts) for PRIME mode compared to standard mode for breast thicknesses from 20 to 69 mm. These technical image quality measures were then supported using a clinically oriented study whereby simulated clusters of microcalcifications and masses were inserted into patient images and read by radiologists in an AFROC study to quantify their detectability. In line with the technical investigation, no significant difference was found between the two imaging modes (p-value 0.95).

  14. Development of fast patient position verification software using 2D-3D image registration and its clinical experience

    PubMed Central

    Mori, Shinichiro; Kumagai, Motoki; Miki, Kentaro; Fukuhara, Riki; Haneishi, Hideaki

    2015-01-01

    To improve treatment workflow, we developed a graphic processing unit (GPU)-based patient positional verification software application and integrated it into carbon-ion scanning beam treatment. Here, we evaluated the basic performance of the software. The algorithm provides 2D/3D registration matching using CT and orthogonal X-ray flat panel detector (FPD) images. The participants were 53 patients with tumors of the head and neck, prostate or lung receiving carbon-ion beam treatment. 2D/3D-ITchi-Gime (ITG) calculation accuracy was evaluated in terms of computation time and registration accuracy. Registration calculation was determined using the similarity measurement metrics gradient difference (GD), normalized mutual information (NMI), zero-mean normalized cross-correlation (ZNCC), and their combination. Registration accuracy was dependent on the particular metric used. Representative examples were determined to have target registration error (TRE) = 0.45 ± 0.23 mm and angular error (AE) = 0.35 ± 0.18° with ZNCC + GD for a head and neck tumor; TRE = 0.12 ± 0.07 mm and AE = 0.16 ± 0.07° with ZNCC for a pelvic tumor; and TRE = 1.19 ± 0.78 mm and AE = 0.83 ± 0.61° with ZNCC for lung tumor. Calculation time was less than 7.26 s.The new registration software has been successfully installed and implemented in our treatment process. We expect that it will improve both treatment workflow and treatment accuracy. PMID:26081313

  15. Development of fast patient position verification software using 2D-3D image registration and its clinical experience.

    PubMed

    Mori, Shinichiro; Kumagai, Motoki; Miki, Kentaro; Fukuhara, Riki; Haneishi, Hideaki

    2015-09-01

    To improve treatment workflow, we developed a graphic processing unit (GPU)-based patient positional verification software application and integrated it into carbon-ion scanning beam treatment. Here, we evaluated the basic performance of the software. The algorithm provides 2D/3D registration matching using CT and orthogonal X-ray flat panel detector (FPD) images. The participants were 53 patients with tumors of the head and neck, prostate or lung receiving carbon-ion beam treatment. 2D/3D-ITchi-Gime (ITG) calculation accuracy was evaluated in terms of computation time and registration accuracy. Registration calculation was determined using the similarity measurement metrics gradient difference (GD), normalized mutual information (NMI), zero-mean normalized cross-correlation (ZNCC), and their combination. Registration accuracy was dependent on the particular metric used. Representative examples were determined to have target registration error (TRE) = 0.45 ± 0.23 mm and angular error (AE) = 0.35 ± 0.18° with ZNCC + GD for a head and neck tumor; TRE = 0.12 ± 0.07 mm and AE = 0.16 ± 0.07° with ZNCC for a pelvic tumor; and TRE = 1.19 ± 0.78 mm and AE = 0.83 ± 0.61° with ZNCC for lung tumor. Calculation time was less than 7.26 s.The new registration software has been successfully installed and implemented in our treatment process. We expect that it will improve both treatment workflow and treatment accuracy.

  16. Photon counting imaging and centroiding with an electron-bombarded CCD using single molecule localisation software

    NASA Astrophysics Data System (ADS)

    Hirvonen, Liisa M.; Barber, Matthew J.; Suhling, Klaus

    2016-06-01

    Photon event centroiding in photon counting imaging and single-molecule localisation in super-resolution fluorescence microscopy share many traits. Although photon event centroiding has traditionally been performed with simple single-iteration algorithms, we recently reported that iterative fitting algorithms originally developed for single-molecule localisation fluorescence microscopy work very well when applied to centroiding photon events imaged with an MCP-intensified CMOS camera. Here, we have applied these algorithms for centroiding of photon events from an electron-bombarded CCD (EBCCD). We find that centroiding algorithms based on iterative fitting of the photon events yield excellent results and allow fitting of overlapping photon events, a feature not reported before and an important aspect to facilitate an increased count rate and shorter acquisition times.

  17. Photon counting imaging and centroiding with an electron-bombarded CCD using single molecule localisation software

    PubMed Central

    Hirvonen, Liisa M.; Barber, Matthew J.; Suhling, Klaus

    2016-01-01

    Photon event centroiding in photon counting imaging and single-molecule localisation in super-resolution fluorescence microscopy share many traits. Although photon event centroiding has traditionally been performed with simple single-iteration algorithms, we recently reported that iterative fitting algorithms originally developed for single-molecule localisation fluorescence microscopy work very well when applied to centroiding photon events imaged with an MCP-intensified CMOS camera. Here, we have applied these algorithms for centroiding of photon events from an electron-bombarded CCD (EBCCD). We find that centroiding algorithms based on iterative fitting of the photon events yield excellent results and allow fitting of overlapping photon events, a feature not reported before and an important aspect to facilitate an increased count rate and shorter acquisition times. PMID:27274604

  18. Color voltage contrast: A new method of implementing fault contrast with color imaging software

    SciTech Connect

    Colvin, J.

    1995-12-31

    Although voltage contrast and fault contrast methods are well established, the current methods of implementation are frequently tedious. A new method of mapping voltage contrast (VC) images in a qualitative (stroboscopic) color mode allows multiple logic states to be simultaneously viewed and updated in color. A shortcoming of image subtraction is that only one direction of logic change is represented unless the frames are exclusive OR`ed together. Although this gives fault information it does not include the VC of neighboring unchanged nodes. When tracking failures such as a saturated transistor resulting from a logic short somewhere else, all logic states; both static and transitional need to be understood and viewed simultaneously if an expedient analysis is desired.

  19. The Java Image Science Toolkit (JIST) for rapid prototyping and publishing of neuroimaging software.

    PubMed

    Lucas, Blake C; Bogovic, John A; Carass, Aaron; Bazin, Pierre-Louis; Prince, Jerry L; Pham, Dzung L; Landman, Bennett A

    2010-03-01

    Non-invasive neuroimaging techniques enable extraordinarily sensitive and specific in vivo study of the structure, functional response and connectivity of biological mechanisms. With these advanced methods comes a heavy reliance on computer-based processing, analysis and interpretation. While the neuroimaging community has produced many excellent academic and commercial tool packages, new tools are often required to interpret new modalities and paradigms. Developing custom tools and ensuring interoperability with existing tools is a significant hurdle. To address these limitations, we present a new framework for algorithm development that implicitly ensures tool interoperability, generates graphical user interfaces, provides advanced batch processing tools, and, most importantly, requires minimal additional programming or computational overhead. Java-based rapid prototyping with this system is an efficient and practical approach to evaluate new algorithms since the proposed system ensures that rapidly constructed prototypes are actually fully-functional processing modules with support for multiple GUI's, a broad range of file formats, and distributed computation. Herein, we demonstrate MRI image processing with the proposed system for cortical surface extraction in large cross-sectional cohorts, provide a system for fully automated diffusion tensor image analysis, and illustrate how the system can be used as a simulation framework for the development of a new image analysis method. The system is released as open source under the Lesser GNU Public License (LGPL) through the Neuroimaging Informatics Tools and Resources Clearinghouse (NITRC).

  20. Reliability and reproducibility of macular segmentation using a custom-built optical coherence tomography retinal image analysis software

    NASA Astrophysics Data System (ADS)

    Cabrera Debuc, Delia; Somfai, Gábor Márk; Ranganathan, Sudarshan; Tátrai, Erika; Ferencz, Mária; Puliafito, Carmen A.

    2009-11-01

    We determine the reliability and reproducibility of retinal thickness measurements with a custom-built OCT retinal image analysis software (OCTRIMA). Ten eyes of five healthy subjects undergo repeated standard macular thickness map scan sessions by two experienced examiners using a Stratus OCT device. Automatic/semi automatic thickness quantification of the macula and intraretinal layers is performed using OCTRIMA software. Intraobserver, interobserver, and intervisit repeatability and reproducibility coefficients, and intraclass correlation coefficients (ICCs) per scan are calculated. Intraobserver, interobserver, and intervisit variability combined account for less than 5% of total variability for the total retinal thickness measurements and less than 7% for the intraretinal layers except the outer segment/ retinal pigment epithelium (RPE) junction. There is no significant difference between scans acquired by different observers or during different visits. The ICCs obtained for the intraobserver and intervisit variability tests are greater than 0.75 for the total retina and all intraretinal layers, except the inner nuclear layer intraobserver and interobserver test and the outer plexiform layer, intraobserver, interobserver, and intervisit test. Our results indicate that thickness measurements for the total retina and all intraretinal layers (except the outer segment/RPE junction) performed using OCTRIMA are highly repeatable and reproducible.

  1. ORBS, ORCS, OACS, a Software Suite for Data Reduction and Analysis of the Hyperspectral Imagers SITELLE and SpIOMM

    NASA Astrophysics Data System (ADS)

    Martin, T.; Drissen, L.; Joncas, G.

    2015-09-01

    SITELLE (installed in 2015 at the Canada-France-Hawaii Telescope) and SpIOMM (a prototype attached to the Observatoire du Mont-Mégantic) are the first Imaging Fourier Transform Spectrometers (IFTS) capable of obtaining a hyperspectral data cube which samples a 12 arc minutes field of view into four millions of visible spectra. The result of each observation is made up of two interferometric data cubes which need to be merged, corrected, transformed and calibrated in order to get a spectral cube of the observed region ready to be analysed. ORBS is a fully automatic data reduction software that has been entirely designed for this purpose. The data size (up to 68 Gb for larger science cases) and the computational needs have been challenging and the highly parallelized object-oriented architecture of ORBS reflects the solutions adopted which made possible to process 68 Gb of raw data in less than 11 hours using 8 cores and 22.6 Gb of RAM. It is based on a core framework (ORB) that has been designed to support the whole software suite for data analysis (ORCS and OACS), data simulation (ORUS) and data acquisition (IRIS). They all aim to provide a strong basis for the creation and development of specialized analysis modules that could benefit the scientific community working with SITELLE and SpIOMM.

  2. Object oriented design provides flexible framework for electrophysiolgy software toolbox - biomed 2010.

    PubMed

    Gruner, Charlotte M

    2010-01-01

    This work describes a software platform to support an expandible toolbox for electrophysiology data analysis. The current focus of the toolbox, known as NeuroMAX is spike-sorting and spike-time analysis tools. A key feature of the toolbox is the ability for a user to connect tools into a workspace toolchain in a flexible, intelligent feed-forward manner that allows a tool to use any previously computed data set as input. Tool parameters can be saved and applied to other data sets. Tools and workspaces can be accessed to process data either directly from the MATLAB command line or from the NeuroMAX GUI. This work discusses the object-oriented design of the toolbox, including the data classes, workspace classes, and tool classes created to achieve this functionality.

  3. Control Software

    NASA Technical Reports Server (NTRS)

    1997-01-01

    Real-Time Innovations, Inc. (RTI) collaborated with Ames Research Center, the Jet Propulsion Laboratory and Stanford University to leverage NASA research to produce ControlShell software. RTI is the first "graduate" of Ames Research Center's Technology Commercialization Center. The ControlShell system was used extensively on a cooperative project to enhance the capabilities of a Russian-built Marsokhod rover being evaluated for eventual flight to Mars. RTI's ControlShell is complex, real-time command and control software, capable of processing information and controlling mechanical devices. One ControlShell tool is StethoScope. As a real-time data collection and display tool, StethoScope allows a user to see how a program is running without changing its execution. RTI has successfully applied its software savvy in other arenas, such as telecommunications, networking, video editing, semiconductor manufacturing, automobile systems, and medical imaging.

  4. Informatics in radiology (infoRAD): free DICOM image viewing and processing software for the Macintosh computer: what's available and what it can do for you.

    PubMed

    Escott, Edward J; Rubinstein, David

    2004-01-01

    It is often necessary for radiologists to use digital images in presentations and conferences. Most imaging modalities produce images in the Digital Imaging and Communications in Medicine (DICOM) format. The image files tend to be large and thus cannot be directly imported into most presentation software, such as Microsoft PowerPoint; the large files also consume storage space. There are many free programs that allow viewing and processing of these files on a personal computer, including conversion to more common file formats such as the Joint Photographic Experts Group (JPEG) format. Free DICOM image viewing and processing software for computers running on the Microsoft Windows operating system has already been evaluated. However, many people use the Macintosh (Apple Computer) platform, and a number of programs are available for these users. The World Wide Web was searched for free DICOM image viewing or processing software that was designed for the Macintosh platform or is written in Java and is therefore platform independent. The features of these programs and their usability were evaluated. There are many free programs for the Macintosh platform that enable viewing and processing of DICOM images.

  5. High-throughput image analysis of tumor spheroids: a user-friendly software application to measure the size of spheroids automatically and accurately.

    PubMed

    Chen, Wenjin; Wong, Chung; Vosburgh, Evan; Levine, Arnold J; Foran, David J; Xu, Eugenia Y

    2014-07-08

    The increasing number of applications of three-dimensional (3D) tumor spheroids as an in vitro model for drug discovery requires their adaptation to large-scale screening formats in every step of a drug screen, including large-scale image analysis. Currently there is no ready-to-use and free image analysis software to meet this large-scale format. Most existing methods involve manually drawing the length and width of the imaged 3D spheroids, which is a tedious and time-consuming process. This study presents a high-throughput image analysis software application - SpheroidSizer, which measures the major and minor axial length of the imaged 3D tumor spheroids automatically and accurately; calculates the volume of each individual 3D tumor spheroid; then outputs the results in two different forms in spreadsheets for easy manipulations in the subsequent data analysis. The main advantage of this software is its powerful image analysis application that is adapted for large numbers of images. It provides high-throughput computation and quality-control workflow. The estimated time to process 1,000 images is about 15 min on a minimally configured laptop, or around 1 min on a multi-core performance workstation. The graphical user interface (GUI) is also designed for easy quality control, and users can manually override the computer results. The key method used in this software is adapted from the active contour algorithm, also known as Snakes, which is especially suitable for images with uneven illumination and noisy background that often plagues automated imaging processing in high-throughput screens. The complimentary "Manual Initialize" and "Hand Draw" tools provide the flexibility to SpheroidSizer in dealing with various types of spheroids and diverse quality images. This high-throughput image analysis software remarkably reduces labor and speeds up the analysis process. Implementing this software is beneficial for 3D tumor spheroids to become a routine in vitro model

  6. Fundus image fusion in EYEPLAN software: An evaluation of a novel technique for ocular melanoma radiation treatment planning

    SciTech Connect

    Daftari, Inder K.; Mishra, Kavita K.; O'Brien, Joan M.; and others

    2010-10-15

    Purpose: The purpose of this study is to evaluate a novel approach for treatment planning using digital fundus image fusion in EYEPLAN for proton beam radiation therapy (PBRT) planning for ocular melanoma. The authors used a prototype version of EYEPLAN software, which allows for digital registration of high-resolution fundus photographs. The authors examined the improvement in tumor localization by replanning with the addition of fundus photo superimposition in patients with macular area tumors. Methods: The new version of EYEPLAN (v3.05) software allows for the registration of fundus photographs as a background image. This is then used in conjunction with clinical examination, tantalum marker clips, surgeon's mapping, and ultrasound to draw the tumor contour accurately. In order to determine if the fundus image superimposition helps in tumor delineation and treatment planning, the authors identified 79 patients with choroidal melanoma in the macular location that were treated with PBRT. All patients were treated to a dose of 56 GyE in four fractions. The authors reviewed and replanned all 79 macular melanoma cases with superimposition of pretreatment and post-treatment fundus imaging in the new EYEPLAN software. For patients with no local failure, the authors analyzed whether fundus photograph fusion accurately depicted and confirmed tumor volumes as outlined in the original treatment plan. For patients with local failure, the authors determined whether the addition of the fundus photograph might have benefited in terms of more accurate tumor volume delineation. Results: The mean follow-up of patients was 33.6{+-}23 months. Tumor growth was seen in six eyes of the 79 macular lesions. All six patients were marginal failures or tumor miss in the region of dose fall-off, including one patient with both in-field recurrence as well as marginal. Among the six recurrences, three were managed by enucleation and one underwent retreatment with proton therapy. Three

  7. C++ software integration for a high-throughput phase imaging platform

    NASA Astrophysics Data System (ADS)

    Kandel, Mikhail E.; Luo, Zelun; Han, Kevin; Popescu, Gabriel

    2015-03-01

    The multi-shot approach in SLIM requires reliable, synchronous, and parallel operation of three independent hardware devices - not meeting these challenges results in degraded phase and slow acquisition speeds, narrowing applications to holistic statements about complex phenomena. The relative youth of quantitative imaging and the lack of ready-made commercial hardware and tools further compounds the problem as Higher level programming languages result in inflexible, experiment specific instruments limited by ill-fitting computational modules, resulting in a palpable chasm between promised and realized hardware performance. Furthermore, general unfamiliarity with intricacies such as background calibration, objective lens attenuation, along with spatial light modular alignment, makes successful measurements difficult for the inattentive or uninitiated. This poses an immediate challenge for moving our techniques beyond the lab to biologically oriented collaborators and clinical practitioners. To meet these challenges, we present our new Quantitative Phase Imaging pipeline, with improved instrument performance, friendly user interface and robust data processing features, enabling us to acquire and catalog clinical datasets hundreds of gigapixels in size.

  8. FIRE: an open-software suite for real-time 2D/3D image registration for image guided radiotherapy research

    NASA Astrophysics Data System (ADS)

    Furtado, H.; Gendrin, C.; Spoerk, J.; Steiner, E.; Underwood, T.; Kuenzler, T.; Georg, D.; Birkfellner, W.

    2016-03-01

    Radiotherapy treatments have changed at a tremendously rapid pace. Dose delivered to the tumor has escalated while organs at risk (OARs) are better spared. The impact of moving tumors during dose delivery has become higher due to very steep dose gradients. Intra-fractional tumor motion has to be managed adequately to reduce errors in dose delivery. For tumors with large motion such as tumors in the lung, tracking is an approach that can reduce position uncertainty. Tumor tracking approaches range from purely image intensity based techniques to motion estimation based on surrogate tracking. Research efforts are often based on custom designed software platforms which take too much time and effort to develop. To address this challenge we have developed an open software platform especially focusing on tumor motion management. FLIRT is a freely available open-source software platform. The core method for tumor tracking is purely intensity based 2D/3D registration. The platform is written in C++ using the Qt framework for the user interface. The performance critical methods are implemented on the graphics processor using the CUDA extension. One registration can be as fast as 90ms (11Hz). This is suitable to track tumors moving due to respiration (~0.3Hz) or heartbeat (~1Hz). Apart from focusing on high performance, the platform is designed to be flexible and easy to use. Current use cases range from tracking feasibility studies, patient positioning and method validation. Such a framework has the potential of enabling the research community to rapidly perform patient studies or try new methods.

  9. Using Image Pro Plus Software to Develop Particle Mapping on Genesis Solar Wind Collector Surfaces

    NASA Technical Reports Server (NTRS)

    Rodriquez, Melissa C.; Allton, J. H.; Burkett, P. J.

    2012-01-01

    The continued success of the Genesis mission science team in analyzing solar wind collector array samples is partially based on close collaboration of the JSC curation team with science team members who develop cleaning techniques and those who assess elemental cleanliness at the levels of detection. The goal of this collaboration is to develop a reservoir of solar wind collectors of known cleanliness to be available to investigators. The heart and driving force behind this effort is Genesis mission PI Don Burnett. While JSC contributes characterization, safe clean storage, and benign collector cleaning with ultrapure water (UPW) and UV ozone, Burnett has coordinated more exotic and rigorous cleaning which is contributed by science team members. He also coordinates cleanliness assessment requiring expertise and instruments not available in curation, such as XPS, TRXRF [1,2] and synchrotron TRXRF. JSC participates by optically documenting the particle distributions as cleaning steps progress. Thus, optical document supplements SEM imaging and analysis, and elemental assessment by TRXRF.

  10. Cytopathology whole slide images and virtual microscopy adaptive tutorials: A software pilot

    PubMed Central

    Van Es, Simone L.; Pryor, Wendy M.; Belinson, Zack; Salisbury, Elizabeth L.; Velan, Gary M.

    2015-01-01

    Background: The constant growth in the body of knowledge in medicine requires pathologists and pathology trainees to engage in continuing education. Providing them with equitable access to efficient and effective forms of education in pathology (especially in remote and rural settings) is important, but challenging. Methods: We developed three pilot cytopathology virtual microscopy adaptive tutorials (VMATs) to explore a novel adaptive E-learning platform (AeLP) which can incorporate whole slide images for pathology education. We collected user feedback to further develop this educational material and to subsequently deploy randomized trials in both pathology specialist trainee and also medical student cohorts. Cytopathology whole slide images were first acquired then novel VMATs teaching cytopathology were created using the AeLP, an intelligent tutoring system developed by Smart Sparrow. The pilot was run for Australian pathologists and trainees through the education section of Royal College of Pathologists of Australasia website over a period of 9 months. Feedback on the usability, impact on learning and any technical issues was obtained using 5-point Likert scale items and open-ended feedback in online questionnaires. Results: A total of 181 pathologists and pathology trainees anonymously attempted the three adaptive tutorials, a smaller proportion of whom went on to provide feedback at the end of each tutorial. VMATs were perceived as effective and efficient E-learning tools for pathology education. User feedback was positive. There were no significant technical issues. Conclusion: During this pilot, the user feedback on the educational content and interface and the lack of technical issues were helpful. Large scale trials of similar online cytopathology adaptive tutorials were planned for the future. PMID:26605119

  11. An Upgrade of the Imaging for Hypersonic Experimental Aeroheating Testing (IHEAT) Software

    NASA Technical Reports Server (NTRS)

    Mason, Michelle L.; Rufer, Shann J.

    2015-01-01

    The Imaging for Hypersonic Experimental Aeroheating Testing (IHEAT) code is used at NASA Langley Research Center to analyze global aeroheating data on wind tunnel models tested in the Langley Aerothermodynamics Laboratory. One-dimensional, semi-infinite heating data derived from IHEAT are used to design thermal protection systems to mitigate the risks due to the aeroheating loads on hypersonic vehicles, such as re-entry vehicles during descent and landing procedures. This code was originally written in the PV-WAVE programming language to analyze phosphor thermography data from the two-color, relativeintensity system developed at Langley. To increase the efficiency, functionality, and reliability of IHEAT, the code was migrated to MATLAB syntax and compiled as a stand-alone executable file labeled version 4.0. New features of IHEAT 4.0 include the options to batch process all of the data from a wind tunnel run, to map the two-dimensional heating distribution to a three-dimensional computer-aided design model of the vehicle to be viewed in Tecplot, and to extract data from a segmented line that follows an interesting feature in the data. Results from IHEAT 4.0 were compared on a pixel level to the output images from the legacy code to validate the program. The differences between the two codes were on the order of 10-5 to 10-7. IHEAT 4.0 replaces the PV-WAVE version as the production code for aeroheating experiments conducted in the hypersonic facilities at NASA Langley.

  12. WASI-2D: A software tool for regionally optimized analysis of imaging spectrometer data from deep and shallow waters

    NASA Astrophysics Data System (ADS)

    Gege, Peter

    2014-01-01

    An image processing software has been developed which allows quantitative analysis of multi- and hyperspectral data from oceanic, coastal and inland waters. It has been implemented into the Water Colour Simulator WASI, which is a tool for the simulation and analysis of optical properties and light field parameters of deep and shallow waters. The new module WASI-2D can import atmospherically corrected images from airborne sensors and satellite instruments in various data formats and units like remote sensing reflectance or radiance. It can be easily adapted by the user to different sensors and to optical properties of the studied area. Data analysis is done by inverse modelling using established analytical models. The bio-optical model of the water column accounts for gelbstoff (coloured dissolved organic matter, CDOM), detritus, and mixtures of up to 6 phytoplankton classes and 2 spectrally different types of suspended matter. The reflectance of the sea floor is treated as sum of up to 6 substrate types. An analytic model of downwelling irradiance allows wavelength dependent modelling of sun glint and sky glint at the water surface. The provided database covers the spectral range from 350 to 1000 nm in 1 nm intervals. It can be exchanged easily to represent the optical properties of water constituents, bottom types and the atmosphere of the studied area.

  13. Leap Motion Gesture Control With Carestream Software in the Operating Room to Control Imaging: Installation Guide and Discussion.

    PubMed

    Pauchot, Julien; Di Tommaso, Laetitia; Lounis, Ahmed; Benassarou, Mourad; Mathieu, Pierre; Bernot, Dominique; Aubry, Sébastien

    2015-12-01

    Nowadays, routine cross-sectional imaging viewing during a surgical procedure requires physical contact with an interface (mouse or touch-sensitive screen). Such contact risks exposure to aseptic conditions and causes loss of time. Devices such as the recently introduced Leap Motion (Leap Motion Society, San Francisco, CA), which enables interaction with the computer without any physical contact, are of wide interest in the field of surgery, but configuration and ergonomics are key challenges for the practitioner, imaging software, and surgical environment. This article aims to suggest an easy configuration of Leap Motion on a PC for optimized use with Carestream Vue PACS v11.3.4 (Carestream Health, Inc, Rochester, NY) using a plug-in (to download at https://drive.google.com/open?id=0B_F4eBeBQc3yNENvTXlnY09qS00&authuser=0) and a video tutorial (https://www.youtube.com/watch?v=yVPTgxg-SIk). Videos of surgical procedure and discussion about innovative gesture control technology and its various configurations are provided in this article.

  14. Tools for Scientist Engagement in E/PO: NASA SMD Community Workspace and Online Resources

    NASA Astrophysics Data System (ADS)

    Dalton, H.; Shipp, S. S.; Grier, J.; Gross, N. A.; Buxner, S.; Bartolone, L.; Peticolas, L. M.; Woroner, M.; Schwerin, T. G.

    2014-12-01

    The Science Mission Directorate (SMD) Science Education and Public Outreach (E/PO) Forums are here to help you get involved in E/PO! The Forums have been developing several online resources to support scientists who are - or who are interested in becoming - involved in E/PO. These include NASA Wavelength, EarthSpace, and the SMD E/PO online community workspace. NASA Wavelength is the one-stop shop of all peer-reviewed NASA education resources to find materials you - or your audiences - can use. Browse by audience (pre-K through 12, higher education, and informal education) or topic, or choose to search for something specific by keyword and audience. http://nasawavelength.org. EarthSpace, an online clearinghouse of Earth and space materials for use in the higher education classroom, is driven by a powerful search engine that allows you to browse the collection of resources by science topic, audience, type of material or key terms. All materials are peer-reviewed before posting, and because all submissions receive a digital object identifier (doi), submitted materials can be listed as publications. http://www.lpi.usra.edu/earthspace. The SMD E/PO online community workspace contains many resources for scientists. These include one-page guides on how to get involved, tips on how to make the most of your time spent on E/PO, and sample activities, as well as news on funding, policy, and what's happening in the E/PO community. The workspace also provides scientists and the public pathways to find opportunities for participation in E/PO, to learn about SMD E/PO projects and their impacts, to connect with SMD E/PO practitioners, and to explore resources to improve professional E/PO practice, including literature reviews, information about the Next Generation Science Standards, and best practices in evaluation and engaging diverse audiences. http://smdepo.org.

  15. Collecting field data from Mars Exploration Rover Spirit and Opportunity Images: Development of 3-D Visualization and Data-Mining Software

    NASA Astrophysics Data System (ADS)

    Eppes, M. C.; Willis, A.; Zhou, B.

    2010-12-01

    NASA’s two Mars rover spacecraft, Spirit and Opportunity, have collected more than 4 years worth of data from nine imaging instruments producing greater than 200k images. To date, however, the potential ‘field’ data that these images represent has remained relatively untapped because of a lack of software with which to readily analyze the images quantitatively. We have developed prototype software that allows scientists to locate and explore 2D and 3D imagery captured by the NASA's Mars Exploratory Rover (MER) mission robots Spirit and Opportunity. For example, using our software, a person could measure the dimensions of a rock or the strike and dip of a bedding plane. The developed software has three aspects that make it distinct from existing approaches for indexing large sets of imagery: (1) a computationally efficient image search engine capable of locating MER images containing features of interest, rocks in particular, (2) an interface for making measurements (distances and orientations) from stereographic image pairs and (3) remote browsing and storage capabilities that removes the burden of storing and managing these very large image sets. Two methods of search are supported (i) a rock detection algorithm for finding images that contain rock-like structures having a specified size and (ii) a generic query-by-image search which uses exemplar image(s) of a desired object to locate other images within the MER data repository that contain similar structures (i.e. one could search for all images of sand dunes). Query by image capabilities are made possible via a bag-of-features (e.g. Labeznik et. al. 2003; Schmid et. al. 2002) representation of the image data which compresses the image into a small set of features which are robust to changes in illumination and perspective. Searches are then reduced to looking for feature sets which have similar values; a task that is computationally tractable providing quick search results for complex image-based queries

  16. Comparison of software and human observers in reading images of the CDMAM test object to assess digital mammography systems

    NASA Astrophysics Data System (ADS)

    Young, Kenneth C.; Cook, James J. H.; Oduko, Jennifer M.; Bosmans, Hilde

    2006-03-01

    European Guidelines for quality control in digital mammography specify minimum and achievable standards of image quality in terms of threshold contrast, based on readings of images of the CDMAM test object by human observers. However this is time-consuming and has large inter-observer error. To overcome these problems a software program (CDCOM) is available to automatically read CDMAM images, but the optimal method of interpreting the output is not defined. This study evaluates methods of determining threshold contrast from the program, and compares these to human readings for a variety of mammography systems. The methods considered are (A) simple thresholding (B) psychometric curve fitting (C) smoothing and interpolation and (D) smoothing and psychometric curve fitting. Each method leads to similar threshold contrasts but with different reproducibility. Method (A) had relatively poor reproducibility with a standard error in threshold contrast of 18.1 +/- 0.7%. This was reduced to 8.4% by using a contrast-detail curve fitting procedure. Method (D) had the best reproducibility with an error of 6.7%, reducing to 5.1% with curve fitting. A panel of 3 human observers had an error of 4.4% reduced to 2.9 % by curve fitting. All automatic methods led to threshold contrasts that were lower than for humans. The ratio of human to program threshold contrasts varied with detail diameter and was 1.50 +/- .04 (sem) at 0.1mm and 1.82 +/- .06 at 0.25mm for method (D). There were good correlations between the threshold contrast determined by humans and the automated methods.

  17. The Decoding Toolbox (TDT): a versatile software package for multivariate analyses of functional imaging data

    PubMed Central

    Hebart, Martin N.; Görgen, Kai; Haynes, John-Dylan

    2015-01-01

    The multivariate analysis of brain signals has recently sparked a great amount of interest, yet accessible and versatile tools to carry out decoding analyses are scarce. Here we introduce The Decoding Toolbox (TDT) which represents a user-friendly, powerful and flexible package for multivariate analysis of functional brain imaging data. TDT is written in Matlab and equipped with an interface to the widely used brain data analysis package SPM. The toolbox allows running fast whole-brain analyses, region-of-interest analyses and searchlight analyses, using machine learning classifiers, pattern correlation analysis, or representational similarity analysis. It offers automatic creation and visualization of diverse cross-validation schemes, feature scaling, nested parameter selection, a variety of feature selection methods, multiclass capabilities, and pattern reconstruction from classifier weights. While basic users can implement a generic analysis in one line of code, advanced users can extend the toolbox to their needs or exploit the structure to combine it with external high-performance classification toolboxes. The toolbox comes with an example data set which can be used to try out the various analysis methods. Taken together, TDT offers a promising option for researchers who want to employ multivariate analyses of brain activity patterns. PMID:25610393

  18. User's guide for mapIMG 3--Map image re-projection software package

    USGS Publications Warehouse

    Finn, Michael P.; Mattli, David M.

    2012-01-01

    Version 0.0 (1995), Dan Steinwand, U.S. Geological Survey (USGS)/Earth Resources Observation Systems (EROS) Data Center (EDC)--Version 0.0 was a command line version for UNIX that required four arguments: the input metadata, the output metadata, the input data file, and the output destination path. Version 1.0 (2003), Stephen Posch and Michael P. Finn, USGS/Mid-Continent Mapping Center (MCMC--Version 1.0 added a GUI interface that was built using the Qt library for cross platform development. Version 1.01 (2004), Jason Trent and Michael P. Finn, USGS/MCMC--Version 1.01 suggested bounds for the parameters of each projection. Support was added for larger input files, storage of the last used input and output folders, and for TIFF/ GeoTIFF input images. Version 2.0 (2005), Robert Buehler, Jason Trent, and Michael P. Finn, USGS/National Geospatial Technical Operations Center (NGTOC)--Version 2.0 added Resampling Methods (Mean, Mode, Min, Max, and Sum), updated the GUI design, and added the viewer/pre-viewer. The metadata style was changed to XML and was switched to a new naming convention. Version 3.0 (2009), David Mattli and Michael P. Finn, USGS/Center of Excellence for Geospatial Information Science (CEGIS)--Version 3.0 brings optimized resampling methods, an updated GUI, support for less than global datasets, UTM support and the whole codebase was ported to Qt4.

  19. On-line access to weather satellite imagery and image manipulation software

    NASA Technical Reports Server (NTRS)

    Emery, W.; Kelley, T.; Dozier, J.; Rotar, P.

    1995-01-01

    Advanced Very High Resolution Radiometer and Geostationary Operational Environmental Satellite imagery, received by antennas located at the University of Colorado, are made available to the Internet users through an on-line data access system. Created as a 'test bed' system for the National Aeronautics and Space Administration's future Earth Observing System Data and Information System, this test bed provides an opportunity to test both the technical requirements of an on-line data system and the different ways in which the general user community would employ such a system. Initiated in December 1991, the basic data system experiment four major evolutionary changes in response to user requests and requirements. Features added with these changes were the addition of on-line browse, user subsetting, and dynamic image processing/navigation. Over its lifetime the system has grown to a maximum of over 2500 registered users, and after losing many of these users due to hardware changes, the system is once again growing with its own independent mass storage system.

  20. On-Line Access to Weather Satellite Imagery and Image Manipulation Software

    NASA Technical Reports Server (NTRS)

    Emery, William J.; Kelley, T.; Dozier, J.; Rotar, P.

    1995-01-01

    Advanced Very High Resolution Radiometer and Geostationary Operational Environmental Satellite Imagery, received by antennas located at the University of Colorado, are made available to the Internet users through an on-line data access system. Created as a 'test bed' data system for the National Aeronautics and Space Administration's future Earth Observing System Data and Information System, this test bed provides an opportunity to test both the technical requirements of an on-line data system and the different ways in which the general user community would employ such a system. Initiated in December 1991, the basic data system experienced four major evolutionary changes in response to user requests and requirements. Features added with these changes were the addition of on-line browse, user subsetting, and dynamic image processing/navigation. Over its lifetime the system has grown to a maximum of over 2500 registered users, and after losing many of these users due to hardware changes, the system is once again growing with its own independent mass storage system.

  1. Flexible workspace design and ergonomics training: impacts on the psychosocial work environment, musculoskeletal health, and work effectiveness among knowledge workers.

    PubMed

    Robertson, Michelle M; Huang, Yueng-Hsiang; O'Neill, Michael J; Schleifer, Lawrence M

    2008-07-01

    A macroergonomics intervention consisting of flexible workspace design and ergonomics training was conducted to examine the effects on psychosocial work environment, musculoskeletal health, and work effectiveness in a computer-based office setting. Knowledge workers were assigned to one of four conditions: flexible workspace (n=121), ergonomics training (n=92), flexible workspace+ergonomics training (n=31), and a no-intervention control (n=45). Outcome measures were collected 2 months prior to the intervention and 3 and 6 months post-intervention. Overall, the study results indicated positive, significant effects on the outcome variables for the two intervention groups compared to the control group, including work-related musculoskeletal discomfort, job control, environmental satisfaction, sense of community, ergonomic climate, communication and collaboration, and business process efficiency (time and costs). However, attrition of workers in the ergonomics training condition precluded an evaluation of the effects of this intervention. This study suggests that a macroergonomics intervention is effective among knowledge workers in office settings.

  2. Comparison of retinal thickness by Fourier-domain optical coherence tomography and OCT retinal image analysis software segmentation analysis derived from Stratus optical coherence tomography images

    NASA Astrophysics Data System (ADS)

    Tátrai, Erika; Ranganathan, Sudarshan; Ferencz, Mária; Debuc, Delia Cabrera; Somfai, Gábor Márk

    2011-05-01

    Purpose: To compare thickness measurements between Fourier-domain optical coherence tomography (FD-OCT) and time-domain OCT images analyzed with a custom-built OCT retinal image analysis software (OCTRIMA). Methods: Macular mapping (MM) by StratusOCT and MM5 and MM6 scanning protocols by an RTVue-100 FD-OCT device are performed on 11 subjects with no retinal pathology. Retinal thickness (RT) and the thickness of the ganglion cell complex (GCC) obtained with the MM6 protocol are compared for each early treatment diabetic retinopathy study (ETDRS)-like region with corresponding results obtained with OCTRIMA. RT results are compared by analysis of variance with Dunnett post hoc test, while GCC results are compared by paired t-test. Results: A high correlation is obtained for the RT between OCTRIMA and MM5 and MM6 protocols. In all regions, the StratusOCT provide the lowest RT values (mean difference 43 +/- 8 μm compared to OCTRIMA, and 42 +/- 14 μm compared to RTVue MM6). All RTVue GCC measurements were significantly thicker (mean difference between 6 and 12 μm) than the GCC measurements of OCTRIMA. Conclusion: High correspondence of RT measurements is obtained not only for RT but also for the segmentation of intraretinal layers between FD-OCT and StratusOCT-derived OCTRIMA analysis. However, a correction factor is required to compensate for OCT-specific differences to make measurements more comparable to any available OCT device.

  3. TU-AB-303-08: GPU-Based Software Platform for Efficient Image-Guided Adaptive Radiation Therapy

    SciTech Connect

    Park, S; Robinson, A; McNutt, T; Wong, J; Lee, J; Plishker, W; Zaki, G; Shekhar, R

    2015-06-15

    Purpose: In this study, we develop an integrated software platform for adaptive radiation therapy (ART) that combines fast and accurate image registration, segmentation, and dose computation/accumulation methods. Methods: The proposed system consists of three key components; 1) deformable image registration (DIR), 2) automatic segmentation, and 3) dose computation/accumulation. The computationally intensive modules including DIR and dose computation have been implemented on a graphics processing unit (GPU). All required patient-specific data including the planning CT (pCT) with contours, daily cone-beam CTs, and treatment plan are automatically queried and retrieved from their own databases. To improve the accuracy of DIR between pCT and CBCTs, we use the double force demons DIR algorithm in combination with iterative CBCT intensity correction by local intensity histogram matching. Segmentation of daily CBCT is then obtained by propagating contours from the pCT. Daily dose delivered to the patient is computed on the registered pCT by a GPU-accelerated superposition/convolution algorithm. Finally, computed daily doses are accumulated to show the total delivered dose to date. Results: Since the accuracy of DIR critically affects the quality of the other processes, we first evaluated our DIR method on eight head-and-neck cancer cases and compared its performance. Normalized mutual-information (NMI) and normalized cross-correlation (NCC) computed as similarity measures, and our method produced overall NMI of 0.663 and NCC of 0.987, outperforming conventional methods by 3.8% and 1.9%, respectively. Experimental results show that our registration method is more consistent and roust than existing algorithms, and also computationally efficient. Computation time at each fraction took around one minute (30–50 seconds for registration and 15–25 seconds for dose computation). Conclusion: We developed an integrated GPU-accelerated software platform that enables accurate and

  4. Kinematic, workspace and singularity analysis of a new parallel robot used in minimally invasive surgery

    NASA Astrophysics Data System (ADS)

    Stoica, Alin; Pisla, Doina; Andras, Szilaghyi; Gherman, Bogdan; Gyurka, Bela-Zoltan; Plitea, Nicolae

    2013-03-01

    In the last ten years, due to development in robotic assisted surgery, the minimally invasive surgery has greatly changed. Until now, the vast majority of robots used in surgery, have serial structures. Due to the orientation parallel module, the structure is able to reduce the pressure exerted on the entrance point in the patient's abdominal wall. The parallel robot can also handle both a laparoscope as well an active instrument for different surgical procedures. The advantage of this parallel structure is that the geometric model has been obtained through an analytical approach. The kinematic modelling of a new parallel architecture, the inverse and direct geometric model and the inverse and direct kinematic models for velocities and accelerations are being determined. The paper will demonstrate that with this parallel structure, one can obtain the necessary workspace required for a minimally invasive operation. The robot workspace was generated using the inverse geometric model. An indepth study of different types of singularity is performed, allowing the development of safe control algorithms of the experimental model. Some kinematic simulation results and the experimental model of the robot are presented in the paper.

  5. Proposal for future diagnosis and management of vascular tumors by using automatic software for image processing and statistic prediction

    PubMed Central

    Popescu, MD; Draghici, L; Secheli, I; Secheli, M; Codrescu, M; Draghici, I

    2015-01-01

    Rationale. Infantile Hemangiomas (IH) are the most frequent tumors of vascular origin, and the differential diagnosis from vascular malformations is difficult to establish. Specific types of IH due to the location, dimensions and fast evolution, can determine important functional and esthetic sequels. To avoid these unfortunate consequences it is necessary to establish the exact appropriate moment to begin the treatment and decide which the most adequate therapeutic procedure is. Objective. Based on clinical data collected by a serial clinical observations correlated with imaging data, and processed by a computer-aided diagnosis system (CAD), the study intended to develop a treatment algorithm to accurately predict the best final results, from the esthetical and functional point of view, for a certain type of lesion. Methods and Results. The preliminary database was composed of 75 patients divided into 4 groups according to the treatment management they received: medical therapy, sclerotherapy, surgical excision and no treatment. The serial clinical observation was performed each month and all the data was processed by using CAD. Discussions. The project goal was to create a software that incorporated advanced methods to accurately measure the specific IH lesions, integrated medical information, statistical methods and computational methods to correlate this information with that obtained from the processing of images. Based on these correlations, a prediction mechanism of the evolution of hemangioma, which helped determine the best method of therapeutic intervention to minimize further complications, was established. Abbreviations: Infantile Hemangiomas = IH, Computer Aided Diagnosis = CAD, Society for the Study of Vascular Anomalies = ISSVA, Color-coded duplex sonography = CCDS PMID:25914738

  6. Man-machine interactive imaging and data processing using high-speed digital mass storage

    NASA Technical Reports Server (NTRS)

    Alsberg, H.; Nathan, R.

    1975-01-01

    The role of vision in teleoperation has been recognized as an important element in the man-machine control loop. In most applications of remote manipulation, direct vision cannot be used. To overcome this handicap, the human operator's control capabilities are augmented by a television system. This medium provides a practical and useful link between workspace and the control station from which the operator perform his tasks. Human performance deteriorates when the images are degraded as a result of instrumental and transmission limitations. Image enhancement is used to bring out selected qualities in a picture to increase the perception of the observer. A general purpose digital computer, an extensive special purpose software system is used to perform an almost unlimited repertoire of processing operations.

  7. Repeatability and Reproducibility of Quantitative Corneal Shape Analysis after Orthokeratology Treatment Using Image-Pro Plus Software

    PubMed Central

    Mei, Ying; Tang, Zhiping

    2016-01-01

    Purpose. To evaluate the repeatability and reproducibility of quantitative analysis of the morphological corneal changes after orthokeratology treatment using “Image-Pro Plus 6.0” software (IPP). Methods. Three sets of measurements were obtained: two sets by examiner 1 with 5 days apart and one set by examiner 2 on the same day. Parameters of the eccentric distance, eccentric angle, area, and roundness of the corneal treatment zone were measured using IPP. The intraclass correlation coefficient (ICC) and repetitive coefficient (COR) were used to calculate the repeatability and reproducibility of these three sets of measurements. Results. ICC analysis suggested “excellent” reliability of more than 0.885 for all variables, and COR values were less than 10% for all variables within the same examiner. ICC analysis suggested “excellent” reliability for all variables of more than 0.90, and COR values were less than 10% for all variables between different examiners. All extreme values of the eccentric distance and area of the treatment zone pointed to the same material number in three sets of measurements. Conclusions. IPP could be used to acquire the exact data of the characteristic morphological corneal changes after orthokeratology treatment with good repeatability and reproducibility. This trial is registered with trial registration number: ChiCTR-IPR-14005505. PMID:27774312

  8. Neuroimaging reveals enhanced activation in a reach-selective brain area for objects located within participants' typical hand workspaces.

    PubMed

    Gallivan, Jason P; McLean, Adam; Culham, Jody C

    2011-11-01

    In recent years, there has been growing excitement within cognitive neuroscience about the concept of embodiment: How do the capabilities and limitations of our physical bodies affect neural representations in the brain? Neuropsychological and neurophysiological studies show clear evidence that short-term visuomotor experience can influence the encoding of the space around the body in parietal cortex. For example, tool-use may expand the neural representation of peripersonal space. But how is this initial spatial representation influenced by a lifetime of object-related interactions? To examine this question we used functional magnetic resonance imaging (fMRI) to investigate the neural effects of an individual's hand preferences for acting within peripersonal space. Left- and right-handed participants viewed real-world objects at different locations accessible by either the left hand, right hand, or neither hand. The superior parieto-occipital cortex (SPOC), an area most often implicated in reaching actions, showed enhanced visual responses for objects located within the range of space in which each group typically acts. Specifically, in right-handers, who strongly prefer grasping with the right hand, SPOC showed strongest activation for objects located within the range of space for the right hand only. In contrast, in left-handers, who use their two hands comparably often in visuomotor tasks, SPOC showed strongest activation for objects located within the range of space of either hand. These findings show that, even in the absence of overt responses, real 3D objects located in the individual's typical workspace for hand actions automatically invoke enhanced responses in associated visuomotor areas of the brain.

  9. Dental application of novel finite element analysis software for three-dimensional finite element modeling of a dentulous mandible from its computed tomography images.

    PubMed

    Nakamura, Keiko; Tajima, Kiyoshi; Chen, Ker-Kong; Nagamatsu, Yuki; Kakigawa, Hiroshi; Masumi, Shin-ich

    2013-12-01

    This study focused on the application of novel finite-element analysis software for constructing a finite-element model from the computed tomography data of a human dentulous mandible. The finite-element model is necessary for evaluating the mechanical response of the alveolar part of the mandible, resulting from occlusal force applied to the teeth during biting. Commercially available patient-specific general computed tomography-based finite-element analysis software was solely applied to the finite-element analysis for the extraction of computed tomography data. The mandibular bone with teeth was extracted from the original images. Both the enamel and the dentin were extracted after image processing, and the periodontal ligament was created from the segmented dentin. The constructed finite-element model was reasonably accurate using a total of 234,644 nodes and 1,268,784 tetrahedral and 40,665 shell elements. The elastic moduli of the heterogeneous mandibular bone were determined from the bone density data of the computed tomography images. The results suggested that the software applied in this study is both useful and powerful for creating a more accurate three-dimensional finite-element model of a dentulous mandible from the computed tomography data without the need for any other software.

  10. Integration of target and hand position signals in the posterior parietal cortex: effects of workspace and hand vision.

    PubMed

    Buneo, Christopher A; Andersen, Richard A

    2012-07-01

    Previous findings suggest the posterior parietal cortex (PPC) contributes to arm movement planning by transforming target and limb position signals into a desired reach vector. However, the neural mechanisms underlying this transformation remain unclear. In the present study we examined the responses of 109 PPC neurons as movements were planned and executed to visual targets presented over a large portion of the reaching workspace. In contrast to previous studies, movements were made without concurrent visual and somatic cues about the starting position of the hand. For comparison, a subset of neurons was also examined with concurrent visual and somatic hand position cues. We found that single cells integrated target and limb position information in a very consistent manner across the reaching workspace. Approximately two-thirds of the neurons with significantly tuned activity (42/61 and 30/46 for left and right workspaces, respectively) coded targets and initial hand positions separably, indicating no hand-centered encoding, whereas the remaining one-third coded targets and hand positions inseparably, in a manner more consistent with the influence of hand-centered coordinates. The responses of both types of neurons were largely invariant with respect to the presence or absence of visual hand position cues, suggesting their corresponding coordinate frames and gain effects were unaffected by cue integration. The results suggest that the PPC uses a consistent scheme for computing reach vectors in different parts of the workspace that is robust to changes in the availability of somatic and visual cues about hand position.

  11. The Role of a Facilitated Online Workspace Component of a Community of Practice: Knowledge Building and Value Creation for NASA

    ERIC Educational Resources Information Center

    Davey, Bradford Thomas

    2013-01-01

    The purpose of this study was to examine the role of an online workspace component of a community in the work of a community of practice. Much has been studied revealing the importance of communities of practice to organizations, project success, and knowledge management and some of these same successes hold true for virtual communities of…

  12. Computer-aided 2D and 3D quantification of human stem cell fate from in vitro samples using Volocity high performance image analysis software.

    PubMed

    Piltti, Katja M; Haus, Daniel L; Do, Eileen; Perez, Harvey; Anderson, A J; Cummings, B J

    2011-11-01

    Accurate automated cell fate analysis of immunostained human stem cells from 2- and 3-dimensional (2D-3D) images would improve efficiency in the field of stem cell research. Development of an accurate and precise tool that reduces variability and the time needed for human stem cell fate analysis will improve productivity and interpretability of the data across research groups. In this study, we have created protocols for high performance image analysis software Volocity® to classify and quantify cytoplasmic and nuclear cell fate markers from 2D-3D images of human neural stem cells after in vitro differentiation. To enhance 3D image capture efficiency, we optimized the image acquisition settings of an Olympus FV10i® confocal laser scanning microscope to match our quantification protocols and improve cell fate classification. The methods developed in this study will allow for a more time efficient and accurate software based, operator validated, stem cell fate classification and quantification from 2D and 3D images, and yield the highest ≥94.4% correspondence with human recognized objects.

  13. An architectural model of conscious and unconscious brain functions: Global Workspace Theory and IDA.

    PubMed

    Baars, Bernard J; Franklin, Stan

    2007-11-01

    While neural net models have been developed to a high degree of sophistication, they have some drawbacks at a more integrative, "architectural" level of analysis. We describe a "hybrid" cognitive architecture that is implementable in neuronal nets, and which has uniform brainlike features, including activation-passing and highly distributed "codelets," implementable as small-scale neural nets. Empirically, this cognitive architecture accounts qualitatively for the data described by Baars' Global Workspace Theory (GWT), and Franklin's LIDA architecture, including state-of-the-art models of conscious contents in action-planning, Baddeley-style Working Memory, and working models of episodic and semantic longterm memory. These terms are defined both conceptually and empirically for the current theoretical domain. The resulting architecture meets four desirable goals for a unified theory of cognition: practical workability, autonomous agency, a plausible role for conscious cognition, and translatability into plausible neural terms. It also generates testable predictions, both empirical and computational.

  14. Collision-free motion of two robot arms in a common workspace

    NASA Technical Reports Server (NTRS)

    Basta, Robert A.; Mehrotra, Rajiv; Varanasi, Murali R.

    1987-01-01

    Collision-free motion of two robot arms in a common workspace is investigated. A collision-free motion is obtained by detecting collisions along the preplanned trajectories using a sphere model for the wrist of each robot and then modifying the paths and/or trajectories of one or both robots to avoid the collision. Detecting and avoiding collisions are based on the premise that: preplanned trajectories of the robots follow a straight line; collisions are restricted to between the wrists of the two robots (which corresponds to the upper three links of PUMA manipulators); and collisions never occur between the beginning points or end points on the straight line paths. The collision detection algorithm is described and some approaches to collision avoidance are discussed.

  15. The Secure Medical Research Workspace: An IT Infrastructure to Enable Secure Research on Clinical Data

    PubMed Central

    Shoffner, Michael; Owen, Phillips; Mostafa, Javed; Lamm, Brent; Wang, Xiaoshu; Schmitt, Charles P.; Ahalt, Stanley C.

    2013-01-01

    Clinical data has tremendous value for translational research, but only if security and privacy concerns can be addressed satisfactorily. A collaboration of clinical and informatics teams, including RENCI, NC TraCS, UNC’s School of Information and Library Science, Information Technology Service’s Research Computing and other partners at the University of North Carolina at Chapel Hill have developed a system called the Secure Medical Research Workspace (SMRW) that enables researchers to use clinical data securely for research. SMRW significantly minimizes the risk presented when using of identified clinical data, thereby protecting patients, researchers, and institutions associated with the data. The SMRW is built on a novel combination of virtualization and data leakage protection and can be combined with other protection methodologies and scaled to production levels. PMID:23751029

  16. Development of image quality assurance measures of the ExacTrac localization system using commercially available image evaluation software and hardware for image-guided radiotherapy.

    PubMed

    Stanley, Dennis N; Papanikolaou, Nikos; Gutierrez, Alonso N

    2014-11-01

    Quality assurance (QA) of the image quality for image-guided localization systems is crucial to ensure accurate visualization and localization of target volumes. In this study, a methodology was developed to assess and evaluate the constancy of the high-contrast spatial resolution, dose, energy, contrast, and geometrical accuracy of the BrainLAB ExacTrac system. An in-house fixation device was constructed to hold the QCkV-1 phantom firmly and reproducibly against the face of the flat panel detectors. Two image sets per detector were acquired using ExacTrac preset console settings over a period of three months. The image sets were analyzed in PIPSpro and the following metrics were recorded: high-contrast spatial resolution (f30,f40,f50 (lp/mm)), noise, and contrast-to-noise ratio. Geometrical image accuracy was evaluated by assessing the length between to predetermined points of the QCkV-1 phantom. Dose and kVp were recorded using the Unfors RaySafe Xi R/F Detector. The kVp and dose were evaluated for the following: Cranial Standard (CS) (80 kV,80 mA,80 ms), Thorax Standard (TS) (120 kV,160 mA,160 ms), Abdomen Standard (AS) (120 kV,160 mA,130 ms), and Pelvis Standard (PS) (120 kV,160 mA,160 ms). With regard to high-contrast spatial resolution, the mean values of the f30 (lp/mm), f40 (lp/mm) and f50 (lp/mm) for the left detector were 1.39±0.04,1.24±0.05, and 1.09±0.04, respectively, while for the right detector they were 1.38±0.04,1.22±0.05, and 1.09±0.05, respectively. Mean CNRs for the left and right detectors were 148±3 and 143±4, respectively. For geometrical accuracy, both detectors had a measured image length of the QCkV-1 of 57.9±0.5mm. The left detector showed dose measurements of 20.4±0.2μGy(CS), 191.8±0.7μGy(TS), 154.2±0.7μGy(AS), and 192.2±0.6μGy(PS), while the right detector showed 20.3±0.3μGy(CS), 189.7±0.8μGy(TS), 151.0±0.7μGy(AS), and 189.7±0.8μGy(PS), respectively. For X-ray energy, the left detector (right X-ray tube) had

  17. Development of image quality assurance measures of the ExacTrac localization system using commercially available image evaluation software and hardware for image-guided radiotherapy.

    PubMed

    Stanley, Dennis N; Papanikolaou, Nikos; Gutiérrez, Alonso N

    2014-11-08

    Quality assurance (QA) of the image quality for image-guided localization systems is crucial to ensure accurate visualization and localization of target volumes. In this study, a methodology was developed to assess and evaluate the constancy of the high-contrast spatial resolution, dose, energy, contrast, and geometrical accuracy of the BrainLAB ExacTrac system. An in-house fixation device was constructed to hold the QCkV-1 phantom firmly and reproducibly against the face of the flat panel detectors. Two image sets per detector were acquired using ExacTrac preset console settings over a period of three months. The image sets were analyzed in PIPSpro and the following metrics were recorded: high-contrast spatial resolution (f30, f40, f50 (lp/mm)), noise, and contrast-to-noise ratio. Geometrical image accu- racy was evaluated by assessing the length between to predetermined points of the QCkV-1 phantom. Dose and kVp were recorded using the Unfors RaySafe Xi R/F Detector. The kVp and dose were evaluated for the following: Cranial Standard (CS) (80 kV,80 mA,80 ms), Thorax Standard (TS) (120 kV,160 mA,160 ms), Abdomen Standard (AS) (120 kV,160 mA,130 ms), and Pelvis Standard (PS) (120 kV,160 mA,160 ms). With regard to high-contrast spatial resolution, the mean values of the f30 (lp/mm), f40 (lp/mm) and f50 (lp/mm) for the left detector were 1.39 ± 0.04, 1.24 ± 0.05, and 1.09 ± 0.04, respectively, while for the right detector they were 1.38 ± 0.04, 1.22 ± 0.05, and 1.09 ± 0.05, respectively. Mean CNRs for the left and right detectors were 148 ± 3 and 143 ± 4, respectively. For geometrical accuracy, both detectors had a measured image length of the QCkV-1 of 57.9 ± 0.5 mm. The left detector showed dose measurements of 20.4 ± 0.2 μGy (CS), 191.8 ± 0.7 μGy (TS), 154.2 ± 0.7 μGy (AS), and 192.2 ± 0.6 μGy (PS), while the right detector showed 20.3 ± 0.3 μGy (CS), 189.7 ± 0.8 μGy (TS), 151.0 ± 0.7 μGy (AS), and 189.7 ± 0.8 μGy (PS), respectively. For X

  18. Force adaptation transfers to untrained workspace regions in children: evidence for developing inverse dynamic motor models.

    PubMed

    Jansen-Osmann, Petra; Richter, Stefanie; Konczak, Jürgen; Kalveram, Karl-Theodor

    2002-03-01

    When humans perform goal-directed arm movements under the influence of an external damping force, they learn to adapt to these external dynamics. After removal of the external force field, they reveal kinematic aftereffects that are indicative of a neural controller that still compensates the no longer existing force. Such behavior suggests that the adult human nervous system uses a neural representation of inverse arm dynamics to control upper-extremity motion. Central to the notion of an inverse dynamic model (IDM) is that learning generalizes. Consequently, aftereffects should be observable even in untrained workspace regions. Adults have shown such behavior, but the ontogenetic development of this process remains unclear. This study examines the adaptive behavior of children and investigates whether learning a force field in one hemifield of the right arm workspace has an effect on force adaptation in the other hemifield. Thirty children (aged 6-10 years) and ten adults performed 30 degrees elbow flexion movements under two conditions of external damping (negative and null). We found that learning to compensate an external damping force transferred to the opposite hemifield, which indicates that a model of the limb dynamics rather than an association of visited space and experienced force was acquired. Aftereffects were more pronounced in the younger children and readaptation to a null-force condition was prolonged. This finding is consistent with the view that IDMs in children are imprecise neural representations of the actual arm dynamics. It indicates that the acquisition of IDMs is a developmental achievement and that the human motor system is inherently flexible enough to adapt to any novel force within the limits of the organism's biomechanics.

  19. The Role of A Facilitated Online Workspace Component of A Community of Practice: Knowledge Building and Value Creation for NASA

    NASA Astrophysics Data System (ADS)

    Davey, B.

    2014-12-01

    This study examined the role of an online workspace component of a community in the work of a community of practice. Much has been studied revealing the importance of communities of practice to organizations, project success, and knowledge management and some of these same successes hold true for virtual communities of practice. Study participants were 75 Education and Public Outreach community members of NASA's Science Mission Directorate Earth Forum. In this mixed methods study, online workspace metrics were used to track participation and a survey completed by 21 members was used to quantify participation. For a more detailed analysis, 15 community members (5 highly active users, 5 average users, and 5 infrequent users) selected based on survey responses, were interviewed. Finally, survey data was gathered from 7 online facilitators to understand their role in the community. Data collected from these 21 community members and 5 facilitating members suggest that highly active users (logging into the workspace daily), were more likely to have transformative experiences, co-create knowledge, feel ownership of community knowledge, have extended opportunities for community exchange, and find new forms of evaluation. Average users shared some similar characteristics with both the highly active members and infrequent users, representing a group in transition as they become more engaged and active in the online workspace. Inactive users viewed the workspace as having little value, being difficult to navigate, being mainly for gaining basic information about events and community news, and as another demand on their time. Results show the online workspace component of the Earth Science Education and Outreach Forum is playing an important and emerging role for this community by supporting knowledge building and knowledge sharing, and growing in value for those that utilizing it more frequently. The evidence suggests that with increased participation or "usage" comes

  20. Challenges to UK community pharmacy: a bio-photographic study of workspace in relation to professional pharmacy practice.

    PubMed

    Rapport, F L; Doel, M A; Jerzembek, G S

    2009-12-01

    This paper presents a novel, qualitative, bio-photographic study with intertextual analysis highlighting the relationship between community pharmacy workspace and practice. Sixteen pharmacists working across pharmacy types such as independent shops, large and small pharmacy chains and multiple pharmacies such as those in supermarkets participated in data capture and feedback consultation. Findings disclosed workspaces unfit for purpose and a workforce ill at ease with their new professional identity, involving increasingly complex tasks in health provision and retail. There was conflict between delegating to others and taking personal responsibility, and there were pressures from a demanding public within the context of a target-driven, litigious society. The study highlights that innovative, mixed methods in this context reveal nuanced, rich data.

  1. CrossTalk: The Journal of Defense Software Engineering. Volume 21, Number 7

    DTIC Science & Technology

    2008-07-01

    widespread fear among buyers that there might be mal- ware in their new software, for example, would depress sales and tarnish a brand. One only need recall...root zone servers. These DDOS attacks attempt to overwhelm servers with vast num- bers of messages. The use of bot - nets has increased the...Mail With Attachments Single and Group Chats Shared Workspaces Audio Conferencing Video Conferencing Centralized IT (DNS, DHCP) Centralized Backup and

  2. Adapting astronomical source detection software to help detect animals in thermal images obtained by unmanned aerial systems

    NASA Astrophysics Data System (ADS)

    Longmore, S. N.; Collins, R. P.; Pfeifer, S.; Fox, S. E.; Mulero-Pazmany, M.; Bezombes, F.; Goodwind, A.; de Juan Ovelar, M.; Knapen, J. H.; Wich, S. A.

    2017-02-01

    In this paper we describe an unmanned aerial system equipped with a thermal-infrared camera and software pipeline that we have developed to monitor animal populations for conservation purposes. Taking a multi-disciplinary approach to tackle this problem, we use freely available astronomical source detection software and the associated expertise of astronomers, to efficiently and reliably detect humans and animals in aerial thermal-infrared footage. Combining this astronomical detection software with existing machine learning algorithms into a single, automated, end-to-end pipeline, we test the software using aerial video footage taken in a controlled, field-like environment. We demonstrate that the pipeline works reliably and describe how it can be used to estimate the completeness of different observational datasets to objects of a given type as a function of height, observing conditions etc. - a crucial step in converting video footage to scientifically useful information such as the spatial distribution and density of different animal species. Finally, having demonstrated the potential utility of the system, we describe the steps we are taking to adapt the system for work in the field, in particular systematic monitoring of endangered species at National Parks around the world.

  3. Onboard utilization of ground control points for image correction. Volume 3: Ground control point simulation software design

    NASA Technical Reports Server (NTRS)

    1981-01-01

    The software developed to simulate the ground control point navigation system is described. The Ground Control Point Simulation Program (GCPSIM) is designed as an analysis tool to predict the performance of the navigation system. The system consists of two star trackers, a global positioning system receiver, a gyro package, and a landmark tracker.

  4. A novel scanning system using an industrial robot and the workspace measurement and positioning system

    NASA Astrophysics Data System (ADS)

    Zhao, Ziyue; Zhu, Jigui; Yang, Linghui; Lin, Jiarui

    2015-10-01

    The present scanning system consists of an industrial robot and a line-structured laser sensor which uses the industrial robot as a position instrument to guarantee the accuracy. However, the absolute accuracy of an industrial robot is relatively poor compared with the good repeatability in the manufacturing industry. This paper proposes a novel method using the workspace measurement and positioning system (wMPS) to remedy the lack of accuracy of the industrial robot. In order to guarantee the positioning accuracy of the system, the wMPS which is a laser-based measurement technology designed for large-volume metrology applications is brought in. Benefitting from the wMPS, this system can measure different cell-areas by the line-structured laser sensor and fuse the measurement data of different cell-areas by using the wMPS accurately. The system calibration which is the procedure to acquire and optimize the structure parameters of the scanning system is also stated in detail in this paper. In order to verify the feasibility of the system for scanning the large free-form surface, an experiment is designed to scan the internal surface of the door of a car-body in white. The final results show that the measurement data of the whole measuring areas have been jointed perfectly and there is no mismatch in the figure especially in the hole measuring areas. This experiment has verified the rationality of the system scheme, the correctness and effectiveness of the relevant methods.

  5. Effect of Tendon Vibration on Hemiparetic Arm Stability in Unstable Workspaces.

    PubMed

    Conrad, Megan O; Gadhoke, Bani; Scheidt, Robert A; Schmit, Brian D

    2015-01-01

    Sensory stimulation of wrist musculature can enhance stability in the proximal arm and may be a useful therapy aimed at improving arm control post-stroke. Specifically, our prior research indicates tendon vibration can enhance stability during point-to-point arm movements and in tracking tasks. The goal of the present study was to investigate the influence of forearm tendon vibration on endpoint stability, measured at the hand, immediately following forward arm movements in an unstable environment. Both proximal and distal workspaces were tested. Ten hemiparetic stroke subjects and 5 healthy controls made forward arm movements while grasping the handle of a two-joint robotic arm. At the end of each movement, the robot applied destabilizing forces. During some trials, 70 Hz vibration was applied to the forearm flexor muscle tendons. 70 Hz was used as the stimulus frequency as it lies within the range of optimal frequencies that activate the muscle spindles at the highest response rate. Endpoint position, velocity, muscle activity and grip force data were compared before, during and after vibration. Stability at the endpoint was quantified as the magnitude of oscillation about the target position, calculated from the power of the tangential velocity data. Prior to vibration, subjects produced unstable, oscillating hand movements about the target location due to the applied force field. Stability increased during vibration, as evidenced by decreased oscillation in hand tangential velocity.

  6. A Multi-Sensorial Hybrid Control for Robotic Manipulation in Human-Robot Workspaces

    PubMed Central

    Pomares, Jorge; Perea, Ivan; García, Gabriel J.; Jara, Carlos A.; Corrales, Juan A.; Torres, Fernando

    2011-01-01

    Autonomous manipulation in semi-structured environments where human operators can interact is an increasingly common task in robotic applications. This paper describes an intelligent multi-sensorial approach that solves this issue by providing a multi-robotic platform with a high degree of autonomy and the capability to perform complex tasks. The proposed sensorial system is composed of a hybrid visual servo control to efficiently guide the robot towards the object to be manipulated, an inertial motion capture system and an indoor localization system to avoid possible collisions between human operators and robots working in the same workspace, and a tactile sensor algorithm to correctly manipulate the object. The proposed controller employs the whole multi-sensorial system and combines the measurements of each one of the used sensors during two different phases considered in the robot task: a first phase where the robot approaches the object to be grasped, and a second phase of manipulation of the object. In both phases, the unexpected presence of humans is taken into account. This paper also presents the successful results obtained in several experimental setups which verify the validity of the proposed approach. PMID:22163729

  7. Effect of Tendon Vibration on Hemiparetic Arm Stability in Unstable Workspaces

    PubMed Central

    Conrad, Megan O.; Gadhoke, Bani; Scheidt, Robert A.; Schmit, Brian D.

    2015-01-01

    Sensory stimulation of wrist musculature can enhance stability in the proximal arm and may be a useful therapy aimed at improving arm control post-stroke. Specifically, our prior research indicates tendon vibration can enhance stability during point-to-point arm movements and in tracking tasks. The goal of the present study was to investigate the influence of forearm tendon vibration on endpoint stability, measured at the hand, immediately following forward arm movements in an unstable environment. Both proximal and distal workspaces were tested. Ten hemiparetic stroke subjects and 5 healthy controls made forward arm movements while grasping the handle of a two-joint robotic arm. At the end of each movement, the robot applied destabilizing forces. During some trials, 70 Hz vibration was applied to the forearm flexor muscle tendons. 70 Hz was used as the stimulus frequency as it lies within the range of optimal frequencies that activate the muscle spindles at the highest response rate. Endpoint position, velocity, muscle activity and grip force data were compared before, during and after vibration. Stability at the endpoint was quantified as the magnitude of oscillation about the target position, calculated from the power of the tangential velocity data. Prior to vibration, subjects produced unstable, oscillating hand movements about the target location due to the applied force field. Stability increased during vibration, as evidenced by decreased oscillation in hand tangential velocity. PMID:26633892

  8. Human precision manipulation workspace: Effects of object size and number of fingers used.

    PubMed

    Bullock, Ian M; Feix, Thomas; Dollar, Aaron M

    2015-08-01

    Precision manipulation, or moving small objects in the fingertips, is important for daily tasks such as writing and key insertion, as well as medically relevant tasks such as scalpel cuts and surgical teleoperation. While fingertip force coordination has been studied in some detail, few previous works have experimentally studied the kinematics of human precision manipulation with real objects. The present work focuses on studying the effects of varying object size and the number of fingers used on the resulting manipulation workspace, or range of motions that the object can be moved through. To study object size effects, seven bar-shaped objects ranging from 20 to 80 mm length were tested; after scaling object length to the equivalent for a 17.5 cm hand, the peak volume was obtained for 48-59 mm object length range (23% above average), and the minimum volume was obtained for the smallest 17-27 mm range (72% of average). 50 mm and 80 mm circular objects were used to study the effect of using different numbers of fingers; the five-finger manipulation volume dropped to less than half the two-finger volume (p<;0.001). We anticipate these results will be useful in designing devices such as hand held tools, as well as in designing protocols for effectively testing and rehabilitating hand function. Finally, the results can provide a benchmark for the manipulation capability of prosthetic hands.

  9. Osteolytica: An automated image analysis software package that rapidly measures cancer-induced osteolytic lesions in in vivo models with greater reproducibility compared to other commonly used methods☆

    PubMed Central

    Evans, H.R.; Karmakharm, T.; Lawson, M.A.; Walker, R.E.; Harris, W.; Fellows, C.; Huggins, I.D.; Richmond, P.; Chantry, A.D.

    2016-01-01

    Methods currently used to analyse osteolytic lesions caused by malignancies such as multiple myeloma and metastatic breast cancer vary from basic 2-D X-ray analysis to 2-D images of micro-CT datasets analysed with non-specialised image software such as ImageJ. However, these methods have significant limitations. They do not capture 3-D data, they are time-consuming and they often suffer from inter-user variability. We therefore sought to develop a rapid and reproducible method to analyse 3-D osteolytic lesions in mice with cancer-induced bone disease. To this end, we have developed Osteolytica, an image analysis software method featuring an easy to use, step-by-step interface to measure lytic bone lesions. Osteolytica utilises novel graphics card acceleration (parallel computing) and 3-D rendering to provide rapid reconstruction and analysis of osteolytic lesions. To evaluate the use of Osteolytica we analysed tibial micro-CT datasets from murine models of cancer-induced bone disease and compared the results to those obtained using a standard ImageJ analysis method. Firstly, to assess inter-user variability we deployed four independent researchers to analyse tibial datasets from the U266-NSG murine model of myeloma. Using ImageJ, inter-user variability between the bones was substantial (± 19.6%), in contrast to using Osteolytica, which demonstrated minimal variability (± 0.5%). Secondly, tibial datasets from U266-bearing NSG mice or BALB/c mice injected with the metastatic breast cancer cell line 4T1 were compared to tibial datasets from aged and sex-matched non-tumour control mice. Analyses by both Osteolytica and ImageJ showed significant increases in bone lesion area in tumour-bearing mice compared to control mice. These results confirm that Osteolytica performs as well as the current 2-D ImageJ osteolytic lesion analysis method. However, Osteolytica is advantageous in that it analyses over the entirety of the bone volume (as opposed to selected 2-D images

  10. TIA Software User's Manual

    NASA Technical Reports Server (NTRS)

    Cramer, K. Elliott; Syed, Hazari I.

    1995-01-01

    This user's manual describes the installation and operation of TIA, the Thermal-Imaging acquisition and processing Application, developed by the Nondestructive Evaluation Sciences Branch at NASA Langley Research Center, Hampton, Virginia. TIA is a user friendly graphical interface application for the Macintosh 2 and higher series computers. The software has been developed to interface with the Perceptics/Westinghouse Pixelpipe(TM) and PixelStore(TM) NuBus cards and the GW Instruments MacADIOS(TM) input-output (I/O) card for the Macintosh for imaging thermal data. The software is also capable of performing generic image-processing functions.

  11. Magnetic resonance image segmentation using semi-automated software for quantification of knee articular cartilage—initial evaluation of a technique for paired scans

    PubMed Central

    Brem, M. H.; Lang, P. K.; Neumann, G.; Schlechtweg, P. M.; Schneider, E.; Jackson, R.; Yu, J.; Eaton, C. B.; Hennig, F. F.; Yoshioka, H.; Pappas, G.; Duryea, J.

    2010-01-01

    Purpose Software-based image analysis is important for studies of cartilage changes in knee osteoarthritis (OA). This study describes an evaluation of a semi-automated cartilage segmentation software tool capable of quantifying paired images for potential use in longitudinal studies of knee OA. We describe the methodology behind the analysis and demonstrate its use by determination of test–retest analysis precision of duplicate knee magnetic resonance imaging (MRI) data sets. Methods Test–retest knee MR images of 12 subjects with a range of knee health were evaluated from the Osteoarthritis Initiative (OAI) pilot MR study. Each subject was removed from the magnet between the two scans. The 3D DESS (sagittal, 0.456 mm×0.365 mm, 0.7 mm slice thickness, TR 16.5 ms, TE 4.7 ms) images were obtained on a 3-T Siemens Trio MR system with a USA Instruments quadrature transmit–receive extremity coil. Segmentation of one 3D-image series was first performed and then the corresponding retest series was segmented by viewing both image series concurrently in two adjacent windows. After manual registration of the series, the first segmentation cartilage outline served as an initial estimate for the second segmentation. We evaluated morphometric measures of the bone and cartilage surface area (tAB and AC), cartilage volume (VC), and mean thickness (ThC.me) for medial/lateral tibia (MT/LT), total femur (F) and patella (P). Test–retest reproducibility was assessed using the root-mean square coefficient of variation (RMS CV%). Results For the paired analyses, RMS CV % ranged from 0.9% to 1.2% for VC, from 0.3% to 0.7% for AC, from 0.6% to 2.7% for tAB and 0.8% to 1.5% for ThC.me. Conclusion Paired image analysis improved the measurement precision of cartilage segmentation. Our results are in agreement with other publications supporting the use of paired analysis for longitudinal studies of knee OA. PMID:19252907

  12. Software for hyperspectral, joint photographic experts group (.JPG), portable network graphics (.PNG) and tagged image file format (.TIFF) segmentation

    NASA Astrophysics Data System (ADS)

    Bruno, L. S.; Rodrigo, B. P.; Lucio, A. de C. Jorge

    2016-10-01

    This paper presents a system developed by an application of a neural network Multilayer Perceptron for drone acquired agricultural image segmentation. This application allows a supervised user training the classes that will posteriorly be interpreted by neural network. These classes will be generated manually with pre-selected attributes in the application. After the attribute selection a segmentation process is made to allow the relevant information extraction for different types of images, RGB or Hyperspectral. The application allows extracting the geographical coordinates from the image metadata, geo referencing all pixels on the image. In spite of excessive memory consume on hyperspectral images regions of interest, is possible to perform segmentation, using bands chosen by user that can be combined in different ways to obtain different results.

  13. TU-PIS-Exhibit Hall-2: Deformable Image Registration, Contour Propagation, and Dose Mapping inMIM Maestro - MIM Software

    SciTech Connect

    Piper, J.

    2015-06-15

    Brachytherapy devices and software are designed to last for a certain period of time. Due to a number of considerations, such as material factors, wear-and-tear, backwards compatibility, and others, they all reach a date when they are no longer supported by the manufacturer. Most of these products have a limited duration for their use, and the information is provided to the user at time of purchase. Because of issues or concerns determined by the manufacturer, certain products are retired sooner than the anticipated date, and the user is immediately notified. In these situations, the institution is facing some difficult choices: remove these products from the clinic or perform tests and continue their usage. Both of these choices come with a financial burden: replacing the product or assuming a potential medicolegal liability. This session will provide attendees with the knowledge and tools to make better decisions when facing these issues. Learning Objectives: Understand the meaning of “end-of-life or “life expectancy” for brachytherapy devices and software Review items (devices and software) affected by “end-of-life” restrictions Learn how to effectively formulate “end-of-life” policies at your institution Learn about possible implications of “end-of-life” policy Review other possible approaches to “end-of-life” issue.

  14. ChiMS: Open-source instrument control software platform on LabVIEW for imaging/depth profiling mass spectrometers.

    PubMed

    Cui, Yang; Hanley, Luke

    2015-06-01

    ChiMS is an open-source data acquisition and control software program written within LabVIEW for high speed imaging and depth profiling mass spectrometers. ChiMS can also transfer large datasets from a digitizer to computer memory at high repetition rate, save data to hard disk at high throughput, and perform high speed data processing. The data acquisition mode generally simulates a digital oscilloscope, but with peripheral devices integrated for control as well as advanced data sorting and processing capabilities. Customized user-designed experiments can be easily written based on several included templates. ChiMS is additionally well suited to non-laser based mass spectrometers imaging and various other experiments in laser physics, physical chemistry, and surface science.

  15. Facial recognition software success rates for the identification of 3D surface reconstructed facial images: implications for patient privacy and security.

    PubMed

    Mazura, Jan C; Juluru, Krishna; Chen, Joseph J; Morgan, Tara A; John, Majnu; Siegel, Eliot L

    2012-06-01

    Image de-identification has focused on the removal of textual protected health information (PHI). Surface reconstructions of the face have the potential to reveal a subject's identity even when textual PHI is absent. This study assessed the ability of a computer application to match research subjects' 3D facial reconstructions with conventional photographs of their face. In a prospective study, 29 subjects underwent CT scans of the head and had frontal digital photographs of their face taken. Facial reconstructions of each CT dataset were generated on a 3D workstation. In phase 1, photographs of the 29 subjects undergoing CT scans were added to a digital directory and tested for recognition using facial recognition software. In phases 2-4, additional photographs were added in groups of 50 to increase the pool of possible matches and the test for recognition was repeated. As an internal control, photographs of all subjects were tested for recognition against an identical photograph. Of 3D reconstructions, 27.5% were matched correctly to corresponding photographs (95% upper CL, 40.1%). All study subject photographs were matched correctly to identical photographs (95% lower CL, 88.6%). Of 3D reconstructions, 96.6% were recognized simply as a face by the software (95% lower CL, 83.5%). Facial recognition software has the potential to recognize features on 3D CT surface reconstructions and match these with photographs, with implications for PHI.

  16. Hardware, software, and scanning issues encountered during small animal imaging of photodynamic therapy in the athymic nude rat

    NASA Astrophysics Data System (ADS)

    Cross, Nathan; Sharma, Rahul; Varghai, Davood; Spring-Robinson, Chandra; Oleinick, Nancy L.; Muzic, Raymond F., Jr.; Dean, David

    2007-02-01

    Small animal imaging devices are now commonly used to study gene activation and model the effects of potential therapies. We are attempting to develop a protocol that non-invasively tracks the affect of Pc 4-mediated photodynamic therapy (PDT) in a human glioma model using structural image data from micro-CT and/or micro-MR scanning and functional data from 18F-fluorodeoxy-glucose (18F-FDG) micro-PET imaging. Methods: Athymic nude rat U87-derived glioma was imaged by micro-PET and either micro-CT or micro-MR prior to Pc 4-PDT. Difficulty insuring animal anesthesia and anatomic position during the micro-PET, micro-CT, and micro-MR scans required adaptation of the scanning bed hardware. Following Pc 4-PDT the animals were again 18F-FDG micro-PET scanned, euthanized one day later, and their brains were explanted and prepared for H&E histology. Histology provided the gold standard for tumor location and necrosis. The tumor and surrounding brain functional and structural image data were then isolated and coregistered. Results: Surprisingly, both the non-PDT and PDT groups showed an increase in tumor functional activity when we expected this signal to disappear in the group receiving PDT. Co-registration of the functional and structural image data was done manually. Discussion: As expected, micro-MR imaging provided better structural discrimination of the brain tumor than micro-CT. Contrary to expectations, in our preliminary analysis 18F-FDG micro-PET imaging does not readily discriminate the U87 tumors that received Pc 4-PDT. We continue to investigate the utility of micro-PET and other methods of functional imaging to remotely detect the specificity and sensitivity of Pc 4-PDT in deeply placed tumors.

  17. Software Program: Software Management Guidebook

    NASA Technical Reports Server (NTRS)

    1996-01-01

    The purpose of this NASA Software Management Guidebook is twofold. First, this document defines the core products and activities required of NASA software projects. It defines life-cycle models and activity-related methods but acknowledges that no single life-cycle model is appropriate for all NASA software projects. It also acknowledges that the appropriate method for accomplishing a required activity depends on characteristics of the software project. Second, this guidebook provides specific guidance to software project managers and team leaders in selecting appropriate life cycles and methods to develop a tailored plan for a software engineering project.

  18. A New Measurement Technique of the Characteristics of Nutrient Artery Canals in Tibias Using Materialise's Interactive Medical Image Control System Software

    PubMed Central

    Li, Jiantao; Zhang, Hao; Yin, Peng; Su, Xiuyun; Zhao, Zhe; Zhou, Jianfeng; Li, Chen; Li, Zhirui; Zhang, Lihai; Tang, Peifu

    2015-01-01

    We established a novel measurement technique to evaluate the anatomic information of nutrient artery canals using Mimics (Materialise's Interactive Medical Image Control System) software, which will provide full knowledge of nutrient artery canals to assist in the diagnosis of longitudinal fractures of tibia and choosing an optimal therapy. Here we collected Digital Imaging and Communications in Medicine (DICOM) format of 199 patients hospitalized in our hospital. All three-dimensional models of tibia in Mimics were reconstructed. In 3-matic software, we marked five points in tibia which located at intercondylar eminence, tibia tuberosity, outer ostium, inner ostium, and bottom of medial malleolus. We then recorded Z-coordinates values of the five points and performed statistical analysis. Our results indicate that foramen was found to be absent in 9 (2.3%) tibias, and 379 (95.2%) tibias had single nutrient foramen. The double foramina was observed in 10 (2.5%) tibias. The mean of tibia length was 358 ± 22 mm. The mean foraminal index was 31.8%  ± 3%. The mean distance between tibial tuberosity and foramen (TFD) is 66 ± 12 mm. Foraminal index has significant positive correlation with TFD (r = 0.721, P < 0.01). Length of nutrient artery canals has significant negative correlation with TFD (r = −0.340, P < 0.01) and has significant negative correlation with foraminal index (r = −0.541, P < 0.01). PMID:26788498

  19. Comparison of performance of object-based image analysis techniques available in open source software (Spring and Orfeo Toolbox/Monteverdi) considering very high spatial resolution data

    NASA Astrophysics Data System (ADS)

    Teodoro, Ana C.; Araujo, Ricardo

    2016-01-01

    The use of unmanned aerial vehicles (UAVs) for remote sensing applications is becoming more frequent. However, this type of information can result in several software problems related to the huge amount of data available. Object-based image analysis (OBIA) has proven to be superior to pixel-based analysis for very high-resolution images. The main objective of this work was to explore the potentialities of the OBIA methods available in two different open source software applications, Spring and OTB/Monteverdi, in order to generate an urban land cover map. An orthomosaic derived from UAVs was considered, 10 different regions of interest were selected, and two different approaches were followed. The first one (Spring) uses the region growing segmentation algorithm followed by the Bhattacharya classifier. The second approach (OTB/Monteverdi) uses the mean shift segmentation algorithm followed by the support vector machine (SVM) classifier. Two strategies were followed: four classes were considered using Spring and thereafter seven classes were considered for OTB/Monteverdi. The SVM classifier produces slightly better results and presents a shorter processing time. However, the poor spectral resolution of the data (only RGB bands) is an important factor that limits the performance of the classifiers applied.

  20. A New Measurement Technique of the Characteristics of Nutrient Artery Canals in Tibias Using Materialise's Interactive Medical Image Control System Software.

    PubMed

    Li, Jiantao; Zhang, Hao; Yin, Peng; Su, Xiuyun; Zhao, Zhe; Zhou, Jianfeng; Li, Chen; Li, Zhirui; Zhang, Lihai; Tang, Peifu

    2015-01-01

    We established a novel measurement technique to evaluate the anatomic information of nutrient artery canals using Mimics (Materialise's Interactive Medical Image Control System) software, which will provide full knowledge of nutrient artery canals to assist in the diagnosis of longitudinal fractures of tibia and choosing an optimal therapy. Here we collected Digital Imaging and Communications in Medicine (DICOM) format of 199 patients hospitalized in our hospital. All three-dimensional models of tibia in Mimics were reconstructed. In 3-matic software, we marked five points in tibia which located at intercondylar eminence, tibia tuberosity, outer ostium, inner ostium, and bottom of medial malleolus. We then recorded Z-coordinates values of the five points and performed statistical analysis. Our results indicate that foramen was found to be absent in 9 (2.3%) tibias, and 379 (95.2%) tibias had single nutrient foramen. The double foramina was observed in 10 (2.5%) tibias. The mean of tibia length was 358 ± 22 mm. The mean foraminal index was 31.8%  ± 3%. The mean distance between tibial tuberosity and foramen (TFD) is 66 ± 12 mm. Foraminal index has significant positive correlation with TFD (r = 0.721, P < 0.01). Length of nutrient artery canals has significant negative correlation with TFD (r = -0.340, P < 0.01) and has significant negative correlation with foraminal index (r = -0.541, P < 0.01).

  1. Proprietary software

    NASA Technical Reports Server (NTRS)

    Marnock, M. J.

    1971-01-01

    The protection of intellectual property by a patent, a copyright, or trade secrets is reviewed. The present and future use of computers and software are discussed, along with the governmental uses of software. The popularity of contractual agreements for sale or lease of computer programs and software services is also summarized.

  2. Force Sensor-less Workspace Virtual Impedance Control Considering Resonant Vibration for Industrial Robot

    NASA Astrophysics Data System (ADS)

    Tungpataratanawong, Somsawas; Ohishi, Kiyoshi; Miyazaki, Toshimasa; Katsura, Seiichiro

    The motion control paradigm provides sufficient performance in many elementary industrial tasks. However, only stiff motion the robot cannot accommodate the interaction force under constrained motion. In such situation, the robot is required to perform interaction behavior with the environment. The conventional impedance control schemes require force-sensing devices to feedback force signals to the controllers. The force-sensing device is therefore indispensable and the performance of the system also depends on the quality of this device. This paper proposes a novel strategy for force sensor-less impedance control using disturbance observer and dynamic model of the robot to estimate the external force. In motion task, the robust D-PD (derivative-PD) control is used with feedforward inverse-dynamic torque compensation to ensure robustness and high-speed response with flexible joint model. When robot is in contact with environment, the proposed force sensor-less scheme impedance control with inner-loop D-PD control is utilized. D-PD control uses both position and speed as the references to implement the damping and stiffness characteristic of the virtual impedance model. In addition, the gravity and friction force-feedback compensation is computed by the same dynamic model, which is used in external force estimation. The flexible-joint robot model is utilized in both disturbance observer and motion control design. The workspace impedance control for robot interaction with human operator is implemented on the experimental setup three-degree-of-freedom (3-DOF) robot manipulator to assure the ability and performance of the proposed force sensor-less scheme for flexible-joint industrial robot.

  3. Workspace location influences joint coordination during reaching in post-stroke hemiparesis.

    PubMed

    Reisman, Darcy S; Scholz, John P

    2006-04-01

    The purpose of this study was to determine the influence of workspace location on joint coordination in persons with post-stroke hemiparesis when trunk motion was required to complete reaches beyond the arm's functional reach length. Seven subjects with mild right hemiparesis following a stroke and seven age and gender matched control subjects participated. Joint motions and characteristics of hand and trunk movement were measured over multiple repetitions. The variance (across trials) of joint combinations was partitioned into two components at every point in the hand's trajectory using the uncontrolled manifold approach; the first component is a measure of the extent to which equivalent joint combinations are used to control a given hand path, and reflects performance flexibility. The second component of joint variance reflects the use of non-equivalent joint combinations, which lead to hand path error. Compared to the control subjects, persons with hemiparesis demonstrated a significantly greater amount of non-equivalent joint variability related to control of the hand's path and of the hand's position relative to the trunk when reaching toward the hemiparetic side (ipsilaterally), but not when reaching to the less involved side. The relative timing of the hand and trunk was also altered when reaching ipsilaterally. The current findings support the idea that the previously proposed "arm compensatory synergy" may be deficient in subjects with hemiparesis. This deficiency may be due to one or a combination of factors: changes in central commands that are thought to set the gain of the arm compensatory synergy; a limited ability to combine shoulder abduction and elbow extension that limits the expression of an appropriately set arm compensatory synergy; or a reduction of the necessary degrees-of-freedom needed to adequately compensate for poor trunk control when reaching ipsilaterally.

  4. Towards a software profession

    NASA Technical Reports Server (NTRS)

    Berard, Edward V.

    1986-01-01

    An increasing number of programmers have attempted to change their image. They have made it plain that they wish not only to be taken seriously, but they also wish to be regarded as professionals. Many programmers now wish to referred to as software engineers. If programmers wish to be considered professionals in every sense of the word, two obstacles must be overcome: the inability to think of software as a product, and the idea that little or no skill is required to create and handle software throughout its life cycle. The steps to be taken toward professionalization are outlined along with recommendations.

  5. A Guide to Selecting Large Print/Enhanced Image Computer Access Hardware/Software for Persons with Low Vision.

    ERIC Educational Resources Information Center

    DeWitt, J. C.; And Others

    1988-01-01

    Large print/enhanced image computer access products for low-vision individuals are reviewed. The products ("DP-11/Plus,""Large Print DOS,""Lyon Large Print,""PC Lens,""Vista/Vista 2," and "ZoomText") work interactively with most application programs and DOS in IBM environments. Features…

  6. Revealing text in a complexly rolled silver scroll from Jerash with computed tomography and advanced imaging software

    NASA Astrophysics Data System (ADS)

    Hoffmann Barfod, Gry; Larsen, John Møller; Lichtenberger, Achim; Raja, Rubina

    2015-12-01

    Throughout Antiquity magical amulets written on papyri, lead and silver were used for apotropaic reasons. While papyri often can be unrolled and deciphered, metal scrolls, usually very thin and tightly rolled up, cannot easily be unrolled without damaging the metal. This leaves us with unreadable results due to the damage done or with the decision not to unroll the scroll. The texts vary greatly and tell us about the cultural environment and local as well as individual practices at a variety of locations across the Mediterranean. Here we present the methodology and the results of the digital unfolding of a silver sheet from Jerash in Jordan from the mid-8th century CE. The scroll was inscribed with 17 lines in presumed pseudo-Arabic as well as some magical signs. The successful unfolding shows that it is possible to digitally unfold complexly folded scrolls, but that it requires a combination of the know-how of the software and linguistic knowledge.

  7. Design principles for innovative workspaces to increase efficiency in pharmaceutical R&D: lessons learned from the Novartis campus.

    PubMed

    Zoller, Frank A; Boutellier, Roman

    2013-04-01

    When managing R&D departments for increased efficiency and effectiveness the focus has often been on organizational structure. Space is, however, of outstanding importance in an environment of large task uncertainty, which is the case in pharmaceutical R&D. Based on case studies about the Novartis campus in Basel, Switzerland, we propose some design principles for laboratory and office workspace to support the strong and weak ties of scientist networks. We address the diversity of technologies and specialization, as well as the pressure on time-to-market, as major challenges in pharmaceutical R&D.

  8. Software safety

    NASA Technical Reports Server (NTRS)

    Leveson, Nancy

    1987-01-01

    Software safety and its relationship to other qualities are discussed. It is shown that standard reliability and fault tolerance techniques will not solve the safety problem for the present. A new attitude requires: looking at what you do NOT want software to do along with what you want it to do; and assuming things will go wrong. New procedures and changes to entire software development process are necessary: special software safety analysis techniques are needed; and design techniques, especially eliminating complexity, can be very helpful.

  9. Development of kinematic equations and determination of workspace of a 6 DOF end-effector with closed-kinematic chain mechanism

    NASA Technical Reports Server (NTRS)

    Nguyen, Charles C.; Pooran, Farhad J.

    1989-01-01

    This report presents results from the research grant entitled Active Control of Robot Manipulators, funded by the Goddard Space Flight Center, under Grant NAG5-780, for the period July 1, 1988 to January 1, 1989. An analysis is presented of a 6 degree-of-freedom robot end-effector built to study telerobotic assembly of NASA hardware in space. Since the end-effector is required to perform high precision motion in a limited workspace, closed-kinematic mechanisms are chosen for its design. A closed-form solution is obtained for the inverse kinematic problem and an iterative procedure employing Newton-Raphson method is proposed to solve the forward kinematic problem. A study of the end-effector workspace results in a general procedure for the workspace determination based on link constraints. Computer simulation results are presented.

  10. IDATEN and G-SITENNO: GUI-assisted software for coherent X-ray diffraction imaging experiments and data analyses at SACLA.

    PubMed

    Sekiguchi, Yuki; Yamamoto, Masaki; Oroguchi, Tomotaka; Takayama, Yuki; Suzuki, Shigeyuki; Nakasako, Masayoshi

    2014-11-01

    Using our custom-made diffraction apparatus KOTOBUKI-1 and two multiport CCD detectors, cryogenic coherent X-ray diffraction imaging experiments have been undertaken at the SPring-8 Angstrom Compact free electron LAser (SACLA) facility. To efficiently perform experiments and data processing, two software suites with user-friendly graphical user interfaces have been developed. The first is a program suite named IDATEN, which was developed to easily conduct four procedures during experiments: aligning KOTOBUKI-1, loading a flash-cooled sample into the cryogenic goniometer stage inside the vacuum chamber of KOTOBUKI-1, adjusting the sample position with respect to the X-ray beam using a pair of telescopes, and collecting diffraction data by raster scanning the sample with X-ray pulses. Named G-SITENNO, the other suite is an automated version of the original SITENNO suite, which was designed for processing diffraction data. These user-friendly software suites are now indispensable for collecting a large number of diffraction patterns and for processing the diffraction patterns immediately after collecting data within a limited beam time.

  11. Quantification of regions of interest in swath sonar backscatter images using grey-level and shape geometry descriptors: the TargAn software

    NASA Astrophysics Data System (ADS)

    Fakiris, Elias; Papatheodorou, George

    2012-06-01

    In this paper, the TargAn software package that is dedicated to parameterizing regions of interest (ROIs) in greyscale images that reflect backscatter information derived by marine geo-acoustical instrumentation (e.g. Side Scan Sonar and Multi-Beam Echo-Sounder) is presented. The ROIs, whose boundaries are marked out either manually or via simple segmentation techniques, are analyzed for as many as 37 features. The adopted and developed methodologies lead to the extraction of: (1) grey-level intensity (1st order) and texture analysis statistics estimated from the inner ROI, (2) descriptors that measure the separation of the ROI in comparison to the intensity characteristics of the peripheral seabed, (3) shape geometry descriptors of the ROI's boundary itself and (4) regional statistics of distinct (segmented) objects possibly included in the ROI. TargAn is implemented in Matlab with a graphical user interface that helps the user to have control over the digitization, segmentation and feature extraction processes involved. It also provides tools for the construction of compact geo-databases, suitable for geostatistical analysis and visualization in popular Geographical Information Systems, concerning the extracted descriptors and the geographic features (e.g. ROIs' boundaries, skeletons, segmented objects) that have been considered for ROIs' analyses. The TargAn software is particularly useful when large amounts of image ROIs need to be objectively quantified and is demonstrated through two case studies regarding Side Scan Sonar imageries. The first one concerns the quantification of marine biohabitats (coralline formations) while the second exhibits the geometrical analysis of pockmarks.

  12. Revealing text in a complexly rolled silver scroll from Jerash with computed tomography and advanced imaging software

    PubMed Central

    Hoffmann Barfod, Gry; Larsen, John Møller; Raja, Rubina

    2015-01-01

    Throughout Antiquity magical amulets written on papyri, lead and silver were used for apotropaic reasons. While papyri often can be unrolled and deciphered, metal scrolls, usually very thin and tightly rolled up, cannot easily be unrolled without damaging the metal. This leaves us with unreadable results due to the damage done or with the decision not to unroll the scroll. The texts vary greatly and tell us about the cultural environment and local as well as individual practices at a variety of locations across the Mediterranean. Here we present the methodology and the results of the digital unfolding of a silver sheet from Jerash in Jordan from the mid-8th century CE. The scroll was inscribed with 17 lines in presumed pseudo-Arabic as well as some magical signs. The successful unfolding shows that it is possible to digitally unfold complexly folded scrolls, but that it requires a combination of the know-how of the software and linguistic knowledge. PMID:26648504

  13. Multimodal evaluation of 2-D and 3-D ultrasound, computed tomography and magnetic resonance imaging in measurements of the thyroid volume using universally applicable cross-sectional imaging software: a phantom study.

    PubMed

    Freesmeyer, Martin; Wiegand, Steffen; Schierz, Jan-Henning; Winkens, Thomas; Licht, Katharina

    2014-07-01

    A precise estimate of thyroid volume is necessary for making adequate therapeutic decisions and planning, as well as for monitoring therapy response. The goal of this study was to compare the precision of different volumetry methods. Thyroid-shaped phantoms were subjected to volumetry via 2-D and 3-D ultrasonography (US), computed tomography (CT) and magnetic resonance imaging (MRI). The 3-D US scans were performed using sensor navigation and mechanical sweeping methods. Volumetry calculation ensued with the conventional ellipsoid model and the manual tracing method. The study confirmed the superiority of manual tracing with CT and MRI volumetry of the thyroid, but extended this knowledge also to the superiority of the 3-D US method, regardless of whether sensor navigation or mechanical sweeping is used. A novel aspect was successful use of the same universally applicable cross-imaging software for all modalities.

  14. Design, modeling, analysis and testing of a novel piezo-actuated XY compliant mechanism for large workspace nano-positioning

    NASA Astrophysics Data System (ADS)

    Zhu, Wu-Le; Zhu, Zhiwei; Shi, Yi; Wang, Xinwei; Guan, Kaimin; Ju, Bing-Feng

    2016-11-01

    In this paper, a new piezo-actuated XY parallel compliant mechanism for large workspace nano-positioning with decoupled motions is developed by incorporating a novel Z-shaped flexure hinge (ZFH)-based mechanism into the mirror-symmetrically distributed structure. The bridge-type mechanism and two-stage leverage mechanisms serve as preliminary displacement amplifiers, while further amplification with motion transfer and decoupled output motions are achieved by means of the ZFH mechanism. Based on finite element theory, a high-precision analytical model of the XY compliant mechanism is established by considering all the connecting linkages as flexible components. Through the improved differential evolution algorithm, the optimized compliant mechanism is capable of performing millimeter-scale workspace nano-positioning with decoupled motions. In addition, the input displacement unbalance, resulting from the lateral force which has potential to damage the piezoelectric actuators, is markedly lowered to a negligible value. The performance of the fabricated compliant mechanism with optimized parameters is investigated to well agree with both the analytical model and ANSYS simulation. In addition, based on the inverse kinematics derived from the model and experimental results, different elliptical vibration trajectories are accurately acquired.

  15. Software Bridge

    NASA Technical Reports Server (NTRS)

    1995-01-01

    I-Bridge is a commercial version of software developed by I-Kinetics under a NASA Small Business Innovation Research (SBIR) contract. The software allows users of Windows applications to gain quick, easy access to databases, programs and files on UNIX services. Information goes directly onto spreadsheets and other applications; users need not manually locate, transfer and convert data.

  16. Software Reviews.

    ERIC Educational Resources Information Center

    Wulfson, Stephen

    1988-01-01

    Presents reviews of six computer software programs for teaching science. Provides the publisher, grade level, cost, and descriptions of software, including: (1) "Recycling Logic"; (2) "Introduction to Biochemistry"; (3) "Food for Thought"; (4) "Watts in a Home"; (5) "Geology in Action"; and (6)…

  17. Software Reviews.

    ERIC Educational Resources Information Center

    Science and Children, 1988

    1988-01-01

    Reviews six software packages for the Apple II family. Programs reviewed include "Science Courseware: Earth Science Series"; "Heat and Light"; "In Search of Space: Introduction to Model Rocketry"; "Drug Education Series: Drugs--Their Effects on You'"; "Uncertainties and Measurement"; and "Software Films: Learning about Science Series," which…

  18. Software Reviews.

    ERIC Educational Resources Information Center

    Miller, Anne, Ed.; Radziemski, Cathy, Ed.

    1988-01-01

    Reviews two software packages for the Macintosh series. "Course Builder 2.0," a courseware authoring system, allows the user to create programs which stand alone and may be used independently in the classroom. "World Builder," an artificial intelligence software package, allows creative thinking, problem-solving, and…

  19. Application of Technical Measures and Software in Constructing Photorealistic 3D Models of Historical Building Using Ground-Based and Aerial (UAV) Digital Images

    NASA Astrophysics Data System (ADS)

    Zarnowski, Aleksander; Banaszek, Anna; Banaszek, Sebastian

    2015-12-01

    Preparing digital documentation of historical buildings is a form of protecting cultural heritage. Recently there have been several intensive studies using non-metric digital images to construct realistic 3D models of historical buildings. Increasingly often, non-metric digital images are obtained with unmanned aerial vehicles (UAV). Technologies and methods of UAV flights are quite different from traditional photogrammetric approaches. The lack of technical guidelines for using drones inhibits the process of implementing new methods of data acquisition. This paper presents the results of experiments in the use of digital images in the construction of photo-realistic 3D model of a historical building (Raphaelsohns' Sawmill in Olsztyn). The aim of the study at the first stage was to determine the meteorological and technical conditions for the acquisition of aerial and ground-based photographs. At the next stage, the technology of 3D modelling was developed using only ground-based or only aerial non-metric digital images. At the last stage of the study, an experiment was conducted to assess the possibility of 3D modelling with the comprehensive use of aerial (UAV) and ground-based digital photographs in terms of their labour intensity and precision of development. Data integration and automatic photo-realistic 3D construction of the models was done with Pix4Dmapper and Agisoft PhotoScan software Analyses have shown that when certain parameters established in an experiment are kept, the process of developing the stock-taking documentation for a historical building moves from the standards of analogue to digital technology with considerably reduced cost.

  20. Accuracy of software designed for automated localization of the inferior alveolar nerve canal on cone beam CT images

    PubMed Central

    Zamani, Ali; Kashkouli, Sadegh; Soltanimehr, Elham; Ghofrani Jahromi, Mohsen; Sanaeian Pourshirazi, Zahra

    2016-01-01

    Objectives: The aim of this study was to design and evaluate a new method for automated localization of the inferior alveolar nerve canal on CBCT images. Methods: The proposed method is based on traversing both panoramic and cross-sectional slices. For the panoramic slices, morphological skeletonization is imposed, and a modified Hough transform is used while traversing the cross-sectional slices. A total of 40 CBCT images were randomly selected. Two experts twice located the inferior alveolar nerve canal during two examinations set 6 weeks apart. Agreement between experts was achieved, and the result of this manual technique was considered the gold standard for our study. The distances for the automated method and those determined using the gold standard method were calculated and recorded. The mean time required for the automated detection was also recorded. Results: The average mean distance error from the baseline was 0.75 ± 0.34 mm. In all, 86% of the detected points had a mean error of <1 mm compared with those determined by the manual gold standard method. Conclusions: The proposed method is far more accurate and faster than previous methods. It also provides more accuracy than human annotation within a shorter time. PMID:26652929

  1. ThermalTracker Software

    SciTech Connect

    2016-08-10

    The software processes recorded thermal video and detects the flight tracks of birds and bats that passed through the camera's field of view. The output is a set of images that show complete flight tracks for any detections, with the direction of travel indicated and the thermal image of the animal delineated. A report of the descriptive features of each detected track is also output in the form of a comma-separated value text file.

  2. Automated rejection of contaminated surface measurements for improved surface registration in image guided neurosurgery.

    PubMed

    Bucholz, R; Macneil, W; Fewings, P; Ravindra, A; McDurmont, L; Baumann, C

    2000-01-01

    Most image guided Neurosurgery employs adhesively mounted external fiducials for registration of medical images to the surgical workspace. Due to high logistical costs associated with these artificial landmarks, we strive to eliminate the need for these markers. At our institution, we developed a handheld laser stripe triangulation device to capture the surface contours of the patient's head while oriented for surgery. Anatomical surface registration algorithms rely on the assumption that the patient's anatomy bears the same geometry as the 3D model of the patient constructed from the imaging modality employed. During the time interval from which the patient is imaged and placed in the Mayfield head clamp in the operating room, the skin of the head bulges at the pinsite and the skull fixation equipment itself optically interferes with the image capture laser. We have developed software to reject points belonging to objects of known geometry while calculating the registration. During the course of development of the laser scanning unit, we have acquired surface contours of 13 patients and 2 cadavers. Initial analysis revealed that this automated rejection of points improved the registrations in all cases, but the accuracy of the fiducial method was not surpassed. Only points belonging to the offending instrument are removed. Skin bulges caused by the clamps and instruments remain in the data. We anticipate that careful removal of the points in these skin bulges will yield registrations that at least match the accuracy of the fiducial method.

  3. The National Alliance for Medical Image Computing, a roadmap initiative to build a free and open source software infrastructure for translational research in medical image analysis.

    PubMed

    Kapur, Tina; Pieper, Steve; Whitaker, Ross; Aylward, Stephen; Jakab, Marianna; Schroeder, Will; Kikinis, Ron

    2012-01-01

    The National Alliance for Medical Image Computing (NA-MIC), is a multi-institutional, interdisciplinary community of researchers, who share the recognition that modern health care demands improved technologies to ease suffering and prolong productive life. Organized under the National Centers for Biomedical Computing 7 years ago, the mission of NA-MIC is to implement a robust and flexible open-source infrastructure for developing and applying advanced imaging technologies across a range of important biomedical research disciplines. A measure of its success, NA-MIC is now applying this technology to diseases that have immense impact on the duration and quality of life: cancer, heart disease, trauma, and degenerative genetic diseases. The targets of this technology range from group comparisons to subject-specific analysis.

  4. Data reduction software for the Mid-Infrared E-ELT Imager and Spectrograph (METIS) for the European Extremely Large Telescope (E-ELT)

    NASA Astrophysics Data System (ADS)

    Mach, Michael; Köhler, Rainer; Czoske, Oliver; Leschinski, Kieran; Zeilinger, Werner W.; Kausch, Wolfgang; Ratzka, Thorsten; Leitzinger, Martin; Greimel, Robert; Przybilla, Norbert; Schaffenroth, Veronika; Güdel, Manuel; Brandl, Bernhard R.

    2016-07-01

    We present the current status of the design of the science data reduction pipeline and the corresponding dataflow system for METIS. It will be one of the first three instruments for the E-ELT and work at wavelengths between 3-19 μm (L/M/N/Q1 bands). We will deliver software which is compliant to standards of the European Southern Observatory (ESO), and will employ state of the art techniques to produce science grade data, master calibration frames, quality control parameters and to handle instrument effects. The Instrument currently offers a wealth of observing modes that are listed in this paper. Data reduction for a ground based instrument at these wavelengths is particularly challenging because of the massive influence of thermal radiation from various sources. We will give a comprehensive overview of the data ow system for the imaging modes that the instrument offers and discuss a single recipe versus a multi recipe approach for the different observing modes for imaging.

  5. Software Smarts

    NASA Technical Reports Server (NTRS)

    1998-01-01

    Under an SBIR (Small Business Innovative Research) contract with Johnson Space Center, Knowledge Based Systems Inc. (KBSI) developed an intelligent software environment for modeling and analyzing mission planning activities, simulating behavior, and, using a unique constraint propagation mechanism, updating plans with each change in mission planning activities. KBSI developed this technology into a commercial product, PROJECTLINK, a two-way bridge between PROSIm, KBSI's process modeling and simulation software and leading project management software like Microsoft Project and Primavera's SureTrak Project Manager.

  6. Global Workspace Dynamics: Cortical “Binding and Propagation” Enables Conscious Contents

    PubMed Central

    Baars, Bernard J.; Franklin, Stan; Ramsoy, Thomas Zoega

    2013-01-01

    A global workspace (GW) is a functional hub of binding and propagation in a population of loosely coupled signaling elements. In computational applications, GW architectures recruit many distributed, specialized agents to cooperate in resolving focal ambiguities. In the brain, conscious experiences may reflect a GW function. For animals, the natural world is full of unpredictable dangers and opportunities, suggesting a general adaptive pressure for brains to resolve focal ambiguities quickly and accurately. GW theory aims to understand the differences between conscious and unconscious brain events. In humans and related species the cortico-thalamic (C-T) core is believed to underlie conscious aspects of perception, thinking, learning, feelings of knowing (FOK), felt emotions, visual imagery, working memory, and executive control. Alternative theoretical perspectives are also discussed. The C-T core has many anatomical hubs, but conscious percepts are unitary and internally consistent at any given moment. Over time, conscious contents constitute a very large, open set. This suggests that a brain-based GW capacity cannot be localized in a single anatomical hub. Rather, it should be sought in a functional hub – a dynamic capacity for binding and propagation of neural signals over multiple task-related networks, a kind of neuronal cloud computing. In this view, conscious contents can arise in any region of the C-T core when multiple input streams settle on a winner-take-all equilibrium. The resulting conscious gestalt may ignite an any-to-many broadcast, lasting ∼100–200 ms, and trigger widespread adaptation in previously established networks. To account for the great range of conscious contents over time, the theory suggests an open repertoire of binding1 coalitions that can broadcast via theta/gamma or alpha/gamma phase coupling, like radio channels competing for a narrow frequency band. Conscious moments are thought to hold only 1–4 unrelated items; this

  7. CSAM Metrology Software Tool

    NASA Technical Reports Server (NTRS)

    Vu, Duc; Sandor, Michael; Agarwal, Shri

    2005-01-01

    CSAM Metrology Software Tool (CMeST) is a computer program for analysis of false-color CSAM images of plastic-encapsulated microcircuits. (CSAM signifies C-mode scanning acoustic microscopy.) The colors in the images indicate areas of delamination within the plastic packages. Heretofore, the images have been interpreted by human examiners. Hence, interpretations have not been entirely consistent and objective. CMeST processes the color information in image-data files to detect areas of delamination without incurring inconsistencies of subjective judgement. CMeST can be used to create a database of baseline images of packages acquired at given times for comparison with images of the same packages acquired at later times. Any area within an image can be selected for analysis, which can include examination of different delamination types by location. CMeST can also be used to perform statistical analyses of image data. Results of analyses are available in a spreadsheet format for further processing. The results can be exported to any data-base-processing software.

  8. Remote Viewer for Maritime Robotics Software

    NASA Technical Reports Server (NTRS)

    Kuwata, Yoshiaki; Wolf, Michael; Huntsberger, Terrance L.; Howard, Andrew B.

    2013-01-01

    This software is a viewer program for maritime robotics software that provides a 3D visualization of the boat pose, its position history, ENC (Electrical Nautical Chart) information, camera images, map overlay, and detected tracks.

  9. Software Reviews.

    ERIC Educational Resources Information Center

    McGrath, Diane

    1990-01-01

    Reviews two programs: (1) "The Weather Machine" on understanding weather and weather forecasting and (2) "The Mystery of the Hotel Victoria" on problem solving in mathematics. Presents the descriptions, advantages, and weaknesses of the software. (YP)

  10. Software Reviews.

    ERIC Educational Resources Information Center

    Wulfson, Stephen, Ed.

    1987-01-01

    Reviews seven computer software programs that can be used in science education programs. Describes courseware which deals with muscles and bones, terminology, classifying animals without backbones, molecular structures, drugs, genetics, and shaping the earth's surface. (TW)

  11. Software Reviews.

    ERIC Educational Resources Information Center

    Mathematics and Computer Education, 1988

    1988-01-01

    Presents reviews of six software packages. Includes (1) "Plain Vanilla Statistics"; (2) "MathCAD 2.0"; (3) "GrFx"; (4) "Trigonometry"; (5) "Algebra II"; (6) "Algebra Drill and Practice I, II, and III." (PK)

  12. Software Reviews.

    ERIC Educational Resources Information Center

    Wulfson, Stephen, Ed.

    1987-01-01

    Provides a review of four science software programs. Includes topics such as plate tectonics, laboratory experiment simulations, the human body, and light and temperature. Contains information on ordering and reviewers' comments. (ML)

  13. Software Reviews.

    ERIC Educational Resources Information Center

    Wulfson, Stephen, Ed.

    1990-01-01

    Reviewed are six computer software packages including "Lunar Greenhouse,""Dyno-Quest,""How Weather Works,""Animal Trackers,""Personal Science Laboratory," and "The Skeletal and Muscular Systems." Availability, functional, and hardware requirements are discussed. (CW)

  14. Software Reviews.

    ERIC Educational Resources Information Center

    Classroom Computer Learning, 1990

    1990-01-01

    Reviewed are three computer software packages including "Martin Luther King, Jr.: Instant Replay of History,""Weeds to Trees," and "The New Print Shop, School Edition." Discussed are hardware requirements, costs, grade levels, availability, emphasis, strengths, and weaknesses. (CW)

  15. Software Reviews.

    ERIC Educational Resources Information Center

    Wulfson, Eugene T., Ed.

    1988-01-01

    Presents reviews by classroom teachers of software for teaching science. Includes material on the work of geologists, genetics, earth science, classification of living things, astronomy, endangered species, skeleton, drugs, and heartbeat. Provides information on availability and equipment needed. (RT)

  16. Software Reviews.

    ERIC Educational Resources Information Center

    Dwyer, Donna; And Others

    1989-01-01

    Reviewed are seven software packages for Apple and IBM computers. Included are: "Toxicology"; "Science Corner: Space Probe"; "Alcohol and Pregnancy"; "Science Tool Kit Plus"; Computer Investigations: Plant Growth"; "Climatrolls"; and "Animal Watch: Whales." (CW)

  17. SU-E-I-63: Quantitative Evaluation of the Effects of Orthopedic Metal Artifact Reduction (OMAR) Software On CT Images for Radiotherapy Simulation

    SciTech Connect

    Jani, S

    2014-06-01

    Purpose: CT simulation for patients with metal implants can often be challenging due to artifacts that obscure tumor/target delineation and normal organ definition. Our objective was to evaluate the effectiveness of Orthopedic Metal Artifact Reduction (OMAR), a commercially available software, in reducing metal-induced artifacts and its effect on computed dose during treatment planning. Methods: CT images of water surrounding metallic cylindrical rods made of aluminum, copper and iron were studied in terms of Hounsfield Units (HU) spread. Metal-induced artifacts were characterized in terms of HU/Volume Histogram (HVH) using the Pinnacle treatment planning system. Effects of OMAR on enhancing our ability to delineate organs on CT and subsequent dose computation were examined in nine (9) patients with hip implants and two (2) patients with breast tissue expanders. Results: Our study characterized water at 1000 HU with a standard deviation (SD) of about 20 HU. The HVHs allowed us to evaluate how the presence of metal changed the HU spread. For example, introducing a 2.54 cm diameter copper rod in water increased the SD in HU of the surrounding water from 20 to 209, representing an increase in artifacts. Subsequent use of OMAR brought the SD down to 78. Aluminum produced least artifacts whereas Iron showed largest amount of artifacts. In general, an increase in kVp and mA during CT scanning showed better effectiveness of OMAR in reducing artifacts. Our dose analysis showed that some isodose contours shifted by several mm with OMAR but infrequently and were nonsignificant in planning process. Computed volumes of various dose levels showed <2% change. Conclusions: In our experience, OMAR software greatly reduced the metal-induced CT artifacts for the majority of patients with implants, thereby improving our ability to delineate tumor and surrounding organs. OMAR had a clinically negligible effect on computed dose within tissues. Partially funded by unrestricted

  18. EXSdetect: an end-to-end software for extended source detection in X-ray images: application to Swift-XRT data

    NASA Astrophysics Data System (ADS)

    Liu, T.; Tozzi, P.; Tundo, E.; Moretti, A.; Wang, J.-X.; Rosati, P.; Guglielmetti, F.

    2013-01-01

    Aims: We present a stand-alone software (named EXSdetect) for the detection of extended sources in X-ray images. Our goal is to provide a flexible tool capable of detecting extended sources down to the lowest flux levels attainable within instrumental limitations, while maintaining robust photometry, high completeness, and low contamination, regardless of source morphology. EXSdetect was developed mainly to exploit the ever-increasing wealth of archival X-ray data, but is also ideally suited to explore the scientific capabilities of future X-ray facilities, with a strong focus on investigations of distant groups and clusters of galaxies. Methods: EXSdetect combines a fast Voronoi tessellation code with a friends-of-friends algorithm and an automated deblending procedure. The values of key parameters are matched to fundamental telescope properties such as angular resolution and instrumental background. In addition, the software is designed to permit extensive tests of its performance via simulations of a wide range of observational scenarios. Results: We applied EXSdetect to simulated data fields modeled to realistically represent the Swift X-ray Cluster Survey (SXCS), which is based on archival data obtained by the X-ray telescope onboard the Swift satellite. We achieve more than 90% completeness for extended sources comprising at least 80 photons in the 0.5-2 keV band, a limit that corresponds to 10-14 erg cm-2 s-1 for the deepest SXCS fields. This detection limit is comparable to the one attained by the most sensitive cluster surveys conducted with much larger X-ray telescopes. While evaluating the performance of EXSdetect, we also explored the impact of improved angular resolution and discuss the ideal properties of the next generation of X-ray survey missions. The Phyton code EXSdetect is available on the SXCS website http://adlibitum.oats.inaf.it/sxcs

  19. Reflight certification software design specifications

    NASA Technical Reports Server (NTRS)

    1984-01-01

    The PDSS/IMC Software Design Specification for the Payload Development Support System (PDSS)/Image Motion Compensator (IMC) is contained. The PDSS/IMC is to be used for checkout and verification of the IMC flight hardware and software by NASA/MSFC.

  20. Software Based Supernova Recognition

    NASA Astrophysics Data System (ADS)

    Walters, Stephen M.

    2014-05-01

    This paper describes software for detecting Supernova (SN) in images. The software can operate in real-time to discover SN while data is being collected so the instrumentation can immediately be re-tasked to perform spectroscopy or photometry of a discovery. Because the instrumentation captures two images per minute, the realtime budget is constrained to 30 seconds per target, a challenging goal. Using a set of two to four images, the program creates a "Reference" (REF) image and a "New" (NEW) image where all images are used in both NEW and REF but any SN survives the combination process only in the NEW image. This process produces good quality images having similar noise characteristics but without artifacts that might be interpreted as SN. The images are then adjusted for seeing and brightness differences using a variant of Tomaney and Crotts method of Point Spread Function (PSF) matching after which REF is subtracted from NEW to produce a Difference (DIF) image. A Classifier is then trained on a grid of artificial SN to estimate the statistical properties of four attributes and used in a process to mask false positives that can be clearly identified as such. Further training to avoid any remaining false positives sets the range, in standard deviations for each attribute, that the Classifier will accept as a valid SN. This training enables the Classifier to discriminate between SN and most subtraction residue. Lastly, the DIF image is scanned and measured by the Classifier to find locations where all four properties fall within their acceptance ranges. If multiple locations are found, the one best conforming to the training estimates is chosen. This location is then declared as a Candidate SN, the instrumentation re-tasked and the operator notified.

  1. Workspace design for crane cabins applying a combined traditional approach and the Taguchi method for design of experiments.

    PubMed

    Spasojević Brkić, Vesna K; Veljković, Zorica A; Golubović, Tamara; Brkić, Aleksandar Dj; Kosić Šotić, Ivana

    2016-01-01

    Procedures in the development process of crane cabins are arbitrary and subjective. Since approximately 42% of incidents in the construction industry are linked to them, there is a need to collect fresh anthropometric data and provide additional recommendations for design. In this paper, dimensioning of the crane cabin interior space was carried out using a sample of 64 crane operators' anthropometric measurements, in the Republic of Serbia, by measuring workspace with 10 parameters using nine measured anthropometric data from each crane operator. This paper applies experiments run via full factorial designs using a combined traditional and Taguchi approach. The experiments indicated which design parameters are influenced by which anthropometric measurements and to what degree. The results are expected to be of use for crane cabin designers and should assist them to design a cabin that may lead to less strenuous sitting postures and fatigue for operators, thus improving safety and accident prevention.

  2. Designing, Supporting, and Sustaining an Online Community of Practice: NASA EPO Workspace as an Ongoing Exploration of the Value of Community

    NASA Astrophysics Data System (ADS)

    Davey, B.; Davis, H. B.

    2015-12-01

    Increasingly, geographically diverse organizations, like NASA's Science Mission Directorate Education and Public Outreach personnel (SMD EPO), are looking for ways to facilitate group interactions in meaningful ways while limiting costs. Towards this end, of particular interest, and showing great potential are communities of practice. Communities of practice represent relationships in real-time between and among people sharing a common practice. They facilitate the sharing of information, building collective knowledge, and growing of the principles of practice. In 2010-11, SMD EPO established a website to support EPO professionals, facilitate headquarters reporting, and foster a community of practice. The purpose of this evaluation is to examine the design and use of the workspace and the value created for both individual community members and SMD EPO, the sponsoring organization. The online workspace was launched in 2010-11 for the members of NASA's SMDEPO community. The online workspace was designed to help facilitate the efficient sharing of information, be a central repository for resources, help facilitate and support knowledge creation, and ultimately lead to the development of an online community of practice. This study examines the role of the online workspace component of a community in the work of a community of practice. Much has been studied revealing the importance of communities of practice to organizations, project success, and knowledge management and some of these same successes hold true for virtual communities of practice. Additionally, we look at the outcomes of housting the online community for these past years in respect to knowledge building and personal and organizational value, the affects on professional dvelopment opportunities, how community members have benefited, and how the workspace has evolved to better serve the community.

  3. Software reengineering

    NASA Technical Reports Server (NTRS)

    Fridge, Ernest M., III

    1991-01-01

    Today's software systems generally use obsolete technology, are not integrated properly with other software systems, and are difficult and costly to maintain. The discipline of reverse engineering is becoming prominent as organizations try to move their systems up to more modern and maintainable technology in a cost effective manner. JSC created a significant set of tools to develop and maintain FORTRAN and C code during development of the Space Shuttle. This tool set forms the basis for an integrated environment to re-engineer existing code into modern software engineering structures which are then easier and less costly to maintain and which allow a fairly straightforward translation into other target languages. The environment will support these structures and practices even in areas where the language definition and compilers do not enforce good software engineering. The knowledge and data captured using the reverse engineering tools is passed to standard forward engineering tools to redesign or perform major upgrades to software systems in a much more cost effective manner than using older technologies. A beta vision of the environment was released in Mar. 1991. The commercial potential for such re-engineering tools is very great. CASE TRENDS magazine reported it to be the primary concern of over four hundred of the top MIS executives.

  4. Software engineering

    NASA Technical Reports Server (NTRS)

    Fridge, Ernest M., III; Hiott, Jim; Golej, Jim; Plumb, Allan

    1993-01-01

    Today's software systems generally use obsolete technology, are not integrated properly with other software systems, and are difficult and costly to maintain. The discipline of reverse engineering is becoming prominent as organizations try to move their systems up to more modern and maintainable technology in a cost effective manner. The Johnson Space Center (JSC) created a significant set of tools to develop and maintain FORTRAN and C code during development of the space shuttle. This tool set forms the basis for an integrated environment to reengineer existing code into modern software engineering structures which are then easier and less costly to maintain and which allow a fairly straightforward translation into other target languages. The environment will support these structures and practices even in areas where the language definition and compilers do not enforce good software engineering. The knowledge and data captured using the reverse engineering tools is passed to standard forward engineering tools to redesign or perform major upgrades to software systems in a much more cost effective manner than using older technologies. The latest release of the environment was in Feb. 1992.

  5. Antiterrorist Software

    NASA Technical Reports Server (NTRS)

    Clark, David A.

    1998-01-01

    In light of the escalation of terrorism, the Department of Defense spearheaded the development of new antiterrorist software for all Government agencies by issuing a Broad Agency Announcement to solicit proposals. This Government-wide competition resulted in a team that includes NASA Lewis Research Center's Computer Services Division, who will develop the graphical user interface (GUI) and test it in their usability lab. The team launched a program entitled Joint Sphere of Security (JSOS), crafted a design architecture (see the following figure), and is testing the interface. This software system has a state-ofthe- art, object-oriented architecture, with a main kernel composed of the Dynamic Information Architecture System (DIAS) developed by Argonne National Laboratory. DIAS will be used as the software "breadboard" for assembling the components of explosions, such as blast and collapse simulations.

  6. FMT (Flight Software Memory Tracker) For Cassini Spacecraft-Software Engineering Using JAVA

    NASA Technical Reports Server (NTRS)

    Kan, Edwin P.; Uffelman, Hal; Wax, Allan H.

    1997-01-01

    The software engineering design of the Flight Software Memory Tracker (FMT) Tool is discussed in this paper. FMT is a ground analysis software set, consisting of utilities and procedures, designed to track the flight software, i.e., images of memory load and updatable parameters of the computers on-board Cassini spacecraft. FMT is implemented in Java.

  7. Data processing software suite SITENNO for coherent X-ray diffraction imaging using the X-ray free-electron laser SACLA

    PubMed Central

    Sekiguchi, Yuki; Oroguchi, Tomotaka; Takayama, Yuki; Nakasako, Masayoshi

    2014-01-01

    Coherent X-ray diffraction imaging is a promising technique for visualizing the structures of non-crystalline particles with dimensions of micrometers to sub-micrometers. Recently, X-ray free-electron laser sources have enabled efficient experiments in the ‘diffraction before destruction’ scheme. Diffraction experiments have been conducted at SPring-8 Angstrom Compact free-electron LAser (SACLA) using the custom-made diffraction apparatus KOTOBUKI-1 and two multiport CCD detectors. In the experiments, ten thousands of single-shot diffraction patterns can be collected within several hours. Then, diffraction patterns with significant levels of intensity suitable for structural analysis must be found, direct-beam positions in diffraction patterns determined, diffraction patterns from the two CCD detectors merged, and phase-retrieval calculations for structural analyses performed. A software suite named SITENNO has been developed to semi-automatically apply the four-step processing to a huge number of diffraction data. Here, details of the algorithm used in the suite are described and the performance for approximately 9000 diffraction patterns collected from cuboid-shaped copper oxide particles reported. Using the SITENNO suite, it is possible to conduct experiments with data processing immediately after the data collection, and to characterize the size distribution and internal structures of the non-crystalline particles. PMID:24763651

  8. Integrating NASA's Land Analysis System (LAS) image processing software with an appropriate Geographic Information System (GIS): A review of candidates in the public domain

    NASA Technical Reports Server (NTRS)

    Rochon, Gilbert L.

    1989-01-01

    A user requirements analysis (URA) was undertaken to determine and appropriate public domain Geographic Information System (GIS) software package for potential integration with NASA's LAS (Land Analysis System) 5.0 image processing system. The necessity for a public domain system was underscored due to the perceived need for source code access and flexibility in tailoring the GIS system to the needs of a heterogenous group of end-users, and to specific constraints imposed by LAS and its user interface, Transportable Applications Executive (TAE). Subsequently, a review was conducted of a variety of public domain GIS candidates, including GRASS 3.0, MOSS, IEMIS, and two university-based packages, IDRISI and KBGIS. The review method was a modified version of the GIS evaluation process, development by the Federal Interagency Coordinating Committee on Digital Cartography. One IEMIS-derivative product, the ALBE (AirLand Battlefield Environment) GIS, emerged as the most promising candidate for integration with LAS. IEMIS (Integrated Emergency Management Information System) was developed by the Federal Emergency Management Agency (FEMA). ALBE GIS is currently under development at the Pacific Northwest Laboratory under contract with the U.S. Army Corps of Engineers' Engineering Topographic Laboratory (ETL). Accordingly, recommendations are offered with respect to a potential LAS/ALBE GIS linkage and with respect to further system enhancements, including coordination with the development of the Spatial Analysis and Modeling System (SAMS) GIS in Goddard's IDM (Intelligent Data Management) developments in Goddard's National Space Science Data Center.

  9. Data processing software suite SITENNO for coherent X-ray diffraction imaging using the X-ray free-electron laser SACLA.

    PubMed

    Sekiguchi, Yuki; Oroguchi, Tomotaka; Takayama, Yuki; Nakasako, Masayoshi

    2014-05-01

    Coherent X-ray diffraction imaging is a promising technique for visualizing the structures of non-crystalline particles with dimensions of micrometers to sub-micrometers. Recently, X-ray free-electron laser sources have enabled efficient experiments in the `diffraction before destruction' scheme. Diffraction experiments have been conducted at SPring-8 Angstrom Compact free-electron LAser (SACLA) using the custom-made diffraction apparatus KOTOBUKI-1 and two multiport CCD detectors. In the experiments, ten thousands of single-shot diffraction patterns can be collected within several hours. Then, diffraction patterns with significant levels of intensity suitable for structural analysis must be found, direct-beam positions in diffraction patterns determined, diffraction patterns from the two CCD detectors merged, and phase-retrieval calculations for structural analyses performed. A software suite named SITENNO has been developed to semi-automatically apply the four-step processing to a huge number of diffraction data. Here, details of the algorithm used in the suite are described and the performance for approximately 9000 diffraction patterns collected from cuboid-shaped copper oxide particles reported. Using the SITENNO suite, it is possible to conduct experiments with data processing immediately after the data collection, and to characterize the size distribution and internal structures of the non-crystalline particles.

  10. [Software version and medical device software supervision].

    PubMed

    Peng, Liang; Liu, Xiaoyan

    2015-01-01

    The importance of software version in the medical device software supervision does not cause enough attention at present. First of all, the effect of software version in the medical device software supervision is discussed, and then the necessity of software version in the medical device software supervision is analyzed based on the discussion of the misunderstanding of software version. Finally the concrete suggestions on software version naming rules, software version supervision for the software in medical devices, and software version supervision scheme are proposed.

  11. Educational Software.

    ERIC Educational Resources Information Center

    Northwest Regional Educational Lab., Portland, OR.

    The third session of IT@EDU98 consisted of five papers on educational software and was chaired by Tran Van Hao (University of Education, Ho Chi Minh City, Vietnam). "Courseware Engineering" (Nguyen Thanh Son, Ngo Ngoc Bao Tran, Quan Thanh Tho, Nguyen Hong Lam) briefly describes the use of courseware. "Machine Discovery Theorems in Geometry: A…

  12. Software Reviews.

    ERIC Educational Resources Information Center

    Sidwell, Joseph C.; And Others

    1988-01-01

    Gives a review of four software packages including "Science Toolkit: Module 3--Body Lab" for measuring heart rate, lung capacity, and response time; "Project Zoo: Adventures with Charts and Graphs" for developing process skills; "The Body Electric" for explaining electrical activity in the body; and "M-ss-ng…

  13. Software Reviews.

    ERIC Educational Resources Information Center

    Classroom Computer Learning, 1990

    1990-01-01

    Reviewed are computer software packages: "Where in Time Is Carmen Sandiego,""The Bio Sci Videodisc," and "Bio Sci Stacks." Included are hardware requirements, costs, emphasis, grade level, and availability. Functions of the packages are discussed including strengths and weaknesses and teaching suggestions. (CW)

  14. Software Reviews.

    ERIC Educational Resources Information Center

    History Microcomputer Review, 1988

    1988-01-01

    Reviews seven educational computer software packages covering such topics as presidential elections, the American Revolution, the Vietnam War, the construction of historical time lines, and general U.S. history. Also reviews a program designed to help tailor data entry files. Provides ordering information, price, and computer compatibility…

  15. Reviews: Software.

    ERIC Educational Resources Information Center

    Mackenzie, Norma N.; And Others

    1988-01-01

    Reviews four computer software packages including: "The Physical Science Series: Sound" which demonstrates making waves, speed of sound, doppler effect, and human hearing; "Andromeda" depicting celestial motions in any direction; "Biology Quiz: Humans" covering chemistry, cells, viruses, and human biology; and…

  16. Software Reviews.

    ERIC Educational Resources Information Center

    McGrath, Diane, Ed.

    1990-01-01

    Reviews two computer software programs: (1) "Conquering Ratios and Proportions" using a medieval theme for guided practice in identifying and forming ratios for grades 5-8, and (2) "Percent Word Problems" providing problems for finding a percentage of a number and a number from a percentage. (YP)

  17. Software Update.

    ERIC Educational Resources Information Center

    Currents, 2000

    2000-01-01

    A chart of 40 alumni-development database systems provides information on vendor/Web site, address, contact/phone, software name, price range, minimum suggested workstation/suggested server, standard reports/reporting tools, minimum/maximum record capacity, and number of installed sites/client type. (DB)

  18. Software Reviews.

    ERIC Educational Resources Information Center

    Classroom Computer Learning, 1990

    1990-01-01

    Reviewed are two computer software packages: "Super Solvers Midnight Rescue!" a problem-solving program for IBM PCs; and "Interactive Physics," a simulation program for the Macintosh computer. The functions of the package are discussed including strengths and weaknesses and teaching suggestions. (CW)

  19. Software Reviews.

    ERIC Educational Resources Information Center

    Bitter, Gary G., Ed.

    1989-01-01

    Describes three software packages: (1) "MacMendeleev"--database/graphic display for chemistry, grades 10-12, Macintosh; (2) "Geometry One: Foundations"--geometry tutorial, grades 7-12, IBM; (3) "Mathematics Exploration Toolkit"--algebra and calculus tutorial, grades 8-12, IBM. (MVL)

  20. Software Reviews.

    ERIC Educational Resources Information Center

    Classroom Computer Learning, 1988

    1988-01-01

    Reviewed three computer software packages for Apple II series computers. Includes "The Right Job," a career counseling program; "Zoyon Patrol," a problem-solving program; and "Adventures with Charts and Graphs: Project Zoo," a graphing, mathematics, and science skills program. Each review includes strengths, weaknesses, and suggestions for use.…

  1. Reviews, Software.

    ERIC Educational Resources Information Center

    Science Teacher, 1988

    1988-01-01

    Reviews two software programs for Apple series computers. Includes "Orbital Mech," a basic planetary orbital simulation for the Macintosh, and "START: Stimulus and Response Tools for Experiments in Memory, Learning, Cognition, and Perception," a program that demonstrates basic psychological principles and experiments. (CW)

  2. Software Reviews.

    ERIC Educational Resources Information Center

    Kinnaman, Daniel E.; And Others

    1988-01-01

    Reviews four educational software packages for Apple, IBM, and Tandy computers. Includes "How the West was One + Three x Four,""Mavis Beacon Teaches Typing,""Math and Me," and "Write On." Reviews list hardware requirements, emphasis, levels, publisher, purchase agreements, and price. Discusses the strengths…

  3. Software Reviews.

    ERIC Educational Resources Information Center

    McGrath, Diane, Ed.

    1989-01-01

    Reviewed are two computer software programs for Apple II computers on weather for upper elementary and middle school grades. "Weather" introduces the major factors (temperature, humidity, wind, and air pressure) affecting weather. "How Weather Works" uses simulation and auto-tutorial formats on sun, wind, fronts, clouds, and…

  4. Star Software.

    ERIC Educational Resources Information Center

    Kloza, Brad

    2000-01-01

    Presents a collection of computer software programs designed to spark learning enthusiasm at every grade level and across the curriculum. They include Reader Rabbit's Learn to Read, Spelling Power, Mind Twister Math, Community Construction Kit, Breaking the Code, Encarta Africana 2000, Virtual Serengeti, Operation: Frog (Deluxe), and My First…

  5. Software Reviews.

    ERIC Educational Resources Information Center

    Smith, Richard L., Ed.

    1988-01-01

    Reviews three computer software programs: (1) "Discovery! Experiences with Scientific Reasoning"--problem solving for grades 4-12 (Apple II); (2) "Organic Stereochemistry"--a tutorial for organic chemistry for advanced secondary/college level (Apple II); and (3) "SHOW PARTNER (2.01)"--a graphics utility tool for…

  6. Software Reviews.

    ERIC Educational Resources Information Center

    Mackenzie, Norma N.; And Others

    1988-01-01

    Describes computer software for use with various age groups. Topics include activities involving temperature, simulations, earth science, the circulatory system, human body, reading in science, and ecology. Provides information on equipment needed, availability, package contents, and price. Comments of reviews are presented by classroom teachers.…

  7. Software Patents.

    ERIC Educational Resources Information Center

    Burke, Edmund B.

    1994-01-01

    Outlines basic patent law information that pertains to computer software programs. Topics addressed include protection in other countries; how to obtain patents; kinds of patents; duration; classes of patentable subject matter, including machines and processes; patentability searches; experimental use prior to obtaining a patent; and patent…

  8. Software Reviews.

    ERIC Educational Resources Information Center

    Smith, Richard L., Ed.

    1988-01-01

    Contains evaluations of two computer software packages, "Simulation Experiments 45-48 in Epstein's Laboratory Manual for Chemistry" and "Maps and Legends--the Cartographer (Ver 3.0)." Includes a brief description, applications, and the perceived strengths and weaknesses for each package. (CW)

  9. Statistical Software.

    ERIC Educational Resources Information Center

    Callamaras, Peter

    1983-01-01

    This buyer's guide to seven major types of statistics software packages for microcomputers reviews Edu-Ware Statistics 3.0; Financial Planning; Speed Stat; Statistics with DAISY; Human Systems Dynamics package of Stats Plus, ANOVA II, and REGRESS II; Maxistat; and Moore-Barnes' MBC Test Construction and MBC Correlation. (MBR)

  10. Software Reviews.

    ERIC Educational Resources Information Center

    Wulfson, Stephen, Ed.

    1990-01-01

    Reviewed are seven computer software packages including "Frog Dissection Lab Report,""Backyard Birds,""LEGO TC Logo,""Alcohol--Four Interactive Programs,""Windows on Science--Life Science,""Climate and Weather/Our Town Database," and "Weeds to Trees." Discussed are availability, features, strengths, and weaknesses. (CW)

  11. Software Reviews.

    ERIC Educational Resources Information Center

    Teles, Elizabeth, Ed.; And Others

    1990-01-01

    Reviewed are two computer software packages for Macintosh microcomputers including "Phase Portraits," an exploratory graphics tool for studying first-order planar systems; and "MacMath," a set of programs for exploring differential equations, linear algebra, and other mathematical topics. Features, ease of use, cost, availability, and hardware…

  12. Software Reviews.

    ERIC Educational Resources Information Center

    Science and Children, 1989

    1989-01-01

    Reviews of seven software packages are presented including "The Environment I: Habitats and EcoSystems; II Cycles and Interactions"; "Super Sign Maker"; "The Great Knowledge Race: Substance Abuse"; "Exploring Science: Temperature"; "Fast Food Calculator and RD Aide"; "The Human Body:…

  13. Software Reviews.

    ERIC Educational Resources Information Center

    Science and Children, 1990

    1990-01-01

    Reviewed are seven computer software packages for IBM and/or Apple Computers. Included are "Windows on Science: Volume 1--Physical Science"; "Science Probe--Physical Science"; "Wildlife Adventures--Grizzly Bears"; "Science Skills--Development Programs"; "The Clean Machine"; "Rock Doctor";…

  14. Software Review.

    ERIC Educational Resources Information Center

    McGrath, Diane, Ed.

    1989-01-01

    Reviewed is a computer software package entitled "Audubon Wildlife Adventures: Grizzly Bears" for Apple II and IBM microcomputers. Included are availability, hardware requirements, cost, and a description of the program. The murder-mystery flavor of the program is stressed in this program that focuses on illegal hunting and game…

  15. Software Reviews.

    ERIC Educational Resources Information Center

    Wulfson, Stephen, Ed.

    1990-01-01

    Reviewed are six software packages for Apple and/or IBM computers. Included are "Autograph,""The New Game Show,""Science Probe-Earth Science,""Pollution Patrol,""Investigating Plant Growth," and "AIDS: The Investigation." Discussed are the grade level, function, availability, cost, and hardware requirements of each. (CW)

  16. Software Reviews.

    ERIC Educational Resources Information Center

    Science and Children, 1988

    1988-01-01

    Reviews five software packages for use with school age children. Includes "Science Toolkit Module 2: Earthquake Lab"; "Adaptations and Identification"; "Geoworld"; "Body Systems II Series: The Blood System: A Liquid of Life," all for Apple II, and "Science Courseware: Life Science/Biology" for…

  17. Software Reviews.

    ERIC Educational Resources Information Center

    Mathematics and Computer Education, 1987

    1987-01-01

    Presented are reviews of several microcomputer software programs. Included are reviews of: (1) Microstat (Zenith); (2) MathCAD (MathSoft); (3) Discrete Mathematics (True Basic); (4) CALCULUS (True Basic); (5) Linear-Kit (John Wiley); and (6) Geometry Sensei (Broderbund). (RH)

  18. Software Reviews.

    ERIC Educational Resources Information Center

    Bitter, Gary G., Ed.

    1990-01-01

    Reviews three computer software: (1) "Elastic Lines: The Electronic Geoboard" on elementary geometry; (2) "Wildlife Adventures: Whales" on environmental science; and (3) "What Do You Do with a Broken Calculator?" on computation and problem solving. Summarizes the descriptions, strengths and weaknesses, and…

  19. Software Epistemology

    DTIC Science & Technology

    2016-03-01

    corpuses at scale using deep neural networks, i.e., Deep Machine Learning, on high quality features computed from canonical representations of...the application of Deep Learning on software features to support automated vulnerability identification and repair. 1.2 Overview Draper’s...referenced in Table 2. Several web -based tools were maintained to show cluster processing status. Figure 10 shows a snapshot of the build inventory

  20. Know Your Software Options.

    ERIC Educational Resources Information Center

    Moxley, Doug

    1986-01-01

    Advice on how to find the best software for institutional needs is presented. Purchasing prewritten software, acquiring custom-written software, and improving ready-made software are discussed. Questions to ask before buying software are provided. (MLW)

  1. Space Station Software Issues

    NASA Technical Reports Server (NTRS)

    Voigt, S. (Editor); Beskenis, S. (Editor)

    1985-01-01

    Issues in the development of software for the Space Station are discussed. Software acquisition and management, software development environment, standards, information system support for software developers, and a future software advisory board are addressed.

  2. Scientific Software

    NASA Technical Reports Server (NTRS)

    1995-01-01

    The Interactive Data Language (IDL), developed by Research Systems, Inc., is a tool for scientists to investigate their data without having to write a custom program for each study. IDL is based on the Mariners Mars spectral Editor (MMED) developed for studies from NASA's Mars spacecraft flights. The company has also developed Environment for Visualizing Images (ENVI), an image processing system for easily analyzing remotely sensed data written in IDL. The Visible Human CD, another Research Systems product, is the first complete digital reference of photographic images for exploring human anatomy.

  3. Strategy for the lowering and the assessment of exposure to nanoparticles at workspace - Case of study concerning the potential emission of nanoparticles of Lead in an epitaxy laboratory

    NASA Astrophysics Data System (ADS)

    Artous, Sébastien; Zimmermann, Eric; Douissard, Paul-Antoine; Locatelli, Dominique; Motellier, Sylvie; Derrough, Samir

    2015-05-01

    The implementation in many products of manufactured nanoparticles is growing fast and raises new questions. For this purpose, the CEA - NanoSafety Platform is developing various research topics for health and safety, environment and nanoparticles exposure in professional activities. The containment optimisation for the exposition lowering, then the exposure assessment to nanoparticles is a strategy for safety improvement at workplace and workspace. The lowering step consists in an optimisation of dynamic and static containment at workplace and/or workspace. Generally, the exposure risk due to the presence of nanoparticles substances does not allow modifying the parameters of containment at workplace and/or workspace. Therefore, gaseous or nanoparticulate tracers are used to evaluate performances of containment. Using a tracer allows to modify safely the parameters of the dynamic containment (ventilation, flow, speed) and to study several configurations of static containment. Moreover, a tracer allows simulating accidental or incidental situation. As a result, a safety procedure can be written more easily in order to manage this type of situation. The step of measurement and characterization of aerosols can therefore be used to assess the exposition at workplace and workspace. The case of study, aim of this paper, concerns the potential emission of Lead nanoparticles at the exhaust of a furnace in an epitaxy laboratory. The use of Helium tracer to evaluate the performance of containment is firstly studied. Secondly, the exposure assessment is characterised in accordance with the French guide “recommendations for characterizing potential emissions and exposure to aerosols released from nanomaterials in workplace operations”. Thirdly the aerosols are sampled, on several places, using collection membranes to try to detect traces of Lead in air.

  4. Software Prototyping

    PubMed Central

    Del Fiol, Guilherme; Hanseler, Haley; Crouch, Barbara Insley; Cummins, Mollie R.

    2016-01-01

    Summary Background Health information exchange (HIE) between Poison Control Centers (PCCs) and Emergency Departments (EDs) could improve care of poisoned patients. However, PCC information systems are not designed to facilitate HIE with EDs; therefore, we are developing specialized software to support HIE within the normal workflow of the PCC using user-centered design and rapid prototyping. Objective To describe the design of an HIE dashboard and the refinement of user requirements through rapid prototyping. Methods Using previously elicited user requirements, we designed low-fidelity sketches of designs on paper with iterative refinement. Next, we designed an interactive high-fidelity prototype and conducted scenario-based usability tests with end users. Users were asked to think aloud while accomplishing tasks related to a case vignette. After testing, the users provided feedback and evaluated the prototype using the System Usability Scale (SUS). Results Survey results from three users provided useful feedback that was then incorporated into the design. After achieving a stable design, we used the prototype itself as the specification for development of the actual software. Benefits of prototyping included having 1) subject-matter experts heavily involved with the design; 2) flexibility to make rapid changes, 3) the ability to minimize software development efforts early in the design stage; 4) rapid finalization of requirements; 5) early visualization of designs; 6) and a powerful vehicle for communication of the design to the programmers. Challenges included 1) time and effort to develop the prototypes and case scenarios; 2) no simulation of system performance; 3) not having all proposed functionality available in the final product; and 4) missing needed data elements in the PCC information system. PMID:27081404

  5. Software Surrogate

    NASA Technical Reports Server (NTRS)

    1999-01-01

    In 1994, Blackboard Technology received a NASA Phase I SBIR award entitled "A Blackboard-Based Framework for Mixed-Initiative, Crewed- Space-System Applications." This research continued in Phase II at JSC, where a generic architecture was developed in which a software surrogate serves as the operator's representative in the fast-paced realm of nearly autonomous, intelligent systems. This SBIR research effort addressed the need to support human-operator monitoring and intervention with intelligent systems such as those being developed for NASA's crewed space program.

  6. Analysis Software

    NASA Technical Reports Server (NTRS)

    1994-01-01

    General Purpose Boundary Element Solution Technology (GPBEST) software employs the boundary element method of mechanical engineering analysis, as opposed to finite element. It is, according to one of its developers, 10 times faster in data preparation and more accurate than other methods. Its use results in less expensive products because the time between design and manufacturing is shortened. A commercial derivative of a NASA-developed computer code, it is marketed by Best Corporation to solve problems in stress analysis, heat transfer, fluid analysis and yielding and cracking of solids. Other applications include designing tractor and auto parts, household appliances and acoustic analysis.

  7. Simulation Software

    NASA Technical Reports Server (NTRS)

    1996-01-01

    Various NASA Small Business Innovation Research grants from Marshall Space Flight Center, Langley Research Center and Ames Research Center were used to develop the 'kernel' of COMCO's modeling and simulation software, the PHLEX finite element code. NASA needed it to model designs of flight vehicles; one of many customized commercial applications is UNISIM, a PHLEX-based code for analyzing underground flows in oil reservoirs for Texaco, Inc. COMCO's products simulate a computational mechanics problem, estimate the solution's error and produce the optimal hp-adapted mesh for the accuracy the user chooses. The system is also used as a research or training tool in universities and in mechanical design in industrial corporations.

  8. The most relevant human brain regions for functional connectivity: Evidence for a dynamical workspace of binding nodes from whole-brain computational modelling.

    PubMed

    Deco, Gustavo; Van Hartevelt, Tim J; Fernandes, Henrique M; Stevner, Angus; Kringelbach, Morten L

    2017-02-01

    In order to promote survival through flexible cognition and goal-directed behaviour, the brain has to optimize segregation and integration of information into coherent, distributed dynamical states. Certain organizational features of the brain have been proposed to be essential to facilitate cognitive flexibility, especially hub regions in the so-called rich club which show dense interconnectivity. These structural hubs have been suggested to be vital for integration and segregation of information. Yet, this has not been evaluated in terms of resulting functional temporal dynamics. A complementary measure covering the temporal aspects of functional connectivity could thus bring new insights into a more complete picture of the integrative nature of brain networks. Here, we use causal whole-brain computational modelling to determine the functional dynamical significance of the rich club and compare this to a new measure of the most functionally relevant brain regions for binding information over time ("dynamical workspace of binding nodes"). We found that removal of the iteratively generated workspace of binding nodes impacts significantly more on measures of integration and encoding of information capability than the removal of the rich club regions. While the rich club procedure produced almost half of the binding nodes, the remaining nodes have low degree yet still play a significant role in the workspace essential for binding information over time and as such goes beyond a description of the structural backbone.

  9. Software system safety

    NASA Technical Reports Server (NTRS)

    Uber, James G.

    1988-01-01

    Software itself is not hazardous, but since software and hardware share common interfaces there is an opportunity for software to create hazards. Further, these software systems are complex, and proven methods for the design, analysis, and measurement of software safety are not yet available. Some past software failures, future NASA software trends, software engineering methods, and tools and techniques for various software safety analyses are reviewed. Recommendations to NASA are made based on this review.

  10. Kinematics of the Shoulder Girdle During Pointing: Coordination Between Joints and their Contribution to the Peri-Personal Workspace.

    PubMed

    Roby-Brami, Agnès; Robertson, Johanna V G; Roren, Alexandra; Lefèvre-Colau, Marie-Martine

    2016-08-19

    This study explored the coordination between the components of the shoulder girdle (clavicle, scapula and humerus), and how they contribute to hand movement in the peri-personal space. Shoulder girdle motion was recorded in 10 healthy subjects during pointing movements to 9 targets in the peri-personal space, using electromagnetic sensors fixed to the trunk, scapula and upper arm. Most of the 9 degrees of freedom (DoF) of the shoulder girdle were finely scaled to target position. Principle component analysis revealed that the 6 DoF of scapula-thoracic motion were coordinated in three elementary patterns (protraction, shrug and lateral rotation). The ratio of gleno-humeral to scapulo-thoracic global motion was close to 2:1. A direct kinematic procedure showed that if no scapular motion occurred, the workspace would be reduced by 15.8 cm laterally, 13.7 cm vertically and 4.8 cm anteriorly. Scapulo-thoracic motion should be taken into account when investigating the physiology of upper-limb movements.

  11. Differences in granular materials for analogue modelling: Insights from repeated compression tests analyzed with X-ray Computed Tomography and image analysis software

    NASA Astrophysics Data System (ADS)

    Klinkmueller, M.; Schreurs, G.

    2009-12-01

    Six different granular materials for analogue modelling have been investigated using a sandbox with a compressional set-up and X-ray computed tomography (XRCT). The evolving structures were evaluated with image analysis software. The sandbox has one movable sidewall that is driven by a computer-controlled servomotor at 20 cm/h. A 12 cm wide and 20 cm long sheet of hard cardboard was placed on the base of the sandbox and attached to the moving sidewall creating a velocity discontinuity. The whole sandbox was covered on the inside with Alkor foil to reduce sidewall friction. Computed Tomography was used to scan the whole volume in 3 mm increments of shortening until 15 mm maximum deformation was reached. The second approach was a scanning procedure to a maximum deformation of 80 mm in 2 mm increments of shortening for the first 10 mm and in 5 mm increments for the last 70 mm. The short deformation scans were repeated three times to investigate reproducibility. The long deformation scans were performed twice. The physical properties of the materials (table 1) have been described in a previous material benchmark. Four natural quartz sands and two artificial granular materials, corundum brown sand and glass beads, have been used. The two artificial materials were used for this experimental series as examples for very angular and very rounded sands in contrast to the sub-rounded to angular natural quartz sands. The short deformation experiments show partly large differences in thrust angles of both front and back-thrust, in timing of thrust initiation, and in the degree of undulation of thrusts. The coarse-grained sands show smooth and low undulating thrusts that are only affected by the sidewall friction whereas the thrusts in fine-grained sands undulate significantly and partly divide and merge in an anastomosing fashion. The coarse-grained sand thrusts are clearer visualized by XRCT, which indicates a wider shear zone where the material dilates. Furthermore, the

  12. PIV Data Validation Software Package

    NASA Technical Reports Server (NTRS)

    Blackshire, James L.

    1997-01-01

    A PIV data validation and post-processing software package was developed to provide semi-automated data validation and data reduction capabilities for Particle Image Velocimetry data sets. The software provides three primary capabilities including (1) removal of spurious vector data, (2) filtering, smoothing, and interpolating of PIV data, and (3) calculations of out-of-plane vorticity, ensemble statistics, and turbulence statistics information. The software runs on an IBM PC/AT host computer working either under Microsoft Windows 3.1 or Windows 95 operating systems.

  13. Phoenix landing site and sample context images from the Surface Stereo Imager

    NASA Astrophysics Data System (ADS)

    Lemmon, M. T.; Arvidson, R.; Blaney, D.; Dejong, E.; Madsen, M. B.; Malin, M.; Mellon, M.; Morris, R.; Pike, W. T.; Smith, P. H.; Stoker, C.; Team, P. S.

    2008-12-01

    Phoenix landed in the northern plains of Mars in an area with low rock abundance dominated by few-meter- scale polygonal patterned ground with decimeter scale troughs. The Phoenix Surface Stereo Imager (SSI) provides geomorphic and spectral information about the Phoenix landing site for scales that range from site- wide to context for samples analyzed by other Phoenix instruments. The SSI is a multispectral stereo camera with properties that are comparable to the Mars Exploration Rover Pancam. It has MER-heritage 1024x1024 pixel detectors, a 14-degree field of view for individual images, and resolution as high as 1-2 mm for near- field terrain (0.24 mrad/pixel). Images are taken through one of 24 filters, including 13 unique spectral bandpasses, 2 stereo bandpasses, 2 filters paired with lenses for best focus on the lander deck, 6 solar filters for atmospheric dust and water vapor and ice measurements, and 1 polarizer. The stereo separation of the eyes is 15 cm, and the focus and toe-in are optimized at 3 to 3.5 m to support Robotic Arm (RA) operations. SSI can image from the camera bar at -72 degrees to the zenith, and through 360 degrees of azimuth. As with Pancam, panoramic images are built on the ground from a number of individual frames. SSI provided geomorphic information through a set of campaigns. Three major site panoramas were acquired: on sols 1 and 3, a low-resolution monochromatic site panorama provided context for higher- resolution images in the RA workspace; a color-stereo panorama was completed on sol 43; a multispectral high-resolution panorama is currently underway. High resolution detail observations were conducted throughout the mission for high priority targets in and beyond the workspace. These campaigns show a landscape dominated by polygons with typical diameters of 2 to 4 meters. Troughs between the polygons have depths of typically 5-20 cm relative to the polygon centers. Phoenix landed with access to a trough and parts of two polygons

  14. Software Engineering Support of the Third Round of Scientific Grand Challenge Investigations: Earth System Modeling Software Framework Survey

    NASA Technical Reports Server (NTRS)

    Talbot, Bryan; Zhou, Shu-Jia; Higgins, Glenn; Zukor, Dorothy (Technical Monitor)

    2002-01-01

    One of the most significant challenges in large-scale climate modeling, as well as in high-performance computing in other scientific fields, is that of effectively integrating many software models from multiple contributors. A software framework facilitates the integration task, both in the development and runtime stages of the simulation. Effective software frameworks reduce the programming burden for the investigators, freeing them to focus more on the science and less on the parallel communication implementation. while maintaining high performance across numerous supercomputer and workstation architectures. This document surveys numerous software frameworks for potential use in Earth science modeling. Several frameworks are evaluated in depth, including Parallel Object-Oriented Methods and Applications (POOMA), Cactus (from (he relativistic physics community), Overture, Goddard Earth Modeling System (GEMS), the National Center for Atmospheric Research Flux Coupler, and UCLA/UCB Distributed Data Broker (DDB). Frameworks evaluated in less detail include ROOT, Parallel Application Workspace (PAWS), and Advanced Large-Scale Integrated Computational Environment (ALICE). A host of other frameworks and related tools are referenced in this context. The frameworks are evaluated individually and also compared with each other.

  15. The Analysis of the Patterns of Radiation-Induced DNA Damage Foci by a Stochastic Monte Carlo Model of DNA Double Strand Breaks Induction by Heavy Ions and Image Segmentation Software

    NASA Technical Reports Server (NTRS)

    Ponomarev, Artem; Cucinotta, F.

    2011-01-01

    To create a generalized mechanistic model of DNA damage in human cells that will generate analytical and image data corresponding to experimentally observed DNA damage foci and will help to improve the experimental foci yields by simulating spatial foci patterns and resolving problems with quantitative image analysis. Material and Methods: The analysis of patterns of RIFs (radiation-induced foci) produced by low- and high-LET (linear energy transfer) radiation was conducted by using a Monte Carlo model that combines the heavy ion track structure with characteristics of the human genome on the level of chromosomes. The foci patterns were also simulated in the maximum projection plane for flat nuclei. Some data analysis was done with the help of image segmentation software that identifies individual classes of RIFs and colocolized RIFs, which is of importance to some experimental assays that assign DNA damage a dual phosphorescent signal. Results: The model predicts the spatial and genomic distributions of DNA DSBs (double strand breaks) and associated RIFs in a human cell nucleus for a particular dose of either low- or high-LET radiation. We used the model to do analyses for different irradiation scenarios. In the beam-parallel-to-the-disk-of-a-flattened-nucleus scenario we found that the foci appeared to be merged due to their high density, while, in the perpendicular-beam scenario, the foci appeared as one bright spot per hit. The statistics and spatial distribution of regions of densely arranged foci, termed DNA foci chains, were predicted numerically using this model. Another analysis was done to evaluate the number of ion hits per nucleus, which were visible from streaks of closely located foci. In another analysis, our image segmentaiton software determined foci yields directly from images with single-class or colocolized foci. Conclusions: We showed that DSB clustering needs to be taken into account to determine the true DNA damage foci yield, which helps to

  16. Sandia software guidelines: Software quality planning

    SciTech Connect

    Not Available

    1987-08-01

    This volume is one in a series of Sandia Software Guidelines intended for use in producing quality software within Sandia National Laboratories. In consonance with the IEEE Standard for Software Quality Assurance Plans, this volume identifies procedures to follow in producing a Software Quality Assurance Plan for an organization or a project, and provides an example project SQA plan. 2 figs., 4 tabs.

  17. Office Computer Software: A Comprehensive Review of Software Programs.

    ERIC Educational Resources Information Center

    Secretary, 1992

    1992-01-01

    Describes types of software including system software, application software, spreadsheets, accounting software, graphics packages, desktop publishing software, database, desktop and personal information management software, project and records management software, groupware, and shareware. (JOW)

  18. Going to where the users are! Making the collaborative resource management and science workspace mobile

    NASA Astrophysics Data System (ADS)

    Osti, D.; Osti, A.

    2013-12-01

    People are very busy today and getting stakeholders the information they need is an important part of our jobs. The BDL application is the mobile extension of the California collaborative resource management portal www.baydeltalive.com. BDL has been visited by more than 250,000 unique visitors this past year from various areas of water use and management including state and federal agencies, agriculture, scientists, policy makers, water consumers, voters, operations management and more. The audience is a qualified user group of more than 15,000 individuals participating in California hydrological ecosystem science, water management and policy. This is an important effort aimed to improve how scientists and policy makers are working together to understand this complicated and divisive system and how they are becoming better managers of that system. The BayDetaLive mobile application gives California watershed management stakeholders and water user community unprecedented access to real time natural resource management information. The application provides user with the following: 1. Access to Real Time Environmental Conditions from the more than the 600 California Data Exchange Sensors including hydrodynamic, water quality and meteorological data. Save important stations as favorites for easy access later. 2. Daily Delta Operations Data including estimated hydrology, daily exports, status of infrastructure operations, reservoir storage, salvage data, major stations, drinking water quality reports, weather forecasts and more. 3. Photos/Videos/Documents: Browse and share from the more than 1000 current documents in the BDL library. Relevant images, videos, science journals, presentations and articles. 4. Science: Access the latest science articles, news, projects and journals. 5. Data Visualizations: View recently published real time data interpolations of Delta Conditions. From 30-day turbidity models to daily forecasts. This service is published as conditions

  19. A dedicated software application for treatment verification with off-line PET/CT imaging at the Heidelberg Ion Beam Therapy Center

    NASA Astrophysics Data System (ADS)

    Chen, W.; Bauer, J.; Kurz, C.; Tessonnier, T.; Handrack, J.; Haberer, T.; Debus, J.; Parodi, K.

    2017-01-01

    We present the workflow of the offline-PET based range verification method used at the Heidelberg Ion Beam Therapy Center, detailing the functionalities of an in-house developed software application, SimInterface14, with which range analysis is performed. Moreover, we introduce the design of a decision support system assessing uncertainties and facilitating physicians in decisions making for plan adaptation.

  20. Software Model Of Software-Development Process

    NASA Technical Reports Server (NTRS)

    Lin, Chi Y.; Synott, Debra J.; Levary, Reuven R.

    1990-01-01

    Collection of computer programs constitutes software tool for simulation of medium- to large-scale software-development projects. Necessary to include easily identifiable and more-readily quantifiable characteristics like costs, times, and numbers of errors. Mathematical model incorporating these and other factors of dynamics of software-development process implemented in the Software Life Cycle Simulator (SLICS) computer program. Simulates dynamics of software-development process. In combination with input and output expert software systems and knowledge-based management software system, develops information for use in managing large software-development project. Intended to aid managers in planning, managing, and controlling software-development processes by reducing uncertainties in budgets, required personnel, and schedules.

  1. PROMOTIONS: PROper MOTION Software

    NASA Astrophysics Data System (ADS)

    Caleb Wherry, John; Sahai, R.

    2009-05-01

    We report on the development of a software tool (PROMOTIONS) to streamline the process of measuring proper motions of material in expanding nebulae. Our tool makes use of IDL's widget programming capabilities to design a unique GUI that is used to compare images of the objects from two epochs. The software allows us to first orient and register the images to a common frame of reference and pixel scale, using field stars in each of the images. We then cross-correlate specific morphological features in order to determine their proper motions, which consist of the proper motion of the nebula as a whole (PM-neb), and expansion motions of the features relative to the center. If the central star is not visible (quite common in bipolar nebulae with dense dusty waists), point-symmetric expansion is assumed and we use the average motion of high-quality symmetric pairs of features on opposite sides of the nebular center to compute PM-neb. This is then subtracted out to determine the individual movements of these and additional features relative to the nebular center. PROMOTIONS should find wide applicability in measuring proper motions in astrophysical objects such as the expanding outflows/jets commonly seen around young and dying stars. We present first results from using PROMOTIONS to successfully measure proper motions in several pre-planetary nebulae (transition objects between the red giant and planetary nebula phases), using images taken 7-10 years apart with the WFPC2 and ACS instruments on board HST. The authors are grateful to NASA's Undergradute Scholars Research Program (USRP) for supporting this research.

  2. MORPH-II, a software package for the analysis of scanning-electron-micrograph images for the assessment of the fractal dimension of exposed stone surfaces

    USGS Publications Warehouse

    Mossotti, Victor G.; Eldeeb, A. Raouf

    2000-01-01

    Turcotte, 1997, and Barton and La Pointe, 1995, have identified many potential uses for the fractal dimension in physicochemical models of surface properties. The image-analysis program described in this report is an extension of the program set MORPH-I (Mossotti and others, 1998), which provided the fractal analysis of electron-microscope images of pore profiles (Mossotti and Eldeeb, 1992). MORPH-II, an integration of the modified kernel of the program MORPH-I with image calibration and editing facilities, was designed to measure the fractal dimension of the exposed surfaces of stone specimens as imaged in cross section in an electron microscope.

  3. IRCAMDR: IRCAM3 Data Reduction Software

    NASA Astrophysics Data System (ADS)

    Aspin, Colin; McCaughrean, Mark; Bridger, Alan B.; Baines, Dave; Beard, Steven; Chan, S.; Giddings, Jack; Hartley, K. F.; Horsfield, A. P.; Kelly, B. D.; Emerson, J. P.; Currie, Malcolm J.; Economou, Frossie

    2014-06-01

    The UKIRT IRCAM3 data reduction and analysis software package, IRCAMDR (formerly ircam_clred) analyzes and displays any 2D data image stored in the standard Starlink (ascl:1110.012) NDF data format. It reduces and analyzes IRCAM1/2 data images of 62x58 pixels and IRCAM3 images of 256x256 size. Most of the applications will work on NDF images of any physical (pixel) dimensions, for example, 1024x1024 CCD images can be processed.

  4. Flash Cards and Animation Software for Education.

    ERIC Educational Resources Information Center

    Byers, John A.

    1999-01-01

    Describes how a software program for DOS/Windows manages a collection of pictures such as photographic slides, overheads, or computer images in one or more databases. Explains how it transforms image files to raw binary files that can then be displayed like flash cards, or as an animated series of images. (Author/LRW)

  5. Book and Software Review.

    ERIC Educational Resources Information Center

    Wissick, Cheryl

    2000-01-01

    This introductory column on books and software concerned with special education technology presents an article by JuHye Yook on the software design process. It discusses the rationale for developing new software for students with reading disabilities, the design and development process, and analysis of the software design. Software use by two…

  6. Other People's Software

    NASA Astrophysics Data System (ADS)

    Mandel, E.; Murray, S. S.

    Why do we continually re-invent the astronomical software wheel? Why is it so difficult to use ``other people's software''? Leaving aside issues such as money, power, and control, we need to investigate practically how we can remove barriers to software sharing. This paper will offer a starting point for software cooperation, centered on the concept of ``minimal software buy-in''.

  7. Software attribute visualization for high integrity software

    SciTech Connect

    Pollock, G.M.

    1998-03-01

    This report documents a prototype tool developed to investigate the use of visualization and virtual reality technologies for improving software surety confidence. The tool is utilized within the execution phase of the software life cycle. It provides a capability to monitor an executing program against prespecified requirements constraints provided in a program written in the requirements specification language SAGE. The resulting Software Attribute Visual Analysis Tool (SAVAnT) also provides a technique to assess the completeness of a software specification.

  8. Space Flight Software Development Software for Intelligent System Health Management

    NASA Technical Reports Server (NTRS)

    Trevino, Luis C.; Crumbley, Tim

    2004-01-01

    The slide presentation examines the Marshall Space Flight Center Flight Software Branch, including software development projects, mission critical space flight software development, software technical insight, advanced software development technologies, and continuous improvement in the software development processes and methods.

  9. Controlling Software Piracy.

    ERIC Educational Resources Information Center

    King, Albert S.

    1992-01-01

    Explains what software manufacturers are doing to combat software piracy, recommends how managers should deal with this problem, and provides a role-playing exercise to help students understand the issues in software piracy. (SR)

  10. Report: Scientific Software.

    ERIC Educational Resources Information Center

    Borman, Stuart A.

    1985-01-01

    Discusses various aspects of scientific software, including evaluation and selection of commercial software products; program exchanges, catalogs, and other information sources; major data analysis packages; statistics and chemometrics software; and artificial intelligence. (JN)

  11. [The Development of a Normal Database of Elderly People for Use with the Statistical Analysis Software Easy Z-score Imaging System with 99mTc-ECD SPECT].

    PubMed

    Nemoto, Hirobumi; Iwasaka, Akemi; Hashimoto, Shingo; Hara, Tadashi; Nemoto, Kiyotaka; Asada, Takashi

    2015-11-01

    We created a new normal database of elderly individuals (Tsukuba-NDB) for easy Z-score Imaging System (eZIS), a statistical imaging analysis software, comprised of 44 healthy individuals aged 75 to 89 years. The Tsukuba-NDB was compared with a conventional NDB (Musashi-NDB) using Statistical Parametric Mapping (SPM8), eZIS analysis, mean images, standard deviation (SD) images, SD values, specific volume of interest analysis (SVA). Furthermore, the association of the mean cerebral blood flow (mCBF) with various clinical indicators was statistically analyzed. A group comparison using SPM8 indicated that the t-value of the Tsukuba-NDB was lower in the frontoparietal region but tended to be higher in the bilateral temporal lobes and the base of the brain than that of the Musashi-NDB. The results of eZIS analysis by Musashi-NDB in 48 subjects indicated the presence of mild decreases in cerebral blood flow in the bilateral frontoparietal lobes of 9 subjects, precuneus and posterior cingulate gyrus of 5 subjects, lingual gyrus of 4 subjects, and near the left frontal gyrus, temporal lobe, superior temporal gyrus, and lenticular nucleus of 12 subjects. The mean images showed that there were no visual differences between both NDBs. The SD images intensities and SD values were lower in Tsukuba-NDB. Clinical case comparison and visual evaluation demonstrated that the sites of decrease in blood flow were more clearly indicated by the Tsukuba-NDB. Furthermore, mCBF was 40.87 ± 0.52 ml/100 g/min (mean ± SE), and tended to decrease with age. The tendency was stronger in male subjects than female subjects. Among various clinical indicators, the platelet count was statistically significantly correlated with CBF. In conclusion, our results suggest that Tsukuba-NDB, which is incorporated into a statistical imaging analysis software, eZIS, is sensitive to changes in cerebral blood flow caused by Cranial nerve disease, dementia and cerebrovascular accidents, and can provide precise

  12. Analytical approaches to image orientation and stereo digitization applied in the Budnlab software. (Polish Title: Rozwiazania analityczne zwiazane z obsluga procesu orientacji zdjec oraz wykonywaniem opracowan wektorowych w programie Bundlab)

    NASA Astrophysics Data System (ADS)

    Kolecki, J.

    2015-12-01

    The Bundlab software has been developed mainly for academic and research application. This work can be treated as a kind of a report describing the current state of the development of this computer program, focusing especially on the analytical solutions. Firstly, the overall characteristics of the software are provided. Then the description of the image orientation procedure starting from the relative orientation is addressed. The applied solution is based on the coplanarity equation parametrized with the essential matrix. The problem is reformulated in order to solve it using methods of algebraic geometry. The solution is followed by the optimization involving the least square criterion. The formation of the image block from the oriented models as well as the absolute orientation procedure were implemented using the Horn approach as a base algorithm. The second part of the paper is devoted to the tools and methods applied in the stereo digitization module. The solutions that support the user and improve the accuracy are given. Within the paper a few exemplary applications and products are mentioned. The work finishes with the concepts of development and improvements of existing functions.

  13. Software Engineering Guidebook

    NASA Technical Reports Server (NTRS)

    Connell, John; Wenneson, Greg

    1993-01-01

    The Software Engineering Guidebook describes SEPG (Software Engineering Process Group) supported processes and techniques for engineering quality software in NASA environments. Three process models are supported: structured, object-oriented, and evolutionary rapid-prototyping. The guidebook covers software life-cycles, engineering, assurance, and configuration management. The guidebook is written for managers and engineers who manage, develop, enhance, and/or maintain software under the Computer Software Services Contract.

  14. Redundantly piezo-actuated XYθ z compliant mechanism for nano-positioning featuring simple kinematics, bi-directional motion and enlarged workspace

    NASA Astrophysics Data System (ADS)

    Zhu, Wu-Le; Zhu, Zhiwei; To, Suet; Liu, Qiang; Ju, Bing-Feng; Zhou, Xiaoqin

    2016-12-01

    This paper presents a novel redundantly piezo-actuated three-degree-of-freedom XYθ z compliant mechanism for nano-positioning, driven by four mirror-symmetrically configured piezoelectric actuators (PEAs). By means of differential motion principle, linearized kinematics and physically bi-directional motions in all the three directions are achieved. Meanwhile, the decoupled delivering of three-directional independent motions at the output end is accessible, and the essential parallel and mirror symmetric configuration guarantees large output stiffness, high natural frequencies, high accuracy as well as high structural compactness of the mechanism. Accurate kinematics analysis with consideration of input coupling indicates that the proposed redundantly actuated compliant mechanism can generate three-dimensional (3D) symmetric polyhedral workspace envelope with enlarged reachable workspace, as compared with the most common parallel XYθ z mechanism driven by three PEAs. Keeping a high consistence with both analytical and numerical models, the experimental results show the working ranges of ±6.21 μm and ±12.41 μm in X- and Y-directions, and that of ±873.2 μrad in θ z-direction with nano-positioning capability can be realized. The superior performances and easily achievable structure well facilitate practical applications of the proposed XYθ z compliant mechanism in nano-positioning systems.

  15. Revision and product generation software

    USGS Publications Warehouse

    ,

    1997-01-01

    The U.S. Geological Survey (USGS) developed revision and product generation (RevPG) software for updating digital line graph (DLG) data and producing maps from such data. This software is based on ARC/INFO, a geographic information system from Environmental Systems Resource Institute (ESRI). RevPG consists of ARC/INFO Arc Macro Language (AML) programs, C routines, and interface menus that permit operators to collect vector data using aerial images, to symbolize the data on-screen, and to produce plots and color-separated files for use in printing maps.

  16. Revision and Product Generation Software

    USGS Publications Warehouse

    ,

    1999-01-01

    The U.S. Geological Survey (USGS) developed revision and product generation (RevPG) software for updating digital line graph (DLG) data and producing maps from such data. This software is based on ARC/INFO, a geographic information system from Environmental Systems Resource Institute (ESRI). RevPG consists of ARC/INFO Arc Macro Language (AML) programs, C routines, and interface menus that permit operators to collect vector data using aerial images, to symbolize the data onscreen, and to produce plots and color-separated files for use in printing maps.

  17. Software Configuration Management Guidebook

    NASA Technical Reports Server (NTRS)

    1995-01-01

    The growth in cost and importance of software to NASA has caused NASA to address the improvement of software development across the agency. One of the products of this program is a series of guidebooks that define a NASA concept of the assurance processes which are used in software development. The Software Assurance Guidebook, SMAP-GB-A201, issued in September, 1989, provides an overall picture of the concepts and practices of NASA in software assurance. Lower level guidebooks focus on specific activities that fall within the software assurance discipline, and provide more detailed information for the manager and/or practitioner. This is the Software Configuration Management Guidebook which describes software configuration management in a way that is compatible with practices in industry and at NASA Centers. Software configuration management is a key software development process, and is essential for doing software assurance.

  18. Method comparison of automated matching software-assisted cone-beam CT and stereoscopic kilovoltage x-ray positional verification image-guided radiation therapy for head and neck cancer: a prospective analysis

    PubMed Central

    Fuller, Clifton D; Scarbrough, Todd J; Sonke, Jan-Jakob; Rasch, Coen R N; Choi, Mehee; Ting, Joe Y; Wang, Samuel J; Papanikolaou, Niko; Rosenthal, David I

    2017-01-01

    We sought to characterize interchangeability and agreement between cone-beam computed tomography (CBCT) and digital stereoscopic kV x-ray (KVX) acquisition, two methods of isocenter positional verification currently used for IGRT of head and neck cancers (HNC). A cohort of 33 patients were near-simultaneously imaged by in-room KVX and CBCT. KVX and CBCT shifts were suggested using manufacturer software for the lateral (X), vertical (Y) and longitudinal (Z) dimensions. Intra-method repeatability, systematic and random error components were calculated for each imaging modality, as were recipe-based PTV expansion margins. Inter-method agreement in each axis was compared using limits of agreement (LOA) methodology, concordance analysis and orthogonal regression. 100 daily positional assessments were performed before daily therapy in 33 patients with head and neck cancer. Systematic error was greater for CBCT in all axes, with larger random error components in the Y- and Z-axis. Repeatability ranged from 9 to 14 mm for all axes, with CBCT showing greater repeatability in 2/3 axes. LOA showed paired shifts to agree 95% of the time within ±11.3 mm in the X-axis, ±9.4 mm in the Y-axis and ±5.5 mm in the Z-axis. Concordance ranged from ‘mediocre’ to ‘satisfactory’. Proportional bias was noted between paired X- and Z-axis measures, with a constant bias component in the Z-axis. Our data suggest non-negligible differences in software-derived CBCT and KVX image-guided directional shifts using formal method comparison statistics. PMID:19934488

  19. Method comparison of automated matching software-assisted cone-beam CT and stereoscopic kilovoltage x-ray positional verification image-guided radiation therapy for head and neck cancer: a prospective analysis

    NASA Astrophysics Data System (ADS)

    Fuller, Clifton D.; Scarbrough, Todd J.; Sonke, Jan-Jakob; Rasch, Coen R. N.; Choi, Mehee; Ting, Joe Y.; Wang, Samuel J.; Papanikolaou, Niko; Rosenthal, David I.

    2009-12-01

    We sought to characterize interchangeability and agreement between cone-beam computed tomography (CBCT) and digital stereoscopic kV x-ray (KVX) acquisition, two methods of isocenter positional verification currently used for IGRT of head and neck cancers (HNC). A cohort of 33 patients were near-simultaneously imaged by in-room KVX and CBCT. KVX and CBCT shifts were suggested using manufacturer software for the lateral (X), vertical (Y) and longitudinal (Z) dimensions. Intra-method repeatability, systematic and random error components were calculated for each imaging modality, as were recipe-based PTV expansion margins. Inter-method agreement in each axis was compared using limits of agreement (LOA) methodology, concordance analysis and orthogonal regression. 100 daily positional assessments were performed before daily therapy in 33 patients with head and neck cancer. Systematic error was greater for CBCT in all axes, with larger random error components in the Y- and Z-axis. Repeatability ranged from 9 to 14 mm for all axes, with CBCT showing greater repeatability in 2/3 axes. LOA showed paired shifts to agree 95% of the time within ±11.3 mm in the X-axis, ±9.4 mm in the Y-axis and ±5.5 mm in the Z-axis. Concordance ranged from 'mediocre' to 'satisfactory'. Proportional bias was noted between paired X- and Z-axis measures, with a constant bias component in the Z-axis. Our data suggest non-negligible differences in software-derived CBCT and KVX image-guided directional shifts using formal method comparison statistics. A correction was made to the first line of page 7404 of this article on 26 November 2009. The corrected electronic version is identical to the print version.

  20. Decentralized Software Evolution

    DTIC Science & Technology

    2003-09-01

    Institute for Software Research University of California, Irvine www.isr.uci.edu/tech-reports.html Peyman Oreizy University of California, Irvine... Peyman Oreizy and Richard N. Taylor Institute for Software Research University of California, Irvine Irvine, CA 92697-3425 USA {peymano, taylor...mechanisms that enforce cooperation among Decentralized Software Evolution Peyman Oreizy and Richard N. Taylor Institute for Software Research

  1. Agile Software Development

    ERIC Educational Resources Information Center

    Biju, Soly Mathew

    2008-01-01

    Many software development firms are now adopting the agile software development method. This method involves the customer at every level of software development, thus reducing the impact of change in the requirement at a later stage. In this article, the principles of the agile method for software development are explored and there is a focus on…

  2. Complexity, Systems, and Software

    DTIC Science & Technology

    2014-08-14

    2014 Carnegie Mellon University Complexity, Systems, and Software Software Engineering Institute Carnegie Mellon University Pittsburgh, PA...NUMBER 7. PERFORMING ORGANIZATION NAME(S) AND ADDRESS(ES) Software Engineering Institute Carnegie Mellon University Pittsburgh, PA 15213 8...for the operation of the Software Engineering Institute, a federally funded research and development center sponsored by the United States

  3. Finding Helpful Software Reviews.

    ERIC Educational Resources Information Center

    Kruse, Ted, Comp.

    1987-01-01

    Provides a list of evaluation services currently producing critical reviews of educational software. Includes information about The Apple K-12 Curriculum Software Reference, The Educational Software Preview, The Educational Software Selector, MicroSIFT, and Only The Best: The Discriminating Guide for Preschool-Grade 12. (TW)

  4. Software distribution using xnetlib

    SciTech Connect

    Dongarra, J.J. |; Rowan, T.H.; Wade, R.C.

    1993-06-01

    Xnetlib is a new tool for software distribution. Whereas its predecessor netlib uses e-mail as the user interface to its large collection of public-domain mathematical software, xnetlib uses an X Window interface and socket-based communication. Xnetlib makes it easy to search through a large distributed collection of software and to retrieve requested software in seconds.

  5. Software productivity improvement through software engineering technology

    NASA Technical Reports Server (NTRS)

    Mcgarry, F. E.

    1985-01-01

    It has been estimated that NASA expends anywhere from 6 to 10 percent of its annual budget on the acquisition, implementation and maintenance of computer software. Although researchers have produced numerous software engineering approaches over the past 5-10 years; each claiming to be more effective than the other, there is very limited quantitative information verifying the measurable impact htat any of these technologies may have in a production environment. At NASA/GSFC, an extended research effort aimed at identifying and measuring software techniques that favorably impact productivity of software development, has been active over the past 8 years. Specific, measurable, software development technologies have been applied and measured in a production environment. Resulting software development approaches have been shown to be effective in both improving quality as well as productivity in this one environment.

  6. Software Formal Inspections Standard

    NASA Technical Reports Server (NTRS)

    1993-01-01

    This Software Formal Inspections Standard (hereinafter referred to as Standard) is applicable to NASA software. This Standard defines the requirements that shall be fulfilled by the software formal inspections process whenever this process is specified for NASA software. The objective of this Standard is to define the requirements for a process that inspects software products to detect and eliminate defects as early as possible in the software life cycle. The process also provides for the collection and analysis of inspection data to improve the inspection process as well as the quality of the software.

  7. Mapping of dose distribution from IMRT onto MRI-guided high dose rate brachytherapy using deformable image registration for cervical cancer treatments: preliminary study with commercially available software

    PubMed Central

    Huq, M. Saiful; Houser, Chris; Beriwal, Sushil; Michalski, Dariusz

    2014-01-01

    Purpose For patients undergoing external beam radiation therapy (EBRT) and brachytherapy, recommendations for target doses and constraints are based on calculation of the equivalent dose in 2 Gy fractions (EQD2) from each phase. At present, the EBRT dose distribution is assumed to be uniform throughout the pelvis. We performed a preliminary study to determine whether deformable dose distribution mapping from the EBRT onto magnetic resonance (MR) images for the brachytherapy would yield differences in doses for organs at risk (OARs) and high-risk clinical target volume (HR-CTV). Material and methods Nine cervical cancer patients were treated to a total dose of 45 Gy in 25 fractions using intensity-modulated radiation therapy (IMRT), followed by MRI-based 3D high dose rate (HDR) brachytherapy. Retrospectively, the IMRT planning CT images were fused with the MR image for each fraction of brachytherapy using deformable image registration. The deformed IMRT dose onto MR images were converted to EQD2 and compared to the uniform dose assumption. Results For all patients, the EQD2 from the EBRT phase was significantly higher with deformable registration than with the conventional uniform dose distribution assumption. The mean EQD2 ± SD for HR-CTV D90 was 45.7 ± 0.7 Gy vs. 44.3 Gy for deformable vs. uniform dose distribution, respectively (p < 0.001). The dose to 2 cc of the bladder, rectum, and sigmoid was 46.4 ± 1.2 Gy, 46.2 ± 1.0 Gy, and 48.0 ± 2.5 Gy, respectively with deformable dose distribution, and was significantly higher than with uniform dose distribution (43.2 Gy for all OAR, p < 0.001). Conclusions This study reveals that deformed EBRT dose distribution to HR-CTV and OARs in MR images for brachytherapy is technically feasible, and achieves differences compared to a uniform dose distribution. Therefore, the assumption that EBRT contributes the same dose value may need to be carefully investigated further based on deformable image registration. PMID:25097559

  8. Lack of exposure to natural light in the workspace is associated with physiological, sleep and depressive symptoms.

    PubMed

    Harb, Francine; Hidalgo, Maria Paz; Martau, Betina

    2015-04-01

    The diurnal light cycle has a crucial influence on all life on earth. Unfortunately, modern society has modified this life-governing cycle by stressing maximum production and by giving insufficient attention to the ecological balance and homeostasis of the human metabolism. The aim of this study is to evaluate the effects of exposure or lack of exposure to natural light in a rest/activity rhythm on cortisol and melatonin levels, as well as on psychological variables in humans under natural conditions. This is a cross-sectional study. The subjects were allocated split into two groups according to their workspace (10 employees in the "with window" group and 10 in the "without window" group). All participants were women and wore anactigraph (Actiwatch 2, Philips Respironics), which measures activity and ambient light exposure, for seven days. Concentrations of melatonin and cortisol were measured from the saliva samples. Participants were instructed to collect saliva during the last day of use of the actigraph at 08:00 am, 4:00 pm and 10:00 pm. The subjects answered the Self-Reporting Questionnaire-20 (SRQ-20) to measure the presence of minor psychiatric disorders; the Montgomery-Asberg (MA) scale was used to measure depression symptoms, and the Pittsburgh Sleep Quality Index questionnaire (PSQI) was used to evaluate the quality of sleep. The Rayleigh analysis indicates that the two groups, "with window" an d "without window", exhibited similar activities and light acrophases. In relation to light exposure, the mesor was significantly higher (t = -2.651, p = 0.023) in t he "with window" group (191.04 ± 133.36) than in the "without window" group (73.8 ± 42.05). Additionally, the "with window" group presented the highest amplitude of light exposure (298.07 ± 222.97). Cortisol levels were significantly different between the groups at 10:00 pm (t = 3.009, p = 0.008; "without window" (4.01 ± 0.91) "with window" (3.10 ± 0.30)). In

  9. Software component quality evaluation

    NASA Technical Reports Server (NTRS)

    Clough, A. J.

    1991-01-01

    The paper describes a software inspection process that can be used to evaluate the quality of software components. Quality criteria, process application, independent testing of the process and proposed associated tool support are covered. Early results indicate that this technique is well suited for assessing software component quality in a standardized fashion. With automated machine assistance to facilitate both the evaluation and selection of software components, such a technique should promote effective reuse of software components.

  10. Responsbility for unreliable software

    SciTech Connect

    Wahl, N.J.

    1994-12-31

    Unreliable software exposes software developers and distributors to legal risks. Under certain circumstances, the developer and distributor of unreliable software can be sued. To avoid lawsuits, software developers should do the following: determine what the risks am, understand the extent of the risks, and identify ways of avoiding the risks and lessening the consequences of the risks. Liability issues associated with unreliable software are explored in this article.

  11. Software and tools for microarray data analysis.

    PubMed

    Mehta, Jai Prakash; Rani, Sweta

    2011-01-01

    A typical microarray experiment results in series of images, depending on the experimental design and number of samples. Software analyses the images to obtain the intensity at each spot and quantify the expression for each transcript. This is followed by normalization, and then various data analysis techniques are applied on the data. The whole analysis pipeline requires a large number of software to accurately handle the massive amount of data. Fortunately, there are large number of freely available and commercial software to churn the massive amount of data to manageable sets of differentially expressed genes, functions, and pathways. This chapter describes the software and tools which can be used to analyze the gene expression data right from the image analysis to gene list, ontology, and pathways.

  12. Imaging sciences workshop

    SciTech Connect

    Candy, J.V.

    1994-11-15

    This workshop on the Imaging Sciences sponsored by Lawrence Livermore National Laboratory contains short abstracts/articles submitted by speakers. The topic areas covered include the following: Astronomical Imaging; biomedical imaging; vision/image display; imaging hardware; imaging software; Acoustic/oceanic imaging; microwave/acoustic imaging; computed tomography; physical imaging; imaging algorithms. Selected papers are indexed separately for inclusion in the Energy Science and Technology Database.

  13. Software Quality Assurance Metrics

    NASA Technical Reports Server (NTRS)

    McRae, Kalindra A.

    2004-01-01

    Software Quality Assurance (SQA) is a planned and systematic set of activities that ensures conformance of software life cycle processes and products conform to requirements, standards and procedures. In software development, software quality means meeting requirements and a degree of excellence and refinement of a project or product. Software Quality is a set of attributes of a software product by which its quality is described and evaluated. The set of attributes includes functionality, reliability, usability, efficiency, maintainability, and portability. Software Metrics help us understand the technical process that is used to develop a product. The process is measured to improve it and the product is measured to increase quality throughout the life cycle of software. Software Metrics are measurements of the quality of software. Software is measured to indicate the quality of the product, to assess the productivity of the people who produce the product, to assess the benefits derived from new software engineering methods and tools, to form a baseline for estimation, and to help justify requests for new tools or additional training. Any part of the software development can be measured. If Software Metrics are implemented in software development, it can save time, money, and allow the organization to identify the caused of defects which have the greatest effect on software development. The summer of 2004, I worked with Cynthia Calhoun and Frank Robinson in the Software Assurance/Risk Management department. My task was to research and collect, compile, and analyze SQA Metrics that have been used in other projects that are not currently being used by the SA team and report them to the Software Assurance team to see if any metrics can be implemented in their software assurance life cycle process.

  14. Postprocessing classification images

    NASA Technical Reports Server (NTRS)

    Kan, E. P.

    1979-01-01

    Program cleans up remote-sensing maps. It can be used with existing image-processing software. Remapped images closely resemble familiar resource information maps and can replace or supplement classification images not postprocessed by this program.

  15. Segmentation of cervical cell nuclei in high-resolution microscopic images: A new algorithm and a web-based software framework.

    PubMed

    Bergmeir, Christoph; García Silvente, Miguel; Benítez, José Manuel

    2012-09-01

    In order to automate cervical cancer screening tests, one of the most important and longstanding challenges is the segmentation of cell nuclei in the stained specimens. Though nuclei of isolated cells in high-quality acquisitions often are easy to segment, the problem lies in the segmentation of large numbers of nuclei with various characteristics under differing acquisition conditions in high-resolution scans of the complete microscope slides. We implemented a system that enables processing of full resolution images, and proposes a new algorithm for segmenting the nuclei under adequate control of the expert user. The system can work automatically or interactively guided, to allow for segmentation within the whole range of slide and image characteristics. It facilitates data storage and interaction of technical and medical experts, especially with its web-based architecture. The proposed algorithm localizes cell nuclei using a voting scheme and prior knowledge, before it determines the exact shape of the nuclei by means of an elastic segmentation algorithm. After noise removal with a mean-shift and a median filtering takes place, edges are extracted with a Canny edge detection algorithm. Motivated by the observation that cell nuclei are surrounded by cytoplasm and their shape is roughly elliptical, edges adjacent to the background are removed. A randomized Hough transform for ellipses finds candidate nuclei, which are then processed by a level set algorithm. The algorithm is tested and compared to other algorithms on a database containing 207 images acquired from two different microscope slides, with promising results.

  16. Software Defined Radio with Parallelized Software Architecture

    NASA Technical Reports Server (NTRS)

    Heckler, Greg

    2013-01-01

    This software implements software-defined radio procession over multi-core, multi-CPU systems in a way that maximizes the use of CPU resources in the system. The software treats each processing step in either a communications or navigation modulator or demodulator system as an independent, threaded block. Each threaded block is defined with a programmable number of input or output buffers; these buffers are implemented using POSIX pipes. In addition, each threaded block is assigned a unique thread upon block installation. A modulator or demodulator system is built by assembly of the threaded blocks into a flow graph, which assembles the processing blocks to accomplish the desired signal processing. This software architecture allows the software to scale effortlessly between single CPU/single-core computers or multi-CPU/multi-core computers without recompilation. NASA spaceflight and ground communications systems currently rely exclusively on ASICs or FPGAs. This software allows low- and medium-bandwidth (100 bps to .50 Mbps) software defined radios to be designed and implemented solely in C/C++ software, while lowering development costs and facilitating reuse and extensibility.

  17. Software Defined Radio with Parallelized Software Architecture

    NASA Technical Reports Server (NTRS)

    Heckler, Greg

    2013-01-01

    This software implements software-defined radio procession over multicore, multi-CPU systems in a way that maximizes the use of CPU resources in the system. The software treats each processing step in either a communications or navigation modulator or demodulator system as an independent, threaded block. Each threaded block is defined with a programmable number of input or output buffers; these buffers are implemented using POSIX pipes. In addition, each threaded block is assigned a unique thread upon block installation. A modulator or demodulator system is built by assembly of the threaded blocks into a flow graph, which assembles the processing blocks to accomplish the desired signal processing. This software architecture allows the software to scale effortlessly between single CPU/single-core computers or multi-CPU/multi-core computers without recompilation. NASA spaceflight and ground communications systems currently rely exclusively on ASICs or FPGAs. This software allows low- and medium-bandwidth (100 bps to approx.50 Mbps) software defined radios to be designed and implemented solely in C/C++ software, while lowering development costs and facilitating reuse and extensibility.

  18. Payload software technology: Software technology development plan

    NASA Technical Reports Server (NTRS)

    1977-01-01

    Programmatic requirements for the advancement of software technology are identified for meeting the space flight requirements in the 1980 to 1990 time period. The development items are described, and software technology item derivation worksheets are presented along with the cost/time/priority assessments.

  19. Software Engineering Program: Software Process Improvement Guidebook

    NASA Technical Reports Server (NTRS)

    1996-01-01

    The purpose of this document is to provide experience-based guidance in implementing a software process improvement program in any NASA software development or maintenance community. This guidebook details how to define, operate, and implement a working software process improvement program. It describes the concept of the software process improvement program and its basic organizational components. It then describes the structure, organization, and operation of the software process improvement program, illustrating all these concepts with specific NASA examples. The information presented in the document is derived from the experiences of several NASA software organizations, including the SEL, the SEAL, and the SORCE. Their experiences reflect many of the elements of software process improvement within NASA. This guidebook presents lessons learned in a form usable by anyone considering establishing a software process improvement program within his or her own environment. This guidebook attempts to balance general and detailed information. It provides material general enough to be usable by NASA organizations whose characteristics do not directly match those of the sources of the information and models presented herein. It also keeps the ideas sufficiently close to the sources of the practical experiences that have generated the models and information.

  20. Space Station Software Recommendations

    NASA Technical Reports Server (NTRS)

    Voigt, S. (Editor)

    1985-01-01

    Four panels of invited experts and NASA representatives focused on the following topics: software management, software development environment, languages, and software standards. Each panel deliberated in private, held two open sessions with audience participation, and developed recommendations for the NASA Space Station Program. The major thrusts of the recommendations were as follows: (1) The software management plan should establish policies, responsibilities, and decision points for software acquisition; (2) NASA should furnish a uniform modular software support environment and require its use for all space station software acquired (or developed); (3) The language Ada should be selected for space station software, and NASA should begin to address issues related to the effective use of Ada; and (4) The space station software standards should be selected (based upon existing standards where possible), and an organization should be identified to promulgate and enforce them. These and related recommendations are described in detail in the conference proceedings.