Science.gov

Sample records for imaging workspace software

  1. Open source software projects of the caBIG In Vivo Imaging Workspace Software special interest group.

    PubMed

    Prior, Fred W; Erickson, Bradley J; Tarbox, Lawrence

    2007-11-01

    The Cancer Bioinformatics Grid (caBIG) program was created by the National Cancer Institute to facilitate sharing of IT infrastructure, data, and applications among the National Cancer Institute-sponsored cancer research centers. The program was launched in February 2004 and now links more than 50 cancer centers. In April 2005, the In Vivo Imaging Workspace was added to promote the use of imaging in cancer clinical trials. At the inaugural meeting, four special interest groups (SIGs) were established. The Software SIG was charged with identifying projects that focus on open-source software for image visualization and analysis. To date, two projects have been defined by the Software SIG. The eXtensible Imaging Platform project has produced a rapid application development environment that researchers may use to create targeted workflows customized for specific research projects. The Algorithm Validation Tools project will provide a set of tools and data structures that will be used to capture measurement information and associated needed to allow a gold standard to be defined for the given database against which change analysis algorithms can be tested. Through these and future efforts, the caBIG In Vivo Imaging Workspace Software SIG endeavors to advance imaging informatics and provide new open-source software tools to advance cancer research. PMID:17846835

  2. Teacher Workspaces

    ERIC Educational Resources Information Center

    Gordon, Douglas

    2010-01-01

    Well-designed and -equipped teacher workspaces provide the opportunity to improve student achievement at every step of their K-12 education. Shared workspace enhances communication among teachers as they evaluate student performance individually and collectively, and share insights with one another. This paper addresses the key elements found in…

  3. SAPHIRE 8 Volume 5 - Workspaces

    SciTech Connect

    C. L. Smith; J. K. Knudsen; D. O'Neal

    2011-03-01

    The Systems Analysis Programs for Hands-on Integrated Reliability Evaluations (SAPHIRE) Version 8 is a software application developed for performing a complete probabilistic risk assessment using a personal computer running the Microsoft Windows™ operating system. SAPHIRE 8 is funded by the U.S. Nuclear Regulatory Commission (NRC). The role of the Idaho National Laboratory (INL) in this project is that of software developer and tester. In older versions of SAPHIRE, the model creation and analysis functions were intermingled. However, in SAPHIRE 8, the act of creating a model has been separated from the analysis of that model in order to improve the quality of both the model (e.g., by avoiding inadvertent changes) and the analysis. Consequently, in SAPHIRE 8, the analysis of models is performed by using what are called Workspaces. Currently, there are Workspaces for three types of analyses: (1) the NRC’s Accident Sequence Precursor program, where the workspace is called “Events and Condition Assessment (ECA);” (2) the NRC’s Significance Determination Process (SDP); and (3) the General Analysis (GA) workspace. Workspaces for each type are created and saved separately from the base model which keeps the original database intact. Workspaces are independent of each other and modifications or calculations made within one workspace will not affect another. In addition, each workspace has a user interface and reports tailored for their intended uses.

  4. Cathodoluminescence Spectrum Imaging Software

    Energy Science and Technology Software Center (ESTSC)

    2011-04-07

    The software developed for spectrum imaging is applied to the analysis of the spectrum series generated by our cathodoluminescence instrumentation. This software provides advanced processing capabilities s such: reconstruction of photon intensity (resolved in energy) and photon energy maps, extraction of the spectrum from selected areas, quantitative imaging mode, pixel-to-pixel correlation spectrum line scans, ASCII, output, filling routines, drift correction, etc.

  5. Biological Imaging Software Tools

    PubMed Central

    Eliceiri, Kevin W.; Berthold, Michael R.; Goldberg, Ilya G.; Ibáñez, Luis; Manjunath, B.S.; Martone, Maryann E.; Murphy, Robert F.; Peng, Hanchuan; Plant, Anne L.; Roysam, Badrinath; Stuurman, Nico; Swedlow, Jason R.; Tomancak, Pavel; Carpenter, Anne E.

    2013-01-01

    Few technologies are more widespread in modern biological laboratories than imaging. Recent advances in optical technologies and instrumentation are providing hitherto unimagined capabilities. Almost all these advances have required the development of software to enable the acquisition, management, analysis, and visualization of the imaging data. We review each computational step that biologists encounter when dealing with digital images, the challenges in that domain, and the overall status of available software for bioimage informatics, focusing on open source options. PMID:22743775

  6. Image Processing Software

    NASA Technical Reports Server (NTRS)

    1992-01-01

    To convert raw data into environmental products, the National Weather Service and other organizations use the Global 9000 image processing system marketed by Global Imaging, Inc. The company's GAE software package is an enhanced version of the TAE, developed by Goddard Space Flight Center to support remote sensing and image processing applications. The system can be operated in three modes and is combined with HP Apollo workstation hardware.

  7. Flyover Animation of Phoenix Workspace

    NASA Technical Reports Server (NTRS)

    2008-01-01

    [figure removed for brevity, see original site] Click on image for animation

    This animated 'flyover' of the workspace of NASA's Phoenix Mars Lander's was created from images taken by the Surface Stereo Imager on Sol 14 (June 8, 2008), or the 14th Martian day after landing.

    The visualization uses both of the camera's 'eyes' to provide depth perception and ranging. The camera is looking north over the workspace.

    The Phoenix Mission is led by the University of Arizona, Tucson, on behalf of NASA. Project management of the mission is by NASA's Jet Propulsion Laboratory, Pasadena, Calif. Spacecraft development is by Lockheed Martin Space Systems, Denver.

  8. Confined Space Imager (CSI) Software

    SciTech Connect

    Karelilz, David

    2013-07-03

    The software provides real-time image capture, enhancement, and display, and sensor control for the Confined Space Imager (CSI) sensor system The software captures images over a Cameralink connection and provides the following image enhancements: camera pixel to pixel non-uniformity correction, optical distortion correction, image registration and averaging, and illumination non-uniformity correction. The software communicates with the custom CSI hardware over USB to control sensor parameters and is capable of saving enhanced sensor images to an external USB drive. The software provides sensor control, image capture, enhancement, and display for the CSI sensor system. It is designed to work with the custom hardware.

  9. IMAGE Software Suite

    NASA Technical Reports Server (NTRS)

    Gallagher, Dennis L.; Rose, M. Franklin (Technical Monitor)

    2000-01-01

    The IMAGE Mission is generating a truely unique set of magnetospheric measurement through a first-of-its-kind complement of remote, global observations. These data are being distributed in the Universal Data Format (UDF), which consists of data, calibration, and documentation. This is an open dataset, available to all by request to the National Space Science Data Center (NSSDC) at NASA Goddard Space Flight Center. Browse data, which consists of summary observations, is also available through the NSSDC in the Common Data Format (CDF) and graphic representations of the browse data. Access to the browse data can be achieved through the NSSDC CDAWeb services or by use of NSSDC provided software tools. This presentation documents the software tools, being provided by the IMAGE team, for use in viewing and analyzing the UDF telemetry data. Like the IMAGE data, these tools are openly available. What these tools can do, how they can be obtained, and how they are expected to evolve will be discussed.

  10. Software thermal imager simulator

    NASA Astrophysics Data System (ADS)

    Le Noc, Loic; Pancrati, Ovidiu; Doucet, Michel; Dufour, Denis; Debaque, Benoit; Turbide, Simon; Berthiaume, Francois; Saint-Laurent, Louis; Marchese, Linda; Bolduc, Martin; Bergeron, Alain

    2014-10-01

    A software application, SIST, has been developed for the simulation of the video at the output of a thermal imager. The approach offers a more suitable representation than current identification (ID) range predictors do: the end user can evaluate the adequacy of a virtual camera as if he was using it in real operating conditions. In particular, the ambiguity in the interpretation of ID range is cancelled. The application also allows for a cost-efficient determination of the optimal design of an imager and of its subsystems without over- or under-specification: the performances are known early in the development cycle, for targets, scene and environmental conditions of interest. The simulated image is also a powerful method for testing processing algorithms. Finally, the display, which can be a severe system limitation, is also fully considered in the system by the use of real hardware components. The application consists in Matlabtm routines that simulate the effect of the subsystems atmosphere, optical lens, detector, and image processing algorithms. Calls to MODTRAN® for the atmosphere modeling and to Zemax for the optical modeling have been implemented. The realism of the simulation depends on the adequacy of the input scene for the application and on the accuracy of the subsystem parameters. For high accuracy results, measured imager characteristics such as noise can be used with SIST instead of less accurate models. The ID ranges of potential imagers were assessed for various targets, backgrounds and atmospheric conditions. The optimal specifications for an optical design were determined by varying the Seidel aberration coefficients to find the worst MTF that still respects the desired ID range.

  11. Confined Space Imager (CSI) Software

    Energy Science and Technology Software Center (ESTSC)

    2013-07-03

    The software provides real-time image capture, enhancement, and display, and sensor control for the Confined Space Imager (CSI) sensor system The software captures images over a Cameralink connection and provides the following image enhancements: camera pixel to pixel non-uniformity correction, optical distortion correction, image registration and averaging, and illumination non-uniformity correction. The software communicates with the custom CSI hardware over USB to control sensor parameters and is capable of saving enhanced sensor images to anmore » external USB drive. The software provides sensor control, image capture, enhancement, and display for the CSI sensor system. It is designed to work with the custom hardware.« less

  12. Image processing software for imaging spectrometry

    NASA Technical Reports Server (NTRS)

    Mazer, Alan S.; Martin, Miki; Lee, Meemong; Solomon, Jerry E.

    1988-01-01

    The paper presents a software system, Spectral Analysis Manager (SPAM), which has been specifically designed and implemented to provide the exploratory analysis tools necessary for imaging spectrometer data, using only modest computational resources. The basic design objectives are described as well as the major algorithms designed or adapted for high-dimensional images. Included in a discussion of system implementation are interactive data display, statistical analysis, image segmentation and spectral matching, and mixture analysis.

  13. Phoenix Robotic Arm's Workspace After 90 Sols

    NASA Technical Reports Server (NTRS)

    2008-01-01

    During the first 90 Martian days, or sols, after its May 25, 2008, landing on an arctic plain of Mars, NASA's Phoenix Mars Lander dug several trenches in the workspace reachable with the lander's robotic arm.

    The lander's Surface Stereo Imager camera recorded this view of the workspace on Sol 90, early afternoon local Mars time (overnight Aug. 25 to Aug. 26, 2008). The shadow of the the camera itself, atop its mast, is just left of the center of the image and roughly a third of a meter (one foot) wide.

    The workspace is on the north side of the lander. The trench just to the right of center is called 'Neverland.'

    The Phoenix Mission is led by the University of Arizona, Tucson, on behalf of NASA. Project management of the mission is by NASA's Jet Propulsion Laboratory, Pasadena, Calif. Spacecraft development is by Lockheed Martin Space Systems, Denver.

  14. Spotlight-8 Image Analysis Software

    NASA Technical Reports Server (NTRS)

    Klimek, Robert; Wright, Ted

    2006-01-01

    Spotlight is a cross-platform GUI-based software package designed to perform image analysis on sequences of images generated by combustion and fluid physics experiments run in a microgravity environment. Spotlight can perform analysis on a single image in an interactive mode or perform analysis on a sequence of images in an automated fashion. Image processing operations can be employed to enhance the image before various statistics and measurement operations are performed. An arbitrarily large number of objects can be analyzed simultaneously with independent areas of interest. Spotlight saves results in a text file that can be imported into other programs for graphing or further analysis. Spotlight can be run on Microsoft Windows, Linux, and Apple OS X platforms.

  15. Automatic AVHRR image navigation software

    NASA Technical Reports Server (NTRS)

    Baldwin, Dan; Emery, William

    1992-01-01

    This is the final report describing the work done on the project entitled Automatic AVHRR Image Navigation Software funded through NASA-Washington, award NAGW-3224, Account 153-7529. At the onset of this project, we had developed image navigation software capable of producing geo-registered images from AVHRR data. The registrations were highly accurate but required a priori knowledge of the spacecraft's axes alignment deviations, commonly known as attitude. The three angles needed to describe the attitude are called roll, pitch, and yaw, and are the components of the deviations in the along scan, along track and about center directions. The inclusion of the attitude corrections in the navigation software results in highly accurate georegistrations, however, the computation of the angles is very tedious and involves human interpretation for several steps. The technique also requires easily identifiable ground features which may not be available due to cloud cover or for ocean data. The current project was motivated by the need for a navigation system which was automatic and did not require human intervention or ground control points. The first step in creating such a system must be the ability to parameterize the spacecraft's attitude. The immediate goal of this project was to study the attitude fluctuations and determine if they displayed any systematic behavior which could be modeled or parameterized. We chose a period in 1991-1992 to study the attitude of the NOAA 11 spacecraft using data from the Tiros receiving station at the Colorado Center for Astrodynamic Research (CCAR) at the University of Colorado.

  16. Easy and Accessible Imaging Software

    NASA Technical Reports Server (NTRS)

    2003-01-01

    DATASTAR, Inc., of Picayune, Mississippi, has taken NASA s award-winning Earth Resources Laboratory Applications Software (ELAS) program and evolved it into a user-friendly desktop application and Internet service to perform processing, analysis, and manipulation of remotely sensed imagery data. NASA s Stennis Space Center developed ELAS in the early 1980s to process satellite and airborne sensor imagery data of the Earth s surface into readable and accessible information. Since then, ELAS information has been applied worldwide to determine soil content, rainfall levels, and numerous other variances of topographical information. However, end-users customarily had to depend on scientific or computer experts to provide the results, because the imaging processing system was intricate and labor intensive.

  17. Software for Automated Image-to-Image Co-registration

    NASA Technical Reports Server (NTRS)

    Benkelman, Cody A.; Hughes, Heidi

    2007-01-01

    The project objectives are: a) Develop software to fine-tune image-to-image co-registration, presuming images are orthorectified prior to input; b) Create a reusable software development kit (SDK) to enable incorporation of these tools into other software; d) provide automated testing for quantitative analysis; and e) Develop software that applies multiple techniques to achieve subpixel precision in the co-registration of image pairs.

  18. Workspaces that move people.

    PubMed

    Waber, Ben; Magnolfi, Jennifer; Lindsay, Greg

    2014-10-01

    Few companies measure whether the design of their workspaces helps or hurts performance, but they should. The authors have collected data that capture individuals' interactions, communications, and location information. They've learned that face-to-face interactions are by far the most important activity in an office; creating chance encounters between knowledge workers, both inside and outside the organization, improves performance. The Norwegian telecom company Telenor was ahead of its time in 2003, when it incorporated "hot desking" (no assigned seats) and spaces that could easily be reconfigured for different tasks and evolving teams. The CEO credits the design of the offices with helping Telenor shift from a state-run monopoly to a competitive multinational carrier with 150 million subscribers. In another example, data collected at one pharmaceuticals company showed that when a salesperson increased interactions with coworkers on other teams by 10%, his or her sales increased by 10%. To get the sales staff running into colleagues from other departments, management shifted from one coffee machine for every six employees to one for every 120 and created a new large cafeteria for everyone. Sales rose by 20%, or $200 million, afterjust one quarter, quickly justifying the capital investment in the redesign. PMID:25509577

  19. False Color Terrain Model of Phoenix Workspace

    NASA Technical Reports Server (NTRS)

    2008-01-01

    This is a terrain model of Phoenix's Robotic Arm workspace. It has been color coded by depth with a lander model for context. The model has been derived using images from the depth perception feature from Phoenix's Surface Stereo Imager (SSI). Red indicates low-lying areas that appear to be troughs. Blue indicates higher areas that appear to be polygons.

    The Phoenix Mission is led by the University of Arizona, Tucson, on behalf of NASA. Project management of the mission is by NASA's Jet Propulsion Laboratory, Pasadena, Calif. Spacecraft development is by Lockheed Martin Space Systems, Denver.

  20. Image analysis library software development

    NASA Technical Reports Server (NTRS)

    Guseman, L. F., Jr.; Bryant, J.

    1977-01-01

    The Image Analysis Library consists of a collection of general purpose mathematical/statistical routines and special purpose data analysis/pattern recognition routines basic to the development of image analysis techniques for support of current and future Earth Resources Programs. Work was done to provide a collection of computer routines and associated documentation which form a part of the Image Analysis Library.

  1. Software Operates On Bit-Map Images

    NASA Technical Reports Server (NTRS)

    Choi, Diana

    1992-01-01

    PIXTOOLS is software for Silicon Graphics IRIS consisting of thirteen programs plus library for operating on bit-map images. Enables user to create, edit, and save high-resolution images in forms in which displayed on video screens, resize them, and capture them. Eleven programs print information and read and write files. Two offer graphical interfaces. Menus enable manipulation of images and background color and saving of an image screen to file. Written in C.

  2. Infrared Imaging Data Reduction Software and Techniques

    NASA Astrophysics Data System (ADS)

    Sabbey, C. N.; McMahon, R. G.; Lewis, J. R.; Irwin, M. J.

    Developed to satisfy certain design requirements not met in existing packages (e.g., full weight map handling) and to optimize the software for large data sets (non-interactive tasks that are CPU and disk efficient), the InfraRed Data Reduction software package is a small ANSI C library of fast image processing routines for automated pipeline reduction of infrared (dithered) observations. The software includes stand-alone C programs for tasks such as running sky frame subtraction with object masking, image registration and co-addition with weight maps, dither offset measurement using cross-correlation, and object mask dilation. Although currently used for near-IR mosaic images, the modular software is concise and readily adaptable for reuse in other work. IRDR, available via anonymous ftp at ftp.ast.cam.ac.uk in pub/sabbey

  3. Software for Simulation of Hyperspectral Images

    NASA Technical Reports Server (NTRS)

    Richtsmeier, Steven C.; Singer-Berk, Alexander; Bernstein, Lawrence S.

    2002-01-01

    A package of software generates simulated hyperspectral images for use in validating algorithms that generate estimates of Earth-surface spectral reflectance from hyperspectral images acquired by airborne and spaceborne instruments. This software is based on a direct simulation Monte Carlo approach for modeling three-dimensional atmospheric radiative transport as well as surfaces characterized by spatially inhomogeneous bidirectional reflectance distribution functions. In this approach, 'ground truth' is accurately known through input specification of surface and atmospheric properties, and it is practical to consider wide variations of these properties. The software can treat both land and ocean surfaces and the effects of finite clouds with surface shadowing. The spectral/spatial data cubes computed by use of this software can serve both as a substitute for and a supplement to field validation data.

  4. Video Image Stabilization and Registration (VISAR) Software

    NASA Technical Reports Server (NTRS)

    1999-01-01

    Two scientists at NASA's Marshall Space Flight Center,atmospheric scientist Paul Meyer and solar physicist Dr. David Hathaway, developed promising new software, called Video Image stabilization and Registration (VISAR), which is illustrated in this Quick Time movie. VISAR is a computer algorithm that stabilizes camera motion in the horizontal and vertical as well as rotation and zoom effects producing clearer images of moving objects, smoothes jagged edges, enhances still images, and reduces video noise or snow. It could steady images of ultrasounds, which are infamous for their grainy, blurred quality. VISAR could also have applications in law enforcement, medical, and meteorological imaging. The software can be used for defense application by improving reconnaissance video imagery made by military vehicles, aircraft, and ships traveling in harsh, rugged environments.

  5. Video Image Stabilization and Registration (VISAR) Software

    NASA Technical Reports Server (NTRS)

    1999-01-01

    Two scientists at NASA's Marshall Space Flight Center, atmospheric scientist Paul Meyer and solar physicist Dr. David Hathaway, developed promising new software, called Video Image Stabilization and Registration (VISAR), which is illustrated in this Quick Time movie. VISAR is a computer algorithm that stabilizes camera motion in the horizontal and vertical as well as rotation and zoom effects producing clearer images of moving objects, smoothes jagged edges, enhances still images, and reduces video noise or snow. It could steady images of ultrasounds, which are infamous for their grainy, blurred quality. VISAR could also have applications in law enforcement, medical, and meteorological imaging. The software can be used for defense application by improving reconnaissance video imagery made by military vehicles, aircraft, and ships traveling in harsh, rugged environments.

  6. Video Image Stabilization and Registration (VISAR) Software

    NASA Technical Reports Server (NTRS)

    1999-01-01

    Two scientists at NASA's Marshall Space Flight Center,atmospheric scientist Paul Meyer and solar physicist Dr. David Hathaway, developed promising new software, called Video Image Stabilization and Registration (VISAR). VISAR may help law enforcement agencies catch criminals by improving the quality of video recorded at crime scenes. In this photograph, the single frame at left, taken at night, was brightened in order to enhance details and reduce noise or snow. To further overcome the video defects in one frame, Law enforcement officials can use VISAR software to add information from multiple frames to reveal a person. Images from less than a second of videotape were added together to create the clarified image at right. VISAR stabilizes camera motion in the horizontal and vertical as well as rotation and zoom effects producing clearer images of moving objects, smoothes jagged edges, enhances still images, and reduces video noise or snow. VISAR could also have applications in medical and meteorological imaging. It could steady images of ultrasounds, which are infamous for their grainy, blurred quality. The software can be used for defense application by improving recornaissance video imagery made by military vehicles, aircraft, and ships traveling in harsh, rugged environments.

  7. Video Image Stabilization and Registration (VISAR) Software

    NASA Technical Reports Server (NTRS)

    1999-01-01

    Two scientists at NASA Marshall Space Flight Center, atmospheric scientist Paul Meyer (left) and solar physicist Dr. David Hathaway, have developed promising new software, called Video Image Stabilization and Registration (VISAR), that may help law enforcement agencies to catch criminals by improving the quality of video recorded at crime scenes, VISAR stabilizes camera motion in the horizontal and vertical as well as rotation and zoom effects; produces clearer images of moving objects; smoothes jagged edges; enhances still images; and reduces video noise of snow. VISAR could also have applications in medical and meteorological imaging. It could steady images of Ultrasounds which are infamous for their grainy, blurred quality. It would be especially useful for tornadoes, tracking whirling objects and helping to determine the tornado's wind speed. This image shows two scientists reviewing an enhanced video image of a license plate taken from a moving automobile.

  8. Product review: lucis image processing software.

    PubMed

    Johnson, J E

    1999-04-01

    Lucis is a software program that allows the manipulation of images through the process of selective contrast pattern emphasis. Using an image-processing algorithm called Differential Hysteresis Processing (DHP), Lucis extracts and highlights patterns based on variations in image intensity (luminance). The result is that details can be seen that would otherwise be hidden in deep shadow or excessive brightness. The software is contained on a single floppy disk, is easy to install on a PC, simple to use, and runs on Windows 95, Windows 98, and Windows NT operating systems. The cost is $8,500 for a license, but is estimated to save a great deal of money in photographic materials, time, and labor that would have otherwise been spent in the darkroom. Superb images are easily obtained from unstained (no lead or uranium) sections, and stored image files sent to laser printers are of publication quality. The software can be used not only for all types of microscopy, including color fluorescence light microscopy, biological and materials science electron microscopy (TEM and SEM), but will be beneficial in medicine, such as X-ray films (pending approval by the FDA), and in the arts. PMID:10206154

  9. SUPRIM: easily modified image processing software.

    PubMed

    Schroeter, J P; Bretaudiere, J P

    1996-01-01

    A flexible, modular software package intended for the processing of electron microscopy images is presented. The system consists of a set of image processing tools or filters, written in the C programming language, and a command line style user interface based on the UNIX shell. The pipe and filter structure of UNIX and the availability of command files in the form of shell scripts eases the construction of complex image processing procedures from the simpler tools. Implementation of a new image processing algorithm in SUPRIM may often be performed by construction of a new shell script, using already existing tools. Currently, the package has been used for two- and three-dimensional image processing and reconstruction of macromolecules and other structures of biological interest. PMID:8742734

  10. Salvo: Seismic imaging software for complex geologies

    SciTech Connect

    OBER,CURTIS C.; GJERTSEN,ROB; WOMBLE,DAVID E.

    2000-03-01

    This report describes Salvo, a three-dimensional seismic-imaging software for complex geologies. Regions of complex geology, such as overthrusts and salt structures, can cause difficulties for many seismic-imaging algorithms used in production today. The paraxial wave equation and finite-difference methods used within Salvo can produce high-quality seismic images in these difficult regions. However this approach comes with higher computational costs which have been too expensive for standard production. Salvo uses improved numerical algorithms and methods, along with parallel computing, to produce high-quality images and to reduce the computational and the data input/output (I/O) costs. This report documents the numerical algorithms implemented for the paraxial wave equation, including absorbing boundary conditions, phase corrections, imaging conditions, phase encoding, and reduced-source migration. This report also describes I/O algorithms for large seismic data sets and images and parallelization methods used to obtain high efficiencies for both the computations and the I/O of seismic data sets. Finally, this report describes the required steps to compile, port and optimize the Salvo software, and describes the validation data sets used to help verify a working copy of Salvo.

  11. Objective facial photograph analysis using imaging software.

    PubMed

    Pham, Annette M; Tollefson, Travis T

    2010-05-01

    Facial analysis is an integral part of the surgical planning process. Clinical photography has long been an invaluable tool in the surgeon's practice not only for accurate facial analysis but also for enhancing communication between the patient and surgeon, for evaluating postoperative results, for medicolegal documentation, and for educational and teaching opportunities. From 35-mm slide film to the digital technology of today, clinical photography has benefited greatly from technological advances. With the development of computer imaging software, objective facial analysis becomes easier to perform and less time consuming. Thus, while the original purpose of facial analysis remains the same, the process becomes much more efficient and allows for some objectivity. Although clinical judgment and artistry of technique is never compromised, the ability to perform objective facial photograph analysis using imaging software may become the standard in facial plastic surgery practices in the future. PMID:20511080

  12. Imaging Sensor Flight and Test Equipment Software

    NASA Technical Reports Server (NTRS)

    Freestone, Kathleen; Simeone, Louis; Robertson, Byran; Frankford, Maytha; Trice, David; Wallace, Kevin; Wilkerson, DeLisa

    2007-01-01

    The Lightning Imaging Sensor (LIS) is one of the components onboard the Tropical Rainfall Measuring Mission (TRMM) satellite, and was designed to detect and locate lightning over the tropics. The LIS flight code was developed to run on a single onboard digital signal processor, and has operated the LIS instrument since 1997 when the TRMM satellite was launched. The software provides controller functions to the LIS Real-Time Event Processor (RTEP) and onboard heaters, collects the lightning event data from the RTEP, compresses and formats the data for downlink to the satellite, collects housekeeping data and formats the data for downlink to the satellite, provides command processing and interface to the spacecraft communications and data bus, and provides watchdog functions for error detection. The Special Test Equipment (STE) software was designed to operate specific test equipment used to support the LIS hardware through development, calibration, qualification, and integration with the TRMM spacecraft. The STE software provides the capability to control instrument activation, commanding (including both data formatting and user interfacing), data collection, decompression, and display and image simulation. The LIS STE code was developed for the DOS operating system in the C programming language. Because of the many unique data formats implemented by the flight instrument, the STE software was required to comprehend the same formats, and translate them for the test operator. The hardware interfaces to the LIS instrument using both commercial and custom computer boards, requiring that the STE code integrate this variety into a working system. In addition, the requirement to provide RTEP test capability dictated the need to provide simulations of background image data with short-duration lightning transients superimposed. This led to the development of unique code used to control the location, intensity, and variation above background for simulated lightning strikes

  13. User questionnaire to evaluate the radiological workspace.

    PubMed

    van Ooijen, Peter M A; Koesoema, Allya P; Oudkerk, Matthijs

    2006-01-01

    Over the past few years, an increase in digitalization of radiology departments can be seen, which has a large impact on the work of the radiologists. This impact is not only demonstrated by the increased use of digital images but also by changing demands on the whole reading environment. In this study, we evaluated the satisfaction of our radiologists with our digital Picture Archival and Communication System environment and their workspace. This evaluation was performed by distribution of a questionnaire consisting of a score sheet and some open questions to all radiologists and residents. Out of 25 questionnaires, 12 were adequately answered and returned. Results clearly showed that most problems were present in the area of reading room design and layout and comfort and ergonomics. Based on the results from this study, adaptations were made and the results were also used in the planning of the redesign of the entire department of radiology. PMID:16767350

  14. Workspaces in the Semantic Web

    NASA Technical Reports Server (NTRS)

    Wolfe, Shawn R.; Keller, RIchard M.

    2005-01-01

    Due to the recency and relatively limited adoption of Semantic Web technologies. practical issues related to technology scaling have received less attention than foundational issues. Nonetheless, these issues must be addressed if the Semantic Web is to realize its full potential. In particular, we concentrate on the lack of scoping methods that reduce the size of semantic information spaces so they are more efficient to work with and more relevant to an agent's needs. We provide some intuition to motivate the need for such reduced information spaces, called workspaces, give a formal definition, and suggest possible methods of deriving them.

  15. ICER-3D Hyperspectral Image Compression Software

    NASA Technical Reports Server (NTRS)

    Xie, Hua; Kiely, Aaron; Klimesh, matthew; Aranki, Nazeeh

    2010-01-01

    Software has been developed to implement the ICER-3D algorithm. ICER-3D effects progressive, three-dimensional (3D), wavelet-based compression of hyperspectral images. If a compressed data stream is truncated, the progressive nature of the algorithm enables reconstruction of hyperspectral data at fidelity commensurate with the given data volume. The ICER-3D software is capable of providing either lossless or lossy compression, and incorporates an error-containment scheme to limit the effects of data loss during transmission. The compression algorithm, which was derived from the ICER image compression algorithm, includes wavelet-transform, context-modeling, and entropy coding subalgorithms. The 3D wavelet decomposition structure used by ICER-3D exploits correlations in all three dimensions of sets of hyperspectral image data, while facilitating elimination of spectral ringing artifacts, using a technique summarized in "Improving 3D Wavelet-Based Compression of Spectral Images" (NPO-41381), NASA Tech Briefs, Vol. 33, No. 3 (March 2009), page 7a. Correlation is further exploited by a context-modeling subalgorithm, which exploits spectral dependencies in the wavelet-transformed hyperspectral data, using an algorithm that is summarized in "Context Modeler for Wavelet Compression of Hyperspectral Images" (NPO-43239), which follows this article. An important feature of ICER-3D is a scheme for limiting the adverse effects of loss of data during transmission. In this scheme, as in the similar scheme used by ICER, the spatial-frequency domain is partitioned into rectangular error-containment regions. In ICER-3D, the partitions extend through all the wavelength bands. The data in each partition are compressed independently of those in the other partitions, so that loss or corruption of data from any partition does not affect the other partitions. Furthermore, because compression is progressive within each partition, when data are lost, any data from that partition received

  16. Image processing software for imaging spectrometry data analysis

    NASA Technical Reports Server (NTRS)

    Mazer, Alan; Martin, Miki; Lee, Meemong; Solomon, Jerry E.

    1988-01-01

    Imaging spectrometers simultaneously collect image data in hundreds of spectral channels, from the near-UV to the IR, and can thereby provide direct surface materials identification by means resembling laboratory reflectance spectroscopy. Attention is presently given to a software system, the Spectral Analysis Manager (SPAM) for the analysis of imaging spectrometer data. SPAM requires only modest computational resources and is composed of one main routine and a set of subroutine libraries. Additions and modifications are relatively easy, and special-purpose algorithms have been incorporated that are tailored to geological applications.

  17. IMCAT: Image and Catalogue Manipulation Software

    NASA Astrophysics Data System (ADS)

    Kaiser, Nick

    2011-08-01

    The IMCAT software was developed initially to do faint galaxy photometry for weak lensing studies, and provides a fairly complete set of tools for this kind of work. Unlike most packages for doing data analysis, the tools are standalone unix commands which you can invoke from the shell, via shell scripts or from perl scripts. The tools are arranges in a tree of directories. One main branch is the 'imtools'. These deal only with fits files. The most important imtool is the 'image calculator' 'ic' which allows one to do rather general operations on fits images. A second branch is the 'catools' which operate only on catalogues. The key cattool is 'lc'; this effectively defines the format of IMCAT catalogues, and allows one to do very general operations on and filtering of such catalogues. A third branch is the 'imcattools'. These tend to be much more specialised than the cattools and imcattools and are focussed on faint galaxy photometry.

  18. Collaborative workspace for multimedia medical conferencing.

    PubMed

    Kim, H J; Park, E S; Lee, S G; Shin, Y G

    1998-01-01

    We propose an approach for collaborative workspace management in medical conferencing. A collaborative workspace is a virtual data space shared between medical experts for working out solutions collaboratively while conferencing. Our approach provides medical users with an integrated view of various kinds of multimedia patient data and a unified control over the workspace. For data navigation and conferencing, a tree-like navigation tool, which we named the patient record tree, is provided. And we classify patient data, which is the object of medical collaborative works, into six basic types, and provide a view template for displaying each of these types. PMID:10384472

  19. A Global Workspace perspective on mental disorders

    PubMed Central

    Wallace, Rodrick

    2005-01-01

    Background Recent developments in Global Workspace theory suggest that human consciousness can suffer interpenetrating dysfunctions of mutual and reciprocal interaction with embedding environments which will have early onset and often insidious staged developmental progression, possibly according to a cancer model, in which a set of long-evolved control strategies progressively fails. Methods and results A rate distortion argument implies that, if an external information source carries a damaging 'message', then sufficient exposure to it, particularly during critical developmental periods, is sure to write a sufficiently accurate image of it on mind and body in a punctuated manner so as to initiate or promote similarly progressively punctuated developmental disorder, in essence either a staged failure affecting large-scale brain connectivity, which is the sine qua non of human consciousness, or else damaging the ability of embedding goal contexts to contain conscious dynamics. Conclusion The key intervention, at the population level, is clearly to limit exposure to factors triggering developmental disorders, a question of proper environmental sanitation, in a large sense, primarily a matter of social justice which has long been known to be determined almost entirely by the interactions of cultural trajectory, group power relations, and economic structure, with public policy. Intervention at the individual level appears limited to triggering or extending periods of remission, representing reestablishment of an extensive, but largely unexplored, spectrum of evolved control strategies, in contrast with the far better-understood case of cancer. PMID:16371149

  20. Metamodels for Planar 3R Workspace Optimization.

    SciTech Connect

    Turner, C. J.

    2002-01-01

    Robotic workspace optimization is a central element of robot system design. To formulate the optimization problem, the complex relationships between design variables, tuning parameters, and performance indices need to be accurately and efficiently represented. The nature of the relationships suggests that metamodels, or models of the models, should be used to derive suitable objective functions. A comparison of two metamodeling techniques for robotic workspace optimization problems for several trial cases suggests that non-uniform rational B-spline models, derived from computer graphics and computer-aided design techniques, are as suitable as response surface models to solve planar 3R workspace optimization problems. Promising nonlinear modeling results with B-spline models suggest future work is justified and performance gains can be realized.

  1. Prototype for remotely shared textual workspaces

    SciTech Connect

    Abdel-Wahab, H.M.; Guan, S.U.; Nievergelt, J.

    1987-01-01

    Computer-based collaboration between geographically dispersed users is still limited primarily to electronic mail and file transfer, but there is increasing interest in computer support for real-time interaction between remote users. The problem of implementing remotely shared workspaces, which allow users to operate simultaneously on the same objects, is of broad interest. Our objective is to show textual workspaces that allow real-time collaboration can be implemented efficiently by using existing operating systems and communications primitives. This paper documents our experiences in implementing a prototype under Berkeley UNIX, using the programming language C, and Berkeley Interprocess Communication facilities. The authors describe design alternatives that take into account the communications bandwidth between the different sites of the network, and they introduce an efficient protocol that regulates user access to the shared workspace.

  2. IMAGE information monitoring and applied graphics software environment. Volume 2. Software description

    SciTech Connect

    Hallam, J.W.; Ng, K.B.; Upham, G.L.

    1986-09-01

    The EPRI Information Monitoring and Applied Graphics Environment (IMAGE) system is designed for 'fast proto-typing' of advanced concepts for computer-aided plant operations tools. It is a flexible software system which can be used for rapidly creating, dynamically driving and evaluating advanced operator aid displays. The software is written to be both host computer and graphic device independent.

  3. Using Shared Workspaces in Higher Education.

    ERIC Educational Resources Information Center

    Sikkel, Klaas; Gommer, Lisa; Veen, Jan van der

    2002-01-01

    Evaluates use of BSCW (a groupware system) shared workspaces in higher education by means of a comparison of seven courses in which this environment was used. Identifies different functions for which the BSCW environment has been used and discusses the relative success of these functions across the cases. (AEF)

  4. Earth Observation Services (Image Processing Software)

    NASA Technical Reports Server (NTRS)

    1992-01-01

    San Diego State University and Environmental Systems Research Institute, with other agencies, have applied satellite imaging and image processing techniques to geographic information systems (GIS) updating. The resulting images display land use and are used by a regional planning agency for applications like mapping vegetation distribution and preserving wildlife habitats. The EOCAP program provides government co-funding to encourage private investment in, and to broaden the use of NASA-developed technology for analyzing information about Earth and ocean resources.

  5. Software Helps Extract Information From Astronomical Images

    NASA Technical Reports Server (NTRS)

    Hartley, Booth; Ebert, Rick; Laughlin, Gaylin

    1995-01-01

    PAC Skyview 2.0 is interactive program for display and analysis of astronomical images. Includes large set of functions for display, analysis and manipulation of images. "Man" pages with descriptions of functions and examples of usage included. Skyview used interactively or in "server" mode, in which another program calls Skyview and executes commands itself. Skyview capable of reading image data files of four types, including those in FITS, S, IRAF, and Z formats. Written in C.

  6. Development and implementation of software systems for imaging spectroscopy

    USGS Publications Warehouse

    Boardman, J.W.; Clark, R.N.; Mazer, A.S.; Biehl, L.L.; Kruse, F.A.; Torson, J.; Staenz, K.

    2006-01-01

    Specialized software systems have played a crucial role throughout the twenty-five year course of the development of the new technology of imaging spectroscopy, or hyperspectral remote sensing. By their very nature, hyperspectral data place unique and demanding requirements on the computer software used to visualize, analyze, process and interpret them. Often described as a marriage of the two technologies of reflectance spectroscopy and airborne/spaceborne remote sensing, imaging spectroscopy, in fact, produces data sets with unique qualities, unlike previous remote sensing or spectrometer data. Because of these unique spatial and spectral properties hyperspectral data are not readily processed or exploited with legacy software systems inherited from either of the two parent fields of study. This paper provides brief reviews of seven important software systems developed specifically for imaging spectroscopy.

  7. MaZda--a software package for image texture analysis.

    PubMed

    Szczypiński, Piotr M; Strzelecki, Michał; Materka, Andrzej; Klepaczko, Artur

    2009-04-01

    MaZda, a software package for 2D and 3D image texture analysis is presented. It provides a complete path for quantitative analysis of image textures, including computation of texture features, procedures for feature selection and extraction, algorithms for data classification, various data visualization and image segmentation tools. Initially, MaZda was aimed at analysis of magnetic resonance image textures. However, it revealed its effectiveness in analysis of other types of textured images, including X-ray and camera images. The software was utilized by numerous researchers in diverse applications. It was proven to be an efficient and reliable tool for quantitative image analysis, even in more accurate and objective medical diagnosis. MaZda was also successfully used in food industry to assess food product quality. MaZda can be downloaded for public use from the Institute of Electronics, Technical University of Lodz webpage. PMID:18922598

  8. Software for Viewing Landsat Mosaic Images

    NASA Technical Reports Server (NTRS)

    Watts, Jack; Farve, Catherine L.; Harvey, Craig

    2002-01-01

    A Windows-based computer program has been written to enable novice users (especially educators and students) to view images of large areas of the Earth (e.g., the continental United States) generated from image data acquired in the Landsat observations performed circa the year 1990. The large-area images are constructed as mosaics from the original Landsat images, which were acquired in several wavelength bands and each of which spans an area (in effect, one tile of a mosaic) of 5 in latitude by approximately equal to 6 degrees in longitude. Whereas the original Landsat data are registered on a universal transverse Mercator (UTM) grid, the program converts the UTM coordinates of a mouse pointer in the image to latitude and longitude, which are continuously updated and displayed as the pointer is moved. The mosaic image currently on display can be exported as a Windows bit-map file. Other images (e.g., of state boundaries or interstate highways) can be overlaid on Landsat mosaics. The program interacts with the user via standard toolbar, keyboard, and mouse user interfaces. The program is supplied on a compact disk along with tutorial and educational information.

  9. Software for Viewing Landsat Mosaic Images

    NASA Technical Reports Server (NTRS)

    2002-01-01

    A Windows-based computer program has been written to enable novice users (especially educators and students) to view images of large areas of the Earth (e.g., the continental United States) generated from image data acquired in the Landsat observations performed circa the year 1990. The large-area images are constructed as mosaics from the original Landsat images, which were acquired in several wavelength bands and each of which spans an area (in effect, one tile of a mosaic) of approx. 5 in latitude by approx. 6 deg in longitude. Whereas the original Landsat data are registered on a universal transverse Mercator (UTM) grid, the program converts the UTM coordinates of a mouse pointer in the image to latitude and longitude, which are continuously updated and displayed as the pointer is moved. The mosaic image currently on display can be exported as a Windows bit-map file. Other images (e.g., of state boundaries or interstate highways) can be overlaid on Landsat mosaics. The program interacts with the user via standard toolbar, keyboard, and mouse user interfaces. The program is supplied on a compact disk along with tutorial and educational information.

  10. Software for Viewing Landsat Mosaic Images

    NASA Technical Reports Server (NTRS)

    Watts, Zack; Farve, Catharine L.; Harvey, Craig

    2003-01-01

    A Windows-based computer program has been written to enable novice users (especially educators and students) to view images of large areas of the Earth (e.g., the continental United States) generated from image data acquired in the Landsat observations performed circa the year 1990. The large-area images are constructed as mosaics from the original Landsat images, which were acquired in several wavelength bands and each of which spans an area (in effect, one tile of a mosaic) of .5 in latitude by .6 in longitude. Whereas the original Landsat data are registered on a universal transverse Mercator (UTM) grid, the program converts the UTM coordinates of a mouse pointer in the image to latitude and longitude, which are continuously updated and displayed as the pointer is moved. The mosaic image currently on display can be exported as a Windows bitmap file. Other images (e.g., of state boundaries or interstate highways) can be overlaid on Landsat mosaics. The program interacts with the user via standard toolbar, keyboard, and mouse user interfaces. The program is supplied on a compact disk along with tutorial and educational information.

  11. The image related services of the HELIOS software engineering environment.

    PubMed

    Engelmann, U; Meinzer, H P; Schröter, A; Günnel, U; Demiris, A M; Makabe, M; Evers, H; Jean, F C; Degoulet, P

    1995-01-01

    This paper describes the approach of the European HELIOS project to integrate image processing tools into ward information systems. The image processing tools are the result of the basic research in image analysis in the Department Medical and Biological Informatics at the German Cancer Research Center. These tools for the analysis of two-dimensional images and three-dimensional data volumes with 3D reconstruction and visualization ae part of the Image Related Services of HELIOS. The HELIOS software engineering environment allows to use the image processing functionality in integrated applications. PMID:7743775

  12. Software Graphical User Interface For Analysis Of Images

    NASA Technical Reports Server (NTRS)

    Leonard, Desiree M.; Nolf, Scott R.; Avis, Elizabeth L.; Stacy, Kathryn

    1992-01-01

    CAMTOOL software provides graphical interface between Sun Microsystems workstation and Eikonix Model 1412 digitizing camera system. Camera scans and digitizes images, halftones, reflectives, transmissives, rigid or flexible flat material, or three-dimensional objects. Users digitize images and select from three destinations: work-station display screen, magnetic-tape drive, or hard disk. Written in C.

  13. FITSH: Software Package for Image Processing

    NASA Astrophysics Data System (ADS)

    Pál, András

    2011-11-01

    FITSH provides a standalone environment for analysis of data acquired by imaging astronomical detectors. The package provides utilities both for the full pipeline of subsequent related data processing steps (including image calibration, astrometry, source identification, photometry, differential analysis, low-level arithmetic operations, multiple image combinations, spatial transformations and interpolations, etc.) and for aiding the interpretation of the (mainly photometric and/or astrometric) results. The package also features a consistent implementation of photometry based on image subtraction, point spread function fitting and aperture photometry and provides easy-to-use interfaces for comparisons and for picking the most suitable method for a particular problem. The utilities in the package are built on the top of the commonly used UNIX/POSIX shells (hence the name of the package), therefore both frequently used and well-documented tools for such environments can be exploited and managing massive amount of data is rather convenient.

  14. gr-MRI: A software package for magnetic resonance imaging using software defined radios

    NASA Astrophysics Data System (ADS)

    Hasselwander, Christopher J.; Cao, Zhipeng; Grissom, William A.

    2016-09-01

    The goal of this work is to develop software that enables the rapid implementation of custom MRI spectrometers using commercially-available software defined radios (SDRs). The developed gr-MRI software package comprises a set of Python scripts, flowgraphs, and signal generation and recording blocks for GNU Radio, an open-source SDR software package that is widely used in communications research. gr-MRI implements basic event sequencing functionality, and tools for system calibrations, multi-radio synchronization, and MR signal processing and image reconstruction. It includes four pulse sequences: a single-pulse sequence to record free induction signals, a gradient-recalled echo imaging sequence, a spin echo imaging sequence, and an inversion recovery spin echo imaging sequence. The sequences were used to perform phantom imaging scans with a 0.5 Tesla tabletop MRI scanner and two commercially-available SDRs. One SDR was used for RF excitation and reception, and the other for gradient pulse generation. The total SDR hardware cost was approximately 2000. The frequency of radio desynchronization events and the frequency with which the software recovered from those events was also measured, and the SDR's ability to generate frequency-swept RF waveforms was validated and compared to the scanner's commercial spectrometer. The spin echo images geometrically matched those acquired using the commercial spectrometer, with no unexpected distortions. Desynchronization events were more likely to occur at the very beginning of an imaging scan, but were nearly eliminated if the user invoked the sequence for a short period before beginning data recording. The SDR produced a 500 kHz bandwidth frequency-swept pulse with high fidelity, while the commercial spectrometer produced a waveform with large frequency spike errors. In conclusion, the developed gr-MRI software can be used to develop high-fidelity, low-cost custom MRI spectrometers using commercially-available SDRs.

  15. gr-MRI: A software package for magnetic resonance imaging using software defined radios.

    PubMed

    Hasselwander, Christopher J; Cao, Zhipeng; Grissom, William A

    2016-09-01

    The goal of this work is to develop software that enables the rapid implementation of custom MRI spectrometers using commercially-available software defined radios (SDRs). The developed gr-MRI software package comprises a set of Python scripts, flowgraphs, and signal generation and recording blocks for GNU Radio, an open-source SDR software package that is widely used in communications research. gr-MRI implements basic event sequencing functionality, and tools for system calibrations, multi-radio synchronization, and MR signal processing and image reconstruction. It includes four pulse sequences: a single-pulse sequence to record free induction signals, a gradient-recalled echo imaging sequence, a spin echo imaging sequence, and an inversion recovery spin echo imaging sequence. The sequences were used to perform phantom imaging scans with a 0.5Tesla tabletop MRI scanner and two commercially-available SDRs. One SDR was used for RF excitation and reception, and the other for gradient pulse generation. The total SDR hardware cost was approximately $2000. The frequency of radio desynchronization events and the frequency with which the software recovered from those events was also measured, and the SDR's ability to generate frequency-swept RF waveforms was validated and compared to the scanner's commercial spectrometer. The spin echo images geometrically matched those acquired using the commercial spectrometer, with no unexpected distortions. Desynchronization events were more likely to occur at the very beginning of an imaging scan, but were nearly eliminated if the user invoked the sequence for a short period before beginning data recording. The SDR produced a 500kHz bandwidth frequency-swept pulse with high fidelity, while the commercial spectrometer produced a waveform with large frequency spike errors. In conclusion, the developed gr-MRI software can be used to develop high-fidelity, low-cost custom MRI spectrometers using commercially-available SDRs. PMID:27394165

  16. MOSAIC: Software for creating mosaics from collections of images

    NASA Technical Reports Server (NTRS)

    Varosi, F.; Gezari, D. Y.

    1992-01-01

    We have developed a powerful, versatile image processing and analysis software package called MOSAIC, designed specifically for the manipulation of digital astronomical image data obtained with (but not limited to) two-dimensional array detectors. The software package is implemented using the Interactive Data Language (IDL), and incorporates new methods for processing, calibration, analysis, and visualization of astronomical image data, stressing effective methods for the creation of mosaic images from collections of individual exposures, while at the same time preserving the photometric integrity of the original data. Since IDL is available on many computers, the MOSAIC software runs on most UNIX and VAX workstations with the X-Windows or Sun View graphics interface.

  17. MOPEX: a software package for astronomical image processing and visualization

    NASA Astrophysics Data System (ADS)

    Makovoz, David; Roby, Trey; Khan, Iffat; Booth, Hartley

    2006-06-01

    We present MOPEX - a software package for astronomical image processing and display. The package is a combination of command-line driven image processing software written in C/C++ with a Java-based GUI. The main image processing capabilities include creating mosaic images, image registration, background matching, point source extraction, as well as a number of minor image processing tasks. The combination of the image processing and display capabilities allows for much more intuitive and efficient way of performing image processing. The GUI allows for the control over the image processing and display to be closely intertwined. Parameter setting, validation, and specific processing options are entered by the user through a set of intuitive dialog boxes. Visualization feeds back into further processing by providing a prompt feedback of the processing results. The GUI also allows for further analysis by accessing and displaying data from existing image and catalog servers using a virtual observatory approach. Even though originally designed for the Spitzer Space Telescope mission, a lot of functionalities are of general usefulness and can be used for working with existing astronomical data and for new missions. The software used in the package has undergone intensive testing and benefited greatly from effective software reuse. The visualization part has been used for observation planning for both the Spitzer and Herschel Space Telescopes as part the tool Spot. The visualization capabilities of Spot have been enhanced and integrated with the image processing functionality of the command-line driven MOPEX. The image processing software is used in the Spitzer automated pipeline processing, which has been in operation for nearly 3 years. The image processing capabilities have also been tested in off-line processing by numerous astronomers at various institutions around the world. The package is multi-platform and includes automatic update capabilities. The software

  18. Image-Processing Software For A Hypercube Computer

    NASA Technical Reports Server (NTRS)

    Lee, Meemong; Mazer, Alan S.; Groom, Steven L.; Williams, Winifred I.

    1992-01-01

    Concurrent Image Processing Executive (CIPE) is software system intended to develop and use image-processing application programs on concurrent computing environment. Designed to shield programmer from complexities of concurrent-system architecture, it provides interactive image-processing environment for end user. CIPE utilizes architectural characteristics of particular concurrent system to maximize efficiency while preserving architectural independence from user and programmer. CIPE runs on Mark-IIIfp 8-node hypercube computer and associated SUN-4 host computer.

  19. Uses of software in digital image analysis: a forensic report

    NASA Astrophysics Data System (ADS)

    Sharma, Mukesh; Jha, Shailendra

    2010-02-01

    Forensic image analysis is required an expertise to interpret the content of an image or the image itself in legal matters. Major sub-disciplines of forensic image analysis with law enforcement applications include photo-grammetry, photographic comparison, content analysis and image authentication. It has wide applications in forensic science range from documenting crime scenes to enhancing faint or indistinct patterns such as partial fingerprints. The process of forensic image analysis can involve several different tasks, regardless of the type of image analysis performed. Through this paper authors have tried to explain these tasks, which are described in to three categories: Image Compression, Image Enhancement & Restoration and Measurement Extraction. With the help of examples like signature comparison, counterfeit currency comparison and foot-wear sole impression using the software Canvas and Corel Draw.

  20. NASA's MERBoard: An Interactive Collaborative Workspace Platform. Chapter 4

    NASA Technical Reports Server (NTRS)

    Trimble, Jay; Wales, Roxana; Gossweiler, Rich

    2003-01-01

    This chapter describes the ongoing process by which a multidisciplinary group at NASA's Ames Research Center is designing and implementing a large interactive work surface called the MERBoard Collaborative Workspace. A MERBoard system involves several distributed, large, touch-enabled, plasma display systems with custom MERBoard software. A centralized server and database back the system. We are continually tuning MERBoard to support over two hundred scientists and engineers during the surface operations of the Mars Exploration Rover Missions. These scientists and engineers come from various disciplines and are working both in small and large groups over a span of space and time. We describe the multidisciplinary, human-centered process by which this h4ERBoard system is being designed, the usage patterns and social interactions that we have observed, and issues we are currently facing.

  1. Performance of Personal Workspace Controls Final Report

    SciTech Connect

    Rubinstein, Francis; Kiliccote, Sila; Loffeld, John; Pettler,Pete; Snook, Joel

    2004-12-01

    One of the key deliverables for the DOE-funded controls research at LBNL for FY04 was the development of a prototype Personal Workspace Control system. The successful development of this system is a critical milestone for the LBNL Lighting Controls Research effort because this system demonstrates how IBECS can add value to today's Task Ambient lighting systems. LBNL has argued that by providing both the occupant and the facilities manager with the ability to precisely control the operation of overhead lighting and all task lighting in a coordinated manner, that task ambient lighting can optimize energy performance and occupant comfort simultaneously [Reference Task Ambient Foundation Document]. The Personal Workspace Control system is the application of IBECS to this important lighting problem. This report discusses the development of the Personal Workspace Control to date including descriptions of the different fixture types that have been converted to IBECS operation and a detailed description of the operation of PWC Scene Controller, which provides the end user with precise control of his task ambient lighting system. The objective, from the Annual Plan, is to demonstrate improvements in efficiency, lighting quality and occupant comfort realized using Personal Workspace Controls (PWC) designed to optimize the delivery of lighting to the individual's workstation regardless of which task-ambient lighting solution is chosen. The PWC will be capable of controlling floor-mounted, desk lamps, furniture-mounted and overhead lighting fixtures from a personal computer and handheld remote. The PWC will use an environmental sensor to automatically monitor illuminance, temperature and occupancy and to appropriately modulate ambient lighting according to daylight availability and to switch off task lighting according to local occupancy. [Adding occupancy control to the system would blunt the historical criticism of occupant-controlled lighting - the tendency of the occupant

  2. Single-molecule localization software applied to photon counting imaging.

    PubMed

    Hirvonen, Liisa M; Kilfeather, Tiffany; Suhling, Klaus

    2015-06-01

    Centroiding in photon counting imaging has traditionally been accomplished by a single-step, noniterative algorithm, often implemented in hardware. Single-molecule localization techniques in superresolution fluorescence microscopy are conceptually similar, but use more sophisticated iterative software-based fitting algorithms to localize the fluorophore. Here, we discuss common features and differences between single-molecule localization and photon counting imaging and investigate the suitability of single-molecule localization software for photon event localization. We find that single-molecule localization software packages designed for superresolution microscopy-QuickPALM, rapidSTORM, and ThunderSTORM-can work well when applied to photon counting imaging with a microchannel-plate-based intensified camera system: photon event recognition can be excellent, fixed pattern noise can be low, and the microchannel plate pores can easily be resolved. PMID:26192667

  3. Modified control software for imaging ultracold atomic clouds

    NASA Astrophysics Data System (ADS)

    Whitaker, D. L.; Sharma, A.; Brown, J. M.

    2006-12-01

    A charge-coupled device (CCD) camera capable of taking high-quality images of ultracold atomic samples can often represent a significant portion of the equipment costs in atom trapping experiment. We have modified the commercial control software of a CCD camera designed for astronomical imaging to take absorption images of ultracold rubidium clouds. This camera is sensitive at 780 nm and has been modified to take three successive 16-bit images at full resolution. The control software can be integrated into a Matlab graphical user interface with fitting routines written as Matlab functions. This camera is capable of recording high-quality images at a fraction of the cost of similar cameras typically used in atom trapping experiments.

  4. Parallel algorithm for computing 3-D reachable workspaces

    NASA Astrophysics Data System (ADS)

    Alameldin, Tarek K.; Sobh, Tarek M.

    1992-03-01

    The problem of computing the 3-D workspace for redundant articulated chains has applications in a variety of fields such as robotics, computer aided design, and computer graphics. The computational complexity of the workspace problem is at least NP-hard. The recent advent of parallel computers has made practical solutions for the workspace problem possible. Parallel algorithms for computing the 3-D workspace for redundant articulated chains with joint limits are presented. The first phase of these algorithms computes workspace points in parallel. The second phase uses workspace points that are computed in the first phase and fits a 3-D surface around the volume that encompasses the workspace points. The second phase also maps the 3- D points into slices, uses region filling to detect the holes and voids in the workspace, extracts the workspace boundary points by testing the neighboring cells, and tiles the consecutive contours with triangles. The proposed algorithms are efficient for computing the 3-D reachable workspace for articulated linkages, not only those with redundant degrees of freedom but also those with joint limits.

  5. Analyzing huge pathology images with open source software

    PubMed Central

    2013-01-01

    Background Digital pathology images are increasingly used both for diagnosis and research, because slide scanners are nowadays broadly available and because the quantitative study of these images yields new insights in systems biology. However, such virtual slides build up a technical challenge since the images occupy often several gigabytes and cannot be fully opened in a computer’s memory. Moreover, there is no standard format. Therefore, most common open source tools such as ImageJ fail at treating them, and the others require expensive hardware while still being prohibitively slow. Results We have developed several cross-platform open source software tools to overcome these limitations. The NDPITools provide a way to transform microscopy images initially in the loosely supported NDPI format into one or several standard TIFF files, and to create mosaics (division of huge images into small ones, with or without overlap) in various TIFF and JPEG formats. They can be driven through ImageJ plugins. The LargeTIFFTools achieve similar functionality for huge TIFF images which do not fit into RAM. We test the performance of these tools on several digital slides and compare them, when applicable, to standard software. A statistical study of the cells in a tissue sample from an oligodendroglioma was performed on an average laptop computer to demonstrate the efficiency of the tools. Conclusions Our open source software enables dealing with huge images with standard software on average computers. They are cross-platform, independent of proprietary libraries and very modular, allowing them to be used in other open source projects. They have excellent performance in terms of execution speed and RAM requirements. They open promising perspectives both to the clinician who wants to study a single slide and to the research team or data centre who do image analysis of many slides on a computer cluster. Virtual slides The virtual slide(s) for this article can be found here: http

  6. Computer Software Configuration Item-Specific Flight Software Image Transfer Script Generator

    NASA Technical Reports Server (NTRS)

    Bolen, Kenny; Greenlaw, Ronald

    2010-01-01

    A K-shell UNIX script enables the International Space Station (ISS) Flight Control Team (FCT) operators in NASA s Mission Control Center (MCC) in Houston to transfer an entire or partial computer software configuration item (CSCI) from a flight software compact disk (CD) to the onboard Portable Computer System (PCS). The tool is designed to read the content stored on a flight software CD and generate individual CSCI transfer scripts that are capable of transferring the flight software content in a given subdirectory on the CD to the scratch directory on the PCS. The flight control team can then transfer the flight software from the PCS scratch directory to the Electronically Erasable Programmable Read Only Memory (EEPROM) of an ISS Multiplexer/ Demultiplexer (MDM) via the Indirect File Transfer capability. The individual CSCI scripts and the CSCI Specific Flight Software Image Transfer Script Generator (CFITSG), when executed a second time, will remove all components from their original execution. The tool will identify errors in the transfer process and create logs of the transferred software for the purposes of configuration management.

  7. Software development for a Ring Imaging Detector

    NASA Astrophysics Data System (ADS)

    Torisky, Benjamin; Benmokhtar, Fatiha

    2015-04-01

    Jefferson Lab (Jlab) is performing a large-scale upgrade to their Continuous Electron Beam Accelerator Facility (CEBAF) up to 12 GeV beam. The Large Acceptance Spectrometer (CLAS12) in Hall B is being upgraded and a new Ring Imaging CHerenkov (RICH) detector is being developed to provide better kaon - pion separation throughout the 3 to 12 GeV range. With this addition, when the electron beam hits the target, the resulting pions, kaons, and other particles will pass through a wall of translucent aerogel tiles and create Cherenkov radiation. This light can then be accurately detected by a large array of Multi-Anode PhotoMultiplier Tubes (MA-PMT). I am presenting my work on the implementation of Java based reconstruction programs for the RICH in the CLAS12 main analysis package.

  8. Software Development for Ring Imaging Detector

    NASA Astrophysics Data System (ADS)

    Torisky, Benjamin

    2016-03-01

    Jefferson Lab (Jlab) is performing a large-scale upgrade to their Continuous Electron Beam Accelerator Facility (CEBAF) up to 12GeV beam. The Large Acceptance Spectrometer (CLAS12) in Hall B is being upgraded and a new Ring Imaging Cherenkov (RICH) detector is being developed to provide better kaon - pion separation throughout the 3 to 12 GeV range. With this addition, when the electron beam hits the target, the resulting pions, kaons, and other particles will pass through a wall of translucent aerogel tiles and create Cherenkov radiation. This light can then be accurately detected by a large array of Multi-Anode PhotoMultiplier Tubes (MA-PMT). I am presenting an update on my work on the implementation of Java based reconstruction programs for the RICH in the CLAS12 main analysis package.

  9. Software to model AXAF image quality

    NASA Technical Reports Server (NTRS)

    Ahmad, Anees

    1993-01-01

    This draft final report describes the work performed under this delivery order from May 1992 through June 1993. The purpose of this contract was to enhance and develop an integrated optical performance modeling software for complex x-ray optical systems such as AXAF. The GRAZTRACE program developed by the MSFC Optical Systems Branch for modeling VETA-I was used as the starting baseline program. The original program was a large single file program and, therefore, could not be modified very efficiently. The original source code has been reorganized, and a 'Make Utility' has been written to update the original program. The new version of the source code consists of 36 small source files to make it easier for the code developer to manage and modify the program. A user library has also been built and a 'Makelib' utility has been furnished to update the library. With the user library, the users can easily access the GRAZTRACE source files and build a custom library. A user manual for the new version of GRAZTRACE has been compiled. The plotting capability for the 3-D point spread functions and contour plots has been provided in the GRAZTRACE using the graphics package DISPLAY. The Graphics emulator over the network has been set up for programming the graphics routine. The point spread function and the contour plot routines have also been modified to display the plot centroid, and to allow the user to specify the plot range, and the viewing angle options. A Command Mode version of GRAZTRACE has also been developed. More than 60 commands have been implemented in a Code-V like format. The functions covered in this version include data manipulation, performance evaluation, and inquiry and setting of internal parameters. The user manual for these commands has been formatted as in Code-V, showing the command syntax, synopsis, and options. An interactive on-line help system for the command mode has also been accomplished to allow the user to find valid commands, command syntax

  10. Open environment for image processing and software development

    NASA Astrophysics Data System (ADS)

    Rasure, John R.; Young, Mark

    1992-04-01

    The main goal of the Khoros software project is to create and provide an integrated software development environment for information processing and data visualization. The Khoros software system is now being used as a foundation to improve productivity and promote software reuse in a wide variety of application domain. A powerful feature of the Khoros system is the high-level, abstract visual language that can be employed to significantly boost the productivity of the researcher. Central to the Khoros system is the need for a consistent yet flexible user interface development system that provides cohesiveness to the vast number of programs that make up the Khoros system. Automated tools assist in maintenance as well as development of programs. The software structure that embodies this system provides for extensibility and portability, and allows for easy tailoring to target specific application domains and processing environments. First, an overview of the Khoros software environment is given. Then this paper presents the abstract applications programmer interface, API, the data services that are provided in Khoros to support it, and the Khoros visualization and image file format. The authors contend that Khoros is an excellent environment for the exploration and implementation of imaging standards.

  11. Non-Imaging Software/Data Analysis Requirements

    NASA Technical Reports Server (NTRS)

    1984-01-01

    The analysis software needs of the non-imaging planetary data user are discussed. Assumptions as to the nature of the planetary science data centers where the data are physically stored are advanced, the scope of the non-imaging data is outlined, and facilities that users are likely to need to define and access data are identified. Data manipulation and analysis needs and display graphics are discussed.

  12. Software to model AXAF-I image quality

    NASA Technical Reports Server (NTRS)

    Ahmad, Anees; Feng, Chen

    1995-01-01

    A modular user-friendly computer program for the modeling of grazing-incidence type x-ray optical systems has been developed. This comprehensive computer software GRAZTRACE covers the manipulation of input data, ray tracing with reflectivity and surface deformation effects, convolution with x-ray source shape, and x-ray scattering. The program also includes the capabilities for image analysis, detector scan modeling, and graphical presentation of the results. A number of utilities have been developed to interface the predicted Advanced X-ray Astrophysics Facility-Imaging (AXAF-I) mirror structural and thermal distortions with the ray-trace. This software is written in FORTRAN 77 and runs on a SUN/SPARC station. An interactive command mode version and a batch mode version of the software have been developed.

  13. SIMA: Python software for analysis of dynamic fluorescence imaging data

    PubMed Central

    Kaifosh, Patrick; Zaremba, Jeffrey D.; Danielson, Nathan B.; Losonczy, Attila

    2014-01-01

    Fluorescence imaging is a powerful method for monitoring dynamic signals in the nervous system. However, analysis of dynamic fluorescence imaging data remains burdensome, in part due to the shortage of available software tools. To address this need, we have developed SIMA, an open source Python package that facilitates common analysis tasks related to fluorescence imaging. Functionality of this package includes correction of motion artifacts occurring during in vivo imaging with laser-scanning microscopy, segmentation of imaged fields into regions of interest (ROIs), and extraction of signals from the segmented ROIs. We have also developed a graphical user interface (GUI) for manual editing of the automatically segmented ROIs and automated registration of ROIs across multiple imaging datasets. This software has been designed with flexibility in mind to allow for future extension with different analysis methods and potential integration with other packages. Software, documentation, and source code for the SIMA package and ROI Buddy GUI are freely available at http://www.losonczylab.org/sima/. PMID:25295002

  14. The Khoros software development environment for image and signal processing.

    PubMed

    Konstantinides, K; Rasure, J R

    1994-01-01

    Data flow visual language systems allow users to graphically create a block diagram of their applications and interactively control input, output, and system variables. Khoros is an integrated software development environment for information processing and visualization. It is particularly attractive for image processing because of its rich collection of tools for image and digital signal processing. This paper presents a general overview of Khoros with emphasis on its image processing and DSP tools. Various examples are presented and the future direction of Khoros is discussed. PMID:18291923

  15. Stromatoporoid biometrics using image analysis software: A first order approach

    NASA Astrophysics Data System (ADS)

    Wolniewicz, Pawel

    2010-04-01

    Strommetric is a new image analysis computer program that performs morphometric measurements of stromatoporoid sponges. The program measures 15 features of skeletal elements (pillars and laminae) visible in both longitudinal and transverse thin sections. The software is implemented in C++, using the Open Computer Vision (OpenCV) library. The image analysis system distinguishes skeletal elements from sparry calcite using Otsu's method for image thresholding. More than 150 photos of thin sections were used as a test set, from which 36,159 measurements were obtained. The software provided about one hundred times more data than the current method applied until now. The data obtained are reproducible, even if the work is repeated by different workers. Thus the method makes the biometric studies of stromatoporoids objective.

  16. [Utility of noise addition image made by using water phantom and image addition and subtraction software].

    PubMed

    Watanabe, Ryo; Ogawa, Masato; Mituzono, Hiroki; Aoki, Takahiro; Hayano, Mizuho; Watanabe, Yuka

    2010-08-20

    In optimizing exposures, it is very important to evaluate the impact of image noise on image quality. To realize this, there is a need to evaluate how much image noise will make the subject disease invisible. But generally it is very difficult to shoot images of different quality in a clinical examination. Thus, a method to create a noise addition image by adding the image noise to raw data has been reported. However, this approach requires a special system, so it is difficult to implement in many facilities. We have invented a method to easily create a noise addition image by using the water phantom and image add-subtract software that accompanies the device. To create a noise addition image, first we made a noise image by subtracting the water phantom with different SD. A noise addition image was then created by adding the noise image to the original image. By using this method, a simulation image with intergraded SD can be created from the original. Moreover, the noise frequency component of the created noise addition image is as same as the real image. Thus, the relationship of image quality to SD in the clinical image can be evaluated. Although this method is an easy method of LDSI creation on image data, a noise addition image can be easily created by using image addition and subtraction software and water phantom, and this can be implemented in many facilities. PMID:20953102

  17. Image Fusion Software in the Clearpem-Sonic Project

    NASA Astrophysics Data System (ADS)

    Pizzichemi, M.; di Vara, N.; Cucciati, G.; Ghezzi, A.; Paganoni, M.; Farina, F.; Frisch, B.; Bugalho, R.

    2012-08-01

    ClearPEM-Sonic is a mammography scanner that combines Positron Emission Tomography with 3D ultrasound echographic and elastographic imaging. It has been developed to improve early stage detection of breast cancer by combining metabolic and anatomical information. The PET system has been developed by the Crystal Clear Collaboration, while the 3D ultrasound probe has been provided by SuperSonic Imagine. In this framework, the visualization and fusion software is an essential tool for the radiologists in the diagnostic process. This contribution discusses the design choices, the issues faced during the implementation, and the commissioning of the software tools developed for ClearPEM-Sonic.

  18. The application of image processing software: Photoshop in environmental design

    NASA Astrophysics Data System (ADS)

    Dong, Baohua; Zhang, Chunmi; Zhuo, Chen

    2011-02-01

    In the process of environmental design and creation, the design sketch holds a very important position in that it not only illuminates the design's idea and concept but also shows the design's visual effects to the client. In the field of environmental design, computer aided design has made significant improvement. Many types of specialized design software for environmental performance of the drawings and post artistic processing have been implemented. Additionally, with the use of this software, working efficiency has greatly increased and drawings have become more specific and more specialized. By analyzing the application of photoshop image processing software in environmental design and comparing and contrasting traditional hand drawing and drawing with modern technology, this essay will further explore the way for computer technology to play a bigger role in environmental design.

  19. Software for visualization, analysis, and manipulation of laser scan images

    NASA Astrophysics Data System (ADS)

    Burnsides, Dennis B.

    1997-03-01

    The recent introduction of laser surface scanning to scientific applications presents a challenge to computer scientists and engineers. Full utilization of this two- dimensional (2-D) and three-dimensional (3-D) data requires advances in techniques and methods for data processing and visualization. This paper explores the development of software to support the visualization, analysis and manipulation of laser scan images. Specific examples presented are from on-going efforts at the Air Force Computerized Anthropometric Research and Design (CARD) Laboratory.

  20. Towards a cognitive neuroscience of consciousness: basic evidence and a workspace framework.

    PubMed

    Dehaene, S; Naccache, L

    2001-04-01

    brain-imaging data strongly argue for a major role of prefrontal cortex, anterior cingulate, and the areas that connect to them, in creating the postulated brain-scale workspace. PMID:11164022

  1. Parallel-Processing Software for Creating Mosaic Images

    NASA Technical Reports Server (NTRS)

    Klimeck, Gerhard; Deen, Robert; McCauley, Michael; DeJong, Eric

    2008-01-01

    A computer program implements parallel processing for nearly real-time creation of panoramic mosaics of images of terrain acquired by video cameras on an exploratory robotic vehicle (e.g., a Mars rover). Because the original images are typically acquired at various camera positions and orientations, it is necessary to warp the images into the reference frame of the mosaic before stitching them together to create the mosaic. [Also see "Parallel-Processing Software for Correlating Stereo Images," Software Supplement to NASA Tech Briefs, Vol. 31, No. 9 (September 2007) page 26.] The warping algorithm in this computer program reflects the considerations that (1) for every pixel in the desired final mosaic, a good corresponding point must be found in one or more of the original images and (2) for this purpose, one needs a good mathematical model of the cameras and a good correlation of individual pixels with respect to their positions in three dimensions. The desired mosaic is divided into slices, each of which is assigned to one of a number of central processing units (CPUs) operating simultaneously. The results from the CPUs are gathered and placed into the final mosaic. The time taken to create the mosaic depends upon the number of CPUs, the speed of each CPU, and whether a local or a remote data-staging mechanism is used.

  2. Designing multistatic ultrasound imaging systems using software analysis

    NASA Astrophysics Data System (ADS)

    Lee, Michael; Singh, Rahul S.; Culjat, Martin O.; Stubbs, Scott; Natarajan, Shyam; Brown, Elliott R.; Grundfest, Warren S.; Lee, Hua

    2010-03-01

    This paper describes the method of using the finite-element analysis software, PZFlex, to direct the design of a novel ultrasound imaging system which uses conformal transducer arrays. Current challenges in ultrasound array technology, including 2D array processing, have motivated exploration into new data acquisition and reconstruction techniques. Ultimately, these efforts encourage a broader examination of the processes used to effectively validate new array configurations and image formation procedures. Commercial software available today is capable of efficiently and accurately modeling detailed operational aspects of customized arrays. Combining quality simulated data with prototyped reconstruction techniques presents a valuable tool for testing novel schemes before committing more costly resources. To investigate this practice, we modeled three 1D ultrasound arrays operating multistatically instead of by the conventional phased-array approach. They are: a simple linear array, a half-circle array with 180-degree coverage, and a full circular array for inward imaging. We present the process used to create unique array models in PZFlex, simulate operation and obtain data, and subsequently generate images by inputting data into a reconstruction algorithm in MATLAB. Further discussion describes the tested reconstruction algorithm and includes resulting images.

  3. Software components for medical image visualization and surgical planning

    NASA Astrophysics Data System (ADS)

    Starreveld, Yves P.; Gobbi, David G.; Finnis, Kirk; Peters, Terence M.

    2001-05-01

    Purpose: The development of new applications in medical image visualization and surgical planning requires the completion of many common tasks such as image reading and re-sampling, segmentation, volume rendering, and surface display. Intra-operative use requires an interface to a tracking system and image registration, and the application requires basic, easy to understand user interface components. Rapid changes in computer and end-application hardware, as well as in operating systems and network environments make it desirable to have a hardware and operating system as an independent collection of reusable software components that can be assembled rapidly to prototype new applications. Methods: Using the OpenGL based Visualization Toolkit as a base, we have developed a set of components that implement the above mentioned tasks. The components are written in both C++ and Python, but all are accessible from Python, a byte compiled scripting language. The components have been used on the Red Hat Linux, Silicon Graphics Iris, Microsoft Windows, and Apple OS X platforms. Rigorous object-oriented software design methods have been applied to ensure hardware independence and a standard application programming interface (API). There are components to acquire, display, and register images from MRI, MRA, CT, Computed Rotational Angiography (CRA), Digital Subtraction Angiography (DSA), 2D and 3D ultrasound, video and physiological recordings. Interfaces to various tracking systems for intra-operative use have also been implemented. Results: The described components have been implemented and tested. To date they have been used to create image manipulation and viewing tools, a deep brain functional atlas, a 3D ultrasound acquisition and display platform, a prototype minimally invasive robotic coronary artery bypass graft planning system, a tracked neuro-endoscope guidance system and a frame-based stereotaxy neurosurgery planning tool. The frame-based stereotaxy module has been

  4. Woods Hole Image Processing System Software implementation; using NetCDF as a software interface for image processing

    USGS Publications Warehouse

    Paskevich, Valerie F.

    1992-01-01

    The Branch of Atlantic Marine Geology has been involved in the collection, processing and digital mosaicking of high, medium and low-resolution side-scan sonar data during the past 6 years. In the past, processing and digital mosaicking has been accomplished with a dedicated, shore-based computer system. With the need to process sidescan data in the field with increased power and reduced cost of major workstations, a need to have an image processing package on a UNIX based computer system which could be utilized in the field as well as be more generally available to Branch personnel was identified. This report describes the initial development of that package referred to as the Woods Hole Image Processing System (WHIPS). The software was developed using the Unidata NetCDF software interface to allow data to be more readily portable between different computer operating systems.

  5. Development of Software to Model AXAF-I Image Quality

    NASA Technical Reports Server (NTRS)

    Ahmad, Anees; Hawkins, Lamar

    1996-01-01

    This draft final report describes the work performed under the delivery order number 145 from May 1995 through August 1996. The scope of work included a number of software development tasks for the performance modeling of AXAF-I. A number of new capabilities and functions have been added to the GT software, which is the command mode version of the GRAZTRACE software, originally developed by MSFC. A structural data interface has been developed for the EAL (old SPAR) finite element analysis FEA program, which is being used by MSFC Structural Analysis group for the analysis of AXAF-I. This interface utility can read the structural deformation file from the EAL and other finite element analysis programs such as NASTRAN and COSMOS/M, and convert the data to a suitable format that can be used for the deformation ray-tracing to predict the image quality for a distorted mirror. There is a provision in this utility to expand the data from finite element models assuming 180 degrees symmetry. This utility has been used to predict image characteristics for the AXAF-I HRMA, when subjected to gravity effects in the horizontal x-ray ground test configuration. The development of the metrology data processing interface software has also been completed. It can read the HDOS FITS format surface map files, manipulate and filter the metrology data, and produce a deformation file, which can be used by GT for ray tracing for the mirror surface figure errors. This utility has been used to determine the optimum alignment (axial spacing and clocking) for the four pairs of AXAF-I mirrors. Based on this optimized alignment, the geometric images and effective focal lengths for the as built mirrors were predicted to cross check the results obtained by Kodak.

  6. Software for Verifying Image-Correlation Tie Points

    NASA Technical Reports Server (NTRS)

    Klimeck, Gerhard; Yagi, Gary

    2008-01-01

    A computer program enables assessment of the quality of tie points in the image-correlation processes of the software described in the immediately preceding article. Tie points are computed in mappings between corresponding pixels in the left and right images of a stereoscopic pair. The mappings are sometimes not perfect because image data can be noisy and parallax can cause some points to appear in one image but not the other. The present computer program relies on the availability of a left- right correlation map in addition to the usual right left correlation map. The additional map must be generated, which doubles the processing time. Such increased time can now be afforded in the data-processing pipeline, since the time for map generation is now reduced from about 60 to 3 minutes by the parallelization discussed in the previous article. Parallel cluster processing time, therefore, enabled this better science result. The first mapping is typically from a point (denoted by coordinates x,y) in the left image to a point (x',y') in the right image. The second mapping is from (x',y') in the right image to some point (x",y") in the left image. If (x,y) and(x",y") are identical, then the mapping is considered perfect. The perfect-match criterion can be relaxed by introducing an error window that admits of round-off error and a small amount of noise. The mapping procedure can be repeated until all points in each image not connected to points in the other image are eliminated, so that what remains are verified correlation data.

  7. Reachable Workspace in Facioscapulohumeral muscular dystrophy (FSHD) by Kinect

    PubMed Central

    Han, Jay J.; Kurillo, Gregorij; Abresch, Richard T.; de Bie, Evan; Nicorici, Alina; Bajcsy, Ruzena

    2014-01-01

    Introduction A depth-ranging sensor (Kinect) based upper extremity motion analysis system was applied to determine the spectrum of reachable workspace encountered in facioscapulohumeral muscular dystrophy (FSHD). Methods Reachable workspaces were obtained from 22 individuals with FSHD and 24 age- and height-matched healthy controls. To allow comparison, total and quadrant reachable workspace relative surface areas (RSA) were obtained by normalizing the acquired reachable workspace by each individual’s arm length. Results Significantly contracted reachable workspace and reduced RSAs were noted for the FSHD cohort compared to controls (0.473±0.188 vs. 0.747±0.082; P<0.0001). With worsening upper extremity function as categorized by the FSHD evaluation subscale II+III, the upper quadrant RSAs decreased progressively, while the lower quadrant RSAs were relatively preserved. There were no side-to-side differences in reachable workspace based on hand-dominance. Discussion This study demonstrates the feasibility and potential of using an innovative Kinect-based reachable workspace outcome measure in FSHD. PMID:24828906

  8. Reducing depth uncertainty in large surgical workspaces, with applications to veterinary medicine

    NASA Astrophysics Data System (ADS)

    Audette, Michel A.; Kolahi, Ahmad; Enquobahrie, Andinet; Gatti, Claudio; Cleary, Kevin

    2010-02-01

    This paper presents on-going research that addresses uncertainty along the Z-axis in image-guided surgery, for applications to large surgical workspaces, including those found in veterinary medicine. Veterinary medicine lags human medicine in using image guidance, despite MR and CT data scanning of animals. The positional uncertainty of a surgical tracking device can be modeled as an octahedron with one long axis coinciding with the depth axis of the sensor, where the short axes are determined by pixel resolution and workspace dimensions. The further a 3D point is from this device, the more elongated is this long axis, and the greater the uncertainty along Z of this point's position, in relation to its components along X and Y. Moreover, for a triangulation-based tracker, its position error degrades with the square of distance. Our approach is to use two or more Micron Trackers to communicate with each other, and combine this feature with flexible positioning. Prior knowledge of the type of surgical procedure, and if applicable, the species of animal that determines the scale of the workspace, would allow the surgeon to pre-operatively configure the trackers in the OR for optimal accuracy. Our research also leverages the open-source Image-guided Surgery Toolkit (IGSTK).

  9. Software and Algorithms for Biomedical Image Data Processing and Visualization

    NASA Technical Reports Server (NTRS)

    Talukder, Ashit; Lambert, James; Lam, Raymond

    2004-01-01

    A new software equipped with novel image processing algorithms and graphical-user-interface (GUI) tools has been designed for automated analysis and processing of large amounts of biomedical image data. The software, called PlaqTrak, has been specifically used for analysis of plaque on teeth of patients. New algorithms have been developed and implemented to segment teeth of interest from surrounding gum, and a real-time image-based morphing procedure is used to automatically overlay a grid onto each segmented tooth. Pattern recognition methods are used to classify plaque from surrounding gum and enamel, while ignoring glare effects due to the reflection of camera light and ambient light from enamel regions. The PlaqTrak system integrates these components into a single software suite with an easy-to-use GUI (see Figure 1) that allows users to do an end-to-end run of a patient s record, including tooth segmentation of all teeth, grid morphing of each segmented tooth, and plaque classification of each tooth image. The automated and accurate processing of the captured images to segment each tooth [see Figure 2(a)] and then detect plaque on a tooth-by-tooth basis is a critical component of the PlaqTrak system to do clinical trials and analysis with minimal human intervention. These features offer distinct advantages over other competing systems that analyze groups of teeth or synthetic teeth. PlaqTrak divides each segmented tooth into eight regions using an advanced graphics morphing procedure [see results on a chipped tooth in Figure 2(b)], and a pattern recognition classifier is then used to locate plaque [red regions in Figure 2(d)] and enamel regions. The morphing allows analysis within regions of teeth, thereby facilitating detailed statistical analysis such as the amount of plaque present on the biting surfaces on teeth. This software system is applicable to a host of biomedical applications, such as cell analysis and life detection, or robotic applications, such

  10. Special Software for Planetary Image Processing and Research

    NASA Astrophysics Data System (ADS)

    Zubarev, A. E.; Nadezhdina, I. E.; Kozlova, N. A.; Brusnikin, E. S.; Karachevtseva, I. P.

    2016-06-01

    The special modules of photogrammetric processing of remote sensing data that provide the opportunity to effectively organize and optimize the planetary studies were developed. As basic application the commercial software package PHOTOMOD™ is used. Special modules were created to perform various types of data processing: calculation of preliminary navigation parameters, calculation of shape parameters of celestial body, global view image orthorectification, estimation of Sun illumination and Earth visibilities from planetary surface. For photogrammetric processing the different types of data have been used, including images of the Moon, Mars, Mercury, Phobos, Galilean satellites and Enceladus obtained by frame or push-broom cameras. We used modern planetary data and images that were taken over the years, shooting from orbit flight path with various illumination and resolution as well as obtained by planetary rovers from surface. Planetary data image processing is a complex task, and as usual it can take from few months to years. We present our efficient pipeline procedure that provides the possibilities to obtain different data products and supports a long way from planetary images to celestial body maps. The obtained data - new three-dimensional control point networks, elevation models, orthomosaics - provided accurate maps production: a new Phobos atlas (Karachevtseva et al., 2015) and various thematic maps that derived from studies of planetary surface (Karachevtseva et al., 2016a).

  11. Validity, Reliability, and Sensitivity of a 3D Vision Sensor-based Upper Extremity Reachable Workspace Evaluation in Neuromuscular Diseases

    PubMed Central

    Han, Jay J.; Kurillo, Gregorij; Abresch, R. Ted; Nicorici, Alina; Bajcsy, Ruzena

    2013-01-01

    Introduction: One of the major challenges in the neuromuscular field has been lack of upper extremity outcome measures that can be useful for clinical therapeutic efficacy studies. Using vision-based sensor system and customized software, 3-dimensional (3D) upper extremity motion analysis can reconstruct a reachable workspace as a valid, reliable and sensitive outcome measure in various neuromuscular conditions where proximal upper extremity range of motion and function is impaired. Methods: Using a stereo-camera sensor system, 3D reachable workspace envelope surface area normalized to an individual’s arm length (relative surface area: RSA) to allow comparison between subjects was determined for 20 healthy controls and 9 individuals with varying degrees of upper extremity dysfunction due to neuromuscular conditions. All study subjects were classified based on Brooke upper extremity function scale. Right and left upper extremity reachable workspaces were determined based on three repeated measures. The RSAs for each frontal hemi-sphere quadrant and total reachable workspaces were determined with and without loading condition (500 gram wrist weight). Data were analyzed for assessment of the developed system and validity, reliability, and sensitivity to change of the reachable workspace outcome. Results: The mean total RSAs of the reachable workspace for the healthy controls and individuals with NMD were significantly different (0.586 ± 0.085 and 0.299 ± 0.198 respectively; p<0.001). All quadrant RSAs were reduced for individuals with NMDs compared to the healthy controls and these reductions correlated with reduced upper limb function as measured by Brooke grade. The upper quadrants of reachable workspace (above the shoulder level) demonstrated greatest reductions in RSA among subjects with progressive severity in upper extremity impairment. Evaluation of the developed outcomes system with the Bland-Altman method demonstrated narrow 95% limits of agreement (LOA

  12. 'Face value': new medical imaging software in commercial view.

    PubMed

    Coopmans, Catelijne

    2011-04-01

    Based on three ethnographic vignettes describing the engagements of a small start-up company with prospective competitors, partners and customers, this paper shows how commercial considerations are folded into the ways visual images become 'seeable'. When company members mount demonstrations of prototype mammography software, they seek to generate interest but also to protect their intellectual property. Pivotal to these efforts to manage revelation and concealment is the visual interface, which is variously performed as obstacle and ally in the development of a profitable product. Using the concept of 'face value', the paper seeks to develop further insight into contemporary dynamics of seeing and showing by tracing the way techno-visual presentations and commercial considerations become entangled in practice. It also draws attention to the salience and significance of enactments of surface and depth in image-based practices. PMID:21998921

  13. Software considerations in the design of an image archive

    NASA Astrophysics Data System (ADS)

    Seshadri, Sridhar B.; Kishore, Sheel; Khalsa, Satjeet S.; Stevens, John F.; Arenson, Ronald L.

    1990-08-01

    The Radiology Department at the Hospital of the University of Pennsylvania is currently expanding its prototype Picture Archiving and Communications System (PACS) into a fully functional clinical system. The first phase of this expansion involves three major efforts: the upgrade of the 10-Mbit token-ring to an 80-Mbit backbone with associated sub-nets, the implementation of a large-scale image archive, and, an interface between the PACS and the Department's Radiology Information System. Upon the completion of this phase, the PACS will serve the storage and display needs of four MRI scanners and four of the Hospital's Intensive Care Units. This paper addresses the implementation of a software suite designed to duplicate and enhance conventional Film Library functions on a PACS. The structure of an electronic 'folder' based upon the ACR/NEMA Digital Imaging and Communication Standard is also introduced.

  14. Automatic Image Registration Using Free and Open Source Software

    NASA Astrophysics Data System (ADS)

    Giri Babu, D.; Raja Shekhar, S. S.; Chandrasekar, K.; Sesha Sai, M. V. R.; Diwakar, P. G.; Dadhwal, V. K.

    2014-11-01

    Image registration is the most critical operation in remote sensing applications to enable location based referencing and analysis of earth features. This is the first step for any process involving identification, time series analysis or change detection using a large set of imagery over a region. Most of the reliable procedures involve time consuming and laborious manual methods of finding the corresponding matching features of the input image with respect to reference. Also the process, as it involves human interaction, does not converge with multiple operations at different times. Automated procedures rely on accurately determining the matching locations or points from both the images under comparison and the procedures are robust and consistent over time. Different algorithms are available to achieve this, based on pattern recognition, feature based detection, similarity techniques etc. In the present study and implementation, Correlation based methods have been used with a improvement over newly developed technique of identifying and pruning the false points of match. Free and Open Source Software (FOSS) have been used to develop the methodology to reach a wider audience, without any dependency on COTS (Commercially off the shelf) software. Standard deviation from foci of the ellipse of correlated points, is a statistical means of ensuring the best match of the points of interest based on both intensity values and location correspondence. The methodology is developed and standardised by enhancements to meet the registration requirements of remote sensing imagery. Results have shown a performance improvement, nearly matching the visual techniques and have been implemented in remote sensing operational projects. The main advantage of the proposed methodology is its viability in production mode environment. This paper also shows that the visualization capabilities of MapWinGIS, GDAL's image handling abilities and OSSIM's correlation facility can be efficiently

  15. Image Settings Affecting Nuchal Translucency Measurement Using Volume NT™ Software

    PubMed Central

    Cho, Hee Young; Kim, Young Han; Park, Yong Won; Kim, Sung Yoon; Lee, Kwang Hee; Yoo, Joon Sang

    2015-01-01

    Purpose To evaluate the effects of the deviation from the mid-sagittal plane, fetal image size, tissue harmonic imaging (THI), and speckle reduction filter (SRF) on the measurement of the nuchal translucency (NT) thickness using Volume NT™ software. Materials and Methods In 79 pregnant women, NT was measured using Volume NT™. Firstly, the three-dimensional volumes were categorized based on the angle of deviation in 10° intervals from the mid-sagittal plane. Secondly, the operator downsized the fetal image to less than 50% of the screen (Method A) and by magnifying the image (Method B). Next, the image was magnified until the fetal head and thorax occupied 75% of the screen, and the NT was measured (Method C). Lastly, NT values were acquired with THI and SRF functions on, with each function alternately on, and with both functions off. Results The mean differences in NT measurements were -0.09 mm (p<0.01) between two-dimensional (2D) and a deviation of 31-40° and -0.10 mm (p<0.01) between 2D and 41-50°. The intraclass correlation coefficients (ICC) for 2D-NT and NT according to image size were 0.858, 0.923, and 0.928 for methods A, B, and C, respectively. The ICC for 2D-NT and NT with respect to the THI and SRF were 0.786, 0.761, 0.740, and 0.731 with both functions on, THI only, SRF only, and with both functions off, respectively. Conclusion NT measurements made using Volume NT™ are affected by angle deviation from the mid-sagittal plane and fetal image size. Additionally, the highest correlation with 2D-NT was achieved when THI and SRF functions were used. PMID:26256978

  16. Workspace zone differentiation and visualization for virtual humans.

    PubMed

    Yang, J; Sinokrot, T; Abdel-Malek, K; Beck, S; Nebel, K

    2008-03-01

    Human performance measures such as discomfort and joint displacement play an important role in product design. The virtual human Santos, a new generation of virtual humans developed at the University of Iowa, goes directly to the computer-aided design model to evaluate a design, saving time and money. This paper presents an optimization-based workspace zone differentiation and visualization. Around the workspace of virtual humans, a volume is discretized to small zones and the posture prediction on each central point of the zone will determine whether the points are outside the workspace as well as the values of different objective functions. Visualization of zone differentiation is accomplished by showing different colours based on values of human performance measures on points that are located inside the workspace. The proposed method can subsequently help ergonomic design. For example, in a vehicle's interior, the controls should not only lie inside the workspace, but also in the zone that encloses the most comfortable points. Using the palette of colours inside the workspace as a visual guide, a designer can obtain a reading of the discomfort level of product users. PMID:18311614

  17. How to Frame Universal Workspace Lighting.

    PubMed

    Mathiasen, Nanet; Frandsen, Anne Kathrine

    2016-01-01

    In 2012 the headquarters for the umbrella organisation 'Disabled people's organisation Denmark' opened, an office building that offers workspace for the administrations of all the member organisations. The ambition for the building was to be the most accessible office building in the world; regardless of disability everybody should be able to move around in the house and work in any of the offices. One, of many ambitions, was to design a functional and effective lighting scheme using as much daylight as possible, and integrating the artificial lighting design and daylight design. The lighting was intended to support all work stations in both one-man offices and open-plan offices with a functional and comfortable visual environment, fit for all users, regardless of disability. Based on a post occupancy evaluation conducted 2 years after the organisations moved in, the present paper evaluates the lighting design in the offices. It reveals that not all the people working in the offices have the same needs and preferences of lighting conditions; these differ even among users with the same disability. Accordingly the findings lead to a discussion on how to understand the concept of Universal Design. Based on the lighting theory of Peter Boyce, the paper discusses the idea of encompassing everyone in the same solution. PMID:27534330

  18. Development of image-processing software for automatic segmentation of brain tumors in MR images

    PubMed Central

    Vijayakumar, C.; Gharpure, Damayanti Chandrashekhar

    2011-01-01

    Most of the commercially available software for brain tumor segmentation have limited functionality and frequently lack the careful validation that is required for clinical studies. We have developed an image-analysis software package called ‘Prometheus,’ which performs neural system–based segmentation operations on MR images using pre-trained information. The software also has the capability to improve its segmentation performance by using the training module of the neural system. The aim of this article is to present the design and modules of this software. The segmentation module of Prometheus can be used primarily for image analysis in MR images. Prometheus was validated against manual segmentation by a radiologist and its mean sensitivity and specificity was found to be 85.71±4.89% and 93.2±2.87%, respectively. Similarly, the mean segmentation accuracy and mean correspondence ratio was found to be 92.35±3.37% and 0.78±0.046, respectively. PMID:21897560

  19. WorkstationJ: workstation emulation software for medical image perception and technology evaluation research

    NASA Astrophysics Data System (ADS)

    Schartz, Kevin M.; Berbaum, Kevin S.; Caldwell, Robert T.; Madsen, Mark T.

    2007-03-01

    We developed image presentation software that mimics the functionality available in the clinic, but also records time-stamped, observer-display interactions and is readily deployable on diverse workstations making it possible to collect comparable observer data at multiple sites. Commercial image presentation software for clinical use has limited application for research on image perception, ergonomics, computer-aids and informatics because it does not collect observer responses, or other information on observer-display interactions, in real time. It is also very difficult to collect observer data from multiple institutions unless the same commercial software is available at different sites. Our software not only records observer reports of abnormalities and their locations, but also inspection time until report, inspection time for each computed radiograph and for each slice of tomographic studies, window/level, and magnification settings used by the observer. The software is a modified version of the open source ImageJ software available from the National Institutes of Health. Our software involves changes to the base code and extensive new plugin code. Our free software is currently capable of displaying computed tomography and computed radiography images. The software is packaged as Java class files and can be used on Windows, Linux, or Mac systems. By deploying our software together with experiment-specific script files that administer experimental procedures and image file handling, multi-institutional studies can be conducted that increase reader and/or case sample sizes or add experimental conditions.

  20. Design and validation of Segment - freely available software for cardiovascular image analysis

    PubMed Central

    2010-01-01

    Background Commercially available software for cardiovascular image analysis often has limited functionality and frequently lacks the careful validation that is required for clinical studies. We have already implemented a cardiovascular image analysis software package and released it as freeware for the research community. However, it was distributed as a stand-alone application and other researchers could not extend it by writing their own custom image analysis algorithms. We believe that the work required to make a clinically applicable prototype can be reduced by making the software extensible, so that researchers can develop their own modules or improvements. Such an initiative might then serve as a bridge between image analysis research and cardiovascular research. The aim of this article is therefore to present the design and validation of a cardiovascular image analysis software package (Segment) and to announce its release in a source code format. Results Segment can be used for image analysis in magnetic resonance imaging (MRI), computed tomography (CT), single photon emission computed tomography (SPECT) and positron emission tomography (PET). Some of its main features include loading of DICOM images from all major scanner vendors, simultaneous display of multiple image stacks and plane intersections, automated segmentation of the left ventricle, quantification of MRI flow, tools for manual and general object segmentation, quantitative regional wall motion analysis, myocardial viability analysis and image fusion tools. Here we present an overview of the validation results and validation procedures for the functionality of the software. We describe a technique to ensure continued accuracy and validity of the software by implementing and using a test script that tests the functionality of the software and validates the output. The software has been made freely available for research purposes in a source code format on the project home page http

  1. 49. TILE PACKING AREA AND APPRENTICE WORKSPACE, SECOND FLOOR, SOUTH ...

    Library of Congress Historic Buildings Survey, Historic Engineering Record, Historic Landscapes Survey

    49. TILE PACKING AREA AND APPRENTICE WORKSPACE, SECOND FLOOR, SOUTH END OF EAST WING. THE SKYLIGHT, ADDED IN 1976. COVERS A ROOF OPENING LEFT FOR THE CHIMNEY OF A POSSIBLE THIRD BISCUIT KILN. - Moravian Pottery & Tile Works, Southwest side of State Route 313 (Swamp Road), Northwest of East Court Street, Doylestown, Bucks County, PA

  2. Toward an Information Visualization Workspace: Combining Multiple Means of Expression.

    ERIC Educational Resources Information Center

    Roth, Steven F.; Chuah, Mei C.; Kerpedjiev, Stephan; Kolojejchick, John A.; Lucas, Peter

    1997-01-01

    Presents an information visualization workspace in which users are able to explore data presented in diverse but coordinated formats using related systems called Visage, SAGE, and selective dynamic manipulation (SDM). Discusses four dimensions for analyzing user interfaces that reveal the combination of design approaches needed for visualizations…

  3. Random issues in workspace analysis for a mobile robot

    NASA Astrophysics Data System (ADS)

    Stǎnescu, Tony; Dolga, Valer; Mondoc, Alina

    2014-12-01

    Evolution of the mobile robot is currently characterized by multiple applications in dynamic workspaces and low initial knowledge. In this paper presents aspects of approaching random processes of evolution of a mobile robot in an unstructured environment . The experimental results are used for modeling an infrared sensor (integrated in the mobile robot structure) and to assess the probability of locating obstacles in the environment.

  4. Appropriation of a Shared Workspace: Organizing Principles and Their Application

    ERIC Educational Resources Information Center

    Overdijk, Maarten; van Diggelen, Wouter

    2008-01-01

    The use and effects of a CSCL-tool are not always predictable from the properties of the tool alone, but depend on how that tool is appropriated. This paper presents the findings from a case study about the appropriation of a graphical shared workspace. When students are presented with a new tool they may encounter competing constraints and…

  5. NEIGHBOUR-IN: Image processing software for spatial analysis of animal grouping

    PubMed Central

    Caubet, Yves; Richard, Freddie-Jeanne

    2015-01-01

    Abstract Animal grouping is a very complex process that occurs in many species, involving many individuals under the influence of different mechanisms. To investigate this process, we have created an image processing software, called NEIGHBOUR-IN, designed to analyse individuals’ coordinates belonging to up to three different groups. The software also includes statistical analysis and indexes to discriminate aggregates based on spatial localisation of individuals and their neighbours. After the description of the software, the indexes computed by the software are illustrated using both artificial patterns and case studies using the spatial distribution of woodlice. The added strengths of this software and methods are also discussed. PMID:26261448

  6. Platform-independent software for medical image processing on the Internet

    NASA Astrophysics Data System (ADS)

    Mancuso, Michael E.; Pathak, Sayan D.; Kim, Yongmin

    1997-05-01

    We have developed a software tool for image processing over the Internet. The tool is a general purpose, easy to use, flexible, platform independent image processing software package with functions most commonly used in medical image processing.It provides for processing of medical images located wither remotely on the Internet or locally. The software was written in Java - the new programming language developed by Sun Microsystems. It was compiled and tested using Microsoft's Visual Java 1.0 and Microsoft's Just in Time Compiler 1.00.6211. The software is simple and easy to use. In order to use the tool, the user needs to download the software from our site before he/she runs it using any Java interpreter, such as those supplied by Sun, Symantec, Borland or Microsoft. Future versions of the operating systems supplied by Sun, Microsoft, Apple, IBM, and others will include Java interpreters. The software is then able to access and process any image on the iNternet or on the local computer. Using a 512 X 512 X 8-bit image, a 3 X 3 convolution took 0.88 seconds on an Intel Pentium Pro PC running at 200 MHz with 64 Mbytes of memory. A window/level operation took 0.38 seconds while a 3 X 3 median filter took 0.71 seconds. These performance numbers demonstrate the feasibility of using this software interactively on desktop computes. Our software tool supports various image processing techniques commonly used in medical image processing and can run without the need of any specialized hardware. It can become an easily accessible resource over the Internet to promote the learning and of understanding image processing algorithms. Also, it could facilitate sharing of medical image databases and collaboration amongst researchers and clinicians, regardless of location.

  7. IMAGE information monitoring and applied graphics software environment. Volume 1. Executive overview

    SciTech Connect

    Hallam, J.W.; Ng, K.B.; Upham, G.L.

    1986-09-01

    The EPRI Information Monitoring and Applied Graphics Environment (IMAGE) system is designed for 'fast proto-typing' of advanced concepts for computer-aided plant operations tools. It is a flexible software system which can be used for rapidly creating, dynamically driving and evaluating advanced operator aid displays. The software is written to be both host computer and graphic device independent. This four volume report includes an Executive Overview of the IMAGE package (Volume 1), followed by Software Description (Volume II), User's Guide (Volume III), and Description of Example Applications (Volume IV).

  8. IMAGE information monitoring and applied graphics software environment. Volume 4. Applications description

    SciTech Connect

    Hallam, J.W.; Ng, K.B.; Upham, G.L.

    1986-09-01

    The EPRI Information Monitoring and Applied Graphics Environment (IMAGE) system is designed for 'fast proto-typing' of advanced concepts for computer-aided plant operations tools. It is a flexible software system which can be used for rapidly creating, dynamically driving and evaluating advanced operator aid displays. The software is written to be both host computer and graphic device independent. This four volume report includes an Executive Overview of the IMAGE package (Volume 1), followed by Software Description (Volume II), User's Guide (Volume III), and Description of Example Applications (Volume IV).

  9. IMAGE information monitoring and applied graphics software environment. Volume 3. User's guide

    SciTech Connect

    Hallam, J.W.; Ng, K.B.; Upham, G.L.

    1986-09-01

    The EPRI Information Monitoring and Applied Graphics Environment (IMAGE) system is designed for 'fast proto-typing' of advanced concepts for computer-aided plant operations tools. It is a flexible software system which can be used for rapidly creating, dynamically driving and evaluating advanced operator aid displays. The software is written to be host computer and graphic device independent. This four volume report includes an Executive Overview of the IMAGE package (Volume 1), followed by Software Description (Volume II), User's Guide (Volume III), and Description of Example Applications (Volume IV).

  10. GILDAS: Grenoble Image and Line Data Analysis Software

    NASA Astrophysics Data System (ADS)

    Gildas Team

    2013-05-01

    GILDAS is a collection of software oriented toward (sub-)millimeter radioastronomical applications (either single-dish or interferometer). It has been adopted as the IRAM standard data reduction package and is jointly maintained by IRAM & CNRS. GILDAS contains many facilities, most of which are oriented towards spectral line mapping and many kinds of 3-dimensional data. The code, written in Fortran-90 with a few parts in C/C++ (mainly keyboard interaction, plotting, widgets), is easily extensible.

  11. ESO C Library for an Image Processing Software Environment (eclipse)

    NASA Astrophysics Data System (ADS)

    Devillard, N.

    Written in ANSI C, eclipse is a library offering numerous services related to astronomical image processing: FITS data access, various image and cube loading methods, binary image handling and filtering (including convolution and morphological filters), 2-D cross-correlation, connected components, cube and image arithmetic, dead pixel detection and correction, object detection, data extraction, flat-fielding with robust fit, image generation, statistics, photometry, image-space resampling, image combination, and cube stacking. It also contains support for mathematical tools like random number generation, FFT, curve fitting, matrices, fast median computation, and point-pattern matching. The main feature of this library is its ability to handle large amounts of input data (up to 2 GB in the current version) regardless of the amount of memory and swap available on the local machine. Another feature is the very high speed allowed by optimized C, making it an ideal base tool for programming efficient number-crunching applications, e.g., on parallel (Beowulf) systems. Running on all Unix-like platforms, eclipse is portable. A high-level interface to Python is foreseen that would allow programmers to prototype their applications much faster than through C programs.

  12. Eclipse: ESO C Library for an Image Processing Software Environment

    NASA Astrophysics Data System (ADS)

    Devillard, Nicolas

    2011-12-01

    Written in ANSI C, eclipse is a library offering numerous services related to astronomical image processing: FITS data access, various image and cube loading methods, binary image handling and filtering (including convolution and morphological filters), 2-D cross-correlation, connected components, cube and image arithmetic, dead pixel detection and correction, object detection, data extraction, flat-fielding with robust fit, image generation, statistics, photometry, image-space resampling, image combination, and cube stacking. It also contains support for mathematical tools like random number generation, FFT, curve fitting, matrices, fast median computation, and point-pattern matching. The main feature of this library is its ability to handle large amounts of input data (up to 2GB in the current version) regardless of the amount of memory and swap available on the local machine. Another feature is the very high speed allowed by optimized C, making it an ideal base tool for programming efficient number-crunching applications, e.g., on parallel (Beowulf) systems.

  13. Polarization information processing and software system design for simultaneously imaging polarimetry

    NASA Astrophysics Data System (ADS)

    Wang, Yahui; Liu, Jing; Jin, Weiqi; Wen, Renjie

    2015-08-01

    Simultaneous imaging polarimetry can realize real-time polarization imaging of the dynamic scene, which has wide application prospect. This paper first briefly illustrates the design of the double separate Wollaston Prism simultaneous imaging polarimetry, and then emphases are put on the polarization information processing methods and software system design for the designed polarimetry. Polarization information processing methods consist of adaptive image segmentation, high-accuracy image registration, instrument matrix calibration. Morphological image processing was used for image segmentation by taking dilation of an image; The accuracy of image registration can reach 0.1 pixel based on the spatial and frequency domain cross-correlation; Instrument matrix calibration adopted four-point calibration method. The software system was implemented under Windows environment based on C++ programming language, which realized synchronous polarization images acquisition and preservation, image processing and polarization information extraction and display. Polarization data obtained with the designed polarimetry shows that: the polarization information processing methods and its software system effectively performs live realize polarization measurement of the four Stokes parameters of a scene. The polarization information processing methods effectively improved the polarization detection accuracy.

  14. BIRP: Software for interactive search and retrieval of image engineering data

    NASA Technical Reports Server (NTRS)

    Arvidson, R. E.; Bolef, L. K.; Guinness, E. A.; Norberg, P.

    1980-01-01

    Better Image Retrieval Programs (BIRP), a set of programs to interactively sort through and to display a database, such as engineering data for images acquired by spacecraft is described. An overview of the philosophy of BIRP design, the structure of BIRP data files, and examples that illustrate the capabilities of the software are provided.

  15. Software for Analyzing Sequences of Flow-Related Images

    NASA Technical Reports Server (NTRS)

    Klimek, Robert; Wright, Ted

    2004-01-01

    Spotlight is a computer program for analysis of sequences of images generated in combustion and fluid physics experiments. Spotlight can perform analysis of a single image in an interactive mode or a sequence of images in an automated fashion. The primary type of analysis is tracking of positions of objects over sequences of frames. Features and objects that are typically tracked include flame fronts, particles, droplets, and fluid interfaces. Spotlight automates the analysis of object parameters, such as centroid position, velocity, acceleration, size, shape, intensity, and color. Images can be processed to enhance them before statistical and measurement operations are performed. An unlimited number of objects can be analyzed simultaneously. Spotlight saves results of analyses in a text file that can be exported to other programs for graphing or further analysis. Spotlight is a graphical-user-interface-based program that at present can be executed on Microsoft Windows and Linux operating systems. A version that runs on Macintosh computers is being considered.

  16. A Review of Diffusion Tensor Magnetic Resonance Imaging Computational Methods and Software Tools

    PubMed Central

    Hasan, Khader M.; Walimuni, Indika S.; Abid, Humaira; Hahn, Klaus R.

    2010-01-01

    In this work we provide an up-to-date short review of computational magnetic resonance imaging (MRI) and software tools that are widely used to process and analyze diffusion-weighted MRI data. A review of different methods used to acquire, model and analyze diffusion-weighted imaging data (DWI) is first provided with focus on diffusion tensor imaging (DTI). The major preprocessing, processing and post-processing procedures applied to DTI data are discussed. A list of freely available software packages to analyze diffusion MRI data is also provided. PMID:21087766

  17. A review of diffusion tensor magnetic resonance imaging computational methods and software tools.

    PubMed

    Hasan, Khader M; Walimuni, Indika S; Abid, Humaira; Hahn, Klaus R

    2011-12-01

    In this work we provide an up-to-date short review of computational magnetic resonance imaging (MRI) and software tools that are widely used to process and analyze diffusion-weighted MRI data. A review of different methods used to acquire, model and analyze diffusion-weighted imaging data (DWI) is first provided with focus on diffusion tensor imaging (DTI). The major preprocessing, processing and post-processing procedures applied to DTI data are discussed. A list of freely available software packages to analyze diffusion MRI data is also provided. PMID:21087766

  18. Spatial data software integration - Merging CAD/CAM/mapping with GIS and image processing

    NASA Technical Reports Server (NTRS)

    Logan, Thomas L.; Bryant, Nevin A.

    1987-01-01

    The integration of CAD/CAM/mapping with image processing using geographic information systems (GISs) as the interface is examined. Particular emphasis is given to the development of software interfaces between JPL's Video Image Communication and Retrieval (VICAR)/Imaged Based Information System (IBIS) raster-based GIS and the CAD/CAM/mapping system. The design and functions of the VICAR and IBIS are described. Vector data capture and editing are studied. Various software programs for interfacing between the VICAR/IBIS and CAD/CAM/mapping are presented and analyzed.

  19. The design of real time infrared image generation software based on Creator and Vega

    NASA Astrophysics Data System (ADS)

    Wang, Rui-feng; Wu, Wei-dong; Huo, Jun-xiu

    2013-09-01

    Considering the requirement of high reality and real-time quality dynamic infrared image of an infrared image simulation, a method to design real-time infrared image simulation application on the platform of VC++ is proposed. This is based on visual simulation software Creator and Vega. The functions of Creator are introduced simply, and the main features of Vega developing environment are analyzed. The methods of infrared modeling and background are offered, the designing flow chart of the developing process of IR image real-time generation software and the functions of TMM Tool and MAT Tool and sensor module are explained, at the same time, the real-time of software is designed.

  20. Software optimization for electrical conductivity imaging in polycrystalline diamond cutters

    SciTech Connect

    Bogdanov, G.; Ludwig, R.; Wiggins, J.; Bertagnolli, K.

    2014-02-18

    We previously reported on an electrical conductivity imaging instrument developed for measurements on polycrystalline diamond cutters. These cylindrical cutters for oil and gas drilling feature a thick polycrystalline diamond layer on a tungsten carbide substrate. The instrument uses electrical impedance tomography to profile the conductivity in the diamond table. Conductivity images must be acquired quickly, on the order of 5 sec per cutter, to be useful in the manufacturing process. This paper reports on successful efforts to optimize the conductivity reconstruction routine, porting major portions of it to NVIDIA GPUs, including a custom CUDA kernel for Jacobian computation.

  1. IDP: Image and data processing (software) in C++

    SciTech Connect

    Lehman, S.

    1994-11-15

    IDP++(Image and Data Processing in C++) is a complied, multidimensional, multi-data type, signal processing environment written in C++. It is being developed within the Radar Ocean Imaging group and is intended as a partial replacement for View. IDP++ takes advantage of the latest object-oriented compiler technology to provide `information hiding.` Users need only know C, not C++. Signals are treated like any other variable with a defined set of operators and functions in an intuitive manner. IDP++ is being designed for real-time environment where interpreted signal processing packages are less efficient.

  2. Image compression software for the SOHO LASCO and EIT experiments

    NASA Technical Reports Server (NTRS)

    Grunes, Mitchell R.; Howard, Russell A.; Hoppel, Karl; Mango, Stephen A.; Wang, Dennis

    1994-01-01

    This paper describes the lossless and lossy image compression algorithms to be used on board the Solar Heliospheric Observatory (SOHO) in conjunction with the Large Angle Spectrometric Coronograph and Extreme Ultraviolet Imaging Telescope experiments. It also shows preliminary results obtained using similar prior imagery and discusses the lossy compression artifacts which will result. This paper is in part intended for the use of SOHO investigators who need to understand the results of SOHO compression in order to better allocate the transmission bits which they have been allocated.

  3. Workspace Analysis and Optimization of 3-PUU Parallel Mechanism in Medicine Base on Genetic Algorithm.

    PubMed

    Hou, Yongchao; Zhao, Yang

    2015-01-01

    A novel 3-PUU parallel robot was put forward, on which kinematic analysis was conducted to obtain its inverse kinematics solution, and on this basis, the limitations of the sliding pair and the Hooke joint on the workspace were analyzed. Moreover, the workspace was solved through the three dimensional limit search method, and then optimization analysis was performed on the workspace of this parallel robot, which laid the foundations for the configuration design and further analysis of the parallel mechanism, with the result indicated that this type of robot was equipped with promising application prospect. In addition that, the workspace after optimization can meet more requirements of patients. PMID:26628930

  4. Workspace Analysis and Optimization of 3-PUU Parallel Mechanism in Medicine Base on Genetic Algorithm

    PubMed Central

    Hou, Yongchao; Zhao, Yang

    2015-01-01

    A novel 3-PUU parallel robot was put forward, on which kinematic analysis was conducted to obtain its inverse kinematics solution, and on this basis, the limitations of the sliding pair and the Hooke joint on the workspace were analyzed. Moreover, the workspace was solved through the three dimensional limit search method, and then optimization analysis was performed on the workspace of this parallel robot, which laid the foundations for the configuration design and further analysis of the parallel mechanism, with the result indicated that this type of robot was equipped with promising application prospect. In addition that, the workspace after optimization can meet more requirements of patients. PMID:26628930

  5. Image analysis software for following progression of peripheral neuropathy

    NASA Astrophysics Data System (ADS)

    Epplin-Zapf, Thomas; Miller, Clayton; Larkin, Sean; Hermesmeyer, Eduardo; Macy, Jenny; Pellegrini, Marco; Luccarelli, Saverio; Staurenghi, Giovanni; Holmes, Timothy

    2009-02-01

    A relationship has been reported by several research groups [1 - 4] between the density and shapes of nerve fibers in the cornea and the existence and severity of peripheral neuropathy. Peripheral neuropathy is a complication of several prevalent diseases or conditions, which include diabetes, HIV, prolonged alcohol overconsumption and aging. A common clinical technique for confirming the condition is intramuscular electromyography (EMG), which is invasive, so a noninvasive technique like the one proposed here carries important potential advantages for the physician and patient. A software program that automatically detects the nerve fibers, counts them and measures their shapes is being developed and tested. Tests were carried out with a database of subjects with levels of severity of diabetic neuropathy as determined by EMG testing. Results from this testing, that include a linear regression analysis are shown.

  6. Development of Software to Model AXAF-I Image Quality

    NASA Technical Reports Server (NTRS)

    Geary, Joseph; Hawkins, Lamar; Ahmad, Anees; Gong, Qian

    1997-01-01

    This report describes work conducted on Delivery Order 181 between October 1996 through June 1997. During this period software was written to: compute axial PSD's from RDOS AXAF-I mirror surface maps; plot axial surface errors and compute PSD's from HDOS "Big 8" axial scans; plot PSD's from FITS format PSD files; plot band-limited RMS vs axial and azimuthal position for multiple PSD files; combine and organize PSD's from multiple mirror surface measurements formatted as input to GRAZTRACE; modify GRAZTRACE to read FITS formatted PSD files; evaluate AXAF-I test results; improve and expand the capabilities of the GT x-ray mirror analysis package. During this period work began on a more user-friendly manual for the GT program, and improvements were made to the on-line help manual.

  7. Validated novel software to measure the conspicuity index of lesions in DICOM images

    NASA Astrophysics Data System (ADS)

    Szczepura, K. R.; Manning, D. J.

    2016-03-01

    A novel software programme and associated Excel spreadsheet has been developed to provide an objective measure of the expected visual detectability of focal abnormalities within DICOM images. ROIs are drawn around the abnormality, the software then fits the lesion using a least squares method to recognize the edges of the lesion based on the full width half maximum. 180 line profiles are then plotted around the lesion, giving 360 edge profiles.

  8. Assessment of using Imaging software Image J to determine percentage woody cover from half meter resolution satellite images

    NASA Astrophysics Data System (ADS)

    Mace, W. D.; Cerling, T. E.

    2010-12-01

    The percentage of woody cover over a landscape has been shown to be related to the d13C in soil organic matter because of the difference in carbon isotope discrimination between plants using C3 and C4 photosynthetic pathways. Woody plants such as those found in dense forests predominantly use the C3 pathway; whereas plants that are found to grow in arid grasslands are predominantly using the C4 pathway. Therefore it has also been shown that it is possible to determine the vegetation of current and past ecosystems using d13C in soil organic matter. With the introduction of very high resolution remote sensing it is becoming possible to make detailed maps based on d13C and estimate percentage woody cover. Using these maps it may be possible to create large scale representations of prehistoric ecosystems. Here we asses the use of a widely available imaging software, Image J to survey the percentage of woody cover of tropical ecosystems in East Africa. These results are compared with canopy gap fraction that has been calculated from in-situ ground-up circular fisheye images. We find that in areas where the percentage woody cover is less than 0.5, Image J is an effective method of analysis; however as the percentage cover becomes greater than 0.5 it becomes difficult to distinguish between true canopy and shadows.

  9. Parallel-Processing Software for Correlating Stereo Images

    NASA Technical Reports Server (NTRS)

    Klimeck, Gerhard; Deen, Robert; Mcauley, Michael; DeJong, Eric

    2007-01-01

    A computer program implements parallel- processing algorithms for cor relating images of terrain acquired by stereoscopic pairs of digital stereo cameras on an exploratory robotic vehicle (e.g., a Mars rove r). Such correlations are used to create three-dimensional computatio nal models of the terrain for navigation. In this program, the scene viewed by the cameras is segmented into subimages. Each subimage is assigned to one of a number of central processing units (CPUs) opera ting simultaneously.

  10. DEIReconstructor: a software for diffraction enhanced imaging processing and tomography reconstruction

    NASA Astrophysics Data System (ADS)

    Zhang, Kai; Yuan, Qing-Xi; Huang, Wan-Xia; Zhu, Pei-Ping; Wu, Zi-Yu

    2014-10-01

    Diffraction enhanced imaging (DEI) has been widely applied in many fields, especially when imaging low-Z samples or when the difference in the attenuation coefficient between different regions in the sample is too small to be detected. Recent developments of this technique have presented a need for a new software package for data analysis. Here, the Diffraction Enhanced Image Reconstructor (DEIReconstructor), developed in Matlab, is presented. DEIReconstructor has a user-friendly graphical user interface and runs under any of the 32-bit or 64-bit Microsoft Windows operating systems including XP and Win7. Many of its features are integrated to support imaging preprocessing, extract absorption, refractive and scattering information of diffraction enhanced imaging and allow for parallel-beam tomography reconstruction for DEI-CT. Furthermore, many other useful functions are also implemented in order to simplify the data analysis and the presentation of results. The compiled software package is freely available.

  11. Image processing software for providing radiometric inputs to land surface climatology models

    NASA Technical Reports Server (NTRS)

    Newcomer, Jeffrey A.; Goetz, Scott J.; Strebel, Donald E.; Hall, Forrest G.

    1989-01-01

    During the First International Land Surface Climatology Project (ISLSCP) Field Experiment (FIFE), 80 gigabytes of image data were generated from a variety of satellite and airborne sensors in a multidisciplinary attempt to study energy and mass exchange between the land surface and the atmosphere. To make these data readily available to researchers with a range of image data handling experience and capabilities, unique image-processing software was designed to perform a variety of nonstandard image-processing manipulations and to derive a set of standard-format image products. The nonconventional features of the software include: (1) adding new layers of geographic coordinates, and solar and viewing conditions to existing data; (2) providing image polygon extraction and calibration of data to at-sensor radiances; and, (3) generating standard-format derived image products that can be easily incorporated into radiometric or climatology models. The derived image products consist of easily handled ASCII descriptor files, byte image data files, and additional per-pixel integer data files (e.g., geographic coordinates, and sun and viewing conditions). Details of the solutions to the image-processing problems, the conventions adopted for handling a variety of satellite and aircraft image data, and the applicability of the output products to quantitative modeling are presented. They should be of general interest to future experiment and data-handling design considerations.

  12. An image-processing software package: UU and Fig for optical metrology applications

    NASA Astrophysics Data System (ADS)

    Chen, Lujie

    2013-06-01

    Modern optical metrology applications are largely supported by computational methods, such as phase shifting [1], Fourier Transform [2], digital image correlation [3], camera calibration [4], etc, in which image processing is a critical and indispensable component. While it is not too difficult to obtain a wide variety of image-processing programs from the internet; few are catered for the relatively special area of optical metrology. This paper introduces an image-processing software package: UU (data processing) and Fig (data rendering) that incorporates many useful functions to process optical metrological data. The cross-platform programs UU and Fig are developed based on wxWidgets. At the time of writing, it has been tested on Windows, Linux and Mac OS. The userinterface is designed to offer precise control of the underline processing procedures in a scientific manner. The data input/output mechanism is designed to accommodate diverse file formats and to facilitate the interaction with other independent programs. In terms of robustness, although the software was initially developed for personal use, it is comparably stable and accurate to most of the commercial software of similar nature. In addition to functions for optical metrology, the software package has a rich collection of useful tools in the following areas: real-time image streaming from USB and GigE cameras, computational geometry, computer vision, fitting of data, 3D image processing, vector image processing, precision device control (rotary stage, PZT stage, etc), point cloud to surface reconstruction, volume rendering, batch processing, etc. The software package is currently used in a number of universities for teaching and research.

  13. Designing Tracking Software for Image-Guided Surgery Applications: IGSTK Experience

    PubMed Central

    Enquobahrie, Andinet; Gobbi, David; Turek, Matt; Cheng, Patrick; Yaniv, Ziv; Lindseth, Frank; Cleary, Kevin

    2009-01-01

    Objective Many image-guided surgery applications require tracking devices as part of their core functionality. The Image-Guided Surgery Toolkit (IGSTK) was designed and developed to interface tracking devices with software applications incorporating medical images. Methods IGSTK was designed as an open source C++ library that provides the basic components needed for fast prototyping and development of image-guided surgery applications. This library follows a component-based architecture with several components designed for specific sets of image-guided surgery functions. At the core of the toolkit is the tracker component that handles communication between a control computer and navigation device to gather pose measurements of surgical instruments present in the surgical scene. The representations of the tracked instruments are superimposed on anatomical images to provide visual feedback to the clinician during surgical procedures. Results The initial version of the IGSTK toolkit has been released in the public domain and several trackers are supported. The toolkit and related information are available at www.igstk.org. Conclusion With the increased popularity of minimally invasive procedures in health care, several tracking devices have been developed for medical applications. Designing and implementing high-quality and safe software to handle these different types of trackers in a common framework is a challenging task. It requires establishing key software design principles that emphasize abstraction, extensibility, reusability, fault-tolerance, and portability. IGSTK is an open source library that satisfies these needs for the image-guided surgery community. PMID:20037671

  14. Pore Size Distribution Estimates Compared: Available software applied to soil CT and synthetic images.

    NASA Astrophysics Data System (ADS)

    Houston, Alasdair N.; Falconer, Ruth E.; Otten, Wilfred; Hapca, Simona M.

    2015-04-01

    The Pore Size Distribution (PSD) has been widely used as a means of characterising porous media and, in conjunction with knowledge of pore space connectivity, has been used to infer hydrological properties. There exist various strategies to estimate PSD from a segmented image and each strategy typically involves a sequence of algorithms that transform image information. Some of these algorithms may be explicitly parameterised, requiring decisions by a knowledgeable operator. As a result PSD estimates may be quite variable between software applications and operators. In order to better understand these differences, a constrained boolean model was used to construct synthetic images whose pore structure is without ambiguity and whose properties can be analytically determined. Applying to such images a selection of analysis procedures in the form of readily available software applications, reveals differences between PSD estimates and analytic information. In some cases it is possible to attribute these differences to artifacts visible within map images generated by the analysis procedures, permitting correction procedures to be devised. In the case of soil CT images which exhibit complex interconnected pore structure, differences in the PSD estimate between analysis procedures are very great in some cases. Inspection of map images can again help in identifying the cause of such problems, but this may result from a fundamental property of the procedure with respect to complex pore structure. Based on the evidence presented, we conclude that some readily available software will produce PSD estimates that can usefully characterise geomaterials.

  15. Monte Carlo PENRADIO software for dose calculation in medical imaging

    NASA Astrophysics Data System (ADS)

    Adrien, Camille; Lòpez Noriega, Mercedes; Bonniaud, Guillaume; Bordy, Jean-Marc; Le Loirec, Cindy; Poumarede, Bénédicte

    2014-06-01

    The increase on the collective radiation dose due to the large number of medical imaging exams has led the medical physics community to deeply consider the amount of dose delivered and its associated risks in these exams. For this purpose we have developed a Monte Carlo tool, PENRADIO, based on a modified version of PENELOPE code 2006 release, to obtain an accurate individualized radiation dose in conventional and interventional radiography and in computed tomography (CT). This tool has been validated showing excellent agreement between the measured and simulated organ doses in the case of a hip conventional radiography and a coronography. We expect the same accuracy in further results for other localizations and CT examinations.

  16. Web-based interactive 2D/3D medical image processing and visualization software.

    PubMed

    Mahmoudi, Seyyed Ehsan; Akhondi-Asl, Alireza; Rahmani, Roohollah; Faghih-Roohi, Shahrooz; Taimouri, Vahid; Sabouri, Ahmad; Soltanian-Zadeh, Hamid

    2010-05-01

    There are many medical image processing software tools available for research and diagnosis purposes. However, most of these tools are available only as local applications. This limits the accessibility of the software to a specific machine, and thus the data and processing power of that application are not available to other workstations. Further, there are operating system and processing power limitations which prevent such applications from running on every type of workstation. By developing web-based tools, it is possible for users to access the medical image processing functionalities wherever the internet is available. In this paper, we introduce a pure web-based, interactive, extendable, 2D and 3D medical image processing and visualization application that requires no client installation. Our software uses a four-layered design consisting of an algorithm layer, web-user-interface layer, server communication layer, and wrapper layer. To compete with extendibility of the current local medical image processing software, each layer is highly independent of other layers. A wide range of medical image preprocessing, registration, and segmentation methods are implemented using open source libraries. Desktop-like user interaction is provided by using AJAX technology in the web-user-interface. For the visualization functionality of the software, the VRML standard is used to provide 3D features over the web. Integration of these technologies has allowed implementation of our purely web-based software with high functionality without requiring powerful computational resources in the client side. The user-interface is designed such that the users can select appropriate parameters for practical research and clinical studies. PMID:20022133

  17. Comparison of two academic software packages for analyzing two-dimensional gel images.

    PubMed

    Wu, Yukun; Zhang, Le

    2011-12-01

    One of the key limitations for proteomic studies using two-dimensional (2D) gel is the lack of automatic, fast, robust, and reliable methods for detecting, matching, and quantifying protein spots. Although there are commercial software packages for 2D gel image analysis, extensive human intervention is still needed for spot detection and matching, which is time-consuming and error-prone. Moreover, the commercial software packages are usually expensive and non-open source. Thus, it is very beneficial for researchers to have free software that is fast, fully automatic, and robust. In this paper, we review and compare two recently developed and publicly available software packages, RegStatGel and Pinnacle, for analyzing 2D gel images. These two software packages share some common features and also have some fundamental difference in the aspects of spot detection and quantification. Based on our experience, RegStatGel is much better in terms of spot detection and matching. It also contains more advanced statistical tools and is more user-friendly. In contrast, Pinnacle is quite sensitive to background noise and relies on external statistical software packages for statistical analysis. PMID:22084013

  18. MMX-I: data-processing software for multimodal X-ray imaging and tomography

    PubMed Central

    Bergamaschi, Antoine; Medjoubi, Kadda; Messaoudi, Cédric; Marco, Sergio; Somogyi, Andrea

    2016-01-01

    A new multi-platform freeware has been developed for the processing and reconstruction of scanning multi-technique X-ray imaging and tomography datasets. The software platform aims to treat different scanning imaging techniques: X-ray fluorescence, phase, absorption and dark field and any of their combinations, thus providing an easy-to-use data processing tool for the X-ray imaging user community. A dedicated data input stream copes with the input and management of large datasets (several hundred GB) collected during a typical multi-technique fast scan at the Nanoscopium beamline and even on a standard PC. To the authors’ knowledge, this is the first software tool that aims at treating all of the modalities of scanning multi-technique imaging and tomography experiments. PMID:27140159

  19. MMX-I: data-processing software for multimodal X-ray imaging and tomography.

    PubMed

    Bergamaschi, Antoine; Medjoubi, Kadda; Messaoudi, Cédric; Marco, Sergio; Somogyi, Andrea

    2016-05-01

    A new multi-platform freeware has been developed for the processing and reconstruction of scanning multi-technique X-ray imaging and tomography datasets. The software platform aims to treat different scanning imaging techniques: X-ray fluorescence, phase, absorption and dark field and any of their combinations, thus providing an easy-to-use data processing tool for the X-ray imaging user community. A dedicated data input stream copes with the input and management of large datasets (several hundred GB) collected during a typical multi-technique fast scan at the Nanoscopium beamline and even on a standard PC. To the authors' knowledge, this is the first software tool that aims at treating all of the modalities of scanning multi-technique imaging and tomography experiments. PMID:27140159

  20. Development of HydroImage, A User Friendly Hydrogeophysical Characterization Software

    SciTech Connect

    Mok, Chin Man; Hubbard, Susan; Chen, Jinsong; Suribhatla, Raghu; Kaback, Dawn Samara

    2014-01-29

    HydroImage, user friendly software that utilizes high-resolution geophysical data for estimating hydrogeological parameters in subsurface strate, was developed under this grant. HydroImage runs on a personal computer platform to promote broad use by hydrogeologists to further understanding of subsurface processes that govern contaminant fate, transport, and remediation. The unique software provides estimates of hydrogeological properties over continuous volumes of the subsurface, whereas previous approaches only allow estimation of point locations. thus, this unique tool can be used to significantly enhance site conceptual models and improve design and operation of remediation systems. The HydroImage technical approach uses statistical models to integrate geophysical data with borehole geological data and hydrological measurements to produce hydrogeological parameter estimates as 2-D or 3-D images.

  1. A User Assessment of Workspaces in Selected Music Education Computer Laboratories.

    ERIC Educational Resources Information Center

    Badolato, Michael Jeremy

    A study of 120 students selected from the user populations of four music education computer laboratories was conducted to determine the applicability of current ergonomic and environmental design guidelines in satisfying the needs of users of educational computing workspaces. Eleven categories of workspace factors were organized into a…

  2. Shadow netWorkspace: An Open Source Intranet for Learning Communities

    ERIC Educational Resources Information Center

    Laffey, James M.; Musser, Dale

    2006-01-01

    Shadow netWorkspace (SNS) is a web application system that allows a school or any type of community to establish an intranet with network workspaces for all members and groups. The goal of SNS has been to make it easy for schools and other educational organizations to provide network services in support of implementing a learning community. SNS is…

  3. The image-guided surgery toolkit IGSTK: an open source C++ software toolkit.

    PubMed

    Enquobahrie, Andinet; Cheng, Patrick; Gary, Kevin; Ibanez, Luis; Gobbi, David; Lindseth, Frank; Yaniv, Ziv; Aylward, Stephen; Jomier, Julien; Cleary, Kevin

    2007-11-01

    This paper presents an overview of the image-guided surgery toolkit (IGSTK). IGSTK is an open source C++ software library that provides the basic components needed to develop image-guided surgery applications. It is intended for fast prototyping and development of image-guided surgery applications. The toolkit was developed through a collaboration between academic and industry partners. Because IGSTK was designed for safety-critical applications, the development team has adopted lightweight software processes that emphasizes safety and robustness while, at the same time, supporting geographically separated developers. A software process that is philosophically similar to agile software methods was adopted emphasizing iterative, incremental, and test-driven development principles. The guiding principle in the architecture design of IGSTK is patient safety. The IGSTK team implemented a component-based architecture and used state machine software design methodologies to improve the reliability and safety of the components. Every IGSTK component has a well-defined set of features that are governed by state machines. The state machine ensures that the component is always in a valid state and that all state transitions are valid and meaningful. Realizing that the continued success and viability of an open source toolkit depends on a strong user community, the IGSTK team is following several key strategies to build an active user community. These include maintaining a users and developers' mailing list, providing documentation (application programming interface reference document and book), presenting demonstration applications, and delivering tutorial sessions at relevant scientific conferences. PMID:17703338

  4. Digital processing of side-scan sonar data with the Woods Hole image processing system software

    USGS Publications Warehouse

    Paskevich, Valerie F.

    1992-01-01

    Since 1985, the Branch of Atlantic Marine Geology has been involved in collecting, processing and digitally mosaicking high and low-resolution side-scan sonar data. Recent development of a UNIX-based image-processing software system includes a series of task specific programs for processing side-scan sonar data. This report describes the steps required to process the collected data and to produce an image that has equal along- and across-track resol

  5. Digital image measurement of specimen deformation based on CCD cameras and Image J software: an application to human pelvic biomechanics

    NASA Astrophysics Data System (ADS)

    Jia, Yongwei; Cheng, Liming; Yu, Guangrong; Lou, Yongjian; Yu, Yan; Chen, Bo; Ding, Zuquan

    2008-03-01

    A method of digital image measurement of specimen deformation based on CCD cameras and Image J software was developed. This method was used to measure the biomechanics behavior of human pelvis. Six cadaveric specimens from the third lumbar vertebra to the proximal 1/3 part of femur were tested. The specimens without any structural abnormalities were dissected of all soft tissue, sparing the hip joint capsules and the ligaments of the pelvic ring and floor. Markers with black dot on white background were affixed to the key regions of the pelvis. Axial loading from the proximal lumbar was applied by MTS in the gradient of 0N to 500N, which simulated the double feet standing stance. The anterior and lateral images of the specimen were obtained through two CCD cameras. Based on Image J software, digital image processing software, which can be freely downloaded from the National Institutes of Health, digital 8-bit images were processed. The procedure includes the recognition of digital marker, image invert, sub-pixel reconstruction, image segmentation, center of mass algorithm based on weighted average of pixel gray values. Vertical displacements of S1 (the first sacral vertebrae) in front view and micro-angular rotation of sacroiliac joint in lateral view were calculated according to the marker movement. The results of digital image measurement showed as following: marker image correlation before and after deformation was excellent. The average correlation coefficient was about 0.983. According to the 768 × 576 pixels image (pixel size 0.68mm × 0.68mm), the precision of the displacement detected in our experiment was about 0.018 pixels and the comparatively error could achieve 1.11\\perthou. The average vertical displacement of S1 of the pelvis was 0.8356+/-0.2830mm under vertical load of 500 Newtons and the average micro-angular rotation of sacroiliac joint in lateral view was 0.584+/-0.221°. The load-displacement curves obtained from our optical measure system

  6. 3-dimensional root phenotyping with a novel imaging and software platform

    Technology Transfer Automated Retrieval System (TEKTRAN)

    A novel imaging and software platform was developed for the high-throughput phenotyping of 3-dimensional root traits during seedling development. To demonstrate the platform’s capacity, plants of two rice (Oryza sativa) genotypes, Azucena and IR64, were grown in a transparent gellan gum system and ...

  7. Onboard utilization of ground control points for image correction. Volume 4: Correlation analysis software design

    NASA Technical Reports Server (NTRS)

    1981-01-01

    The software utilized for image correction accuracy measurement is described. The correlation analysis program is written to allow the user various tools to analyze different correlation algorithms. The algorithms were tested using LANDSAT imagery in two different spectral bands. Three classification algorithms are implemented.

  8. WHIPPET: a collaborative software environment for medical image processing and analysis

    NASA Astrophysics Data System (ADS)

    Hu, Yangqiu; Haynor, David R.; Maravilla, Kenneth R.

    2007-03-01

    While there are many publicly available software packages for medical image processing, making them available to end users in clinical and research labs remains non-trivial. An even more challenging task is to mix these packages to form pipelines that meet specific needs seamlessly, because each piece of software usually has its own input/output formats, parameter sets, and so on. To address these issues, we are building WHIPPET (Washington Heterogeneous Image Processing Pipeline EnvironmenT), a collaborative platform for integrating image analysis tools from different sources. The central idea is to develop a set of Python scripts which glue the different packages together and make it possible to connect them in processing pipelines. To achieve this, an analysis is carried out for each candidate package for WHIPPET, describing input/output formats, parameters, ROI description methods, scripting and extensibility and classifying its compatibility with other WHIPPET components as image file level, scripting level, function extension level, or source code level. We then identify components that can be connected in a pipeline directly via image format conversion. We set up a TWiki server for web-based collaboration so that component analysis and task request can be performed online, as well as project tracking, knowledge base management, and technical support. Currently WHIPPET includes the FSL, MIPAV, FreeSurfer, BrainSuite, Measure, DTIQuery, and 3D Slicer software packages, and is expanding. Users have identified several needed task modules and we report on their implementation.

  9. CellProfiler Analyst: data exploration and analysis software for complex image-based screens

    PubMed Central

    Jones, Thouis R; Kang, In Han; Wheeler, Douglas B; Lindquist, Robert A; Papallo, Adam; Sabatini, David M; Golland, Polina; Carpenter, Anne E

    2008-01-01

    Background Image-based screens can produce hundreds of measured features for each of hundreds of millions of individual cells in a single experiment. Results Here, we describe CellProfiler Analyst, open-source software for the interactive exploration and analysis of multidimensional data, particularly data from high-throughput, image-based experiments. Conclusion The system enables interactive data exploration for image-based screens and automated scoring of complex phenotypes that require combinations of multiple measured features per cell. PMID:19014601

  10. New Software Developments for Quality Mesh Generation and Optimization from Biomedical Imaging Data

    PubMed Central

    Yu, Zeyun; Wang, Jun; Gao, Zhanheng; Xu, Ming; Hoshijima, Masahiko

    2013-01-01

    In this paper we present a new software toolkit for generating and optimizing surface and volumetric meshes from three-dimensional (3D) biomedical imaging data, targeted at image-based finite element analysis of some biomedical activities in a single material domain. Our toolkit includes a series of geometric processing algorithms including surface re-meshing and quality-guaranteed tetrahedral mesh generation and optimization. All methods described have been encapsulated into a user-friendly graphical interface for easy manipulation and informative visualization of biomedical images and mesh models. Numerous examples are presented to demonstrate the effectiveness and efficiency of the described methods and toolkit. PMID:24252469

  11. Software for MR image overlay guided needle insertions: the clinical translation process

    NASA Astrophysics Data System (ADS)

    Ungi, Tamas; U-Thainual, Paweena; Fritz, Jan; Iordachita, Iulian I.; Flammang, Aaron J.; Carrino, John A.; Fichtinger, Gabor

    2013-03-01

    PURPOSE: Needle guidance software using augmented reality image overlay was translated from the experimental phase to support preclinical and clinical studies. Major functional and structural changes were needed to meet clinical requirements. We present the process applied to fulfill these requirements, and selected features that may be applied in the translational phase of other image-guided surgical navigation systems. METHODS: We used an agile software development process for rapid adaptation to unforeseen clinical requests. The process is based on iterations of operating room test sessions, feedback discussions, and software development sprints. The open-source application framework of 3D Slicer and the NA-MIC kit provided sufficient flexibility and stable software foundations for this work. RESULTS: All requirements were addressed in a process with 19 operating room test iterations. Most features developed in this phase were related to workflow simplification and operator feedback. CONCLUSION: Efficient and affordable modifications were facilitated by an open source application framework and frequent clinical feedback sessions. Results of cadaver experiments show that software requirements were successfully solved after a limited number of operating room tests.

  12. IHE cross-enterprise document sharing for imaging: interoperability testing software

    PubMed Central

    2010-01-01

    Background With the deployments of Electronic Health Records (EHR), interoperability testing in healthcare is becoming crucial. EHR enables access to prior diagnostic information in order to assist in health decisions. It is a virtual system that results from the cooperation of several heterogeneous distributed systems. Interoperability between peers is therefore essential. Achieving interoperability requires various types of testing. Implementations need to be tested using software that simulates communication partners, and that provides test data and test plans. Results In this paper we describe a software that is used to test systems that are involved in sharing medical images within the EHR. Our software is used as part of the Integrating the Healthcare Enterprise (IHE) testing process to test the Cross Enterprise Document Sharing for imaging (XDS-I) integration profile. We describe its architecture and functionalities; we also expose the challenges encountered and discuss the elected design solutions. Conclusions EHR is being deployed in several countries. The EHR infrastructure will be continuously evolving to embrace advances in the information technology domain. Our software is built on a web framework to allow for an easy evolution with web technology. The testing software is publicly available; it can be used by system implementers to test their implementations. It can also be used by site integrators to verify and test the interoperability of systems, or by developers to understand specifications ambiguities, or to resolve implementations difficulties. PMID:20858241

  13. Capturing a failure of an ASIC in-situ, using infrared radiometry and image processing software

    NASA Technical Reports Server (NTRS)

    Ruiz, Ronald P.

    2003-01-01

    Failures in electronic devices can sometimes be tricky to locate-especially if they are buried inside radiation-shielded containers designed to work in outer space. Such was the case with a malfunctioning ASIC (Application Specific Integrated Circuit) that was drawing excessive power at a specific temperature during temperature cycle testing. To analyze the failure, infrared radiometry (thermography) was used in combination with image processing software to locate precisely where the power was being dissipated at the moment the failure took place. The IR imaging software was used to make the image of the target and background, appear as unity. As testing proceeded and the failure mode was reached, temperature changes revealed the precise location of the fault. The results gave the design engineers the information they needed to fix the problem. This paper describes the techniques and equipment used to accomplish this failure analysis.

  14. A software to digital image processing to be used in the voxel phantom development.

    PubMed

    Vieira, J W; Lima, F R A

    2009-01-01

    Anthropomorphic models used in computational dosimetry, also denominated phantoms, are based on digital images recorded from scanning of real people by Computed Tomography (CT) or Magnetic Resonance Imaging (MRI). The voxel phantom construction requests computational processing for transformations of image formats, to compact two-dimensional (2-D) images forming of three-dimensional (3-D) matrices, image sampling and quantization, image enhancement, restoration and segmentation, among others. Hardly the researcher of computational dosimetry will find all these available abilities in single software, and almost always this difficulty presents as a result the decrease of the rhythm of his researches or the use, sometimes inadequate, of alternative tools. The need to integrate the several tasks mentioned above to obtain an image that can be used in an exposure computational model motivated the development of the Digital Image Processing (DIP) software, mainly to solve particular problems in Dissertations and Thesis developed by members of the Grupo de Pesquisa em Dosimetria Numérica (GDN/CNPq). Because of this particular objective, the software uses the Portuguese idiom in their implementations and interfaces. This paper presents the second version of the DIP, whose main changes are the more formal organization on menus and menu items, and menu for digital image segmentation. Currently, the DIP contains the menus Fundamentos, Visualizações, Domínio Espacial, Domínio de Frequências, Segmentações and Estudos. Each menu contains items and sub-items with functionalities that, usually, request an image as input and produce an image or an attribute in the output. The DIP reads edits and writes binary files containing the 3-D matrix corresponding to a stack of axial images from a given geometry that can be a human body or other volume of interest. It also can read any type of computational image and to make conversions. When the task involves only an output image

  15. A near-infrared fluorescence-based surgical navigation system imaging software for sentinel lymph node detection

    NASA Astrophysics Data System (ADS)

    Ye, Jinzuo; Chi, Chongwei; Zhang, Shuang; Ma, Xibo; Tian, Jie

    2014-02-01

    Sentinel lymph node (SLN) in vivo detection is vital in breast cancer surgery. A new near-infrared fluorescence-based surgical navigation system (SNS) imaging software, which has been developed by our research group, is presented for SLN detection surgery in this paper. The software is based on the fluorescence-based surgical navigation hardware system (SNHS) which has been developed in our lab, and is designed specifically for intraoperative imaging and postoperative data analysis. The surgical navigation imaging software consists of the following software modules, which mainly include the control module, the image grabbing module, the real-time display module, the data saving module and the image processing module. And some algorithms have been designed to achieve the performance of the software, for example, the image registration algorithm based on correlation matching. Some of the key features of the software include: setting the control parameters of the SNS; acquiring, display and storing the intraoperative imaging data in real-time automatically; analysis and processing of the saved image data. The developed software has been used to successfully detect the SLNs in 21 cases of breast cancer patients. In the near future, we plan to improve the software performance and it will be extensively used for clinical purpose.

  16. OsiriX: an open-source software for navigating in multidimensional DICOM images.

    PubMed

    Rosset, Antoine; Spadola, Luca; Ratib, Osman

    2004-09-01

    A multidimensional image navigation and display software was designed for display and interpretation of large sets of multidimensional and multimodality images such as combined PET-CT studies. The software is developed in Objective-C on a Macintosh platform under the MacOS X operating system using the GNUstep development environment. It also benefits from the extremely fast and optimized 3D graphic capabilities of the OpenGL graphic standard widely used for computer games optimized for taking advantage of any hardware graphic accelerator boards available. In the design of the software special attention was given to adapt the user interface to the specific and complex tasks of navigating through large sets of image data. An interactive jog-wheel device widely used in the video and movie industry was implemented to allow users to navigate in the different dimensions of an image set much faster than with a traditional mouse or on-screen cursors and sliders. The program can easily be adapted for very specific tasks that require a limited number of functions, by adding and removing tools from the program's toolbar and avoiding an overwhelming number of unnecessary tools and functions. The processing and image rendering tools of the software are based on the open-source libraries ITK and VTK. This ensures that all new developments in image processing that could emerge from other academic institutions using these libraries can be directly ported to the OsiriX program. OsiriX is provided free of charge under the GNU open-source licensing agreement at http://homepage.mac.com/rossetantoine/osirix. PMID:15534753

  17. Applying Workspace Limitations in a Velocity-Controlled Robotic Mechanism

    NASA Technical Reports Server (NTRS)

    Abdallah, Muhammad E. (Inventor); Hargrave, Brian (Inventor); Platt, Robert J., Jr. (Inventor)

    2014-01-01

    A robotic system includes a robotic mechanism responsive to velocity control signals, and a permissible workspace defined by a convex-polygon boundary. A host machine determines a position of a reference point on the mechanism with respect to the boundary, and includes an algorithm for enforcing the boundary by automatically shaping the velocity control signals as a function of the position, thereby providing smooth and unperturbed operation of the mechanism along the edges and corners of the boundary. The algorithm is suited for application with higher speeds and/or external forces. A host machine includes an algorithm for enforcing the boundary by shaping the velocity control signals as a function of the reference point position, and a hardware module for executing the algorithm. A method for enforcing the convex-polygon boundary is also provided that shapes a velocity control signal via a host machine as a function of the reference point position.

  18. The i5k Workspace@NAL--enabling genomic data access, visualization and curation of arthropod genomes.

    PubMed

    Poelchau, Monica; Childers, Christopher; Moore, Gary; Tsavatapalli, Vijaya; Evans, Jay; Lee, Chien-Yueh; Lin, Han; Lin, Jun-Wei; Hackett, Kevin

    2015-01-01

    The 5000 arthropod genomes initiative (i5k) has tasked itself with coordinating the sequencing of 5000 insect or related arthropod genomes. The resulting influx of data, mostly from small research groups or communities with little bioinformatics experience, will require visualization, dissemination and curation, preferably from a centralized platform. The National Agricultural Library (NAL) has implemented the i5k Workspace@NAL (http://i5k.nal.usda.gov/) to help meet the i5k initiative's genome hosting needs. Any i5k member is encouraged to contact the i5k Workspace with their genome project details. Once submitted, new content will be accessible via organism pages, genome browsers and BLAST search engines, which are implemented via the open-source Tripal framework, a web interface for the underlying Chado database schema. We also implement the Web Apollo software for groups that choose to curate gene models. New content will add to the existing body of 35 arthropod species, which include species relevant for many aspects of arthropod genomic research, including agriculture, invasion biology, systematics, ecology and evolution, and developmental research. PMID:25332403

  19. The i5k Workspace@NAL—enabling genomic data access, visualization and curation of arthropod genomes

    PubMed Central

    Poelchau, Monica; Childers, Christopher; Moore, Gary; Tsavatapalli, Vijaya; Evans, Jay; Lee, Chien-Yueh; Lin, Han; Lin, Jun-Wei; Hackett, Kevin

    2015-01-01

    The 5000 arthropod genomes initiative (i5k) has tasked itself with coordinating the sequencing of 5000 insect or related arthropod genomes. The resulting influx of data, mostly from small research groups or communities with little bioinformatics experience, will require visualization, dissemination and curation, preferably from a centralized platform. The National Agricultural Library (NAL) has implemented the i5k Workspace@NAL (http://i5k.nal.usda.gov/) to help meet the i5k initiative's genome hosting needs. Any i5k member is encouraged to contact the i5k Workspace with their genome project details. Once submitted, new content will be accessible via organism pages, genome browsers and BLAST search engines, which are implemented via the open-source Tripal framework, a web interface for the underlying Chado database schema. We also implement the Web Apollo software for groups that choose to curate gene models. New content will add to the existing body of 35 arthropod species, which include species relevant for many aspects of arthropod genomic research, including agriculture, invasion biology, systematics, ecology and evolution, and developmental research. PMID:25332403

  20. The accuracy of a designed software for automated localization of craniofacial landmarks on CBCT images

    PubMed Central

    2014-01-01

    Background Two-dimensional projection radiographs have been traditionally considered the modality of choice for cephalometric analysis. To overcome the shortcomings of two-dimensional images, three-dimensional computed tomography (CT) has been used to evaluate craniofacial structures. However, manual landmark detection depends on medical expertise, and the process is time-consuming. The present study was designed to produce software capable of automated localization of craniofacial landmarks on cone beam (CB) CT images based on image registration and to evaluate its accuracy. Methods The software was designed using MATLAB programming language. The technique was a combination of feature-based (principal axes registration) and voxel similarity-based methods for image registration. A total of 8 CBCT images were selected as our reference images for creating a head atlas. Then, 20 CBCT images were randomly selected as the test images for evaluating the method. Three experts twice located 14 landmarks in all 28 CBCT images during two examinations set 6 weeks apart. The differences in the distances of coordinates of each landmark on each image between manual and automated detection methods were calculated and reported as mean errors. Results The combined intraclass correlation coefficient for intraobserver reliability was 0.89 and for interobserver reliability 0.87 (95% confidence interval, 0.82 to 0.93). The mean errors of all 14 landmarks were <4 mm. Additionally, 63.57% of landmarks had a mean error of <3 mm compared with manual detection (gold standard method). Conclusion The accuracy of our approach for automated localization of craniofacial landmarks, which was based on combining feature-based and voxel similarity-based methods for image registration, was acceptable. Nevertheless we recommend repetition of this study using other techniques, such as intensity-based methods. PMID:25223399

  1. Upper Extremity 3D Reachable Workspace Assessment in ALS by Kinect sensor

    PubMed Central

    Oskarsson, Bjorn; Joyce, Nanette C.; de Bie, Evan; Nicorici, Alina; Bajcsy, Ruzena; Kurillo, Gregorij; Han, Jay J.

    2016-01-01

    Introduction Reachable workspace is a measure that provides clinically meaningful information regarding arm function. In this study, a Kinect sensor was used to determine the spectrum of 3D reachable workspace encountered in a cross-sectional cohort of individuals with ALS. Method Bilateral 3D reachable workspace was recorded from 10 subjects with ALS and 23 healthy controls. The data were normalized by each individual's arm length to obtain a reachable workspace relative surface area (RSA). Concurrent validity was assessed by correlation with ALSFRSr scores. Results The Kinect-measured reachable workspace RSA differed significantly between the ALS and control subjects (0.579±0.226 vs. 0.786±0.069; P<0.001). The RSA demonstrated correlation with ALSFRSr upper extremity items (Spearman correlation ρ=0.569; P=0.009). With worsening upper extremity function as categorized by the ALSFRSr, the reachable workspace also decreased progressively. Conclusions This study demonstrates the feasibility and potential of using a novel Kinect-based reachable workspace outcome measure in ALS. PMID:25965847

  2. Determination of 6D workspaces of Gough-type parallel manipulator and comparison between different geometries

    SciTech Connect

    Merlet, J.P.

    1999-09-01

    The author considers in this paper a Gough-type parallel robot whose leg length values are constrained to lie within some fixed ranges and for which there may be mechanical limits for the motion of the passive joints. The purpose of this paper is to present algorithms to determine: the constant orientation workspace; all the possible locations of the center of the platform that can be reached with a fixed orientation; the total orientation workspace: all the possible locations of the center of the platform that can be reached with any orientation in a set defined by three ranges for the orientation angles; the inclusive orientation workspace: all the possible locations of the center of the platform that can be reached with at least one orientation among a set defined by three ranges for the orientation angles. Most of these algorithms are based on a basic methods: approximation of the results by a set of 3D or 6D boxes obtained from an initial estimation through a bisection process. The boxes in the result will either fully or partially lie inside the workspace: the bisection stops as soon as all the boxes that do not lie fully inside the workspace have a size that is lower than a fixed threshold. The paper includes a comparison between the workspace volumes of four different robot geometries, which shows that for robots of similar dimensions the joints layout has a large influence on the workspace volume.

  3. An open-source deconvolution software package for 3-D quantitative fluorescence microscopy imaging

    PubMed Central

    SUN, Y.; DAVIS, P.; KOSMACEK, E. A.; IANZINI, F.; MACKEY, M. A.

    2010-01-01

    Summary Deconvolution techniques have been widely used for restoring the 3-D quantitative information of an unknown specimen observed using a wide-field fluorescence microscope. Deconv, an open-source deconvolution software package, was developed for 3-D quantitative fluorescence microscopy imaging and was released under the GNU Public License. Deconv provides numerical routines for simulation of a 3-D point spread function and deconvolution routines implemented three constrained iterative deconvolution algorithms: one based on a Poisson noise model and two others based on a Gaussian noise model. These algorithms are presented and evaluated using synthetic images and experimentally obtained microscope images, and the use of the library is explained. Deconv allows users to assess the utility of these deconvolution algorithms and to determine which are suited for a particular imaging application. The design of Deconv makes it easy for deconvolution capabilities to be incorporated into existing imaging applications. PMID:19941558

  4. MedXViewer: an extensible web-enabled software package for medical imaging

    NASA Astrophysics Data System (ADS)

    Looney, P. T.; Young, K. C.; Mackenzie, Alistair; Halling-Brown, Mark D.

    2014-03-01

    MedXViewer (Medical eXtensible Viewer) is an application designed to allow workstation-independent, PACS-less viewing and interaction with anonymised medical images (e.g. observer studies). The application was initially implemented for use in digital mammography and tomosynthesis but the flexible software design allows it to be easily extended to other imaging modalities. Regions of interest can be identified by a user and any associated information about a mark, an image or a study can be added. The questions and settings can be easily configured depending on the need of the research allowing both ROC and FROC studies to be performed. The extensible nature of the design allows for other functionality and hanging protocols to be available for each study. Panning, windowing, zooming and moving through slices are all available while modality-specific features can be easily enabled e.g. quadrant zooming in mammographic studies. MedXViewer can integrate with a web-based image database allowing results and images to be stored centrally. The software and images can be downloaded remotely from this centralised data-store. Alternatively, the software can run without a network connection where the images and results can be encrypted and stored locally on a machine or external drive. Due to the advanced workstation-style functionality, the simple deployment on heterogeneous systems over the internet without a requirement for administrative access and the ability to utilise a centralised database, MedXViewer has been used for running remote paper-less observer studies and is capable of providing a training infrastructure and co-ordinating remote collaborative viewing sessions (e.g. cancer reviews, interesting cases).

  5. Higher-order continuation for the determination of robot workspace boundaries

    NASA Astrophysics Data System (ADS)

    Hentz, Gauthier; Charpentier, Isabelle; Renaud, Pierre

    2016-02-01

    In the medical and surgical fields, robotics may be of great interest for safer and more accurate procedures. Space constraints for a robotic assistant are however strict. Therefore, roboticists study non-conventional mechanisms with advantageous size/workspace ratios. The determination of mechanism workspace, and primarily its boundaries, is thus of major importance. This Note builds on boundary equation definition, continuation and automatic differentiation to propose a general, accurate, fast and automated method for the determination of mechanism workspace. The method is illustrated with a planar RRR mechanism and a three-dimensional Orthoglide parallel mechanism.

  6. The Performance Evaluation of Multi-Image 3d Reconstruction Software with Different Sensors

    NASA Astrophysics Data System (ADS)

    Mousavi, V.; Khosravi, M.; Ahmadi, M.; Noori, N.; Naveh, A. Hosseini; Varshosaz, M.

    2015-12-01

    Today, multi-image 3D reconstruction is an active research field and generating three dimensional model of the objects is one the most discussed issues in Photogrammetry and Computer Vision that can be accomplished using range-based or image-based methods. Very accurate and dense point clouds generated by range-based methods such as structured light systems and laser scanners has introduced them as reliable tools in the industry. Image-based 3D digitization methodologies offer the option of reconstructing an object by a set of unordered images that depict it from different viewpoints. As their hardware requirements are narrowed down to a digital camera and a computer system, they compose an attractive 3D digitization approach, consequently, although range-based methods are generally very accurate, image-based methods are low-cost and can be easily used by non-professional users. One of the factors affecting the accuracy of the obtained model in image-based methods is the software and algorithm used to generate three dimensional model. These algorithms are provided in the form of commercial software, open source and web-based services. Another important factor in the accuracy of the obtained model is the type of sensor used. Due to availability of mobile sensors to the public, popularity of professional sensors and the advent of stereo sensors, a comparison of these three sensors plays an effective role in evaluating and finding the optimized method to generate three-dimensional models. Lots of research has been accomplished to identify a suitable software and algorithm to achieve an accurate and complete model, however little attention is paid to the type of sensors used and its effects on the quality of the final model. The purpose of this paper is deliberation and the introduction of an appropriate combination of a sensor and software to provide a complete model with the highest accuracy. To do this, different software, used in previous studies, were compared and

  7. Despeckle filtering software toolbox for ultrasound imaging of the common carotid artery.

    PubMed

    Loizou, Christos P; Theofanous, Charoula; Pantziaris, Marios; Kasparis, Takis

    2014-04-01

    Ultrasound imaging of the common carotid artery (CCA) is a non-invasive tool used in medicine to assess the severity of atherosclerosis and monitor its progression through time. It is also used in border detection and texture characterization of the atherosclerotic carotid plaque in the CCA, the identification and measurement of the intima-media thickness (IMT) and the lumen diameter that all are very important in the assessment of cardiovascular disease (CVD). Visual perception, however, is hindered by speckle, a multiplicative noise, that degrades the quality of ultrasound B-mode imaging. Noise reduction is therefore essential for improving the visual observation quality or as a pre-processing step for further automated analysis, such as image segmentation of the IMT and the atherosclerotic carotid plaque in ultrasound images. In order to facilitate this preprocessing step, we have developed in MATLAB(®) a unified toolbox that integrates image despeckle filtering (IDF), texture analysis and image quality evaluation techniques to automate the pre-processing and complement the disease evaluation in ultrasound CCA images. The proposed software, is based on a graphical user interface (GUI) and incorporates image normalization, 10 different despeckle filtering techniques (DsFlsmv, DsFwiener, DsFlsminsc, DsFkuwahara, DsFgf, DsFmedian, DsFhmedian, DsFad, DsFnldif, DsFsrad), image intensity normalization, 65 texture features, 15 quantitative image quality metrics and objective image quality evaluation. The software is publicly available in an executable form, which can be downloaded from http://www.cs.ucy.ac.cy/medinfo/. It was validated on 100 ultrasound images of the CCA, by comparing its results with quantitative visual analysis performed by a medical expert. It was observed that the despeckle filters DsFlsmv, and DsFhmedian improved image quality perception (based on the expert's assessment and the image texture and quality metrics). It is anticipated that the

  8. The Spectral Image Processing System (SIPS): Software for integrated analysis of AVIRIS data

    NASA Technical Reports Server (NTRS)

    Kruse, F. A.; Lefkoff, A. B.; Boardman, J. W.; Heidebrecht, K. B.; Shapiro, A. T.; Barloon, P. J.; Goetz, A. F. H.

    1992-01-01

    The Spectral Image Processing System (SIPS) is a software package developed by the Center for the Study of Earth from Space (CSES) at the University of Colorado, Boulder, in response to a perceived need to provide integrated tools for analysis of imaging spectrometer data both spectrally and spatially. SIPS was specifically designed to deal with data from the Airborne Visible/Infrared Imaging Spectrometer (AVIRIS) and the High Resolution Imaging Spectrometer (HIRIS), but was tested with other datasets including the Geophysical and Environmental Research Imaging Spectrometer (GERIS), GEOSCAN images, and Landsat TM. SIPS was developed using the 'Interactive Data Language' (IDL). It takes advantage of high speed disk access and fast processors running under the UNIX operating system to provide rapid analysis of entire imaging spectrometer datasets. SIPS allows analysis of single or multiple imaging spectrometer data segments at full spatial and spectral resolution. It also allows visualization and interactive analysis of image cubes derived from quantitative analysis procedures such as absorption band characterization and spectral unmixing. SIPS consists of three modules: SIPS Utilities, SIPS_View, and SIPS Analysis. SIPS version 1.1 is described below.

  9. Oxygen octahedra picker: A software tool to extract quantitative information from STEM images.

    PubMed

    Wang, Yi; Salzberger, Ute; Sigle, Wilfried; Eren Suyolcu, Y; van Aken, Peter A

    2016-09-01

    In perovskite oxide based materials and hetero-structures there are often strong correlations between oxygen octahedral distortions and functionality. Thus, atomistic understanding of the octahedral distortion, which requires accurate measurements of atomic column positions, will greatly help to engineer their properties. Here, we report the development of a software tool to extract quantitative information of the lattice and of BO6 octahedral distortions from STEM images. Center-of-mass and 2D Gaussian fitting methods are implemented to locate positions of individual atom columns. The precision of atomic column distance measurements is evaluated on both simulated and experimental images. The application of the software tool is demonstrated using practical examples. PMID:27344044

  10. The Image-Guided Surgery ToolKit IGSTK: an open source C++ software toolkit

    NASA Astrophysics Data System (ADS)

    Cheng, Peng; Ibanez, Luis; Gobbi, David; Gary, Kevin; Aylward, Stephen; Jomier, Julien; Enquobahrie, Andinet; Zhang, Hui; Kim, Hee-su; Blake, M. Brian; Cleary, Kevin

    2007-03-01

    The Image-Guided Surgery Toolkit (IGSTK) is an open source C++ software library that provides the basic components needed to develop image-guided surgery applications. The focus of the toolkit is on robustness using a state machine architecture. This paper presents an overview of the project based on a recent book which can be downloaded from igstk.org. The paper includes an introduction to open source projects, a discussion of our software development process and the best practices that were developed, and an overview of requirements. The paper also presents the architecture framework and main components. This presentation is followed by a discussion of the state machine model that was incorporated and the associated rationale. The paper concludes with an example application.

  11. Using MATLAB software with Tomcat server and Java platform for remote image analysis in pathology

    PubMed Central

    2011-01-01

    Background The Matlab software is a one of the most advanced development tool for application in engineering practice. From our point of view the most important is the image processing toolbox, offering many built-in functions, including mathematical morphology, and implementation of a many artificial neural networks as AI. It is very popular platform for creation of the specialized program for image analysis, also in pathology. Based on the latest version of Matlab Builder Java toolbox, it is possible to create the software, serving as a remote system for image analysis in pathology via internet communication. The internet platform can be realized based on Java Servlet Pages with Tomcat server as servlet container. Methods In presented software implementation we propose remote image analysis realized by Matlab algorithms. These algorithms can be compiled to executable jar file with the help of Matlab Builder Java toolbox. The Matlab function must be declared with the set of input data, output structure with numerical results and Matlab web figure. Any function prepared in that manner can be used as a Java function in Java Servlet Pages (JSP). The graphical user interface providing the input data and displaying the results (also in graphical form) must be implemented in JSP. Additionally the data storage to database can be implemented within algorithm written in Matlab with the help of Matlab Database Toolbox directly with the image processing. The complete JSP page can be run by Tomcat server. Results The proposed tool for remote image analysis was tested on the Computerized Analysis of Medical Images (CAMI) software developed by author. The user provides image and case information (diagnosis, staining, image parameter etc.). When analysis is initialized, input data with image are sent to servlet on Tomcat. When analysis is done, client obtains the graphical results as an image with marked recognized cells and also the quantitative output. Additionally, the

  12. Development of a software based automatic exposure control system for use in image guided radiation therapy

    NASA Astrophysics Data System (ADS)

    Morton, Daniel R.

    Modern image guided radiation therapy involves the use of an isocentrically mounted imaging system to take radiographs of a patient's position before the start of each treatment. Image guidance helps to minimize errors associated with a patients setup, but the radiation dose received by patients from imaging must be managed to ensure no additional risks. The Varian On-Board Imager (OBI) (Varian Medical Systems, Inc., Palo Alto, CA) does not have an automatic exposure control system and therefore requires exposure factors to be manually selected. Without patient specific exposure factors, images may become saturated and require multiple unnecessary exposures. A software based automatic exposure control system has been developed to predict optimal, patient specific exposure factors. The OBI system was modelled in terms of the x-ray tube output and detector response in order to calculate the level of detector saturation for any exposure situation. Digitally reconstructed radiographs are produced via ray-tracing through the patients' volumetric datasets that are acquired for treatment planning. The ray-trace determines the attenuation of the patient and subsequent x-ray spectra incident on the imaging detector. The resulting spectra are used in the detector response model to determine the exposure levels required to minimize detector saturation. Images calculated for various phantoms showed good agreement with the images that were acquired on the OBI. Overall, regions of detector saturation were accurately predicted and the detector response for non-saturated regions in images of an anthropomorphic phantom were calculated to generally be within 5 to 10 % of the measured values. Calculations were performed on patient data and found similar results as the phantom images, with the calculated images being able to determine detector saturation with close agreement to images that were acquired during treatment. Overall, it was shown that the system model and calculation

  13. Performing Quantitative Imaging Acquisition, Analysis and Visualization Using the Best of Open Source and Commercial Software Solutions

    PubMed Central

    Shenoy, Shailesh M.

    2016-01-01

    A challenge in any imaging laboratory, especially one that uses modern techniques, is to achieve a sustainable and productive balance between using open source and commercial software to perform quantitative image acquisition, analysis and visualization. In addition to considering the expense of software licensing, one must consider factors such as the quality and usefulness of the software’s support, training and documentation. Also, one must consider the reproducibility with which multiple people generate results using the same software to perform the same analysis, how one may distribute their methods to the community using the software and the potential for achieving automation to improve productivity. PMID:27516727

  14. TiLIA: a software package for image analysis of firefly flash patterns.

    PubMed

    Konno, Junsuke; Hatta-Ohashi, Yoko; Akiyoshi, Ryutaro; Thancharoen, Anchana; Silalom, Somyot; Sakchoowong, Watana; Yiu, Vor; Ohba, Nobuyoshi; Suzuki, Hirobumi

    2016-05-01

    As flash signaling patterns of fireflies are species specific, signal-pattern analysis is important for understanding this system of communication. Here, we present time-lapse image analysis (TiLIA), a free open-source software package for signal and flight pattern analyses of fireflies that uses video-recorded image data. TiLIA enables flight path tracing of individual fireflies and provides frame-by-frame coordinates and light intensity data. As an example of TiLIA capabilities, we demonstrate flash pattern analysis of the fireflies Luciola cruciata and L. lateralis during courtship behavior. PMID:27069594

  15. 2D-CELL: image processing software for extraction and analysis of 2-dimensional cellular structures

    NASA Astrophysics Data System (ADS)

    Righetti, F.; Telley, H.; Leibling, Th. M.; Mocellin, A.

    1992-01-01

    2D-CELL is a software package for the processing and analyzing of photographic images of cellular structures in a largely interactive way. Starting from a binary digitized image, the programs extract the line network (skeleton) of the structure and determine the graph representation that best models it. Provision is made for manually correcting defects such as incorrect node positions or dangling bonds. Then a suitable algorithm retrieves polygonal contours which define individual cells — local boundary curvatures are neglected for simplicity. Using elementary analytical geometry relations, a range of metric and topological parameters describing the population are then computed, organized into statistical distributions and graphically displayed.

  16. Vobi One: a data processing software package for functional optical imaging

    PubMed Central

    Takerkart, Sylvain; Katz, Philippe; Garcia, Flavien; Roux, Sébastien; Reynaud, Alexandre; Chavane, Frédéric

    2014-01-01

    Optical imaging is the only technique that allows to record the activity of a neuronal population at the mesoscopic scale. A large region of the cortex (10–20 mm diameter) is directly imaged with a CCD camera while the animal performs a behavioral task, producing spatio-temporal data with an unprecedented combination of spatial and temporal resolutions (respectively, tens of micrometers and milliseconds). However, researchers who have developed and used this technique have relied on heterogeneous software and methods to analyze their data. In this paper, we introduce Vobi One, a software package entirely dedicated to the processing of functional optical imaging data. It has been designed to facilitate the processing of data and the comparison of different analysis methods. Moreover, it should help bring good analysis practices to the community because it relies on a database and a standard format for data handling and it provides tools that allow producing reproducible research. Vobi One is an extension of the BrainVISA software platform, entirely written with the Python programming language, open source and freely available for download at https://trac.int.univ-amu.fr/vobi_one. PMID:24478623

  17. Vobi One: a data processing software package for functional optical imaging.

    PubMed

    Takerkart, Sylvain; Katz, Philippe; Garcia, Flavien; Roux, Sébastien; Reynaud, Alexandre; Chavane, Frédéric

    2014-01-01

    Optical imaging is the only technique that allows to record the activity of a neuronal population at the mesoscopic scale. A large region of the cortex (10-20 mm diameter) is directly imaged with a CCD camera while the animal performs a behavioral task, producing spatio-temporal data with an unprecedented combination of spatial and temporal resolutions (respectively, tens of micrometers and milliseconds). However, researchers who have developed and used this technique have relied on heterogeneous software and methods to analyze their data. In this paper, we introduce Vobi One, a software package entirely dedicated to the processing of functional optical imaging data. It has been designed to facilitate the processing of data and the comparison of different analysis methods. Moreover, it should help bring good analysis practices to the community because it relies on a database and a standard format for data handling and it provides tools that allow producing reproducible research. Vobi One is an extension of the BrainVISA software platform, entirely written with the Python programming language, open source and freely available for download at https://trac.int.univ-amu.fr/vobi_one. PMID:24478623

  18. A software tool for automatic classification and segmentation of 2D/3D medical images

    NASA Astrophysics Data System (ADS)

    Strzelecki, Michal; Szczypinski, Piotr; Materka, Andrzej; Klepaczko, Artur

    2013-02-01

    Modern medical diagnosis utilizes techniques of visualization of human internal organs (CT, MRI) or of its metabolism (PET). However, evaluation of acquired images made by human experts is usually subjective and qualitative only. Quantitative analysis of MR data, including tissue classification and segmentation, is necessary to perform e.g. attenuation compensation, motion detection, and correction of partial volume effect in PET images, acquired with PET/MR scanners. This article presents briefly a MaZda software package, which supports 2D and 3D medical image analysis aiming at quantification of image texture. MaZda implements procedures for evaluation, selection and extraction of highly discriminative texture attributes combined with various classification, visualization and segmentation tools. Examples of MaZda application in medical studies are also provided.

  19. SOFI Simulation Tool: A Software Package for Simulating and Testing Super-Resolution Optical Fluctuation Imaging.

    PubMed

    Girsault, Arik; Lukes, Tomas; Sharipov, Azat; Geissbuehler, Stefan; Leutenegger, Marcel; Vandenberg, Wim; Dedecker, Peter; Hofkens, Johan; Lasser, Theo

    2016-01-01

    Super-resolution optical fluctuation imaging (SOFI) allows one to perform sub-diffraction fluorescence microscopy of living cells. By analyzing the acquired image sequence with an advanced correlation method, i.e. a high-order cross-cumulant analysis, super-resolution in all three spatial dimensions can be achieved. Here we introduce a software tool for a simple qualitative comparison of SOFI images under simulated conditions considering parameters of the microscope setup and essential properties of the biological sample. This tool incorporates SOFI and STORM algorithms, displays and describes the SOFI image processing steps in a tutorial-like fashion. Fast testing of various parameters simplifies the parameter optimization prior to experimental work. The performance of the simulation tool is demonstrated by comparing simulated results with experimentally acquired data. PMID:27583365

  20. AIRS: The Medical Imaging Software for Segmentation and Registration in SPECT/CT

    NASA Astrophysics Data System (ADS)

    Widita, R.; Kurniadi, R.; Haryanto, F.; Darma, Y.; Perkasa, Y. S.; Zasneda, S. S.

    2010-06-01

    We have been successfully developed a new software, Automated Image Registration and Segmentation (AIRS), to fuse the CT and SPECT images. It is designed to solve different registration and segmentation problems that arises in tomographic data sets. AIRS is addressed to obtain anatomic information to be applied to NanoSpect system which is imaging for nano-tissues or small animals. It will be demonstrated that the information obtained by SPECT/CT is more accurate in evaluating patients/objects than that obtained from either SPECT or CT alone. The registration methods developed here are for both two-dimensional and three-dimensional registration. We used normalized mutual information (NMI) which is amenable for images produced by different modalities and having unclear boundaries between tissues. The segmentation components used in this software is region growing algorithms which have proven to be an effective approach for image segmentation. The implementations of region growing developed here are connected threshold and neighborhood connected. Our method is designed to perform with clinically acceptable speed, using accelerated techniques (multiresolution).

  1. Image 100 procedures manual development: Applications system library definition and Image 100 software definition

    NASA Technical Reports Server (NTRS)

    Guseman, L. F., Jr.; Decell, H. P., Jr.

    1975-01-01

    An outline for an Image 100 procedures manual for Earth Resources Program image analysis was developed which sets forth guidelines that provide a basis for the preparation and updating of an Image 100 Procedures Manual. The scope of the outline was limited to definition of general features of a procedures manual together with special features of an interactive system. Computer programs were identified which should be implemented as part of an applications oriented library for the system.

  2. [Development of DICOM image viewing software for efficient image reading and evaluation of distributed server system for diagnostic environment].

    PubMed

    Ishikawa, K

    2000-12-01

    To construct an efficient diagnostic environment using computer displays, the author investigated the time of network transmission using clinical images. In our hospital, we introduced optical-fiber 100Base-Fx Ethernet connections between 22 HIS-segments and one RIS-segment. Although Ethernet architecture is inexpensive, the speed of image transmission becomes 2371 KB/sec. (4.6 CT-slice/sec.) in the RIS-segment and 996 KB/sec. (1.9 CT-slice/sec.) from the RIS-segment to HIS-segments. Because one examination is transmitted in one minute, it does not disturb image reading. Otherwise, a distributed server system using inexpensive personal computers helps in constructing an efficient system. This investigation showed that commercially based Digital Imaging and Communications in Medicine(DICOM) servers and RSNA Central Test Node servers are not so different in transmission speed. The author programmed and developed DICOM transmission and viewing software for Macintosh computers. This viewer includes two inventions, dynamic tiling window system (DTWS) and window binding mode(WBM). On DTWS, windows, tiles, and images are independent objects, which are movable and resizable. The tile-matrix is changeable by mouse dragging, which realizes suitable tile rectangles for wide-low or narrow-high images. The arranging window tool prevents windows from scattering. Using WBM, any operation affects each window similarly. This means that the relationship of compared images is always equivalent. DTWS and WBM contribute greatly to a filmless diagnostic environment. PMID:11197836

  3. BAMS2 workspace: a comprehensive and versatile neuroinformatic platform for collating and processing neuroanatomical connections.

    PubMed

    Bota, Mihail; Talpalaru, Stefan; Hintiryan, Houri; Dong, Hong-Wei; Swanson, Larry W

    2014-10-01

    We describe a novel neuroinformatic platform, the BAMS2 Workspace (http://brancusi1.usc.edu), designed for storing and processing information on gray matter region axonal connections. This de novo constructed module allows registered users to collate their data directly by using a simple and versatile visual interface. It also allows construction and analysis of sets of connections associated with gray matter region nomenclatures from any designated species. The Workspace includes a set of tools allowing the display of data in matrix and networks formats and the uploading of processed information in visual, PDF, CSV, and Excel formats. Finally, the Workspace can be accessed anonymously by third-party systems to create individualized connectivity networks. All features of the BAMS2 Workspace are described in detail and are demonstrated with connectivity reports collated in BAMS and associated with the rat sensory-motor cortex, medial frontal cortex, and amygdalar regions. PMID:24668342

  4. BAMS2 Workspace: a comprehensive and versatile neuroinformatic platform for collating and processing neuroanatomical connections

    PubMed Central

    Bota, Mihail; Talpalaru, Ştefan; Hintiryan, Houri; Dong, Hong-Wei; Swanson, Larry W.

    2014-01-01

    We present in this paper a novel neuroinformatic platform, the BAMS2 Workspace (http://brancusi1.usc.edu), designed for storing and processing information about gray matter region axonal connections. This de novo constructed module allows registered users to directly collate their data by using a simple and versatile visual interface. It also allows construction and analysis of sets of connections associated with gray matter region nomenclatures from any designated species. The Workspace includes a set of tools allowing the display of data in matrix and networks formats, and the uploading of processed information in visual, PDF, CSV, and Excel formats. Finally, the Workspace can be accessed anonymously by third party systems to create individualized connectivity networks. All features of the BAMS2 Workspace are described in detail, and are demonstrated with connectivity reports collated in BAMS and associated with the rat sensory-motor cortex, medial frontal cortex, and amygdalar regions. PMID:24668342

  5. CPG-inspired workspace trajectory generation and adaptive locomotion control for quadruped robots.

    PubMed

    Liu, Chengju; Chen, Qijun; Wang, Danwei

    2011-06-01

    This paper deals with the locomotion control of quadruped robots inspired by the biological concept of central pattern generator (CPG). A control architecture is proposed with a 3-D workspace trajectory generator and a motion engine. The workspace trajectory generator generates adaptive workspace trajectories based on CPGs, and the motion engine realizes joint motion imputes. The proposed architecture is able to generate adaptive workspace trajectories online by tuning the parameters of the CPG network to adapt to various terrains. With feedback information, a quadruped robot can walk through various terrains with adaptive joint control signals. A quadruped platform AIBO is used to validate the proposed locomotion control system. The experimental results confirm the effectiveness of the proposed control architecture. A comparison by experiments shows the superiority of the proposed method against the traditional CPG-joint-space control method. PMID:21216715

  6. A collaborative resource management workspace and project management application for data collection, analysis and visualization: OpenNRM

    NASA Astrophysics Data System (ADS)

    Osti, A.

    2013-12-01

    During the process of research and design for OpenNRM, we imagined a place where diverse groups of people and communities could effectively and efficiently collaborate to manage large-scale environmental problems and projects. Our research revealed the need to combine a variety of software components. Users can explore and analyze a topic while simultaneously develop stories and solve problems in a way that the end result is consumable by their colleagues and the general public. To do this we brought together software modules that are typically separate: Document and Asset Management, GIS and Interactive Mapping, WIKI and Information Libraries, Data Catalogs and Services, Project Management Tools and Storytelling templates. These components, along with others are supported by extensive data catalogs (NWIS, Storet, CDEC, Cuahsi), data analysis tools and web services for a turn-key workspace that will allow you to quickly build project communities and data stories. In this presentation we will show you how our investigation into these collaborative efforts are implemented and working for some of our clients, including the State of California's Sacramento San Joaquin Bay-Delta and San Joaquin River Basin. The case study will display the use of the OpenNRM workspace for real time environmental conditions management, data visualization, project operations, environmental restoration, high frequency monitoring and data reporting. We will demonstrate how scientists and policy makers are working together to tell the story of this complicated and divisive system and how they are becoming better managers of that system. Using the genius of web services, we will show you how OpenNRM was designed to allow you to build your own community while easily sharing data stories, project data, monitoring results, document libraries, interactive maps and datasets with others. We will get into more technical detail by presenting how our data interpolation tools can show high frequency

  7. A Survey of DICOM Viewer Software to Integrate Clinical Research and Medical Imaging.

    PubMed

    Haak, Daniel; Page, Charles-E; Deserno, Thomas M

    2016-04-01

    The digital imaging and communications in medicine (DICOM) protocol is the leading standard for image data management in healthcare. Imaging biomarkers and image-based surrogate endpoints in clinical trials and medical registries require DICOM viewer software with advanced functionality for visualization and interfaces for integration. In this paper, a comprehensive evaluation of 28 DICOM viewers is performed. The evaluation criteria are obtained from application scenarios in clinical research rather than patient care. They include (i) platform, (ii) interface, (iii) support, (iv) two-dimensional (2D), and (v) three-dimensional (3D) viewing. On the average, 4.48 and 1.43 of overall 8 2D and 5 3D image viewing criteria are satisfied, respectively. Suitable DICOM interfaces for central viewing in hospitals are provided by GingkoCADx, MIPAV, and OsiriX Lite. The viewers ImageJ, MicroView, MIPAV, and OsiriX Lite offer all included 3D-rendering features for advanced viewing. Interfaces needed for decentral viewing in web-based systems are offered by Oviyam, Weasis, and Xero. Focusing on open source components, MIPAV is the best candidate for 3D imaging as well as DICOM communication. Weasis is superior for workflow optimization in clinical trials. Our evaluation shows that advanced visualization and suitable interfaces can also be found in the open source field and not only in commercial products. PMID:26482912

  8. Upper extremity 3D reachable workspace analysis in dystrophinopathy using Kinect

    PubMed Central

    Han, Jay J.; Kurillo, Gregorij; Abresch, Richard T.; de Bie, Evan; Nicorici, Alina; Bajcsy, Ruzena

    2015-01-01

    Introduction An innovative upper extremity 3D reachable workspace outcome measure acquired using Kinect sensor is applied towards Duchenne/Becker muscular dystrophy (DMD/BMD). The validity, sensitivity, and clinical meaningfulness of the novel outcome is examined. Methods Upper extremity function assessment (Brooke scale, NeuroQOL questionnaire) and Kinect-based reachable workspace analyses were conducted in 43 individuals with dystrophinopathy (30-DMD, 13-BMD; ages 7–60) and 46 controls (ages 6–68). Results The reachable workspace measure reliably captured a wide-range of upper extremity impairments encountered in both pediatric and adult, as well as ambulatory and non-ambulatory individuals with dystrophinopathy. Reduced reachable workspaces were noted for the dystrophinopathy cohort compared to controls, and they correlated with Brooke grades. Additionally, progressive reduction in reachable workspace directly correlated with worsening ability to perform activities of daily living, as self-reported on the NeuroQOL. Discussion This study demonstrates the utility and potential of the novel sensor-acquired reachable workspace outcome measure in dystrophinopathy. PMID:25597487

  9. Optimal Design of a 3-Leg 6-DOF Parallel Manipulator for a Specific Workspace

    NASA Astrophysics Data System (ADS)

    Fu, Jianxun; Gao, Feng

    2016-04-01

    Researchers seldom study optimum design of a six-degree-of-freedom(DOF) parallel manipulator with three legs based upon the given workspace. An optimal design method of a novel three-leg six-DOF parallel manipulator(TLPM) is presented. The mechanical structure of this robot is introduced, with this structure the kinematic constrain equations is decoupled. Analytical solutions of the forward kinematics are worked out, one configuration of this robot, including position and orientation of the end-effector are graphically displayed. Then, on the basis of several extreme positions of the kinematic performances, the task workspace is given. An algorithm of optimal designing is introduced to find the smallest dimensional parameters of the proposed robot. Examples illustrate the design results, and a design stability index is introduced, which ensures that the robot remains a safe distance from the boundary of sits actual workspace. Finally, one prototype of the robot is developed based on this method. This method can easily find appropriate kinematic parameters that can size a robot having the smallest workspace enclosing a predefined task workspace. It improves the design efficiency, ensures that the robot has a small mechanical size possesses a large given workspace volume, and meets the lightweight design requirements.

  10. Software and hardware integration of a microprogrammable state machine for NMR imaging.

    PubMed

    Stewart, B K; Pratt, R G; Thomas, S R; Dieckman, S L; Ridgway, T H

    1991-01-01

    We have integrated a commercially available microprogrammable state machine (Tecmag PULSkit) for use as a magnetic resonance pulse programmer. Providing the capability for active research environment imaging protocols, it features timing resolution of 100 nsec, ten 16-bit loop counters, and individually addressable look-up tables. This integration involved hardware and software integration with a VAX 11/750 at several levels. Hardware: Each of the three gradient channels employs three digital-to-analog converters (DACs). An 8-bit, 4-quadrant, multiplying DAC generates the gradient waveform shape. A 12-bit DAC generates the multiplying DAC scaling voltage, controlling gradient amplitude and sign. A third 12-bit DAC produces a gradient offset (shim) voltage. An eddy current compensation network is present for each gradient channel. Software: The software design philosophy was to create a flexible interface (interactive window environment), while not constraining complex manipulation of the hardware (direct use of the pulse-sequence compiler primitives and microprogramming). The software levels include (a) pulse-sequence microprogramming, (b) pulse-sequence compiler, (c) interactive parameter specification, and (d) canned pulse-sequence microcode library. PMID:1779734

  11. Development of an Open Source Image-Based Flow Modeling Software - SimVascular

    NASA Astrophysics Data System (ADS)

    Updegrove, Adam; Merkow, Jameson; Schiavazzi, Daniele; Wilson, Nathan; Marsden, Alison; Shadden, Shawn

    2014-11-01

    SimVascular (www.simvascular.org) is currently the only comprehensive software package that provides a complete pipeline from medical image data segmentation to patient specific blood flow simulation. This software and its derivatives have been used in hundreds of conference abstracts and peer-reviewed journal articles, as well as the foundation of medical startups. SimVascular was initially released in August 2007, yet major challenges and deterrents for new adopters were the requirement of licensing three expensive commercial libraries utilized by the software, a complicated build process, and a lack of documentation, support and organized maintenance. In the past year, the SimVascular team has made significant progress to integrate open source alternatives for the linear solver, solid modeling, and mesh generation commercial libraries required by the original public release. In addition, the build system, available distributions, and graphical user interface have been significantly enhanced. Finally, the software has been updated to enable users to directly run simulations using models and boundary condition values, included in the Vascular Model Repository (vascularmodel.org). In this presentation we will briefly overview the capabilities of the new SimVascular 2.0 release. National Science Foundation.

  12. CONRAD—A software framework for cone-beam imaging in radiology

    SciTech Connect

    Maier, Andreas; Choi, Jang-Hwan; Riess, Christian; Keil, Andreas; Fahrig, Rebecca; Hofmann, Hannes G.; Berger, Martin; Fischer, Peter; Schwemmer, Chris; Wu, Haibo; Müller, Kerstin; Hornegger, Joachim

    2013-11-15

    Purpose: In the community of x-ray imaging, there is a multitude of tools and applications that are used in scientific practice. Many of these tools are proprietary and can only be used within a certain lab. Often the same algorithm is implemented multiple times by different groups in order to enable comparison. In an effort to tackle this problem, the authors created CONRAD, a software framework that provides many of the tools that are required to simulate basic processes in x-ray imaging and perform image reconstruction with consideration of nonlinear physical effects.Methods: CONRAD is a Java-based state-of-the-art software platform with extensive documentation. It is based on platform-independent technologies. Special libraries offer access to hardware acceleration such as OpenCL. There is an easy-to-use interface for parallel processing. The software package includes different simulation tools that are able to generate up to 4D projection and volume data and respective vector motion fields. Well known reconstruction algorithms such as FBP, DBP, and ART are included. All algorithms in the package are referenced to a scientific source.Results: A total of 13 different phantoms and 30 processing steps have already been integrated into the platform at the time of writing. The platform comprises 74.000 nonblank lines of code out of which 19% are used for documentation. The software package is available for download at http://conrad.stanford.edu. To demonstrate the use of the package, the authors reconstructed images from two different scanners, a table top system and a clinical C-arm system. Runtimes were evaluated using the RabbitCT platform and demonstrate state-of-the-art runtimes with 2.5 s for the 256 problem size and 12.4 s for the 512 problem size.Conclusions: As a common software framework, CONRAD enables the medical physics community to share algorithms and develop new ideas. In particular this offers new opportunities for scientific collaboration and

  13. CAVASS: a computer-assisted visualization and analysis software system - image processing aspects

    NASA Astrophysics Data System (ADS)

    Udupa, Jayaram K.; Grevera, George J.; Odhner, Dewey; Zhuge, Ying; Souza, Andre; Mishra, Shipra; Iwanaga, Tad

    2007-03-01

    The development of the concepts within 3DVIEWNIX and of the software system 3DVIEWNIX itself dates back to the 1970s. Since then, a series of software packages for Computer Assisted Visualization and Analysis (CAVA) of images came out from our group, 3DVIEWNIX released in 1993, being the most recent, and all were distributed with source code. CAVASS, an open source system, is the latest in this series, and represents the next major incarnation of 3DVIEWNIX. It incorporates four groups of operations: IMAGE PROCESSING (including ROI, interpolation, filtering, segmentation, registration, morphological, and algebraic operations), VISUALIZATION (including slice display, reslicing, MIP, surface rendering, and volume rendering), MANIPULATION (for modifying structures and surgery simulation), ANALYSIS (various ways of extracting quantitative information). CAVASS is designed to work on all platforms. Its key features are: (1) most major CAVA operations incorporated; (2) very efficient algorithms and their highly efficient implementations; (3) parallelized algorithms for computationally intensive operations; (4) parallel implementation via distributed computing on a cluster of PCs; (5) interface to other systems such as CAD/CAM software, ITK, and statistical packages; (6) easy to use GUI. In this paper, we focus on the image processing operations and compare the performance of CAVASS with that of ITK. Our conclusions based on assessing performance by utilizing a regular (6 MB), large (241 MB), and a super (873 MB) 3D image data set are as follows: CAVASS is considerably more efficient than ITK, especially in those operations which are computationally intensive. It can handle considerably larger data sets than ITK. It is easy and ready to use in applications since it provides an easy to use GUI. The users can easily build a cluster from ordinary inexpensive PCs and reap the full power of CAVASS inexpensively compared to expensive multiprocessing systems which are less

  14. 76 FR 43724 - In the Matter of Certain Digital Imaging Devices and Related Software; Notice of Commission...

    Federal Register 2010, 2011, 2012, 2013, 2014

    2011-07-21

    ... Cupertino, California (``Apple''). 75 FR 28058 (May 19, 2010). The complaint alleged ] violations of section... COMMISSION In the Matter of Certain Digital Imaging Devices and Related Software; Notice of Commission... related software by reason of infringement of various claims of United States Patent Nos. 6,031,964 and...

  15. The role of camera-bundled image management software in the consumer digital imaging value chain

    NASA Astrophysics Data System (ADS)

    Mueller, Milton; Mundkur, Anuradha; Balasubramanian, Ashok; Chirania, Virat

    2005-02-01

    This research was undertaken by the Convergence Center at the Syracuse University School of Information Studies (www.digital-convergence.info). Project ICONICA, the name for the research, focuses on the strategic implications of digital Images and the CONvergence of Image management and image CApture. Consumer imaging - the activity that we once called "photography" - is now recognized as in the throes of a digital transformation. At the end of 2003, market researchers estimated that about 30% of the households in the U.S. and 40% of the households in Japan owned digital cameras. In 2004, of the 86 million new cameras sold (excluding one-time use cameras), a majority (56%) were estimated to be digital cameras. Sales of photographic film, while still profitable, are declining precipitously.

  16. Creation of 4D imaging data using open source image registration software

    NASA Astrophysics Data System (ADS)

    Wong, Kenneth H.; Ibanez, Luis; Popa, Teo; Cleary, Kevin

    2006-03-01

    4D images (3 spatial dimensions plus time) using CT or MRI will play a key role in radiation medicine as techniques for respiratory motion compensation become more widely available. Advance knowledge of the motion of a tumor and its surrounding anatomy will allow the creation of highly conformal dose distributions in organs such as the lung, liver, and pancreas. However, many of the current investigations into 4D imaging rely on synchronizing the image acquisition with an external respiratory signal such as skin motion, tidal flow, or lung volume, which typically requires specialized hardware and modifications to the scanner. We propose a novel method for 4D image acquisition that does not require any specific gating equipment and is based solely on open source image registration algorithms. Specifically, we use the Insight Toolkit (ITK) to compute the normalized mutual information (NMI) between images taken at different times and use that value as an index of respiratory phase. This method has the advantages of (1) being able to be implemented without any hardware modification to the scanner, and (2) basing the respiratory phase on changes in internal anatomy rather than external signal. We have demonstrated the capabilities of this method with CT fluoroscopy data acquired from a swine model.

  17. Comparison of Perfusion- and Diffusion-weighted Imaging Parameters in Brain Tumor Studies Processed Using Different Software Platforms

    PubMed Central

    Milchenko, Mikhail V.; Rajderkar, Dhanashree; LaMontagne, Pamela; Massoumzadeh, Parinaz; Bogdasarian, Ronald; Schweitzer, Gordon; Benzinger, Tammie; Marcus, Dan; Shimony, Joshua S.; Fouke, Sarah Jost

    2015-01-01

    Rationale and Objectives To compare quantitative imaging parameter measures from diffusion- and perfusion-weighted imaging magnetic resonance imaging (MRI) sequences in subjects with brain tumors that have been processed with different software platforms. Materials and Methods Scans from 20 subjects with primary brain tumors were selected from the Comprehensive Neuro-oncology Data Repository at Washington University School of Medicine (WUSM) and the Swedish Neuroscience Institute. MR images were coregistered, and each subject's data set was processed by three software packages: 1) vendor-specific scanner software, 2) research software developed at WUSM, and 3) a commercially available, Food and Drug Administration–approved, processing platform (Nordic Ice). Regions of interest (ROIs) were chosen within the brain tumor and normal nontumor tissue. The results obtained using these methods were compared. Results For diffusion parameters, including mean diffusivity and fractional anisotropy, concordance was high when comparing different processing methods. For perfusion-imaging parameters, a significant variance in cerebral blood volume, cerebral blood flow, and mean transit time (MTT) values was seen when comparing the same raw data processed using different software platforms. Correlation was better with larger ROIs (radii ≥ 5 mm). Greatest variance was observed in MTT. Conclusions Diffusion parameter values were consistent across different software processing platforms. Perfusion parameter values were more variable and were influenced by the software used. Variation in the MTT was especially large suggesting that MTT estimation may be unreliable in tumor tissues using current MRI perfusion methods. PMID:25088833

  18. Mississippi Company Using NASA Software Program to Provide Unique Imaging Service: DATASTAR Success Story

    NASA Technical Reports Server (NTRS)

    2001-01-01

    DATASTAR, Inc., of Picayune, Miss., has taken NASA's award-winning Earth Resources Laboratory Applications (ELAS) software program and evolved it to the point that the company is now providing a unique, spatial imagery service over the Internet. ELAS was developed in the early 80's to process satellite and airborne sensor imagery data of the Earth's surface into readable and useable information. While there are several software packages on the market that allow the manipulation of spatial data into useable products, this is usually a laborious task. The new program, called the DATASTAR Image Processing Exploitation, or DIPX, Delivery Service, is a subscription service available over the Internet that takes the work out of the equation and provides normalized geo-spatial data in the form of decision products.

  19. A comprehensive software system for image processing and programming. Final report

    SciTech Connect

    Rasure, J.; Hallett, S.; Jordan, R.

    1994-12-31

    XVision is an example of a comprehensive software system dedicated to the processing of multidimensional scientific data. Because it is comprehensive it is necessarily complex. This design complexity is dealt with by considering XVision as nine overlapping software systems, their components and the required standards. The complexity seen by a user of XVision is minimized by the different interfaces providing access to the image processing routines as well as an interface to ease the incorporation of new routines. The XVision project has stressed the importance of having: (1) interfaces to accommodate users with differing preferences and backgrounds and (2) tools to support the programmer and the scientist. The result is a system that provides a framework for building a powerful research, education and development tool.

  20. Advances in hardware, software, and automation for 193nm aerial image measurement systems

    NASA Astrophysics Data System (ADS)

    Zibold, Axel M.; Schmid, R.; Seyfarth, A.; Waechter, M.; Harnisch, W.; Doornmalen, H. v.

    2005-05-01

    A new, second generation AIMS fab 193 system has been developed which is capable of emulating lithographic imaging of any type of reticles such as binary and phase shift masks (PSM) including resolution enhancement technologies (RET) such as optical proximity correction (OPC) or scatter bars. The system emulates the imaging process by adjustment of the lithography equivalent illumination and imaging conditions of 193nm wafer steppers including circular, annular, dipole and quadrupole type illumination modes. The AIMS fab 193 allows a rapid prediction of wafer printability of critical mask features, including dense patterns and contacts, defects or repairs by acquiring through-focus image stacks by means of a CCD camera followed by quantitative image analysis. Moreover the technology can be readily applied to directly determine the process window of a given mask under stepper imaging conditions. Since data acquisition is performed electronically, AIMS in many applications replaces the need for costly and time consuming wafer prints using a wafer stepper/ scanner followed by CD SEM resist or wafer analysis. The AIMS fab 193 second generation system is designed for 193nm lithography mask printing predictability down to the 65nm node. In addition to hardware improvements a new modular AIMS software is introduced allowing for a fully automated operation mode. Multiple pre-defined points can be visited and through-focus AIMS measurements can be executed automatically in a recipe based mode. To increase the effectiveness of the automated operation mode, the throughput of the system to locate the area of interest, and to acquire the through-focus images is increased by almost a factor of two in comparison with the first generation AIMS systems. In addition a new software plug-in concept is realised for the tools. One new feature has been successfully introduced as "Global CD Map", enabling automated investigation of global mask quality based on the local determination of

  1. Real-time telemedicine using shared three-dimensional workspaces over ATM

    NASA Astrophysics Data System (ADS)

    Cahoon, Peter; Forsey, David R.; Hutchison, Susan

    1999-03-01

    During the past five years a high speed ATM network has been developed at UBC that provides a campus testbed, a local testbed to the hospitals, and a National testbed between here and the BADLAB in Ottawa. This testbed has been developed to combine a commercial shared audio/video/whiteboard environment coupled with a shared interactive 3-dimensional solid model. This solid model ranges from a skull reconstructed from a CT scan with muscles and an overlying skin, to a model of the ventricle system of the human brain. Typical interactions among surgeon, radiologist and modeler consist of having image slices of the original scan shared by all and the ability to adjust the surface of the model to conform to each individuals perception of what the final object should look like. The purpose of this interaction can range from forensic reconstruction from partial remains to pre-maxillofacial surgery. A joint project with the forensic unit of the R.C.M.P. in Ottawa using the BADLAB is now in the stages of testing this methodology on a real case beginning with a CT scan of partial remains. A second study underway with the department of Maxiofacial reconstruction at Dalhousie University in Halifax Nova Scotia and concerns a subject who is about to undergo orthognathic surgery, in particular a mandibular advancement. This subject has been MRI scanned, a solid model constructed of the mandible and the virtual surgery constructed on the model. This model and the procedure have been discussed and modified by the modeler and the maxillofacial specialist using these shared workspaces. The procedure will be repeated after the actual surgery to verify the modeled procedure. The advantage of this technique is that none of the specialists need be in the same room, or city. Given the scarcity of time and specialists this methodology shows great promise. In November of this last year a shared live demonstration of this facial modeler was done between Vancouver and Dalhousie University in

  2. Features of the Upgraded Imaging for Hypersonic Experimental Aeroheating Testing (IHEAT) Software

    NASA Technical Reports Server (NTRS)

    Mason, Michelle L.; Rufer, Shann J.

    2016-01-01

    The Imaging for Hypersonic Experimental Aeroheating Testing (IHEAT) software is used at the NASA Langley Research Center to analyze global aeroheating data on wind tunnel models tested in the Langley Aerothermodynamics Laboratory. One-dimensional, semi-infinite heating data derived from IHEAT are used in the design of thermal protection systems for hypersonic vehicles that are exposed to severe aeroheating loads, such as reentry vehicles during descent and landing procedures. This software program originally was written in the PV-WAVE(Registered Trademark) programming language to analyze phosphor thermography data from the two-color, relative-intensity system developed at Langley. To increase the efficiency, functionality, and reliability of IHEAT, the program was migrated to MATLAB(Registered Trademark) syntax and compiled as a stand-alone executable file labeled version 4.0. New features of IHEAT 4.0 include the options to perform diagnostic checks of the accuracy of the acquired data during a wind tunnel test, to extract data along a specified multi-segment line following a feature such as a leading edge or a streamline, and to batch process all of the temporal frame data from a wind tunnel run. Results from IHEAT 4.0 were compared on a pixel level to the output images from the legacy software to validate the program. The absolute differences between the heat transfer data output from the two programs were on the order of 10(exp -5) to 10(exp -7). IHEAT 4.0 replaces the PV-WAVE(Registered Trademark) version as the production software for aeroheating experiments conducted in the hypersonic facilities at NASA Langley.

  3. Scalable, High-performance 3D Imaging Software Platform: System Architecture and Application to Virtual Colonoscopy

    PubMed Central

    Yoshida, Hiroyuki; Wu, Yin; Cai, Wenli; Brett, Bevin

    2013-01-01

    One of the key challenges in three-dimensional (3D) medical imaging is to enable the fast turn-around time, which is often required for interactive or real-time response. This inevitably requires not only high computational power but also high memory bandwidth due to the massive amount of data that need to be processed. In this work, we have developed a software platform that is designed to support high-performance 3D medical image processing for a wide range of applications using increasingly available and affordable commodity computing systems: multi-core, clusters, and cloud computing systems. To achieve scalable, high-performance computing, our platform (1) employs size-adaptive, distributable block volumes as a core data structure for efficient parallelization of a wide range of 3D image processing algorithms; (2) supports task scheduling for efficient load distribution and balancing; and (3) consists of a layered parallel software libraries that allow a wide range of medical applications to share the same functionalities. We evaluated the performance of our platform by applying it to an electronic cleansing system in virtual colonoscopy, with initial experimental results showing a 10 times performance improvement on an 8-core workstation over the original sequential implementation of the system. PMID:23366803

  4. Digital map and situation surface: a team-oriented multidisplay workspace for network enabled situation analysis

    NASA Astrophysics Data System (ADS)

    Peinsipp-Byma, E.; Geisler, Jürgen; Bader, Thomas

    2009-05-01

    System concepts for network enabled image-based ISR (intelligence, surveillance, reconnaissance) is the major mission of Fraunhofer IITB's applied research in the area of defence and security solutions. For the TechDemo08 as part of the NATO CNAD POW Defence against terrorism Fraunhofer IITB advanced a new multi display concept to handle the shear amount and high complexity of ISR data acquired by networked, distributed surveillance systems with the objective to support the generation of a common situation picture. Amount and Complexity of ISR data demands an innovative man-machine interface concept for humans to deal with it. The IITB's concept is the Digital Map & Situation Surface. This concept offers to the user a coherent multi display environment combining a horizontal surface for the situation overview from the bird's eye view, an attached vertical display for collateral information and so-called foveatablets as personalized magic lenses in order to obtain high resolved and role-specific information about a focused areaof- interest and to interact with it. In the context of TechDemo08 the Digital Map & Situation Surface served as workspace for team-based situation visualization and analysis. Multiple sea- and landside surveillance components were connected to the system.

  5. A complete software application for automatic registration of x-ray mammography and magnetic resonance images

    SciTech Connect

    Solves-Llorens, J. A.; Rupérez, M. J. Monserrat, C.; Lloret, M.

    2014-08-15

    Purpose: This work presents a complete and automatic software application to aid radiologists in breast cancer diagnosis. The application is a fully automated method that performs a complete registration of magnetic resonance (MR) images and x-ray (XR) images in both directions (from MR to XR and from XR to MR) and for both x-ray mammograms, craniocaudal (CC), and mediolateral oblique (MLO). This new approximation allows radiologists to mark points in the MR images and, without any manual intervention, it provides their corresponding points in both types of XR mammograms and vice versa. Methods: The application automatically segments magnetic resonance images and x-ray images using the C-Means method and the Otsu method, respectively. It compresses the magnetic resonance images in both directions, CC and MLO, using a biomechanical model of the breast that distinguishes the specific biomechanical behavior of each one of its three tissues (skin, fat, and glandular tissue) separately. It makes a projection of both compressions and registers them with the original XR images using affine transformations and nonrigid registration methods. Results: The application has been validated by two expert radiologists. This was carried out through a quantitative validation on 14 data sets in which the Euclidean distance between points marked by the radiologists and the corresponding points obtained by the application were measured. The results showed a mean error of 4.2 ± 1.9 mm for the MRI to CC registration, 4.8 ± 1.3 mm for the MRI to MLO registration, and 4.1 ± 1.3 mm for the CC and MLO to MRI registration. Conclusions: A complete software application that automatically registers XR and MR images of the breast has been implemented. The application permits radiologists to estimate the position of a lesion that is suspected of being a tumor in an imaging modality based on its position in another different modality with a clinically acceptable error. The results show that the

  6. Gemini planet imager integration to the Gemini South telescope software environment

    NASA Astrophysics Data System (ADS)

    Rantakyrö, Fredrik T.; Cardwell, Andrew; Chilcote, Jeffrey; Dunn, Jennifer; Goodsell, Stephen; Hibon, Pascale; Macintosh, Bruce; Quiroz, Carlos; Perrin, Marshall D.; Sadakuni, Naru; Saddlemyer, Leslie; Savransky, Dmitry; Serio, Andrew; Winge, Claudia; Galvez, Ramon; Gausachs, Gaston; Hardie, Kayla; Hartung, Markus; Luhrs, Javier; Poyneer, Lisa; Thomas, Sandrine

    2014-08-01

    The Gemini Planet Imager is an extreme AO instrument with an integral field spectrograph (IFS) operating in Y, J, H, and K bands. Both the Gemini telescope and the GPI instrument are very complex systems. Our goal is that the combined telescope and instrument system may be run by one observer operating the instrument, and one operator controlling the telescope and the acquisition of light to the instrument. This requires a smooth integration between the two systems and easily operated control interfaces. We discuss the definition of the software and hardware interfaces, their implementation and testing, and the integration of the instrument with the telescope environment.

  7. Multithreaded real-time 3D image processing software architecture and implementation

    NASA Astrophysics Data System (ADS)

    Ramachandra, Vikas; Atanassov, Kalin; Aleksic, Milivoje; Goma, Sergio R.

    2011-03-01

    Recently, 3D displays and videos have generated a lot of interest in the consumer electronics industry. To make 3D capture and playback popular and practical, a user friendly playback interface is desirable. Towards this end, we built a real time software 3D video player. The 3D video player displays user captured 3D videos, provides for various 3D specific image processing functions and ensures a pleasant viewing experience. Moreover, the player enables user interactivity by providing digital zoom and pan functionalities. This real time 3D player was implemented on the GPU using CUDA and OpenGL. The player provides user interactive 3D video playback. Stereo images are first read by the player from a fast drive and rectified. Further processing of the images determines the optimal convergence point in the 3D scene to reduce eye strain. The rationale for this convergence point selection takes into account scene depth and display geometry. The first step in this processing chain is identifying keypoints by detecting vertical edges within the left image. Regions surrounding reliable keypoints are then located on the right image through the use of block matching. The difference in the positions between the corresponding regions in the left and right images are then used to calculate disparity. The extrema of the disparity histogram gives the scene disparity range. The left and right images are shifted based upon the calculated range, in order to place the desired region of the 3D scene at convergence. All the above computations are performed on one CPU thread which calls CUDA functions. Image upsampling and shifting is performed in response to user zoom and pan. The player also consists of a CPU display thread, which uses OpenGL rendering (quad buffers). This also gathers user input for digital zoom and pan and sends them to the processing thread.

  8. Position, singularity and workspace analysis of 3-PSR-O spatial parallel manipulator

    NASA Astrophysics Data System (ADS)

    Shao, Jiejie; Chen, Wenyu; Fu, Xin

    2015-04-01

    Although the parallel mechanisms have the advantages of high accuracy, velocity, stiffness, and payload capacity, the shortcomings of the space utilization and workspace limit the applications in the confined space. A novel 3 degrees of freedom spatial parallel manipulator 3-PSR-O (prismatic-spherical-revolute) is proposed, which possesses a compact architecture and extended workspace while maintaining the inherent advantages of the parallel mechanisms. The direct-inverse position, singularity and workspace are investigated. The mapping method is adopted in the position analysis, and the closed form solution is derived in the form of a six order equation. The singularity analysis of the mechanism is also carried out based on the geometrical constraints, including six singularity boundaries. A feature boundary, which is independent of the prismatic joints' stroke limit, is obtained by integrating the six singularity boundaries. According to the formation of the reachable workspace, a concept of basic workspace is also introduced and presented in the analytical way. By demarcating the basic workspace along the central height with the feature boundary, the reachable workspace can be derived and analyzed more efficiently. Finally, a comparative study on the space utilization between the 3-PSP parallel mechanism and the new mechanism is also presented. The area of feature boundary of the new mechanism is about 140% of the 3-PSP parallel mechanism, while its installation radius is only 1/2 of the 3-PSP parallel mechanism. The proposed parallel mechanism shows great space utilization, and is ideally suited for applications in confined space occasions such as immersion lithography, nano-imprint etc.

  9. Review of free software tools for image analysis of fluorescence cell micrographs.

    PubMed

    Wiesmann, V; Franz, D; Held, C; Münzenmayer, C; Palmisano, R; Wittenberg, T

    2015-01-01

    An increasing number of free software tools have been made available for the evaluation of fluorescence cell micrographs. The main users are biologists and related life scientists with no or little knowledge of image processing. In this review, we give an overview of available tools and guidelines about which tools the users should use to segment fluorescence micrographs. We selected 15 free tools and divided them into stand-alone, Matlab-based, ImageJ-based, free demo versions of commercial tools and data sharing tools. The review consists of two parts: First, we developed a criteria catalogue and rated the tools regarding structural requirements, functionality (flexibility, segmentation and image processing filters) and usability (documentation, data management, usability and visualization). Second, we performed an image processing case study with four representative fluorescence micrograph segmentation tasks with figure-ground and cell separation. The tools display a wide range of functionality and usability. In the image processing case study, we were able to perform figure-ground separation in all micrographs using mainly thresholding. Cell separation was not possible with most of the tools, because cell separation methods are provided only by a subset of the tools and are difficult to parametrize and to use. Most important is that the usability matches the functionality of a tool. To be usable, specialized tools with less functionality need to fulfill less usability criteria, whereas multipurpose tools need a well-structured menu and intuitive graphical user interface. PMID:25359577

  10. MIA - A free and open source software for gray scale medical image analysis

    PubMed Central

    2013-01-01

    Background Gray scale images make the bulk of data in bio-medical image analysis, and hence, the main focus of many image processing tasks lies in the processing of these monochrome images. With ever improving acquisition devices, spatial and temporal image resolution increases, and data sets become very large. Various image processing frameworks exists that make the development of new algorithms easy by using high level programming languages or visual programming. These frameworks are also accessable to researchers that have no background or little in software development because they take care of otherwise complex tasks. Specifically, the management of working memory is taken care of automatically, usually at the price of requiring more it. As a result, processing large data sets with these tools becomes increasingly difficult on work station class computers. One alternative to using these high level processing tools is the development of new algorithms in a languages like C++, that gives the developer full control over how memory is handled, but the resulting workflow for the prototyping of new algorithms is rather time intensive, and also not appropriate for a researcher with little or no knowledge in software development. Another alternative is in using command line tools that run image processing tasks, use the hard disk to store intermediate results, and provide automation by using shell scripts. Although not as convenient as, e.g. visual programming, this approach is still accessable to researchers without a background in computer science. However, only few tools exist that provide this kind of processing interface, they are usually quite task specific, and don’t provide an clear approach when one wants to shape a new command line tool from a prototype shell script. Results The proposed framework, MIA, provides a combination of command line tools, plug-ins, and libraries that make it possible to run image processing tasks interactively in a command shell

  11. Software-Assisted Depth Analysis of Optic Nerve Stereoscopic Images in Telemedicine.

    PubMed

    Xia, Tian; Patel, Shriji N; Szirth, Ben C; Kolomeyer, Anton M; Khouri, Albert S

    2016-01-01

    Background. Software guided optic nerve assessment can assist in process automation and reduce interobserver disagreement. We tested depth analysis software (DAS) in assessing optic nerve cup-to-disc ratio (VCD) from stereoscopic optic nerve images (SONI) of normal eyes. Methods. In a prospective study, simultaneous SONI from normal subjects were collected during telemedicine screenings using a Kowa 3Wx nonmydriatic simultaneous stereoscopic retinal camera (Tokyo, Japan). VCD was determined from SONI pairs and proprietary pixel DAS (Kowa Inc., Tokyo, Japan) after disc and cup contour line placement. A nonstereoscopic VCD was determined using the right channel of a stereo pair. Mean, standard deviation, t-test, and the intraclass correlation coefficient (ICCC) were calculated. Results. 32 patients had mean age of 40 ± 14 years. Mean VCD on SONI was 0.36 ± 0.09, with DAS 0.38 ± 0.08, and with nonstereoscopic 0.29 ± 0.12. The difference between stereoscopic and DAS assisted was not significant (p = 0.45). ICCC showed agreement between stereoscopic and software VCD assessment. Mean VCD difference was significant between nonstereoscopic and stereoscopic (p < 0.05) and nonstereoscopic and DAS (p < 0.005) recordings. Conclusions. DAS successfully assessed SONI and showed a high degree of correlation to physician-determined stereoscopic VCD. PMID:27190507

  12. Software-Assisted Depth Analysis of Optic Nerve Stereoscopic Images in Telemedicine

    PubMed Central

    Xia, Tian; Patel, Shriji N.; Szirth, Ben C.

    2016-01-01

    Background. Software guided optic nerve assessment can assist in process automation and reduce interobserver disagreement. We tested depth analysis software (DAS) in assessing optic nerve cup-to-disc ratio (VCD) from stereoscopic optic nerve images (SONI) of normal eyes. Methods. In a prospective study, simultaneous SONI from normal subjects were collected during telemedicine screenings using a Kowa 3Wx nonmydriatic simultaneous stereoscopic retinal camera (Tokyo, Japan). VCD was determined from SONI pairs and proprietary pixel DAS (Kowa Inc., Tokyo, Japan) after disc and cup contour line placement. A nonstereoscopic VCD was determined using the right channel of a stereo pair. Mean, standard deviation, t-test, and the intraclass correlation coefficient (ICCC) were calculated. Results. 32 patients had mean age of 40 ± 14 years. Mean VCD on SONI was 0.36 ± 0.09, with DAS 0.38 ± 0.08, and with nonstereoscopic 0.29 ± 0.12. The difference between stereoscopic and DAS assisted was not significant (p = 0.45). ICCC showed agreement between stereoscopic and software VCD assessment. Mean VCD difference was significant between nonstereoscopic and stereoscopic (p < 0.05) and nonstereoscopic and DAS (p < 0.005) recordings. Conclusions. DAS successfully assessed SONI and showed a high degree of correlation to physician-determined stereoscopic VCD. PMID:27190507

  13. Software development for ACR-approved phantom-based nuclear medicine tomographic image quality control with cross-platform compatibility

    NASA Astrophysics Data System (ADS)

    Oh, Jungsu S.; Choi, Jae Min; Nam, Ki Pyo; Chae, Sun Young; Ryu, Jin-Sook; Moon, Dae Hyuk; Kim, Jae Seung

    2015-07-01

    Quality control and quality assurance (QC/QA) have been two of the most important issues in modern nuclear medicine (NM) imaging for both clinical practices and academic research. Whereas quantitative QC analysis software is common to modern positron emission tomography (PET) scanners, the QC of gamma cameras and/or single-photon-emission computed tomography (SPECT) scanners has not been sufficiently addressed. Although a thorough standard operating process (SOP) for mechanical and software maintenance may help the QC/QA of a gamma camera and SPECT-computed tomography (CT), no previous study has addressed a unified platform or process to decipher or analyze SPECT phantom images acquired from various scanners thus far. In addition, a few approaches have established cross-platform software to enable the technologists and physicists to assess the variety of SPECT scanners from different manufacturers. To resolve these issues, we have developed Interactive Data Language (IDL)-based in-house software for crossplatform (in terms of not only operating systems (OS) but also manufacturers) analyses of the QC data on an ACR SPECT phantom, which is essential for assessing and assuring the tomographical image quality of SPECT. We applied our devised software to our routine quarterly QC of ACR SPECT phantom images acquired from a number of platforms (OS/manufacturers). Based on our experience, we suggest that our devised software can offer a unified platform that allows images acquired from various types of scanners to be analyzed with great precision and accuracy.

  14. Holistic approach to design and implementation of a medical teleconsultation workspace.

    PubMed

    Czekierda, Łukasz; Malawski, Filip; Wyszkowski, Przemysław

    2015-10-01

    While there are many state-of-the-art approaches to introducing telemedical services in the area of medical imaging, it is hard to point to studies which would address all relevant aspects in a complete and comprehensive manner. In this paper we describe our approach to design and implementation of a universal platform for imaging medicine which is based on our longstanding experience in this area. We claim it is holistic, because, contrary to most of the available studies it addresses all aspects related to creation and utilization of a medical teleconsultation workspace. We present an extensive analysis of requirements, including possible usage scenarios, user needs, organizational and security issues and infrastructure components. We enumerate and analyze multiple usage scenarios related to medical imaging data in treatment, research and educational applications - with typical teleconsultations treated as just one of many possible options. Certain phases common to all these scenarios have been identified, with the resulting classification distinguishing several modes of operation (local vs. remote, collaborative vs. non-interactive etc.). On this basis we propose a system architecture which addresses all of the identified requirements, applying two key concepts: Service Oriented Architecture (SOA) and Virtual Organizations (VO). The SOA paradigm allows us to decompose the functionality of the system into several distinct building blocks, ensuring flexibility and reliability. The VO paradigm defines the cooperation model for all participating healthcare institutions. Our approach is validated by an ICT platform called TeleDICOM II which implements the proposed architecture. All of its main elements are described in detail and cross-checked against the listed requirements. A case study presents the role and usage of the platform in a specific scenario. Finally, our platform is compared with similar systems described into-date studies and available on the market

  15. Hierarchical Image Segmentation of Remotely Sensed Data using Massively Parallel GNU-LINUX Software

    NASA Technical Reports Server (NTRS)

    Tilton, James C.

    2003-01-01

    A hierarchical set of image segmentations is a set of several image segmentations of the same image at different levels of detail in which the segmentations at coarser levels of detail can be produced from simple merges of regions at finer levels of detail. In [1], Tilton, et a1 describes an approach for producing hierarchical segmentations (called HSEG) and gave a progress report on exploiting these hierarchical segmentations for image information mining. The HSEG algorithm is a hybrid of region growing and constrained spectral clustering that produces a hierarchical set of image segmentations based on detected convergence points. In the main, HSEG employs the hierarchical stepwise optimization (HSWO) approach to region growing, which was described as early as 1989 by Beaulieu and Goldberg. The HSWO approach seeks to produce segmentations that are more optimized than those produced by more classic approaches to region growing (e.g. Horowitz and T. Pavlidis, [3]). In addition, HSEG optionally interjects between HSWO region growing iterations, merges between spatially non-adjacent regions (i.e., spectrally based merging or clustering) constrained by a threshold derived from the previous HSWO region growing iteration. While the addition of constrained spectral clustering improves the utility of the segmentation results, especially for larger images, it also significantly increases HSEG s computational requirements. To counteract this, a computationally efficient recursive, divide-and-conquer, implementation of HSEG (RHSEG) was devised, which includes special code to avoid processing artifacts caused by RHSEG s recursive subdivision of the image data. The recursive nature of RHSEG makes for a straightforward parallel implementation. This paper describes the HSEG algorithm, its recursive formulation (referred to as RHSEG), and the implementation of RHSEG using massively parallel GNU-LINUX software. Results with Landsat TM data are included comparing RHSEG with classic

  16. AnaSP: a software suite for automatic image analysis of multicellular spheroids.

    PubMed

    Piccinini, Filippo

    2015-04-01

    Today, more and more biological laboratories use 3D cell cultures and tissues grown in vitro as a 3D model of in vivo tumours and metastases. In the last decades, it has been extensively established that multicellular spheroids represent an efficient model to validate effects of drugs and treatments for human care applications. However, a lack of methods for quantitative analysis limits the usage of spheroids as models for routine experiments. Several methods have been proposed in literature to perform high throughput experiments employing spheroids by automatically computing different morphological parameters, such as diameter, volume and sphericity. Nevertheless, these systems are typically grounded on expensive automated technologies, that make the suggested solutions affordable only for a limited subset of laboratories, frequently performing high content screening analysis. In this work we propose AnaSP, an open source software suitable for automatically estimating several morphological parameters of spheroids, by simply analyzing brightfield images acquired with a standard widefield microscope, also not endowed with a motorized stage. The experiments performed proved sensitivity and precision of the segmentation method proposed, and excellent reliability of AnaSP to compute several morphological parameters of spheroids imaged in different conditions. AnaSP is distributed as an open source software tool. Its modular architecture and graphical user interface make it attractive also for researchers who do not work in areas of computer vision and suitable for both high content screenings and occasional spheroid-based experiments. PMID:25737369

  17. ASAP (Automatic Software for ASL Processing): A toolbox for processing Arterial Spin Labeling images.

    PubMed

    Mato Abad, Virginia; García-Polo, Pablo; O'Daly, Owen; Hernández-Tamames, Juan Antonio; Zelaya, Fernando

    2016-04-01

    The method of Arterial Spin Labeling (ASL) has experienced a significant rise in its application to functional imaging, since it is the only technique capable of measuring blood perfusion in a truly non-invasive manner. Currently, there are no commercial packages for processing ASL data and there is no recognized standard for normalizing ASL data to a common frame of reference. This work describes a new Automated Software for ASL Processing (ASAP) that can automatically process several ASL datasets. ASAP includes functions for all stages of image pre-processing: quantification, skull-stripping, co-registration, partial volume correction and normalization. To assess the applicability and validity of the toolbox, this work shows its application in the study of hypoperfusion in a sample of healthy subjects at risk of progressing to Alzheimer's disease. ASAP requires limited user intervention, minimizing the possibility of random and systematic errors, and produces cerebral blood flow maps that are ready for statistical group analysis. The software is easy to operate and results in excellent quality of spatial normalization. The results found in this evaluation study are consistent with previous studies that find decreased perfusion in Alzheimer's patients in similar regions and demonstrate the applicability of ASAP. PMID:26612079

  18. A medical software system for volumetric analysis of cerebral pathologies in magnetic resonance imaging (MRI) data.

    PubMed

    Egger, Jan; Kappus, Christoph; Freisleben, Bernd; Nimsky, Christopher

    2012-08-01

    In this contribution, a medical software system for volumetric analysis of different cerebral pathologies in magnetic resonance imaging (MRI) data is presented. The software system is based on a semi-automatic segmentation algorithm and helps to overcome the time-consuming process of volume determination during monitoring of a patient. After imaging, the parameter settings-including a seed point-are set up in the system and an automatic segmentation is performed by a novel graph-based approach. Manually reviewing the result leads to reseeding, adding seed points or an automatic surface mesh generation. The mesh is saved for monitoring the patient and for comparisons with follow-up scans. Based on the mesh, the system performs a voxelization and volume calculation, which leads to diagnosis and therefore further treatment decisions. The overall system has been tested with different cerebral pathologies-glioblastoma multiforme, pituitary adenomas and cerebral aneurysms- and evaluated against manual expert segmentations using the Dice Similarity Coefficient (DSC). Additionally, intra-physician segmentations have been performed to provide a quality measure for the presented system. PMID:21384268

  19. SNARK09 - a software package for reconstruction of 2D images from 1D projections.

    PubMed

    Klukowska, Joanna; Davidi, Ran; Herman, Gabor T

    2013-06-01

    The problem of reconstruction of slices and volumes from 1D and 2D projections has arisen in a large number of scientific fields (including computerized tomography, electron microscopy, X-ray microscopy, radiology, radio astronomy and holography). Many different methods (algorithms) have been suggested for its solution. In this paper we present a software package, SNARK09, for reconstruction of 2D images from their 1D projections. In the area of image reconstruction, researchers often desire to compare two or more reconstruction techniques and assess their relative merits. SNARK09 provides a uniform framework to implement algorithms and evaluate their performance. It has been designed to treat both parallel and divergent projection geometries and can either create test data (with or without noise) for use by reconstruction algorithms or use data collected by another software or a physical device. A number of frequently-used classical reconstruction algorithms are incorporated. The package provides a means for easy incorporation of new algorithms for their testing, comparison and evaluation. It comes with tools for statistical analysis of the results and ten worked examples. PMID:23414602

  20. Measuring the Pain Area: An Intra- and Inter-Rater Reliability Study Using Image Analysis Software.

    PubMed

    Dos Reis, Felipe Jose Jandre; de Barros E Silva, Veronica; de Lucena, Raphaela Nunes; Mendes Cardoso, Bruno Alexandre; Nogueira, Leandro Calazans

    2016-01-01

    Pain drawings have frequently been used for clinical information and research. The aim of this study was to investigate intra- and inter-rater reliability of area measurements performed on pain drawings. Our secondary objective was to verify the reliability when using computers with different screen sizes, both with and without mouse hardware. Pain drawings were completed by patients with chronic neck pain or neck-shoulder-arm pain. Four independent examiners participated in the study. Examiners A and B used the same computer with a 16-inch screen and wired mouse hardware. Examiner C used a notebook with a 16-inch screen and no mouse hardware, and Examiner D used a computer with an 11.6-inch screen and a wireless mouse. Image measurements were obtained using GIMP and NIH ImageJ computer programs. The length of all the images was measured using GIMP software to a set scale in ImageJ. Thus, each marked area was encircled and the total surface area (cm(2) ) was calculated for each pain drawing measurement. A total of 117 areas were identified and 52 pain drawings were analyzed. The intrarater reliability between all examiners was high (ICC = 0.989). The inter-rater reliability was also high. No significant differences were observed when using different screen sizes or when using or not using the mouse hardware. This suggests that the precision of these measurements is acceptable for the use of this method as a measurement tool in clinical practice and research. PMID:25490926

  1. Imaging C. elegans Embryos using an Epifluorescent Microscope and Open Source Software

    PubMed Central

    Verbrugghe, Koen J. C.; Chan, Raymond C.

    2011-01-01

    Cellular processes, such as chromosome assembly, segregation and cytokinesis,are inherently dynamic. Time-lapse imaging of living cells, using fluorescent-labeled reporter proteins or differential interference contrast (DIC) microscopy, allows for the examination of the temporal progression of these dynamic events which is otherwise inferred from analysis of fixed samples1,2. Moreover, the study of the developmental regulations of cellular processes necessitates conducting time-lapse experiments on an intact organism during development. The Caenorhabiditis elegans embryo is light-transparent and has a rapid, invariant developmental program with a known cell lineage3, thus providing an ideal experiment model for studying questions in cell biology4,5and development6-9. C. elegans is amendable to genetic manipulation by forward genetics (based on random mutagenesis10,11) and reverse genetics to target specific genes (based on RNAi-mediated interference and targeted mutagenesis12-15). In addition, transgenic animals can be readily created to express fluorescently tagged proteins or reporters16,17. These traits combine to make it easy to identify the genetic pathways regulating fundamental cellular and developmental processes in vivo18-21. In this protocol we present methods for live imaging of C. elegans embryos using DIC optics or GFP fluorescence on a compound epifluorescent microscope. We demonstrate the ease with which readily available microscopes, typically used for fixed sample imaging, can also be applied for time-lapse analysis using open-source software to automate the imaging process. PMID:21490567

  2. Image pixel guided tours: a software platform for non-destructive x-ray imaging

    NASA Astrophysics Data System (ADS)

    Lam, K. P.; Emery, R.

    2009-02-01

    Multivariate analysis seeks to describe the relationship between an arbitrary number of variables. To explore highdimensional data sets, projections are often used for data visualisation to aid discovering structure or patterns that lead to the formation of statistical hypothesis. The basic concept necessitates a systematic search for lower-dimensional representations of the data that might show interesting structure(s). Motivated by the recent research on the Image Grand Tour (IGT), which can be adapted to view guided projections by using objective indexes that are capable of revealing latent structures of the data, this paper presents a signal processing perspective on constructing such indexes under the unifying exploratory frameworks of Independent Component Analysis (ICA) and Projection Pursuit (PP). Our investigation begins with an overview of dimension reduction techniques by means of orthogonal transforms, including the classical procedure of Principal Component Analysis (PCA), and extends to an application of the more powerful techniques of ICA in the context of our recent work on non-destructive testing technology by element specific x-ray imaging.

  3. New AIRS: The medical imaging software for segmentation and registration of elastic organs in SPECT/CT

    NASA Astrophysics Data System (ADS)

    Widita, R.; Kurniadi, R.; Darma, Y.; Perkasa, Y. S.; Trianti, N.

    2012-06-01

    We have been successfully improved our software, Automated Image Registration and Segmentation (AIRS), to fuse the CT and SPECT images of elastic organs. Segmentation and registration of elastic organs presents many challenges. Many artifacts can arise in SPECT/CT scans. Also, different organs and tissues have very similar gray levels, which consign thresholding to limited utility. We have been developed a new software to solve different registration and segmentation problems that arises in tomographic data sets. It will be demonstrated that the information obtained by SPECT/CT is more accurate in evaluating patients/objects than that obtained from either SPECT or CT alone. We used multi-modality registration which is amenable for images produced by different modalities and having unclear boundaries between tissues. The segmentation components used in this software is region growing algorithms which have proven to be an effective approach for image segmentation. Our method is designed to perform with clinically acceptable speed, using accelerated techniques (multiresolution).

  4. RegStatGel: proteomic software for identifying differentially expressed proteins based on 2D gel images

    PubMed Central

    Li, Feng; Seillier-Moiseiwitsch, Françoise

    2011-01-01

    Image analysis of two-dimensional gel electrophoresis is a key step in proteomic workflow for identifying proteins that change under different experimental conditions. Since there are usually large amount of proteins and variations shown in the gel images, the use of software for analysis of 2D gel images is inevitable. We developed open-source software with graphical user interface for differential analysis of 2D gel images. The user-friendly software, RegStatGel, contains fully automated as well as interactive procedures. It was developed and has been tested under Matlab 7.01. Availability The database is available for free at http://www.mediafire.com/FengLi/2DGelsoftware PMID:21904427

  5. An Effective On-line Polymer Characterization Technique by Using SALS Image Processing Software and Wavelet Analysis

    PubMed Central

    Xian, Guang-ming; Qu, Jin-ping; Zeng, Bi-qing

    2008-01-01

    This paper describes an effective on-line polymer characterization technique by using small-angle light-scattering (SALS) image processing software and wavelet analysis. The phenomenon of small-angle light scattering has been applied to give information about transparent structures on morphology. Real-time visualization of various scattered light image and light intensity matrices is performed by the optical image real-time processing software for SALS. The software can measure the signal intensity of light scattering images, draw the frequency-intensity curves and the amplitude-intensity curves to indicate the variation of the intensity of scattered light in different processing conditions, and estimate the parameters. The current study utilizes a one-dimensional wavelet to delete noise from the original SALS signal and estimate the variation trend of maximum intensity area of the scattered light. So, the system brought the qualitative analysis of the structural information of transparent film success. PMID:19229343

  6. NeuronMetrics: Software for Semi-Automated Processing of Cultured-Neuron Images

    PubMed Central

    Narro, Martha L.; Yang, Fan; Kraft, Robert; Wenk, Carola; Efrat, Alon; Restifo, Linda L.

    2007-01-01

    Using primary cell culture to screen for changes in neuronal morphology requires specialized analysis software. We developed NeuronMetrics™ for semi-automated, quantitative analysis of two-dimensional (2D) images of fluorescently labeled cultured neurons. It skeletonizes the neuron image using two complementary image-processing techniques, capturing fine terminal neurites with high fidelity. An algorithm was devised to span wide gaps in the skeleton. NeuronMetrics uses a novel strategy based on geometric features called faces to extract a branch-number estimate from complex arbors with numerous neurite-to-neurite contacts, without creating a precise, contact-free representation of the neurite arbor. It estimates total neurite length, branch number, primary neurite number, territory (the area of the convex polygon bounding the skeleton and cell body), and Polarity Index (a measure of neuronal polarity). These parameters provide fundamental information about the size and shape of neurite arbors, which are critical factors for neuronal function. NeuronMetrics streamlines optional manual tasks such as removing noise, isolating the largest primary neurite, and correcting length for self-fasciculating neurites. Numeric data are output in a single text file, readily imported into other applications for further analysis. Written as modules for ImageJ, NeuronMetrics provides practical analysis tools that are easy to use and support batch processing. Depending on the need for manual intervention, processing time for a batch of ~60 2D images is 1.0–2.5 hours, from a folder of images to a table of numeric data. NeuronMetrics’ output accelerates the quantitative detection of mutations and chemical compounds that alter neurite morphology in vitro, and will contribute to the use of cultured neurons for drug discovery. PMID:17270152

  7. An image-based software tool for screening retinal fundus images using vascular morphology and network transport analysis

    NASA Astrophysics Data System (ADS)

    Clark, Richard D.; Dickrell, Daniel J.; Meadows, David L.

    2014-03-01

    As the number of digital retinal fundus images taken each year grows at an increasing rate, there exists a similarly increasing need for automatic eye disease detection through image-based analysis. A new method has been developed for classifying standard color fundus photographs into both healthy and diseased categories. This classification was based on the calculated network fluid conductance, a function of the geometry and connectivity of the vascular segments. To evaluate the network resistance, the retinal vasculature was first manually separated from the background to ensure an accurate representation of the geometry and connectivity. The arterial and venous networks were then semi-automatically separated into two separate binary images. The connectivity of the arterial network was then determined through a series of morphological image operations. The network comprised of segments of vasculature and points of bifurcation, with each segment having a characteristic geometric and fluid properties. Based on the connectivity and fluid resistance of each vascular segment, an arterial network flow conductance was calculated, which described the ease with which blood can pass through a vascular system. In this work, 27 eyes (13 healthy and 14 diabetic) from patients roughly 65 years in age were evaluated using this methodology. Healthy arterial networks exhibited an average fluid conductance of 419 ± 89 μm3/mPa-s while the average network fluid conductance of the diabetic set was 165 ± 87 μm3/mPa-s (p < 0.001). The results of this new image-based software demonstrated an ability to automatically, quantitatively and efficiently screen diseased eyes from color fundus imagery.

  8. Software-based high-level synthesis design of FPGA beamformers for synthetic aperture imaging.

    PubMed

    Amaro, Joao; Yiu, Billy Y S; Falcao, Gabriel; Gomes, Marco A C; Yu, Alfred C H

    2015-05-01

    Field-programmable gate arrays (FPGAs) can potentially be configured as beamforming platforms for ultrasound imaging, but a long design time and skilled expertise in hardware programming are typically required. In this article, we present a novel approach to the efficient design of FPGA beamformers for synthetic aperture (SA) imaging via the use of software-based high-level synthesis techniques. Software kernels (coded in OpenCL) were first developed to stage-wise handle SA beamforming operations, and their corresponding FPGA logic circuitry was emulated through a high-level synthesis framework. After design space analysis, the fine-tuned OpenCL kernels were compiled into register transfer level descriptions to configure an FPGA as a beamformer module. The processing performance of this beamformer was assessed through a series of offline emulation experiments that sought to derive beamformed images from SA channel-domain raw data (40-MHz sampling rate, 12 bit resolution). With 128 channels, our FPGA-based SA beamformer can achieve 41 frames per second (fps) processing throughput (3.44 × 10(8) pixels per second for frame size of 256 × 256 pixels) at 31.5 W power consumption (1.30 fps/W power efficiency). It utilized 86.9% of the FPGA fabric and operated at a 196.5 MHz clock frequency (after optimization). Based on these findings, we anticipate that FPGA and high-level synthesis can together foster rapid prototyping of real-time ultrasound processor modules at low power consumption budgets. PMID:25965680

  9. Workspace Safe Operation of a Force- or Impedance-Controlled Robot

    NASA Technical Reports Server (NTRS)

    Abdallah, Muhammad E. (Inventor); Hargrave, Brian (Inventor); Yamokoski, John D. (Inventor); Strawser, Philip A. (Inventor)

    2013-01-01

    A method of controlling a robotic manipulator of a force- or impedance-controlled robot within an unstructured workspace includes imposing a saturation limit on a static force applied by the manipulator to its surrounding environment, and may include determining a contact force between the manipulator and an object in the unstructured workspace, and executing a dynamic reflex when the contact force exceeds a threshold to thereby alleviate an inertial impulse not addressed by the saturation limited static force. The method may include calculating a required reflex torque to be imparted by a joint actuator to a robotic joint. A robotic system includes a robotic manipulator having an unstructured workspace and a controller that is electrically connected to the manipulator, and which controls the manipulator using force- or impedance-based commands. The controller, which is also disclosed herein, automatically imposes the saturation limit and may execute the dynamic reflex noted above.

  10. Creating the optimal workspace for hospital staff using human centred design.

    PubMed

    Cawood, T; Saunders, E; Drennan, C; Cross, N; Nicholl, D; Kenny, A; Meates, D; Laing, R

    2016-07-01

    We were tasked with creating best possible non-clinical workspace solutions for approximately 450 hospital staff across 11 departments encompassing medical, nursing, allied health, administrative and other support staff. We used a Human-Centred Design process, involving 'Hear, Create and Deliver' stages. We used observations, contextual enquiry and role-specific workshops to understand needs, key interactions and drivers of behaviour. Co-design workshops were then used to explore and prototype-test concepts for the final design. With extensive employee engagement and design process expertise, an innovative solution was created that focussed on meeting the functional workspace needs of a diverse group of staff requiring a range of different spaces, incorporating space constraints and equity. This project demonstrated the strength of engaging employees in an expert-led Human-Centred Design process. We believe this is a successful blueprint process for other institutions to embrace when facing similar workspace design challenges. PMID:27405891

  11. Pre-Hardware Optimization of Spacecraft Image Processing Software Algorithms and Hardware Implementation

    NASA Technical Reports Server (NTRS)

    Kizhner, Semion; Flatley, Thomas P.; Hestnes, Phyllis; Jentoft-Nilsen, Marit; Petrick, David J.; Day, John H. (Technical Monitor)

    2001-01-01

    Spacecraft telemetry rates have steadily increased over the last decade presenting a problem for real-time processing by ground facilities. This paper proposes a solution to a related problem for the Geostationary Operational Environmental Spacecraft (GOES-8) image processing application. Although large super-computer facilities are the obvious heritage solution, they are very costly, making it imperative to seek a feasible alternative engineering solution at a fraction of the cost. The solution is based on a Personal Computer (PC) platform and synergy of optimized software algorithms and re-configurable computing hardware technologies, such as Field Programmable Gate Arrays (FPGA) and Digital Signal Processing (DSP). It has been shown in [1] and [2] that this configuration can provide superior inexpensive performance for a chosen application on the ground station or on-board a spacecraft. However, since this technology is still maturing, intensive pre-hardware steps are necessary to achieve the benefits of hardware implementation. This paper describes these steps for the GOES-8 application, a software project developed using Interactive Data Language (IDL) (Trademark of Research Systems, Inc.) on a Workstation/UNIX platform. The solution involves converting the application to a PC/Windows/RC platform, selected mainly by the availability of low cost, adaptable high-speed RC hardware. In order for the hybrid system to run, the IDL software was modified to account for platform differences. It was interesting to examine the gains and losses in performance on the new platform, as well as unexpected observations before implementing hardware. After substantial pre-hardware optimization steps, the necessity of hardware implementation for bottleneck code in the PC environment became evident and solvable beginning with the methodology described in [1], [2], and implementing a novel methodology for this specific application [6]. The PC-RC interface bandwidth problem for the

  12. JHelioviewer: Open-Source Software for Discovery and Image Access in the Petabyte Age (Invited)

    NASA Astrophysics Data System (ADS)

    Mueller, D.; Dimitoglou, G.; Langenberg, M.; Pagel, S.; Dau, A.; Nuhn, M.; Garcia Ortiz, J. P.; Dietert, H.; Schmidt, L.; Hughitt, V. K.; Ireland, J.; Fleck, B.

    2010-12-01

    The unprecedented torrent of data returned by the Solar Dynamics Observatory is both a blessing and a barrier: a blessing for making available data with significantly higher spatial and temporal resolution, but a barrier for scientists to access, browse and analyze them. With such staggering data volume, the data is bound to be accessible only from a few repositories and users will have to deal with data sets effectively immobile and practically difficult to download. From a scientist's perspective this poses three challenges: accessing, browsing and finding interesting data while avoiding the proverbial search for a needle in a haystack. To address these challenges, we have developed JHelioviewer, an open-source visualization software that lets users browse large data volumes both as still images and movies. We did so by deploying an efficient image encoding, storage, and dissemination solution using the JPEG 2000 standard. This solution enables users to access remote images at different resolution levels as a single data stream. Users can view, manipulate, pan, zoom, and overlay JPEG 2000 compressed data quickly, without severe network bandwidth penalties. Besides viewing data, the browser provides third-party metadata and event catalog integration to quickly locate data of interest, as well as an interface to the Virtual Solar Observatory to download science-quality data. As part of the Helioviewer Project, JHelioviewer offers intuitive ways to browse large amounts of heterogeneous data remotely and provides an extensible and customizable open-source platform for the scientific community.

  13. HydroImage: A New Software for HydroGeophysical and BioGeophysical Data Integration

    NASA Astrophysics Data System (ADS)

    Suribhatla, R. M.; Mok, C. M.; Kaback, D.; Chen, J.; Hubbard, S. S.

    2011-12-01

    Hydrogeophysical and biogeophysical data integration have recently emerged as cost-effective and rapid techniques for improving subsurface characterization and monitoring. In a Bayesian framework for integration, borehole based data provide prior distribution and geophysical information serve as data to update the prior through likelihood functions obtained from petrophysical models between borehole and cross-well data. We present the application of a Windows-based software called HydroImage that uses this Bayesian framework for data integration and visualization. HydroImage can be used for geostatistical estimation, geophysical tomographic inversion, petrophysical model development, and Bayesian integration. We demonstrate HydroImage using three different field datasets to estimate different subsurface states or parameters. The first example combines wellbore flowmeter test data and crosshole seismic and ground penetrating radar (GPR) data to estimate hydraulic conductivity at the DOE Bacterial Transport Site in Oyster, Virginia. The second example focuses on using time-lapse radar data to estimate moisture content dynamics associated with a desiccation test performed to remediate the deep vadose zone in Hanford, Washington. The third example demonstrates the use of spectral induced polarization data to estimate the spatial and temporal distribution of geochemical parameters that are indicative of the redox state of a contaminated aquifer.

  14. Computation of robot configuration and workspaces via the Fourier transform on the discrete-motion group

    SciTech Connect

    Kyatkin, A.B.; Chirikjian, G.S.

    1999-06-01

    The authors apply the Fourier transform on the discrete-motion group to the problem of computing the configuration-space obstacles of mobile robots which move among static obstacles, the problem of finding the workspace density of binary manipulators with many actuators, and the problem of determining workspace boundaries of manipulators with continuous-motion actuators. They develop and implement Fourier transforms for the discrete-motion group of the plane. These transforms allow them to apply fast Fourier transform methods to the computation of convolution-like integrals that arise in robot kinematics and motion planning. The results of the implementation are discussed for particular examples.

  15. Quantitative Neuroimaging Software for Clinical Assessment of Hippocampal Volumes on MR Imaging

    PubMed Central

    Ahdidan, Jamila; Raji, Cyrus A.; DeYoe, Edgar A.; Mathis, Jedidiah; Noe, Karsten Ø.; Rimestad, Jens; Kjeldsen, Thomas K.; Mosegaard, Jesper; Becker, James T.; Lopez, Oscar

    2015-01-01

    Background: Multiple neurological disorders including Alzheimer’s disease (AD), mesial temporal sclerosis, and mild traumatic brain injury manifest with volume loss on brain MRI. Subtle volume loss is particularly seen early in AD. While prior research has demonstrated the value of this additional information from quantitative neuroimaging, very few applications have been approved for clinical use. Here we describe a US FDA cleared software program, NeuroreaderTM, for assessment of clinical hippocampal volume on brain MRI. Objective: To present the validation of hippocampal volumetrics on a clinical software program. Method: Subjects were drawn (n = 99) from the Alzheimer Disease Neuroimaging Initiative study. Volumetric brain MR imaging was acquired in both 1.5 T (n = 59) and 3.0 T (n = 40) scanners in participants with manual hippocampal segmentation. Fully automated hippocampal segmentation and measurement was done using a multiple atlas approach. The Dice Similarity Coefficient (DSC) measured the level of spatial overlap between NeuroreaderTM and gold standard manual segmentation from 0 to 1 with 0 denoting no overlap and 1 representing complete agreement. DSC comparisons between 1.5 T and 3.0 T scanners were done using standard independent samples T-tests. Results: In the bilateral hippocampus, mean DSC was 0.87 with a range of 0.78–0.91 (right hippocampus) and 0.76–0.91 (left hippocampus). Automated segmentation agreement with manual segmentation was essentially equivalent at 1.5 T (DSC = 0.879) versus 3.0 T (DSC = 0.872). Conclusion: This work provides a description and validation of a software program that can be applied in measuring hippocampal volume, a biomarker that is frequently abnormal in AD and other neurological disorders. PMID:26484924

  16. Measuring the area of tear film break-up by image analysis software

    NASA Astrophysics Data System (ADS)

    Pena-Verdeal, Hugo; García-Resúa, Carlos; Ramos, Lucía.; Mosquera, Antonio; Yebra-Pimentel, Eva; Giráldez, María. Jesús

    2013-11-01

    Tear film breakup time (BUT) test only examines the first break in the tear film, but subsequent tear film events are not monitored. We present a method of measuring the area of breakup after the appearance of the first breakup by using open source software. Furthermore, the speed of the rupture was determined. 84 subjects participated in the study. 2 μl volume of 2% sodium fluorescein was instilled using a micropipette. The subject was seated behind a slit-lamp using a cobalt blue filter together with a Wratten 12 yellow filter. Then, the tear film was recorded by a camera attached to the slit lamp. 4 frames of each video was extracted, the first rupture (BUT_0), breakup after 1 second (BUT_1), rupture after 2 seconds (BUT_2) and breakup before the last blink (BUT_F). Open source software of measurement based on Java (NIH ImageJ) was used to measure the number of pixels in areas of breakup. These areas were divided by the area of exposed cornea to obtain the percentage of ruptures. Instantaneous breakup speed was calculated for second 1 as the difference between BUT_1 - BUT_0, whereas instant speed for second 2 was BUT_2 - BUT_1. Mean area of breakup obtained was: BUT_0 = 0.26%, BUT_1 = 0.48%, BUT_2 = 0.79% and BUT_F = 1.61%. Break speed was 0.22 area/sec for second 1 and 0.31 area/sec for second 2, showing a statistical difference between them (p = 0.007). Post BUT analysis may be easily monitoring with the aid of this software.

  17. Digital mapping of side-scan sonar data with the Woods Hole Image Processing System software

    USGS Publications Warehouse

    Paskevich, Valerie F.

    1992-01-01

    Since 1985, the Branch of Atlantic Marine Geology has been involved in collecting, processing and digitally mosaicking high and low resolution sidescan sonar data. In the past, processing and digital mosaicking has been accomplished with a dedicated, shore-based computer system. Recent development of a UNIX-based image-processing software system includes a series of task specific programs for pre-processing sidescan sonar data. To extend the capabilities of the UNIX-based programs, development of digital mapping techniques have been developed. This report describes the initial development of an automated digital mapping procedure. Included is a description of the programs and steps required to complete the digital mosaicking on a UNIXbased computer system, and a comparison of techniques that the user may wish to select.

  18. Integration of XNAT/PACS, DICOM, and research software for automated multi-modal image analysis

    NASA Astrophysics Data System (ADS)

    Gao, Yurui; Burns, Scott S.; Lauzon, Carolyn B.; Fong, Andrew E.; James, Terry A.; Lubar, Joel F.; Thatcher, Robert W.; Twillie, David A.; Wirt, Michael D.; Zola, Marc A.; Logan, Bret W.; Anderson, Adam W.; Landman, Bennett A.

    2013-03-01

    Traumatic brain injury (TBI) is an increasingly important public health concern. While there are several promising avenues of intervention, clinical assessments are relatively coarse and comparative quantitative analysis is an emerging field. Imaging data provide potentially useful information for evaluating TBI across functional, structural, and microstructural phenotypes. Integration and management of disparate data types are major obstacles. In a multi-institution collaboration, we are collecting electroencephalogy (EEG), structural MRI, diffusion tensor MRI (DTI), and single photon emission computed tomography (SPECT) from a large cohort of US Army service members exposed to mild or moderate TBI who are undergoing experimental treatment. We have constructed a robust informatics backbone for this project centered on the DICOM standard and eXtensible Neuroimaging Archive Toolkit (XNAT) server. Herein, we discuss (1) optimization of data transmission, validation and storage, (2) quality assurance and workflow management, and (3) integration of high performance computing with research software.

  19. SU-E-J-264: Comparison of Two Commercially Available Software Platforms for Deformable Image Registration

    SciTech Connect

    Tuohy, R; Stathakis, S; Mavroidis, P; Bosse, C; Papanikolaou, N

    2014-06-01

    Purpose: To evaluate and compare the deformable image registration algorithms available in the Velocity (Velocity Medical Solutions, Atlanta, GA) and RayStation (RaySearch Americas, Inc., Garden city NY). Methods: Ten consecutive patient cone beam CTs (CBCT) for each fraction were collected. The CBCTs along with the simulation CT were exported to the Velocity and the RayStation software. Each CBCT was registered using deformable image registration to the simulation CT and the resulting deformable vector matrix was generated. Each registration was visually inspected by a physicist and the prescribing physician. The volumes of the critical organs were calculated for each deformable CT and used for comparison. Results: The resulting deformable registrations revealed differences between the two algorithms. These differences were realized when the organs at risk were contoured on each deformed CBCT. Differences in the order of 10% ±30% in volume were observed for bladder, 17 ±21% for rectum and 16±10% for sigmoid. The prostate and PTV volume differences were in the order of 3±5%. The volumetric differences observed had a respective impact on the DVHs of all organs at risk. Differences of 8–10% in the mean dose were observed for all organs above. Conclusion: Deformable registration is a powerful tool that aids in the definition of critical structures and is often used for the evaluation of daily dose delivered to the patient. It should be noted that extended QA should be performed before clinical implementation of the software and the users should be aware of advantages and limitations of the methods.

  20. Comparison between three methods to value lower tear meniscus measured by image software

    NASA Astrophysics Data System (ADS)

    García-Resúa, Carlos; Pena-Verdeal, Hugo; Lira, Madalena; Oliveira, M. Elisabete Real; Giráldez, María. Jesús; Yebra-Pimentel, Eva

    2013-11-01

    To measure different parameters of lower tear meniscus height (TMH) by using photography with open software of measurement. TMH was addressed from lower eyelid to the top of the meniscus (absolute TMH) and to the brightest meniscus reflex (reflex TMH). 121 young healthy subjects were included in the study. The lower tear meniscus was videotaped by a digital camera attached to a slit lamp. Three videos were recorded in central meniscus portion on three different methods: slit lamp without fluorescein instillation, slit lamp with fluorescein instillation and TearscopeTM without fluorescein instillation. Then, a masked observed obtained an image from each video and measured TMH by using open source software of measurement based on Java (NIH ImageJ). Absolute central (TMH-CA), absolute with fluorescein (TMH-F) and absolute using the Tearscope (TMH-Tc) were compared each other as well as reflex central (TMH-CR) and reflex Tearscope (TMH-TcR). Mean +/- S.D. values of TMH-CA, TMH-CR, TMH-F, TMH-Tc and TMH-TcR of 0.209 +/- 0.049, 0.139 +/- 0.031, 0.222 +/- 0.058, 0.175 +/- 0.045 and 0.109 +/- 0.029 mm, respectively were found. Paired t-test was performed for the relationship between TMH-CA - TMH-CR, TMH-CA - TMH-F, TMH-CA - TMH-Tc, TMH-F - TMH-Tc, TMH-Tc - TMH-TcR and TMH-CR - TMH-TcR. In all cases, it was found a significant difference between both variables (all p < 0.008). This study showed a useful tool to objectively measure TMH by photography. Eye care professionals should maintain the same TMH parameter in the follow-up visits, due to the difference between them.

  1. Reliability evaluation of I-123 ADAM SPECT imaging using SPM software and AAL ROI methods

    NASA Astrophysics Data System (ADS)

    Yang, Bang-Hung; Tsai, Sung-Yi; Wang, Shyh-Jen; Su, Tung-Ping; Chou, Yuan-Hwa; Chen, Chia-Chieh; Chen, Jyh-Cheng

    2011-08-01

    The level of serotonin was regulated by serotonin transporter (SERT), which is a decisive protein in regulation of serotonin neurotransmission system. Many psychiatric disorders and therapies were also related to concentration of cerebral serotonin. I-123 ADAM was the novel radiopharmaceutical to image SERT in brain. The aim of this study was to measure reliability of SERT densities of healthy volunteers by automated anatomical labeling (AAL) method. Furthermore, we also used statistic parametric mapping (SPM) on a voxel by voxel analysis to find difference of cortex between test and retest of I-123 ADAM single photon emission computed tomography (SPECT) images.Twenty-one healthy volunteers were scanned twice with SPECT at 4 h after intravenous administration of 185 MBq of 123I-ADAM. The image matrix size was 128×128 and pixel size was 3.9 mm. All images were obtained through filtered back-projection (FBP) reconstruction algorithm. Region of interest (ROI) definition was performed based on the AAL brain template in PMOD version 2.95 software package. ROI demarcations were placed on midbrain, pons, striatum, and cerebellum. All images were spatially normalized to the SPECT MNI (Montreal Neurological Institute) templates supplied with SPM2. And each image was transformed into standard stereotactic space, which was matched to the Talairach and Tournoux atlas. Then differences across scans were statistically estimated on a voxel by voxel analysis using paired t-test (population main effect: 2 cond's, 1 scan/cond.), which was applied to compare concentration of SERT between the test and retest cerebral scans.The average of specific uptake ratio (SUR: target/cerebellum-1) of 123I-ADAM binding to SERT in midbrain was 1.78±0.27, pons was 1.21±0.53, and striatum was 0.79±0.13. The cronbach's α of intra-class correlation coefficient (ICC) was 0.92. Besides, there was also no significant statistical finding in cerebral area using SPM2 analysis. This finding might help us

  2. A flexible software architecture for scalable real-time image and video processing applications

    NASA Astrophysics Data System (ADS)

    Usamentiaga, Rubén; Molleda, Julio; García, Daniel F.; Bulnes, Francisco G.

    2012-06-01

    Real-time image and video processing applications require skilled architects, and recent trends in the hardware platform make the design and implementation of these applications increasingly complex. Many frameworks and libraries have been proposed or commercialized to simplify the design and tuning of real-time image processing applications. However, they tend to lack flexibility because they are normally oriented towards particular types of applications, or they impose specific data processing models such as the pipeline. Other issues include large memory footprints, difficulty for reuse and inefficient execution on multicore processors. This paper presents a novel software architecture for real-time image and video processing applications which addresses these issues. The architecture is divided into three layers: the platform abstraction layer, the messaging layer, and the application layer. The platform abstraction layer provides a high level application programming interface for the rest of the architecture. The messaging layer provides a message passing interface based on a dynamic publish/subscribe pattern. A topic-based filtering in which messages are published to topics is used to route the messages from the publishers to the subscribers interested in a particular type of messages. The application layer provides a repository for reusable application modules designed for real-time image and video processing applications. These modules, which include acquisition, visualization, communication, user interface and data processing modules, take advantage of the power of other well-known libraries such as OpenCV, Intel IPP, or CUDA. Finally, we present different prototypes and applications to show the possibilities of the proposed architecture.

  3. Global workspace dynamics: cortical "binding and propagation" enables conscious contents.

    PubMed

    Baars, Bernard J; Franklin, Stan; Ramsoy, Thomas Zoega

    2013-01-01

    A global workspace (GW) is a functional hub of binding and propagation in a population of loosely coupled signaling elements. In computational applications, GW architectures recruit many distributed, specialized agents to cooperate in resolving focal ambiguities. In the brain, conscious experiences may reflect a GW function. For animals, the natural world is full of unpredictable dangers and opportunities, suggesting a general adaptive pressure for brains to resolve focal ambiguities quickly and accurately. GW theory aims to understand the differences between conscious and unconscious brain events. In humans and related species the cortico-thalamic (C-T) core is believed to underlie conscious aspects of perception, thinking, learning, feelings of knowing (FOK), felt emotions, visual imagery, working memory, and executive control. Alternative theoretical perspectives are also discussed. The C-T core has many anatomical hubs, but conscious percepts are unitary and internally consistent at any given moment. Over time, conscious contents constitute a very large, open set. This suggests that a brain-based GW capacity cannot be localized in a single anatomical hub. Rather, it should be sought in a functional hub - a dynamic capacity for binding and propagation of neural signals over multiple task-related networks, a kind of neuronal cloud computing. In this view, conscious contents can arise in any region of the C-T core when multiple input streams settle on a winner-take-all equilibrium. The resulting conscious gestalt may ignite an any-to-many broadcast, lasting ∼100-200 ms, and trigger widespread adaptation in previously established networks. To account for the great range of conscious contents over time, the theory suggests an open repertoire of binding coalitions that can broadcast via theta/gamma or alpha/gamma phase coupling, like radio channels competing for a narrow frequency band. Conscious moments are thought to hold only 1-4 unrelated items; this small

  4. Software for Collaborative Use of Large Interactive Displays

    NASA Technical Reports Server (NTRS)

    Trimble, Jay; Shab, Thodore; Wales, Roxana; Vera, Alonso; Tollinger, Irene; McCurdy, Michael; Lyubimov, Dmitriy

    2006-01-01

    The MERBoard Collaborative Workspace, which is currently being deployed to support the Mars Exploration Rover (MER) Missions, is the first instantiation of a new computing architecture designed to support collaborative and group computing using computing devices situated in NASA mission operations room. It is a software system for generation of large-screen interactive displays by multiple users

  5. SU-E-I-13: Evaluation of Metal Artifact Reduction (MAR) Software On Computed Tomography (CT) Images

    SciTech Connect

    Huang, V; Kohli, K

    2015-06-15

    Purpose: A new commercially available metal artifact reduction (MAR) software in computed tomography (CT) imaging was evaluated with phantoms in the presence of metals. The goal was to assess the ability of the software to restore the CT number in the vicinity of the metals without impacting the image quality. Methods: A Catphan 504 was scanned with a GE Optima RT 580 CT scanner (GE Healthcare, Milwaukee, WI) and the images were reconstructed with and without the MAR software. Both datasets were analyzed with Image Owl QA software (Image Owl Inc, Greenwich, NY). CT number sensitometry, MTF, low contrast, uniformity, noise and spatial accuracy were compared for scans with and without MAR software. In addition, an in-house made phantom was scanned with and without a stainless steel insert at three different locations. The accuracy of the CT number and metal insert dimension were investigated as well. Results: Comparisons between scans with and without MAR algorithm on the Catphan phantom demonstrate similar results for image quality. However, noise was slightly higher for the MAR algorithm. Evaluation of the CT number at various locations of the in-house made phantom was also performed. The baseline HU, obtained from the scan without metal insert, was compared to scans with the stainless steel insert at 3 different locations. The HU difference between the baseline scan versus metal scan was improved when the MAR algorithm was applied. In addition, the physical diameter of the stainless steel rod was over-estimated by the MAR algorithm by 0.9 mm. Conclusion: This work indicates with the presence of metal in CT scans, the MAR algorithm is capable of providing a more accurate CT number without compromising the overall image quality. Future work will include the dosimetric impact on the MAR algorithm.

  6. Fostering Collaborative Knowledge Construction in a Video-Based Learning Setting: Effects of a Shared Workspace and a Content-Specific Graphical Representation

    ERIC Educational Resources Information Center

    Hron, Aemilian; Cress, Ulrike; Hammer, Karsten; Friedrich, Helmut-Felix

    2007-01-01

    This study examined means of fostering videoconference-based collaborative learning. An experiment was conducted with 15 learning dyads divided into three conditions of videoconference-based learning: without shared workspace, with shared workspace and with shared workspace plus a content-specific graphical representation. Compared with those with…

  7. Xmipp 3.0: an improved software suite for image processing in electron microscopy.

    PubMed

    de la Rosa-Trevín, J M; Otón, J; Marabini, R; Zaldívar, A; Vargas, J; Carazo, J M; Sorzano, C O S

    2013-11-01

    Xmipp is a specialized software package for image processing in electron microscopy, and that is mainly focused on 3D reconstruction of macromolecules through single-particles analysis. In this article we present Xmipp 3.0, a major release which introduces several improvements and new developments over the previous version. A central improvement is the concept of a project that stores the entire processing workflow from data import to final results. It is now possible to monitor, reproduce and restart all computing tasks as well as graphically explore the complete set of interrelated tasks associated to a given project. Other graphical tools have also been improved such as data visualization, particle picking and parameter "wizards" that allow the visual selection of some key parameters. Many standard image formats are transparently supported for input/output from all programs. Additionally, results have been standardized, facilitating the interoperation between different Xmipp programs. Finally, as a result of a large code refactoring, the underlying C++ libraries are better suited for future developments and all code has been optimized. Xmipp is an open-source package that is freely available for download from: http://xmipp.cnb.csic.es. PMID:24075951

  8. The Importance of Structuring Information and Resources within Shared Workspaces during Collaborative Design Learning

    ERIC Educational Resources Information Center

    Nicol, David; Littlejohn, Allison; Grierson, Hilary

    2005-01-01

    This paper investigates how the organization or structure of information and resources in shared workspaces influences team sharing and design learning. Two groupware products, BSCW and TikiWiki, were configured so that teams could structure and share resources. In BSCW the resources were structured hierarchically using folders and subfolders…

  9. Revolute manipulator workspace optimization using a modified bacteria foraging algorithm: A comparative study

    NASA Astrophysics Data System (ADS)

    Panda, S.; Mishra, D.; Biswal, B. B.; Tripathy, M.

    2014-02-01

    Robotic manipulators with three-revolute (3R) motions to attain desired positional configurations are very common in industrial robots. The capability of these robots depends largely on the workspace of the manipulator in addition to other parameters. In this study, an evolutionary optimization algorithm based on the foraging behaviour of the Escherichia coli bacteria present in the human intestine is utilized to optimize the workspace volume of a 3R manipulator. The new optimization method is modified from the original algorithm for faster convergence. This method is also useful for optimization problems in a highly constrained environment, such as robot workspace optimization. The new approach for workspace optimization of 3R manipulators is tested using three cases. The test results are compared with standard results available using other optimization algorithms, i.e. the differential evolution algorithm, the genetic algorithm and the particle swarm optimization algorithm. The present method is found to be superior to the other methods in terms of computational efficiency.

  10. A Neuro-genetic Control Scheme Application for Industrial R 3 Workspaces

    NASA Astrophysics Data System (ADS)

    Irigoyen, E.; Larrea, M.; Valera, J.; Gómez, V.; Artaza, F.

    This work presents a neuro-genetic control scheme for a R 3 workspace application. The solution is based on a Multi Objective Genetic Algorithm reference generator and an Adaptive Predictive Neural Network Controller. Crane position control is presented as an application of the proposed control scheme.

  11. Investigating Uses and Perceptions of an Online Collaborative Workspace for the Dissertation Process

    ERIC Educational Resources Information Center

    Rockinson-Szapkiw, Amanda J.

    2012-01-01

    The intent of this study was to investigate 93 doctoral candidates' perceptions and use of an online collaboration workspace and content management server, Microsoft Office SharePoint, for dissertation process. All candidates were enrolled in an Ed.D. programme in the United States. Descriptive statistics demonstrate that candidates frequently use…

  12. An adaptive workspace hypothesis about the neural correlates of consciousness: insights from neuroscience and meditation studies.

    PubMed

    Raffone, Antonino; Srinivasan, Narayanan

    2009-01-01

    While enormous progress has been made to identify neural correlates of consciousness (NCC), crucial NCC aspects are still very controversial. A major hurdle is the lack of an adequate definition and characterization of different aspects of conscious experience and also its relationship to attention and metacognitive processes like monitoring. In this paper, we therefore attempt to develop a unitary theoretical framework for NCC, with an interdependent characterization of endogenous attention, access consciousness, phenomenal awareness, metacognitive consciousness, and a non-referential form of unified consciousness. We advance an adaptive workspace hypothesis about the NCC based on the global workspace model emphasizing transient resonant neurodynamics and prefrontal cortex function, as well as meditation-related characterizations of conscious experiences. In this hypothesis, transient dynamic links within an adaptive coding net in prefrontal cortex, especially in anterior prefrontal cortex, and between it and the rest of the brain, in terms of ongoing intrinsic and long-range signal exchanges, flexibly regulate the interplay between endogenous attention, access consciousness, phenomenal awareness, and metacognitive consciousness processes. Such processes are established in terms of complementary aspects of an ongoing transition between context-sensitive global workspace assemblies, modulated moment-to-moment by body and environment states. Brain regions associated to momentary interoceptive and exteroceptive self-awareness, or first-person experiential perspective as emphasized in open monitoring meditation, play an important modulatory role in adaptive workspace transitions. PMID:19733756

  13. JHelioviewer: Open-Source Software for Discovery and Image Access in the Petabyte Age

    NASA Astrophysics Data System (ADS)

    Mueller, D.; Dimitoglou, G.; Garcia Ortiz, J.; Langenberg, M.; Nuhn, M.; Dau, A.; Pagel, S.; Schmidt, L.; Hughitt, V. K.; Ireland, J.; Fleck, B.

    2011-12-01

    The unprecedented torrent of data returned by the Solar Dynamics Observatory is both a blessing and a barrier: a blessing for making available data with significantly higher spatial and temporal resolution, but a barrier for scientists to access, browse and analyze them. With such staggering data volume, the data is accessible only from a few repositories and users have to deal with data sets effectively immobile and practically difficult to download. From a scientist's perspective this poses three challenges: accessing, browsing and finding interesting data while avoiding the proverbial search for a needle in a haystack. To address these challenges, we have developed JHelioviewer, an open-source visualization software that lets users browse large data volumes both as still images and movies. We did so by deploying an efficient image encoding, storage, and dissemination solution using the JPEG 2000 standard. This solution enables users to access remote images at different resolution levels as a single data stream. Users can view, manipulate, pan, zoom, and overlay JPEG 2000 compressed data quickly, without severe network bandwidth penalties. Besides viewing data, the browser provides third-party metadata and event catalog integration to quickly locate data of interest, as well as an interface to the Virtual Solar Observatory to download science-quality data. As part of the ESA/NASA Helioviewer Project, JHelioviewer offers intuitive ways to browse large amounts of heterogeneous data remotely and provides an extensible and customizable open-source platform for the scientific community. In addition, the easy-to-use graphical user interface enables the general public and educators to access, enjoy and reuse data from space missions without barriers.

  14. Communities of Practice Transition Online - Lessons learned from NASA's EPO Online Workspace

    NASA Astrophysics Data System (ADS)

    Davey, B.

    2012-12-01

    The Earth Forum Education and Public Outreach (EP/O) community has long interacted to better their practice as a community as well as individually. Working together to share knowledge and grow, they function as a community of practice. In 2009, NASA designed and implemented an online workspace in hopes of promoting the communities continued interactions. This study examines the role of an online workspace component of a community in the work of a community of practice. Much has been studied revealing the importance of communities of practice to organizations, project success, and knowledge management and some of these same successes hold true for virtual communities of practice. Study participants were 75 Education and Public Outreach community members of NASA's Science Mission Directorate Earth Forum. In this mixed methods study, online workspace metrics were used to track participation and a survey completed by 21 members was used to quantify participation. For a more detailed analysis, 15 community members (five highly active users, five average users, and five infrequent users) selected based on survey responses, were interviewed. Finally, survey data was gathered from seven online facilitators to understand their role in the community. Data collected from these 21 community members and five facilitating members suggest that highly active users (logging into the workspace daily), were more likely to have transformative experiences, co-create knowledge, feel ownership of community knowledge, have extended opportunities for community exchange, and find new forms of evaluation. Average users shared some similar characteristics with both the highly active members and infrequent users, representing a group in transition as they become more engaged and active in the online workspace. Inactive users viewed the workspace as having little value, being difficult to navigate, being mainly for gaining basic information about events and community news, and as another demand

  15. Caltech/JPL Conference on Image Processing Technology, Data Sources and Software for Commercial and Scientific Applications

    NASA Technical Reports Server (NTRS)

    Redmann, G. H.

    1976-01-01

    Recent advances in image processing and new applications are presented to the user community to stimulate the development and transfer of this technology to industrial and commercial applications. The Proceedings contains 37 papers and abstracts, including many illustrations (some in color) and provides a single reference source for the user community regarding the ordering and obtaining of NASA-developed image-processing software and science data.

  16. Evaluation of linear array human papillomavirus genotyping using automatic optical imaging software.

    PubMed

    Jeronimo, J; Wentzensen, N; Long, R; Schiffman, M; Dunn, S T; Allen, R A; Walker, J L; Gold, M A; Zuna, R E; Sherman, M E; Wacholder, S; Wang, S S

    2008-08-01

    Variations in biological behavior suggest that each carcinogenic human papillomavirus (HPV) type should be considered individually in etiologic studies. HPV genotyping assays might have clinical applications if they are approved for use by the FDA. A widely used genotyping assay is the Roche Linear Array HPV genotyping test (LA). We used LA to genotype the HPV isolates from cervical specimens from women with the full spectrum of cervical disease: cervical cancer, cervical intraepithelial neoplasia (CIN), and HPV infections. To explore the feasibility and value of the automated reading of the LA results, we custom-designed novel, optical imaging software that provides optical density measurements of LA bands. We compared unmagnified visual examination with the automated measurements. The two measurements were highly associated. By either method, the threshold between a negative and a positive result was fairly sharp, with a clear bimodal distribution. Visually, most positive results were judged to be strong or medium, with fewer equivocal results categorized as weak (9.5% of positive samples), very weak (6.5% of positive samples), or extremely weak (7.7% of positive samples). The automated measurements of the intensities were significantly associated with the strength of the visual categories (P < 0.001). At the extremes of the automated signal intensities (< or = 20 units or > or = 120 units), the bands were almost always categorized visually as negative and positive, respectively. In the equivocal zone (20 to 119 units), specimens were more increasingly likely to be judged to be visually positive as the number of other, definite infections on the same strip increased (P for trend < 0.001). Multiple, concurrent infections comprise > or = 25% of HPV infections; thus, any systematic visual tendency that influences their evaluation when the result is equivocal should be minimized. Therefore, automated reading is probably worth development if easy-to-calibrate hardware

  17. Development of a viability standard curve for microencapsulated probiotic bacteria using confocal microscopy and image analysis software.

    PubMed

    Moore, Sarah; Kailasapathy, Kasipathy; Phillips, Michael; Jones, Mark R

    2015-07-01

    Microencapsulation is proposed to protect probiotic strains from food processing procedures and to maintain probiotic viability. Little research has described the in situ viability of microencapsulated probiotics. This study successfully developed a real-time viability standard curve for microencapsulated bacteria using confocal microscopy, fluorescent dyes and image analysis software. PMID:25887694

  18. Experiments with a novel content-based image retrieval software: can we eliminate classification systems in adolescent idiopathic scoliosis?

    PubMed

    Menon, K Venugopal; Kumar, Dinesh; Thomas, Tessamma

    2014-02-01

    Study Design Preliminary evaluation of new tool. Objective To ascertain whether the newly developed content-based image retrieval (CBIR) software can be used successfully to retrieve images of similar cases of adolescent idiopathic scoliosis (AIS) from a database to help plan treatment without adhering to a classification scheme. Methods Sixty-two operated cases of AIS were entered into the newly developed CBIR database. Five new cases of different curve patterns were used as query images. The images were fed into the CBIR database that retrieved similar images from the existing cases. These were analyzed by a senior surgeon for conformity to the query image. Results Within the limits of variability set for the query system, all the resultant images conformed to the query image. One case had no similar match in the series. The other four retrieved several images that were matching with the query. No matching case was left out in the series. The postoperative images were then analyzed to check for surgical strategies. Broad guidelines for treatment could be derived from the results. More precise query settings, inclusion of bending films, and a larger database will enhance accurate retrieval and better decision making. Conclusion The CBIR system is an effective tool for accurate documentation and retrieval of scoliosis images. Broad guidelines for surgical strategies can be made from the postoperative images of the existing cases without adhering to any classification scheme. PMID:24494177

  19. Design and evaluation of a THz time domain imaging system using standard optical design software.

    PubMed

    Brückner, Claudia; Pradarutti, Boris; Müller, Ralf; Riehemann, Stefan; Notni, Gunther; Tünnermann, Andreas

    2008-09-20

    A terahertz (THz) time domain imaging system is analyzed and optimized with standard optical design software (ZEMAX). Special requirements to the illumination optics and imaging optics are presented. In the optimized system, off-axis parabolic mirrors and lenses are combined. The system has a numerical aperture of 0.4 and is diffraction limited for field points up to 4 mm and wavelengths down to 750 microm. ZEONEX is used as the lens material. Higher aspherical coefficients are used for correction of spherical aberration and reduction of lens thickness. The lenses were manufactured by ultraprecision machining. For optimization of the system, ray tracing and wave-optical methods were combined. We show how the ZEMAX Gaussian beam analysis tool can be used to evaluate illumination optics. The resolution of the THz system was tested with a wire and a slit target, line gratings of different period, and a Siemens star. The behavior of the temporal line spread function can be modeled with the polychromatic coherent line spread function feature in ZEMAX. The spectral and temporal resolutions of the line gratings are compared with the respective modulation transfer function of ZEMAX. For maximum resolution, the system has to be diffraction limited down to the smallest wavelength of the spectrum of the THz pulse. Then, the resolution on time domain analysis of the pulse maximum can be estimated with the spectral resolution of the center of gravity wavelength. The system resolution near the optical axis on time domain analysis of the pulse maximum is 1 line pair/mm with an intensity contrast of 0.22. The Siemens star is used for estimation of the resolution of the whole system. An eight channel electro-optic sampling system was used for detection. The resolution on time domain analysis of the pulse maximum of all eight channels could be determined with the Siemens star to be 0.7 line pairs/mm. PMID:18806862

  20. A user-friendly LabVIEW software platform for grating based X-ray phase-contrast imaging.

    PubMed

    Wang, Shenghao; Han, Huajie; Gao, Kun; Wang, Zhili; Zhang, Can; Yang, Meng; Wu, Zhao; Wu, Ziyu

    2015-01-01

    X-ray phase-contrast imaging can provide greatly improved contrast over conventional absorption-based imaging for weakly absorbing samples, such as biological soft tissues and fibre composites. In this study, we introduced an easy and fast way to develop a user-friendly software platform dedicated to the new grating-based X-ray phase-contrast imaging setup at the National Synchrotron Radiation Laboratory of the University of Science and Technology of China. The control of 21 motorized stages, of a piezoelectric stage and of an X-ray tube are achieved with this software, it also covers image acquisition with a flat panel detector for automatic phase stepping scan. Moreover, a data post-processing module for signals retrieval and other custom features are in principle available. With a seamless integration of all the necessary functions in one software package, this platform greatly facilitate users' activities during experimental runs with this grating based X-ray phase contrast imaging setup. PMID:25882730

  1. Analyses of requirements for computer control and data processing experiment subsystems: Image data processing system (IDAPS) software description (7094 version), volume 2

    NASA Technical Reports Server (NTRS)

    1973-01-01

    A description of each of the software modules of the Image Data Processing System (IDAPS) is presented. The changes in the software modules are the result of additions to the application software of the system and an upgrade of the IBM 7094 Mod(1) computer to a 1301 disk storage configuration. Necessary information about IDAPS sofware is supplied to the computer programmer who desires to make changes in the software system or who desires to use portions of the software outside of the IDAPS system. Each software module is documented with: module name, purpose, usage, common block(s) description, method (algorithm of subroutine) flow diagram (if needed), subroutines called, and storage requirements.

  2. I-SPINE: a software package for advances in image-guided and minimally invasive spine procedures

    NASA Astrophysics Data System (ADS)

    Choi, Jae Jeong; Cleary, Kevin R.; Zeng, Jianchao; Gary, Kevin A.; Freedman, Matthew T.; Watson, Vance; Lindisch, David; Mun, Seong K.

    2000-05-01

    While image guidance is now routinely used in the brain in the form of frameless stereotaxy, it is beginning to be more widely used in other clinical areas such as the spine. At Georgetown University Medical Center, we are developing a program to provide advanced visualization and image guidance for minimally invasive spine procedures. This is a collaboration between an engineering-based research group and physicians from the radiology, neurosurgery, and orthopaedics departments. A major component of this work is the ISIS Center Spine Procedures Imaging and Navigation Engine, which is a software package under development as the base platform for technical advances.

  3. Army technology development. IBIS query. Software to support the Image Based Information System (IBIS) expansion for mapping, charting and geodesy

    NASA Technical Reports Server (NTRS)

    Friedman, S. Z.; Walker, R. E.; Aitken, R. B.

    1986-01-01

    The Image Based Information System (IBIS) has been under development at the Jet Propulsion Laboratory (JPL) since 1975. It is a collection of more than 90 programs that enable processing of image, graphical, tabular data for spatial analysis. IBIS can be utilized to create comprehensive geographic data bases. From these data, an analyst can study various attributes describing characteristics of a given study area. Even complex combinations of disparate data types can be synthesized to obtain a new perspective on spatial phenomena. In 1984, new query software was developed enabling direct Boolean queries of IBIS data bases through the submission of easily understood expressions. An improved syntax methodology, a data dictionary, and display software simplified the analysts' tasks associated with building, executing, and subsequently displaying the results of a query. The primary purpose of this report is to describe the features and capabilities of the new query software. A secondary purpose of this report is to compare this new query software to the query software developed previously (Friedman, 1982). With respect to this topic, the relative merits and drawbacks of both approaches are covered.

  4. ORBS: A data reduction software for the imaging Fourier transform spectrometers SpIOMM and SITELLE

    NASA Astrophysics Data System (ADS)

    Martin, T.; Drissen, L.; Joncas, G.

    2012-09-01

    SpIOMM (Spectromètre-Imageur de l'Observatoire du Mont Mégantic) is still the only operational astronomical Imaging Fourier Transform Spectrometer (IFTS) capable of obtaining the visible spectrum of every source of light in a field of view of 12 arc-minutes. Even if it has been designed to work with both outputs of the Michelson interferometer, up to now only one output has been used. Here we present ORBS (Outils de Réduction Binoculaire pour SpIOMM/SITELLE), the reduction software we designed in order to take advantage of the two output data. ORBS will also be used to reduce the data of SITELLE (Spectromètre-Imageur pour l' Étude en Long et en Large des raies d' Émissions) { the direct successor of SpIOMM, which will be in operation at the Canada-France- Hawaii Telescope (CFHT) in early 2013. SITELLE will deliver larger data cubes than SpIOMM (up to 2 cubes of 34 Go each). We thus have made a strong effort in optimizing its performance efficiency in terms of speed and memory usage in order to ensure the best compliance with the quality characteristics discussed with the CFHT team. As a result ORBS is now capable of reducing 68 Go of data in less than 20 hours using only 5 Go of random-access memory (RAM).

  5. 3D reconstruction of SEM images by use of optical photogrammetry software.

    PubMed

    Eulitz, Mona; Reiss, Gebhard

    2015-08-01

    Reconstruction of the three-dimensional (3D) surface of an object to be examined is widely used for structure analysis in science and many biological questions require information about their true 3D structure. For Scanning Electron Microscopy (SEM) there has been no efficient non-destructive solution for reconstruction of the surface morphology to date. The well-known method of recording stereo pair images generates a 3D stereoscope reconstruction of a section, but not of the complete sample surface. We present a simple and non-destructive method of 3D surface reconstruction from SEM samples based on the principles of optical close range photogrammetry. In optical close range photogrammetry a series of overlapping photos is used to generate a 3D model of the surface of an object. We adapted this method to the special SEM requirements. Instead of moving a detector around the object, the object itself was rotated. A series of overlapping photos was stitched and converted into a 3D model using the software commonly used for optical photogrammetry. A rabbit kidney glomerulus was used to demonstrate the workflow of this adaption. The reconstruction produced a realistic and high-resolution 3D mesh model of the glomerular surface. The study showed that SEM micrographs are suitable for 3D reconstruction by optical photogrammetry. This new approach is a simple and useful method of 3D surface reconstruction and suitable for various applications in research and teaching. PMID:26073969

  6. New image processing software for analyzing object size-frequency distributions, geometry, orientation, and spatial distribution

    NASA Astrophysics Data System (ADS)

    Beggan, Ciarán; Hamilton, Christopher W.

    2010-04-01

    Geological Image Analysis Software (GIAS) combines basic tools for calculating object area, abundance, radius, perimeter, eccentricity, orientation, and centroid location, with the first automated method for characterizing the aerial distribution of objects using sample-size-dependent nearest neighbor (NN) statistics. The NN analyses include tests for (1) Poisson, (2) Normalized Poisson, (3) Scavenged k=1, and (4) Scavenged k=2 NN distributions. GIAS is implemented in MATLAB with a Graphical User Interface (GUI) that is available as pre-parsed pseudocode for use with MATLAB, or as a stand-alone application that runs on Windows and Unix systems. GIAS can process raster data (e.g., satellite imagery, photomicrographs, etc.) and tables of object coordinates to characterize the size, geometry, orientation, and spatial organization of a wide range of geological features. This information expedites quantitative measurements of 2D object properties, provides criteria for validating the use of stereology to transform 2D object sections into 3D models, and establishes a standardized NN methodology that can be used to compare the results of different geospatial studies and identify objects using non-morphological parameters.

  7. ViewDEX: A java-based software for presentation and evaluation of medical images in observer performance studies

    NASA Astrophysics Data System (ADS)

    Håkansson, Markus; Svensson, Sune; Båth, Magnus; Månsson, Lars Gunnar

    2007-03-01

    Observer performance studies are time-consuming tasks, both for the participating observers and for the scientists collecting and analyzing the data. A possible way to optimize such studies is to perform the study in a completely digital environment. A software tool - ViewDEX (Viewer for Digital Evaluation of X-ray images) - has been developed in Java, enabling it to function on almost any computer. ViewDEX is a DICOM-compatible software tool that can be used to display medical images with simultaneous registration of the observer's response. ViewDEX is designed so that the user in a simple way can alter the types of questions and images presented to the observers, enabling ROC, MAFC and visual grading studies to be conducted in a fast and efficient way. The software can also be used for bench marking and for educational purposes. The results from each observer are saved in a log file, which can be exported for further analysis. The software is freely available for non-commercial purposes.

  8. Parameter-based estimation of CT dose index and image quality using an in-house android™-based software

    NASA Astrophysics Data System (ADS)

    Mubarok, S.; Lubis, L. E.; Pawiro, S. A.

    2016-03-01

    Compromise between radiation dose and image quality is essential in the use of CT imaging. CT dose index (CTDI) is currently the primary dosimetric formalisms in CT scan, while the low and high contrast resolutions are aspects indicating the image quality. This study was aimed to estimate CTDIvol and image quality measures through a range of exposure parameters variation. CTDI measurements were performed using PMMA (polymethyl methacrylate) phantom of 16 cm diameter, while the image quality test was conducted by using catphan ® 600. CTDI measurements were carried out according to IAEA TRS 457 protocol using axial scan mode, under varied parameters of tube voltage, collimation or slice thickness, and tube current. Image quality test was conducted accordingly under the same exposure parameters with CTDI measurements. An Android™ based software was also result of this study. The software was designed to estimate the value of CTDIvol with maximum difference compared to actual CTDIvol measurement of 8.97%. Image quality can also be estimated through CNR parameter with maximum difference to actual CNR measurement of 21.65%.

  9. ImageMiner: a software system for comparative analysis of tissue microarrays using content-based image retrieval, high-performance computing, and grid technology

    PubMed Central

    Foran, David J; Yang, Lin; Hu, Jun; Goodell, Lauri A; Reiss, Michael; Wang, Fusheng; Kurc, Tahsin; Pan, Tony; Sharma, Ashish; Saltz, Joel H

    2011-01-01

    Objective and design The design and implementation of ImageMiner, a software platform for performing comparative analysis of expression patterns in imaged microscopy specimens such as tissue microarrays (TMAs), is described. ImageMiner is a federated system of services that provides a reliable set of analytical and data management capabilities for investigative research applications in pathology. It provides a library of image processing methods, including automated registration, segmentation, feature extraction, and classification, all of which have been tailored, in these studies, to support TMA analysis. The system is designed to leverage high-performance computing machines so that investigators can rapidly analyze large ensembles of imaged TMA specimens. To support deployment in collaborative, multi-institutional projects, ImageMiner features grid-enabled, service-based components so that multiple instances of ImageMiner can be accessed remotely and federated. Results The experimental evaluation shows that: (1) ImageMiner is able to support reliable detection and feature extraction of tumor regions within imaged tissues; (2) images and analysis results managed in ImageMiner can be searched for and retrieved on the basis of image-based features, classification information, and any correlated clinical data, including any metadata that have been generated to describe the specified tissue and TMA; and (3) the system is able to reduce computation time of analyses by exploiting computing clusters, which facilitates analysis of larger sets of tissue samples. PMID:21606133

  10. Biological Visualization, Imaging and Simulation(Bio-VIS) at NASA Ames Research Center: Developing New Software and Technology for Astronaut Training and Biology Research in Space

    NASA Technical Reports Server (NTRS)

    Smith, Jeffrey

    2003-01-01

    The Bio- Visualization, Imaging and Simulation (BioVIS) Technology Center at NASA's Ames Research Center is dedicated to developing and applying advanced visualization, computation and simulation technologies to support NASA Space Life Sciences research and the objectives of the Fundamental Biology Program. Research ranges from high resolution 3D cell imaging and structure analysis, virtual environment simulation of fine sensory-motor tasks, computational neuroscience and biophysics to biomedical/clinical applications. Computer simulation research focuses on the development of advanced computational tools for astronaut training and education. Virtual Reality (VR) and Virtual Environment (VE) simulation systems have become important training tools in many fields from flight simulation to, more recently, surgical simulation. The type and quality of training provided by these computer-based tools ranges widely, but the value of real-time VE computer simulation as a method of preparing individuals for real-world tasks is well established. Astronauts routinely use VE systems for various training tasks, including Space Shuttle landings, robot arm manipulations and extravehicular activities (space walks). Currently, there are no VE systems to train astronauts for basic and applied research experiments which are an important part of many missions. The Virtual Glovebox (VGX) is a prototype VE system for real-time physically-based simulation of the Life Sciences Glovebox where astronauts will perform many complex tasks supporting research experiments aboard the International Space Station. The VGX consists of a physical display system utilizing duel LCD projectors and circular polarization to produce a desktop-sized 3D virtual workspace. Physically-based modeling tools (Arachi Inc.) provide real-time collision detection, rigid body dynamics, physical properties and force-based controls for objects. The human-computer interface consists of two magnetic tracking devices

  11. Web-based spatial analysis with the ILWIS open source GIS software and satellite images from GEONETCast

    NASA Astrophysics Data System (ADS)

    Lemmens, R.; Maathuis, B.; Mannaerts, C.; Foerster, T.; Schaeffer, B.; Wytzisk, A.

    2009-12-01

    This paper involves easy accessible integrated web-based analysis of satellite images with a plug-in based open source software. The paper is targeted to both users and developers of geospatial software. Guided by a use case scenario, we describe the ILWIS software and its toolbox to access satellite images through the GEONETCast broadcasting system. The last two decades have shown a major shift from stand-alone software systems to networked ones, often client/server applications using distributed geo-(web-)services. This allows organisations to combine without much effort their own data with remotely available data and processing functionality. Key to this integrated spatial data analysis is a low-cost access to data from within a user-friendly and flexible software. Web-based open source software solutions are more often a powerful option for developing countries. The Integrated Land and Water Information System (ILWIS) is a PC-based GIS & Remote Sensing software, comprising a complete package of image processing, spatial analysis and digital mapping and was developed as commercial software from the early nineties onwards. Recent project efforts have migrated ILWIS into a modular, plug-in-based open source software, and provide web-service support for OGC-based web mapping and processing. The core objective of the ILWIS Open source project is to provide a maintainable framework for researchers and software developers to implement training components, scientific toolboxes and (web-) services. The latest plug-ins have been developed for multi-criteria decision making, water resources analysis and spatial statistics analysis. The development of this framework is done since 2007 in the context of 52°North, which is an open initiative that advances the development of cutting edge open source geospatial software, using the GPL license. GEONETCast, as part of the emerging Global Earth Observation System of Systems (GEOSS), puts essential environmental data at the

  12. Influence of kinematic redundancy on the singularity-free workspace of parallel kinematic machines

    NASA Astrophysics Data System (ADS)

    Kotlarski, Jens; Heimann, Bodo; Ortmaier, Tobias

    2012-06-01

    In this paper the effect of kinematic redundancy in order to reduce the singularity loci of the direct kinematics and to increase the operational, i.e., singularityfree, workspace is demonstrated. The proposed approach consists of additional prismatic actuators allowing one or more base joints to move linearly. As a result, a selective reconfiguration can be performed in order to avoid singular configurations. Exemplarily, kinematically redundant schemes of four structures, the 3 RRR, the 3R PR, the 6U PS, and the 6 RUS, are considered. The relationship between the redundancy and the operational workspace is studied and several analysis examples demonstrate the effectiveness of the proposed concept. Furthermore, the additional benefit of an increasing number of redundant actuators is discussed.

  13. Quantitative comparison and evaluation of software packages for assessment of abdominal adipose tissue distribution by magnetic resonance imaging

    PubMed Central

    Bonekamp, S; Ghosh, P; Crawford, S; Solga, SF; Horska, A; Brancati, FL; Diehl, AM; Smith, S; Clark, JM

    2009-01-01

    Objective To examine five available software packages for the assessment of abdominal adipose tissue with magnetic resonance imaging, compare their features and assess the reliability of measurement results. Design Feature evaluation and test–retest reliability of softwares (NIHImage, SliceOmatic, Analyze, HippoFat and EasyVision) used in manual, semi-automated or automated segmentation of abdominal adipose tissue. Subjects A random sample of 15 obese adults with type 2 diabetes. Measurements Axial T1-weighted spin echo images centered at vertebral bodies of L2–L3 were acquired at 1.5 T. Five software packages were evaluated (NIHImage, SliceOmatic, Analyze, HippoFat and EasyVision), comparing manual, semi-automated and automated segmentation approaches. Images were segmented into cross-sectional area (CSA), and the areas of visceral (VAT) and subcutaneous adipose tissue (SAT). Ease of learning and use and the design of the graphical user interface (GUI) were rated. Intra-observer accuracy and agreement between the software packages were calculated using intra-class correlation. Intra-class correlation coefficient was used to obtain test–retest reliability. Results Three of the five evaluated programs offered a semi-automated technique to segment the images based on histogram values or a user-defined threshold. One software package allowed manual delineation only. One fully automated program demonstrated the drawbacks of uncritical automated processing. The semi-automated approaches reduced variability and measurement error, and improved reproducibility. There was no significant difference in the intra-observer agreement in SAT and CSA. The VAT measurements showed significantly lower test–retest reliability. There were some differences between the software packages in qualitative aspects, such as user friendliness. Conclusion Four out of five packages provided essentially the same results with respect to the inter- and intra-rater reproducibility. Our

  14. Online Workspace to Connect Scientists with NASA's Science E/PO Efforts and Practitioners

    NASA Astrophysics Data System (ADS)

    Shipp, Stephanie; Bartolone , Lindsay; Peticolas, Laura; Woroner, Morgan; Dalton, Heather; Schwerin, Theresa; Smith, Denise

    2014-11-01

    There is a growing awareness of the need for a scientifically literate public in light of challenges facing society today, and also a growing concern about the preparedness of our future workforce to meet those challenges. Federal priorities for science, technology, engineering, and math (STEM) education call for improvement of teacher training, increased youth and public engagement, greater involvement of underrepresented populations, and investment in undergraduate and graduate education. How can planetary scientists contribute to these priorities? How can they “make their work and findings comprehensible, appealing, and available to the public” as called for in the Planetary Decadal Survey?NASA’s Science Mission Directorate (SMD) Education and Public Outreach (E/PO) workspace provides the SMD E/PO community of practice - scientists and educators funded to conduct SMD E/PO or those using NASA’s science discoveries in E/PO endeavors - with an online environment in which to communicate, collaborate, and coordinate activities, thus helping to increase effectiveness of E/PO efforts. The workspace offers interested scientists avenues to partner with SMD E/PO practitioners and learn about E/PO projects and impacts, as well as to advertise their own efforts to reach a broader audience. Through the workspace, scientists can become aware of opportunities for involvement and explore resources to improve professional practice, including literature reviews of best practices for program impact, mechanisms for engaging diverse audiences, and large- and small-scale program evaluation. Scientists will find “how to” manuals for getting started and increasing impact with public presentations, classroom visits, and other audiences, as well as primers with activity ideas and resources that can augment E/PO interactions with different audiences. The poster will introduce the workspace to interested scientists and highlight pathways to resources of interest that can help

  15. Towards a Collaborative Online Workspace and Unified Standards for Geochemical Data

    NASA Astrophysics Data System (ADS)

    Mernagh, T. P.; Treloar, A.; Wyborn, L. A.

    2011-12-01

    stores at the institution or elsewhere. They are developing a national discovery service that enables access to data in institutional stores with rich context. No data is stored in this system, only metadata with pointers back to the original data. This enables researchers to keep their own data but also enables access to many repositories at once. Such a system will require standardisation at all phases of the process of analytical geochemistry. The geochemistry community needs to work together to develop standards for attributes as the data are collected from the instrument, to develop more standardised processing of the raw data and to agree on what is required for publishing. An online-collaborative workspace such as this would be ideal for geochemical data and the provision of standardised, open source software would greatly enhance the persistence of individual geochemistry data collections and facilitate reuse and repurposing. This conforms to the guidelines from Geoinformatics for Geochemistry (http://www.geoinfogeochem.org/) which requires metadata on how the samples were analysed.

  16. 3DVIEWNIX-AVS: a software package for the separate visualization of arteries and veins in CE-MRA images.

    PubMed

    Lei, Tianhu; Udupa, Jayaram K; Odhner, Dewey; Nyúl, László G; Saha, Punam K

    2003-01-01

    Our earlier study developed a computerized method, based on fuzzy connected object delineation principles and algorithms, for artery and vein separation in contrast enhanced Magnetic Resonance Angiography (CE-MRA) images. This paper reports its current development-a software package-for routine clinical use. The software package, termed 3DVIEWNIX-AVS, consists of the following major operational parts: (1) converting data from DICOM3 to 3DVIEWNIX format, (2) previewing slices and creating VOI and MIP Shell, (3) segmenting vessel, (4) separating artery and vein, (5) shell rendering vascular structures and creating animations. This package has been applied to EPIX Medical Inc's CE-MRA data (AngioMark MS-325). One hundred and thirty-five original CE-MRA data sets (of 52 patients) from 6 hospitals have been processed. In all case studies, unified parameter settings produce correct artery-vein separation. The current package is running on a Pentium PC under Linux and the total computation time per study is about 3 min. The strengths of this software package are (1) minimal user interaction, (2) minimal anatomic knowledge requirements on human vascular system, (3) clinically required speed, (4) free entry to any operational stages, (5) reproducible, reliable, high quality of results, and (6) cost effective computer implementation. To date, it seems to be the only software package (using an image processing approach) available for artery and vein separation of the human vascular system for routine use in a clinical setting. PMID:12821028

  17. EVALUATION OF DOSE REDUCTION POTENTIALS OF A NOVEL SCATTER CORRECTION SOFTWARE FOR BEDSIDE CHEST X-RAY IMAGING.

    PubMed

    Renger, Bernhard; Brieskorn, Carina; Toth, Vivien; Mentrup, Detlef; Jockel, Sascha; Lohöfer, Fabian; Schwarz, Martin; Rummeny, Ernst J; Noël, Peter B

    2016-06-01

    Bedside chest X-rays (CXR) for catheter position control may add up to a considerable radiation dose for patients in the intensive care unit (ICU). In this study, image quality and dose reduction potentials of a novel X-ray scatter correction software (SkyFlow, Philips Healthcare, Hamburg, Germany) were evaluated. CXRs of a 'LUNGMAN' (Kyoto Kagaku Co., LTD, Kyoto, Japan) thoracic phantom with a portacath system, a central venous line and a dialysis catheter were performed in an experimental set-up with multiple tube voltage and tube current settings without and with an antiscatter grid. Images with diagnostic exposure index (EI) 250-500 were evaluated for the difference in applied mAs with and without antiscatter grid. Three radiologists subjectively assessed the diagnostic image quality of grid and non-grid images. Compared with a non-grid image, usage of an antiscatter grid implied twice as high mAs in order to reach diagnostic EI. SkyFlow significantly improved the image quality of images acquired without grid. CXR with grid provided better image contrast than grid-less imaging with scatter correction. PMID:26977074

  18. Implementation of a real-time software-only image smoothing filter for a block-transform video codec

    NASA Astrophysics Data System (ADS)

    Miaw, Wesley F.; Rowe, Lawrence A.

    2003-05-01

    The JPEG compression standard is a popular image format. However, at high compression ratios JPEG compression, which uses block-transform coding, can produce blocking artifacts, or artificially introduced edges within the image. Several post-processing algorithms have been developed to remove these artifacts. This paper describes an implementation of a post-processing algorithm developed by Ramchandran, Chou, and Crouse (RCC) which is fast enough for real-time software-only video applications. The original implementation of the RCC algorithm involved calculating thresholds to identify artificial edges. These calculations proved too expensive for use in real-time software-only applications. We replaced these calculations with a linear scale approximating ideal threshold values based on a combination of peak signal-to-noise ratio calculations and subjective visual quality. The resulting filter implementation is available in the widely-deployed Open Mash streaming media toolkit.

  19. Fundamentally Distributed Information Processing Integrates the Motor Network into the Mental Workspace during Mental Rotation.

    PubMed

    Schlegel, Alexander; Konuthula, Dedeepya; Alexander, Prescott; Blackwood, Ethan; Tse, Peter U

    2016-08-01

    The manipulation of mental representations in the human brain appears to share similarities with the physical manipulation of real-world objects. In particular, some neuroimaging studies have found increased activity in motor regions during mental rotation, suggesting that mental and physical operations may involve overlapping neural populations. Does the motor network contribute information processing to mental rotation? If so, does it play a similar computational role in both mental and manual rotation, and how does it communicate with the wider network of areas involved in the mental workspace? Here we used multivariate methods and fMRI to study 24 participants as they mentally rotated 3-D objects or manually rotated their hands in one of four directions. We find that information processing related to mental rotations is distributed widely among many cortical and subcortical regions, that the motor network becomes tightly integrated into a wider mental workspace network during mental rotation, and that motor network activity during mental rotation only partially resembles that involved in manual rotation. Additionally, these findings provide evidence that the mental workspace is organized as a distributed core network that dynamically recruits specialized subnetworks for specific tasks as needed. PMID:27054403

  20. Mathematically gifted adolescents mobilize enhanced workspace configuration of theta cortical network during deductive reasoning.

    PubMed

    Zhang, L; Gan, J Q; Wang, H

    2015-03-19

    Previous studies have established the importance of the fronto-parietal brain network in the information processing of reasoning. At the level of cortical source analysis, this eletroencepalogram (EEG) study investigates the functional reorganization of the theta-band (4-8Hz) neurocognitive network of mathematically gifted adolescents during deductive reasoning. Depending on the dense increase of long-range phase synchronizations in the reasoning process, math-gifted adolescents show more significant adaptive reorganization and enhanced "workspace" configuration in the theta network as compared with average-ability control subjects. The salient areas are mainly located in the anterior cortical vertices of the fronto-parietal network. Further correlation analyses have shown that the enhanced workspace configuration with respect to the global topological metrics of the theta network in math-gifted subjects is correlated with the intensive frontal midline theta (fm theta) response that is related to strong neural effort for cognitive events. These results suggest that by investing more cognitive resources math-gifted adolescents temporally mobilize an enhanced task-related global neuronal workspace, which is manifested as a highly integrated fronto-parietal information processing network during the reasoning process. PMID:25595993

  1. Three-Dimensional Root Phenotyping with a Novel Imaging and Software Platform1[C][W][OA

    PubMed Central

    Clark, Randy T.; MacCurdy, Robert B.; Jung, Janelle K.; Shaff, Jon E.; McCouch, Susan R.; Aneshansley, Daniel J.; Kochian, Leon V.

    2011-01-01

    A novel imaging and software platform was developed for the high-throughput phenotyping of three-dimensional root traits during seedling development. To demonstrate the platform’s capacity, plants of two rice (Oryza sativa) genotypes, Azucena and IR64, were grown in a transparent gellan gum system and imaged daily for 10 d. Rotational image sequences consisting of 40 two-dimensional images were captured using an optically corrected digital imaging system. Three-dimensional root reconstructions were generated and analyzed using a custom-designed software, RootReader3D. Using the automated and interactive capabilities of RootReader3D, five rice root types were classified and 27 phenotypic root traits were measured to characterize these two genotypes. Where possible, measurements from the three-dimensional platform were validated and were highly correlated with conventional two-dimensional measurements. When comparing gellan gum-grown plants with those grown under hydroponic and sand culture, significant differences were detected in morphological root traits (P < 0.05). This highly flexible platform provides the capacity to measure root traits with a high degree of spatial and temporal resolution and will facilitate novel investigations into the development of entire root systems or selected components of root systems. In combination with the extensive genetic resources that are now available, this platform will be a powerful resource to further explore the molecular and genetic determinants of root system architecture. PMID:21454799

  2. Interference-free ultrasound imaging during HIFU therapy, using software tools

    NASA Technical Reports Server (NTRS)

    Vaezy, Shahram (Inventor); Held, Robert (Inventor); Sikdar, Siddhartha (Inventor); Managuli, Ravi (Inventor); Zderic, Vesna (Inventor)

    2010-01-01

    Disclosed herein is a method for obtaining a composite interference-free ultrasound image when non-imaging ultrasound waves would otherwise interfere with ultrasound imaging. A conventional ultrasound imaging system is used to collect frames of ultrasound image data in the presence of non-imaging ultrasound waves, such as high-intensity focused ultrasound (HIFU). The frames are directed to a processor that analyzes the frames to identify portions of the frame that are interference-free. Interference-free portions of a plurality of different ultrasound image frames are combined to generate a single composite interference-free ultrasound image that is displayed to a user. In this approach, a frequency of the non-imaging ultrasound waves is offset relative to a frequency of the ultrasound imaging waves, such that the interference introduced by the non-imaging ultrasound waves appears in a different portion of the frames.

  3. SU-E-J-42: Customized Deformable Image Registration Using Open-Source Software SlicerRT

    SciTech Connect

    Gaitan, J Cifuentes; Chin, L; Pignol, J; Kirby, N; Pouliot, J; Lasso, A; Pinter, C; Fichtinger, G

    2014-06-01

    Purpose: SlicerRT is a flexible platform that allows the user to incorporate the necessary images registration and processing tools to improve clinical workflow. This work validates the accuracy and the versatility of the deformable image registration algorithm of the free open-source software SlicerRT using a deformable physical pelvic phantom versus available commercial image fusion algorithms. Methods: Optical camera images of nonradiopaque markers implanted in an anatomical pelvic phantom were used to measure the ground-truth deformation and evaluate the theoretical deformations for several DIR algorithms. To perform the registration, full and empty bladder computed tomography (CT) images of the phantom were obtained and used as fixed and moving images, respectively. The DIR module, found in SlicerRT, used a B-spline deformable image registration with multiple optimization parameters that allowed customization of the registration including a regularization term that controlled the amount of local voxel displacement. The virtual deformation field at the center of the phantom was obtained and compared to the experimental ground-truth values. The parameters of SlicerRT were then varied to improve spatial accuracy. To quantify image similarity, the mean absolute difference (MAD) parameter using Hounsfield units was calculated. In addition, the Dice coefficient of the contoured rectum was evaluated to validate the strength of the algorithm to transfer anatomical contours. Results: Overall, SlicerRT achieved one of the lowest MAD values across the algorithm spectrum, but slightly smaller mean spatial errors in comparison to MIM software (MIM). On the other hand, SlicerRT created higher mean spatial errors than Velocity Medical Solutions (VEL), although obtaining an improvement on the DICE to 0.91. The large spatial errors were attributed to the poor contrast in the prostate bladder interface of the phantom. Conclusion: Based phantom validation, SlicerRT is capable of

  4. Nuquantus: Machine learning software for the characterization and quantification of cell nuclei in complex immunofluorescent tissue images

    PubMed Central

    Gross, Polina; Honnorat, Nicolas; Varol, Erdem; Wallner, Markus; Trappanese, Danielle M.; Sharp, Thomas E.; Starosta, Timothy; Duran, Jason M.; Koller, Sarah; Davatzikos, Christos; Houser, Steven R.

    2016-01-01

    Determination of fundamental mechanisms of disease often hinges on histopathology visualization and quantitative image analysis. Currently, the analysis of multi-channel fluorescence tissue images is primarily achieved by manual measurements of tissue cellular content and sub-cellular compartments. Since the current manual methodology for image analysis is a tedious and subjective approach, there is clearly a need for an automated analytical technique to process large-scale image datasets. Here, we introduce Nuquantus (Nuclei quantification utility software) - a novel machine learning-based analytical method, which identifies, quantifies and classifies nuclei based on cells of interest in composite fluorescent tissue images, in which cell borders are not visible. Nuquantus is an adaptive framework that learns the morphological attributes of intact tissue in the presence of anatomical variability and pathological processes. Nuquantus allowed us to robustly perform quantitative image analysis on remodeling cardiac tissue after myocardial infarction. Nuquantus reliably classifies cardiomyocyte versus non-cardiomyocyte nuclei and detects cell proliferation, as well as cell death in different cell classes. Broadly, Nuquantus provides innovative computerized methodology to analyze complex tissue images that significantly facilitates image analysis and minimizes human bias. PMID:27005843

  5. Nuquantus: Machine learning software for the characterization and quantification of cell nuclei in complex immunofluorescent tissue images

    NASA Astrophysics Data System (ADS)

    Gross, Polina; Honnorat, Nicolas; Varol, Erdem; Wallner, Markus; Trappanese, Danielle M.; Sharp, Thomas E.; Starosta, Timothy; Duran, Jason M.; Koller, Sarah; Davatzikos, Christos; Houser, Steven R.

    2016-03-01

    Determination of fundamental mechanisms of disease often hinges on histopathology visualization and quantitative image analysis. Currently, the analysis of multi-channel fluorescence tissue images is primarily achieved by manual measurements of tissue cellular content and sub-cellular compartments. Since the current manual methodology for image analysis is a tedious and subjective approach, there is clearly a need for an automated analytical technique to process large-scale image datasets. Here, we introduce Nuquantus (Nuclei quantification utility software) - a novel machine learning-based analytical method, which identifies, quantifies and classifies nuclei based on cells of interest in composite fluorescent tissue images, in which cell borders are not visible. Nuquantus is an adaptive framework that learns the morphological attributes of intact tissue in the presence of anatomical variability and pathological processes. Nuquantus allowed us to robustly perform quantitative image analysis on remodeling cardiac tissue after myocardial infarction. Nuquantus reliably classifies cardiomyocyte versus non-cardiomyocyte nuclei and detects cell proliferation, as well as cell death in different cell classes. Broadly, Nuquantus provides innovative computerized methodology to analyze complex tissue images that significantly facilitates image analysis and minimizes human bias.

  6. Nuquantus: Machine learning software for the characterization and quantification of cell nuclei in complex immunofluorescent tissue images.

    PubMed

    Gross, Polina; Honnorat, Nicolas; Varol, Erdem; Wallner, Markus; Trappanese, Danielle M; Sharp, Thomas E; Starosta, Timothy; Duran, Jason M; Koller, Sarah; Davatzikos, Christos; Houser, Steven R

    2016-01-01

    Determination of fundamental mechanisms of disease often hinges on histopathology visualization and quantitative image analysis. Currently, the analysis of multi-channel fluorescence tissue images is primarily achieved by manual measurements of tissue cellular content and sub-cellular compartments. Since the current manual methodology for image analysis is a tedious and subjective approach, there is clearly a need for an automated analytical technique to process large-scale image datasets. Here, we introduce Nuquantus (Nuclei quantification utility software) - a novel machine learning-based analytical method, which identifies, quantifies and classifies nuclei based on cells of interest in composite fluorescent tissue images, in which cell borders are not visible. Nuquantus is an adaptive framework that learns the morphological attributes of intact tissue in the presence of anatomical variability and pathological processes. Nuquantus allowed us to robustly perform quantitative image analysis on remodeling cardiac tissue after myocardial infarction. Nuquantus reliably classifies cardiomyocyte versus non-cardiomyocyte nuclei and detects cell proliferation, as well as cell death in different cell classes. Broadly, Nuquantus provides innovative computerized methodology to analyze complex tissue images that significantly facilitates image analysis and minimizes human bias. PMID:27005843

  7. Software workflow for the automatic tagging of medieval manuscript images (SWATI)

    NASA Astrophysics Data System (ADS)

    Chandna, Swati; Tonne, Danah; Jejkal, Thomas; Stotzka, Rainer; Krause, Celia; Vanscheidt, Philipp; Busch, Hannah; Prabhune, Ajinkya

    2015-01-01

    Digital methods, tools and algorithms are gaining in importance for the analysis of digitized manuscript collections in the arts and humanities. One example is the BMBF-funded research project "eCodicology" which aims to design, evaluate and optimize algorithms for the automatic identification of macro- and micro-structural layout features of medieval manuscripts. The main goal of this research project is to provide better insights into high-dimensional datasets of medieval manuscripts for humanities scholars. The heterogeneous nature and size of the humanities data and the need to create a database of automatically extracted reproducible features for better statistical and visual analysis are the main challenges in designing a workflow for the arts and humanities. This paper presents a concept of a workflow for the automatic tagging of medieval manuscripts. As a starting point, the workflow uses medieval manuscripts digitized within the scope of the project Virtual Scriptorium St. Matthias". Firstly, these digitized manuscripts are ingested into a data repository. Secondly, specific algorithms are adapted or designed for the identification of macro- and micro-structural layout elements like page size, writing space, number of lines etc. And lastly, a statistical analysis and scientific evaluation of the manuscripts groups are performed. The workflow is designed generically to process large amounts of data automatically with any desired algorithm for feature extraction. As a result, a database of objectified and reproducible features is created which helps to analyze and visualize hidden relationships of around 170,000 pages. The workflow shows the potential of automatic image analysis by enabling the processing of a single page in less than a minute. Furthermore, the accuracy tests of the workflow on a small set of manuscripts with respect to features like page size and text areas show that automatic and manual analysis are comparable. The usage of a computer

  8. NiftyFit: a Software Package for Multi-parametric Model-Fitting of 4D Magnetic Resonance Imaging Data.

    PubMed

    Melbourne, Andrew; Toussaint, Nicolas; Owen, David; Simpson, Ivor; Anthopoulos, Thanasis; De Vita, Enrico; Atkinson, David; Ourselin, Sebastien

    2016-07-01

    Multi-modal, multi-parametric Magnetic Resonance (MR) Imaging is becoming an increasingly sophisticated tool for neuroimaging. The relationships between parameters estimated from different individual MR modalities have the potential to transform our understanding of brain function, structure, development and disease. This article describes a new software package for such multi-contrast Magnetic Resonance Imaging that provides a unified model-fitting framework. We describe model-fitting functionality for Arterial Spin Labeled MRI, T1 Relaxometry, T2 relaxometry and Diffusion Weighted imaging, providing command line documentation to generate the figures in the manuscript. Software and data (using the nifti file format) used in this article are simultaneously provided for download. We also present some extended applications of the joint model fitting framework applied to diffusion weighted imaging and T2 relaxometry, in order to both improve parameter estimation in these models and generate new parameters that link different MR modalities. NiftyFit is intended as a clear and open-source educational release so that the user may adapt and develop their own functionality as they require. PMID:26972806

  9. Evaluation of three methods for retrospective correction of vignetting on medical microscopy images utilizing two open source software tools.

    PubMed

    Babaloukas, Georgios; Tentolouris, Nicholas; Liatis, Stavros; Sklavounou, Alexandra; Perrea, Despoina

    2011-12-01

    Correction of vignetting on images obtained by a digital camera mounted on a microscope is essential before applying image analysis. The aim of this study is to evaluate three methods for retrospective correction of vignetting on medical microscopy images and compare them with a prospective correction method. One digital image from four different tissues was used and a vignetting effect was applied on each of these images. The resulted vignetted image was replicated four times and in each replica a different method for vignetting correction was applied with fiji and gimp software tools. The highest peak signal-to-noise ratio from the comparison of each method to the original image was obtained from the prospective method in all tissues. The morphological filtering method provided the highest peak signal-to-noise ratio value amongst the retrospective methods. The prospective method is suggested as the method of choice for correction of vignetting and if it is not applicable, then the morphological filtering may be suggested as the retrospective alternative method. PMID:21950542

  10. Filtering Chromatic Aberration for Wide Acceptance Angle Electrostatic Lenses II--Experimental Evaluation and Software-Based Imaging Energy Analyzer.

    PubMed

    Fazekas, Ádám; Daimon, Hiroshi; Matsuda, Hiroyuki; Tóth, László

    2016-03-01

    Here, the experimental results of the method of filtering the effect of chromatic aberration for wide acceptance angle electrostatic lens-based system are described. This method can eliminate the effect of chromatic aberration from the images of a measured spectral image sequence by determining and removing the effect of higher and lower kinetic energy electrons on each different energy image, which leads to significant improvement of image and spectral quality. The method is based on the numerical solution of a large system of linear equations and equivalent with a multivariate strongly nonlinear deconvolution method. A matrix whose elements describe the strongly nonlinear chromatic aberration-related transmission function of the lens system acts on the vector of the ordered pixels of the distortion free spectral image sequence, and produces the vector of the ordered pixels of the measured spectral image sequence. Since the method can be applied not only on 2D real- and $k$ -space diffraction images, but also along a third dimension of the image sequence that is along the optical or in the 3D parameter space, the energy axis, it functions as a software-based imaging energy analyzer (SBIEA). It can also be applied in cases of light or other type of optics for different optical aberrations and distortions. In case of electron optics, the SBIEA method makes possible the spectral imaging without the application of any other energy filter. It is notable that this method also eliminates the disturbing background significantly in the present investigated case of reflection electron energy loss spectra. It eliminates the instrumental effects and makes possible to measure the real physical processes better. PMID:26863662

  11. Astronomy Software

    NASA Technical Reports Server (NTRS)

    1995-01-01

    Software Bisque's TheSky, SkyPro and Remote Astronomy Software incorporate technology developed for the Hubble Space Telescope. TheSky and SkyPro work together to orchestrate locating, identifying and acquiring images of deep sky objects. With all three systems, the user can directly control computer-driven telescopes and charge coupled device (CCD) cameras through serial ports. Through the systems, astronomers and students can remotely operate a telescope at the Mount Wilson Observatory Institute.

  12. SU-E-J-104: Evaluation of Accuracy for Various Deformable Image Registrations with Virtual Deformation QA Software

    SciTech Connect

    Han, S; Kim, K; Kim, M; Jung, H; Ji, Y; Choi, S; Park, S

    2015-06-15

    Purpose: The accuracy of deformable image registration (DIR) has a significant dosimetric impact in radiation treatment planning. We evaluated accuracy of various DIR algorithms using virtual deformation QA software (ImSimQA, Oncology System Limited, UK). Methods: The reference image (Iref) and volume (Vref) was first generated with IMSIMQA software. We deformed Iref with axial movement of deformation point and Vref depending on the type of deformation that are the deformation1 is to increase the Vref (relaxation) and the deformation 2 is to decrease the Vref (contraction) .The deformed image (Idef) and volume (Vdef) were inversely deformed to Iref and Vref using DIR algorithms. As a Result, we acquired deformed image (Iid) and volume (Vid). The DIR algorithms were optical flow (HS, IOF) and demons (MD, FD) of the DIRART. The image similarity evaluation between Iref and Iid was calculated by Normalized Mutual Information (NMI) and Normalized Cross Correlation (NCC). The value of Dice Similarity Coefficient (DSC) was used for evaluation of volume similarity. Results: When moving distance of deformation point was 4 mm, the value of NMI was above 1.81 and NCC was above 0.99 in all DIR algorithms. Since the degree of deformation was increased, the degree of image similarity was decreased. When the Vref increased or decreased about 12%, the difference between Vref and Vid was within ±5% regardless of the type of deformation. The value of DSC was above 0.95 in deformation1 except for the MD algorithm. In case of deformation 2, that of DSC was above 0.95 in all DIR algorithms. Conclusion: The Idef and Vdef have not been completely restored to Iref and Vref and the accuracy of DIR algorithms was different depending on the degree of deformation. Hence, the performance of DIR algorithms should be verified for the desired applications.

  13. Assessing angulation on digital images of radiographs of fractures of the distal radius: visual estimation versus computer software measurement.

    PubMed

    Robertson, G A J; Robertson, B F M; Thomas, B; McEachan, J; Davidson, D M

    2011-03-01

    We assessed the reliability of visual estimation of angles on computer images of radiographs, and compared their accuracy with the measurement of angles using computer software for ten distal radius fractures. We asked 73 clinicians to visually estimate the dorsal angulation on ten computerized radiographs of fractures of the distal radius. The reliability of these estimations was calculated. Their accuracy was compared to a 'gold standard' obtained by consensus agreement between three consultants measuring these angles using the software. Inter-observer reliability was calculated as ICC = 0.51 and intra-observer reliability as r = 0.76. The visual estimations were less accurate with a mean percentage error of 31% (range, 7-83%). As angulation increased the estimation accuracy improved. Although reliability and accuracy of such estimation was better for clinicians with greater experience, actual measurement was more reliable and accurate. PMID:21169298

  14. Mission planning for Shuttle Imaging Radar-C (SIR-C) with a real-time interactive planning software

    NASA Astrophysics Data System (ADS)

    Potts, Su K.

    1993-03-01

    The Shuttle Imaging Radar-C (SIR-C) mission will operate from the payload bay of the space shuttle for 8 days, gathering Synthetic Aperture Radar (SAR) data over specific sites on the Earth. The short duration of the mission and the requirement for realtime planning offer challenges in mission planning and in the design of the Planning and Analysis Subsystem (PAS). The PAS generates shuttle ephemerides and mission planning data and provides an interactive real-time tool for quick mission replanning. It offers a multi-user and multiprocessing environment, and it is able to keep multiple versions of the mission timeline data while maintaining data integrity and security. Its flexible design allows one software to provide different menu options based on the user's operational function, and makes it easy to tailor the software for other Earth orbiting missions.

  15. Mission planning for Shuttle Imaging Radar-C (SIR-C) with a real-time interactive planning software

    NASA Technical Reports Server (NTRS)

    Potts, Su K.

    1993-01-01

    The Shuttle Imaging Radar-C (SIR-C) mission will operate from the payload bay of the space shuttle for 8 days, gathering Synthetic Aperture Radar (SAR) data over specific sites on the Earth. The short duration of the mission and the requirement for realtime planning offer challenges in mission planning and in the design of the Planning and Analysis Subsystem (PAS). The PAS generates shuttle ephemerides and mission planning data and provides an interactive real-time tool for quick mission replanning. It offers a multi-user and multiprocessing environment, and it is able to keep multiple versions of the mission timeline data while maintaining data integrity and security. Its flexible design allows one software to provide different menu options based on the user's operational function, and makes it easy to tailor the software for other Earth orbiting missions.

  16. Preliminary studies for a CBCT imaging protocol for offline organ motion analysis: registration software validation and CTDI measurements.

    PubMed

    Falco, Maria Daniela; Fontanarosa, Davide; Miceli, Roberto; Carosi, Alessandra; Santoni, Riccardo; D'Andrea, Marco

    2011-01-01

    Cone-beam X-ray volumetric imaging in the treatment room, allows online correction of set-up errors and offline assessment of residual set-up errors and organ motion. In this study the registration algorithm of the X-ray volume imaging software (XVI, Elekta, Crawley, United Kingdom), which manages a commercial cone-beam computed tomography (CBCT)-based positioning system, has been tested using a homemade and an anthropomorphic phantom to: (1) assess its performance in detecting known translational and rotational set-up errors and (2) transfer the transformation matrix of its registrations into a commercial treatment planning system (TPS) for offline organ motion analysis. Furthermore, CBCT dose index has been measured for a particular site (prostate: 120 kV, 1028.8 mAs, approximately 640 frames) using a standard Perspex cylindrical body phantom (diameter 32 cm, length 15 cm) and a 10-cm-long pencil ionization chamber. We have found that known displacements were correctly calculated by the registration software to within 1.3 mm and 0.4°. For the anthropomorphic phantom, only translational displacements have been considered. Both studies have shown errors within the intrinsic uncertainty of our system for translational displacements (estimated as 0.87 mm) and rotational displacements (estimated as 0.22°). The resulting table translations proposed by the system to correct the displacements were also checked with portal images and found to place the isocenter of the plan on the linac isocenter within an error of 1 mm, which is the dimension of the spherical lead marker inserted at the center of the homemade phantom. The registration matrix translated into the TPS image fusion module correctly reproduced the alignment between planning CT scans and CBCT scans. Finally, measurements on the CBCT dose index indicate that CBCT acquisition delivers less dose than conventional CT scans and electronic portal imaging device portals. The registration software was found to be

  17. Preliminary Studies for a CBCT Imaging Protocol for Offline Organ Motion Analysis: Registration Software Validation and CTDI Measurements

    SciTech Connect

    Falco, Maria Daniela; Fontanarosa, Davide; Miceli, Roberto; Carosi, Alessandra; Santoni, Riccardo; D'Andrea, Marco

    2011-04-01

    Cone-beam X-ray volumetric imaging in the treatment room, allows online correction of set-up errors and offline assessment of residual set-up errors and organ motion. In this study the registration algorithm of the X-ray volume imaging software (XVI, Elekta, Crawley, United Kingdom), which manages a commercial cone-beam computed tomography (CBCT)-based positioning system, has been tested using a homemade and an anthropomorphic phantom to: (1) assess its performance in detecting known translational and rotational set-up errors and (2) transfer the transformation matrix of its registrations into a commercial treatment planning system (TPS) for offline organ motion analysis. Furthermore, CBCT dose index has been measured for a particular site (prostate: 120 kV, 1028.8 mAs, approximately 640 frames) using a standard Perspex cylindrical body phantom (diameter 32 cm, length 15 cm) and a 10-cm-long pencil ionization chamber. We have found that known displacements were correctly calculated by the registration software to within 1.3 mm and 0.4{sup o}. For the anthropomorphic phantom, only translational displacements have been considered. Both studies have shown errors within the intrinsic uncertainty of our system for translational displacements (estimated as 0.87 mm) and rotational displacements (estimated as 0.22{sup o}). The resulting table translations proposed by the system to correct the displacements were also checked with portal images and found to place the isocenter of the plan on the linac isocenter within an error of 1 mm, which is the dimension of the spherical lead marker inserted at the center of the homemade phantom. The registration matrix translated into the TPS image fusion module correctly reproduced the alignment between planning CT scans and CBCT scans. Finally, measurements on the CBCT dose index indicate that CBCT acquisition delivers less dose than conventional CT scans and electronic portal imaging device portals. The registration software was

  18. Evaluation of the Image-Pro Plus 4.5 software for automatic counting of labeled nuclei by PCNA immunohistochemistry.

    PubMed

    Francisco, Jairo Silva; Moraes, Heleno Pinto de; Dias, Eliane Pedra

    2004-01-01

    The objective of this study was to create and evaluate a routine (macro) using Image-Pro Plus 4.5 software (Media Cybernetics, Silver Spring, USA) for automatic counting of labeled nuclei by proliferating cell nuclear antigen (PCNA) immunohistochemistry. A total of 154 digital color images were obtained from eleven sections of reticular oral lichen planus stained by PCNA immunohistochemistry. Mean density (gray-level), red density, green density, blue density, area, minor axis, perimeter rate and roundness were parameters used for PCNA labeled nuclei discrimination, followed by their outlined presentation and counting in each image by the macro. Mean density and area thresholds were automatically defined based, respectively, on mean density and mean area of PCNA labeled nuclei in the assessed image. The reference method consisted in visual counting of manually outlined labeled nuclei. Statistical analysis of macro results versus reference countings showed a very significant correlation (rs = 0.964, p < 0.001) for general results and a high level (89.8 +/- 3.8%) of correctly counted labeled nuclei. We conclude that the main parameters associated with a high correlation between macro and reference results were mean density (gray-level) and area thresholds based on image profiles; and that Image-Pro Plus 4.5 using a routine with automatic definition of mean density and area thresholds can be considered a valid alternative to visual counting of PCNA labeled nuclei. PMID:15311310

  19. Development and Application of Stereo Camera-Based Upper Extremity Workspace Evaluation in Patients with Neuromuscular Diseases

    PubMed Central

    Abresch, Richard T.; Nicorici, Alina; Yan, Posu; Bajcsy, Ruzena

    2012-01-01

    Background The concept of reachable workspace is closely tied to upper limb joint range of motion and functional capability. Currently, no practical and cost-effective methods are available in clinical and research settings to provide arm-function evaluation using an individual’s three-dimensional (3D) reachable workspace. A method to intuitively display and effectively analyze reachable workspace would not only complement traditional upper limb functional assessments, but also provide an innovative approach to quantify and monitor upper limb function. Methodology/Principal Findings A simple stereo camera-based reachable workspace acquisition system combined with customized 3D workspace analysis algorithm was developed and compared against a sub-millimeter motion capture system. The stereo camera-based system was robust, with minimal loss of data points, and with the average hand trajectory error of about 40 mm, which resulted to ∼5% error of the total arm distance. As a proof-of-concept, a pilot study was undertaken with healthy individuals (n = 20) and a select group of patients with various neuromuscular diseases and varying degrees of shoulder girdle weakness (n = 9). The workspace envelope surface areas generated from the 3D hand trajectory captured by the stereo camera were compared. Normalization of acquired reachable workspace surface areas to the surface area of the unit hemi-sphere allowed comparison between subjects. The healthy group’s relative surface areas were 0.618±0.09 and 0.552±0.092 (right and left), while the surface areas for the individuals with neuromuscular diseases ranged from 0.03 and 0.09 (the most severely affected individual) to 0.62 and 0.50 (very mildly affected individual). Neuromuscular patients with severe arm weakness demonstrated movement largely limited to the ipsilateral lower quadrant of their reachable workspace. Conclusions/Significance The findings indicate that the proposed stereo camera-based reachable

  20. Novel collaboration and situational awareness environment for leaders and their support staff via self assembling software.

    SciTech Connect

    Bouchard, Ann Marie; Osbourn, Gordon Cecil; Bartholomew, John Warren

    2008-02-01

    This is the final report on the Sandia Fellow LDRD, project 117865, 08-0281. This presents an investigation of self-assembling software intended to create shared workspace environment to allow online collaboration and situational awareness for use by high level managers and their teams.

  1. Tools for Scientist Engagement in E/PO: NASA SMD Community Workspace and Online Resources

    NASA Astrophysics Data System (ADS)

    Dalton, H.; Shipp, S. S.; Grier, J.; Gross, N. A.; Buxner, S.; Bartolone, L.; Peticolas, L. M.; Woroner, M.; Schwerin, T. G.

    2014-12-01

    The Science Mission Directorate (SMD) Science Education and Public Outreach (E/PO) Forums are here to help you get involved in E/PO! The Forums have been developing several online resources to support scientists who are - or who are interested in becoming - involved in E/PO. These include NASA Wavelength, EarthSpace, and the SMD E/PO online community workspace. NASA Wavelength is the one-stop shop of all peer-reviewed NASA education resources to find materials you - or your audiences - can use. Browse by audience (pre-K through 12, higher education, and informal education) or topic, or choose to search for something specific by keyword and audience. http://nasawavelength.org. EarthSpace, an online clearinghouse of Earth and space materials for use in the higher education classroom, is driven by a powerful search engine that allows you to browse the collection of resources by science topic, audience, type of material or key terms. All materials are peer-reviewed before posting, and because all submissions receive a digital object identifier (doi), submitted materials can be listed as publications. http://www.lpi.usra.edu/earthspace. The SMD E/PO online community workspace contains many resources for scientists. These include one-page guides on how to get involved, tips on how to make the most of your time spent on E/PO, and sample activities, as well as news on funding, policy, and what's happening in the E/PO community. The workspace also provides scientists and the public pathways to find opportunities for participation in E/PO, to learn about SMD E/PO projects and their impacts, to connect with SMD E/PO practitioners, and to explore resources to improve professional E/PO practice, including literature reviews, information about the Next Generation Science Standards, and best practices in evaluation and engaging diverse audiences. http://smdepo.org.

  2. Prostate tumour volumes: evaluation of the agreement between magnetic resonance imaging and histology using novel co-registration software

    PubMed Central

    Le Nobin, Julien; Orczyk, Clément; Deng, Fang-Ming; Melamed, Jonathan; Rusinek, Henry; Taneja, Samir S.; Rosenkrantz, Andrew B.

    2016-01-01

    Objective To evaluate the agreement between prostate tumour volume determined using multiparametric magnetic resonance imaging (MRI) and that determined by histological assessment, using detailed software-assisted co-registration. Materials and Methods A total of 37 patients who underwent 3T multiparametric MRI (T2-weighted imaging [T2WI], diffusion-weighted imaging [DWI]/apparent diffusion coefficient [ADC], dynamic contrast-enhanced [DCE] imaging) were included. A radiologist traced the borders of suspicious lesions on T2WI and ADC and assigned a suspicion score of between 2 and 5, while a uropathologist traced the borders of tumours on histopathological photographs. Software was used to co-register MRI and three-dimensional digital reconstructions of radical prostatectomy specimens and to compute imaging and histopathological volumes. Agreement in volumes between MRI and histology was assessed using Bland–Altman plots and stratified by tumour characteristics. Results Among 50 tumours, the mean differences (95% limits of agreement) in MRI relative to histology were −32% (−128 to +65%) on T2WI and −47% (−143 to +49%) on ADC. For all tumour subsets, volume underestimation was more marked on ADC maps (mean difference ranging from −57 to −16%) than on T2WI (mean difference ranging from −45 to +2%). The 95% limits of agreement were wide for all comparisons, with the lower 95% limit ranging between −77 and −143% across assessments. Volume underestimation was more marked for tumours with a Gleason score ≥7 or a MRI suspicion score 4 or 5. Conclusion Volume estimates of prostate cancer using MRI tended to substantially underestimate histopathological volumes, with a wide variability in extent of underestimation across cases. These findings have implications for efforts to use MRI to guide risk assessment. PMID:24673731

  3. Blended Interaction Design: A Spatial Workspace Supporting HCI and Design Practice

    NASA Astrophysics Data System (ADS)

    Geyer, Florian

    This research investigates novel methods and techniques along with tool support that result from a conceptual blend of human-computer interaction with design practice. Using blending theory with material anchors as a theoretical framework, we frame both input spaces and explore emerging structures within technical, cognitive, and social aspects. Based on our results, we will describe a framework of the emerging structures and will design and evaluate tool support within a spatial, studio-like workspace to support collaborative creativity in interaction design.

  4. Analysis of a marine phototrophic biofilm by confocal laser scanning microscopy using the new image quantification software PHLIP

    PubMed Central

    Mueller, Lukas N; de Brouwer, Jody FC; Almeida, Jonas S; Stal, Lucas J; Xavier, João B

    2006-01-01

    Background Confocal laser scanning microscopy (CLSM) is the method of choice to study interfacial biofilms and acquires time-resolved three-dimensional data of the biofilm structure. CLSM can be used in a multi-channel modus where the different channels map individual biofilm components. This communication presents a novel image quantification tool, PHLIP, for the quantitative analysis of large amounts of multichannel CLSM data in an automated way. PHLIP can be freely downloaded from Results PHLIP is an open source public license Matlab toolbox that includes functions for CLSM imaging data handling and ten image analysis operations describing various aspects of biofilm morphology. The use of PHLIP is here demonstrated by a study of the development of a natural marine phototrophic biofilm. It is shown how the examination of the individual biofilm components using the multi-channel capability of PHLIP allowed the description of the dynamic spatial and temporal separation of diatoms, bacteria and organic and inorganic matter during the shift from a bacteria-dominated to a diatom-dominated phototrophic biofilm. Reflection images and weight measurements complementing the PHLIP analyses suggest that a large part of the biofilm mass consisted of inorganic mineral material. Conclusion The presented case study reveals new insight into the temporal development of a phototrophic biofilm where multi-channel imaging allowed to parallel monitor the dynamics of the individual biofilm components over time. This application of PHLIP presents the power of biofilm image analysis by multi-channel CLSM software and demonstrates the importance of PHLIP for the scientific community as a flexible and extendable image analysis platform for automated image processing. PMID:16412253

  5. ImaSim, a software tool for basic education of medical x-ray imaging in radiotherapy and radiology

    NASA Astrophysics Data System (ADS)

    Landry, Guillaume; deBlois, François; Verhaegen, Frank

    2013-11-01

    Introduction: X-ray imaging is an important part of medicine and plays a crucial role in radiotherapy. Education in this field is mostly limited to textbook teaching due to equipment restrictions. A novel simulation tool, ImaSim, for teaching the fundamentals of the x-ray imaging process based on ray-tracing is presented in this work. ImaSim is used interactively via a graphical user interface (GUI). Materials and methods: The software package covers the main x-ray based medical modalities: planar kilo voltage (kV), planar (portal) mega voltage (MV), fan beam computed tomography (CT) and cone beam CT (CBCT) imaging. The user can modify the photon source, object to be imaged and imaging setup with three-dimensional editors. Objects are currently obtained by combining blocks with variable shapes. The imaging of three-dimensional voxelized geometries is currently not implemented, but can be added in a later release. The program follows a ray-tracing approach, ignoring photon scatter in its current implementation. Simulations of a phantom CT scan were generated in ImaSim and were compared to measured data in terms of CT number accuracy. Spatial variations in the photon fluence and mean energy from an x-ray tube caused by the heel effect were estimated from ImaSim and Monte Carlo simulations and compared. Results: In this paper we describe ImaSim and provide two examples of its capabilities. CT numbers were found to agree within 36 Hounsfield Units (HU) for bone, which corresponds to a 2% attenuation coefficient difference. ImaSim reproduced the heel effect reasonably well when compared to Monte Carlo simulations. Discussion: An x-ray imaging simulation tool is made available for teaching and research purposes. ImaSim provides a means to facilitate the teaching of medical x-ray imaging.

  6. PCID and ASPIRE 2.0 - The Next Generation of AMOS Image Processing Software

    NASA Astrophysics Data System (ADS)

    Matson, C.; Soo Hoo, T.; Murphy, M.; Calef, B.; Beckner, C.; You, S.

    One of the missions of the Air Force Maui Optical and Supercomputing (AMOS) site is to generate high-resolution images of space objects using the Air Force telescopes located on Haleakala. Because atmospheric turbulence greatly reduces the resolution of space object images collected with ground-based telescopes, methods for overcoming atmospheric blurring are necessary. One such method is the use of adaptive optics systems to measure and compensate for atmospheric blurring in real time. A second method is to use image restoration algorithms on one or more short-exposure images of the space object under consideration. At AMOS, both methods are used routinely. In the case of adaptive optics, rarely can all atmospheric turbulence effects be removed from the imagery, so image restoration algorithms are useful even for adaptive-optics-corrected images. Historically, the bispectrum algorithm has been the primary image restoration algorithm used at AMOS. It has the advantages of being extremely fast (processing times of less than one second) and insensitive to atmospheric phase distortions. In addition, multi-frame blind deconvolution (MFBD) algorithms have also been used for image restoration. It has been observed empirically and with the use of computer simulation studies that MFBD algorithms produce higher-resolution image restorations than does the bispectrum algorithm. MFBD algorithms also do not need separate measurements of a star in order to work. However, in the past, MFBD algorithms have been factors of one hundred or more slower than the bispectrum algorithm, limiting their use to non-time-critical image restorations. Recently, with the financial support of AMOS and the High-Performance Computing Modernization Office, an MFBD algorithm called Physically-Constrained Iterative Deconvolution (PCID) has been efficiently parallelized and is able to produce image restorations in only a few seconds. In addition, with the financial support of AFOSR, it has been shown

  7. Software-based turbulence mitigation of short exposure image data with motion detection and background segmentation

    NASA Astrophysics Data System (ADS)

    Huebner, Claudia S.

    2011-11-01

    The degree of image degradation due to atmospheric turbulence is particularly severe when imaging over long horizontal paths since the turbulence is strongest close to the ground. The most pronounced effects include image blurring and image dancing and in case of strong turbulence image distortion as well. To mitigate these effects a number of methods from the field of image processing have been proposed most of which aim exclusively at the restoration of static scenes. But there is also an increasing interest in advancing turbulence mitigation to encompass moving objects as well. Therefore, in this paper a procedure is described that employs block-matching for the segmentation of static scene elements and moving objects such that image restoration can be carried out for both separately. This way motion blurring is taken into account in addition to atmospheric blurring, effectively reducing motion artefacts and improving the overall restoration result. Motion-compensated averaging with subsequent blind deconvolution is used for the actual image restoration.

  8. Software-based mitigation of image degradation due to atmospheric turbulence

    NASA Astrophysics Data System (ADS)

    Huebner, Claudia S.; Scheifling, Corinne

    2010-10-01

    Motion-Compensated Averaging (MCA) with blind deconvolution has proven successful in mitigating turbulence effects like image dancing and blurring. In this paper an image quality control according to the "Lucky Imaging" principle is combined with the MCA-procedure, weighting good frames more heavily than bad ones, skipping a given percentage of extremely degraded frames entirely. To account for local isoplanatism, when image dancing will effect local displacements between consecutive frames rather than global shifts only, a locally operating MCA variant with block matching, proposed in earlier work, is employed. In order to reduce loss of detail due to normal averaging, various combinations of temporal mode, median and mean are tested as reference image. The respective restoration results by means of a weighted blind deconvolution algorithm are presented and evaluated.

  9. Analysis of visual acuity and motion resolvability as measures for optimal visual perception of the workspace.

    PubMed

    Janabi-Sharifi, Farrokh; Vakanski, Aleksandar

    2011-03-01

    For working tasks with high visual demand, ergonomic design of the working stations requires defining criteria for comparative evaluation and analysis of the visual perceptibility in different regions of the workspace. This paper provides kinematic models of visual acuity and motion resolvability as adopted measures of visual perceptibility of the workspace. The proposed models have been examined through two sets of experiments. The first experiment is designed to compare the models outputs with those from experiments. Time measurements of the participants' response to visual events are employed for calculation of the perceptibility measures. The overall comparison results show similar patterns and moderate statistical errors of the measured and kinematically modeled values of the parameters. In the second experiment, the proposed set of visual perceptibility measures are examined for a simulated industrial task of inserting electronic chips into slots of a working table, resembling a fine assembly line of transponders manufacturing. The results from ANOVA tests for the visual acuity and the motion resolvability justify the postures adopted by the participants using visual perceptibility measures for completing the insertion tasks. PMID:20947063

  10. Automated Scoring of Chromogenic Media for Detection of Methicillin-Resistant Staphylococcus aureus by Use of WASPLab Image Analysis Software

    PubMed Central

    Faron, Matthew L.; Vismara, Chiara; Lacchini, Carla; Bielli, Alessandra; Gesu, Giovanni; Liebregts, Theo; van Bree, Anita; Jansz, Arjan; Soucy, Genevieve; Korver, John

    2015-01-01

    Recently, systems have been developed to create total laboratory automation for clinical microbiology. These systems allow for the automation of specimen processing, specimen incubation, and imaging of bacterial growth. In this study, we used the WASPLab to validate software that discriminates and segregates positive and negative chromogenic methicillin-resistant Staphylococcus aureus (MRSA) plates by recognition of pigmented colonies. A total of 57,690 swabs submitted for MRSA screening were enrolled in the study. Four sites enrolled specimens following their standard of care. Chromogenic agar used at these sites included MRSASelect (Bio-Rad Laboratories, Redmond, WA), chromID MRSA (bioMérieux, Marcy l'Etoile, France), and CHROMagar MRSA (BD Diagnostics, Sparks, MD). Specimens were plated and incubated using the WASPLab. The digital camera took images at 0 and 16 to 24 h and the WASPLab software determined the presence of positive colonies based on a hue, saturation, and value (HSV) score. If the HSV score fell within a defined threshold, the plate was called positive. The performance of the digital analysis was compared to manual reading. Overall, the digital software had a sensitivity of 100% and a specificity of 90.7% with the specificity ranging between 90.0 and 96.0 across all sites. The results were similar using the three different agars with a sensitivity of 100% and specificity ranging between 90.7 and 92.4%. These data demonstrate that automated digital analysis can be used to accurately sort positive from negative chromogenic agar cultures regardless of the pigmentation produced. PMID:26719443

  11. Automated Scoring of Chromogenic Media for Detection of Methicillin-Resistant Staphylococcus aureus by Use of WASPLab Image Analysis Software.

    PubMed

    Faron, Matthew L; Buchan, Blake W; Vismara, Chiara; Lacchini, Carla; Bielli, Alessandra; Gesu, Giovanni; Liebregts, Theo; van Bree, Anita; Jansz, Arjan; Soucy, Genevieve; Korver, John; Ledeboer, Nathan A

    2016-03-01

    Recently, systems have been developed to create total laboratory automation for clinical microbiology. These systems allow for the automation of specimen processing, specimen incubation, and imaging of bacterial growth. In this study, we used the WASPLab to validate software that discriminates and segregates positive and negative chromogenic methicillin-resistant Staphylococcus aureus (MRSA) plates by recognition of pigmented colonies. A total of 57,690 swabs submitted for MRSA screening were enrolled in the study. Four sites enrolled specimens following their standard of care. Chromogenic agar used at these sites included MRSASelect (Bio-Rad Laboratories, Redmond, WA), chromID MRSA (bioMérieux, Marcy l'Etoile, France), and CHROMagar MRSA (BD Diagnostics, Sparks, MD). Specimens were plated and incubated using the WASPLab. The digital camera took images at 0 and 16 to 24 h and the WASPLab software determined the presence of positive colonies based on a hue, saturation, and value (HSV) score. If the HSV score fell within a defined threshold, the plate was called positive. The performance of the digital analysis was compared to manual reading. Overall, the digital software had a sensitivity of 100% and a specificity of 90.7% with the specificity ranging between 90.0 and 96.0 across all sites. The results were similar using the three different agars with a sensitivity of 100% and specificity ranging between 90.7 and 92.4%. These data demonstrate that automated digital analysis can be used to accurately sort positive from negative chromogenic agar cultures regardless of the pigmentation produced. PMID:26719443

  12. Development of software for digital image processing for analysis of neuroangiogenesis

    NASA Astrophysics Data System (ADS)

    Gonzalez, M. A.; Ballarin, V. L.; Celín, A. R.; Rapacioli, M.; López-Costa, J. J.; Flores, V.

    2011-12-01

    The process of formation, growth and distribution of vessels within the developing central nervous system is difficult to analyze due to the complexity of the paths and branches within the system. The study of images of this area poses particular problems because the high levels of noise, blurring and poor contrast often prevent the objects of interest detected correctly. The design of algorithms for digital image processing suitable for this type of imagery remains a constant challenge. The aim of this work is to develop a computer tool to assist the specialist in processing these images. This paper proposes the use of morphological grayscale reconstruction and other morphological operators in order to segment the images properly. The results show that the algorithms allow a suitable segmentation of the objects of interest. Moreover, the interface developed for processing enables easy and simple analysis of them by the specialists.

  13. TGS[underscore]FIT: Image reconstruction software for quantitative, low-resolution tomographic assays

    SciTech Connect

    Estep, R J

    1993-01-01

    We developed the computer program TGS[underscore]FIT to aid in researching the tomographic gamma scanner method of nondestructive assay. This software, written in C-programming, language, implements a full Beer's Law attenuation correction in reconstructing low-resolution emission tomograms. The attenuation coefficients for the corrections are obtained by reconstructing a transmission tomogram of the same resolution. The command-driven interface, combined with (crude) simulation capabilities and command file control, allows design studies to be performed in a semi-automated manner.

  14. Analysis of image sharpness reproducibility on a novel engineered micro-CT scanner with variable geometry and embedded recalibration software.

    PubMed

    Panetta, D; Belcari, N; Del Guerra, A; Bartolomei, A; Salvadori, P A

    2012-04-01

    This study investigates the reproducibility of the reconstructed image sharpness, after modifications of the geometry setup, for a variable magnification micro-CT (μCT) scanner. All the measurements were performed on a novel engineered μCT scanner for in vivo imaging of small animals (Xalt), which has been recently built at the Institute of Clinical Physiology of the National Research Council (IFC-CNR, Pisa, Italy), in partnership with the University of Pisa. The Xalt scanner is equipped with an integrated software for on-line geometric recalibration, which will be used throughout the experiments. In order to evaluate the losses of image quality due to modifications of the geometry setup, we have made 22 consecutive acquisitions by changing alternatively the system geometry between two different setups (Large FoV - LF, and High Resolution - HR). For each acquisition, the tomographic images have been reconstructed before and after the on-line geometric recalibration. For each reconstruction, the image sharpness was evaluated using two different figures of merit: (i) the percentage contrast on a small bar pattern of fixed frequency (f = 5.5 lp/mm for the LF setup and f = 10 lp/mm for the HR setup) and (ii) the image entropy. We have found that, due to the small-scale mechanical uncertainty (in the order of the voxel size), a recalibration is necessary for each geometric setup after repositioning of the system's components; the resolution losses due to the lack of recalibration are worse for the HR setup (voxel size = 18.4 μm). The integrated on-line recalibration algorithm of the Xalt scanner allowed to perform the recalibration quickly, by restoring the spatial resolution of the system to the reference resolution obtained after the initial (off-line) calibration. PMID:21501966

  15. Familiarity effects in the construction of facial-composite images using modern software systems.

    PubMed

    Frowd, Charlie D; Skelton, Faye C; Butt, Neelam; Hassan, Amal; Fields, Stephen; Hancock, Peter J B

    2011-12-01

    We investigate the effect of target familiarity on the construction of facial composites, as used by law enforcement to locate criminal suspects. Two popular software construction methods were investigated. Participants were shown a target face that was either familiar or unfamiliar to them and constructed a composite of it from memory using a typical 'feature' system, involving selection of individual facial features, or one of the newer 'holistic' types, involving repeated selection and breeding from arrays of whole faces. This study found that composites constructed of a familiar face were named more successfully than composites of an unfamiliar face; also, naming of composites of internal and external features was equivalent for construction of unfamiliar targets, but internal features were better named than the external features for familiar targets. These findings applied to both systems, although benefit emerged for the holistic type due to more accurate construction of internal features and evidence for a whole-face advantage. STATEMENT OF RELEVANCE: This work is of relevance to practitioners who construct facial composites with witnesses to and victims of crime, as well as for software designers to help them improve the effectiveness of their composite systems. PMID:22103723

  16. Seismic reflection imaging of underground cavities using open-source software

    SciTech Connect

    Mellors, R J

    2011-12-20

    The Comprehensive Nuclear Test Ban Treaty (CTBT) includes provisions for an on-site inspection (OSI), which allows the use of specific techniques to detect underground anomalies including cavities and rubble zones. One permitted technique is active seismic surveys such as seismic refraction or reflection. The purpose of this report is to conduct some simple modeling to evaluate the potential use of seismic reflection in detecting cavities and to test the use of open-source software in modeling possible scenarios. It should be noted that OSI inspections are conducted under specific constraints regarding duration and logistics. These constraints are likely to significantly impact active seismic surveying, as a seismic survey typically requires considerable equipment, effort, and expertise. For the purposes of this study, which is a first-order feasibility study, these issues will not be considered. This report provides a brief description of the seismic reflection method along with some commonly used software packages. This is followed by an outline of a simple processing stream based on a synthetic model, along with results from a set of models representing underground cavities. A set of scripts used to generate the models are presented in an appendix. We do not consider detection of underground facilities in this work and the geologic setting used in these tests is an extremely simple one.

  17. Integration of instrumentation and processing software of a laser speckle contrast imaging system

    NASA Astrophysics Data System (ADS)

    Carrick, Jacob J.

    Laser speckle contrast imaging (LSCI) has the potential to be a powerful tool in medicine, but more research in the field is required so it can be used properly. To help in the progression of Michigan Tech's research in the field, a graphical user interface (GUI) was designed in Matlab to control the instrumentation of the experiments as well as process the raw speckle images into contrast images while they are being acquired. The design of the system was successful and is currently being used by Michigan Tech's Biomedical Engineering department. This thesis describes the development of the LSCI GUI as well as offering a full introduction into the history, theory and applications of LSCI.

  18. Improved modified pressure imaging and software for egg micro-crack detection and egg quality grading

    Technology Transfer Automated Retrieval System (TEKTRAN)

    Cracks in the egg shell increase a food safety risk. Especially, eggs with very fine, hairline cracks (micro-cracks) are often undetected during the grading process because they are almost impossible to detect visually. A modified pressure imaging system was developed to detect eggs with micro-crack...

  19. 3D Imaging for hand gesture recognition: Exploring the software-hardware interaction of current technologies

    NASA Astrophysics Data System (ADS)

    Periverzov, Frol; Ilieş, Horea T.

    2012-09-01

    Interaction with 3D information is one of the fundamental and most familiar tasks in virtually all areas of engineering and science. Several recent technological advances pave the way for developing hand gesture recognition capabilities available to all, which will lead to more intuitive and efficient 3D user interfaces (3DUI). These developments can unlock new levels of expression and productivity in all activities concerned with the creation and manipulation of virtual 3D shapes and, specifically, in engineering design. Building fully automated systems for tracking and interpreting hand gestures requires robust and efficient 3D imaging techniques as well as potent shape classifiers. We survey and explore current and emerging 3D imaging technologies, and focus, in particular, on those that can be used to build interfaces between the users' hands and the machine. The purpose of this paper is to categorize and highlight the relevant differences between these existing 3D imaging approaches in terms of the nature of the information provided, output data format, as well as the specific conditions under which these approaches yield reliable data. Furthermore we explore the impact of each of these approaches on the computational cost and reliability of the required image processing algorithms. Finally we highlight the main challenges and opportunities in developing natural user interfaces based on hand gestures, and conclude with some promising directions for future research. [Figure not available: see fulltext.

  20. Parallel software requirements to the design of a general architecture: application to the image processing

    NASA Astrophysics Data System (ADS)

    Bonnin, Patrick J.; Hoeltzener-Douarin, Brigitte; Aubin, N.; Cartier, S.; Porcher, Thierry; Fiorini, P.; Zavidovique, Bertrand

    1993-10-01

    A great number of parallel computer architectures have been proposed, whether they are SIMD machines (Single Instruction Multiple Data) with lots of quite simple processors, or MIMD machines (Multiple Instruction Multiple Data) containing few, but powerful processors. Each one claims to offer some kind of an optimality at the hardware level. But implementing parallel image processing algorithms to make them run in real time will remain a real challenge; it addresses rather the control of communication networks between processors (message passing, circuit switching..) or the computing model (e.g. data parallel model). In that respect, our goal here is to point out some algorithmic needs to distribute image processing operators. They will be translated first in terms of programming models, more general then image processing applications, and then as hardware properties of the processor network. In that way, we do not design yet another parallel machine dedicated to image processing, but a more general parallel architecture which one will be able to efficiently implement different kinds of programming models.

  1. Full-sun synchronic EUV and coronal hole mapping using multi-instrument images: Data and software made available

    NASA Astrophysics Data System (ADS)

    Caplan, R. M.; Downs, C.; Linker, J.

    2015-12-01

    A method for the automatic generation of EUV and coronal hole (CH) maps using simultaneous multi-instrument imaging data is described. Synchronized EUV images from STEREO/EUVI A&B 195Å and SDO/AIA 193Å undergo preprocessing steps that include PSF-deconvolution and the application of nonlinear data-derived intensity corrections that account for center-to-limb variations (limb-brightening) and inter-instrument intensity normalization. The latter two corrections are derived using a robust, systematic approach that takes advantage of unbiased long-term averages of data and serve to flatten the images by converting all pixel intensities to a unified disk center equivalent. While the number of applications are broad, we demonstrate how this technique is very useful for CH detection as it enables the use of a fast and simplified image segmentation algorithm to obtain consistent detection results. The multi-instrument nature of the technique also allows one to track evolving features consistently for longer periods than is possible with a single instrument, and preliminary results quantifying CH area and shape evolution are shown.Most importantly, several data and software products are made available to the community for use. For the ~4 year period of 6/10/2010 to 8/18/2014, we provide synchronic EUV and coronal hole maps at 6-hour cadence as well as the data-derived limb brightening and inter-instrument correction factors that we applied. We also make available a ready-to-use MATLAB script EUV2CHM used to generate the maps, which loads EUV images, applies our preprocessing steps, and then uses our GPU-accelerated/CPU-multithreaded segmentation algorithm EZSEG to detect coronal holes.

  2. Gamma-H2AX foci counting: image processing and control software for high-content screening

    NASA Astrophysics Data System (ADS)

    Barber, P. R.; Locke, R. J.; Pierce, G. P.; Rothkamm, K.; Vojnovic, B.

    2007-02-01

    Phosphorylation of the chromatin protein H2AX (forming γH2AX) is implicated in the repair of DNA double strand breaks (DSB's); a large number of H2AX molecules become phosphorylated at the sites of DSB's. Fluorescent staining of the cell nuclei for γH2AX, via an antibody, visualises the formation of these foci, allowing the quantification of DNA DSB's and forming the basis for a sensitive biological dosimeter of ionising radiation. We describe an automated fluorescence microscopy system, including automated image processing, to count γH2AX foci. The image processing is performed by a Hough transform based algorithm, CHARM, which has wide applicability for the detection and analysis of cells and cell colonies. This algorithm and its applications for cell nucleus and foci detection will be described. The system also relies heavily on robust control software, written using multi-threaded cbased modules in LabWindows/CVI that adapt to the timing requirements of a particular experiment for optimised slide/plate scanning and mosaicing, making use of modern multi-core processors. The system forms the basis of a general purpose high-content screening platform with wide ranging applications in live and fixed cell imaging and tissue micro arrays, that in future, can incorporate spectrally and time-resolved information.

  3. Grid-less imaging with antiscatter correction software in 2D mammography: the effects on image quality and MGD under a partial virtual clinical validation study

    NASA Astrophysics Data System (ADS)

    Van Peteghem, Nelis; Bemelmans, Frédéric; Bramaje Adversalo, Xenia; Salvagnini, Elena; Marshall, Nicholas; Bosmans, Hilde; Van Ongeval, Chantal

    2016-03-01

    This work investigated the effect of the grid-less acquisition mode with scatter correction software developed by Siemens Healthcare (PRIME mode) on image quality and mean glandular dose (MGD) in a comparative study against a standard mammography system with grid. Image quality was technically quantified with contrast-detail (c-d) analysis and by calculating detectability indices (d') using a non-prewhitening with eye filter model observer (NPWE). MGD was estimated technically using slabs of PMMA and clinically on a set of 11439 patient images. The c-d analysis gave similar results for all mammographic systems examined, although the d' values were slightly lower for the system with PRIME mode when compared to the same system in standard mode (-2.8% to -5.7%, depending on the PMMA thickness). The MGD values corresponding to the PMMA measurements with automatic exposure control indicated a dose reduction from 11.0% to 20.8% for the system with PRIME mode compared to the same system without PRIME mode. The largest dose reductions corresponded to the thinnest PMMA thicknesses. The results from the clinical dosimetry study showed an overall population-averaged dose reduction of 11.6% (up to 27.7% for thinner breasts) for PRIME mode compared to standard mode for breast thicknesses from 20 to 69 mm. These technical image quality measures were then supported using a clinically oriented study whereby simulated clusters of microcalcifications and masses were inserted into patient images and read by radiologists in an AFROC study to quantify their detectability. In line with the technical investigation, no significant difference was found between the two imaging modes (p-value 0.95).

  4. A Multimodality Imaging and Software System for Combining an Anatomical and Physiological Assessment of Skin and Underlying Tissue Conditions

    PubMed Central

    Langemo, Diane; Spahn, James G.

    2016-01-01

    ABSTRACT OBJECTIVE: The timely and accurate assessment of skin and underlying tissue is crucial for making informed decisions relating to wound development and existing wounds. The study objective was to determine within- and between-reader agreement of Scout Visual-to-Thermal Overlay (WoundVision LLC, Indianapolis, Indiana) placement (moving the wound edge trace from the visual image onto the wound edge signature of the infrared image). MATERIALS AND METHODS: For establishing within- and between-reader agreement of the Scout Visual-to-Thermal Overlay feature, 5 different readers overlaid a wound edge trace from the visual image and placed it onto the congruent thermal representation of the wound on a thermal image 3 independent times. Forty different wound image pairs were evaluated by each reader. All readers were trained by the same trainer on the operation of the Scout prior to using the software features. The Scout Visual-to-Thermal Overlay feature allows clinicians to use an anatomical measurement of the wound on the visual image (area and perimeter) to extract a congruent physiological measurement of the wound on the thermal image (thermal intensity variation data) by taking the wound edge trace from the visual image and overlaying it onto the corresponding thermal signature of the same wound edge. RESULTS: The results are very similar both within- and between-readers. The coefficient of variation (CV) for the mean PV both within- and between-readers averages less than 1%, 0.89 and 0.77 respectively. When converted into degrees Celsius across all 5 readers and all 3 wound replicates, the average temperature differential is 0.28° C (Table 2). The largest difference observed was 0.63° C and the smallest difference observed was 0.04° C. CONCLUSIONS: The Scout software’s Visual-to-Thermal Overlay procedure, as implemented in this study, is very precise. This study demonstrates that the thermal signature of wounds may be delineated repeatedly by the same

  5. Experiences with the BSCW Shared Workspace System as the Backbone of a Virtual Learning Environment for Students.

    ERIC Educational Resources Information Center

    Appelt, Wolfgang; Mambrey, Peter

    The GMD (German National Research Center for Information Technology) has developed the BSCW (Basic Support for Cooperative Work) Shared Workspace system within the last four years with the goal of transforming the Web from a primarily passive information repository to an active cooperation medium. The BSCW system is a Web-based groupware tool for…

  6. The Role of a Facilitated Online Workspace Component of a Community of Practice: Knowledge Building and Value Creation for NASA

    ERIC Educational Resources Information Center

    Davey, Bradford Thomas

    2013-01-01

    The purpose of this study was to examine the role of an online workspace component of a community in the work of a community of practice. Much has been studied revealing the importance of communities of practice to organizations, project success, and knowledge management and some of these same successes hold true for virtual communities of…

  7. Using a Shared Workspace and Wireless Laptops to Improve Collaborative Project Learning in an Engineering Design Class

    ERIC Educational Resources Information Center

    Nicol, David J.; MacLeod, Iain A.

    2005-01-01

    Two different technologies, groupware (a shared workspace) and shared wireless laptop computers, were implemented in a project design class in a civil engineering course. The research interest was in the way these technologies supported resource sharing within and across project groups and in the forms of group collaboration that resulted. The…

  8. Development of fast patient position verification software using 2D-3D image registration and its clinical experience.

    PubMed

    Mori, Shinichiro; Kumagai, Motoki; Miki, Kentaro; Fukuhara, Riki; Haneishi, Hideaki

    2015-09-01

    To improve treatment workflow, we developed a graphic processing unit (GPU)-based patient positional verification software application and integrated it into carbon-ion scanning beam treatment. Here, we evaluated the basic performance of the software. The algorithm provides 2D/3D registration matching using CT and orthogonal X-ray flat panel detector (FPD) images. The participants were 53 patients with tumors of the head and neck, prostate or lung receiving carbon-ion beam treatment. 2D/3D-ITchi-Gime (ITG) calculation accuracy was evaluated in terms of computation time and registration accuracy. Registration calculation was determined using the similarity measurement metrics gradient difference (GD), normalized mutual information (NMI), zero-mean normalized cross-correlation (ZNCC), and their combination. Registration accuracy was dependent on the particular metric used. Representative examples were determined to have target registration error (TRE) = 0.45 ± 0.23 mm and angular error (AE) = 0.35 ± 0.18° with ZNCC + GD for a head and neck tumor; TRE = 0.12 ± 0.07 mm and AE = 0.16 ± 0.07° with ZNCC for a pelvic tumor; and TRE = 1.19 ± 0.78 mm and AE = 0.83 ± 0.61° with ZNCC for lung tumor. Calculation time was less than 7.26 s.The new registration software has been successfully installed and implemented in our treatment process. We expect that it will improve both treatment workflow and treatment accuracy. PMID:26081313

  9. Development of fast patient position verification software using 2D-3D image registration and its clinical experience

    PubMed Central

    Mori, Shinichiro; Kumagai, Motoki; Miki, Kentaro; Fukuhara, Riki; Haneishi, Hideaki

    2015-01-01

    To improve treatment workflow, we developed a graphic processing unit (GPU)-based patient positional verification software application and integrated it into carbon-ion scanning beam treatment. Here, we evaluated the basic performance of the software. The algorithm provides 2D/3D registration matching using CT and orthogonal X-ray flat panel detector (FPD) images. The participants were 53 patients with tumors of the head and neck, prostate or lung receiving carbon-ion beam treatment. 2D/3D-ITchi-Gime (ITG) calculation accuracy was evaluated in terms of computation time and registration accuracy. Registration calculation was determined using the similarity measurement metrics gradient difference (GD), normalized mutual information (NMI), zero-mean normalized cross-correlation (ZNCC), and their combination. Registration accuracy was dependent on the particular metric used. Representative examples were determined to have target registration error (TRE) = 0.45 ± 0.23 mm and angular error (AE) = 0.35 ± 0.18° with ZNCC + GD for a head and neck tumor; TRE = 0.12 ± 0.07 mm and AE = 0.16 ± 0.07° with ZNCC for a pelvic tumor; and TRE = 1.19 ± 0.78 mm and AE = 0.83 ± 0.61° with ZNCC for lung tumor. Calculation time was less than 7.26 s.The new registration software has been successfully installed and implemented in our treatment process. We expect that it will improve both treatment workflow and treatment accuracy. PMID:26081313

  10. Photon counting imaging and centroiding with an electron-bombarded CCD using single molecule localisation software

    NASA Astrophysics Data System (ADS)

    Hirvonen, Liisa M.; Barber, Matthew J.; Suhling, Klaus

    2016-06-01

    Photon event centroiding in photon counting imaging and single-molecule localisation in super-resolution fluorescence microscopy share many traits. Although photon event centroiding has traditionally been performed with simple single-iteration algorithms, we recently reported that iterative fitting algorithms originally developed for single-molecule localisation fluorescence microscopy work very well when applied to centroiding photon events imaged with an MCP-intensified CMOS camera. Here, we have applied these algorithms for centroiding of photon events from an electron-bombarded CCD (EBCCD). We find that centroiding algorithms based on iterative fitting of the photon events yield excellent results and allow fitting of overlapping photon events, a feature not reported before and an important aspect to facilitate an increased count rate and shorter acquisition times.

  11. Photon counting imaging and centroiding with an electron-bombarded CCD using single molecule localisation software

    PubMed Central

    Hirvonen, Liisa M.; Barber, Matthew J.; Suhling, Klaus

    2016-01-01

    Photon event centroiding in photon counting imaging and single-molecule localisation in super-resolution fluorescence microscopy share many traits. Although photon event centroiding has traditionally been performed with simple single-iteration algorithms, we recently reported that iterative fitting algorithms originally developed for single-molecule localisation fluorescence microscopy work very well when applied to centroiding photon events imaged with an MCP-intensified CMOS camera. Here, we have applied these algorithms for centroiding of photon events from an electron-bombarded CCD (EBCCD). We find that centroiding algorithms based on iterative fitting of the photon events yield excellent results and allow fitting of overlapping photon events, a feature not reported before and an important aspect to facilitate an increased count rate and shorter acquisition times. PMID:27274604

  12. Improved structure, function and compatibility for CellProfiler: modular high-throughput image analysis software

    PubMed Central

    Kamentsky, Lee; Jones, Thouis R.; Fraser, Adam; Bray, Mark-Anthony; Logan, David J.; Madden, Katherine L.; Ljosa, Vebjorn; Rueden, Curtis; Eliceiri, Kevin W.; Carpenter, Anne E.

    2011-01-01

    Summary: There is a strong and growing need in the biology research community for accurate, automated image analysis. Here, we describe CellProfiler 2.0, which has been engineered to meet the needs of its growing user base. It is more robust and user friendly, with new algorithms and features to facilitate high-throughput work. ImageJ plugins can now be run within a CellProfiler pipeline. Availability and Implementation: CellProfiler 2.0 is free and open source, available at http://www.cellprofiler.org under the GPL v. 2 license. It is available as a packaged application for Macintosh OS X and Microsoft Windows and can be compiled for Linux. Contact: anne@broadinstitute.org Supplementary information: Supplementary data are available at Bioinformatics online. PMID:21349861

  13. Reliability and reproducibility of macular segmentation using a custom-built optical coherence tomography retinal image analysis software

    NASA Astrophysics Data System (ADS)

    Cabrera Debuc, Delia; Somfai, Gábor Márk; Ranganathan, Sudarshan; Tátrai, Erika; Ferencz, Mária; Puliafito, Carmen A.

    2009-11-01

    We determine the reliability and reproducibility of retinal thickness measurements with a custom-built OCT retinal image analysis software (OCTRIMA). Ten eyes of five healthy subjects undergo repeated standard macular thickness map scan sessions by two experienced examiners using a Stratus OCT device. Automatic/semi automatic thickness quantification of the macula and intraretinal layers is performed using OCTRIMA software. Intraobserver, interobserver, and intervisit repeatability and reproducibility coefficients, and intraclass correlation coefficients (ICCs) per scan are calculated. Intraobserver, interobserver, and intervisit variability combined account for less than 5% of total variability for the total retinal thickness measurements and less than 7% for the intraretinal layers except the outer segment/ retinal pigment epithelium (RPE) junction. There is no significant difference between scans acquired by different observers or during different visits. The ICCs obtained for the intraobserver and intervisit variability tests are greater than 0.75 for the total retina and all intraretinal layers, except the inner nuclear layer intraobserver and interobserver test and the outer plexiform layer, intraobserver, interobserver, and intervisit test. Our results indicate that thickness measurements for the total retina and all intraretinal layers (except the outer segment/RPE junction) performed using OCTRIMA are highly repeatable and reproducible.

  14. ORBS, ORCS, OACS, a Software Suite for Data Reduction and Analysis of the Hyperspectral Imagers SITELLE and SpIOMM

    NASA Astrophysics Data System (ADS)

    Martin, T.; Drissen, L.; Joncas, G.

    2015-09-01

    SITELLE (installed in 2015 at the Canada-France-Hawaii Telescope) and SpIOMM (a prototype attached to the Observatoire du Mont-Mégantic) are the first Imaging Fourier Transform Spectrometers (IFTS) capable of obtaining a hyperspectral data cube which samples a 12 arc minutes field of view into four millions of visible spectra. The result of each observation is made up of two interferometric data cubes which need to be merged, corrected, transformed and calibrated in order to get a spectral cube of the observed region ready to be analysed. ORBS is a fully automatic data reduction software that has been entirely designed for this purpose. The data size (up to 68 Gb for larger science cases) and the computational needs have been challenging and the highly parallelized object-oriented architecture of ORBS reflects the solutions adopted which made possible to process 68 Gb of raw data in less than 11 hours using 8 cores and 22.6 Gb of RAM. It is based on a core framework (ORB) that has been designed to support the whole software suite for data analysis (ORCS and OACS), data simulation (ORUS) and data acquisition (IRIS). They all aim to provide a strong basis for the creation and development of specialized analysis modules that could benefit the scientific community working with SITELLE and SpIOMM.

  15. Software for Sunspots Automatic Detection, Heliographic Location and Area Measurement for Soho Images

    NASA Astrophysics Data System (ADS)

    Rivero Gavilán, H.; Guevara Day, W.

    2006-06-01

    Active regions (ARs) are the manifestation of the magnetic flux tubes because of the buoyancy action these emerge in the typical letter Greek Ω shape. The tracking and the respective study ARs permit us to study the global properties of the flow tubes (which form the active regions) and provide important information about the origin (formation and transports in the convective zone) and how the magnetic helicity is taken along the corona by pho-tos-phe-ric movements. In order to initiate an study of these behaviors we are developing a programming algorithm using IDL as base, moreover taking routines developed in SOLARSOFT, will allow us to pursue of some interesting active region. The program has obtained the year 2005 magnetograms data base provided by MDI-SOHO, in which we selected the ARs of interest to determine the location of the region in function of its heliographic coordinates. At the time of selecting this image, the level of intensity of the interest field is selected and the program calculates the position of different polarities and his geometric area (given in arcsec), these values are stored in a text file as well as a support image which shows the contour lines of magnetic field intensities chosen by the user. As a test of the algorithm we have taken several images MDI-SOHO of the 10715 NOAA region from 01 to 03 of January of the present year; we have used up to 43 images. These results by are part of an study of active zones evolution for the purpose of determining the origin of the RA's formation.

  16. The Java Image Science Toolkit (JIST) for Rapid Prototyping and Publishing of Neuroimaging Software

    PubMed Central

    Lucas, Blake C.; Bogovic, John A.; Carass, Aaron; Bazin, Pierre-Louis; Prince, Jerry L.; Pham, Dzung

    2010-01-01

    Non-invasive neuroimaging techniques enable extraordinarily sensitive and specific in vivo study of the structure, functional response and connectivity of biological mechanisms. With these advanced methods comes a heavy reliance on computer-based processing, analysis and interpretation. While the neuroimaging community has produced many excellent academic and commercial tool packages, new tools are often required to interpret new modalities and paradigms. Developing custom tools and ensuring interoperability with existing tools is a significant hurdle. To address these limitations, we present a new framework for algorithm development that implicitly ensures tool interoperability, generates graphical user interfaces, provides advanced batch processing tools, and, most importantly, requires minimal additional programming or computational overhead. Java-based rapid prototyping with this system is an efficient and practical approach to evaluate new algorithms since the proposed system ensures that rapidly constructed prototypes are actually fully-functional processing modules with support for multiple GUI's, a broad range of file formats, and distributed computation. Herein, we demonstrate MRI image processing with the proposed system for cortical surface extraction in large cross-sectional cohorts, provide a system for fully automated diffusion tensor image analysis, and illustrate how the system can be used as a simulation framework for the development of a new image analysis method. The system is released as open source under the Lesser GNU Public License (LGPL) through the Neuroimaging Informatics Tools and Resources Clearinghouse (NITRC). PMID:20077162

  17. Design of a hardware/software FPGA-based driver system for a large area high resolution CCD image sensor

    NASA Astrophysics Data System (ADS)

    Chen, Ying; Xu, Wanpeng; Zhao, Rongsheng; Chen, Xiangning

    2014-09-01

    A hardware/software field programmable gate array (FPGA)-based driver system was proposed and demonstrated for the KAF-39000 large area high resolution charge coupled device (CCD). The requirements of the KAF-39000 driver system were analyzed. The structure of "microprocessor with application specific integrated circuit (ASIC) chips" was implemented to design the driver system. The system test results showed that dual channels of imaging analog data were obtained with a frame rate of 0.87 frame/s. The frequencies of horizontal timing and vertical timing were 22.9 MHz and 28.7 kHz, respectively, which almost reached the theoretical value of 24 MHz and 30 kHz, respectively.

  18. The secure medical research workspace: an IT infrastructure to enable secure research on clinical data.

    PubMed

    Shoffner, Michael; Owen, Phillips; Mostafa, Javed; Lamm, Brent; Wang, Xiaoshu; Schmitt, Charles P; Ahalt, Stanley C

    2013-06-01

    Clinical data have tremendous value for translational research, but only if security and privacy concerns can be addressed satisfactorily. A collaboration of clinical and informatics teams, including RENCI, NC TraCS, UNC's School of Information and Library Science, Information Technology Service's Research Computing and other partners at the University of North Carolina at Chapel Hill have developed a system called the Secure Medical Research Workspace (SMRW) that enables researchers to use clinical data securely for research. SMRW significantly minimizes the risk presented when using identified clinical data, thereby protecting patients, researchers, and institutions associated with the data. The SMRW is built on a novel combination of virtualization and data leakage protection and can be combined with other protection methodologies and scaled to production levels. PMID:23751029

  19. Collision-free motion of two robot arms in a common workspace

    NASA Technical Reports Server (NTRS)

    Basta, Robert A.; Mehrotra, Rajiv; Varanasi, Murali R.

    1987-01-01

    Collision-free motion of two robot arms in a common workspace is investigated. A collision-free motion is obtained by detecting collisions along the preplanned trajectories using a sphere model for the wrist of each robot and then modifying the paths and/or trajectories of one or both robots to avoid the collision. Detecting and avoiding collisions are based on the premise that: preplanned trajectories of the robots follow a straight line; collisions are restricted to between the wrists of the two robots (which corresponds to the upper three links of PUMA manipulators); and collisions never occur between the beginning points or end points on the straight line paths. The collision detection algorithm is described and some approaches to collision avoidance are discussed.

  20. The Secure Medical Research Workspace: An IT Infrastructure to Enable Secure Research on Clinical Data

    PubMed Central

    Shoffner, Michael; Owen, Phillips; Mostafa, Javed; Lamm, Brent; Wang, Xiaoshu; Schmitt, Charles P.; Ahalt, Stanley C.

    2013-01-01

    Clinical data has tremendous value for translational research, but only if security and privacy concerns can be addressed satisfactorily. A collaboration of clinical and informatics teams, including RENCI, NC TraCS, UNC’s School of Information and Library Science, Information Technology Service’s Research Computing and other partners at the University of North Carolina at Chapel Hill have developed a system called the Secure Medical Research Workspace (SMRW) that enables researchers to use clinical data securely for research. SMRW significantly minimizes the risk presented when using of identified clinical data, thereby protecting patients, researchers, and institutions associated with the data. The SMRW is built on a novel combination of virtualization and data leakage protection and can be combined with other protection methodologies and scaled to production levels. PMID:23751029

  1. Extracting and Utilizing Social Networks from Log Files of Shared Workspaces

    NASA Astrophysics Data System (ADS)

    Nasirifard, Peyman; Peristeras, Vassilios; Hayes, Conor; Decker, Stefan

    Log files of online shared workspaces contain rich information that can be further analyzed. In this paper, log-file information is used to extract object-centric and user-centric social networks. The object-centric social networks are used as a means for assigning concept-based expertise elements to users based on the documents that they created, revised or read. The user-centric social networks are derived from users working on common documents. Weights, called the Cooperation Index, are assigned to links between users in a user-centric social network, which indicates how closely two people have collaborated together, based on their history. We also present a set of tools that was developed to realize our approach.

  2. Control Software

    NASA Technical Reports Server (NTRS)

    1997-01-01

    Real-Time Innovations, Inc. (RTI) collaborated with Ames Research Center, the Jet Propulsion Laboratory and Stanford University to leverage NASA research to produce ControlShell software. RTI is the first "graduate" of Ames Research Center's Technology Commercialization Center. The ControlShell system was used extensively on a cooperative project to enhance the capabilities of a Russian-built Marsokhod rover being evaluated for eventual flight to Mars. RTI's ControlShell is complex, real-time command and control software, capable of processing information and controlling mechanical devices. One ControlShell tool is StethoScope. As a real-time data collection and display tool, StethoScope allows a user to see how a program is running without changing its execution. RTI has successfully applied its software savvy in other arenas, such as telecommunications, networking, video editing, semiconductor manufacturing, automobile systems, and medical imaging.

  3. Orbit Determination and Gravity Field Estimation of the Dawn spacecraft at Vesta Using Radiometric and Image Constraints with GEODYN Software

    NASA Astrophysics Data System (ADS)

    Centinello, F. J.; Zuber, M. T.; Mazarico, E.

    2013-12-01

    The Dawn spacecraft orbited the protoplanet Vesta from May 3, 2011 to July 25, 2012. Precise orbit determination was critical for the geophysical investigation, as well as the definition of the Vesta-fixed reference frame and the subsequent registration of datasets to the surface. GEODYN, the orbit determination and geodetic parameter estimation software of NASA Goddard Spaceflight Center, was used to compute the orbit of the Dawn spacecraft and estimate the gravity field of Vesta. GEODYN utilizes radiometric Doppler and range measurements, and was modified to process image data from Dawn's cameras. X-band radiometric measurements were acquired by the NASA Deep Space Network (DSN). The addition of the capability to process image constraints decreases position uncertainty in the along- and cross-orbit track directions because of their geometric strengths compared with radiometric measurements. This capability becomes critical for planetary missions such as Dawn due to the weak gravity environment, where non-conservative forces affect the orbit more than typical of orbits at larger planetary bodies. Radiometric measurements were fit to less than 0.1 mm/s and 5 m for Doppler and range during the Survey orbit phase (compared with measurement noise RMS of about 0.05 mm/s and 2 m for Doppler and range). Image constraint RMS was fit to less than 100 m (resolution is 5 - 150 m/pixel, depending on the spacecraft altitude). Orbits computed using GEODYN were used to estimate a 20th degree and order gravity field of Vesta. The quality of the orbit determination and estimated gravity field with and without image constraints was assessed through comparison with the spacecraft trajectory and gravity model provided by the Dawn Science Team.

  4. Man-machine interactive imaging and data processing using high-speed digital mass storage

    NASA Technical Reports Server (NTRS)

    Alsberg, H.; Nathan, R.

    1975-01-01

    The role of vision in teleoperation has been recognized as an important element in the man-machine control loop. In most applications of remote manipulation, direct vision cannot be used. To overcome this handicap, the human operator's control capabilities are augmented by a television system. This medium provides a practical and useful link between workspace and the control station from which the operator perform his tasks. Human performance deteriorates when the images are degraded as a result of instrumental and transmission limitations. Image enhancement is used to bring out selected qualities in a picture to increase the perception of the observer. A general purpose digital computer, an extensive special purpose software system is used to perform an almost unlimited repertoire of processing operations.

  5. Fundus image fusion in EYEPLAN software: An evaluation of a novel technique for ocular melanoma radiation treatment planning

    SciTech Connect

    Daftari, Inder K.; Mishra, Kavita K.; O'Brien, Joan M.; and others

    2010-10-15

    Purpose: The purpose of this study is to evaluate a novel approach for treatment planning using digital fundus image fusion in EYEPLAN for proton beam radiation therapy (PBRT) planning for ocular melanoma. The authors used a prototype version of EYEPLAN software, which allows for digital registration of high-resolution fundus photographs. The authors examined the improvement in tumor localization by replanning with the addition of fundus photo superimposition in patients with macular area tumors. Methods: The new version of EYEPLAN (v3.05) software allows for the registration of fundus photographs as a background image. This is then used in conjunction with clinical examination, tantalum marker clips, surgeon's mapping, and ultrasound to draw the tumor contour accurately. In order to determine if the fundus image superimposition helps in tumor delineation and treatment planning, the authors identified 79 patients with choroidal melanoma in the macular location that were treated with PBRT. All patients were treated to a dose of 56 GyE in four fractions. The authors reviewed and replanned all 79 macular melanoma cases with superimposition of pretreatment and post-treatment fundus imaging in the new EYEPLAN software. For patients with no local failure, the authors analyzed whether fundus photograph fusion accurately depicted and confirmed tumor volumes as outlined in the original treatment plan. For patients with local failure, the authors determined whether the addition of the fundus photograph might have benefited in terms of more accurate tumor volume delineation. Results: The mean follow-up of patients was 33.6{+-}23 months. Tumor growth was seen in six eyes of the 79 macular lesions. All six patients were marginal failures or tumor miss in the region of dose fall-off, including one patient with both in-field recurrence as well as marginal. Among the six recurrences, three were managed by enucleation and one underwent retreatment with proton therapy. Three

  6. C++ software integration for a high-throughput phase imaging platform

    NASA Astrophysics Data System (ADS)

    Kandel, Mikhail E.; Luo, Zelun; Han, Kevin; Popescu, Gabriel

    2015-03-01

    The multi-shot approach in SLIM requires reliable, synchronous, and parallel operation of three independent hardware devices - not meeting these challenges results in degraded phase and slow acquisition speeds, narrowing applications to holistic statements about complex phenomena. The relative youth of quantitative imaging and the lack of ready-made commercial hardware and tools further compounds the problem as Higher level programming languages result in inflexible, experiment specific instruments limited by ill-fitting computational modules, resulting in a palpable chasm between promised and realized hardware performance. Furthermore, general unfamiliarity with intricacies such as background calibration, objective lens attenuation, along with spatial light modular alignment, makes successful measurements difficult for the inattentive or uninitiated. This poses an immediate challenge for moving our techniques beyond the lab to biologically oriented collaborators and clinical practitioners. To meet these challenges, we present our new Quantitative Phase Imaging pipeline, with improved instrument performance, friendly user interface and robust data processing features, enabling us to acquire and catalog clinical datasets hundreds of gigapixels in size.

  7. FIRE: an open-software suite for real-time 2D/3D image registration for image guided radiotherapy research

    NASA Astrophysics Data System (ADS)

    Furtado, H.; Gendrin, C.; Spoerk, J.; Steiner, E.; Underwood, T.; Kuenzler, T.; Georg, D.; Birkfellner, W.

    2016-03-01

    Radiotherapy treatments have changed at a tremendously rapid pace. Dose delivered to the tumor has escalated while organs at risk (OARs) are better spared. The impact of moving tumors during dose delivery has become higher due to very steep dose gradients. Intra-fractional tumor motion has to be managed adequately to reduce errors in dose delivery. For tumors with large motion such as tumors in the lung, tracking is an approach that can reduce position uncertainty. Tumor tracking approaches range from purely image intensity based techniques to motion estimation based on surrogate tracking. Research efforts are often based on custom designed software platforms which take too much time and effort to develop. To address this challenge we have developed an open software platform especially focusing on tumor motion management. FLIRT is a freely available open-source software platform. The core method for tumor tracking is purely intensity based 2D/3D registration. The platform is written in C++ using the Qt framework for the user interface. The performance critical methods are implemented on the graphics processor using the CUDA extension. One registration can be as fast as 90ms (11Hz). This is suitable to track tumors moving due to respiration (~0.3Hz) or heartbeat (~1Hz). Apart from focusing on high performance, the platform is designed to be flexible and easy to use. Current use cases range from tracking feasibility studies, patient positioning and method validation. Such a framework has the potential of enabling the research community to rapidly perform patient studies or try new methods.

  8. Leap Motion Gesture Control With Carestream Software in the Operating Room to Control Imaging: Installation Guide and Discussion.

    PubMed

    Pauchot, Julien; Di Tommaso, Laetitia; Lounis, Ahmed; Benassarou, Mourad; Mathieu, Pierre; Bernot, Dominique; Aubry, Sébastien

    2015-12-01

    Nowadays, routine cross-sectional imaging viewing during a surgical procedure requires physical contact with an interface (mouse or touch-sensitive screen). Such contact risks exposure to aseptic conditions and causes loss of time. Devices such as the recently introduced Leap Motion (Leap Motion Society, San Francisco, CA), which enables interaction with the computer without any physical contact, are of wide interest in the field of surgery, but configuration and ergonomics are key challenges for the practitioner, imaging software, and surgical environment. This article aims to suggest an easy configuration of Leap Motion on a PC for optimized use with Carestream Vue PACS v11.3.4 (Carestream Health, Inc, Rochester, NY) using a plug-in (to download at https://drive.google.com/open?id=0B_F4eBeBQc3yNENvTXlnY09qS00&authuser=0) and a video tutorial (https://www.youtube.com/watch?v=yVPTgxg-SIk). Videos of surgical procedure and discussion about innovative gesture control technology and its various configurations are provided in this article. PMID:26002115

  9. Using Image Pro Plus Software to Develop Particle Mapping on Genesis Solar Wind Collector Surfaces

    NASA Technical Reports Server (NTRS)

    Rodriquez, Melissa C.; Allton, J. H.; Burkett, P. J.

    2012-01-01

    The continued success of the Genesis mission science team in analyzing solar wind collector array samples is partially based on close collaboration of the JSC curation team with science team members who develop cleaning techniques and those who assess elemental cleanliness at the levels of detection. The goal of this collaboration is to develop a reservoir of solar wind collectors of known cleanliness to be available to investigators. The heart and driving force behind this effort is Genesis mission PI Don Burnett. While JSC contributes characterization, safe clean storage, and benign collector cleaning with ultrapure water (UPW) and UV ozone, Burnett has coordinated more exotic and rigorous cleaning which is contributed by science team members. He also coordinates cleanliness assessment requiring expertise and instruments not available in curation, such as XPS, TRXRF [1,2] and synchrotron TRXRF. JSC participates by optically documenting the particle distributions as cleaning steps progress. Thus, optical document supplements SEM imaging and analysis, and elemental assessment by TRXRF.

  10. Cytopathology whole slide images and virtual microscopy adaptive tutorials: A software pilot

    PubMed Central

    Van Es, Simone L.; Pryor, Wendy M.; Belinson, Zack; Salisbury, Elizabeth L.; Velan, Gary M.

    2015-01-01

    Background: The constant growth in the body of knowledge in medicine requires pathologists and pathology trainees to engage in continuing education. Providing them with equitable access to efficient and effective forms of education in pathology (especially in remote and rural settings) is important, but challenging. Methods: We developed three pilot cytopathology virtual microscopy adaptive tutorials (VMATs) to explore a novel adaptive E-learning platform (AeLP) which can incorporate whole slide images for pathology education. We collected user feedback to further develop this educational material and to subsequently deploy randomized trials in both pathology specialist trainee and also medical student cohorts. Cytopathology whole slide images were first acquired then novel VMATs teaching cytopathology were created using the AeLP, an intelligent tutoring system developed by Smart Sparrow. The pilot was run for Australian pathologists and trainees through the education section of Royal College of Pathologists of Australasia website over a period of 9 months. Feedback on the usability, impact on learning and any technical issues was obtained using 5-point Likert scale items and open-ended feedback in online questionnaires. Results: A total of 181 pathologists and pathology trainees anonymously attempted the three adaptive tutorials, a smaller proportion of whom went on to provide feedback at the end of each tutorial. VMATs were perceived as effective and efficient E-learning tools for pathology education. User feedback was positive. There were no significant technical issues. Conclusion: During this pilot, the user feedback on the educational content and interface and the lack of technical issues were helpful. Large scale trials of similar online cytopathology adaptive tutorials were planned for the future. PMID:26605119

  11. An Upgrade of the Imaging for Hypersonic Experimental Aeroheating Testing (IHEAT) Software

    NASA Technical Reports Server (NTRS)

    Mason, Michelle L.; Rufer, Shann J.

    2015-01-01

    The Imaging for Hypersonic Experimental Aeroheating Testing (IHEAT) code is used at NASA Langley Research Center to analyze global aeroheating data on wind tunnel models tested in the Langley Aerothermodynamics Laboratory. One-dimensional, semi-infinite heating data derived from IHEAT are used to design thermal protection systems to mitigate the risks due to the aeroheating loads on hypersonic vehicles, such as re-entry vehicles during descent and landing procedures. This code was originally written in the PV-WAVE programming language to analyze phosphor thermography data from the two-color, relativeintensity system developed at Langley. To increase the efficiency, functionality, and reliability of IHEAT, the code was migrated to MATLAB syntax and compiled as a stand-alone executable file labeled version 4.0. New features of IHEAT 4.0 include the options to batch process all of the data from a wind tunnel run, to map the two-dimensional heating distribution to a three-dimensional computer-aided design model of the vehicle to be viewed in Tecplot, and to extract data from a segmented line that follows an interesting feature in the data. Results from IHEAT 4.0 were compared on a pixel level to the output images from the legacy code to validate the program. The differences between the two codes were on the order of 10-5 to 10-7. IHEAT 4.0 replaces the PV-WAVE version as the production code for aeroheating experiments conducted in the hypersonic facilities at NASA Langley.

  12. The Decoding Toolbox (TDT): a versatile software package for multivariate analyses of functional imaging data

    PubMed Central

    Hebart, Martin N.; Görgen, Kai; Haynes, John-Dylan

    2015-01-01

    The multivariate analysis of brain signals has recently sparked a great amount of interest, yet accessible and versatile tools to carry out decoding analyses are scarce. Here we introduce The Decoding Toolbox (TDT) which represents a user-friendly, powerful and flexible package for multivariate analysis of functional brain imaging data. TDT is written in Matlab and equipped with an interface to the widely used brain data analysis package SPM. The toolbox allows running fast whole-brain analyses, region-of-interest analyses and searchlight analyses, using machine learning classifiers, pattern correlation analysis, or representational similarity analysis. It offers automatic creation and visualization of diverse cross-validation schemes, feature scaling, nested parameter selection, a variety of feature selection methods, multiclass capabilities, and pattern reconstruction from classifier weights. While basic users can implement a generic analysis in one line of code, advanced users can extend the toolbox to their needs or exploit the structure to combine it with external high-performance classification toolboxes. The toolbox comes with an example data set which can be used to try out the various analysis methods. Taken together, TDT offers a promising option for researchers who want to employ multivariate analyses of brain activity patterns. PMID:25610393

  13. The Decoding Toolbox (TDT): a versatile software package for multivariate analyses of functional imaging data.

    PubMed

    Hebart, Martin N; Görgen, Kai; Haynes, John-Dylan

    2014-01-01

    The multivariate analysis of brain signals has recently sparked a great amount of interest, yet accessible and versatile tools to carry out decoding analyses are scarce. Here we introduce The Decoding Toolbox (TDT) which represents a user-friendly, powerful and flexible package for multivariate analysis of functional brain imaging data. TDT is written in Matlab and equipped with an interface to the widely used brain data analysis package SPM. The toolbox allows running fast whole-brain analyses, region-of-interest analyses and searchlight analyses, using machine learning classifiers, pattern correlation analysis, or representational similarity analysis. It offers automatic creation and visualization of diverse cross-validation schemes, feature scaling, nested parameter selection, a variety of feature selection methods, multiclass capabilities, and pattern reconstruction from classifier weights. While basic users can implement a generic analysis in one line of code, advanced users can extend the toolbox to their needs or exploit the structure to combine it with external high-performance classification toolboxes. The toolbox comes with an example data set which can be used to try out the various analysis methods. Taken together, TDT offers a promising option for researchers who want to employ multivariate analyses of brain activity patterns. PMID:25610393

  14. User interface software development for the WIYN One Degree Imager (ODI)

    NASA Astrophysics Data System (ADS)

    Ivens, John; Yeatts, Andrey; Harbeck, Daniel; Martin, Pierre

    2010-07-01

    User interfaces (UIs) are a necessity for almost any data acquisition system. The development team for the WIYN One Degree Imager (ODI) chose to develop a user interface that allows access to most of the instrument control for both scientists and engineers through the World Wide Web, because of the web's ease of use and accessibility around the world. Having a web based UI allows ODI to grow from a visitor-mode instrument to a queue-managed instrument and also facilitate remote servicing and troubleshooting. The challenges of developing such a system involve the difficulties of browser inter-operability, speed, presentation, and the choices involved with integrating browser and server technologies. To this end, the team has chosen a combination of Java, JBOSS, AJAX technologies, XML data descriptions, Oracle XML databases, and an emerging technology called the Google Web Toolkit (GWT) that compiles Java into Javascript for presentation in a browser. Advantages of using GWT include developing the front end browser code in Java, GWT's native support for AJAX, the use of XML to describe the user interface, the ability to profile code speed and discover bottlenecks, the ability to efficiently communicate with application servers such as JBOSS, and the ability to optimize and test code for multiple browsers. We discuss the inter-operation of all of these technologies to create fast, flexible, and robust user interfaces that are scalable, manageable, separable, and as much as possible allow maintenance of all code in Java.

  15. Analysis of Endoscopic Electronic Image of Intramucosal Gastric Carcinoma Using a Software Program for Calculating Hemoglobin Index

    PubMed Central

    Kim, Gwang Ha; Kim, Kwang Baek; Lim, Eun Kyung; Choi, Seong Ho; Kim, Tae Oh; Heo, Jeong; Kang, Dae Hwan; Cho, Mong; Park, Do Youn

    2006-01-01

    Hemoglobin is the predominent pigment in the gastrointestinal mucosa, and the development of electronic endoscopy has made it possible to quantitatively measure the mucosal hemoglobin volume, by using a hemoglobin index (IHb). The aims of this study were to make a software program to calculate the IHb and then to investigate whether the mucosal IHb determined from the electronic endoscopic data is a useful marker for evaluating the color of intramucosal gastric carcinoma with regard to its value for discriminating between the histologic types. We made a software program for calculating the IHb in the endoscopic images. By using this program, the mean values of the IHb for the carcinoma (IHb-C) and those of the IHb for the surrounding non-cancerous mucosa (IHb-N) were calculated in 75 intestinal-type and 34 diffuse-type intramucosal gastric carcinomas. We then analyzed the ratio of the IHb-C to the IHb-N (C/N ratio). The C/N ratio in the intestinal-type carcinoma group was higher than that in the diffuse-type carcinoma group (p<0.001). In the diffuse-type carcinoma group, the C/N ratio in the body was lower than that in the antrum (p=0.022). The accuracy rate, sensitivity, specificity, and the positive and negative predictive values for the differential diagnosis of the diffuse-type carcinoma from the intestinal-type carcinoma were 94.5%, 94.1%, 94.7%, 88.9% and 97.3%, respectively. IHb is useful for making quantitative measurement of the endoscopic color in the intramucosal gastric carcinoma, and the C/N ratio by using the IHb would be helpful for distinguishing the diffuse-type carcinoma from the intestinal-type carcinoma. PMID:17179684

  16. The Role of A Facilitated Online Workspace Component of A Community of Practice: Knowledge Building and Value Creation for NASA

    NASA Astrophysics Data System (ADS)

    Davey, B.

    2014-12-01

    This study examined the role of an online workspace component of a community in the work of a community of practice. Much has been studied revealing the importance of communities of practice to organizations, project success, and knowledge management and some of these same successes hold true for virtual communities of practice. Study participants were 75 Education and Public Outreach community members of NASA's Science Mission Directorate Earth Forum. In this mixed methods study, online workspace metrics were used to track participation and a survey completed by 21 members was used to quantify participation. For a more detailed analysis, 15 community members (5 highly active users, 5 average users, and 5 infrequent users) selected based on survey responses, were interviewed. Finally, survey data was gathered from 7 online facilitators to understand their role in the community. Data collected from these 21 community members and 5 facilitating members suggest that highly active users (logging into the workspace daily), were more likely to have transformative experiences, co-create knowledge, feel ownership of community knowledge, have extended opportunities for community exchange, and find new forms of evaluation. Average users shared some similar characteristics with both the highly active members and infrequent users, representing a group in transition as they become more engaged and active in the online workspace. Inactive users viewed the workspace as having little value, being difficult to navigate, being mainly for gaining basic information about events and community news, and as another demand on their time. Results show the online workspace component of the Earth Science Education and Outreach Forum is playing an important and emerging role for this community by supporting knowledge building and knowledge sharing, and growing in value for those that utilizing it more frequently. The evidence suggests that with increased participation or "usage" comes

  17. Comparison of retinal thickness by Fourier-domain optical coherence tomography and OCT retinal image analysis software segmentation analysis derived from Stratus optical coherence tomography images

    NASA Astrophysics Data System (ADS)

    Tátrai, Erika; Ranganathan, Sudarshan; Ferencz, Mária; Debuc, Delia Cabrera; Somfai, Gábor Márk

    2011-05-01

    Purpose: To compare thickness measurements between Fourier-domain optical coherence tomography (FD-OCT) and time-domain OCT images analyzed with a custom-built OCT retinal image analysis software (OCTRIMA). Methods: Macular mapping (MM) by StratusOCT and MM5 and MM6 scanning protocols by an RTVue-100 FD-OCT device are performed on 11 subjects with no retinal pathology. Retinal thickness (RT) and the thickness of the ganglion cell complex (GCC) obtained with the MM6 protocol are compared for each early treatment diabetic retinopathy study (ETDRS)-like region with corresponding results obtained with OCTRIMA. RT results are compared by analysis of variance with Dunnett post hoc test, while GCC results are compared by paired t-test. Results: A high correlation is obtained for the RT between OCTRIMA and MM5 and MM6 protocols. In all regions, the StratusOCT provide the lowest RT values (mean difference 43 +/- 8 μm compared to OCTRIMA, and 42 +/- 14 μm compared to RTVue MM6). All RTVue GCC measurements were significantly thicker (mean difference between 6 and 12 μm) than the GCC measurements of OCTRIMA. Conclusion: High correspondence of RT measurements is obtained not only for RT but also for the segmentation of intraretinal layers between FD-OCT and StratusOCT-derived OCTRIMA analysis. However, a correction factor is required to compensate for OCT-specific differences to make measurements more comparable to any available OCT device.

  18. A user's guide for the signal processing software for image and speech compression developed in the Communications and Signal Processing Laboratory (CSPL), version 1

    NASA Technical Reports Server (NTRS)

    Kumar, P.; Lin, F. Y.; Vaishampayan, V.; Farvardin, N.

    1986-01-01

    A complete documentation of the software developed in the Communication and Signal Processing Laboratory (CSPL) during the period of July 1985 to March 1986 is provided. Utility programs and subroutines that were developed for a user-friendly image and speech processing environment are described. Additional programs for data compression of image and speech type signals are included. Also, programs for the zero-memory and block transform quantization in the presence of channel noise are described. Finally, several routines for simulating the perfromance of image compression algorithms are included.

  19. Computer Software.

    ERIC Educational Resources Information Center

    Kay, Alan

    1984-01-01

    Discusses the nature and development of computer software. Programing, programing languages, types of software (including dynamic spreadsheets), and software of the future are among the topics considered. (JN)

  20. Flexible workspace design and ergonomics training: impacts on the psychosocial work environment, musculoskeletal health, and work effectiveness among knowledge workers.

    PubMed

    Robertson, Michelle M; Huang, Yueng-Hsiang; O'Neill, Michael J; Schleifer, Lawrence M

    2008-07-01

    A macroergonomics intervention consisting of flexible workspace design and ergonomics training was conducted to examine the effects on psychosocial work environment, musculoskeletal health, and work effectiveness in a computer-based office setting. Knowledge workers were assigned to one of four conditions: flexible workspace (n=121), ergonomics training (n=92), flexible workspace+ergonomics training (n=31), and a no-intervention control (n=45). Outcome measures were collected 2 months prior to the intervention and 3 and 6 months post-intervention. Overall, the study results indicated positive, significant effects on the outcome variables for the two intervention groups compared to the control group, including work-related musculoskeletal discomfort, job control, environmental satisfaction, sense of community, ergonomic climate, communication and collaboration, and business process efficiency (time and costs). However, attrition of workers in the ergonomics training condition precluded an evaluation of the effects of this intervention. This study suggests that a macroergonomics intervention is effective among knowledge workers in office settings. PMID:18462704

  1. A novel scanning system using an industrial robot and the workspace measurement and positioning system

    NASA Astrophysics Data System (ADS)

    Zhao, Ziyue; Zhu, Jigui; Yang, Linghui; Lin, Jiarui

    2015-10-01

    The present scanning system consists of an industrial robot and a line-structured laser sensor which uses the industrial robot as a position instrument to guarantee the accuracy. However, the absolute accuracy of an industrial robot is relatively poor compared with the good repeatability in the manufacturing industry. This paper proposes a novel method using the workspace measurement and positioning system (wMPS) to remedy the lack of accuracy of the industrial robot. In order to guarantee the positioning accuracy of the system, the wMPS which is a laser-based measurement technology designed for large-volume metrology applications is brought in. Benefitting from the wMPS, this system can measure different cell-areas by the line-structured laser sensor and fuse the measurement data of different cell-areas by using the wMPS accurately. The system calibration which is the procedure to acquire and optimize the structure parameters of the scanning system is also stated in detail in this paper. In order to verify the feasibility of the system for scanning the large free-form surface, an experiment is designed to scan the internal surface of the door of a car-body in white. The final results show that the measurement data of the whole measuring areas have been jointed perfectly and there is no mismatch in the figure especially in the hole measuring areas. This experiment has verified the rationality of the system scheme, the correctness and effectiveness of the relevant methods.

  2. Assessment of Workspace Attributes Under Simulated Index Finger Proximal Interphalangeal Arthrodesis.

    PubMed

    Arauz, Paul G; Sisto, Sue A; Kao, Imin

    2016-05-01

    This article presented an assessment of quantitative measures of workspace (WS) attributes under simulated proximal interphalangeal (PIP) joint arthrodesis of the index finger. Seven healthy subjects were tested with the PIP joint unconstrained (UC) and constrained to selected angles using a motion analysis system. A model of the constrained finger was developed in order to address the impact of the inclusion of prescribed joint arthrodesis angles on WS attributes. Model parameters were obtained from system identification experiments involving flexion-extension (FE) movements of the UC and constrained finger. The data of experimental FE movements of the constrained finger were used to generate the two-dimensional (2D) WS boundaries and to validate the model. A weighted criterion was formulated to define an optimal constraint angle among several system parameters. Results indicated that a PIP joint immobilization angle of 40-50 deg of flexion maximized the 2D WS. The analysis of the aspect ratio of the 2D WS indicated that the WS was more evenly distributed as the imposed PIP joint constraint angle increased. With the imposed PIP joint constraint angles of 30 deg, 40 deg, 50 deg, and 60 deg of flexion, the normalized maximum distance of fingertip reach was reduced by approximately 3%, 4%, 7%, and 9%, respectively. PMID:26974649

  3. Effect of Tendon Vibration on Hemiparetic Arm Stability in Unstable Workspaces

    PubMed Central

    Conrad, Megan O.; Gadhoke, Bani; Scheidt, Robert A.; Schmit, Brian D.

    2015-01-01

    Sensory stimulation of wrist musculature can enhance stability in the proximal arm and may be a useful therapy aimed at improving arm control post-stroke. Specifically, our prior research indicates tendon vibration can enhance stability during point-to-point arm movements and in tracking tasks. The goal of the present study was to investigate the influence of forearm tendon vibration on endpoint stability, measured at the hand, immediately following forward arm movements in an unstable environment. Both proximal and distal workspaces were tested. Ten hemiparetic stroke subjects and 5 healthy controls made forward arm movements while grasping the handle of a two-joint robotic arm. At the end of each movement, the robot applied destabilizing forces. During some trials, 70 Hz vibration was applied to the forearm flexor muscle tendons. 70 Hz was used as the stimulus frequency as it lies within the range of optimal frequencies that activate the muscle spindles at the highest response rate. Endpoint position, velocity, muscle activity and grip force data were compared before, during and after vibration. Stability at the endpoint was quantified as the magnitude of oscillation about the target position, calculated from the power of the tangential velocity data. Prior to vibration, subjects produced unstable, oscillating hand movements about the target location due to the applied force field. Stability increased during vibration, as evidenced by decreased oscillation in hand tangential velocity. PMID:26633892

  4. Effect of Tendon Vibration on Hemiparetic Arm Stability in Unstable Workspaces.

    PubMed

    Conrad, Megan O; Gadhoke, Bani; Scheidt, Robert A; Schmit, Brian D

    2015-01-01

    Sensory stimulation of wrist musculature can enhance stability in the proximal arm and may be a useful therapy aimed at improving arm control post-stroke. Specifically, our prior research indicates tendon vibration can enhance stability during point-to-point arm movements and in tracking tasks. The goal of the present study was to investigate the influence of forearm tendon vibration on endpoint stability, measured at the hand, immediately following forward arm movements in an unstable environment. Both proximal and distal workspaces were tested. Ten hemiparetic stroke subjects and 5 healthy controls made forward arm movements while grasping the handle of a two-joint robotic arm. At the end of each movement, the robot applied destabilizing forces. During some trials, 70 Hz vibration was applied to the forearm flexor muscle tendons. 70 Hz was used as the stimulus frequency as it lies within the range of optimal frequencies that activate the muscle spindles at the highest response rate. Endpoint position, velocity, muscle activity and grip force data were compared before, during and after vibration. Stability at the endpoint was quantified as the magnitude of oscillation about the target position, calculated from the power of the tangential velocity data. Prior to vibration, subjects produced unstable, oscillating hand movements about the target location due to the applied force field. Stability increased during vibration, as evidenced by decreased oscillation in hand tangential velocity. PMID:26633892

  5. A multi-sensorial hybrid control for robotic manipulation in human-robot workspaces.

    PubMed

    Pomares, Jorge; Perea, Ivan; García, Gabriel J; Jara, Carlos A; Corrales, Juan A; Torres, Fernando

    2011-01-01

    Autonomous manipulation in semi-structured environments where human operators can interact is an increasingly common task in robotic applications. This paper describes an intelligent multi-sensorial approach that solves this issue by providing a multi-robotic platform with a high degree of autonomy and the capability to perform complex tasks. The proposed sensorial system is composed of a hybrid visual servo control to efficiently guide the robot towards the object to be manipulated, an inertial motion capture system and an indoor localization system to avoid possible collisions between human operators and robots working in the same workspace, and a tactile sensor algorithm to correctly manipulate the object. The proposed controller employs the whole multi-sensorial system and combines the measurements of each one of the used sensors during two different phases considered in the robot task: a first phase where the robot approaches the object to be grasped, and a second phase of manipulation of the object. In both phases, the unexpected presence of humans is taken into account. This paper also presents the successful results obtained in several experimental setups which verify the validity of the proposed approach. PMID:22163729

  6. Images.

    ERIC Educational Resources Information Center

    Barr, Catherine, Ed.

    1997-01-01

    The theme of this month's issue is "Images"--from early paintings and statuary to computer-generated design. Resources on the theme include Web sites, CD-ROMs and software, videos, books, and others. A page of reproducible activities is also provided. Features include photojournalism, inspirational Web sites, art history, pop art, and myths. (AEF)

  7. Comparison of retinal thickness by Fourier-domain optical coherence tomography and OCT retinal image analysis software segmentation analysis derived from Stratus optical coherence tomography images

    PubMed Central

    Tátrai, Erika; Ranganathan, Sudarshan; Ferencz, Mária; DeBuc, Delia Cabrera; Somfai, Gábor Márk

    2011-01-01

    Purpose: To compare thickness measurements between Fourier-domain optical coherence tomography (FD-OCT) and time-domain OCT images analyzed with a custom-built OCT retinal image analysis software (OCTRIMA). Methods: Macular mapping (MM) by StratusOCT and MM5 and MM6 scanning protocols by an RTVue-100 FD-OCT device are performed on 11 subjects with no retinal pathology. Retinal thickness (RT) and the thickness of the ganglion cell complex (GCC) obtained with the MM6 protocol are compared for each early treatment diabetic retinopathy study (ETDRS)-like region with corresponding results obtained with OCTRIMA. RT results are compared by analysis of variance with Dunnett post hoc test, while GCC results are compared by paired t-test. Results: A high correlation is obtained for the RT between OCTRIMA and MM5 and MM6 protocols. In all regions, the StratusOCT provide the lowest RT values (mean difference 43 ± 8 μm compared to OCTRIMA, and 42 ± 14 μm compared to RTVue MM6). All RTVue GCC measurements were significantly thicker (mean difference between 6 and 12 μm) than the GCC measurements of OCTRIMA. Conclusion: High correspondence of RT measurements is obtained not only for RT but also for the segmentation of intraretinal layers between FD-OCT and StratusOCT-derived OCTRIMA analysis. However, a correction factor is required to compensate for OCT-specific differences to make measurements more comparable to any available OCT device. PMID:21639572

  8. Semiquantitative analysis of in-situ hybridization results using IMAGE software: a rapid method for counting reduced silver grains over mRNA-positive cells.

    PubMed

    Lucas, L R; Mize, R R; Harlan, R E

    1994-06-01

    The advent of microcomputers has brought about a revolution in the computing power available to the average user. Image analysis is a very resource-intensive process, making great demands on computing power, memory, and display capabilities of most computers. Thus, in the past, dedicated, single-use hardware and software had to be custom made for environments requiring image analysis. We present here an easy-to-use image analysis protocol available to most users with a Macintosh II series computer and access to IMAGE (a public domain image analysis program). The protocol allows for semi-quantitation of silver grains over cells used in the interpretation of in-situ hybridization results. We show that the method provides a quick and reliable means of counting grains over mRNA-positive cells in an automated fashion. We also provide evidence that the method can be used to detect differences between experimental treatments. PMID:7526081

  9. NeuronGrowth, a software for automatic quantification of neurite and filopodial dynamics from time-lapse sequences of digital images.

    PubMed

    Fanti, Zian; Martinez-Perez, M Elena; De-Miguel, Francisco F

    2011-10-01

    We developed NeuronGrowth, a software for the automatic quantification of extension and retraction of neurites and filopodia, from time-lapse sequences of two-dimensional digital micrographs. NeuronGrowth requires a semiautomatic characterization of individual neurites in a reference frame, which is then used for automatic tracking and measurement of every neurite over the whole image sequence. Modules for sequence alignment, background subtraction, flat field correction, light normalization, and cropping have been integrated to improve the quality of the analysis. Moreover, NeuronGrowth incorporates a deconvolution filter that corrects the shadow-cast effect of differential interference contrast (DIC) images. NeuronGrowth was tested by analyzing the formation of outgrowth patterns by individual leech neurons cultured under two different conditions. Phase contrast images were obtained from neurons plated on CNS homogenates and DIC images were obtained from similar neurons plated on ganglion capsules as substrates. Filopodia were measured from fluorescent growth-cones of chick dorsal root ganglion cells. Quantitative data of neurite extension and retraction obtained by three different users applying NeuronGrowth and two other manually operated software packages were similar. However, NeuronGrowth required less user participation and had a better time performance when compared with the other software packages. NeuronGrowth may be used in general to quantify the dynamics of tubular structures such as blood vessels. NeuronGrowth is a free plug-in for the free software ImageJ and can be downloaded along with a user manual, a troubleshooting section and other information required for its use from http://www.ifc.unam.mx or http://www.ifc.unam.mx/ffm/index.html. PMID:21913334

  10. Development of image quality assurance measures of the ExacTrac localization system using commercially available image evaluation software and hardware for image-guided radiotherapy.

    PubMed

    Stanley, Dennis N; Papanikolaou, Nikos; Gutiérrez, Alonso N

    2014-01-01

    Quality assurance (QA) of the image quality for image-guided localization systems is crucial to ensure accurate visualization and localization of target volumes. In this study, a methodology was developed to assess and evaluate the constancy of the high-contrast spatial resolution, dose, energy, contrast, and geometrical accuracy of the BrainLAB ExacTrac system. An in-house fixation device was constructed to hold the QCkV-1 phantom firmly and reproducibly against the face of the flat panel detectors. Two image sets per detector were acquired using ExacTrac preset console settings over a period of three months. The image sets were analyzed in PIPSpro and the following metrics were recorded: high-contrast spatial resolution (f30, f40, f50 (lp/mm)), noise, and contrast-to-noise ratio. Geometrical image accu- racy was evaluated by assessing the length between to predetermined points of the QCkV-1 phantom. Dose and kVp were recorded using the Unfors RaySafe Xi R/F Detector. The kVp and dose were evaluated for the following: Cranial Standard (CS) (80 kV,80 mA,80 ms), Thorax Standard (TS) (120 kV,160 mA,160 ms), Abdomen Standard (AS) (120 kV,160 mA,130 ms), and Pelvis Standard (PS) (120 kV,160 mA,160 ms). With regard to high-contrast spatial resolution, the mean values of the f30 (lp/mm), f40 (lp/mm) and f50 (lp/mm) for the left detector were 1.39 ± 0.04, 1.24 ± 0.05, and 1.09 ± 0.04, respectively, while for the right detector they were 1.38 ± 0.04, 1.22 ± 0.05, and 1.09 ± 0.05, respectively. Mean CNRs for the left and right detectors were 148 ± 3 and 143 ± 4, respectively. For geometrical accuracy, both detectors had a measured image length of the QCkV-1 of 57.9 ± 0.5 mm. The left detector showed dose measurements of 20.4 ± 0.2 μGy (CS), 191.8 ± 0.7 μGy (TS), 154.2 ± 0.7 μGy (AS), and 192.2 ± 0.6 μGy (PS), while the right detector showed 20.3 ± 0.3 μGy (CS), 189.7 ± 0.8 μGy (TS), 151.0 ± 0.7 μGy (AS), and 189.7 ± 0.8 μGy (PS), respectively. For X

  11. Model-based software engineering for an imaging CubeSat and its extrapolation to other missions

    NASA Astrophysics Data System (ADS)

    Mohammad, Atif; Straub, Jeremy; Korvald, Christoffer; Grant, Emanuel

    Small satellites with their limited computational capabilities require that software engineering techniques promote efficient use of spacecraft resources. A model-driven approach to software engineering is an excellent solution to this resource maximization challenge as it facilitates visualization of the key solution processes and data elements.

  12. Design principles for innovative workspaces to increase efficiency in pharmaceutical R&D: lessons learned from the Novartis campus.

    PubMed

    Zoller, Frank A; Boutellier, Roman

    2013-04-01

    When managing R&D departments for increased efficiency and effectiveness the focus has often been on organizational structure. Space is, however, of outstanding importance in an environment of large task uncertainty, which is the case in pharmaceutical R&D. Based on case studies about the Novartis campus in Basel, Switzerland, we propose some design principles for laboratory and office workspace to support the strong and weak ties of scientist networks. We address the diversity of technologies and specialization, as well as the pressure on time-to-market, as major challenges in pharmaceutical R&D. PMID:23318251

  13. Selecting Software.

    ERIC Educational Resources Information Center

    Pereus, Steven C.

    2002-01-01

    Describes a comprehensive computer software selection and evaluation process, including documenting district needs, evaluating software packages, weighing the alternatives, and making the purchase. (PKP)

  14. Software-only IR image generation and reticle simulation for the HWIL testing of a single detector frequency modulated reticle seeker

    NASA Astrophysics Data System (ADS)

    Delport, Jan Peet; le Roux, Francois P. J.; du Plooy, Matthys J. U.; Theron, Hendrik J.; Annamalai, Leeandran

    2004-08-01

    Hardware-in-the-Loop (HWIL) testing of seeker systems usually requires a 5-axis flight motion simulator (FMS) coupled to expensive hardware for infrared (IR) scene generation and projection. Similar tests can be conducted by using a 3-axis flight motion simulator, bypassing the seeker optics and injecting a synthetically calculated detector signal directly into the seeker. The constantly increasing speed and memory bandwidth of high-end personal computers make them attractive software rendering platforms. A software OpenGL pipeline provides flexibility in terms of access to the rendered output, colour channel dynamic range and lighting equations. This paper describes how a system was constructed using personal computer hardware to perform closed tracking loop HWIL testing of a single detector frequency modulated reticle seeker. The main parts of the system that are described include: * The software-only implementation of OpenGL used to render the IR image with floating point accuracy directly to system memory. * The software used to inject the detector signal and extract the seeker look position. * The architecture used to control the flight motion simulator.

  15. SynPAnal: Software for Rapid Quantification of the Density and Intensity of Protein Puncta from Fluorescence Microscopy Images of Neurons

    PubMed Central

    Danielson, Eric; Lee, Sang H.

    2014-01-01

    Continuous modification of the protein composition at synapses is a driving force for the plastic changes of synaptic strength, and provides the fundamental molecular mechanism of synaptic plasticity and information storage in the brain. Studying synaptic protein turnover is not only important for understanding learning and memory, but also has direct implication for understanding pathological conditions like aging, neurodegenerative diseases, and psychiatric disorders. Proteins involved in synaptic transmission and synaptic plasticity are typically concentrated at synapses of neurons and thus appear as puncta (clusters) in immunofluorescence microscopy images. Quantitative measurement of the changes in puncta density, intensity, and sizes of specific proteins provide valuable information on their function in synaptic transmission, circuit development, synaptic plasticity, and synaptopathy. Unfortunately, puncta quantification is very labor intensive and time consuming. In this article, we describe a software tool designed for the rapid semi-automatic detection and quantification of synaptic protein puncta from 2D immunofluorescence images generated by confocal laser scanning microscopy. The software, dubbed as SynPAnal (for Synaptic Puncta Analysis), streamlines data quantification for puncta density and average intensity, thereby increases data analysis throughput compared to a manual method. SynPAnal is stand-alone software written using the JAVA programming language, and thus is portable and platform-free. PMID:25531531

  16. SynPAnal: software for rapid quantification of the density and intensity of protein puncta from fluorescence microscopy images of neurons.

    PubMed

    Danielson, Eric; Lee, Sang H

    2014-01-01

    Continuous modification of the protein composition at synapses is a driving force for the plastic changes of synaptic strength, and provides the fundamental molecular mechanism of synaptic plasticity and information storage in the brain. Studying synaptic protein turnover is not only important for understanding learning and memory, but also has direct implication for understanding pathological conditions like aging, neurodegenerative diseases, and psychiatric disorders. Proteins involved in synaptic transmission and synaptic plasticity are typically concentrated at synapses of neurons and thus appear as puncta (clusters) in immunofluorescence microscopy images. Quantitative measurement of the changes in puncta density, intensity, and sizes of specific proteins provide valuable information on their function in synaptic transmission, circuit development, synaptic plasticity, and synaptopathy. Unfortunately, puncta quantification is very labor intensive and time consuming. In this article, we describe a software tool designed for the rapid semi-automatic detection and quantification of synaptic protein puncta from 2D immunofluorescence images generated by confocal laser scanning microscopy. The software, dubbed as SynPAnal (for Synaptic Puncta Analysis), streamlines data quantification for puncta density and average intensity, thereby increases data analysis throughput compared to a manual method. SynPAnal is stand-alone software written using the JAVA programming language, and thus is portable and platform-free. PMID:25531531

  17. Environmental, scanning electron and optical microscope image analysis software for determining volume and occupied area of solid-state fermentation fungal cultures.

    PubMed

    Osma, Johann F; Toca-Herrera, José L; Rodríguez-Couto, Susana

    2011-01-01

    Here we propose a software for the estimation of the occupied area and volume of fungal cultures. This software was developed using a Matlab platform and allows analysis of high-definition images from optical, electronic or atomic force microscopes. In a first step, a single hypha grown on potato dextrose agar was monitored using optical microscopy to estimate the change in occupied area and volume. Weight measurements were carried out to compare them with the estimated volume, revealing a slight difference of less than 1.5%. Similarly, samples from two different solid-state fermentation cultures were analyzed using images from a scanning electron microscope (SEM) and an environmental SEM (ESEM). Occupied area and volume were calculated for both samples, and the results obtained were correlated with the dry weight of the cultures. The difference between the estimated volume ratio and the dry weight ratio of the two cultures showed a difference of 10%. Therefore, this software is a promising non-invasive technique to determine fungal biomass in solid-state cultures. PMID:21154435

  18. Acceptance test of a commercially available software for automatic image registration of computed tomography (CT), magnetic resonance imaging (MRI) and 99mTc-methoxyisobutylisonitrile (MIBI) single-photon emission computed tomography (SPECT) brain images.

    PubMed

    Loi, Gianfranco; Dominietto, Marco; Manfredda, Irene; Mones, Eleonora; Carriero, Alessandro; Inglese, Eugenio; Krengli, Marco; Brambilla, Marco

    2008-09-01

    This note describes a method to characterize the performances of image fusion software (Syntegra) with respect to accuracy and robustness. Computed tomography (CT), magnetic resonance imaging (MRI), and single-photon emission computed tomography (SPECT) studies were acquired from two phantoms and 10 patients. Image registration was performed independently by two couples composed of one radiotherapist and one physicist by means of superposition of anatomic landmarks. Each couple performed jointly and saved the registration. The two solutions were averaged to obtain the gold standard registration. A new set of estimators was defined to identify translation and rotation errors in the coordinate axes, independently from point position in image field of view (FOV). Algorithms evaluated were local correlation (LC) for CT-MRI, normalized mutual information (MI) for CT-MRI, and CT-SPECT registrations. To evaluate accuracy, estimator values were compared to limiting values for the algorithms employed, both in phantoms and in patients. To evaluate robustness, different alignments between images taken from a sample patient were produced and registration errors determined. LC algorithm resulted accurate in CT-MRI registrations in phantoms, but exceeded limiting values in 3 of 10 patients. MI algorithm resulted accurate in CT-MRI and CT-SPECT registrations in phantoms; limiting values were exceeded in one case in CT-MRI and never reached in CT-SPECT registrations. Thus, the evaluation of robustness was restricted to the algorithm of MI both for CT-MRI and CT-SPECT registrations. The algorithm of MI proved to be robust: limiting values were not exceeded with translation perturbations up to 2.5 cm, rotation perturbations up to 10 degrees and roto-translational perturbation up to 3 cm and 5 degrees. PMID:17549564

  19. Development of kinematic equations and determination of workspace of a 6 DOF end-effector with closed-kinematic chain mechanism

    NASA Technical Reports Server (NTRS)

    Nguyen, Charles C.; Pooran, Farhad J.

    1989-01-01

    This report presents results from the research grant entitled Active Control of Robot Manipulators, funded by the Goddard Space Flight Center, under Grant NAG5-780, for the period July 1, 1988 to January 1, 1989. An analysis is presented of a 6 degree-of-freedom robot end-effector built to study telerobotic assembly of NASA hardware in space. Since the end-effector is required to perform high precision motion in a limited workspace, closed-kinematic mechanisms are chosen for its design. A closed-form solution is obtained for the inverse kinematic problem and an iterative procedure employing Newton-Raphson method is proposed to solve the forward kinematic problem. A study of the end-effector workspace results in a general procedure for the workspace determination based on link constraints. Computer simulation results are presented.

  20. Software for 3D radiotherapy dosimetry. Validation

    NASA Astrophysics Data System (ADS)

    Kozicki, Marek; Maras, Piotr; Karwowski, Andrzej C.

    2014-08-01

    The subject of this work is polyGeVero® software (GeVero Co., Poland), which has been developed to fill the requirements of fast calculations of 3D dosimetry data with the emphasis on polymer gel dosimetry for radiotherapy. This software comprises four workspaces that have been prepared for: (i) calculating calibration curves and calibration equations, (ii) storing the calibration characteristics of the 3D dosimeters, (iii) calculating 3D dose distributions in irradiated 3D dosimeters, and (iv) comparing 3D dose distributions obtained from measurements with the aid of 3D dosimeters and calculated with the aid of treatment planning systems (TPSs). The main features and functions of the software are described in this work. Moreover, the core algorithms were validated and the results are presented. The validation was performed using the data of the new PABIGnx polymer gel dosimeter. The polyGeVero® software simplifies and greatly accelerates the calculations of raw 3D dosimetry data. It is an effective tool for fast verification of TPS-generated plans for tumor irradiation when combined with a 3D dosimeter. Consequently, the software may facilitate calculations by the 3D dosimetry community. In this work, the calibration characteristics of the PABIGnx obtained through four calibration methods: multi vial, cross beam, depth dose, and brachytherapy, are discussed as well.

  1. Secure Video Surveillance System Acquisition Software

    SciTech Connect

    2009-12-04

    The SVSS Acquisition Software collects and displays video images from two cameras through a VPN, and store the images onto a collection controller. The software is configured to allow a user to enter a time window to display up to 2 1/2, hours of video review. The software collects images from the cameras at a rate of 1 image per second and automatically deletes images older than 3 hours. The software code operates in a linux environment and can be run in a virtual machine on Windows XP. The Sandia software integrates the different COTS software together to build the video review system.

  2. Secure Video Surveillance System Acquisition Software

    Energy Science and Technology Software Center (ESTSC)

    2009-12-04

    The SVSS Acquisition Software collects and displays video images from two cameras through a VPN, and store the images onto a collection controller. The software is configured to allow a user to enter a time window to display up to 2 1/2, hours of video review. The software collects images from the cameras at a rate of 1 image per second and automatically deletes images older than 3 hours. The software code operates in amore » linux environment and can be run in a virtual machine on Windows XP. The Sandia software integrates the different COTS software together to build the video review system.« less

  3. Use of 3D imaging in CT of the acute trauma patient: impact of a PACS-based software package.

    PubMed

    Soto, Jorge A; Lucey, Brain C; Stuhlfaut, Joshua W; Varghese, Jose C

    2005-04-01

    To evaluate the impact of a picture archiving and communication systems (PACS)-based software package on the requests for 3D reconstructions of multidetector CT (MDCT) data sets in the emergency radiology of a level 1 trauma center, we reviewed the number and type of physician requests for 3D reconstructions of MDCT data sets for patients admitted after sustaining multiple trauma, during a 12-month period (January 2003-December 2003). During the first 5 months of the study, 3D reconstructions were performed in dedicated workstations located separately from the emergency radiology CT interpretation area. During the last 7 months of the study, reconstructions were performed online by the attending radiologist or resident on duty, using a software package directly incorporated into the PACS workstations. The mean monthly number of 3D reconstructions requested during the two time periods was compared using Student's t test. The monthly mean +/- SD of 3D reconstructions performed before and after 3D software incorporation into the PACS was 34+/-7 (95% CI, 10-58) and 132+/-31 (95% CI, 111-153), respectively. This difference was statistically significant (p<0.0001). In the multiple trauma patient, implementation of PACS-integrated software increases utilization of 3D reconstructions of MDCT data sets. PMID:16028324

  4. Onboard utilization of ground control points for image correction. Volume 3: Ground control point simulation software design

    NASA Technical Reports Server (NTRS)

    1981-01-01

    The software developed to simulate the ground control point navigation system is described. The Ground Control Point Simulation Program (GCPSIM) is designed as an analysis tool to predict the performance of the navigation system. The system consists of two star trackers, a global positioning system receiver, a gyro package, and a landmark tracker.

  5. Osteolytica: An automated image analysis software package that rapidly measures cancer-induced osteolytic lesions in in vivo models with greater reproducibility compared to other commonly used methods☆

    PubMed Central

    Evans, H.R.; Karmakharm, T.; Lawson, M.A.; Walker, R.E.; Harris, W.; Fellows, C.; Huggins, I.D.; Richmond, P.; Chantry, A.D.

    2016-01-01

    Methods currently used to analyse osteolytic lesions caused by malignancies such as multiple myeloma and metastatic breast cancer vary from basic 2-D X-ray analysis to 2-D images of micro-CT datasets analysed with non-specialised image software such as ImageJ. However, these methods have significant limitations. They do not capture 3-D data, they are time-consuming and they often suffer from inter-user variability. We therefore sought to develop a rapid and reproducible method to analyse 3-D osteolytic lesions in mice with cancer-induced bone disease. To this end, we have developed Osteolytica, an image analysis software method featuring an easy to use, step-by-step interface to measure lytic bone lesions. Osteolytica utilises novel graphics card acceleration (parallel computing) and 3-D rendering to provide rapid reconstruction and analysis of osteolytic lesions. To evaluate the use of Osteolytica we analysed tibial micro-CT datasets from murine models of cancer-induced bone disease and compared the results to those obtained using a standard ImageJ analysis method. Firstly, to assess inter-user variability we deployed four independent researchers to analyse tibial datasets from the U266-NSG murine model of myeloma. Using ImageJ, inter-user variability between the bones was substantial (± 19.6%), in contrast to using Osteolytica, which demonstrated minimal variability (± 0.5%). Secondly, tibial datasets from U266-bearing NSG mice or BALB/c mice injected with the metastatic breast cancer cell line 4T1 were compared to tibial datasets from aged and sex-matched non-tumour control mice. Analyses by both Osteolytica and ImageJ showed significant increases in bone lesion area in tumour-bearing mice compared to control mice. These results confirm that Osteolytica performs as well as the current 2-D ImageJ osteolytic lesion analysis method. However, Osteolytica is advantageous in that it analyses over the entirety of the bone volume (as opposed to selected 2-D images

  6. Osteolytica: An automated image analysis software package that rapidly measures cancer-induced osteolytic lesions in in vivo models with greater reproducibility compared to other commonly used methods.

    PubMed

    Evans, H R; Karmakharm, T; Lawson, M A; Walker, R E; Harris, W; Fellows, C; Huggins, I D; Richmond, P; Chantry, A D

    2016-02-01

    Methods currently used to analyse osteolytic lesions caused by malignancies such as multiple myeloma and metastatic breast cancer vary from basic 2-D X-ray analysis to 2-D images of micro-CT datasets analysed with non-specialised image software such as ImageJ. However, these methods have significant limitations. They do not capture 3-D data, they are time-consuming and they often suffer from inter-user variability. We therefore sought to develop a rapid and reproducible method to analyse 3-D osteolytic lesions in mice with cancer-induced bone disease. To this end, we have developed Osteolytica, an image analysis software method featuring an easy to use, step-by-step interface to measure lytic bone lesions. Osteolytica utilises novel graphics card acceleration (parallel computing) and 3-D rendering to provide rapid reconstruction and analysis of osteolytic lesions. To evaluate the use of Osteolytica we analysed tibial micro-CT datasets from murine models of cancer-induced bone disease and compared the results to those obtained using a standard ImageJ analysis method. Firstly, to assess inter-user variability we deployed four independent researchers to analyse tibial datasets from the U266-NSG murine model of myeloma. Using ImageJ, inter-user variability between the bones was substantial (±19.6%), in contrast to using Osteolytica, which demonstrated minimal variability (±0.5%). Secondly, tibial datasets from U266-bearing NSG mice or BALB/c mice injected with the metastatic breast cancer cell line 4T1 were compared to tibial datasets from aged and sex-matched non-tumour control mice. Analyses by both Osteolytica and ImageJ showed significant increases in bone lesion area in tumour-bearing mice compared to control mice. These results confirm that Osteolytica performs as well as the current 2-D ImageJ osteolytic lesion analysis method. However, Osteolytica is advantageous in that it analyses over the entirety of the bone volume (as opposed to selected 2-D images), it

  7. Global Workspace Dynamics: Cortical “Binding and Propagation” Enables Conscious Contents

    PubMed Central

    Baars, Bernard J.; Franklin, Stan; Ramsoy, Thomas Zoega

    2013-01-01

    A global workspace (GW) is a functional hub of binding and propagation in a population of loosely coupled signaling elements. In computational applications, GW architectures recruit many distributed, specialized agents to cooperate in resolving focal ambiguities. In the brain, conscious experiences may reflect a GW function. For animals, the natural world is full of unpredictable dangers and opportunities, suggesting a general adaptive pressure for brains to resolve focal ambiguities quickly and accurately. GW theory aims to understand the differences between conscious and unconscious brain events. In humans and related species the cortico-thalamic (C-T) core is believed to underlie conscious aspects of perception, thinking, learning, feelings of knowing (FOK), felt emotions, visual imagery, working memory, and executive control. Alternative theoretical perspectives are also discussed. The C-T core has many anatomical hubs, but conscious percepts are unitary and internally consistent at any given moment. Over time, conscious contents constitute a very large, open set. This suggests that a brain-based GW capacity cannot be localized in a single anatomical hub. Rather, it should be sought in a functional hub – a dynamic capacity for binding and propagation of neural signals over multiple task-related networks, a kind of neuronal cloud computing. In this view, conscious contents can arise in any region of the C-T core when multiple input streams settle on a winner-take-all equilibrium. The resulting conscious gestalt may ignite an any-to-many broadcast, lasting ∼100–200 ms, and trigger widespread adaptation in previously established networks. To account for the great range of conscious contents over time, the theory suggests an open repertoire of binding1 coalitions that can broadcast via theta/gamma or alpha/gamma phase coupling, like radio channels competing for a narrow frequency band. Conscious moments are thought to hold only 1–4 unrelated items; this

  8. TIA Software User's Manual

    NASA Technical Reports Server (NTRS)

    Cramer, K. Elliott; Syed, Hazari I.

    1995-01-01

    This user's manual describes the installation and operation of TIA, the Thermal-Imaging acquisition and processing Application, developed by the Nondestructive Evaluation Sciences Branch at NASA Langley Research Center, Hampton, Virginia. TIA is a user friendly graphical interface application for the Macintosh 2 and higher series computers. The software has been developed to interface with the Perceptics/Westinghouse Pixelpipe(TM) and PixelStore(TM) NuBus cards and the GW Instruments MacADIOS(TM) input-output (I/O) card for the Macintosh for imaging thermal data. The software is also capable of performing generic image-processing functions.

  9. ChiMS: Open-source instrument control software platform on LabVIEW for imaging/depth profiling mass spectrometers

    PubMed Central

    Cui, Yang; Hanley, Luke

    2015-01-01

    ChiMS is an open-source data acquisition and control software program written within LabVIEW for high speed imaging and depth profiling mass spectrometers. ChiMS can also transfer large datasets from a digitizer to computer memory at high repetition rate, save data to hard disk at high throughput, and perform high speed data processing. The data acquisition mode generally simulates a digital oscilloscope, but with peripheral devices integrated for control as well as advanced data sorting and processing capabilities. Customized user-designed experiments can be easily written based on several included templates. ChiMS is additionally well suited to non-laser based mass spectrometers imaging and various other experiments in laser physics, physical chemistry, and surface science. PMID:26133872

  10. ChiMS: Open-source instrument control software platform on LabVIEW for imaging/depth profiling mass spectrometers

    NASA Astrophysics Data System (ADS)

    Cui, Yang; Hanley, Luke

    2015-06-01

    ChiMS is an open-source data acquisition and control software program written within LabVIEW for high speed imaging and depth profiling mass spectrometers. ChiMS can also transfer large datasets from a digitizer to computer memory at high repetition rate, save data to hard disk at high throughput, and perform high speed data processing. The data acquisition mode generally simulates a digital oscilloscope, but with peripheral devices integrated for control as well as advanced data sorting and processing capabilities. Customized user-designed experiments can be easily written based on several included templates. ChiMS is additionally well suited to non-laser based mass spectrometers imaging and various other experiments in laser physics, physical chemistry, and surface science.

  11. ChiMS: Open-source instrument control software platform on LabVIEW for imaging/depth profiling mass spectrometers.

    PubMed

    Cui, Yang; Hanley, Luke

    2015-06-01

    ChiMS is an open-source data acquisition and control software program written within LabVIEW for high speed imaging and depth profiling mass spectrometers. ChiMS can also transfer large datasets from a digitizer to computer memory at high repetition rate, save data to hard disk at high throughput, and perform high speed data processing. The data acquisition mode generally simulates a digital oscilloscope, but with peripheral devices integrated for control as well as advanced data sorting and processing capabilities. Customized user-designed experiments can be easily written based on several included templates. ChiMS is additionally well suited to non-laser based mass spectrometers imaging and various other experiments in laser physics, physical chemistry, and surface science. PMID:26133872

  12. Powerplant software

    SciTech Connect

    Elliott, T.C.

    1995-07-01

    Powerplants need software to thrive and compete. Covered here are many programs and applications -- an overview of the functions, tasks, and problem-solving software is used for today. Software or, more accurately, software-driven systems are pervasive. Their presence is felt in every nook and cranny of the powerplant -- from design and construction through operation and maintenance, even dismantling and decommissioning -- embracing whole systems but also focusing on individual pieces of equipment. No one software supplier or two or three dominates -- powerplant software is the purview of scores if not hundreds of suppliers ranging from the largest corporations to individual consultants and application developers.

  13. A New Measurement Technique of the Characteristics of Nutrient Artery Canals in Tibias Using Materialise's Interactive Medical Image Control System Software.

    PubMed

    Li, Jiantao; Zhang, Hao; Yin, Peng; Su, Xiuyun; Zhao, Zhe; Zhou, Jianfeng; Li, Chen; Li, Zhirui; Zhang, Lihai; Tang, Peifu

    2015-01-01

    We established a novel measurement technique to evaluate the anatomic information of nutrient artery canals using Mimics (Materialise's Interactive Medical Image Control System) software, which will provide full knowledge of nutrient artery canals to assist in the diagnosis of longitudinal fractures of tibia and choosing an optimal therapy. Here we collected Digital Imaging and Communications in Medicine (DICOM) format of 199 patients hospitalized in our hospital. All three-dimensional models of tibia in Mimics were reconstructed. In 3-matic software, we marked five points in tibia which located at intercondylar eminence, tibia tuberosity, outer ostium, inner ostium, and bottom of medial malleolus. We then recorded Z-coordinates values of the five points and performed statistical analysis. Our results indicate that foramen was found to be absent in 9 (2.3%) tibias, and 379 (95.2%) tibias had single nutrient foramen. The double foramina was observed in 10 (2.5%) tibias. The mean of tibia length was 358 ± 22 mm. The mean foraminal index was 31.8%  ± 3%. The mean distance between tibial tuberosity and foramen (TFD) is 66 ± 12 mm. Foraminal index has significant positive correlation with TFD (r = 0.721, P < 0.01). Length of nutrient artery canals has significant negative correlation with TFD (r = -0.340, P < 0.01) and has significant negative correlation with foraminal index (r = -0.541, P < 0.01). PMID:26788498

  14. A New Measurement Technique of the Characteristics of Nutrient Artery Canals in Tibias Using Materialise's Interactive Medical Image Control System Software

    PubMed Central

    Li, Jiantao; Zhang, Hao; Yin, Peng; Su, Xiuyun; Zhao, Zhe; Zhou, Jianfeng; Li, Chen; Li, Zhirui; Zhang, Lihai; Tang, Peifu

    2015-01-01

    We established a novel measurement technique to evaluate the anatomic information of nutrient artery canals using Mimics (Materialise's Interactive Medical Image Control System) software, which will provide full knowledge of nutrient artery canals to assist in the diagnosis of longitudinal fractures of tibia and choosing an optimal therapy. Here we collected Digital Imaging and Communications in Medicine (DICOM) format of 199 patients hospitalized in our hospital. All three-dimensional models of tibia in Mimics were reconstructed. In 3-matic software, we marked five points in tibia which located at intercondylar eminence, tibia tuberosity, outer ostium, inner ostium, and bottom of medial malleolus. We then recorded Z-coordinates values of the five points and performed statistical analysis. Our results indicate that foramen was found to be absent in 9 (2.3%) tibias, and 379 (95.2%) tibias had single nutrient foramen. The double foramina was observed in 10 (2.5%) tibias. The mean of tibia length was 358 ± 22 mm. The mean foraminal index was 31.8%  ± 3%. The mean distance between tibial tuberosity and foramen (TFD) is 66 ± 12 mm. Foraminal index has significant positive correlation with TFD (r = 0.721, P < 0.01). Length of nutrient artery canals has significant negative correlation with TFD (r = −0.340, P < 0.01) and has significant negative correlation with foraminal index (r = −0.541, P < 0.01). PMID:26788498

  15. Comparison of performance of object-based image analysis techniques available in open source software (Spring and Orfeo Toolbox/Monteverdi) considering very high spatial resolution data

    NASA Astrophysics Data System (ADS)

    Teodoro, Ana C.; Araujo, Ricardo

    2016-01-01

    The use of unmanned aerial vehicles (UAVs) for remote sensing applications is becoming more frequent. However, this type of information can result in several software problems related to the huge amount of data available. Object-based image analysis (OBIA) has proven to be superior to pixel-based analysis for very high-resolution images. The main objective of this work was to explore the potentialities of the OBIA methods available in two different open source software applications, Spring and OTB/Monteverdi, in order to generate an urban land cover map. An orthomosaic derived from UAVs was considered, 10 different regions of interest were selected, and two different approaches were followed. The first one (Spring) uses the region growing segmentation algorithm followed by the Bhattacharya classifier. The second approach (OTB/Monteverdi) uses the mean shift segmentation algorithm followed by the support vector machine (SVM) classifier. Two strategies were followed: four classes were considered using Spring and thereafter seven classes were considered for OTB/Monteverdi. The SVM classifier produces slightly better results and presents a shorter processing time. However, the poor spectral resolution of the data (only RGB bands) is an important factor that limits the performance of the classifiers applied.

  16. Software Program: Software Management Guidebook

    NASA Technical Reports Server (NTRS)

    1996-01-01

    The purpose of this NASA Software Management Guidebook is twofold. First, this document defines the core products and activities required of NASA software projects. It defines life-cycle models and activity-related methods but acknowledges that no single life-cycle model is appropriate for all NASA software projects. It also acknowledges that the appropriate method for accomplishing a required activity depends on characteristics of the software project. Second, this guidebook provides specific guidance to software project managers and team leaders in selecting appropriate life cycles and methods to develop a tailored plan for a software engineering project.

  17. Proprietary software

    NASA Technical Reports Server (NTRS)

    Marnock, M. J.

    1971-01-01

    The protection of intellectual property by a patent, a copyright, or trade secrets is reviewed. The present and future use of computers and software are discussed, along with the governmental uses of software. The popularity of contractual agreements for sale or lease of computer programs and software services is also summarized.

  18. Workspace design for crane cabins applying a combined traditional approach and the Taguchi method for design of experiments.

    PubMed

    Spasojević Brkić, Vesna K; Veljković, Zorica A; Golubović, Tamara; Brkić, Aleksandar Dj; Kosić Šotić, Ivana

    2016-01-01

    Procedures in the development process of crane cabins are arbitrary and subjective. Since approximately 42% of incidents in the construction industry are linked to them, there is a need to collect fresh anthropometric data and provide additional recommendations for design. In this paper, dimensioning of the crane cabin interior space was carried out using a sample of 64 crane operators' anthropometric measurements, in the Republic of Serbia, by measuring workspace with 10 parameters using nine measured anthropometric data from each crane operator. This paper applies experiments run via full factorial designs using a combined traditional and Taguchi approach. The experiments indicated which design parameters are influenced by which anthropometric measurements and to what degree. The results are expected to be of use for crane cabin designers and should assist them to design a cabin that may lead to less strenuous sitting postures and fatigue for operators, thus improving safety and accident prevention. PMID:26652099

  19. Kinematic Analysis and Synthesis of a 3-URU Pure Rotational Parallel Mechanism with Respect to Singularity and Workspace

    NASA Astrophysics Data System (ADS)

    Huda, Syamsul; Takeda, Yukio

    This paper concerns kinematics and dimensional synthesis of a three universal-revolute-universal (3-URU) pure rotational parallel mechanism. The mechanism is composed of a base, a platform and three symmetric limbs consisting of U-R-U joints. This mechanism is a spatial non-overconstrained mechanism with three degrees of freedom. The joints in each limb are so arranged to perform pure rotational motion of the platform around a specific point. Equations for inverse displacement analysis and singularities were derived to investigate the relationship of the kinematic constants to the solution of the inverse kinematics and singularities. Based on the results, a dimensional synthesis procedure for the 3-URU parallel mechanism considering singularities and the workspace was proposed. A numerical example was also presented to illustrate the synthesis method.

  20. Designing, Supporting, and Sustaining an Online Community of Practice: NASA EPO Workspace as an Ongoing Exploration of the Value of Community

    NASA Astrophysics Data System (ADS)

    Davey, B.; Davis, H. B.

    2015-12-01

    Increasingly, geographically diverse organizations, like NASA's Science Mission Directorate Education and Public Outreach personnel (SMD EPO), are looking for ways to facilitate group interactions in meaningful ways while limiting costs. Towards this end, of particular interest, and showing great potential are communities of practice. Communities of practice represent relationships in real-time between and among people sharing a common practice. They facilitate the sharing of information, building collective knowledge, and growing of the principles of practice. In 2010-11, SMD EPO established a website to support EPO professionals, facilitate headquarters reporting, and foster a community of practice. The purpose of this evaluation is to examine the design and use of the workspace and the value created for both individual community members and SMD EPO, the sponsoring organization. The online workspace was launched in 2010-11 for the members of NASA's SMDEPO community. The online workspace was designed to help facilitate the efficient sharing of information, be a central repository for resources, help facilitate and support knowledge creation, and ultimately lead to the development of an online community of practice. This study examines the role of the online workspace component of a community in the work of a community of practice. Much has been studied revealing the importance of communities of practice to organizations, project success, and knowledge management and some of these same successes hold true for virtual communities of practice. Additionally, we look at the outcomes of housting the online community for these past years in respect to knowledge building and personal and organizational value, the affects on professional dvelopment opportunities, how community members have benefited, and how the workspace has evolved to better serve the community.

  1. The SPoCA-suite: Software for extraction, characterization, and tracking of active regions and coronal holes on EUV images

    NASA Astrophysics Data System (ADS)

    Verbeeck, C.; Delouille, V.; Mampaey, B.; De Visscher, R.

    2014-01-01

    Context. Precise localization and characterization of active regions (AR) and coronal holes (CH) as observed by extreme ultra violet (EUV) imagers are crucial for a wide range of solar and helio-physics studies. Aims: We introduce a set of segmentation procedures (known as the SPoCA-suite) that allows one to retrieve AR and CH properties on EUV images taken from SOHO-EIT, STEREO-EUVI, PROBA2-SWAP, and SDO-AIA. Methods: We build upon our previous work on the Spatial Possibilistic Clustering Algorithm (SPoCA), that we have improved substantially in several ways. Results: We apply our algorithm on the synoptic EIT archive from 1997 to 2011 and decompose this dataset into regions that can clearly be identified as AR, quiet Sun, and CH. An antiphase between AR and CH filling factor is observed, as expected. The SPoCA-suite is next applied to datasets from EUVI, SWAP, and AIA. The time series pertaining to ARs or CHs are presented. Conclusions: The SPoCA-suite enables the extraction of several long time series of AR and CH properties from the data files of EUV imagers and also allows tracking individual ARs or CHs over time. For AIA images, AR and CH catalogs are available in near-real time from the Heliophysics Events Knowledgebase. The full code, which allows processing any EUV images, is available upon request to the authors.

  2. Computer software.

    PubMed

    Rosenthal, L E

    1986-10-01

    Software is the component in a computer system that permits the hardware to perform the various functions that a computer system is capable of doing. The history of software and its development can be traced to the early nineteenth century. All computer systems are designed to utilize the "stored program concept" as first developed by Charles Babbage in the 1850s. The concept was lost until the mid-1940s, when modern computers made their appearance. Today, because of the complex and myriad tasks that a computer system can perform, there has been a differentiation of types of software. There is software designed to perform specific business applications. There is software that controls the overall operation of a computer system. And there is software that is designed to carry out specialized tasks. Regardless of types, software is the most critical component of any computer system. Without it, all one has is a collection of circuits, transistors, and silicone chips. PMID:3536223

  3. Towards a software profession

    NASA Technical Reports Server (NTRS)

    Berard, Edward V.

    1986-01-01

    An increasing number of programmers have attempted to change their image. They have made it plain that they wish not only to be taken seriously, but they also wish to be regarded as professionals. Many programmers now wish to referred to as software engineers. If programmers wish to be considered professionals in every sense of the word, two obstacles must be overcome: the inability to think of software as a product, and the idea that little or no skill is required to create and handle software throughout its life cycle. The steps to be taken toward professionalization are outlined along with recommendations.

  4. Revealing text in a complexly rolled silver scroll from Jerash with computed tomography and advanced imaging software

    NASA Astrophysics Data System (ADS)

    Hoffmann Barfod, Gry; Larsen, John Møller; Lichtenberger, Achim; Raja, Rubina

    2015-12-01

    Throughout Antiquity magical amulets written on papyri, lead and silver were used for apotropaic reasons. While papyri often can be unrolled and deciphered, metal scrolls, usually very thin and tightly rolled up, cannot easily be unrolled without damaging the metal. This leaves us with unreadable results due to the damage done or with the decision not to unroll the scroll. The texts vary greatly and tell us about the cultural environment and local as well as individual practices at a variety of locations across the Mediterranean. Here we present the methodology and the results of the digital unfolding of a silver sheet from Jerash in Jordan from the mid-8th century CE. The scroll was inscribed with 17 lines in presumed pseudo-Arabic as well as some magical signs. The successful unfolding shows that it is possible to digitally unfold complexly folded scrolls, but that it requires a combination of the know-how of the software and linguistic knowledge.

  5. A Guide to Selecting Large Print/Enhanced Image Computer Access Hardware/Software for Persons with Low Vision.

    ERIC Educational Resources Information Center

    DeWitt, J. C.; And Others

    1988-01-01

    Large print/enhanced image computer access products for low-vision individuals are reviewed. The products ("DP-11/Plus,""Large Print DOS,""Lyon Large Print,""PC Lens,""Vista/Vista 2," and "ZoomText") work interactively with most application programs and DOS in IBM environments. Features evaluated include: documentation, set-up, hardware/software…

  6. IDATEN and G-SITENNO: GUI-assisted software for coherent X-ray diffraction imaging experiments and data analyses at SACLA.

    PubMed

    Sekiguchi, Yuki; Yamamoto, Masaki; Oroguchi, Tomotaka; Takayama, Yuki; Suzuki, Shigeyuki; Nakasako, Masayoshi

    2014-11-01

    Using our custom-made diffraction apparatus KOTOBUKI-1 and two multiport CCD detectors, cryogenic coherent X-ray diffraction imaging experiments have been undertaken at the SPring-8 Angstrom Compact free electron LAser (SACLA) facility. To efficiently perform experiments and data processing, two software suites with user-friendly graphical user interfaces have been developed. The first is a program suite named IDATEN, which was developed to easily conduct four procedures during experiments: aligning KOTOBUKI-1, loading a flash-cooled sample into the cryogenic goniometer stage inside the vacuum chamber of KOTOBUKI-1, adjusting the sample position with respect to the X-ray beam using a pair of telescopes, and collecting diffraction data by raster scanning the sample with X-ray pulses. Named G-SITENNO, the other suite is an automated version of the original SITENNO suite, which was designed for processing diffraction data. These user-friendly software suites are now indispensable for collecting a large number of diffraction patterns and for processing the diffraction patterns immediately after collecting data within a limited beam time. PMID:25343809

  7. Software safety

    NASA Technical Reports Server (NTRS)

    Leveson, Nancy

    1987-01-01

    Software safety and its relationship to other qualities are discussed. It is shown that standard reliability and fault tolerance techniques will not solve the safety problem for the present. A new attitude requires: looking at what you do NOT want software to do along with what you want it to do; and assuming things will go wrong. New procedures and changes to entire software development process are necessary: special software safety analysis techniques are needed; and design techniques, especially eliminating complexity, can be very helpful.

  8. Revealing text in a complexly rolled silver scroll from Jerash with computed tomography and advanced imaging software.

    PubMed

    Hoffmann Barfod, Gry; Larsen, John Møller; Lichtenberger, Achim; Raja, Rubina

    2015-01-01

    Throughout Antiquity magical amulets written on papyri, lead and silver were used for apotropaic reasons. While papyri often can be unrolled and deciphered, metal scrolls, usually very thin and tightly rolled up, cannot easily be unrolled without damaging the metal. This leaves us with unreadable results due to the damage done or with the decision not to unroll the scroll. The texts vary greatly and tell us about the cultural environment and local as well as individual practices at a variety of locations across the Mediterranean. Here we present the methodology and the results of the digital unfolding of a silver sheet from Jerash in Jordan from the mid-8(th) century CE. The scroll was inscribed with 17 lines in presumed pseudo-Arabic as well as some magical signs. The successful unfolding shows that it is possible to digitally unfold complexly folded scrolls, but that it requires a combination of the know-how of the software and linguistic knowledge. PMID:26648504

  9. Revealing text in a complexly rolled silver scroll from Jerash with computed tomography and advanced imaging software

    PubMed Central

    Hoffmann Barfod, Gry; Larsen, John Møller; Raja, Rubina

    2015-01-01

    Throughout Antiquity magical amulets written on papyri, lead and silver were used for apotropaic reasons. While papyri often can be unrolled and deciphered, metal scrolls, usually very thin and tightly rolled up, cannot easily be unrolled without damaging the metal. This leaves us with unreadable results due to the damage done or with the decision not to unroll the scroll. The texts vary greatly and tell us about the cultural environment and local as well as individual practices at a variety of locations across the Mediterranean. Here we present the methodology and the results of the digital unfolding of a silver sheet from Jerash in Jordan from the mid-8th century CE. The scroll was inscribed with 17 lines in presumed pseudo-Arabic as well as some magical signs. The successful unfolding shows that it is possible to digitally unfold complexly folded scrolls, but that it requires a combination of the know-how of the software and linguistic knowledge. PMID:26648504

  10. Application of Technical Measures and Software in Constructing Photorealistic 3D Models of Historical Building Using Ground-Based and Aerial (UAV) Digital Images

    NASA Astrophysics Data System (ADS)

    Zarnowski, Aleksander; Banaszek, Anna; Banaszek, Sebastian

    2015-12-01

    Preparing digital documentation of historical buildings is a form of protecting cultural heritage. Recently there have been several intensive studies using non-metric digital images to construct realistic 3D models of historical buildings. Increasingly often, non-metric digital images are obtained with unmanned aerial vehicles (UAV). Technologies and methods of UAV flights are quite different from traditional photogrammetric approaches. The lack of technical guidelines for using drones inhibits the process of implementing new methods of data acquisition. This paper presents the results of experiments in the use of digital images in the construction of photo-realistic 3D model of a historical building (Raphaelsohns' Sawmill in Olsztyn). The aim of the study at the first stage was to determine the meteorological and technical conditions for the acquisition of aerial and ground-based photographs. At the next stage, the technology of 3D modelling was developed using only ground-based or only aerial non-metric digital images. At the last stage of the study, an experiment was conducted to assess the possibility of 3D modelling with the comprehensive use of aerial (UAV) and ground-based digital photographs in terms of their labour intensity and precision of development. Data integration and automatic photo-realistic 3D construction of the models was done with Pix4Dmapper and Agisoft PhotoScan software Analyses have shown that when certain parameters established in an experiment are kept, the process of developing the stock-taking documentation for a historical building moves from the standards of analogue to digital technology with considerably reduced cost.

  11. Diagnostic use of facial image analysis software in endocrine and genetic disorders: review, current results and future perspectives.

    PubMed

    Kosilek, R P; Frohner, R; Würtz, R P; Berr, C M; Schopohl, J; Reincke, M; Schneider, H J

    2015-10-01

    Cushing's syndrome (CS) and acromegaly are endocrine diseases that are currently diagnosed with a delay of several years from disease onset. Novel diagnostic approaches and increased awareness among physicians are needed. Face classification technology has recently been introduced as a promising diagnostic tool for CS and acromegaly in pilot studies. It has also been used to classify various genetic syndromes using regular facial photographs. The authors provide a basic explanation of the technology, review available literature regarding its use in a medical setting, and discuss possible future developments. The method the authors have employed in previous studies uses standardized frontal and profile facial photographs for classification. Image analysis is based on applying mathematical functions evaluating geometry and image texture to a grid of nodes semi-automatically placed on relevant facial structures, yielding a binary classification result. Ongoing research focuses on improving diagnostic algorithms of this method and bringing it closer to clinical use. Regarding future perspectives, the authors propose an online interface that facilitates submission of patient data for analysis and retrieval of results as a possible model for clinical application. PMID:26162404

  12. Software Reviews.

    ERIC Educational Resources Information Center

    Bitter, Gary G., Ed.

    1990-01-01

    Reviews three computer software: (1) "Elastic Lines: The Electronic Geoboard" on elementary geometry; (2) "Wildlife Adventures: Whales" on environmental science; and (3) "What Do You Do with a Broken Calculator?" on computation and problem solving. Summarizes the descriptions, strengths and weaknesses, and applications of each software. (YP)

  13. Software Repository

    NASA Technical Reports Server (NTRS)

    Merwarth, P., D.

    1983-01-01

    The Common Software Module Repository (CSMR) is computerized library system with high product and service visibility to potential users. Online capabilities of system allow both librarian and user to interact with library. Librarian is responsible for maintaining information in CSMR library. User searches library to locate software modules that meet his or her current needs.

  14. Software Reviews.

    ERIC Educational Resources Information Center

    Miller, Anne, Ed.; Radziemski, Cathy, Ed.

    1988-01-01

    Reviews two software packages for the Macintosh series. "Course Builder 2.0," a courseware authoring system, allows the user to create programs which stand alone and may be used independently in the classroom. "World Builder," an artificial intelligence software package, allows creative thinking, problem-solving, and decision-making. (YP)

  15. Software Bridge

    NASA Technical Reports Server (NTRS)

    1995-01-01

    I-Bridge is a commercial version of software developed by I-Kinetics under a NASA Small Business Innovation Research (SBIR) contract. The software allows users of Windows applications to gain quick, easy access to databases, programs and files on UNIX services. Information goes directly onto spreadsheets and other applications; users need not manually locate, transfer and convert data.

  16. Software Reviews.

    ERIC Educational Resources Information Center

    Wulfson, Stephen

    1988-01-01

    Presents reviews of six computer software programs for teaching science. Provides the publisher, grade level, cost, and descriptions of software, including: (1) "Recycling Logic"; (2) "Introduction to Biochemistry"; (3) "Food for Thought"; (4) "Watts in a Home"; (5) "Geology in Action"; and (6) "Biomes." All are for Apple series microcomputers.…

  17. The National Alliance for Medical Image Computing, a roadmap initiative to build a free and open source software infrastructure for translational research in medical image analysis.

    PubMed

    Kapur, Tina; Pieper, Steve; Whitaker, Ross; Aylward, Stephen; Jakab, Marianna; Schroeder, Will; Kikinis, Ron

    2012-01-01

    The National Alliance for Medical Image Computing (NA-MIC), is a multi-institutional, interdisciplinary community of researchers, who share the recognition that modern health care demands improved technologies to ease suffering and prolong productive life. Organized under the National Centers for Biomedical Computing 7 years ago, the mission of NA-MIC is to implement a robust and flexible open-source infrastructure for developing and applying advanced imaging technologies across a range of important biomedical research disciplines. A measure of its success, NA-MIC is now applying this technology to diseases that have immense impact on the duration and quality of life: cancer, heart disease, trauma, and degenerative genetic diseases. The targets of this technology range from group comparisons to subject-specific analysis. PMID:22081219

  18. SAR Product Control Software

    NASA Astrophysics Data System (ADS)

    Meadows, P. J.; Hounam, D.; Rye, A. J.; Rosich, B.; Börner, T.; Closa, J.; Schättler, B.; Smith, P. J.; Zink, M.

    2003-03-01

    As SAR instruments and their operating modes become more complex, as new applications place more and more demands on image quality and as our understanding of their imperfections becomes more sophisticated, there is increasing recognition that SAR data quality has to be controlled more completely to keep pace. The SAR product CONtrol software (SARCON) is a comprehensive SAR product control software suite tailored to the latest generation of SAR sensors. SARCON profits from the most up-to-date thinking on SAR image performance derived from other spaceborne and airborne SAR projects and is based on the newest applications. This paper gives an overview of the structure and the features of this new software tool, which is a product of a co-operation between teams at BAE SYSTEMS Advanced Technology Centre and DLR under contract to ESA (ESRIN). Work on SARCON began in 1999 and is continuing.

  19. Determination of Flow Rates in Capillary Liquid Chromatography Coupled to a Nanoelectrospray Source using Droplet Image Analysis Software.

    PubMed

    Cohen, Alejandro M; Soto, Axel J; Fawcett, James P

    2016-08-01

    Liquid chromatography coupled to electrospray tandem mass spectrometry (LC-ESI-MS/MS) is widely used in proteomic and metabolomic workflows. Considerable analytical improvements have been observed when the components of LC systems are scaled down. Currently, nano-ESI is typically done at capillary LC flow rates ranging from 200 to 300 nL/min. At these flow rates, trouble shooting and leak detection of LC systems has become increasingly challenging. In this paper we present a novel proof-of-concept approach to measure flow rates at the tip of electrospray emitters when the ionization voltage is turned off. This was achieved by estimating the changes in the droplet volume over time using digital image analysis. The results are comparable with the traditional methods of measuring flow rates, with the potential advantages of being fully automatable and nondisruptive. PMID:27351615

  20. Strategy for the lowering and the assessment of exposure to nanoparticles at workspace - Case of study concerning the potential emission of nanoparticles of Lead in an epitaxy laboratory

    NASA Astrophysics Data System (ADS)

    Artous, Sébastien; Zimmermann, Eric; Douissard, Paul-Antoine; Locatelli, Dominique; Motellier, Sylvie; Derrough, Samir

    2015-05-01

    The implementation in many products of manufactured nanoparticles is growing fast and raises new questions. For this purpose, the CEA - NanoSafety Platform is developing various research topics for health and safety, environment and nanoparticles exposure in professional activities. The containment optimisation for the exposition lowering, then the exposure assessment to nanoparticles is a strategy for safety improvement at workplace and workspace. The lowering step consists in an optimisation of dynamic and static containment at workplace and/or workspace. Generally, the exposure risk due to the presence of nanoparticles substances does not allow modifying the parameters of containment at workplace and/or workspace. Therefore, gaseous or nanoparticulate tracers are used to evaluate performances of containment. Using a tracer allows to modify safely the parameters of the dynamic containment (ventilation, flow, speed) and to study several configurations of static containment. Moreover, a tracer allows simulating accidental or incidental situation. As a result, a safety procedure can be written more easily in order to manage this type of situation. The step of measurement and characterization of aerosols can therefore be used to assess the exposition at workplace and workspace. The case of study, aim of this paper, concerns the potential emission of Lead nanoparticles at the exhaust of a furnace in an epitaxy laboratory. The use of Helium tracer to evaluate the performance of containment is firstly studied. Secondly, the exposure assessment is characterised in accordance with the French guide “recommendations for characterizing potential emissions and exposure to aerosols released from nanomaterials in workplace operations”. Thirdly the aerosols are sampled, on several places, using collection membranes to try to detect traces of Lead in air.

  1. Remote Viewer for Maritime Robotics Software

    NASA Technical Reports Server (NTRS)

    Kuwata, Yoshiaki; Wolf, Michael; Huntsberger, Terrance L.; Howard, Andrew B.

    2013-01-01

    This software is a viewer program for maritime robotics software that provides a 3D visualization of the boat pose, its position history, ENC (Electrical Nautical Chart) information, camera images, map overlay, and detected tracks.

  2. Software Smarts

    NASA Technical Reports Server (NTRS)

    1998-01-01

    Under an SBIR (Small Business Innovative Research) contract with Johnson Space Center, Knowledge Based Systems Inc. (KBSI) developed an intelligent software environment for modeling and analyzing mission planning activities, simulating behavior, and, using a unique constraint propagation mechanism, updating plans with each change in mission planning activities. KBSI developed this technology into a commercial product, PROJECTLINK, a two-way bridge between PROSIm, KBSI's process modeling and simulation software and leading project management software like Microsoft Project and Primavera's SureTrak Project Manager.

  3. Software testing

    NASA Astrophysics Data System (ADS)

    Price-Whelan, Adrian M.

    2016-01-01

    Now more than ever, scientific results are dependent on sophisticated software and analysis. Why should we trust code written by others? How do you ensure your own code produces sensible results? How do you make sure it continues to do so as you update, modify, and add functionality? Software testing is an integral part of code validation and writing tests should be a requirement for any software project. I will talk about Python-based tools that make managing and running tests much easier and explore some statistics for projects hosted on GitHub that contain tests.

  4. MpUL-multi: Software for Calculation of Amyloid Fibril Mass per Unit Length from TB-TEM Images

    PubMed Central

    Iadanza, Matthew G.; Jackson, Matthew P.; Radford, Sheena E.; Ranson, Neil A.

    2016-01-01

    Structure determination for amyloid fibrils presents many challenges due to the high variability exhibited by fibrils and heterogeneous morphologies present, even in single samples. Mass per unit length (MPL) estimates can be used to differentiate amyloid fibril morphologies and provide orthogonal evidence for helical symmetry parameters determined by other methods. In addition, MPL data can provide insight on the arrangement of subunits in a fibril, especially for more complex fibrils assembled with multiple parallel copies of the asymmetric unit or multiple twisted protofilaments. By detecting only scattered electrons, which serve as a relative measure of total scattering, and therefore protein mass, dark field imaging gives an approximation of the total mass of protein present in any given length of fibril. When compared with a standard of known MPL, such as Tobacco Mosaic Virus (TMV), MPL of the fibrils in question can be determined. The program suite MpUL-multi was written for rapid semi-automated processing of TB-TEM dark field data acquired using this method. A graphical user interface allows for simple designation of fibrils and standards. A second program averages intensities from multiple TMV molecules for accurate standard determination, makes multiple measurements along a given fibril, and calculates the MPL. PMID:26867957

  5. Comparative evaluation of cephalometric measurements of monitor-displayed images by Nemoceph software and its hard copy by manual tracing

    PubMed Central

    Tikku, Tripti; Khanna, Rohit; Maurya, R.P.; Srivastava, Kamna; Bhushan, Rastra

    2014-01-01

    Objective The aim of this study was to evaluate and compare the cephalometric measurements obtained from computerized tracing of direct digital radiographs and hand tracing of their digital radiographic printouts. Material and methods The soft- and hard-copies of pre-treatment lateral cephalograms of 40 subjects (both males and females) within the age group of 10–30 years, irrespective of the type of malocclusion were taken. Total 26 measurements (13 linear and 13 angular) were obtained using both the manual and the digital technique. Results Amongst the linear measurements, Anterior facial height (AFH), Posterior facial height (PFH), Upper lip length (ULL), Lower lip length (LLL), Anterior cranial base length (ACBL), Posterior cranial base length (PCBL), Maxillary length (MxL), Mandibular length (MdL), Lower incisor to NB line (L1 to NB) and Lower lip protrusion (LLP) showed statistically significant difference between the two techniques but were clinically acceptable (difference between the digital and manual technique were less than 2 units (1 unit = 1 mm for linear measurements and 1° for angular measurements). While amongst the angular measurements, only occlusal plane angle showed statistically significant difference between the two techniques that was not clinically acceptable. Conclusion Digital measurements obtained from monitor-displayed images (soft copy) were found to be reproducible and comparable to the manual method done on its hard copy, for all the measurements except occlusal plane angle (SN-occlusal plane). PMID:25737917

  6. CSAM Metrology Software Tool

    NASA Technical Reports Server (NTRS)

    Vu, Duc; Sandor, Michael; Agarwal, Shri

    2005-01-01

    CSAM Metrology Software Tool (CMeST) is a computer program for analysis of false-color CSAM images of plastic-encapsulated microcircuits. (CSAM signifies C-mode scanning acoustic microscopy.) The colors in the images indicate areas of delamination within the plastic packages. Heretofore, the images have been interpreted by human examiners. Hence, interpretations have not been entirely consistent and objective. CMeST processes the color information in image-data files to detect areas of delamination without incurring inconsistencies of subjective judgement. CMeST can be used to create a database of baseline images of packages acquired at given times for comparison with images of the same packages acquired at later times. Any area within an image can be selected for analysis, which can include examination of different delamination types by location. CMeST can also be used to perform statistical analyses of image data. Results of analyses are available in a spreadsheet format for further processing. The results can be exported to any data-base-processing software.

  7. EXSdetect: an end-to-end software for extended source detection in X-ray images: application to Swift-XRT data

    NASA Astrophysics Data System (ADS)

    Liu, T.; Tozzi, P.; Tundo, E.; Moretti, A.; Wang, J.-X.; Rosati, P.; Guglielmetti, F.

    2013-01-01

    Aims: We present a stand-alone software (named EXSdetect) for the detection of extended sources in X-ray images. Our goal is to provide a flexible tool capable of detecting extended sources down to the lowest flux levels attainable within instrumental limitations, while maintaining robust photometry, high completeness, and low contamination, regardless of source morphology. EXSdetect was developed mainly to exploit the ever-increasing wealth of archival X-ray data, but is also ideally suited to explore the scientific capabilities of future X-ray facilities, with a strong focus on investigations of distant groups and clusters of galaxies. Methods: EXSdetect combines a fast Voronoi tessellation code with a friends-of-friends algorithm and an automated deblending procedure. The values of key parameters are matched to fundamental telescope properties such as angular resolution and instrumental background. In addition, the software is designed to permit extensive tests of its performance via simulations of a wide range of observational scenarios. Results: We applied EXSdetect to simulated data fields modeled to realistically represent the Swift X-ray Cluster Survey (SXCS), which is based on archival data obtained by the X-ray telescope onboard the Swift satellite. We achieve more than 90% completeness for extended sources comprising at least 80 photons in the 0.5-2 keV band, a limit that corresponds to 10-14 erg cm-2 s-1 for the deepest SXCS fields. This detection limit is comparable to the one attained by the most sensitive cluster surveys conducted with much larger X-ray telescopes. While evaluating the performance of EXSdetect, we also explored the impact of improved angular resolution and discuss the ideal properties of the next generation of X-ray survey missions. The Phyton code EXSdetect is available on the SXCS website http://adlibitum.oats.inaf.it/sxcs

  8. SU-E-I-63: Quantitative Evaluation of the Effects of Orthopedic Metal Artifact Reduction (OMAR) Software On CT Images for Radiotherapy Simulation

    SciTech Connect

    Jani, S

    2014-06-01

    Purpose: CT simulation for patients with metal implants can often be challenging due to artifacts that obscure tumor/target delineation and normal organ definition. Our objective was to evaluate the effectiveness of Orthopedic Metal Artifact Reduction (OMAR), a commercially available software, in reducing metal-induced artifacts and its effect on computed dose during treatment planning. Methods: CT images of water surrounding metallic cylindrical rods made of aluminum, copper and iron were studied in terms of Hounsfield Units (HU) spread. Metal-induced artifacts were characterized in terms of HU/Volume Histogram (HVH) using the Pinnacle treatment planning system. Effects of OMAR on enhancing our ability to delineate organs on CT and subsequent dose computation were examined in nine (9) patients with hip implants and two (2) patients with breast tissue expanders. Results: Our study characterized water at 1000 HU with a standard deviation (SD) of about 20 HU. The HVHs allowed us to evaluate how the presence of metal changed the HU spread. For example, introducing a 2.54 cm diameter copper rod in water increased the SD in HU of the surrounding water from 20 to 209, representing an increase in artifacts. Subsequent use of OMAR brought the SD down to 78. Aluminum produced least artifacts whereas Iron showed largest amount of artifacts. In general, an increase in kVp and mA during CT scanning showed better effectiveness of OMAR in reducing artifacts. Our dose analysis showed that some isodose contours shifted by several mm with OMAR but infrequently and were nonsignificant in planning process. Computed volumes of various dose levels showed <2% change. Conclusions: In our experience, OMAR software greatly reduced the metal-induced CT artifacts for the majority of patients with implants, thereby improving our ability to delineate tumor and surrounding organs. OMAR had a clinically negligible effect on computed dose within tissues. Partially funded by unrestricted

  9. Software Reviews.

    ERIC Educational Resources Information Center

    Wulfson, Stephen, Ed.

    1990-01-01

    Reviewed are six computer software packages including "Lunar Greenhouse,""Dyno-Quest,""How Weather Works,""Animal Trackers,""Personal Science Laboratory," and "The Skeletal and Muscular Systems." Availability, functional, and hardware requirements are discussed. (CW)

  10. Software Reviews.

    ERIC Educational Resources Information Center

    McGrath, Diane

    1990-01-01

    Reviews two programs: (1) "The Weather Machine" on understanding weather and weather forecasting and (2) "The Mystery of the Hotel Victoria" on problem solving in mathematics. Presents the descriptions, advantages, and weaknesses of the software. (YP)

  11. Software Reviews.

    ERIC Educational Resources Information Center

    Davis, Shelly J., Ed.; Knaupp, Jon, Ed.

    1984-01-01

    Reviewed is computer software on: (1) classification of living things, a tutorial program for grades 5-10; and (2) polynomial practice using tiles, a drill-and-practice program for algebra students. (MNS)

  12. Software Reviews.

    ERIC Educational Resources Information Center

    Wulfson, Stephen, Ed.

    1987-01-01

    Provides a review of four science software programs. Includes topics such as plate tectonics, laboratory experiment simulations, the human body, and light and temperature. Contains information on ordering and reviewers' comments. (ML)

  13. Software Reviews.

    ERIC Educational Resources Information Center

    Dwyer, Donna; And Others

    1989-01-01

    Reviewed are seven software packages for Apple and IBM computers. Included are: "Toxicology"; "Science Corner: Space Probe"; "Alcohol and Pregnancy"; "Science Tool Kit Plus"; Computer Investigations: Plant Growth"; "Climatrolls"; and "Animal Watch: Whales." (CW)

  14. Software Reviews.

    ERIC Educational Resources Information Center

    Kimball, Jeffrey P.; And Others

    1987-01-01

    Describes a variety of computer software. The packages reviewed include a variety of simulations, a spread sheet, a printer driver and an alternative operating system for DBM.PCs and compatible programs. (BSR)

  15. Software Reviews.

    ERIC Educational Resources Information Center

    Wulfson, Eugene T., Ed.

    1988-01-01

    Presents reviews by classroom teachers of software for teaching science. Includes material on the work of geologists, genetics, earth science, classification of living things, astronomy, endangered species, skeleton, drugs, and heartbeat. Provides information on availability and equipment needed. (RT)

  16. Software Reviews.

    ERIC Educational Resources Information Center

    Classroom Computer Learning, 1990

    1990-01-01

    Reviewed are three computer software packages including "Martin Luther King, Jr.: Instant Replay of History,""Weeds to Trees," and "The New Print Shop, School Edition." Discussed are hardware requirements, costs, grade levels, availability, emphasis, strengths, and weaknesses. (CW)

  17. Software Reviews.

    ERIC Educational Resources Information Center

    Mathematics and Computer Education, 1988

    1988-01-01

    Presents reviews of six software packages. Includes (1) "Plain Vanilla Statistics"; (2) "MathCAD 2.0"; (3) "GrFx"; (4) "Trigonometry"; (5) "Algebra II"; (6) "Algebra Drill and Practice I, II, and III." (PK)

  18. Reflight certification software design specifications

    NASA Technical Reports Server (NTRS)

    1984-01-01

    The PDSS/IMC Software Design Specification for the Payload Development Support System (PDSS)/Image Motion Compensator (IMC) is contained. The PDSS/IMC is to be used for checkout and verification of the IMC flight hardware and software by NASA/MSFC.

  19. The LUCIFER control software

    NASA Astrophysics Data System (ADS)

    Jütte, Marcus; Knierim, Volker; Polsterer, Kai; Lehmitz, Michael; Storz, Clemens; Seifert, Walter; Ageorges, Nancy

    2010-07-01

    The successful roll-out of the control software for a complex NIR imager/spectrograph with MOS calls for flexible development strategies due to changing requirements during different phases of the project. A waterfall strategy used in the beginning has to change to a more iterative and agile process in the later stages. The choice of an appropriate program language as well as suitable software layout is crucial. For example the software has to accomplish multiple demands of different user groups, including a high level of flexibility for later changes and extensions. Different access levels to the instrument are mandatory to afford direct control mechanisms for lab operations and inspections of the instrument as well as tools to accomplish efficient science observations. Our hierarchical software structure with four layers of increasing abstract levels and the use of an object oriented language ideally supports these requirements. Here we describe our software architecture, the software development process, the different access levels and our commissioning experiences with LUCIFER 1.

  20. Software Based Supernova Recognition

    NASA Astrophysics Data System (ADS)

    Walters, Stephen M.

    2014-05-01

    This paper describes software for detecting Supernova (SN) in images. The software can operate in real-time to discover SN while data is being collected so the instrumentation can immediately be re-tasked to perform spectroscopy or photometry of a discovery. Because the instrumentation captures two images per minute, the realtime budget is constrained to 30 seconds per target, a challenging goal. Using a set of two to four images, the program creates a "Reference" (REF) image and a "New" (NEW) image where all images are used in both NEW and REF but any SN survives the combination process only in the NEW image. This process produces good quality images having similar noise characteristics but without artifacts that might be interpreted as SN. The images are then adjusted for seeing and brightness differences using a variant of Tomaney and Crotts method of Point Spread Function (PSF) matching after which REF is subtracted from NEW to produce a Difference (DIF) image. A Classifier is then trained on a grid of artificial SN to estimate the statistical properties of four attributes and used in a process to mask false positives that can be clearly identified as such. Further training to avoid any remaining false positives sets the range, in standard deviations for each attribute, that the Classifier will accept as a valid SN. This training enables the Classifier to discriminate between SN and most subtraction residue. Lastly, the DIF image is scanned and measured by the Classifier to find locations where all four properties fall within their acceptance ranges. If multiple locations are found, the one best conforming to the training estimates is chosen. This location is then declared as a Candidate SN, the instrumentation re-tasked and the operator notified.

  1. FMT (Flight Software Memory Tracker) For Cassini Spacecraft-Software Engineering Using JAVA

    NASA Technical Reports Server (NTRS)

    Kan, Edwin P.; Uffelman, Hal; Wax, Allan H.

    1997-01-01

    The software engineering design of the Flight Software Memory Tracker (FMT) Tool is discussed in this paper. FMT is a ground analysis software set, consisting of utilities and procedures, designed to track the flight software, i.e., images of memory load and updatable parameters of the computers on-board Cassini spacecraft. FMT is implemented in Java.

  2. Autonomous robot software development using simple software components

    NASA Astrophysics Data System (ADS)

    Burke, Thomas M.; Chung, Chan-Jin

    2004-10-01

    Developing software to control a sophisticated lane-following, obstacle-avoiding, autonomous robot can be demanding and beyond the capabilities of novice programmers - but it doesn"t have to be. A creative software design utilizing only basic image processing and a little algebra, has been employed to control the LTU-AISSIG autonomous robot - a contestant in the 2004 Intelligent Ground Vehicle Competition (IGVC). This paper presents a software design equivalent to that used during the IGVC, but with much of the complexity removed. The result is an autonomous robot software design, that is robust, reliable, and can be implemented by programmers with a limited understanding of image processing. This design provides a solid basis for further work in autonomous robot software, as well as an interesting and achievable robotics project for students.

  3. Software engineering

    NASA Technical Reports Server (NTRS)

    Fridge, Ernest M., III; Hiott, Jim; Golej, Jim; Plumb, Allan

    1993-01-01

    Today's software systems generally use obsolete technology, are not integrated properly with other software systems, and are difficult and costly to maintain. The discipline of reverse engineering is becoming prominent as organizations try to move their systems up to more modern and maintainable technology in a cost effective manner. The Johnson Space Center (JSC) created a significant set of tools to develop and maintain FORTRAN and C code during development of the space shuttle. This tool set forms the basis for an integrated environment to reengineer existing code into modern software engineering structures which are then easier and less costly to maintain and which allow a fairly straightforward translation into other target languages. The environment will support these structures and practices even in areas where the language definition and compilers do not enforce good software engineering. The knowledge and data captured using the reverse engineering tools is passed to standard forward engineering tools to redesign or perform major upgrades to software systems in a much more cost effective manner than using older technologies. The latest release of the environment was in Feb. 1992.

  4. Software reengineering

    NASA Technical Reports Server (NTRS)

    Fridge, Ernest M., III

    1991-01-01

    Today's software systems generally use obsolete technology, are not integrated properly with other software systems, and are difficult and costly to maintain. The discipline of reverse engineering is becoming prominent as organizations try to move their systems up to more modern and maintainable technology in a cost effective manner. JSC created a significant set of tools to develop and maintain FORTRAN and C code during development of the Space Shuttle. This tool set forms the basis for an integrated environment to re-engineer existing code into modern software engineering structures which are then easier and less costly to maintain and which allow a fairly straightforward translation into other target languages. The environment will support these structures and practices even in areas where the language definition and compilers do not enforce good software engineering. The knowledge and data captured using the reverse engineering tools is passed to standard forward engineering tools to redesign or perform major upgrades to software systems in a much more cost effective manner than using older technologies. A beta vision of the environment was released in Mar. 1991. The commercial potential for such re-engineering tools is very great. CASE TRENDS magazine reported it to be the primary concern of over four hundred of the top MIS executives.

  5. Integrating NASA's Land Analysis System (LAS) image processing software with an appropriate Geographic Information System (GIS): A review of candidates in the public domain

    NASA Technical Reports Server (NTRS)

    Rochon, Gilbert L.

    1989-01-01

    A user requirements analysis (URA) was undertaken to determine and appropriate public domain Geographic Information System (GIS) software package for potential integration with NASA's LAS (Land Analysis System) 5.0 image processing system. The necessity for a public domain system was underscored due to the perceived need for source code access and flexibility in tailoring the GIS system to the needs of a heterogenous group of end-users, and to specific constraints imposed by LAS and its user interface, Transportable Applications Executive (TAE). Subsequently, a review was conducted of a variety of public domain GIS candidates, including GRASS 3.0, MOSS, IEMIS, and two university-based packages, IDRISI and KBGIS. The review method was a modified version of the GIS evaluation process, development by the Federal Interagency Coordinating Committee on Digital Cartography. One IEMIS-derivative product, the ALBE (AirLand Battlefield Environment) GIS, emerged as the most promising candidate for integration with LAS. IEMIS (Integrated Emergency Management Information System) was developed by the Federal Emergency Management Agency (FEMA). ALBE GIS is currently under development at the Pacific Northwest Laboratory under contract with the U.S. Army Corps of Engineers' Engineering Topographic Laboratory (ETL). Accordingly, recommendations are offered with respect to a potential LAS/ALBE GIS linkage and with respect to further system enhancements, including coordination with the development of the Spatial Analysis and Modeling System (SAMS) GIS in Goddard's IDM (Intelligent Data Management) developments in Goddard's National Space Science Data Center.

  6. Antiterrorist Software

    NASA Technical Reports Server (NTRS)

    Clark, David A.

    1998-01-01

    In light of the escalation of terrorism, the Department of Defense spearheaded the development of new antiterrorist software for all Government agencies by issuing a Broad Agency Announcement to solicit proposals. This Government-wide competition resulted in a team that includes NASA Lewis Research Center's Computer Services Division, who will develop the graphical user interface (GUI) and test it in their usability lab. The team launched a program entitled Joint Sphere of Security (JSOS), crafted a design architecture (see the following figure), and is testing the interface. This software system has a state-ofthe- art, object-oriented architecture, with a main kernel composed of the Dynamic Information Architecture System (DIAS) developed by Argonne National Laboratory. DIAS will be used as the software "breadboard" for assembling the components of explosions, such as blast and collapse simulations.

  7. Software Reviews.

    ERIC Educational Resources Information Center

    Classroom Computer Learning, 1990

    1990-01-01

    Reviewed are two computer software packages: "Super Solvers Midnight Rescue!" a problem-solving program for IBM PCs; and "Interactive Physics," a simulation program for the Macintosh computer. The functions of the package are discussed including strengths and weaknesses and teaching suggestions. (CW)

  8. Reviews, Software.

    ERIC Educational Resources Information Center

    Science Teacher, 1988

    1988-01-01

    Reviews two software programs for Apple series computers. Includes "Orbital Mech," a basic planetary orbital simulation for the Macintosh, and "START: Stimulus and Response Tools for Experiments in Memory, Learning, Cognition, and Perception," a program that demonstrates basic psychological principles and experiments. (CW)

  9. Software Reviews.

    ERIC Educational Resources Information Center

    Wulfson, Stephen, Ed.

    1989-01-01

    Six software packages are described in this review. Included are "Molecules and Atoms: Exploring the Essence of Matter"; "Heart Probe"; "GM Sunraycer"; "Six Puzzles"; "Information Laboratory--Life Science"; and "Science Test Builder." Hardware requirements, prices, and a summary of the abilities of each program are presented. (CW)

  10. Educational Software.

    ERIC Educational Resources Information Center

    Northwest Regional Educational Lab., Portland, OR.

    The third session of IT@EDU98 consisted of five papers on educational software and was chaired by Tran Van Hao (University of Education, Ho Chi Minh City, Vietnam). "Courseware Engineering" (Nguyen Thanh Son, Ngo Ngoc Bao Tran, Quan Thanh Tho, Nguyen Hong Lam) briefly describes the use of courseware. "Machine Discovery Theorems in Geometry: A…

  11. Software Reviews.

    ERIC Educational Resources Information Center

    Computing Teacher, 1985

    1985-01-01

    Reprinted from "The Computing Teacher," this document contains software reviews for 23 computer programs that educators could use in the classroom or for administrative purposes. Each review describes the program by listing the program title, subject, producer, grade level (if applicable), hardware required, cost, and reviewer's name and…

  12. Software Reviews.

    ERIC Educational Resources Information Center

    Bitter, Gary G., Ed.

    1989-01-01

    Reviews three software packages: (1) "The Weather Machine Courseware Kit" for grades 7-12; (2) "Exploring Measurement, Time, and Money--Level I," for primary level mathematics; and (3) "Professor DOS with SmartGuide for DOS" providing an extensive tutorial covering DOS 2.1 to 4.0. Discusses the strengths and weaknesses of each package. (YP)

  13. Software Reviews.

    ERIC Educational Resources Information Center

    Smith, Richard L., Ed.

    1987-01-01

    Reviewed are three computer software programs: the Astronomer (astronomy program for middle school students and older); Hands-on-Statistics: Explorations with a Microcomputer (statistics program for secondary school students and older); and CATGEN (a genetics program for secondary school students and older). Each review provides information on:…

  14. Software Reviews.

    ERIC Educational Resources Information Center

    Science and Children, 1990

    1990-01-01

    Reviewed are seven computer software packages for IBM and/or Apple Computers. Included are "Windows on Science: Volume 1--Physical Science"; "Science Probe--Physical Science"; "Wildlife Adventures--Grizzly Bears"; "Science Skills--Development Programs"; "The Clean Machine"; "Rock Doctor"; and "Geology Search." Cost, quality, hardware, and…

  15. Software Review.

    ERIC Educational Resources Information Center

    McGrath, Diane, Ed.

    1989-01-01

    Reviewed is a computer software package entitled "Audubon Wildlife Adventures: Grizzly Bears" for Apple II and IBM microcomputers. Included are availability, hardware requirements, cost, and a description of the program. The murder-mystery flavor of the program is stressed in this program that focuses on illegal hunting and game management. (CW)

  16. Software Reviews.

    ERIC Educational Resources Information Center

    Wulfson, Stephen, Ed.

    1989-01-01

    Presents comments by classroom teachers on software for science teaching including topics on: the size of a molecule, matter, leaves, vitamins and minerals, dinosaurs, and collecting and measuring data. Each is an Apple computer series. Availability and costs are included. (RT)

  17. Software Comparison

    NASA Technical Reports Server (NTRS)

    Blanchard, D. C.

    1986-01-01

    Software Comparison Package (SCP) compares similar files. Normally, these are 90-character files produced by CDC UPDATE utility from program libraries that contain FORTRAN source code plus identifier. SCP also used to compare load maps, cross-reference outputs, and UPDATE corrections sets. Helps wherever line-by-line comparison of similarly structured files required.

  18. Software Patents.

    ERIC Educational Resources Information Center

    Burke, Edmund B.

    1994-01-01

    Outlines basic patent law information that pertains to computer software programs. Topics addressed include protection in other countries; how to obtain patents; kinds of patents; duration; classes of patentable subject matter, including machines and processes; patentability searches; experimental use prior to obtaining a patent; and patent…

  19. Software reengineering

    NASA Technical Reports Server (NTRS)

    Fridge, Ernest M., III

    1991-01-01

    Programs in use today generally have all of the function and information processing capabilities required to do their specified job. However, older programs usually use obsolete technology, are not integrated properly with other programs, and are difficult to maintain. Reengineering is becoming a prominent discipline as organizations try to move their systems to more modern and maintainable technologies. The Johnson Space Center (JSC) Software Technology Branch (STB) is researching and developing a system to support reengineering older FORTRAN programs into more maintainable forms that can also be more readily translated to a modern languages such as FORTRAN 8x, Ada, or C. This activity has led to the development of maintenance strategies for design recovery and reengineering. These strategies include a set of standards, methodologies, and the concepts for a software environment to support design recovery and reengineering. A brief description of the problem being addressed and the approach that is being taken by the STB toward providing an economic solution to the problem is provided. A statement of the maintenance problems, the benefits and drawbacks of three alternative solutions, and a brief history of the STB experience in software reengineering are followed by the STB new FORTRAN standards, methodology, and the concepts for a software environment.

  20. Software Reviews.

    ERIC Educational Resources Information Center

    Smith, Richard L., Ed.

    1988-01-01

    Reviews two software packages, "Solutions Unlimited" and "BASIC Data Base System." Provides a description, summary, strengths and weaknesses, availability and costs. Includes reviews of three structured BASIC packages: "True BASIC (2.0)"; "Turbo BASIC (1.0)"; and "QuickBASIC (3.0)." Explains significant features such as graphics, costs,…

  1. Design Software

    NASA Technical Reports Server (NTRS)

    1991-01-01

    A NASA contractor and Small Business Innovation Research (SBIR) participant has converted its research into commercial software products for auto design, structural analysis and other applications. ViGYAN, Inc., utilizing the aeronautical research principle of computational fluid dynamics, has created - with VGRID3D and VPLOT3D - an easier alternative to conventional structured grids for fluid dynamic calculations.

  2. Software Reviews.

    ERIC Educational Resources Information Center

    Science and Children, 1988

    1988-01-01

    Reviews six software packages for use with school age children ranging from grade 3 to grade 12. Includes "The Microcomputer Based Lab Project: Motion, Sound"; "Genetics"; "Geologic History"; "The Microscope Simulator"; and "Wiz Works" all for Apple II and "Reading for Information: Level II" for IBM. (CW)

  3. Software Reviews.

    ERIC Educational Resources Information Center

    Science and Children, 1989

    1989-01-01

    Reviews of seven software packages are presented including "The Environment I: Habitats and EcoSystems; II Cycles and Interactions"; "Super Sign Maker"; "The Great Knowledge Race: Substance Abuse"; "Exploring Science: Temperature"; "Fast Food Calculator and RD Aide"; "The Human Body: Circulation and Respiration" and "Forces in Liquids and Gases."…

  4. Software Reviews.

    ERIC Educational Resources Information Center

    Wulfson, Stephen, Ed.

    1990-01-01

    Reviewed are seven computer software packages including "Frog Dissection Lab Report,""Backyard Birds,""LEGO TC Logo,""Alcohol--Four Interactive Programs,""Windows on Science--Life Science,""Climate and Weather/Our Town Database," and "Weeds to Trees." Discussed are availability, features, strengths, and weaknesses. (CW)

  5. Star Software.

    ERIC Educational Resources Information Center

    Kloza, Brad

    2000-01-01

    Presents a collection of computer software programs designed to spark learning enthusiasm at every grade level and across the curriculum. They include Reader Rabbit's Learn to Read, Spelling Power, Mind Twister Math, Community Construction Kit, Breaking the Code, Encarta Africana 2000, Virtual Serengeti, Operation: Frog (Deluxe), and My First…

  6. Software Reviews.

    ERIC Educational Resources Information Center

    McGrath, Diane, Ed.

    1989-01-01

    Reviewed are two computer software programs for Apple II computers on weather for upper elementary and middle school grades. "Weather" introduces the major factors (temperature, humidity, wind, and air pressure) affecting weather. "How Weather Works" uses simulation and auto-tutorial formats on sun, wind, fronts, clouds, and storms. (YP)

  7. Statistical Software.

    ERIC Educational Resources Information Center

    Callamaras, Peter

    1983-01-01

    This buyer's guide to seven major types of statistics software packages for microcomputers reviews Edu-Ware Statistics 3.0; Financial Planning; Speed Stat; Statistics with DAISY; Human Systems Dynamics package of Stats Plus, ANOVA II, and REGRESS II; Maxistat; and Moore-Barnes' MBC Test Construction and MBC Correlation. (MBR)

  8. Software Reviews.

    ERIC Educational Resources Information Center

    Teles, Elizabeth, Ed.; And Others

    1990-01-01

    Reviewed are two computer software packages for Macintosh microcomputers including "Phase Portraits," an exploratory graphics tool for studying first-order planar systems; and "MacMath," a set of programs for exploring differential equations, linear algebra, and other mathematical topics. Features, ease of use, cost, availability, and hardware…

  9. Reviews: Software.

    ERIC Educational Resources Information Center

    Mackenzie, Norma N.; And Others

    1988-01-01

    Reviews four computer software packages including: "The Physical Science Series: Sound" which demonstrates making waves, speed of sound, doppler effect, and human hearing; "Andromeda" depicting celestial motions in any direction; "Biology Quiz: Humans" covering chemistry, cells, viruses, and human biology; and "MacStronomy" covering information on…

  10. Software Reviews.

    ERIC Educational Resources Information Center

    Mackenzie, Norma N.; And Others

    1988-01-01

    Describes computer software for use with various age groups. Topics include activities involving temperature, simulations, earth science, the circulatory system, human body, reading in science, and ecology. Provides information on equipment needed, availability, package contents, and price. Comments of reviews are presented by classroom teachers.…

  11. Going to where the users are! Making the collaborative resource management and science workspace mobile

    NASA Astrophysics Data System (ADS)

    Osti, D.; Osti, A.

    2013-12-01

    People are very busy today and getting stakeholders the information they need is an important part of our jobs. The BDL application is the mobile extension of the California collaborative resource management portal www.baydeltalive.com. BDL has been visited by more than 250,000 unique visitors this past year from various areas of water use and management including state and federal agencies, agriculture, scientists, policy makers, water consumers, voters, operations management and more. The audience is a qualified user group of more than 15,000 individuals participating in California hydrological ecosystem science, water management and policy. This is an important effort aimed to improve how scientists and policy makers are working together to understand this complicated and divisive system and how they are becoming better managers of that system. The BayDetaLive mobile application gives California watershed management stakeholders and water user community unprecedented access to real time natural resource management information. The application provides user with the following: 1. Access to Real Time Environmental Conditions from the more than the 600 California Data Exchange Sensors including hydrodynamic, water quality and meteorological data. Save important stations as favorites for easy access later. 2. Daily Delta Operations Data including estimated hydrology, daily exports, status of infrastructure operations, reservoir storage, salvage data, major stations, drinking water quality reports, weather forecasts and more. 3. Photos/Videos/Documents: Browse and share from the more than 1000 current documents in the BDL library. Relevant images, videos, science journals, presentations and articles. 4. Science: Access the latest science articles, news, projects and journals. 5. Data Visualizations: View recently published real time data interpolations of Delta Conditions. From 30-day turbidity models to daily forecasts. This service is published as conditions

  12. Space Station Software Issues

    NASA Technical Reports Server (NTRS)

    Voigt, S. (Editor); Beskenis, S. (Editor)

    1985-01-01

    Issues in the development of software for the Space Station are discussed. Software acquisition and management, software development environment, standards, information system support for software developers, and a future software advisory board are addressed.

  13. Scientific Software

    NASA Technical Reports Server (NTRS)

    1995-01-01

    The Interactive Data Language (IDL), developed by Research Systems, Inc., is a tool for scientists to investigate their data without having to write a custom program for each study. IDL is based on the Mariners Mars spectral Editor (MMED) developed for studies from NASA's Mars spacecraft flights. The company has also developed Environment for Visualizing Images (ENVI), an image processing system for easily analyzing remotely sensed data written in IDL. The Visible Human CD, another Research Systems product, is the first complete digital reference of photographic images for exploring human anatomy.

  14. Software Engineering Support of the Third Round of Scientific Grand Challenge Investigations: Earth System Modeling Software Framework Survey

    NASA Technical Reports Server (NTRS)

    Talbot, Bryan; Zhou, Shu-Jia; Higgins, Glenn; Zukor, Dorothy (Technical Monitor)

    2002-01-01

    One of the most significant challenges in large-scale climate modeling, as well as in high-performance computing in other scientific fields, is that of effectively integrating many software models from multiple contributors. A software framework facilitates the integration task, both in the development and runtime stages of the simulation. Effective software frameworks reduce the programming burden for the investigators, freeing them to focus more on the science and less on the parallel communication implementation. while maintaining high performance across numerous supercomputer and workstation architectures. This document surveys numerous software frameworks for potential use in Earth science modeling. Several frameworks are evaluated in depth, including Parallel Object-Oriented Methods and Applications (POOMA), Cactus (from (he relativistic physics community), Overture, Goddard Earth Modeling System (GEMS), the National Center for Atmospheric Research Flux Coupler, and UCLA/UCB Distributed Data Broker (DDB). Frameworks evaluated in less detail include ROOT, Parallel Application Workspace (PAWS), and Advanced Large-Scale Integrated Computational Environment (ALICE). A host of other frameworks and related tools are referenced in this context. The frameworks are evaluated individually and also compared with each other.

  15. Analysis Software

    NASA Technical Reports Server (NTRS)

    1994-01-01

    General Purpose Boundary Element Solution Technology (GPBEST) software employs the boundary element method of mechanical engineering analysis, as opposed to finite element. It is, according to one of its developers, 10 times faster in data preparation and more accurate than other methods. Its use results in less expensive products because the time between design and manufacturing is shortened. A commercial derivative of a NASA-developed computer code, it is marketed by Best Corporation to solve problems in stress analysis, heat transfer, fluid analysis and yielding and cracking of solids. Other applications include designing tractor and auto parts, household appliances and acoustic analysis.

  16. Seminar Software

    NASA Technical Reports Server (NTRS)

    1993-01-01

    The Society for Computer Simulation International is a professional technical society that distributes information on methodology techniques and uses of computer simulation. The society uses NETS, a NASA-developed program, to assist seminar participants in learning to use neural networks for computer simulation. NETS is a software system modeled after the human brain; it is designed to help scientists exploring artificial intelligence to solve pattern matching problems. Examples from NETS are presented to seminar participants, who can then manipulate, alter or enhance them for their own applications.

  17. Simulation Software

    NASA Technical Reports Server (NTRS)

    1996-01-01

    Various NASA Small Business Innovation Research grants from Marshall Space Flight Center, Langley Research Center and Ames Research Center were used to develop the 'kernel' of COMCO's modeling and simulation software, the PHLEX finite element code. NASA needed it to model designs of flight vehicles; one of many customized commercial applications is UNISIM, a PHLEX-based code for analyzing underground flows in oil reservoirs for Texaco, Inc. COMCO's products simulate a computational mechanics problem, estimate the solution's error and produce the optimal hp-adapted mesh for the accuracy the user chooses. The system is also used as a research or training tool in universities and in mechanical design in industrial corporations.

  18. Software Surrogate

    NASA Technical Reports Server (NTRS)

    1999-01-01

    In 1994, Blackboard Technology received a NASA Phase I SBIR award entitled "A Blackboard-Based Framework for Mixed-Initiative, Crewed- Space-System Applications." This research continued in Phase II at JSC, where a generic architecture was developed in which a software surrogate serves as the operator's representative in the fast-paced realm of nearly autonomous, intelligent systems. This SBIR research effort addressed the need to support human-operator monitoring and intervention with intelligent systems such as those being developed for NASA's crewed space program.

  19. Software system safety

    NASA Technical Reports Server (NTRS)

    Uber, James G.

    1988-01-01

    Software itself is not hazardous, but since software and hardware share common interfaces there is an opportunity for software to create hazards. Further, these software systems are complex, and proven methods for the design, analysis, and measurement of software safety are not yet available. Some past software failures, future NASA software trends, software engineering methods, and tools and techniques for various software safety analyses are reviewed. Recommendations to NASA are made based on this review.

  20. Differences in granular materials for analogue modelling: Insights from repeated compression tests analyzed with X-ray Computed Tomography and image analysis software

    NASA Astrophysics Data System (ADS)

    Klinkmueller, M.; Schreurs, G.

    2009-12-01

    Six different granular materials for analogue modelling have been investigated using a sandbox with a compressional set-up and X-ray computed tomography (XRCT). The evolving structures were evaluated with image analysis software. The sandbox has one movable sidewall that is driven by a computer-controlled servomotor at 20 cm/h. A 12 cm wide and 20 cm long sheet of hard cardboard was placed on the base of the sandbox and attached to the moving sidewall creating a velocity discontinuity. The whole sandbox was covered on the inside with Alkor foil to reduce sidewall friction. Computed Tomography was used to scan the whole volume in 3 mm increments of shortening until 15 mm maximum deformation was reached. The second approach was a scanning procedure to a maximum deformation of 80 mm in 2 mm increments of shortening for the first 10 mm and in 5 mm increments for the last 70 mm. The short deformation scans were repeated three times to investigate reproducibility. The long deformation scans were performed twice. The physical properties of the materials (table 1) have been described in a previous material benchmark. Four natural quartz sands and two artificial granular materials, corundum brown sand and glass beads, have been used. The two artificial materials were used for this experimental series as examples for very angular and very rounded sands in contrast to the sub-rounded to angular natural quartz sands. The short deformation experiments show partly large differences in thrust angles of both front and back-thrust, in timing of thrust initiation, and in the degree of undulation of thrusts. The coarse-grained sands show smooth and low undulating thrusts that are only affected by the sidewall friction whereas the thrusts in fine-grained sands undulate significantly and partly divide and merge in an anastomosing fashion. The coarse-grained sand thrusts are clearer visualized by XRCT, which indicates a wider shear zone where the material dilates. Furthermore, the

  1. The Analysis of the Patterns of Radiation-Induced DNA Damage Foci by a Stochastic Monte Carlo Model of DNA Double Strand Breaks Induction by Heavy Ions and Image Segmentation Software

    NASA Technical Reports Server (NTRS)

    Ponomarev, Artem; Cucinotta, F.

    2011-01-01

    To create a generalized mechanistic model of DNA damage in human cells that will generate analytical and image data corresponding to experimentally observed DNA damage foci and will help to improve the experimental foci yields by simulating spatial foci patterns and resolving problems with quantitative image analysis. Material and Methods: The analysis of patterns of RIFs (radiation-induced foci) produced by low- and high-LET (linear energy transfer) radiation was conducted by using a Monte Carlo model that combines the heavy ion track structure with characteristics of the human genome on the level of chromosomes. The foci patterns were also simulated in the maximum projection plane for flat nuclei. Some data analysis was done with the help of image segmentation software that identifies individual classes of RIFs and colocolized RIFs, which is of importance to some experimental assays that assign DNA damage a dual phosphorescent signal. Results: The model predicts the spatial and genomic distributions of DNA DSBs (double strand breaks) and associated RIFs in a human cell nucleus for a particular dose of either low- or high-LET radiation. We used the model to do analyses for different irradiation scenarios. In the beam-parallel-to-the-disk-of-a-flattened-nucleus scenario we found that the foci appeared to be merged due to their high density, while, in the perpendicular-beam scenario, the foci appeared as one bright spot per hit. The statistics and spatial distribution of regions of densely arranged foci, termed DNA foci chains, were predicted numerically using this model. Another analysis was done to evaluate the number of ion hits per nucleus, which were visible from streaks of closely located foci. In another analysis, our image segmentaiton software determined foci yields directly from images with single-class or colocolized foci. Conclusions: We showed that DSB clustering needs to be taken into account to determine the true DNA damage foci yield, which helps to

  2. Lack of exposure to natural light in the workspace is associated with physiological, sleep and depressive symptoms.

    PubMed

    Harb, Francine; Hidalgo, Maria Paz; Martau, Betina

    2015-04-01

    The diurnal light cycle has a crucial influence on all life on earth. Unfortunately, modern society has modified this life-governing cycle by stressing maximum production and by giving insufficient attention to the ecological balance and homeostasis of the human metabolism. The aim of this study is to evaluate the effects of exposure or lack of exposure to natural light in a rest/activity rhythm on cortisol and melatonin levels, as well as on psychological variables in humans under natural conditions. This is a cross-sectional study. The subjects were allocated split into two groups according to their workspace (10 employees in the "with window" group and 10 in the "without window" group). All participants were women and wore anactigraph (Actiwatch 2, Philips Respironics), which measures activity and ambient light exposure, for seven days. Concentrations of melatonin and cortisol were measured from the saliva samples. Participants were instructed to collect saliva during the last day of use of the actigraph at 08:00 am, 4:00 pm and 10:00 pm. The subjects answered the Self-Reporting Questionnaire-20 (SRQ-20) to measure the presence of minor psychiatric disorders; the Montgomery-Asberg (MA) scale was used to measure depression symptoms, and the Pittsburgh Sleep Quality Index questionnaire (PSQI) was used to evaluate the quality of sleep. The Rayleigh analysis indicates that the two groups, "with window" an d "without window", exhibited similar activities and light acrophases. In relation to light exposure, the mesor was significantly higher (t = -2.651, p = 0.023) in t he "with window" group (191.04 ± 133.36) than in the "without window" group (73.8 ± 42.05). Additionally, the "with window" group presented the highest amplitude of light exposure (298.07 ± 222.97). Cortisol levels were significantly different between the groups at 10:00 pm (t = 3.009, p = 0.008; "without window" (4.01 ± 0.91) "with window" (3.10 ± 0.30)). In

  3. Sandia software guidelines: Software quality planning

    SciTech Connect

    Not Available

    1987-08-01

    This volume is one in a series of Sandia Software Guidelines intended for use in producing quality software within Sandia National Laboratories. In consonance with the IEEE Standard for Software Quality Assurance Plans, this volume identifies procedures to follow in producing a Software Quality Assurance Plan for an organization or a project, and provides an example project SQA plan. 2 figs., 4 tabs.

  4. Software for surface analysis

    NASA Astrophysics Data System (ADS)

    Watson, D. G.; Doern, F. E.

    1985-04-01

    Two software packages designed to aid in the analysis of digitally stored Secondary Ion Mass Spectrometric (SIMS) and electron spectroscopic data are described. The first, MASS, is a program that normalizes, and allows the application of sensitivity coefficients to SIMS depth profiles. The second, DIP, is a digital image processor designed to enhance secondary, backscattered, and Auger electron spectroscopic (AES) maps. DIP can also provide quantitative area analysis of AES maps. The algorithms are currently optimized to handle data generated by Physical Electronics Industries data acquisition systems, but are generally applicable.

  5. Office Computer Software: A Comprehensive Review of Software Programs.

    ERIC Educational Resources Information Center

    Secretary, 1992

    1992-01-01

    Describes types of software including system software, application software, spreadsheets, accounting software, graphics packages, desktop publishing software, database, desktop and personal information management software, project and records management software, groupware, and shareware. (JOW)

  6. Should software hold data hostage?

    SciTech Connect

    Wiley, H S.; Michaels, George S.

    2004-08-01

    development of facile user interfaces and robust environments. This is where some companies have provided real value to the community, building on the foundation of open source software. Outside of genomics and bioinformatics, there is still a critical need for software tools, particularly in areas such as imaging, biochemistry and cell signaling. The computer skills of investigators in these fields is generally more rudimentary, and thus the open source options are much more limited. Commercial software dominates these areas, but open source has the potential to contribute more in the future.

  7. Software Archive Related Issues

    NASA Technical Reports Server (NTRS)

    Angelini, Lorella

    2008-01-01

    With the archive opening of the major X-ray and Gamma ray missions, the school is intended to provide information on the resource available in the data archive and the public software. This talk reviews the archive content, the data format for the major active missions Chandra, XMM-Newton, Swift, RXTE, Integral and Suzaku and the available software for each of these missions. It will explain the FITS format in general and the specific layout for the most popular mission, explaining the role of keywords and how they fit in the multimission standard approach embrace by the High Energy Community. Specifically, it reviews : the difference data levels and the difference software applicable; the popular/standard method of analysis for high level products such as spectra, timing and images; the role of calibration in the multi mission approach; how to navigate the archive query databases. It will present also how the school is organized and how the information provided will be relevant to each of the afternoon science projects that will be proposed to the students and led by a project leader

  8. The LSST Software Stack

    NASA Astrophysics Data System (ADS)

    Jenness, Timothy; LSST Data Management Team

    2016-01-01

    The Large Synoptic Survey Telescope (LSST) is an 8-m optical ground-based telescope being constructed on Cerro Pachon in Chile. LSST will survey half the sky every few nights in six optical bands. The data will be transferred to the data center in North America and within 60 seconds it will be reduced using difference imaging and an alert list be generated for the community. Additionally, annual data releases will be constructed from all the data during the 10-year mission, producing catalogs and deep co-added images with unprecedented time resolution for such a large region of sky. In the paper we present the current status of the LSST stack including the data processing components, Qserv database and data visualization software, describe how to obtain it, and provide a summary of the development road map.

  9. Software Model Of Software-Development Process

    NASA Technical Reports Server (NTRS)

    Lin, Chi Y.; Synott, Debra J.; Levary, Reuven R.

    1990-01-01

    Collection of computer programs constitutes software tool for simulation of medium- to large-scale software-development projects. Necessary to include easily identifiable and more-readily quantifiable characteristics like costs, times, and numbers of errors. Mathematical model incorporating these and other factors of dynamics of software-development process implemented in the Software Life Cycle Simulator (SLICS) computer program. Simulates dynamics of software-development process. In combination with input and output expert software systems and knowledge-based management software system, develops information for use in managing large software-development project. Intended to aid managers in planning, managing, and controlling software-development processes by reducing uncertainties in budgets, required personnel, and schedules.

  10. PROMOTIONS: PROper MOTION Software

    NASA Astrophysics Data System (ADS)

    Caleb Wherry, John; Sahai, R.

    2009-05-01

    We report on the development of a software tool (PROMOTIONS) to streamline the process of measuring proper motions of material in expanding nebulae. Our tool makes use of IDL's widget programming capabilities to design a unique GUI that is used to compare images of the objects from two epochs. The software allows us to first orient and register the images to a common frame of reference and pixel scale, using field stars in each of the images. We then cross-correlate specific morphological features in order to determine their proper motions, which consist of the proper motion of the nebula as a whole (PM-neb), and expansion motions of the features relative to the center. If the central star is not visible (quite common in bipolar nebulae with dense dusty waists), point-symmetric expansion is assumed and we use the average motion of high-quality symmetric pairs of features on opposite sides of the nebular center to compute PM-neb. This is then subtracted out to determine the individual movements of these and additional features relative to the nebular center. PROMOTIONS should find wide applicability in measuring proper motions in astrophysical objects such as the expanding outflows/jets commonly seen around young and dying stars. We present first results from using PROMOTIONS to successfully measure proper motions in several pre-planetary nebulae (transition objects between the red giant and planetary nebula phases), using images taken 7-10 years apart with the WFPC2 and ACS instruments on board HST. The authors are grateful to NASA's Undergradute Scholars Research Program (USRP) for supporting this research.

  11. MORPH-II, a software package for the analysis of scanning-electron-micrograph images for the assessment of the fractal dimension of exposed stone surfaces

    USGS Publications Warehouse

    Mossotti, Victor G.; Eldeeb, A. Raouf

    2000-01-01

    Turcotte, 1997, and Barton and La Pointe, 1995, have identified many potential uses for the fractal dimension in physicochemical models of surface properties. The image-analysis program described in this report is an extension of the program set MORPH-I (Mossotti and others, 1998), which provided the fractal analysis of electron-microscope images of pore profiles (Mossotti and Eldeeb, 1992). MORPH-II, an integration of the modified kernel of the program MORPH-I with image calibration and editing facilities, was designed to measure the fractal dimension of the exposed surfaces of stone specimens as imaged in cross section in an electron microscope.

  12. Flash Cards and Animation Software for Education.

    ERIC Educational Resources Information Center

    Byers, John A.

    1999-01-01

    Describes how a software program for DOS/Windows manages a collection of pictures such as photographic slides, overheads, or computer images in one or more databases. Explains how it transforms image files to raw binary files that can then be displayed like flash cards, or as an animated series of images. (Author/LRW)

  13. Software attribute visualization for high integrity software

    SciTech Connect

    Pollock, G.M.

    1998-03-01

    This report documents a prototype tool developed to investigate the use of visualization and virtual reality technologies for improving software surety confidence. The tool is utilized within the execution phase of the software life cycle. It provides a capability to monitor an executing program against prespecified requirements constraints provided in a program written in the requirements specification language SAGE. The resulting Software Attribute Visual Analysis Tool (SAVAnT) also provides a technique to assess the completeness of a software specification.

  14. Space Flight Software Development Software for Intelligent System Health Management

    NASA Technical Reports Server (NTRS)

    Trevino, Luis C.; Crumbley, Tim

    2004-01-01

    The slide presentation examines the Marshall Space Flight Center Flight Software Branch, including software development projects, mission critical space flight software development, software technical insight, advanced software development technologies, and continuous improvement in the software development processes and methods.

  15. Analytical approaches to image orientation and stereo digitization applied in the Budnlab software. (Polish Title: Rozwiazania analityczne zwiazane z obsluga procesu orientacji zdjec oraz wykonywaniem opracowan wektorowych w programie Bundlab)

    NASA Astrophysics Data System (ADS)

    Kolecki, J.

    2015-12-01

    The Bundlab software has been developed mainly for academic and research application. This work can be treated as a kind of a report describing the current state of the development of this computer program, focusing especially on the analytical solutions. Firstly, the overall characteristics of the software are provided. Then the description of the image orientation procedure starting from the relative orientation is addressed. The applied solution is based on the coplanarity equation parametrized with the essential matrix. The problem is reformulated in order to solve it using methods of algebraic geometry. The solution is followed by the optimization involving the least square criterion. The formation of the image block from the oriented models as well as the absolute orientation procedure were implemented using the Horn approach as a base algorithm. The second part of the paper is devoted to the tools and methods applied in the stereo digitization module. The solutions that support the user and improve the accuracy are given. Within the paper a few exemplary applications and products are mentioned. The work finishes with the concepts of development and improvements of existing functions.

  16. Software Design for Smile Analysis

    PubMed Central

    Sodagar, A.; Rafatjoo, R.; Gholami Borujeni, D.; Noroozi, H.; Sarkhosh, A.

    2010-01-01

    Introduction: Esthetics and attractiveness of the smile is one of the major demands in contemporary orthodontic treatment. In order to improve a smile design, it is necessary to record “posed smile” as an intentional, non-pressure, static, natural and reproducible smile. The record then should be analyzed to determine its characteristics. In this study, we intended to design and introduce a software to analyze the smile rapidly and precisely in order to produce an attractive smile for the patients. Materials and Methods: For this purpose, a practical study was performed to design multimedia software “Smile Analysis” which can receive patients’ photographs and videographs. After giving records to the software, the operator should mark the points and lines which are displayed on the system’s guide and also define the correct scale for each image. Thirty-three variables are measured by the software and displayed on the report page. Reliability of measurements in both image and video was significantly high (α=0.7–1). Results: In order to evaluate intra- operator and inter-operator reliability, five cases were selected randomly. Statistical analysis showed that calculations performed in smile analysis software were both valid and highly reliable (for both video and photo). Conclusion: The results obtained from smile analysis could be used in diagnosis, treatment planning and evaluation of the treatment progress. PMID:21998792

  17. Report: Scientific Software.

    ERIC Educational Resources Information Center

    Borman, Stuart A.

    1985-01-01

    Discusses various aspects of scientific software, including evaluation and selection of commercial software products; program exchanges, catalogs, and other information sources; major data analysis packages; statistics and chemometrics software; and artificial intelligence. (JN)

  18. Controlling Software Piracy.

    ERIC Educational Resources Information Center

    King, Albert S.

    1992-01-01

    Explains what software manufacturers are doing to combat software piracy, recommends how managers should deal with this problem, and provides a role-playing exercise to help students understand the issues in software piracy. (SR)

  19. Software Engineering Guidebook

    NASA Technical Reports Server (NTRS)

    Connell, John; Wenneson, Greg

    1993-01-01

    The Software Engineering Guidebook describes SEPG (Software Engineering Process Group) supported processes and techniques for engineering quality software in NASA environments. Three process models are supported: structured, object-oriented, and evolutionary rapid-prototyping. The guidebook covers software life-cycles, engineering, assurance, and configuration management. The guidebook is written for managers and engineers who manage, develop, enhance, and/or maintain software under the Computer Software Services Contract.

  20. Revision and product generation software

    USGS Publications Warehouse

    U.S. Geological Survey

    1997-01-01

    The U.S. Geological Survey (USGS) developed revision and product generation (RevPG) software for updating digital line graph (DLG) data and producing maps from such data. This software is based on ARC/INFO, a geographic information system from Environmental Systems Resource Institute (ESRI). RevPG consists of ARC/INFO Arc Macro Language (AML) programs, C routines, and interface menus that permit operators to collect vector data using aerial images, to symbolize the data on-screen, and to produce plots and color-separated files for use in printing maps.